{"id":1066,"date":"2024-12-06T11:50:10","date_gmt":"2024-12-06T02:50:10","guid":{"rendered":"https:\/\/www.aicritique.org\/us\/?post_type=explainable&#038;p=1066"},"modified":"2024-12-06T11:50:10","modified_gmt":"2024-12-06T02:50:10","slug":"seldon","status":"publish","type":"explainable","link":"https:\/\/www.aicritique.org\/us\/explainable\/seldon\/","title":{"rendered":"Seldon"},"content":{"rendered":"\n<ul class=\"wp-block-list\">\n<li><strong>Alibi<\/strong>: An open-source library for enhancing the interpretability of machine learning models, supporting trust in AI decisions.<\/li>\n\n\n\n<li><strong>Seldon Core<\/strong>: Uses SHAP and LIME to provide explanations for deployed AI models on Kubernetes.<\/li>\n\n\n\n<li><strong>Seldon Alibi Detect<\/strong>: Offers anomaly detection with explainability features for deployed AI models.<\/li>\n<\/ul>\n","protected":false},"featured_media":0,"template":"","class_list":["post-1066","explainable","type-explainable","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/www.aicritique.org\/us\/wp-json\/wp\/v2\/explainable\/1066","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.aicritique.org\/us\/wp-json\/wp\/v2\/explainable"}],"about":[{"href":"https:\/\/www.aicritique.org\/us\/wp-json\/wp\/v2\/types\/explainable"}],"wp:attachment":[{"href":"https:\/\/www.aicritique.org\/us\/wp-json\/wp\/v2\/media?parent=1066"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}