Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
-
Updated
May 20, 2024 - Python
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
Fit interpretable models. Explain blackbox machine learning.
A curated list of awesome responsible machine learning resources.
Model interpretability and understanding for PyTorch
Examples of techniques for training interpretable ML models, explaining ML models, and debugging ML models for accuracy, discrimination, and security.
A collection of research materials on explainable AI/ML
H2O.ai Machine Learning Interpretability Resources
Explainable AI framework for data scientists. Explain & debug any blackbox machine learning model with a single line of code. We are looking for co-authors to take this project forward. Reach out @ [email protected]
Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.
Material related to my book Intuitive Machine Learning. Some of this material is also featured in my new book Synthetic Data and Generative AI.
The code of NeurIPS 2021 paper "Scalable Rule-Based Representation Learning for Interpretable Classification" and TPAMI paper "Learning Interpretable Rules for Scalable Data Representation and Classification"
Package towards building Explainable Forecasting and Nowcasting Models with State-of-the-art Deep Neural Networks and Dynamic Factor Model on Time Series data sets with single line of code. Also, provides utilify facility for time-series signal similarities matching, and removing noise from timeseries signals.
In this part, I've introduced and experimented with ways to interpret and evaluate models in the field of image. (Pytorch)
Modular Python Toolbox for Fairness, Accountability and Transparency Forensics
All about explainable AI, algorithmic fairness and more
PyTorch Explain: Interpretable Deep Learning in Python.
Code for NeurIPS 2019 paper ``Self-Critical Reasoning for Robust Visual Question Answering''
ProtoTorch is a PyTorch-based Python toolbox for bleeding-edge research in prototype-based machine learning algorithms.
XAI based human-in-the-loop framework for automatic rule-learning.
Add a description, image, and links to the interpretable-ai topic page so that developers can more easily learn about it.
To associate your repository with the interpretable-ai topic, visit your repo's landing page and select "manage topics."