Skip to content

Causality-Aware Local Interpretable Model Agnostic Explanations

License

Notifications You must be signed in to change notification settings

marti5ini/CALIME

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CALIME

Build Status License: MIT

CALIME is a GitHub repository containing the CALIME algorithm. It refers to the paper titled Causality-Aware Local Interpretable Model-Agnostic Explanations.

Check out the pdf here: [pdf]

Abstract

A significant drawback of eXplainable Artificial Intelligence (XAI) approaches is the assumption of feature independence. This paper focuses on integrating causal knowledge in XAI methods to increase trust and help users assess explanations’ quality. We propose a novel exten- sion to a widely used local and model-agnostic explainer that explicitly encodes causal relationships in the data generated around the input in- stance to explain. Extensive experiments show that our method achieves superior performance comparing the initial one for both the fidelity in mimicking the black-box and the stability of the explanations.

Setup

The packages requires a python version >=3.8, as well as some libraries listed in requirements file. For some additional functionalities, more libraries are needed for these extra functions and options to become available.

git clone https://github.com/marti5ini/CALIME.git
cd CALIME

Dependencies are listed in requirements.txt, a virtual environment is advised:

python3 -m venv ./venv # optional but recommended
pip install -r requirements.txt

Please note that in addition to the dependencies listed in the requirements file, you also need to install a novel version of "fim" package. You can find the package and installation instructions on the following webpage: [https://borgelt.net/pyfim.html]

Citation

If you use CALIME in your research, please cite our paper:

@InProceedings{cinquini2024calime,
author="Cinquini, Martina
and Guidotti, Riccardo",
editor="Longo, Luca
and Lapuschkin, Sebastian
and Seifert, Christin",
title="Causality-Aware Local Interpretable Model-Agnostic Explanations",
booktitle="Explainable Artificial Intelligence",
year="2024",
publisher="Springer Nature Switzerland",
address="Cham",
pages="108--124",
}

About

Causality-Aware Local Interpretable Model Agnostic Explanations

Topics

Resources

License

Stars

Watchers

Forks

Languages