This is a Explainable Artificial Intelligence tool focused in ensemble black box models.
José Ribeiro - site: https://sites.google.com/view/jose-sousa-ribeiro
Lucas Cardoso - site: http://lattes.cnpq.br/9591352011725008
Raíssa Silva - site: https://sites.google.com/site/silvarailors
Vitor Cirilo - site: https://sites.google.com/site/vitorciriloaraujosantos/
Níkolas Carneiro - site: https://br.linkedin.com/in/nikolas-carneiro-62b6568
Ronnie Alves (Leader) - site: https://sites.google.com/site/alvesrco
The measure Explainable based on Item Response Theory - eXirt is one of the XAI measures performed in the developed benchmark. This measure is a new proposal to generate explanations for tree-ensemble models that is based on the generation of attribute relevance ranks by using item response theory.
Just like other XAI measures, eXirt only uses the training data, test data, the model itself together with its outputs, figure 1 (A). Initially, the test data and the model to be explained are passed on to the eXirt; Figure 1 (B) the tool then creates the so-called "loop of models", shown in figure 1 (C) and (D), which is a basic iterative process relevant for the creation of the table with the respondent's answers used in the IRT run, figure 1 (E) and (F).
The creation of the response matrix, figure 1 (E) and (F), contains the answers of all the classifiers (respondents). The columns refer to the instances of the dataset that was passed on, while the rows refer to the different classifiers. Values equal to 0 (zero) are wrong answers of the prediction, while values 1 (one) are correct answers of the prediction, regardless of the number of classes that the problem may have. This matrix is used to calculate the values of the item parameters (discrimination, difficulty and guessing) for each of the instances, in figure 1 (G).
The implementation of the IRT used was Cardoso, L (2020), called decodIRT, in a code developed exclusively for the purpose of this paper, as the code first receives the answer matrix, performs the calculations to generate the item parameter values --- different algorithms can be used to calculate the IRT (such as: ternary, dichotomous, fibonacci, golden, brent, bounded or golden2) using catsim --- and, after this step, generates the rank of most skilled classifiers, figure 1 (G).
**Figure 1 - Process of the eXirt.**Among the new features of decordIRT is a new score calculation that involves the calculated ability of all respondents and their respective hits and misses, called Total Score. Total Score can be understood as an adaptation of the True-Score, whereby the score is calculated by summing up all the hit probabilities for the test items. However, in cases where respondents have a very close ability, the True-Score result can be very similar or even equal, since only the hit chance is considered. To avoid equal score values and to give more robustness to the models' final score, the Total Score also considers the respondent's probability of error, given by:
The Total Score resulting from the execution of decodIRT is not yet the final rank of explainability of the model, because in this case it is necessary to calculate the average of the skills found for each attribute, figure 1 (H), involving the different variations and combinations of attributes used in the previous steps.
Ultimately, figure 1 (I) and (J), an explanation rank is generated where each attribute appears with a skill value. In this case, the lower the ability values, the more the attribute explains the analyzed model.
You must have installed the R environment in local executions. Then install from the project link: https://cloud.r-project.org/.
Now, just run the commands below in a python environment and you will have your installation normally.
Installation of the "catsim" dependency.
pip install catsim
Installation of the "decodIRT" dependency.
!wget https://raw.githubusercontent.com/josesousaribeiro/eXirt-XAI-Benchmark/main/decodIRT/decodIRT_MLtIRT.py
!wget https://raw.githubusercontent.com/josesousaribeiro/eXirt-XAI-Benchmark/main/decodIRT/decodIRT_analysis.py
Installation of the "eXirt"
pip install eXirt
from pyexirt.eXirt import Explainer
explainer = Explainer()
global_explanation_attributes, global_explanation_attributes_scores =
explainer.explainRankByEXirt(model, X_train, X_test, y_train, y_test,dataset_name)
Note:
global_explanation_attributes = name of relevance features
global_explanation_attributes_scores = rank of relevance features with score
The result of running the eXirt is an attribute relevance rank, sorted in ascending order with respect to the calculated skill. The lower the ability of an attribute, the greater its relevance for explaining the model. In figure 2 is presented for the phoneme dataset.
Figure 2 - Result of eXirt to phoneme dataset.
You can cite the package using the following bibtex entry:
@misc{https://doi.org/10.48550/arxiv.2210.09933,
doi = {10.48550/ARXIV.2210.09933},
url = {https://arxiv.org/abs/2210.09933},
author = {Ribeiro, José and Cardoso, Lucas and Silva, Raíssa and Cirilo, Vitor and Carneiro, Níkolas and Alves, Ronnie},
keywords = {Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences, I.2.6},
title = {Global Explanation of Tree-Ensembles Models Based on Item Response Theory},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}