TY - GEN
T1 - Trading-Off Interpretability and Accuracy in Medical Applications
T2 - 2024 IEEE International Conference on Fuzzy Systems, FUZZ-IEEE 2024
AU - Sharma, Arnab
AU - Leite, Daniel
AU - Demir, Caglar
AU - Ngomo, Axel Cyrille Ngonga
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - With an increased number of applications of machine learning models to support decision-making in critical domains, there is a pressing need to understand the internal behavior of these models. Essentially, explaining learning models to humans has expedited the development of methods to extract information and access models' inner components. Researchers have proposed approaches to compute explanations for different types of models. While these approaches are often effective in explaining complex relations and rough input-output maps, studies on model optimality, which trade off accuracy and interpretability are scarce. We conducted a study to understand the relationship between accuracy and interpretability of Hoeffding trees, developed from evolving data streams. We employed formal reasoning techniques, founded on theoretical guarantees, to generate subset-minimal explanations for a set of examples. Rankings of features, according to their importance to the model outcomes, were obtained. After computing model accuracy and interpretability, the least important feature based on the ranking was removed. By repeating this procedure, we leveraged the setup that leads to an optimal (accurate and interpretable) tree model. Application examples considered medical datasets, namely the Parkinson's Disease telemonitoring and the EEG eye-state datasets, to generate Hoeffding regression and classification trees, respectively. The study has shown that as tree interpretability increases, accuracy tends to decrease; however, an optimal solution can be established by balancing conflicting aspects.
AB - With an increased number of applications of machine learning models to support decision-making in critical domains, there is a pressing need to understand the internal behavior of these models. Essentially, explaining learning models to humans has expedited the development of methods to extract information and access models' inner components. Researchers have proposed approaches to compute explanations for different types of models. While these approaches are often effective in explaining complex relations and rough input-output maps, studies on model optimality, which trade off accuracy and interpretability are scarce. We conducted a study to understand the relationship between accuracy and interpretability of Hoeffding trees, developed from evolving data streams. We employed formal reasoning techniques, founded on theoretical guarantees, to generate subset-minimal explanations for a set of examples. Rankings of features, according to their importance to the model outcomes, were obtained. After computing model accuracy and interpretability, the least important feature based on the ranking was removed. By repeating this procedure, we leveraged the setup that leads to an optimal (accurate and interpretable) tree model. Application examples considered medical datasets, namely the Parkinson's Disease telemonitoring and the EEG eye-state datasets, to generate Hoeffding regression and classification trees, respectively. The study has shown that as tree interpretability increases, accuracy tends to decrease; however, an optimal solution can be established by balancing conflicting aspects.
KW - Evolving Intelligent Systems
KW - Explainable AI
KW - Hoeffding Tree
KW - Incremental Learning
UR - http://www.scopus.com/inward/record.url?scp=85201549728&partnerID=8YFLogxK
U2 - 10.1109/FUZZ-IEEE60900.2024.10611982
DO - 10.1109/FUZZ-IEEE60900.2024.10611982
M3 - Conference contribution
AN - SCOPUS:85201549728
T3 - IEEE International Conference on Fuzzy Systems
BT - 2024 IEEE International Conference on Fuzzy Systems, FUZZ-IEEE 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 30 June 2024 through 5 July 2024
ER -