https://www.mdu.se/

mdu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Explainable Artificial Intelligence for Enhancing Transparency in Decision Support Systems
Mälardalen University, School of Innovation, Design and Engineering.ORCID iD: 0000-0003-0730-4405
2024 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Artificial Intelligence (AI) is recognized as advanced technology that assist in decision-making processes with high accuracy and precision. However, many AI models are generally appraised as black boxes due to their reliance on complex inference mechanisms.  The intricacies of how and why these AI models reach a decision are often not comprehensible to human users, resulting in concerns about the acceptability of their decisions. Previous studies have shown that the lack of associated explanation in a human-understandable form makes the decisions unacceptable to end-users. Here, the research domain of Explainable AI (XAI) provides a wide range of methods with the common theme of investigating how AI models reach to a decision or explain it. These explanation methods aim to enhance transparency in Decision Support Systems (DSS), particularly crucial in safety-critical domains like Road Safety (RS) and Air Traffic Flow Management (ATFM). Despite ongoing developments, DSSs are still in the evolving phase for safety-critical applications. Improved transparency, facilitated by XAI, emerges as a key enabler for making these systems operationally viable in real-world applications, addressing acceptability and trust issues. Besides, certification authorities are less likely to approve the systems for general use following the current mandate of Right to Explanation from the European Commission and similar directives from organisations across the world. This urge to permeate the prevailing systems with explanations paves the way for research studies on XAI concentric to DSSs.

To this end, this thesis work primarily developed explainable models for the application domains of RS and ATFM. Particularly, explainable models are developed for assessing drivers' in-vehicle mental workload and driving behaviour through classification and regression tasks. In addition, a novel method is proposed for generating a hybrid feature set from vehicular and electroencephalography (EEG) signals using mutual information (MI). The use of this feature set is successfully demonstrated to reduce the efforts required for complex computations of EEG feature extraction.  The concept of MI was further utilized in generating human-understandable explanations of mental workload classification. For the domain of ATFM, an explainable model for flight take-off time delay prediction from historical flight data is developed and presented in this thesis. The gained insights through the development and evaluation of the explainable applications for the two domains underscore the need for further research on the advancement of XAI methods.

In this doctoral research, the explainable applications for the DSSs are developed with the additive feature attribution (AFA) methods, a class of XAI methods that are popular in current XAI research. Nevertheless, there are several sources from the literature that assert that feature attribution methods often yield inconsistent results that need plausible evaluation. However, the existing body of literature on evaluation techniques is still immature offering numerous suggested approaches without a standardized consensus on their optimal application in various scenarios. To address this issue, comprehensive evaluation criteria are also developed for AFA methods as the literature on XAI suggests. The proposed evaluation process considers the underlying characteristics of the data and utilizes the additive form of Case-based Reasoning, namely AddCBR. The AddCBR is proposed in this thesis and is demonstrated to complement the evaluation process as the baseline to compare the feature attributions produced by the AFA methods. Apart from generating an explanation with feature attribution, this thesis work also proposes the iXGB-interpretable XGBoost. iXGB generates decision rules and counterfactuals to support the output of an XGBoost model thus improving its interpretability. From the functional evaluation, iXGB demonstrates the potential to be used for interpreting arbitrary tree-ensemble methods.

In essence, this doctoral thesis initially contributes to the development of ideally evaluated explainable models tailored for two distinct safety-critical domains. The aim is to augment transparency within the corresponding DSSs. Additionally, the thesis introduces novel methods for generating more comprehensible explanations in different forms, surpassing existing approaches. It also showcases a robust evaluation approach for XAI methods.

Place, publisher, year, edition, pages
Västerås: Mälardalen university , 2024.
Series
Mälardalen University Press Dissertations, ISSN 1651-4238 ; 397
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:mdh:diva-64909ISBN: 978-91-7485-626-2 (print)OAI: oai:DiVA.org:mdh-64909DiVA, id: diva2:1816127
Public defence
2024-01-30, Gamma, Mälardalens universitet, Västerås, 13:15 (English)
Opponent
Supervisors
Available from: 2023-12-04 Created: 2023-12-01 Last updated: 2024-01-09Bibliographically approved
List of papers
1. Deep Learning for Automatic EEG Feature Extraction: An Application in Drivers' Mental Workload Classification
Open this publication in new window or tab >>Deep Learning for Automatic EEG Feature Extraction: An Application in Drivers' Mental Workload Classification
Show others...
2019 (English)In: Communications in Computer and Information Science, Volume 1107, 2019, p. 121-135Conference paper, Published paper (Refereed)
Abstract [en]

In the pursuit of reducing traffic accidents, drivers' mental workload (MWL) has been considered as one of the vital aspects. To measure MWL in different driving situations Electroencephalography (EEG) of the drivers has been studied intensely. However, in the literature, mostly, manual analytic methods are applied to extract and select features from the EEG signals to quantify drivers' MWL. Nevertheless, the amount of time and effort required to perform prevailing feature extraction techniques leverage the need for automated feature extraction techniques. This work investigates deep learning (DL) algorithm to extract and select features from the EEG signals during naturalistic driving situations. Here, to compare the DL based and traditional feature extraction techniques, a number of classifiers have been deployed. Results have shown that the highest value of area under the curve of the receiver operating characteristic (AUC-ROC) is 0.94, achieved using the features extracted by CNN-AE and support vector machine. Whereas, using the features extracted by the traditional method, the highest value of AUC-ROC is 0.78 with the multi-layer perceptron. Thus, the outcome of this study shows that the automatic feature extraction techniques based on CNN-AE can outperform the manual techniques in terms of classification accuracy.

Keywords
Autoencoder, Convolutional Neural Networks, Electroencephalography, Feature Extraction, Mental Workload
National Category
Engineering and Technology Computer Systems
Identifiers
urn:nbn:se:mdh:diva-45059 (URN)10.1007/978-3-030-32423-0_8 (DOI)2-s2.0-85075680380 (Scopus ID)9783030324223 (ISBN)
Conference
The 3rd International Symposium on Human Mental Workload: Models and Applications H-WORKLOAD 2019, 14 Nov 2019, Rome, Italy
Projects
BRAINSAFEDRIVE: A Technology to detect Mental States During Drive for improving the Safety of the road
Available from: 2019-08-22 Created: 2019-08-22 Last updated: 2024-04-15Bibliographically approved
2. A Novel Mutual Information based Feature Set for Drivers’ Mental Workload Evaluation Using Machine Learning
Open this publication in new window or tab >>A Novel Mutual Information based Feature Set for Drivers’ Mental Workload Evaluation Using Machine Learning
Show others...
2020 (English)In: Brain Sciences, E-ISSN 2076-3425, Vol. 10, no 8, p. 1-23, article id 551Article in journal (Refereed) Published
Abstract [en]

Analysis of physiological signals, electroencephalography in more specific notion, is considered as a very promising technique to obtain objective measures for mental workload evaluation, however, it requires complex apparatus to record and thus with poor usability in monitoring in-vehicle drivers’mental workload. This study proposes amethodology of constructing a novel mutual information-based feature set from the fusion of electroencephalography and vehicular signals acquired through real driving experiment and deployed in evaluating drivers’ mental workload. Mutual information of electroencephalography and vehicular signals were used as the prime factor for the fusion of features. In order to assess the reliability of the developed feature set mental workload score prediction, classification and event classification tasks were performed using different machine learning models. Moreover, features extracted from electroencephalography were used to compare the performance. In the prediction of mental workload score, expert-defined scores were used as the target values. For classification tasks, true labels were set from contextual information of the experiment. An extensive evaluation of every prediction tasks was carried out using different validation methods. In predicting mental workload score from the proposed feature set lowest mean absolute error was 0.09 and for classifying mental workload highest accuracy was 94%. According to the outcome of the study, it can be stated that the novel mutual information based features developed through the proposed approach can be employed to classify and monitor in-vehicle drivers’ mental workload.

Place, publisher, year, edition, pages
Switzerland: MDPI AG, 2020
Keywords
electroencephalography, feature extraction, machine learning, mental workload, mutual information, vehicular signal
National Category
Computer Systems
Identifiers
urn:nbn:se:mdh:diva-49988 (URN)10.3390/brainsci10080551 (DOI)000564149000001 ()32823582 (PubMedID)2-s2.0-85089564153 (Scopus ID)
Projects
BRAINSAFEDRIVE: A Technology to detect Mental States During Drive for improving the Safety of the road
Available from: 2020-09-10 Created: 2020-09-10 Last updated: 2024-07-04Bibliographically approved
3. A Systematic Review of Explainable Artificial Intelligence in Terms of Different Application Domains and Tasks
Open this publication in new window or tab >>A Systematic Review of Explainable Artificial Intelligence in Terms of Different Application Domains and Tasks
2022 (English)In: Applied Sciences, E-ISSN 2076-3417, Vol. 12, no 3, article id 1353Article, review/survey (Refereed) Published
Abstract [en]

Artificial intelligence (AI) and machine learning (ML) have recently been radically improved and are now being employed in almost every application domain to develop automated or semi-automated systems. To facilitate greater human acceptability of these systems, explainable artificial intelligence (XAI) has experienced significant growth over the last couple of years with the development of highly accurate models but with a paucity of explainability and interpretability. The literature shows evidence from numerous studies on the philosophy and methodologies of XAI. Nonetheless, there is an evident scarcity of secondary studies in connection with the application domains and tasks, let alone review studies following prescribed guidelines, that can enable researchers’ understanding of the current trends in XAI, which could lead to future research for domain- and application-specific method development. Therefore, this paper presents a systematic literature review (SLR) on the recent developments of XAI methods and evaluation metrics concerning different application domains and tasks. This study considers 137 articles published in recent years and identified through the prominent bibliographic databases. This systematic synthesis of research articles resulted in several analytical findings: XAI methods are mostly developed for safety-critical domains worldwide, deep learning and ensemble models are being exploited more than other types of AI/ML models, visual explanations are more acceptable to end-users and robust evaluation metrics are being developed to assess the quality of explanations. Research studies have been performed on the addition of explanations to widely used AI/ML models for expert users. However, more attention is required to generate explanations for general users from sensitive domains such as finance and the judicial system.

Place, publisher, year, edition, pages
MDPI, 2022
Keywords
Evaluation metrics, Explainability, Explainable artificial intelligence, Systematic literature review
National Category
Computer Sciences
Identifiers
urn:nbn:se:mdh:diva-57253 (URN)10.3390/app12031353 (DOI)000759811300001 ()2-s2.0-85123898494 (Scopus ID)
Available from: 2022-02-09 Created: 2022-02-09 Last updated: 2024-04-15Bibliographically approved
4. Local and Global Interpretability Using Mutual Information in Explainable Artificial Intelligence
Open this publication in new window or tab >>Local and Global Interpretability Using Mutual Information in Explainable Artificial Intelligence
2021 (English)In: 2021 8TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE (ISCMI 2021), IEEE, 2021, p. 191-195Conference paper, Published paper (Refereed)
Abstract [en]

Numerous studies have exploited the potential of Artificial Intelligence (AI) and Machine Learning (ML) models to develop intelligent systems in diverse domains for complex tasks, such as analysing data, extracting features, prediction, recommendation etc. However, presently these systems embrace acceptability issues from the end-users. The models deployed at the back of the systems mostly analyse the correlations or dependencies between the input and output to uncover the important characteristics of the input features, but they lack explainability and interpretability that causing the acceptability issues of intelligent systems and raising the research domain of eXplainable Artificial Intelligence (XAI). In this study, to overcome these shortcomings, a hybrid XAI approach is developed to explain an AI/ML model's inference mechanism as well as the final outcome. The overall approach comprises of 1) a convolutional encoder that extracts deep features from the data and computes their relevancy with features extracted using domain knowledge, 2) a model for classifying data points using the features from autoencoder, and 3) a process of explaining the model's working procedure and decisions using mutual information to provide global and local interpretability. To demonstrate and validate the proposed approach, experimentation was performed using an electroencephalography dataset from road safety to classify drivers' in-vehicle mental workload. The outcome of the experiment was found to be promising that produced a Support Vector Machine classifier for mental workload with approximately 89% performance accuracy. Moreover, the proposed approach can also provide an explanation for the classifier model's behaviour and decisions with the combined illustration of Shapely values and mutual information.

Place, publisher, year, edition, pages
IEEE, 2021
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:mdh:diva-57425 (URN)10.1109/ISCMI53840.2021.9654898 (DOI)000750613100034 ()2-s2.0-85124424608 (Scopus ID)9781728186832 (ISBN)
Conference
8th International Conference on Soft Computing & Machine Intelligence (ISCMI), NOV 26-27 2021, Cairo, Egypt
Available from: 2022-02-16 Created: 2022-02-16 Last updated: 2024-04-15Bibliographically approved
5. Interpretable Machine Learning for Modelling and Explaining Car Drivers' Behaviour: An Exploratory Analysis on Heterogeneous Data
Open this publication in new window or tab >>Interpretable Machine Learning for Modelling and Explaining Car Drivers' Behaviour: An Exploratory Analysis on Heterogeneous Data
2023 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Understanding individual car drivers’ behavioural variations and heterogeneity is a significant aspect of developingcar simulator technologies, which are widely used in transport safety. This also characterizes the heterogeneity in drivers’ behaviour in terms of risk and hurry, using both real-time on-track and in-simulator driving performance features. Machine learning (ML) interpretability has become increasingly crucial for identifying accurate and relevant structural relationships between spatial events and factors that explain drivers’ behaviour while being classified and the explanations for them are evaluated. However, the high predictive power of ML algorithms ignore the characteristics of non-stationary domain relationships in spatiotemporal data (e.g., dependence, heterogeneity), which can lead to incorrect interpretations and poor management decisions. This study addresses this critical issue of ‘interpretability’ in ML-based modelling of structural relationships between the events and corresponding features of the car drivers’ behavioural variations. In this work, an exploratory experiment is described that contains simulator and real driving concurrently with a goal to enhance the simulator technologies. Here, initially, with heterogeneous data, several analytic techniques for simulator bias in drivers’ behaviour have been explored. Afterwards, five different ML classifier models were developed to classify risk and hurry in drivers’ behaviour in real and simulator driving. Furthermore, two different feature attribution-based explanation models were developed to explain the decision from the classifiers. According to the results and observation, among the classifiers, Gradient Boosted Decision Trees performed best with a classification accuracy of 98.62%. After quantitative evaluation, among the feature attribution methods, the explanation from Shapley Additive Explanations (SHAP) was found to be more accurate. The use of different metrics for evaluating explanation methods and their outcome lay the path toward further research in enhancing the feature attribution methods.

Series
International Conference on Agents and Artificial Intelligence, ISSN 21843589
Keywords
Artificial Intelligence, Driving Behaviour, Feature Attribution, Evaluation, Explainable Artificial Intelligence, Interpretability, Road Safety.
National Category
Computer Sciences
Identifiers
urn:nbn:se:mdh:diva-64912 (URN)10.5220/0011801000003393 (DOI)2-s2.0-85182555765 (Scopus ID)
Conference
15th International Conference on Agents and Artificial Intelligence (ICAART 2023)
Available from: 2023-11-30 Created: 2023-11-30 Last updated: 2024-12-04Bibliographically approved
6. Investigating Additive Feature Attribution for Regression
Open this publication in new window or tab >>Investigating Additive Feature Attribution for Regression
2023 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Feature attribution is a class of explainable artificial intelligence (XAI) methods that produce the contributions of data features to a model's decision. There are multiple accounts stating that feature attribution methods produce inconsistent results and should always be evaluated. However, the existing body of literature on evaluation techniques is still immature with multiple proposed techniques and a lack of widely adopted methods, making it difficult to recognize the best approach for each circumstance. This article investigates an approach to creating synthetic data for regression that can be used to evaluate the results of feature attribution methods. From a real-world dataset, the proposed approach describes how to create synthetic data that preserves the patterns of the original data and enables comprehensive evaluation of XAI methods. This research also demonstrates how global and local feature attributions can be represented in the additive form of case-based reasoning as a benchmark method for evaluation. Finally, this work demonstrates the case where a method that includes a standardization step does not produce feature attributions of the same quality as one that does not use standardization in the context of a regression task.

Keywords
Explainability, Additive Feature Attribution, Regression, Additive CBR, CBR, Evaluation, Interpretability, LIME, SHAP, Synthetic Data, XAI.
National Category
Computer Sciences
Identifiers
urn:nbn:se:mdh:diva-64913 (URN)
Note

Submitted to the journal of Artificial Intelligence (AIJ)

Available from: 2023-11-30 Created: 2023-11-30 Last updated: 2024-04-15Bibliographically approved
7. iXGB: improving the interpretability of XGBoost using decision rules and counterfactuals
Open this publication in new window or tab >>iXGB: improving the interpretability of XGBoost using decision rules and counterfactuals
2024 (English)In: Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART, 2024, Vol. 3, p. 1345-1353Conference paper, Published paper (Other academic)
Abstract [en]

Tree-ensemble models, such as Extreme Gradient Boosting (XGBoost), are renowned Machine Learning models which have higher prediction accuracy compared to traditional tree-based models. This higher accuracy, however, comes at the cost of reduced interpretability. Also, the decision path or prediction rule of XGBoost is not explicit like the tree-based models. This paper proposes the iXGB--interpretable XGBoost, an approach to improve the interpretability of XGBoost. iXGB approximates a set of rules from the internal structure of XGBoost and the characteristics of the data. In addition, iXGB generates a set of counterfactuals from the neighbourhood of the test instances to support the understanding of the end-users on their operational relevance. The performance of iXGB in generating rule sets is evaluated with experiments on real and benchmark datasets which demonstrated reasonable interpretability. The evaluation result also supports that the interpretability of XGBoost can be improved without using surrogate methods.

Keywords
Counterfactuals, Explainability, Explainable Artificial Intelligence, Interpretability, Regression, Rule-based Explanation, XGBoost.
National Category
Computer Sciences
Identifiers
urn:nbn:se:mdh:diva-64916 (URN)10.5220/0012474000003636 (DOI)2-s2.0-85190810293 (Scopus ID)978-989-758-680-4 (ISBN)
Conference
16th International Conference on Agents and Artificial Intelligence (ICAART 2024)
Available from: 2023-11-30 Created: 2023-11-30 Last updated: 2024-05-08Bibliographically approved

Open Access in DiVA

fulltext(3089 kB)972 downloads
File information
File name FULLTEXT02.pdfFile size 3089 kBChecksum SHA-512
71d8501b0cf28eef0afd15fbba3a87ad5e4f11ae17c442b1ab9808d540418ef75062704ea2afeb78ccbfa8c1e60fa8757ed1a0d821bc2702108bfe595412d580
Type fulltextMimetype application/pdf

Authority records

Islam, Mir Riyanul

Search in DiVA

By author/editor
Islam, Mir Riyanul
By organisation
School of Innovation, Design and Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 972 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 1973 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf