Search: WFRF:(Raza Shahid 1980 )
> (2020-2024) >
Experimental Analys...
Experimental Analysis of Trustworthy In-Vehicle Intrusion Detection System Using eXplainable Artificial Intelligence (XAI)
-
- Lundberg, Hampus (author)
- Mittuniversitetet,Institutionen för informationssystem och –teknologi
-
- Mowla, Nishat (author)
- RISE,Mobilitet och system
-
- Fakhrul Abedin, Sarder (author)
- Mittuniversitetet,Institutionen för informationssystem och –teknologi
-
show more...
-
- Thar, Kyi (author)
- Mittuniversitetet,Institutionen för informationssystem och –teknologi
-
- Mahmood, Aamir, 1980- (author)
- Mittuniversitetet,Institutionen för informationssystem och –teknologi
-
- Gidlund, Mikael, 1972- (author)
- Mittuniversitetet,Institutionen för informationssystem och –teknologi
-
- Raza, Shahid, 1980- (author)
- RISE,Datavetenskap
-
show less...
-
(creator_code:org_t)
- Institute of Electrical and Electronics Engineers Inc. 2022
- 2022
- English.
-
In: IEEE Access. - : Institute of Electrical and Electronics Engineers Inc.. - 2169-3536. ; 10, s. 102831-102841
- Related links:
-
https://doi.org/10.1...
-
show more...
-
https://urn.kb.se/re...
-
https://doi.org/10.1...
-
https://urn.kb.se/re...
-
show less...
Abstract
Subject headings
Close
- Anomaly-based In-Vehicle Intrusion Detection System (IV-IDS) is one of the protection mechanisms to detect cyber attacks on automotive vehicles. Using artificial intelligence (AI) for anomaly detection to thwart cyber attacks is promising but suffers from generating false alarms and making decisions that are hard to interpret. Consequently, this issue leads to uncertainty and distrust towards such IDS design unless it can explain its behavior, e.g., by using eXplainable AI (XAI). In this paper, we consider the XAI-powered design of such an IV-IDS using CAN bus data from a public dataset, named 'Survival'. Novel features are engineered, and a Deep Neural Network (DNN) is trained over the dataset. A visualization-based explanation, 'VisExp', is created to explain the behavior of the AI-based IV-IDS, which is evaluated by experts in a survey, in relation to a rule-based explanation. Our results show that experts' trust in the AI-based IV-IDS is significantly increased when they are provided with VisExp (more so than the rule-based explanation). These findings confirm the effect, and by extension the need, of explainability in automated systems, and VisExp, being a source of increased explainability, shows promise in helping involved parties gain trust in such systems.
Subject headings
- NATURVETENSKAP -- Data- och informationsvetenskap (hsv//swe)
- NATURAL SCIENCES -- Computer and Information Sciences (hsv//eng)
Keyword
- Automotive
- deep learning
- intrusion detection system
- machine learning
- trustworthiness
- XAI
- Automation
- Behavioral research
- Computer crime
- Crime
- Decision trees
- Deep neural networks
- Network security
- Vehicles
- Automotives
- Behavioral science
- Intrusion Detection Systems
- Intrusion-Detection
- Machine-learning
- Random forests
- Trust management
- Intrusion detection
Publication and Content Type
- ref (subject category)
- art (subject category)
Find in a library
To the university's database
- By the author/editor
-
Lundberg, Hampus
-
Mowla, Nishat
-
Fakhrul Abedin, ...
-
Thar, Kyi
-
Mahmood, Aamir, ...
-
Gidlund, Mikael, ...
-
show more...
-
Raza, Shahid, 19 ...
-
show less...
- About the subject
-
- NATURAL SCIENCES
-
NATURAL SCIENCES
-
and Computer and Inf ...
- Articles in the publication
-
IEEE Access
- By the university
-
RISE
-
Mid Sweden University