MIT Libraries logoDSpace@MIT

MIT
View Item 
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
  • DSpace@MIT Home
  • MIT Open Access Articles
  • MIT Open Access Articles
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

The Situation Awareness Framework for Explainable AI (SAFE-AI) and Human Factors Considerations for XAI Systems

Author(s)
Sanneman, Lindsay; Shah, Julie A
Thumbnail
DownloadPublished version (2.373Mb)
Publisher with Creative Commons License

Publisher with Creative Commons License

Creative Commons Attribution

Terms of use
Creative Commons Attribution-NonCommercial-NoDerivatives https://creativecommons.org/licenses/by-nc-nd/4.0/
Metadata
Show full item record
Abstract
Recent advances in artificial intelligence (AI) have drawn attention to the need for AI systems to beunderstandable to human users. The explainable AI (XAI) literature aims to enhance human under-standing and human-AI team performance by providing users with necessary information about AI sys-tem behavior. Simultaneously, the human factors literature has long addressed importantconsiderations that contribute to human performance, including how to determine human informa-tional needs, human workload, and human trust in autonomous systems. Drawing from the human fac-tors literature, we propose the Situation Awareness Framework for Explainable AI (SAFE-AI), a three-level framework for the development and evaluation of explanations about AI system behavior. Ourproposed levels of XAI are based on the informational needs of human users, which can be deter-mined using the levels of situation awareness (SA) framework from the human factors literature. Basedon our levels of XAI framework, we also suggest a method for assessing the effectiveness of XAI sys-tems. We further detail human workload considerations for determining the content and frequency ofexplanations as well as metrics that can be used to assess human workload. Finally, we discuss theimportance of appropriately calibrating user trust in AI systems through explanations along with othertrust-related considerations for XAI, and we detail metrics that can be used to evaluate user trust inthese systems.
Date issued
2022-06-22
URI
https://hdl.handle.net/1721.1/163525
Department
Massachusetts Institute of Technology. Department of Aeronautics and Astronautics
Journal
International Journal of Human–Computer Interaction
Publisher
Taylor & Francis
Citation
Sanneman, L., & Shah, J. A. (2022). The Situation Awareness Framework for Explainable AI (SAFE-AI) and Human Factors Considerations for XAI Systems. International Journal of Human–Computer Interaction, 38(18–20), 1772–1788. https://doi.org/10.1080/10447318.2022.2081282
Version: Final published version

Collections
  • MIT Open Access Articles

Browse

All of DSpaceCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

My Account

Login

Statistics

OA StatisticsStatistics by CountryStatistics by Department
MIT Libraries
PrivacyPermissionsAccessibilityContact us
MIT
Content created by the MIT Libraries, CC BY-NC unless otherwise noted. Notify us about copyright concerns.