Show simple item record

dc.contributor.authorThao, Phan Nguyen Minh
dc.contributor.authorDao, Cong-Tinh
dc.contributor.authorWu, Chenwei
dc.contributor.authorWang, Jian-Zhe
dc.contributor.authorLiu, Shun
dc.contributor.authorDing, Jun-En
dc.contributor.authorRestrepo, David
dc.contributor.authorLiu, Feng
dc.contributor.authorHung, Fang-Ming
dc.contributor.authorPeng, Wen-Chih
dc.date.accessioned2024-11-14T21:33:51Z
dc.date.available2024-11-14T21:33:51Z
dc.date.issued2024-10-21
dc.identifier.isbn979-8-4007-0436-9
dc.identifier.urihttps://hdl.handle.net/1721.1/157546
dc.descriptionCIKM ’24, October 21–25, 2024, Boise, ID, USAen_US
dc.description.abstractElectronic health records (EHRs) are multimodal by nature, consisting of structured tabular features like lab tests and unstructured clinical notes. In real-life clinical practice, doctors use complementary multimodal EHR data sources to get a clearer picture of patients' health and support clinical decision-making. However, most EHR predictive models do not reflect these procedures, as they either focus on a single modality or overlook the inter-modality interactions/redundancy. In this work, we propose MEDFuse, a Multimodal EHR Data Fusion framework that incorporates masked lab-test modeling and large language models (LLMs) to effectively integrate structured and unstructured medical data. MEDFuse leverages multimodal embeddings extracted from two sources: LLMs fine-tuned on free clinical text and masked tabular transformers trained on structured lab test results. We design a disentangled transformer module, optimized by a mutual information loss to 1) decouple modality-specific and modality-shared information and 2) extract useful joint representation from the noise and redundancy present in clinical notes. Through comprehensive validation on the public MIMIC-III dataset and the in-house FEMH dataset, MEDFuse demonstrates great potential in advancing clinical predictions, achieving over 90% F1 score in the 10-disease multi-label classification task.en_US
dc.publisherACM|Proceedings of the 33rd ACM International Conference on Information and Knowledge Managementen_US
dc.relation.isversionofhttps://doi.org/10.1145/3627673.3679962en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleMEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Modelsen_US
dc.typeArticleen_US
dc.identifier.citationThao, Phan Nguyen Minh, Dao, Cong-Tinh, Wu, Chenwei, Wang, Jian-Zhe, Liu, Shun et al. 2024. "MEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Models."
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-11-01T07:46:35Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-11-01T07:46:35Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record