dc.contributor.author | Thao, Phan Nguyen Minh | |
dc.contributor.author | Dao, Cong-Tinh | |
dc.contributor.author | Wu, Chenwei | |
dc.contributor.author | Wang, Jian-Zhe | |
dc.contributor.author | Liu, Shun | |
dc.contributor.author | Ding, Jun-En | |
dc.contributor.author | Restrepo, David | |
dc.contributor.author | Liu, Feng | |
dc.contributor.author | Hung, Fang-Ming | |
dc.contributor.author | Peng, Wen-Chih | |
dc.date.accessioned | 2024-11-14T21:33:51Z | |
dc.date.available | 2024-11-14T21:33:51Z | |
dc.date.issued | 2024-10-21 | |
dc.identifier.isbn | 979-8-4007-0436-9 | |
dc.identifier.uri | https://hdl.handle.net/1721.1/157546 | |
dc.description | CIKM ’24, October 21–25, 2024, Boise, ID, USA | en_US |
dc.description.abstract | Electronic health records (EHRs) are multimodal by nature, consisting of structured tabular features like lab tests and unstructured clinical notes. In real-life clinical practice, doctors use complementary multimodal EHR data sources to get a clearer picture of patients' health and support clinical decision-making. However, most EHR predictive models do not reflect these procedures, as they either focus on a single modality or overlook the inter-modality interactions/redundancy. In this work, we propose MEDFuse, a Multimodal EHR Data Fusion framework that incorporates masked lab-test modeling and large language models (LLMs) to effectively integrate structured and unstructured medical data. MEDFuse leverages multimodal embeddings extracted from two sources: LLMs fine-tuned on free clinical text and masked tabular transformers trained on structured lab test results. We design a disentangled transformer module, optimized by a mutual information loss to 1) decouple modality-specific and modality-shared information and 2) extract useful joint representation from the noise and redundancy present in clinical notes. Through comprehensive validation on the public MIMIC-III dataset and the in-house FEMH dataset, MEDFuse demonstrates great potential in advancing clinical predictions, achieving over 90% F1 score in the 10-disease multi-label classification task. | en_US |
dc.publisher | ACM|Proceedings of the 33rd ACM International Conference on Information and Knowledge Management | en_US |
dc.relation.isversionof | https://doi.org/10.1145/3627673.3679962 | en_US |
dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
dc.source | Association for Computing Machinery | en_US |
dc.title | MEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Models | en_US |
dc.type | Article | en_US |
dc.identifier.citation | Thao, Phan Nguyen Minh, Dao, Cong-Tinh, Wu, Chenwei, Wang, Jian-Zhe, Liu, Shun et al. 2024. "MEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Models." | |
dc.identifier.mitlicense | PUBLISHER_POLICY | |
dc.eprint.version | Final published version | en_US |
dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
dc.date.updated | 2024-11-01T07:46:35Z | |
dc.language.rfc3066 | en | |
dc.rights.holder | The author(s) | |
dspace.date.submission | 2024-11-01T07:46:35Z | |
mit.license | PUBLISHER_POLICY | |
mit.metadata.status | Authority Work and Publication Information Needed | en_US |