Show simple item record

dc.contributor.authorGhosh, Susobhan
dc.contributor.authorKim, Raphael
dc.contributor.authorChhabria, Prasidh
dc.contributor.authorDwivedi, Raaz
dc.contributor.authorKlasnja, Predrag
dc.contributor.authorLiao, Peng
dc.contributor.authorZhang, Kelly
dc.contributor.authorMurphy, Susan
dc.date.accessioned2025-04-11T16:43:40Z
dc.date.available2025-04-11T16:43:40Z
dc.date.issued2024-04-10
dc.identifier.urihttps://hdl.handle.net/1721.1/159071
dc.description.abstractThere is a growing interest in using reinforcement learning (RL) to personalize sequences of treatments in digital health to support users in adopting healthier behaviors. Such sequential decision-making problems involve decisions about when to treat and how to treat based on the user’s context (e.g., prior activity level, location, etc.). Online RL is a promising data-driven approach for this problem as it learns based on each user’s historical responses and uses that knowledge to personalize these decisions. However, to decide whether the RL algorithm should be included in an “optimized” intervention for real-world deployment, we must assess the data evidence indicating that the RL algorithm is actually personalizing the treatments to its users. Due to the stochasticity in the RL algorithm, one may get a false impression that it is learning in certain states and using this learning to provide specific treatments. We use a working definition of personalization and introduce a resampling-based methodology for investigating whether the personalization exhibited by the RL algorithm is an artifact of the RL algorithm stochasticity. We illustrate our methodology with a case study by analyzing the data from a physical activity clinical trial called HeartSteps, which included the use of an online RL algorithm. We demonstrate how our approach enhances data-driven truth-in-advertising of algorithm personalization both across all users as well as within specific users in the study.en_US
dc.publisherSpringer USen_US
dc.relation.isversionofhttps://doi.org/10.1007/s10994-024-06526-xen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceSpringer USen_US
dc.titleDid we personalize? Assessing personalization by an online reinforcement learning algorithm using resamplingen_US
dc.typeArticleen_US
dc.identifier.citationGhosh, S., Kim, R., Chhabria, P. et al. Did we personalize? Assessing personalization by an online reinforcement learning algorithm using resampling. Mach Learn 113, 3961–3997 (2024).en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.relation.journalMachine Learningen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-04-10T03:31:55Z
dc.language.rfc3066en
dc.rights.holderThe Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature
dspace.embargo.termsY
dspace.date.submission2025-04-10T03:31:55Z
mit.journal.volume113en_US
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record