Reinforcement Learning: A Promising Tool for Predicting Optimal Treatment in Complex Diseases
November 3, 2023
Reinforcement Learning (RL), a crucial component of machine learning (ML), serves as a remarkable framework for identifying a sequence of actions aimed at increasing the likelihood of accomplishing a predetermined goal. RL has been widely used in the fields of gaming and robotics, with notable instances including its utilization in defeating a world champion program in chess, shoji, and Go.1 In recent years, RL has gained significant attention in the healthcare sector, particularly in the domain of clinical decision-making and the optimization of treatment regimens.
Reinforcement learning in healthcare decision-making: What to do for best possible patient outcomes
In the context of healthcare, treatment regimens often involve a dynamic and sequential decision-making process, where the sequence of interventions can significantly influence patient outcomes. For example, a clinician chooses a treatment based on a patient’s health, observes the response to the treatment, and then the process is repeated. The potential of RL in clinical decision-making is gaining recognition and showing great promise. As an example, a novel RL framework developed by Yala et al. to assist personalized breast cancer screening was found to outperform the standard regimen used in clinical practice.2 Additionally, using real-world medical registry data, Liu et al. employed an RL framework to predict the sequence of prevention and treatments for acute and chronic graft versus host disease and demonstrated its potential to enhance long-term outcomes of patients by learning from expert actions.3
RL provides the capacity to adapt and refine diagnostic and treatment strategies in response to evolving patient conditions and characteristics, ensuring the continuous pursuit of an effective course of action. In scenarios where numerous treatment options exist, RL’s ability to navigate this complex decision space is invaluable.
To illustrate, consider a patient with cancer undergoing treatment — RL can assist in determining the optimal sequence of regimens, dosages, and timing to maximize the effectiveness of the treatment. This process typically involves a series of decisions made at each treatment stage, where RL algorithms can learn from the patient’s response and adapt the treatment plans accordingly, making it a truly personalized approach.
A case for the treatment of multiple myeloma
In our upcoming presentation at ISPOR Europe, we will be showcasing a case study using an RL framework to estimate optimal personalized treatment sequencing for patients with multiple myeloma (MM). The established framework may provide a robust and clinically meaningful approach for estimating personalized treatment regimens that can improve patient outcomes and address heterogeneity in MM and similar disease settings using real-world claims data.
In the field of chronic disease management, RL can contribute to optimizing the management of chronic diseases by creating individualized care plans. Additionally, in the realm of pharmacogenomics, RL can aid in the interpretation of large and complex datasets. By analyzing an individual’s genetic makeup and understanding how it influences their response to treatments, RL can help clinicians make informed decisions about the most suitable therapies, dosages, and potential side effects for a patient.
The future of reinforcement learning
RL has emerged as a promising tool in the healthcare sector, particularly in clinical decision-making. By leveraging the utilization of data-driven decision-making and continuous adaptation, RL contributes to a healthcare landscape where treatments and interventions are uniquely tailored to each patient’s biology and circumstances. Its ability to identify optimal treatment sequences, adapt to evolving patient conditions, and personalize treatment plans makes it an asset in the pursuit of improved patient care and outcomes.
Despite its promise, there is still some work to do before one sees large-scale implementation. If used incorrectly, RL frameworks could suggest practices that are not supported by evidence, which may cause harmful patient outcomes. In addition, lack of measurement in relevant variables in real-world datasets may impact the prediction accuracy of RL frameworks.
Future research should focus on developing guidance to facilitate the validation process of RL algorithms intended for healthcare applications, leveraging extensive, prospectively collected datasets and incorporate appropriate RL systems with comprehensive static and time-varying features to ensure prediction accuracy.
Interested in learning more? Contact us!
Notes
- Silver, D., Hubert, T., Schrittwieser, J., Antonoglou, I., Lai, M., Guez, A., … & Hassabis, D. (2018). A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play. Science, 362(6419), 1140-1144.
- Yala, A., Mikhael, P. G., Lehman, C., Lin, G., Strand, F., Wan, Y. L., … & Barzilay, R. (2022). Optimizing risk-based breast cancer screening policies with reinforcement learning. Nature medicine, 28(1), 136-143.
- Liu, Y., Logan, B., Liu, N., Xu, Z., Tang, J., & Wang, Y. (2017, August). Deep reinforcement learning for dynamic treatment regimes on medical registry data. In 2017 IEEE international conference on healthcare informatics (ICHI)(pp. 380-385). IEEE.
Subscribe to our newsletter
Fei Tang
Research Manager
Fei brings over eight years of experience in epidemiologic study design, claims-based and longitudinal data analysis, advanced statistical modeling, and manuscript writing. She has worked extensively with major US RWD sources, JMDC data, and German claims. Additionally, she has expertise in conducting studies using ECA designs and QBA methodologies. Before joining Cytel, Fei served as a Research Scientist at the New York State Office of Mental Health. In this role, she led multiple research studies aimed at identifying risk factors and evaluating service utilization and health outcomes among vulnerable populations with behavioral health needs. Fei holds a Master of Public Health (MPH) degree and a Ph.D. in epidemiology from the University at Albany – State University of New York.
Read full employee bio
Evie Merinopoulou
Senior Director, Real-World Evidence
Evie Merinopoulou is Senior Director, Real-World Evidence, at Cytel. She is a health economist and real-world data scientist working on applications of real-world evidence in support of regulatory and HTA decision-making.
Evie has worked in the healthcare consulting industry for over 10 years, currently leading the design and execution of observational research projects using global real-world data. She particularly focuses on projects involving real-world synthetic control arms, quantitative bias analysis, head-to-head comparisons using target trial emulation, and transportability analysis.
Read full employee bioClaim your free 30-minute strategy session
Book a free, no-obligation strategy session with a Cytel expert to get advice on how to improve your drug’s probability of success and plot a clearer route to market.