Abstract
In recent sequential multiple assignment randomized trials, outcomes were assessed multiple times to evaluate longer-term impacts of the dynamic treatment regimes (DTRs). Q-learning requires a scalar response to identify the optimal DTR. Inverse probability weighting may be used to estimate the optimal outcome trajectory, but it is inefficient, susceptible to model mis-specification, and unable to characterize how treatment effects manifest over time. We propose modified Q-learning with generalized estimating equations to address these limitations and apply it to the M-bridge trial, which evaluates adaptive interventions to prevent problematic drinking among college freshmen. Simulation studies demonstrate our proposed method improves efficiency and robustness.
| Original language | English (US) |
|---|---|
| Pages (from-to) | 976-991 |
| Number of pages | 16 |
| Journal | Journal of the Royal Statistical Society. Series C: Applied Statistics |
| Volume | 72 |
| Issue number | 4 |
| DOIs | |
| State | Published - Aug 2023 |
Bibliographical note
Publisher Copyright:© 2023 The Royal Statistical Society.
Keywords
- Q-learning
- generalized estimating equation
- heterogeneous treatment effect
- longitudinal outcome trajectory
- sequential multiple assignment randomized trial
PubMed: MeSH publication types
- Journal Article
Fingerprint
Dive into the research topics of 'Outcome trajectory estimation for optimal dynamic treatment regimes with repeated measures'. Together they form a unique fingerprint.Cite this
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS