Outcome trajectory estimation for optimal dynamic treatment regimes with repeated measures

Research output: Contribution to journalArticlepeer-review

Abstract

In recent sequential multiple assignment randomized trials, outcomes were assessed multiple times to evaluate longer-term impacts of the dynamic treatment regimes (DTRs). Q-learning requires a scalar response to identify the optimal DTR. Inverse probability weighting may be used to estimate the optimal outcome trajectory, but it is inefficient, susceptible to model mis-specification, and unable to characterize how treatment effects manifest over time. We propose modified Q-learning with generalized estimating equations to address these limitations and apply it to the M-bridge trial, which evaluates adaptive interventions to prevent problematic drinking among college freshmen. Simulation studies demonstrate our proposed method improves efficiency and robustness.

Original languageEnglish (US)
Pages (from-to)976-991
Number of pages16
JournalJournal of the Royal Statistical Society. Series C: Applied Statistics
Volume72
Issue number4
DOIs
StatePublished - Aug 2023

Bibliographical note

Publisher Copyright:
© 2023 The Royal Statistical Society.

Keywords

  • Q-learning
  • generalized estimating equation
  • heterogeneous treatment effect
  • longitudinal outcome trajectory
  • sequential multiple assignment randomized trial

PubMed: MeSH publication types

  • Journal Article

Fingerprint

Dive into the research topics of 'Outcome trajectory estimation for optimal dynamic treatment regimes with repeated measures'. Together they form a unique fingerprint.

Cite this