Fast Penalized Regression and Cross Validation for Tall Data with the oem Package

Jared D. Huling, Peter Chien

Research output: Contribution to journalArticlepeer-review

1 Scopus citations


A large body of research has focused on theory and computation for variable selection techniques for high dimensional data. There has been substantially less work in the big “tall” data paradigm, where the number of variables may be large, but the number of observations is much larger. The orthogonalizing expectation maximization (OEM) algorithm is one approach for computation of penalized models which excels in the big tall data regime. The oem package is an efficient implementation of the OEM algorithm which provides a multitude of computation routines with a focus on big tall data, such as a function for out-of-memory computation, for large-scale parallel computation of penalized regression models. Furthermore, in this paper we propose a specialized implementation of the OEM algorithm for cross validation, dramatically reducing the computing time for cross validation over a naive implementation.

Original languageEnglish (US)
JournalJournal of Statistical Software
Issue number6
StatePublished - 2022

Bibliographical note

Funding Information:
This material is based upon work supported by, or in part by, NSF Grants DMS 1055214 and DMS 1564376, and NIH grant T32HL083806.

Publisher Copyright:
© 2022, American Statistical Association. All rights reserved.


  • C++
  • expectation maximization
  • lasso
  • MCP
  • OpenMP
  • optimization
  • out-of-memory computing
  • parallel computing


Dive into the research topics of 'Fast Penalized Regression and Cross Validation for Tall Data with the oem Package'. Together they form a unique fingerprint.

Cite this