LCBC
  • News
    • Media coverage
  • About
    • Who we are
    • Vacancies
    • Funding and approval
    • Our methods
    • Find us
  • Research Areas
    • Lifespan changes
    • Memory
    • Development
    • Brain plasticity and cognitive training
    • Alzheimer’s Disease
  • Publications
  • Search
  • Menu Menu

The journal of open source software ; pdf

Øystein Sørensen

High Dimensional Regression with Measurement Error

A new research paper

Norwegian links

English links

Many problems in science involve using measured variables to explain an outcome of interest using some statistical regression model. In high-dimensional problems, charac- terized by having a very large number of variables, one often focuses on finding a subset of variables with good explanatory power. An example from cancer research involves finding gene expressions or other biomarkers which can explain disease progression, from a large set of candidates (Kristensen et al., 2014). Another example is customer analytics, where it may be of interest to find out which variables predict whether customers will return or not, and variables of interest include factors like previous purchasing patterns, demographics, and satisfaction measures (Baesens, 2014).

The lasso (Tibshirani, 1996) and the Dantzig selector (Candes & Tao, 2007; James & Rad- chenko, 2009) are popular methods for variable selection in this type of problems, com- bining computational speed with good statistical properties (Bühlmann & Geer, 2011). In many practical applications, the process of measuring the variables of interest is sub- ject to measurement error (Carroll, Ruppert, Stefanski, & Crainiceanu, 2006), but this additional source of noise is neglected by the aforementioned models. Such measurement error has been shown to lead to worse variable selection properties of the lasso (Sørensen, Frigessi, & Thoresen, 2015), typically involving an increased number of false positive se- lections. A corrected lasso has been proposed and analyzed by Loh & Wainwright (2012) for linear models and Sørensen et al. (2015) for generalized linear models. It has been applied by Vasquez, Hu, Roe, Halonen, & Guerra (2019) in a problem involving mea- surement of serum biomarkers. For the Dantzig selector, Rosenbaum & Tsybakov (2010) proposed the Matrix Uncertainty Selector (MUS) for linear models, which was extended to the generalized linear model case by Sørensen, Hellton, Frigessi, & Thoresen (2018) with an algorithm named GMUS (Generalized MUS).

hdme is an R (R Core Team, 2018) package containing implementations of both the corrected lasso and the MU selector for high-dimensional measurement error problems. Its main functions are gmus() and corrected_lasso(). Additional functions pro- vide opportunities for hyperparameter tuning using cross-validation or the elbow rule (Rosenbaum & Tsybakov, 2010), and plotting tools for visualizing the model fit. The underlying numerical procedures are implemented in C++ using the RcppArmadillo pack- age (Eddelbuettel & Sanderson, 2014) and linear programming with Rglpk (Theussl & Hornik, 2019). hdme is available from the comprehensive R archive network (CRAN) at https://CRAN.R-project.org, and the latest development version is available at https: //github.com/osorensen/hdme. The package vignette, which can be opened in R with the command vignette(“hdme”), contains a step-by-step introduction to the models implemented in the package.

Click the image to go to the paper

© Copyright - LCBC - Developed by LCBC (James Michael Roe & Athanasia Mowinckel) & New Element AS
  • Twitter
  • Facebook
  • News
  • About
  • Research Areas
  • Publications
Scroll to top

This site uses cookies. By continuing to browse the site, you are agreeing to our use of cookies.

OKLearn more

Cookie and Privacy Settings



How we use cookies

We may request cookies to be set on your device. We use cookies to let us know when you visit our websites, how you interact with us, to enrich your user experience, and to customize your relationship with our website.

Click on the different category headings to find out more. You can also change some of your preferences. Note that blocking some types of cookies may impact your experience on our websites and the services we are able to offer.

Essential Website Cookies

These cookies are strictly necessary to provide you with services available through our website and to use some of its features.

Because these cookies are strictly necessary to deliver the website, refusing them will have impact how our site functions. You always can block or delete cookies by changing your browser settings and force blocking all cookies on this website. But this will always prompt you to accept/refuse cookies when revisiting our site.

We fully respect if you want to refuse cookies but to avoid asking you again and again kindly allow us to store a cookie for that. You are free to opt out any time or opt in for other cookies to get a better experience. If you refuse cookies we will remove all set cookies in our domain.

We provide you with a list of stored cookies on your computer in our domain so you can check what we stored. Due to security reasons we are not able to show or modify cookies from other domains. You can check these in your browser security settings.

Other external services

We also use different external services like Google Webfonts, Google Maps, and external Video providers. Since these providers may collect personal data like your IP address we allow you to block them here. Please be aware that this might heavily reduce the functionality and appearance of our site. Changes will take effect once you reload the page.

Google Webfont Settings:

Google Map Settings:

Google reCaptcha Settings:

Vimeo and Youtube video embeds:

Accept settingsHide notification only