scispace - formally typeset
Open AccessJournal ArticleDOI

Genomic selection using regularized linear regression models: ridge regression, lasso, elastic net and their extensions

TLDR
The elastic net, lasso, adaptive lasso and the adaptive elastic net all had similar accuracies but outperformed ridge regression and ridge regression BLUP in terms of the Pearson correlation between predicted GEBVs and the true genomic value as well as the root mean squared error.
Abstract
Background Genomic selection (GS) is emerging as an efficient and cost-effective method for estimating breeding values using molecular markers distributed over the entire genome. In essence, it involves estimating the simultaneous effects of all genes or chromosomal segments and combining the estimates to predict the total genomic breeding value (GEBV). Accurate prediction of GEBVs is a central and recurring challenge in plant and animal breeding. The existence of a bewildering array of approaches for predicting breeding values using markers underscores the importance of identifying approaches able to efficiently and accurately predict breeding values. Here, we comparatively evaluate the predictive performance of six regularized linear regression methods-- ridge regression, ridge regression BLUP, lasso, adaptive lasso, elastic net and adaptive elastic net-- for predicting GEBV using dense SNP markers.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

On Hyperparameter Optimization of Machine Learning Algorithms: Theory and Practice

TL;DR: This survey paper will help industrial users, data analysts, and researchers to better develop machine learning models by identifying the proper hyper-parameter configurations effectively and introducing several state-of-the-art optimization techniques.
Journal ArticleDOI

Guidelines for Developing and Reporting Machine Learning Predictive Models in Biomedical Research: A Multidisciplinary View.

TL;DR: A set of guidelines was generated to enable correct application of machine learning models and consistent reporting of model specifications and results in biomedical research and it is believed that such guidelines will accelerate the adoption of big data analysis, particularly with machine learning methods, in the biomedical research community.
Journal ArticleDOI

Genomic selection: genome-wide prediction in plant improvement

TL;DR: Several genomic selection (GS) models are reviewed with respect to both the prediction accuracy and genetic gain from selection.
Journal ArticleDOI

An enhanced extreme learning machine model for river flow forecasting: State-of-the-art, practical applications in water resource engineering area and future research direction

TL;DR: The obtained results reveal that the EELM model is a robust expert model and can be embraced practically in real-life water resources management and river sustainability decisions.
Journal ArticleDOI

A Review of Feature Reduction Techniques in Neuroimaging

TL;DR: Feature reduction is an essential step before training a machine learning model to avoid overfitting and therefore improving model prediction accuracy and generalization ability and in this review, feature reduction techniques used with machine learning in neuroimaging studies are discussed.
References
More filters
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

Regularization and variable selection via the elastic net

TL;DR: It is shown that the elastic net often outperforms the lasso, while enjoying a similar sparsity of representation, and an algorithm called LARS‐EN is proposed for computing elastic net regularization paths efficiently, much like algorithm LARS does for the lamba.
Journal ArticleDOI

Regularization Paths for Generalized Linear Models via Coordinate Descent

TL;DR: In comparative timings, the new algorithms are considerably faster than competing methods and can handle large problems and can also deal efficiently with sparse features.
Journal ArticleDOI

The Elements of Statistical Learning

Eric R. Ziegel
- 01 Aug 2003 - 
TL;DR: Chapter 11 includes more case studies in other areas, ranging from manufacturing to marketing research, and a detailed comparison with other diagnostic tools, such as logistic regression and tree-based methods.
Related Papers (5)