• Aucun résultat trouvé

Model selection via the lasso in conditional logistic regression

N/A
N/A
Protected

Academic year: 2021

Partager "Model selection via the lasso in conditional logistic regression"

Copied!
2
0
0

Texte intégral

(1)

HAL Id: inserm-00402434

https://www.hal.inserm.fr/inserm-00402434

Submitted on 7 Jul 2009

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Model selection via the lasso in conditional logistic

regression

Marta Avalos

To cite this version:

Marta Avalos. Model selection via the lasso in conditional logistic regression. Second International Biometric Society Channel Network Conference, Apr 2009, Ghent, Belgium. pp.Model selection and clustering session. �inserm-00402434�

(2)

Conference abstracts

2nd Channel Network Meeting, Ghent, Belgium, April 6-8, 2009

Model selection via the lasso in conditional logistic

regression

Marta Avalos

Biostatistics team of INSERM U897 and University of Bordeaux 2, France

marta.avalos@isped.u-bordeaux2.fr

We propose a model selection procedure in the context of matched case-control studies and, more specifically, for the conditional logistic regression. The method is based on penalized conditional likelihood with an L1-type penalty of the regression coefficients,

the so-called lasso. This penalty, which shrinks coefficients to improve the accuracy of prediction, is particularly adapted when the number of covariates is large (with respect to the number of events) or in case of collinearity between them. An attractive property of the lasso is that it performs parameter estimation and variable selection simultaneously. The implementation of the method is based on a simple modification of the R-package

penalized for fitting generalized linear and Cox proportional hazard models with L1 or L2

penalties, developed by J. Goeman. The effectiveness of the algorithms allows the use of resampling methods for choosing the regularization term and for evaluating the stability of the selected model. Thus, K-fold likelihood cross-validation and bootstrap are applied, both taking into account the dependent nature of data. The methodology is illustrated with some examples.

Références

Documents relatifs

In this note, we propose an alternative regularization procedure based on the penalized least squares for variable selection in linear regression problem, which combines the L 1

After a general introduction to these partition-based strategies, we study two cases: a classical one in which the conditional density depends in a piecewise polynomial manner of

The purpose of the next two sections is to investigate whether the dimensionality of the models, which is freely available in general, can be used for building a computationally

We show that under a suitable condition on the design matrix, the number of nonzero coefficients of the Lasso solution is an unbiased estimate for the degrees of freedom.. An

But model selection procedures based on penalized maximum likelihood estimators in logistic regression are less studied in the literature.. In this paper we focus on model

A second application of this inequality leads to a new general model selection theorem under very mild assumptions on the statistical setting.. We propose 3 illustrations of

Abstract: We wish to estimate conditional density using Gaussian Mixture Regression model with logistic weights and means depending on the covariate.. We aim at selecting the number

First, the mixture of logistic regression model is presented in Section 2.1, and the prediction step is described thereby. Then, parameter estimation by maximization of the