提交
单变量
Lasso(编程语言)
简单(哲学)
计算机科学
数学优化
最优化问题
卡鲁什-库恩-塔克条件
凸优化
多样性(控制论)
数学
正多边形
机器学习
人工智能
多元统计
几何学
万维网
哲学
认识论
数据库
作者
Robert Tibshirani,Jacob Bien,Jerome H. Friedman,Trevor Hastie,Noah Simon,Jonathan Taylor,Ryan J. Tibshirani
标识
DOI:10.1111/j.1467-9868.2011.01004.x
摘要
We consider rules for discarding predictors in lasso regression and related problems, for computational efficiency. El Ghaoui and his colleagues have propose 'SAFE' rules, based on univariate inner products between each predictor and the outcome, which guarantee that a coefficient will be 0 in the solution vector. This provides a reduction in the number of variables that need to be entered into the optimization. We propose strong rules that are very simple and yet screen out far more predictors than the SAFE rules. This great practical improvement comes at a price: the strong rules are not foolproof and can mistakenly discard active predictors, i.e. predictors that have non-zero coefficients in the solution. We therefore combine them with simple checks of the Karush-Kuhn-Tucker conditions to ensure that the exact solution to the convex problem is delivered. Of course, any (approximate) screening method can be combined with the Karush-Kuhn-Tucker, conditions to ensure the exact solution; the strength of the strong rules lies in the fact that, in practice, they discard a very large number of the inactive predictors and almost never commit mistakes. We also derive conditions under which they are foolproof. Strong rules provide substantial savings in computational time for a variety of statistical optimization problems.
科研通智能强力驱动
Strongly Powered by AbleSci AI