Comparing solution paths of sparse quadratic minimization with a Stieltjes matrix

Author:

He ZiyuORCID,Han Shaoning,Gómez Andrés,Cui Ying,Pang Jong-Shi

Abstract

AbstractThis paper studies several solution paths of sparse quadratic minimization problems as a function of the weighing parameter of the bi-objective of estimation loss versus solution sparsity. Three such paths are considered: the “$$\ell _0$$ 0 -path” where the discontinuous $$\ell _0$$ 0 -function provides the exact sparsity count; the “$$\ell _1$$ 1 -path” where the $$\ell _1$$ 1 -function provides a convex surrogate of sparsity count; and the “capped $$\ell _1$$ 1 -path” where the nonconvex nondifferentiable capped $$\ell _1$$ 1 -function aims to enhance the $$\ell _1$$ 1 -approximation. Serving different purposes, each of these three formulations is different from each other, both analytically and computationally. Our results deepen the understanding of (old and new) properties of the associated paths, highlight the pros, cons, and tradeoffs of these sparse optimization models, and provide numerical evidence to support the practical superiority of the capped $$\ell _1$$ 1 -path. Our study of the capped $$\ell _1$$ 1 -path is interesting in its own right as the path pertains to computable directionally stationary (= strongly locally minimizing in this context, as opposed to globally optimal) solutions of a parametric nonconvex nondifferentiable optimization problem. Motivated by classical parametric quadratic programming theory and reinforced by modern statistical learning studies, both casting an exponential perspective in fully describing such solution paths, we also aim to address the question of whether some of them can be fully traced in strongly polynomial time in the problem dimensions. A major conclusion of this paper is that a path of directional stationary solutions of the capped $$\ell _1$$ 1 -regularized problem offers interesting theoretical properties and practical compromise between the $$\ell _0$$ 0 -path and the $$\ell _1$$ 1 -path. Indeed, while the $$\ell _0$$ 0 -path is computationally prohibitive and greatly handicapped by the repeated solution of mixed-integer nonlinear programs, the quality of $$\ell _1$$ 1 -path, in terms of the two criteria—loss and sparsity—in the estimation objective, is inferior to the capped $$\ell _1$$ 1 -path; the latter can be obtained efficiently by a combination of a parametric pivoting-like scheme supplemented by an algorithm that takes advantage of the Z-matrix structure of the loss function.

Funder

National Science Foundation

Air Force Office of Scientific Research

Publisher

Springer Science and Business Media LLC

Subject

General Mathematics,Software

Reference45 articles.

1. Ahn, M., Pang, J.S., Xin, J.: Difference-of-convex learning: directional stationarity, optimality, and sparsity. SIAM J. Optim. 27(3), 1637–1665 (2017)

2. Akaike, H.: Information theory and an extension of the maximum likelihood principle. Proceeding of IEEE International Symposium on Information Theory 267–281 (1973)

3. Aneja, Y.P., Nair, K.P.K.: Bicriteria transportation problem. Manag. Sci. 25, 73–78 (1979)

4. Atamtürk, A., Gómez, A.: Strong formulations for quadratic optimzation with M-matrices and indicator variables. Math. Program. Series B 170, 141–176 (2018)

5. Atamtürk, A., Gómez, A.: Rank-one convexifications for sparse regression. arXiv:1901.10334 (2019)

Cited by 4 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3