运筹学学报 ›› 2019, Vol. 23 ›› Issue (2): 86-94.doi: 10.15960/j.cnki.issn.1007-6093.2019.02.008

• 运筹学 • 上一篇    下一篇

求解稀疏逻辑回归问题的嵌套BB算法的分裂增广拉格朗日算法

梁仁莉, 白延琴*   

  1. 上海大学理学院数学系, 上海 200444
  • 收稿日期:2016-04-27 出版日期:2019-06-15 发布日期:2019-06-15
  • 通讯作者: 白延琴 E-mail:yqbai@shu.edu.cn
  • 基金资助:
    国家自然科学基金(No.11371242)

A splitting augmented Lagrangian method embedding in the BB method for solving the sparse logistic problem

LIANG Renli, BAI Yanqin*   

  1. Department of Mathematics, Collage of Sciences, Shanghai University, Shanghai 200444, China
  • Received:2016-04-27 Online:2019-06-15 Published:2019-06-15

摘要: 逻辑回归是经典的分类方法,广泛应用于数据挖掘、机器学习和计算机视觉.现研究带有l0模约束的逻辑回归问题.这类问题广泛用于分类问题中的特征提取,且一般是NP-难的.为了求解这类问题,提出了嵌套BB(Barzilai and Borwein)算法的分裂增广拉格朗日算法(SALM-BB).该算法在迭代中交替地求解一个无约束凸优化问题和一个带l0模约束的二次优化问题.然后借助BB算法求解无约束凸优化问题.通过简单的等价变形直接得到带l0模约束二次优化问题的精确解,并且给出了算法的收敛性定理.最后通过数值实验来测试SALM-BB算法对稀疏逻辑回归问题的计算精确性.数据来源包括真实的UCI数据和模拟数据.数值实验表明,相对于一阶算法SLEP,SALM-BB能够得到更低的平均逻辑损失和错分率.

关键词: 稀疏逻辑回归, 分裂增广拉格朗日算法, 特征提取

Abstract: Logistic regression is a promising method that has been used widely in many applications of data mining, machine learning, computer vision. In this paper, we study on the l0 sparse logistic regression problem. This problem has been proposed as a promising method for feature selection in classification problems, and it is in general NP-hard. In order to solve this problem, we develop a splitting augmented Lagrange method with embedding BB (Barzilai and Borwein) method (SALM-BB). At each iteration, SALM-BB solves an unconstrained convex subproblem and a quadratic programming with l0 constraint. We use BB method to solve the unconstrained convex subproblem. For the quadratic programming subproblem, we obtain its exact optimal solution directly. Furthermore, we prove the convergence of SALM-BB, under certain assumptions. Finally, we implement SALM-BB on the l0 sparse logistic regression problem based on the simulation data and the data of University of California, Irvine, Machine Learning Repository. Numerical results show that SALM-BB outperforms the well-known first-order solver SLEP in terms of lower average logistic loss and lower misclassification error rate.

Key words: sparse logistic regression, splitting augmented Lagrange method, feature selection

中图分类号: