Learning With Subquadratic Regularization : A Primal-Dual Approach

Learning With Subquadratic Regularization : A Primal-Dual Approach

Raman Sankaran, Francis Bach, Chiranjib Bhattacharyya

Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence
Main track. Pages 1963-1969. https://doi.org/10.24963/ijcai.2020/272

Subquadratic norms have been studied recently in the context of structured sparsity, which has been shown to be more beneficial than conventional regularizers in applications such as image denoising, compressed sensing, banded covariance estimation, etc. While existing works have been successful in learning structured sparse models such as trees, graphs, their associated optimization procedures have been inefficient because of hard-to-evaluate proximal operators of the norms. In this paper, we study the computational aspects of learning with subquadratic norms in a general setup. Our main contributions are two proximal-operator based algorithms ADMM-η and CP-η, which generically apply to these learning problems with convex loss functions, and achieve a proven rate of convergence of O(1/T) after T iterations. These algorithms are derived in a primal-dual framework, which have not been examined for subquadratic norms. We illustrate the efficiency of the algorithms developed in the context of tree-structured sparsity, where they comprehensively outperform relevant baselines.
Keywords:
Machine Learning: Feature Selection; Learning Sparse Models
Data Mining: Feature Extraction, Selection and Dimensionality Reduction