Uncertainty in Artificial Intelligence
First Name   Last Name   Password   Forgot Password   Log in!
    Proceedings   Proceeding details   Article details         Authors         Search    
High-dimensional Joint Sparsity Random Effects Model for Multi-task Learning
Krishnakumar Balasubramanian, Kai Yu, Tong Zhang
Joint sparsity regularization in multi-task learning has attracted much attention in recent years. The traditional convex formulation employs the group Lasso relaxation to achieve joint sparsity across tasks. Although this approach leads to a simple convex formulation, it suffers from several issues due to the looseness of the relaxation. To remedy this problem, we view jointly sparse multi-task learning as a specialized random effects model, and derive a convex relaxation approach that involves two steps. The first step learns the covariance matrix of the coefficients using a convex formulation which we refer to as sparse covariance coding; the second step solves a ridge regression problem with a sparse quadratic regularizer based on the covariance matrix obtained in the first step. It is shown that this approach produces an asymptotically optimal quadratic regularizer in the multitask learning setting when the number of tasks approaches infinity. Experimental results demonstrate that the convex formulation obtained via the proposed model significantly outperforms group Lasso (and related multi-stage formulations
Pages: 42-51
PS Link:
PDF Link: /papers/13/p42-balasubramanian.pdf
AUTHOR = "Krishnakumar Balasubramanian and Kai Yu and Tong Zhang",
TITLE = "High-dimensional Joint Sparsity Random Effects Model for Multi-task Learning",
BOOKTITLE = "Proceedings of the Twenty-Ninth Conference Annual Conference on Uncertainty in Artificial Intelligence (UAI-13)",
ADDRESS = "Corvallis, Oregon",
YEAR = "2013",
PAGES = "42--51"

hosted by DSL   •   site info   •   help