610 likes | 744 Views
Model Selection via Bilevel Optimization. Kristin P. Bennett, Jing Hu, Xiaoyun Ji, Gautam Kunapuli and Jong-Shi Pang Department of Mathematical Sciences Rensselaer Polytechnic Institute Troy, NY. Convex Machine Learning.
E N D
Model Selection via Bilevel Optimization Kristin P. Bennett, Jing Hu, Xiaoyun Ji, Gautam Kunapuli and Jong-Shi Pang Department of Mathematical Sciences Rensselaer Polytechnic Institute Troy, NY
Convex Machine Learning Convex optimization approaches to machine learning has been major obsession of machine learning for last ten years. But are the problems really convex?
Outline • The myth of convex machine learning • Bilevel Programming Model Selection • Regression • Classification • Extensions to other machine learning tasks • Discussion
Modeler’s Choices Data Function Loss/Regularization CONVEX! Optimization Algorithm w
Many Hidden Choices • Data: • Variable Selection • Scaling • Feature Construction • Missing Data • Outlier removal • Function Family: • linear, kernel (introduces kernel parameters) • Optimization model • loss function • regularization • Parameters/Constraints
Data Function Loss/Regularization Cross-Validation Strategy Generalization Error Cross-Validation: C,ε, [X,y] Optimization Algorithm w NONCONVEX
How does modeler make choices? • Best training set error • Experience/policy • Estimate of generalization error • Cross-validation • Bounds • Optimize generalization error estimate • Fiddle around. • Grid Search • Gradient methods • Bilevel Programming
ε C CV via Grid Search For every C, ε • For every validation set, Solve model on corresp. training set, and to estimate loss for •Estimate generalization error for C, ε Return best values for C,ε Make final model using C,ε
CV as Continuous Optimization Problem • Bilevel Program for T folds • Prior Approaches: Golub et al., 1979, Generalized Cross-Validation for one parameter in Ridge Regression Outer-level validation problem T inner-level training problems
Benefit: More Design Variables Add feature box constraint: in the inner-level problems.
Key Transformation • KKT for the inner level training problems are necessary and sufficient • Replace lower level problems by their KKT Conditions • Problem becomes an Mathematical Programming Problem with Equilibrium Constraints (MPEC)
Bilevel Problem as MPEC Replace T inner-level problems with corresponding optimality conditions
MPEC to NLP via Inexact Cross Validation Relax “hard” equilibrium constraints to “soft” inexact constraints tol is some user-defined tolerance.
Solvers Strategy: Proof of concept using nonlinear general purpose solvers from NEOS on NLP FILTER, SNOPT Sequential Quadratic Programming Methods FILTER results almost always better. Many possible alternatives: Integer Programming Branch and Bound Lagrangian Relaxations
Computational Experiments: DATA Synthetic • (5,10,15)-D Data with Gaussian and Laplacian noise and (3,7,10) relevant features. • NLP: 3-fold CV • Results: 30 to 90 train, 1000 test points, 10 trials QSAR/Drug Design • 4 datasets, 600+ dimensions reduced to 25 top principal components. NLP: 5-fold CV • Results: 40 – 100 train, rest test, 20 trials
Cross-validation Methods Compared • Unconstrained Grid: Try 3 values each for C,ε • Constrained Grid: Try 3 values each for C, ε, and {0, 1} for each component of • Bilevel/FILTER: Nonlinear program solved using off-the-shelf SQP algorithm via NEOS
Classification Cross Validation Given sample data from two classes. Find classification function that minimizes out-of-sample estimate of classification error 1 -1
Lower level - SVM • Define parallel planes • Minimize points on wrong side • Maximize margin of separation
Lower Level Loss Function: Hinge Loss Measures distance of points that violate the appropriate hyperplane constraints,
Outer-level Loss Functions • Misclassification Minimization Loss (MM) • Loss function used in classical CV • Loss = 1, if validation pt misclassified, 0, otherwise(computed using step function, ) • Hinge Loss (HL) • Both inner and outer levels use same loss function • Loss = distance from(computed using max function, )
Hinge Loss is Convex Approx. of Misclassification Minimization Loss
Hinge Loss Bilevel Program (BilevelHL) • Replace max in outer level objective with convex constraints • Replace inner-level problems with KKT conditions
Misclassification Min. Bilevel Program (BilevelMM) Misclassifications are counted using the step function, defined component wise for a n-vector as
The Step Function Mangasarian (1994) showed that and that any solution, , to the LP satisfies
Misclassifications in the Validation Set • Validation point misclassified when the sign of is negative i.e., • This can be recast for all validation points (within the t-th fold) as
Misclassification Minimization Bilevel Program (revisited) Outer-level average misclassification minimization Inner-level problems to determine misclassified validation points Inner-level training problems
Inexact Cross Validation NLP • Both BilevelHL and BilevelMM MPECs are transformed to NLP by relaxing equilibrium constraints (inexact CV) • Solved using FILTER on NEOS • These are compared with classical cross validation: unconstrained and constrained grid.
Experiments: Data sets 3-fold cross validation for model selection Average results for 20 train test splits
Progress • Cross Validation is a bilevel problem solvable by continuous optimization methods • Off-the-shelf NLP algorithm – FILTER solved classification and regression • Bilevel Optimization extendable to many Machine Learning problems
Extending Bilevel Approach to other Machine Learning Problems • Kernel Classification/Regression • Variable Selection/Scaling • Multi-task Learning • Semi-supervised Learning • Generative methods
Semi-supervised Learning • Have labeled data, and unlabeled data • Treat missing labels, , as design variables in the outer level • Lower level problems are still convex
Semi-supervised Regression Outer level minimizes error on labeled data to find optimal parameters and labels -insensitive loss on labeled data in inner level -insensitive loss on unlabeled data in inner level Inner level regularization