Locally robust models for optimization under tail-based data imbalance
In "Seminars and talks"

Speakers

Karthyek Murthy
Karthyek Murthy

Assistant Professor, Singapore University of Technology & Design

Karthyek Murthy serves as an Assistant Professor in Singapore University of Technology & Design. His research interests lie in data-driven operations research. Prior to joining SUTD, he was a postdoctoral researcher in Columbia University IEOR department. His research has been recognised with 2021 INFORMS Junior Faculty Forum (JFIG) Paper competition (Third place), 2019 WSC Best Paper Award, and IBM PhD fellowship. Karthyek serves as an Associate Editor for the INFORMS journal Stochastic Systems and as a judge for the INFORMS Nicholson student paper competition.Karthyek Murthy serves as an Assistant Professor in Singapore University of Technology & Design. His research interests lie in data-driven operations research. Prior to joining SUTD, he was a postdoctoral researcher in Columbia University IEOR department. His research has been recognised with 2021 INFORMS Junior Faculty Forum (JFIG) Paper competition (Third place), 2019 WSC Best Paper Award, and IBM PhD fellowship. Karthyek serves as an Associate Editor for the INFORMS journal Stochastic Systems and as a judge for the INFORMS Nicholson student paper competition.


Date:
Friday, 17 March 2023
Time:
10:00 am - 11:30 am
Venue:
Institute of Data Science
Innovation 4.0 I4-01-03 (Level 1 Seminar Room)
3 Research Link
Singapore 117602 (Map)

Abstract

Many problems in operations and risk management require the familiar “estimate, then optimize” workflow involving a model estimation from data in the first step before plugging in the trained model to solve various optimization tasks. In this talk, we first give the ingredients for constructing locally robust optimization formulations in which the first step model estimation has no effect, locally, on the optimal solutions. Then delving specifically into optimization problems affected by tail-based data imbalance, we show that this local sensitivity translates to improved out-of-sample performance freed from the first-order impact of model errors caused by model selection and misspecification biases that are especially difficult to avoid when performing estimation with imbalanced data.  

 

The key ingredient in achieving this local robustness is a novel debiasing procedure that adds a non-parametric bias correction term to the objective. The debiased objective retains convexity, and the imputation of the correction term relies only on a non-restrictive large deviations behavior conducive for transferring knowledge from representative data-rich regions to the datascarce tail regions suffering from imbalance. The bias correction gets determined by the extent of model error in the estimation step and the specifics of the stochastic program in the optimization step, thereby serving as a scalable “smart-correction” step bridging the disparate goals in estimation and optimization. Besides showing the empirical effectiveness of the proposed formulation in real datasets, the numerical experiments bring out the utility of locally robust solutions in tackling model errors and shifts in distribution between training and deployment.