Abstract
We consider optimal decision-making problems in an uncertain environment. In particular, we consider the case in which the distribution of the input is unknown, yet there is some historical data drawn from the distribution. In this paper, we propose a new type of distributionally robust optimization model called the likelihood robust optimization (LRO) model for this class of problems. In contrast to previous work on distributionally robust optimization that focuses on certain parameters (e.g., mean, variance, etc.) of the input distribution, we exploit the historical data and define the accessible distribution set to contain only those distributions that make the observed data achieve a certain level of likelihood. Then we formulate the targeting problem as one of optimizing the expected value of the objective function under the worst-case distribution in that set. Our model avoids the over-conservativeness of some prior robust approaches by ruling out unrealistic distributions while maintaining robustness of the solution for any statistically likely outcomes. We present statistical analyses of our model using Bayesian statistics and empirical likelihood theory. Specifically, we prove the asymptotic behavior of our distribution set and establish the relationship between our model and other distributionally robust models. To test the performance of our model, we apply it to the newsvendor problem and the portfolio selection problem. The test results show that the solutions of our model indeed have desirable performance.
Similar content being viewed by others
Notes
We note that there is also a vast literature on robust optimization where the worst-case parameter is chosen for each decision made. However, the robust optimization is based on a slightly different philosophy than the distributionally robust optimization and is usually more conservative. It can also be viewed as a special case of the distributionally robust optimization where the distribution set only contains singleton distributions. In view of this, we choose not to include a detailed discussion of this literature in the main text but refer the readers to Ben-Tal et al. (2009) and Bertsimas et al. (2011) for comprehensive reviews.
References
Ben-Tal A, den Hertog D, De Waegenaere A, Melenberg B, Rennen G (2013) Robust solutions of optimization problems affected by uncertain probabilities. Manage Sci 59(2):341–357
Ben-Tal A, El Ghaoui L, Nemirovski A (2009) Robust optimization. Princeton Series in Applied Mathematics. Princeton University Press, Princeton
Ben-Tal A, Nemirovski A (1999) Robust solutions of uncertain linear programs. Oper Res Lett 25(1):1–13
Bertsimas D, Thiele A (2006) Robust and data-driven optimization: modern decision-making under uncertainty. In: Tutorial on Operations Research, INFORMS
Bertsimas D, Brown D, Caramanis C (2011) Theory and applications of robust optimization. SIAM Rev 53(3):464–501
Bertsimas D, Gupta V, Kallus N (2013) Data-driven robust optimization. arXiv:1401.0212
Bertsimas D, Sim M (2004) The price of robustness. Oper Res 52(1):35–53
Bertsimas D, Thiele A (2006) A robust optimization approach to inventory theory. Oper Res 54(1):150–168
Boyd S, Vandenberghe L (2004) Convex optimization. Cambridge University Press, Cambridge
Calafiore G, El Ghaoui L (2006) On distributionally robust chance-constrained linear programs. J Optim Theory Appl 130(1):1–22
Chen X, Sim M, Sun P (2007) A robust optimization perspective on stochastic programming. Oper Res 55(6):1058–1071
Delage E (2009) Distributionally robust optimization in context of data-driven problems. PhD thesis
Delage E, Ye Y (2008) Distributionally robust optimization under moment uncertainty with application to data-driven problems. Oper Res 58(3):595–612
Dupaĉová J (1987) The minimax approach to stochastic programming and an illustrative application. Stochastics 20(1):73–88
Gabrel V, Murat C, Thiele A (2014) Recent advances in robust optimization: An overview. Eur J Oper Res 235(3):471–483
Gallego G, Moon I (1993) The distribution free newsboy problem: Review and extension. J Oper Res Soc 44(8):825–834
Gelman A, Carlin J, Stern HS, Rubin DB (1995) Bayesian data analysis. Chapman and Hall Press, New York
Iyengar G (2005) Robust dynamic programming. Math Oper Res 30(2):257–280
Johnson N, Kotz S, Balakrishnan N (1994) Continuous univariate distributions, vol. 1. Wiley Series in Probability and Statistics
Khouja M (1999) The single period newsvendor problem: literature review and suggestions for future research. Omega 27:537–553
Luenberger D (1997) Investment science. Oxford University Press, Oxford
Mason D, Schuenemeyer J (1983) A modified \(\text{ K }\)olmogorov-\(\text{ S }\)mirnov test sensitive to tail alternatives. Ann Stat 11(3):933–946
Nilim A, El Ghaoui L (2005) Robust control of Markov decision processes with uncertain transition matrices. Oper Res 53(5):780–798
Owen A (2001) Empirical likelihood. Chapman and Hall Press, London
Pardo L (2005) Statistical inference based on divergence measures. Chapman and Hall Press, London
Perakis G, Roels G (2008) Regret in the newsvendor model with partial information. Oper Res 56(1):188–203
Popescu I (2007) Robust mean-covariance solutions for stochastic optimization. Oper Res 55(1):98–112
Scarf H (1958) A min-max solution of an inventory problem. In: Arrow K, Karlin S, Scarf H (eds) Studies in the mathematical theory of inventory and production. Stanford University Press, California, pp 201–209
Scarf H (1959) Bayes solutions of the statistical inventory problem. Ann Math Stat 30(2):490–508
Shapiro A, Dentcheva D, Ruszczynski A (2009) Lectures on stochastic programming: modeling and theory. MPS-SIAM Series on Optimization
Shapiro A, Kleywegt AJ (2002) Minimax analysis of stochastic programs. Optim Methods Softw 17:523–542
So AMS, Zhang J, Ye Y (2009) Stochastic combinatorial optimization with controllable risk aversion level. Math Oper Res 34(3):522–537
Wassaman L (2009) All of statistics: a concise course in statistical inference. Springer Texts in Statistics, New York
Yue J, Chen B, Wang M (2006) Expected value of distribution information for the newsvendor problem. Oper Res 54(6):1128–1136
Ẑáĉková J (1966) On minimax solutions of stochastic linear programming problems. Casopis pro Pêstování Matematiky 91:423–430
Zhu Z, Zhang J, Ye Y (2013) Newsvendor optimization with limited distribution information. Optim Methods Softw 28(3):640–667
Acknowledgments
The authors thank Dongdong Ge and Zhisu Zhu for valuable insights and discussions. The authors also thank Erick Delage for insightful discussions and for sharing useful codes for the numerical experiments. The research of the first author is supported by the National Science Foundation (NSF) under research Grant CMMI-1434541.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
Proof of Proposition 1
First, we show that problem (4) is a convex optimization problem. To see this, we first write (4) in a slightly different but equivalent way:
Note that because the objective is to maximize, (15) is an equivalent representation of (4). It is easy to see that when \(h(\varvec{x}, \varvec{\xi })\) is concave in \(\varvec{x}\), the constraints are all convex constraints. Now we only need to verify that the objective function is jointly concave. Note that the first two terms of the objective function are linear, therefore, we only need to show that the function \(-N\lambda \log {\lambda } + \lambda \sum _{i=1}^n N_i \log {y_i}\) is jointly concave in \(\lambda \) and \(\varvec{y}\). We consider the Hessian matrix of the objective function with respect to \(\lambda \) and \(\varvec{y}\):
Now consider a diagonal matrix A with diagonal elements \(\lambda , y_1,\ldots ,y_n\), then we have
which is a diagonal dominant matrix with all negative diagonal elements. Thus \(A^THA\) is negative semidefinite, which further implies that H is negative semidefinite. Thus the objective function is a concave function and the optimization problem is a convex optimization problem.
Finally, we prove that the worst-case distribution satisfies:
We consider the derivative of the Lagrangian function \(L(\varvec{p},\lambda , \mu )\) with respect to \(p_i\). By the KKT conditions, we must have \(p_i\nabla _{p_i} L(\varvec{p}, \lambda , \mu ) = 0\), which implies that \(p_i(h(\varvec{x}^*, \varvec{\xi }_i) -\mu ^*) = \lambda ^* N_i\). Also, we note that in (4), since there is a term of \(\log {(h(\varvec{x}, \varvec{\xi }_i) -\mu )}\), therefore, \(h(\varvec{x}^*, \varvec{\xi }_i) -\mu ^*\ne 0\). Thus Proposition 1 is proved. \(\square \)
Rights and permissions
About this article
Cite this article
Wang, Z., Glynn, P.W. & Ye, Y. Likelihood robust optimization for data-driven problems. Comput Manag Sci 13, 241–261 (2016). https://doi.org/10.1007/s10287-015-0240-3
Received:
Accepted:
Published:
Issue date:
DOI: https://doi.org/10.1007/s10287-015-0240-3