Adaptive Distributionally Robust Optimization Bertsimas
Adaptive Distributionally Robust Optimization Bertsimas
Melvyn Sim
NUS Business School, National University of Singapore
Meilin Zhang
SUSS Business School, Singapore University of Social Science
We develop a modular and tractable framework for solving an adaptive distributionally robust linear opti-
mization problem, where we minimize the worst-case expected cost over an ambiguity set of probability
distributions. The adaptive distrbutaionally robust optimization framework caters for dynamic decision mak-
ing, where decisions adapt to the uncertain outcomes as they unfold in stages. For tractability considerations,
we focus on a class of second-order conic (SOC) representable ambiguity set, though our results can easily
be extended to more general conic representations. We show that the adaptive distributionally robust linear
optimization problem can be formulated as a classical robust optimization problem. To obtain a tractable
formulation, we approximate the adaptive distributionally robust optimization problem using linear decision
rule (LDR) techniques. More interestingly, by incorporating the primary and auxiliary random variables of
the lifted ambiguity set in the LDR approximation, we can significantly improve the solutions, and for a
class of adaptive distributionally robust optimization problems, exact solutions can also be obtained. Using
the new LDR approximation, we can transform the distributionally adaptive robust optimization problem to
a classical robust optimization problem with an SOC representable uncertainty set. Finally, to demonstrate
the potential for solving management decision problems, we develop an algebraic modeling package and
illustrate how it can be used to facilitate modeling and obtain high quality solutions for medical appointment
scheduling and inventory management problems.
1. Introduction
Addressing uncertainty in many real world optimization problems has often lead to computation-
ally intractable models. As a result, uncertainty is often ignored in optimization models and this
may lead to poor or even unacceptable decisions when implementing them in practice. We charac-
terize uncertainty as risk, whenever the probability distribution is known, or otherwise as ambiguity
(Knight 1921). Traditionally, mathematical optimization models such as stochastic programming
(see, for instance, Birge and Louveaux 1997, Ruszczyński and Shapiro 2003) are based on the
paradigm of risk and they do not incorporate ambiguity in their decision criteria for optimization.
However, with the growing importance of ambiguity in decision making (see, for instance, Ells-
berg 1961, Hsu et al. 2005), research on ambiguity has garnered considerable interest in various
1
Bertsimas et al.: Adaptive Distributionally Robust Optimization
2 Article submitted;
fields including economics, mathematical finance and management science. In particular, robust
optimization is a relatively new approach that deals with ambiguity in mathematical optimization
problems. In classical robust optimization, uncertainty is described by a distribution free uncer-
tainty set, which is typically a conic representable bounded convex set (see, for instance, El Ghaoui
and Lebret 1997, El Ghaoui et al. 1998, Ben-Tal and Nemirovski 1998, 1999, 2000, Bertsimas
and Sim 2004, Bertsimas and Brown 2009, Bertsimas et al. 2011). The key advantage of a robust
optimization model is its computational tractability and it has been successful in providing com-
putationally scalable solutions for a wide variety of management inspired optimization problems.
In evaluating preferences over risk and ambiguity, Gilboa and Schmeidler (1989) propose a deci-
sion criterion that is based on the worst-case expected utility or disutility over an ambiguity set of
probability distributions. Scarf (1958) is first to study a single-product Newsvendor problem where
the precise demand distribution is unknown but is only characterized by its mean and variance.
Subsequently, such models have been extended to minimax stochastic optimization models (see,
for instance, Žáčková 1966, Dupacova 1987, Breton and EI Hachem 1995, Shapiro and Kleywegt
2002, Shapiro and Ahmed 2004), and recently to distributionally robust optimization models (see,
for instance, Chen et al. 2007, Chen and Sim 2009, Popescu 2007, Delage and Ye 2010, Xu and
Mannor 2012). In terms of tractable formulations for a wide variety of single stage convex opti-
mization problems, Wiesemann et al. (2014) propose a broad class of ambiguity sets where the
family of probability distributions are characterized by conic representable expectation constraints
and nested conic representable confidence sets.
Dynamic optimization models, where decisions adapt to the uncertain outcomes as they unfold in
stages, typically suffer from the “curse of dimensionality” and are computationally intractable (see,
for instance, Shapiro and Nemirovski 2005, Dyer and Stougie 2006, Ben-Tal et al. 2004). To yield
tractable models, linear decision rule (LDR), where adaptive decisions are restricted to affine func-
tions of the uncertain parameters, was proposed in the early literature of stochastic programming.
However, the technique had been abandoned due to suboptimality of the decision rule (see Garstka
and Wets 1974). Nevertheless, LDR approximation has been revived by Ben-Tal et al. (2004) in
their seminal work on adaptive robust optimization1 . Subsequently, Bertsimas et al. (2010) establish
the optimality of LDR approximation in some important classes of adaptive robust optimization
problems. Chen and Zhang (2009) further improve the LDR approximation by extending the affine
dependency to the auxiliary variables associated with the support set. For solving adaptive distri-
butionally robust optimization problems, Chen et al. (2007) propose tractable formulations using
LDR approximation techniques. Henceforth, variants of piecewise-linear decision rule approxima-
tion have been proposed to improve the approximation while maintaining the tractability of the
1
Note that we prefer the term “adaptive” over “adjustable” as used in Ben-Tal et al. (2004).
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 3
adaptive distributionally robust optimization models. Such approaches include the deflected and
segregated LDR approximation of Chen et al. (2008), the truncated LDR approximation of See
and Sim (2009), and the bideflected and (generalized) segregated LDR approximation of Goh and
Sim (2010). Interestingly, there is also a revival of using LDR approximation for solving multistage
stochastic optimization problems (Kuhn et al. 2011).
For broader impact, a general purpose optimization framework to address a wide variety of
optimization problems should be implementable in software packages where reliable solutions can be
obtained with reasonable computational effort. Compared to deterministic optimization frameworks
such as linear optimization, classical optimization frameworks that deal with uncertainty such as
stochastic optimization and dynamic optimization have been less successful. Software packages that
facilitate robust and distributionally robust optimization modeling have begun to surface in recent
years. Existing toolboxes include YALMIP (Löberg 2012), AIMMS (http://www.aimms.com/),
ROME (Goh and Sim 2009, 2010). and JuMPeR ( http://jumper.readthedocs.org). Of these,
ROME, AIMMS and JuMPeR have also incorporated LDR approximation.
Our contributions to this paper are as follows:
1. We propose a tractable and scalable framework for solving an adaptive distributionally robust
linear optimization problem, where we minimize the worst-case expected cost over a second-order
conic (SOC) representable ambiguity set. We show that adaptive distributionally robust linear
optimization problem can be formulated as a classical robust optimization problem.
2. To obtain tractable formulation, we approximate the adaptive distributionally robust lin-
ear optimization problem using LDR techniques. Depending on the choice of ambiguity set, the
resulting framework is either a linear optimization problem or a second order conic optimization
problem (SOCP), which can be solved efficiently by general purpose commercial grade solvers such
as CPLEX and Gurobi.
3. We show that we can significantly improve the LDR approximation by incorporating the
auxiliary random variable associated with the lifted ambiguity set. This approach outperforms more
sophisticated decision rule approximations developed in Chen and Zhang (2009), Chen et al. (2008),
See and Sim (2009), Goh and Sim (2010). Using the new LDR approximation, we can transform the
adaptive distributionally robust optimization problem to a classical robust optimization problem
with an SOC representable uncertainty set.
4. We demonstrate our approach for addressing a medical appointment scheduling problem as
well as a multiperiod inventory control problem. In these problems, we also show that by incor-
porating partial cross moments information in the ambiguity set, we can significantly improve the
solutions over alternatives found in recent literature where the ambiguity set is only characterized
by marginal moments.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
4 Article submitted;
Notations. We use [N ], N ∈ N to denote the set of running indices, {1, . . . , N }. We generally use
bold faced characters such as x ∈ RN and A ∈ RM ×N to represent vectors and matrices, respectively
and [x]i or xi to denote the ith element of the vector x. We use (x)+ to denote max{x, 0}. Special
vectors include 0, 1 and ei which are respectively the vector of zeros, the vector of ones and the
standard unit basis vector. We denote RN,M as the space of all measurable functions from RN
to RM that are bounded on compact sets. We use P0 (RI ) to represent the set of all probability
distributions on RI . A random variable, z̃ is denoted with a tilde sign and we use z̃ ∼ P, P ∈ P0 (RI )
to define z̃ as an I dimensional random variable with probability distribution P. We denote EP [·]
as the expectation over the probability distribution P. For a set W ⊆ RI , P [z̃ ∈ W ] represents the
probability of z̃ being in the set W evaluated on the distribution P.
with A0 , A1 , ..., AI1 ∈ RM ×N1 and b0 , b1 , ..., bI1 ∈ RM . The matrix B ∈ RM ×N2 , also known in
stochastic programming as the recourse matrix, and the vector d ∈ RN2 are constants, which corre-
spond to the stochastic programming format known as fixed recourse. Note that Problem (1) may
not aways be feasible and, as in the case of complete recourse, the recourse matrix can influence
the feasibility of the second stage problem.
Definition 1. The second stage problem (1) has complete recourse if and only if for any t ∈ RM ,
there exists y ∈ RN2 such that By ≥ t.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 5
Complete recourse is a strong sufficient condition that guarantees the feasibility of the second
stage problem for all x ∈ RN1 and z ∈ RI1 . Many real-life stochastic programming models, including
the newsvendor problem and its variants are modeled as complete recourse problems, which ensure
that no outcome can produce infeasible results. However, there are also problems that would
generally not satisfy complete recourse, such as a production planning problem where a manager
determines a production plan today to satisfy all uncertain demands for tomorrow instead of
incurring penalty. For more information, we refer interested readers to Birge and Louveaux (1997).
Typically, a weaker condition is assumed in stochastic programming to ensure that the second
stage problem is essentially feasible.
Definition 2. The second stage problem (1) has relatively complete recourse if and only if the
problem is feasible for all x ∈ X and z ∈ W .
Given an ambiguity set of probability distributions, F , the second stage cost is evaluated based
on the worst-case expectation over the ambiguity set given by
Corresponding, the here-and-now decision, x is determined by minimizing the sum of the deter-
ministic first stage cost and the worst-case expected second stage cost over the ambiguity set as
follows:
Z ∗ = min c0 x + β(x)
(3)
s.t. x ∈ X.
of the ambiguity set. In particular the ambiguity set based on information of moments, notwith-
standing its popularity, may not necessarily yield tractable distributionally robust counterparts.
We propose an SOC representable ambiguity set where we restrict only to SOC representation.
For generalization to the ambiguity set of Wiesemann et al. (2014), we refer interested readers to
e-companion EC.5.
Definition 3. An SOC ambiguity set, F is an ambiguity set of probability distributions that
can be expressed as
z̃ ∼ P
EP [Gz̃] = µ
F= P ∈ P0 (RI1 ) (4)
EP [gi (z̃)] ≤ σi ∀i ∈ [I2 ]
P[z̃ ∈ W ] = 1
with parameters G ∈ RL1 ×I1 , µ ∈ RL1 , σ ∈ RI2 , support set W ∈ RI1 and functions gi ∈ RI1 ,1 ,
i ∈ [I2 ]. The support set W is an SOC representable set and the epigraph of each gi , i ∈ [I2 ],
The formulation of SOC representable functions is a process that can be automated in an alge-
braic modeling software package. For more information, we refer interested readers to Ben-Tal and
Nemirovski (2001a) for an excellent reference on the algebra of SOC representable functions. The
SOC ambiguity set provides useful and interesting characterization of distributions including:
• Bounds on mean values: EP [z̃] ∈ [µ, µ].
• Upper bound on absolute deviation: EP [|f 0 z + h|] ≤ σ, for some (f , h) ∈ RI1 +1 .
• Upper bound on variance: EP [(f 0 z + h)2 ] ≤ σ, for some (f , h) ∈ RI1 +1 .
• Upper bound on p-ordered deviation: EP [(|f 0 z + h|)p ] ≤ σ, for some (f , h) ∈ RI1 +1 and
some rational p ≥ 1.
• Upper bound on semi-variance: EP [((f 0 z + h)+ )2 ] ≤ σ, for some (f , h) ∈ RI1 +1 .
• Approximate upper bound on entropy: EP [exp(f 0 z)] ≤ σ, for some f ∈ RI1 . We refer
readers to Ben-Tal and Nemirovski (2001a) for the approximate SOC representation.
• Upper bound on convex piecewise linear function: EP [maxp∈[P ] {f 0p z + hp }] ≤ σ, for
some (f p , hp ) ∈ RI1 +1 , p ∈ [P ].
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 7
An important class of ambiguity set that could be modeled in the ambiguity set of Wiesemann
et al. (2014) but not ours is the cross moment ambiguity set as follows,
z̃ ∼ P
EP [z̃] = µ
FCM = P ∈ P0 RI1
.
EP [(z̃ − µ)(z̃ − µ)0 ] Σ
P[z̃ ∈ W ] = 1
Hence, as a conservative approximation of the cross moment ambiguity set, we propose the partial
cross moment ambiguity set, which is an SOC ambiguity set as follows:
z̃ ∼ P
[z̃] = µ
I1
E P
FP CM = P ∈ P0 R ,
EP [(f 0k (z̃ − µ))2 ] ≤ f 0k Σf k ∀k ∈ [K]
P[z̃ ∈ W ] = 1
for some choice of parameters f 1 , . . . , f K ∈ RI1 . Observe that the approximation never deteriorates
with addition of new vectors, f k , k > K. In our applications to inventory control and appointment
scheduling problems, we will demonstrate how the partial cross moment ambiguity set can yield
tractable models and provide far less conservative solutions than those obtained from the marginal
moment ambiguity set, an ambiguity set that does not consider cross moment information.
As in Wiesemann et al. (2014), we also define the lifted ambiguity set, G that encompasses the
primary random variable z̃ and the lifted or auxiliary random variable ũ as follows:
(z̃, ũ) ∼ P
[Gz̃] = µ
I1 I2
E P
G = P ∈ P0 R × R , (6)
EP [ũ] ≤ σ
P (z̃, ũ) ∈ W̄ = 1
with g(z) = (g1 (z), . . . , gI2 (z)). Observe that the lifted ambiguity set has only linear expectation
constraints and that the corresponding lifted support set is SOC representable.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
8 Article submitted;
Observe that
EPG [ũ] = EPG [g(z̃)] ≤ σ
and
PG (z̃, ũ) ∈ W̄ = 1.
Q
Hence, F ⊆ z̃ G.
Assumption 1 (Slater’s Condition). Suppose the lifted support set has the following explicit
formulation:
W̄ = (z, u) ∈ RI1 × RI2 | ∃v ∈ RI3 : Cz + Du + Ev K h ,
(8)
with C ∈ RL2 ×I1 , D ∈ RL2 ×I2 , E ∈ RL2 ×I3 , h ∈ RL2 and K ⊆ RL2 is a Cartesian product of second-
order cones. There exists z † , u† , v † such that
Gz † = µ
u† < σ
Cz † + Du† + Ev † ≺K h.
Theorem 1. Suppose the second stage problem (1) has relatively complete recourse, then under
Assumption 1, Problem (2) is equivalent to the following robust optimization problem,
β(x) = min r + s0 µ + t0 σ
s.t. r + s0 (Gz) + t0 u ≥ Q(x, z) ∀(z, u) ∈ W̄
(9)
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 9
0
P = p ∈ RM
+ : B p=d .
Unfortunately, despite its simplicity, the LDR approximation may lead to infeasibility even if
the problem has complete recourse (see, for instance, Chen et al. 2008, Goh and Sim 2009). As an
illustration, we consider the following complete recourse problem,
β = min sup EP [y(z̃)]
P∈F
s.t. y(z) ≥ z ∀z ∈ R
(13)
y(z) ≥ −z ∀z ∈ R
y ∈ R1,1
where
F = P ∈ P0 (R) EP [|z̃ |] ≤ 1 .
Clearly, y(z) = |z | is the optimal decision rule that yields β = 1. However, under the regular LDR
approximation, we require y(z) = y0 + y1 z for some y0 , y1 ∈ R, which will be infeasible in the
following set of semi-infinite constraints,
y0 + y1 z ≥ z ∀z ∈ R
(14)
y0 + y1 z ≥ −z ∀z ∈ R.
where
∃y 0 , y 1i , y 2j ∈ RN , ∀i ∈ [I1 ], j ∈ [I2 ] :
N
L̄ = y ∈ RI1 +I2 ,N
X X
1 2
y(z, u) = y + 0
y i zi + y j uj .
i∈[I ] j∈[I ]
1 2
Theorem 2.
β(x) ≤ βE (x) ≤ βL (x)
Proof. Given the flexibility of the new LDR approximation, it follows trivially that βE (x) ≤ βL (x).
Let y ∗ be an optimal decision rule of Problem (15) and we define the decision rule y † ∈ RI1 ,N2 such
that
y † (z) = y ∗ (z, g(z)).
β(x) ≤ sup EP [d0 y † (z̃)] = sup EP [d0 y ∗ (z̃, g(z̃))] ≤ sup EP [d0 y ∗ (z̃, ũ)] = βE (x).
P∈F P∈G P∈G
Proposition 2. Problem (15) is equivalent to the affinely adaptive robust optimization problem
of Ben-Tal et al. (2004) as follows,
βE (x) = min r + s0 µ + t0 σ
s.t. r + s0 (Gz) + t0 u ≥ d0 y(z, u) ∀(z, u) ∈ W̄
A(z)x + By(z, u) ≥ b(z) ∀(z, u) ∈ W̄
(16)
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
y ∈ L̄N2 .
Theorem 3. Suppose Problem (11) has complete recourse and the objective is bounded from
below. For any ambiguity set, F such that
EP [|z̃i |] < ∞ ∀P ∈ F
there exists a lifted ambiguity set, G whose corresponding LDR is feasible in Problem (15).
Bertsimas et al.: Adaptive Distributionally Robust Optimization
12 Article submitted;
Remarks: Note that the class of ambiguity set depicted in Theorem 3 encompasses any random
variable with finite deviation, i.e., EP [|z̃i − µi |pi ] < ∞ for some µi ∈ R, pi ≥ 1, i ∈ [I1 ], since
1/pi
EP [|z̃i |] ≤ EP [|z̃i − µi |] + |µi | ≤ (EP [|z̃i − µi |pi ]) + |µi | < ∞.
More interestingly, we show in the following result that the new LDR approximation can attain the
optimal objective values for a class of adaptive distributionally robust linear optimization problems.
Theorem 4. Suppose Problem (11) is a complete recourse problem with only one second stage
decision variable, i.e. N2 = 1, then
β(x) = βE (x).
Remarks: Note that for complete recourse problem with N2 = 1, Problem (9) becomes a tractable
problem since the number of extreme points of the polyhedron P equals to M . Nevertheless,
notwithstanding the simplicity, we are not aware of other types of decision rules that would yield
tight results for this instance.
A natural question is whether we could extend the results of Theorem 3 and 4 to the case of
relatively complete recourse. However, this is not the case as depicted in the following negative
result even for the case of N2 = 1.
Proposition 3. There exists a relatively complete recourse problem with N2 = 1 for which Prob-
lem (15) is infeasible under any LDR that incorporates both the primary and auxiliary random
variables associated with the lifted ambiguity set.
min 0
s.t. y(z) ≥ z1 − z2 ∀z ∈ W
y(z) ≥ z2 − z1 ∀z ∈ W
(17)
y(z) ≤ z1 + z2 + 2 ∀z ∈ W
y(z) ≤ −z1 − z2 + 2 ∀z ∈ W
y ∈ R2,1
where
W = z ∈ R2
kz k∞ ≤ 1 .
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 13
We verify that Problem (17) is one with relatively complete recourse by showing that y(z) = |z1 − z2 |
is a feasible solution. Indeed, y(z) is feasible in the first two sets of constraints of Problem (17).
Moreover, for all kz k∞ ≤ 1,
|z1 − z2 | + |z1 + z2 |
= max{z1 − z2 + z1 + z2 , −z1 + z2 + z1 + z2 , z1 − z2 − z1 − z2 , −z1 + z2 − z1 − z2 }
= max{2z1 , 2z2 , −2z2 , −2z1 } ≤ 2.
Hence, for all z ∈ W , y(z) ≤ −|z1 + z2 | + 2 and it is therefore feasible in the last two sets of
constraints of Problem (17). We also note that there does not exist a feasible LDR in which y is
affinely dependent on z, i.e. y(z) = y0 + y1 z1 + y2 z2 for some y0 , y1 , y2 ∈ R. Indeed, when substituting
the extreme points of W in Problem (17), we will obtain the following set of infeasible equations,
z1 = z 2 = 1 ⇒ y0 + y1 + y2 = 0
z1 = z2 = −1 ⇒ y0 − y1 − y2 = 0
z1 = −z2 = 1 ⇒ y0 + y1 − y2 = 2
z1 = −z2 = −1 ⇒ y0 − y1 + y2 = 2.
For any lifted ambiguity set, the corresponding lifted support set W̄ is defined in (6), where the
parameter u is associated with the auxiliary random variable. Incorporating the auxiliary random
variable in the LDR, we have
y(z, u) = y0 + y 1 z + y 2 u.
Since u is unbounded from above in the lifted ambiguity set W̄ , the constraints
y(z, u) ≥ z1 − z2 ∀(z, u) ∈ W̄
y(z, u) ≤ z1 + z2 + 2 ∀(z, u) ∈ W̄
imply y 2 ≤ 0. Hence, the LDR is reduced one that is only affinely dependent on z, which, as we
have shown, is infeasible in Problem (17).
Quite surprisingly, by simply extending the LDR approximation to include the auxiliary random
variable of the lifted ambiguity set, we are able to attain the optimum objective values for the
class of complete recourse problems described in Theorem 4. In contrast, the proposed deflected
LDR (DLDR) approximations of Chen et al. (2008), Goh and Sim (2009), albeit more complex, do
not necessarily recover the optimum objective values for this class of complete recourse problems.
In e-companion EC.4, we will further demonstrate that the new LDR approximation can indeed
improve over the more sophisticated DLDR approximations.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
14 Article submitted;
Problem (18) solves for the optimal decision rule y ∈ RI1 ,N2 that minimizes the worst-case expected
objective taking into account of the information dependency requirement. Under the new LDR
approximation, we have
γE (x) = min sup EP [d0 y(z̃, ũ)]
P∈G
s.t. A(z)x + By(z, u)) ≥ b(z) ∀(z, u) ∈ W̄ (19)
yi ∈ L̄(S i , T i ) ∀i ∈ [N2 ],
where
∃y 0 , yi1 , yj2 ∈ R, ∀i ∈ S , j ∈ T :
L̄(S , T ) = y ∈ RI1 +I2 ,1
X X
0 1 2
y(z, u) = y + yi zi + yj uj
i∈S j∈T
and the subsets T i ⊆ [I2 ], i ∈ [N2 ] are consistent with the information restriction imposed by
S i ⊆ [I1 ]. In Section 5, we will illustrate how we can use this model to formulate a multi-period
inventory control problem. We also refer readers to He et al. (2017) for modeling and optimizing
multi-period repositioning decisions in vehicle sharing problems.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 15
Observe that y † (z̃) is feasible in Problem (11) and from the proof of Theorem 2, it follows that
Suppose β(x) = βE (x), which is the case for complete recourse problems and N2 = 1, there is a
tendency to infer the optimality of y † vis-à-vis y ∗ so that
d0 y † (z) = d0 y ∗ (z) ∀z ∈ W .
However, this is not the case and we will demonstrate this fallacy in the following simple example.
Consider the following complete recourse problem,
β = min sup EP [y(z̃)]
P∈F
s.t. y(z) ≥ z ∀z ∈ R
(20)
y(z) ≥ −z ∀z ∈ R
y ∈ R1,1 ,
where
F = P ∈ P0 (R) z̃ ∼ P, EP [z̃] = 0, EP [z̃ 2 ] ≤ 1 .
Bertsimas et al.: Adaptive Distributionally Robust Optimization
16 Article submitted;
Clearly, y ∗ (z) = |z | is the optimal solution and it is also the optimal objective value for all z ∈ R.
1+z 2
However, under the LDR approximation, we obtain y † (z) = 2
, which is essentially greater than
the optimum policy y ∗ (z) except at z = 1 and z = −1. Incidentally, the worst-case distribution
P ∈ F corresponds to the two point distributions with P[z̃ = 1] = P[z̃ = −1] = 1/2, which explains
why their worse-case expectations coincide. This is similar to Delage and Iancu (2015) observation
that the worst-case policy generated by decision rule can be inefficient and such degeneracy is
common in robust multistage decision models.
Another issue with using the optimal decision rule as a policy is the potential violation of
time consistency. In dynamic decision making, time inconsistency arises when an optimal policy
perceived in one time period may not be recognized as optimal in another. Delage and Iancu
(2015), Xin et al. (2015) show that in addressing multiperiod robust or distributionally robust
optimization problems, time consistency may be affected by how the ambiguity sets are being
updated dynamically. While time consistency is a desirable feature in rational decision making,
policies that may violate time consistency have also been justified in the literature (see, for instance,
Basak and Chabakauri 2010, Kydland and Prescott 1977, Richardson 1989, Bajeux-Besnainou and
Portait 1998).
Consequently, when solving the adaptive distributionally robust optimization problem, we cau-
tion against using the optimal decision rule as a policy. In many practical applications of dynamic
decision making, it suffices to implement the here-and-now decision without having to commit to a
policy that dictates how the solution might change as uncertainty unfolds. For a two-stage problem,
the second stage decision should be determined by solving a linear optimization problem after the
uncertainty is resolved. In addressing a multistage decision problem, we advocate using the LDR
approximation to obtain the here-and-now decision, x ∈ X, which accounts for how decisions might
adapt as uncertainty unfolds over the stages. As we proceed to the next stage, we should adopt
the folding horizon approach and solve for new here-and-now decision using the latest available
information as inputs to another adaptive distributionally robust optimization problem.
Software packages
As a proof of concept, we have develop the software package, ROC (
http://www.meilinzhang.com/software) to provide an intuitive environment for formulating
and solving our proposed adaptive distributionally robust optimization models. ROC is developed
in the C++ programming language, which is fast, highly portable and well suited for deployment
in decision support systems. A typical algebraic modeling package provides the standardized
format for declaration of decision variables, transcription of constraints and objective functions,
and interface with external solvers. ROC has additional features including declaration of uncertain
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 17
parameters and linear decision rules, transcriptions of ambiguity sets and automatic reformulation
of standard and distributionally robust counterparts using the techniques described in this paper.
Interestingly, XProg (http://xprog.weebly.com), a new MATLAB based algebraic modeling
package that implements our proposed framework has independently emerged. The design of
XProg is similar to ROC. Since MATLAB platform is a more user friendly environment, XProg
can be used for rapid prototyping of models, while ROC would be better suited for deployment
of the solutions. The examples in our numerical studies below can easily be implemented in both
ROC and XProg.
where T is the scheduled completion time for the physician before overtime commences.
A common decision criterion in the medical appointment schedule is to minimize the expected
total cost of patients waiting and physician overtime, where the cost of a patient waiting is normal-
ized to one per unit delay and the physician’s overtime cost is γ per unit delay. For a given arrival
schedule x ∈ X, and a realization of consultation times z ∈ RN
+ , the total cost can be determined
where yi denotes the waiting time of patient i, i ∈ [N ] and yN +1 represents the overtime of the
physician. Note that the appointment scheduling problem is one that has complete recourse. From
Theorem 1, we can compute the worst-case expectation over an ambiguity set F ,
by enumerating all the extreme points of the corresponding dual feasible set,
( )
p i − p i−1 ≥ − 1 ∀i ∈ { 2, . . . , N }
P = p ∈ RN + .
pN ≤ γ
However, given the exponentially large number of extreme points of P , it would be generally
intractable to obtain exact solutions.
Kong et al. (2013) is first to propose a distributional robust optimization for the appointment
scheduling problem. They consider a cross moment ambiguity set that characterizes the distribu-
tions of all nonnegative random variables with some specified mean values, µ and covariance Σ as
follows:
z̃ ∼ P
EP [z̃] = µ
F̄CM = P ∈ P0 (RN ) EP [(z − µ)(z − µ)0 ] = Σ . (22)
P[z̃ ∈ RN+] = 1
As the problem is intractable, they formulate a SDP relaxation that solves the problem approxi-
mately.
To obtain a tractable formulation, Mak et al. (2014) ignore information on covariance and con-
sider a marginal moment ambiguity set as follows:
z̃ ∼ P
EP [z̃] = µ
F̄M M = P ∈ P0 (RN ) E [(z̃ − µ )2 ] = σ 2 ∀i ∈ [N ] , (23)
P i i i
P[z̃ ∈ RN ] = 1
+
where σi2 , i ∈ [N ] is the variance of z̃i . Surprsingly, Mak et al. (2014) show that the model has
a hidden tractable reformulation, which they have cleverly exploited to obtain exact solutions.
Observe that due to equality constraints on variances, F̄M M is not a SOC ambiguity set. Never-
theless, by relaxing the equality constraints to inequality constraints, we will obtain the following
SOC ambiguity set:
z̃ ∼ P
EP [z̃] = µ
FM M = P ∈ P0 (RN ) EP [(z̃i − µi )2 ] ≤ σi2 ∀i ∈ [N ] . (24)
P[z̃ ∈ RN +] = 1
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 19
where
z≥0
W̄M M = (z, u) ∈ RN × RN .
q
2 ui −1 2 ui +1
(zi − µi ) + 2
≤ 2
∀i ∈ [N ]
In the proof of Mak et al. (2014) Theorem 1, the authors show that the unconstrained dual
variables that are associated with the second moment equality constraints must be positive at
optimality. Incidentally, when the problem is replaced by the relaxed ambiguity set, these variables
would have nonnegative constraints. Hence, optimizing over the relaxed ambiguity set, FM M would
yield the same set of optimal solutions as those obtained via the ambiguity set, F̄M M .
However, since the marginal moment ambiguity set is incapable of incorporating covariance,
despite the ease of computing the optimal solution, it may lead to conservative solutions. As a
compromise, we propose the following partial cross moment SOC ambiguity set,
z̃ ∼ P
EP [z̃] = µ
2 2
FP CM = P ∈ P0 (R ) N E P [(z̃ i − µ i ) ] ≤ σ i ∀i ∈ [N ] . (25)
0 2 0
E P [(1 (z̃ − µ)) ] ≤ 1 Σ1
P z̃ ∈ RN = 1
+
where
z≥0
2
(zi − µi ) ≤ ui ∀i ∈ [N ]
2
W̄P CM = (z, u) ∈ RN × RN +1 (26)
X
(zi − µi ) ≤ uN +1
i∈[N ]
For these SOC ambiguity sets, we can obtain approximate solutions to the appointment schedul-
ing problem based on the new LDR approximation as follows,
X
min sup EP yi (z̃, ũ) + γyN +1 (z̃, ũ)
P∈G
i∈[N ]
where α ∈ [0, 1] is the correlation coefficient between any two different random consultation times.
The evaluation period, T depends on the instance parameters as follows,
v
XN u N
uX
T= µi + 0.5t σi 2 .
i=1 i=1
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 21
For each approach and α ∈ {0.00, 0.25, 0.50, 0.75, 1.00}, we obtain the objective values of seven
randomly generated instances. We report the results in Table 1.
Observe that Regular LDR performs extremely poorly. Indeed, as noted in Model (12), the
regular LDR approximation is unable to incorporate most of the information of the ambiguity set
other than the mean, µ and the nonnegative support set, which leads to the ultra-conservative
result. We note that Exact MM improves marginally over Approx MM, and Approx PCM improves
over Approx MM. For uncorrelated or mildly correlated random variables, i.e., α ∈ {0, 0.25}, Approx
PCM would yield a lower objective then. Under perfect correlation, i.e., α = 1, the objective values
of Approx PCM and Approx MM coincide and they are marginally higher than the objective
values of Exact MM. Hence, while there are benefits from the new LDR approximation, it does
not replicate the optimal solution of Mak et al. (2014).
In Table 2, we show how the size of the four tractable models (Regular LDR, Exact MM, Approx
MM and Approx PCM) scales with the number of jobs, N . For Approx CM, unlike the previous
approaches, we are unable to obtain its optimally verified solutions by all the three SDP solvers
when the correlation is high. Among the instances that the solutions of Approx CM could be
optimally verified by at least one of the solvers, we observe that the corresponding objective values
attained are the lowest values among the approaches.
In Table 3 and Table 4, we increase the number of jobs from N = 12 to N = 30 and report the
objective values and computational times for the different approaches. In this numerical study, we
set α = 0 and = 0.15. The results indicate that the approach with a tighter approximation also
incurs a longer computational time. Observe that it takes significantly longer time to solve Approx
CM and more seriously, its solution may not necessarily be optimally verified by the SDP solvers
as the problem size increases. In contrast, the Approx PCM can be computed quickly and reliably
and its solution consistently improves over Exact MM and Approx MM for the case when α = 0.
Hence, this underscores the importance of having a stable optimization format and reaffirm our
restriction to SOC ambiguity set.
Similar to Mak et al. (2014), we also compare the performance of the approaches in out-of-
sample study using truncated normal and log normal probability distributions. We assume that
the underlying random variables are independently distributed and the parameters of the distri-
butions correspond to the respective ambiguity sets for which the distributionally robust solutions
are obtained. Upon obtaining the solutions from the various methods, we compare the perfor-
mance and present the results in Table 5. These values are evaluated via Monte Carlo simulation
with 10, 000 trials under each specific distribution. Interestingly, despite the differences in objec-
tive values attained, the out-of-sample study alludes to the closeness of results between Exact
Bertsimas et al.: Adaptive Distributionally Robust Optimization
22 Article submitted;
Models N = 12 N = 14 N = 16 N = 18 N = 20 N = 25 N = 30
Exact MM 216.74 278.75 368.62 364.84 520.24 736.24 1129.32
Approx MM 219.65 288.48 374.71 380.50 538.52 796.55 1206.43
Approx PCM 196.64 255.07 329.23 333.67 459.48 608.02 889.51
Approx CM 143.87 179.88 226.84 224.29 301.04 419.72 –
Table 3 Objective values under different instances with increasing jobs, N .
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 23
Models N = 12 N = 14 N = 16 N = 18 N = 20 N = 25 N = 30
Exact MM <1 <1 <1 <1 <1 <1 <1
Approx MM <1 <1 <1 <1 <1 <1 1
Approx PCM <1 <1 <1 <1 <1 <1 2
Approx CM 7 15 31 57 123 378 –
Table 4 Computation time under different instances with increasing jobs, N .
MM and Approx MM as well as between Approx PCM and Approx CM. Since the random vari-
ables are independent, and hence uncorrelated, as we have expected, incorporating covariance or
partial covariance information in the ambiguity would lead to improvement in the out-of-sample
performances.
t
!2
X
ust ≥ zr ∀s ≤ t, s, t ∈ [T ]
W̄P CM = (z, u) ∈ RT × R(T +1)T /2 .
r=s
z ∈ [z, z̄]
This partial cross moment incorporates the variances of the sum of factors leading to the time period
t. We let ũt = (ũrs )1≤r≤s≤t , t ∈ [T ], which, together with z̃ t are associated with the information
available at the end of period t. Consequently, we use ROC to formulate the problem via the new
LDR approximation and solve it using CPLEX.
Following the numerical study of See and Sim (2009), we set the parameters x̄t = 260, ct = 0.1,
ht = 0.02, for all t ∈ [T ], bt = 0.2, for all t ∈ [T − 1] and bT = 2. We assume that random factors z̃t
are uncorrected random variables in [−z̄, z̄] with standard deviations bounded below by √1 z̄. In
3
characterizing the partial cross moment ambiguity set, we have µt = µ, t ∈ [T ] and
(t − s + 1) 2
φ2st = z̄ ∀s ≤ t, s, t ∈ [T ].
3
Observe that iid uniformly distributed random variables in [−z̄, z̄] would be a feasible distribution
in the ambiguity set and we use this to obtain a lower bound to the inventory control problem.
Specifically, we investigate the performance of the multi-period inventory control problem among
different approaches as follows,
• (Lower Bound): A lower bound obtained by using iid uniformly distributed random factors
and solving the dynamic inventory control problem to optimality (for the dynamic programming
implementation, see See and Sim 2009).
• (Approx MM): Solutions based on the new LDR approximation under the marginal moment
ambiguity set (i.e. known mean values, upper bound on variances and nonnegative support).
• (Approx PCM): Solutions based on the new LDR approximation under the partial cross
moment ambiguity set, FP CM .
In Table 6, we report the objective values attained for the different approaches under various
parameters. As in the previous computational study, we observe that by incorporating partial cross
moment information, we can significantly improve the objectives of the adaptive distributionally
robust optimization problems. Moreover, the objectives of Approx PCM are reasonably close to
the lower bounds. It has well been known that despite the gaps from the lower bounds, numerical
experiments on robust inventory control problems have demonstrated that the actual objectives
attained in out-of-sample analysis are often closer to the optimal values than what the model objec-
tives have reflected (see, for instance, Bertsimas and Thiele 2006, See and Sim 2009). Moreover,
the benefit of distributional robustness arises when there is disparity between the actual demand
distribution and the demand distribution in which the optimal policy is derived. In such cases,
the robust solution could perform significantly better than the misspecified optimum policy (see,
numerical experiments of Bertsimas and Thiele 2006).
Bertsimas et al.: Adaptive Distributionally Robust Optimization
26 Article submitted;
6. Future work
In our numerical studies, we show the benefits of the partial cross moments ambiguity set. However,
the choice of such ambiguity set appears ad hoc and it begs an interesting question as to how we
can systematically adapt and improve the partial cross moments ambiguity set. Chen et al. (2016)
have recently proposed a new class of infinitely constrained ambiguity sets where the number of
expectation constraints could be infinite. To solve the problem, they consider a relaxed ambiguity
set with finite number of expectation constraints, as in the case of the partial cross moments ambi-
guity set. More interestingly, for static robust optimization problems, the “violating” expectation
constraint can be identified and added to the relaxed ambiguity set to improve the solution. While
the approach works for static distributionally robust optimization problems, the extension to adap-
tive problems has not been studied. We believe this is an important extension of this framework
that will help us model and solve adaptive distributionally robust optimization problems for a
larger variety of ambiguity sets.
Acknowledgments
The authors would like to thank the Department Editor, Professor Noah Gans, the anonymous AE and the
reviewers for their valuable and insightful comments. The research is funded by NUS Global Asia Institute
and the Singapore Ministry of Education Social Science Research Thematic Grant MOE2016-SSRTG-059.
Any opinions, findings, and conclusions or recommendations expressed in this material are those of the
authors and do not reflect the views of the Singapore Ministry of Education or the Singapore Government.
References
Alizadeh, F., D. Goldfarb (2003) Second-order cone programming. Math. Programming 95(1):2–51.
Bajeux-Besnainou, Isabelle, Roland Portait (1998) Dynamic asset allocation in a mean-variance framework.
Management Science 44.11-part-2 (1998): S79-S95.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 27
Basak, Suleyman, Georgy Chabakauri (2010) Dynamic mean-variance asset allocation. Review of financial
Studies 23(8): 2970–3016.
Ben-Tal, A., den Hertog, D. and Vial, J. (2015) Deriving robust counterparts of nonlinear uncertain inequal-
ities. Math. Programming 149(1):265–299.
Ben-Tal, A., A. Nemirovski (1998) Robust convex optimization. Math. Oper. Res. 23(4):769–805.
Ben-Tal, A., A. Nemirovski (1999) Robust solutions of uncertain linear programs. Oper. Res. Lett 25(1,
Auguest):1–13.
Ben-Tal, A., A. Nemirovski (2000) Robust solutions of linear programming problems contaminated with
uncertain data. Math. Programming Ser. A 88(3):411–424.
Ben-Tal, A., A. Nemirovski (2001a) Lectures on Modern Convex Optimization: Analysis, Algorithms, and
Engineering Applications. SIAM.
Ben-Tal, A., A. Nemirovski (2001b) On polyhedral approximations of the second-order cone Mathematics of
Operations Research 26:193–205
Ben-Tal, A., A. Goryashko, E.Guslitzer, A. Nemirovski (2004) Adjustable robust solutions of uncertain linear
programs. Math. Programming 99:351-376.
Bertsimas, D., D. A. Iancu, P. A. Parrilo (2010) Optimality of Affine Policies in Multistage Robust Opti-
mization. Mathematics of Operations Research 35(2):363–394.
Bertsimas, D., D.B. Brown (2009) Constructing uncertainty sets for robust linear optimization. Operations
Research 57(6) 1483-1495.
Bertsimas, D., Brown, D. B., Caramanis, C. (2011). Theory and applications of robust optimization. SIAM
review, 53(3), 464-501.
Bertsimas, D., Popescu, I. (2005) Optimal inequalities in probability theory: A convex optimization approach.
SIAM Journal on Optimization 15(3):780-804.
Bertsimas, D., M. Sim (2004) The price of robustness. Oper. Res. 52(1):35–53.
Bertsimas, D., A. Thiele (2006) A robust optimization approach to inventory theory. Oper. Res. 54(1):150–
168.
Birge, J. R., F. Louveaux (1997) Introduction to Stochastic Programming. Springer, New York.
Breton, M., S. El Hachem (1995) Algorithms for the solution of stochastic dynamic minimax problems.
Comput. Optim. Appl. 4:317–345.
Chen, Z., M. Sim., H. Xu (2016) Distributionally Robust Optimization with Infinitely Constrained Ambiguity
Sets Optimization online
Chen, X., Y. Zhang (2009) Uncertain linear programs: Extended affinely adjustable robust counterparts.
Oper. Res. 57(6):1469–1482.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
28 Article submitted;
Chen, X., M. Sim, P. Sun (2007) A robust optimization perspective on stochastic programming. Operations
Research, 55(6):1058–1071.
Chen, X., M. Sim, P. Sun, J. Zhang (2008) A linear decision-based approximation approach to stochastic
programming. Oper. Res. 56(2):344–357.
Delage, E., Dan Iancu (2015) Robust multi-stage decision making. INFORMS Tutorials in Operations
Research (2015): 20–46.
Delage, E., Y. Ye (2010) Distributionally robust optimization under moment uncertainty with application
to data-driven problems. Oper. Res. 58(3):596–612.
Dupacova, J.(1987) The minimax approach to stochastic programming and an illustrative application.
Stochastics 20(1):73–88.
Dyer, M., L. Stougie (2006) Computational complexity of stochastic programming problems. Math. Pro-
gramming Ser. A 106:423–432.
Ellsberg, D. (1961) Risk, ambiguity and the Savage axioms. Quarterly Journal of Economics, 75(4), pp.
643-669.
Garstka, S. J., R. J.-B. Wets (1974) On decision rules in stochastic programming. Math. Programming
7(1):117-143.
Ghaoui, El. L., H. Lebret (1997) Robust solutions to least-squares problems with uncertain data. SIAM
J.Matrix Anal. Appl.18(4):1035–1064.
Ghaoui, El. L., F. Oustry, H. Lebret (1998) Robust solutions to uncertain semidefinite programs. SIAM J.
Optim. 9:33–53.
Gilboa, I., D. Schmeidler (1989) Maxmin expected utility with non-unique prior. Journal of Mathematical
Economics 18(2):141–153.
Goh, J., M. Sim (2009) Robust optimization made easy with ROME. Oper. Res. 59(4):973–985.
Goh, J., M. Sim (2010) Distributionally robust optimization and its tractable approximations. Oper. Res.
58(4):902–917.
Graves, S. C. (1999) A single-item inventory model for a nonstationary demand process. Manufacturing &
Service Operations Management 1(1):50–61.
He, Long, Zhenyu Hu, Meilin Zhang (2017) Robust Repositioning for Vehicle Sharing. NUS Working paper
Hsu M., M. Bhatt, R. Adolphs, D. Tranel, C.F. Camerer. (2005). Neural systems responding to degrees of
uncertainty in human decision-making. Science 310 1680–1683.
Isii, K. (1962) On sharpness of tchebycheff-type inequalities. Annals of the Institute of Statistical Mathematics
14(1):185–197.
Knight, F. H. (1921) Risk, uncertainty and profit. Hart, Schaffner and Marx.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
Article submitted; 29
Kong, Q., Lee, C. Y., Teo, C. P., Zheng, Z. (2013). Scheduling arrivals to a stochastic service delivery system
using copositive cones. Oper. Res. 61(3): 711726.
Kuhn, D., W. Wiesemann, A. Georghiou (2011) Primal and dual linear decision rules in stochastic and robust
optimization. Math. Programming 130(1):177–209.
Kydland, Finn E., Edward C. Prescott (1977) Rules rather than discretion: The inconsistency of optimal
plans. The journal of political Economy: 473-491.
Lobo, M., L. Vandenberghe, S. Boyd, H. Lebret (1998) Applications of second-order cone programming.
Linear Algebra and its Applications 284(1-3):193–228.
Löberg, J. (2012) Automatic robust convex programming. Optimization methods and software 27(1): 115–129.
Mak, H. Y., Rong, Y., Zhang, J. (2014) Appointment scheduling with limited distributional information.
Management Science 61(2): 316-334.
F. Meng, J. Qi, M. Zhang, J. Ang, S. Chu, M. Sim (2015) A robust optimization model for managing elective
admission in a public hospital. Operations Research 63(6):1452–1467.
Popescu, I. (2007) Robust mean-covariance solutions for stochastic optimization. Oper. Res. 55(4):98–112.
J. Qi (2015) Mitigating delays and unfairness in appointment systems. Management Science 63(2):566-583.
Richardson, Henry R (1989) A minimum variance result in continuous trading portfolio optimization. Man-
agement Science 35(9): 1045–1055.
Ruszczyński, A., A. Shapiro (2003) Stochastic Programming. Handbooks in Operations Research and Man-
agement Science 10. Elsevier Science, Amsterdam.
Shapiro, A., A. Kleywegt (2002) Minimax analysis of stochastic programs. Optimization Methods and Soft-
ware, 17(3):523–542.
Scarf, H. (1958) A min-max solution of an inventory problem. K. Arrow, ed. Studies in the Mathematical
Theory of Inventory and Production. Stanford University Press, Stanford, CA, 201–209.
See, C.-T., M. Sim (2009) Robust approximation of multiperiod inventory management. Oper. Res. 58(3):583–
594.
Shapiro, A., S. Ahmed (2004) On a class of minimax stochastic programs. SIAM Journal on Optimization
14(4):1237–1249.
Sturm, J. F. (1999) Using SeDuMi 1.02, a MATLAB toolbox for optimization over symmetric cones. Opti-
mization methods and software 11(1-4), 625–653.
Tutuncu, R.H., K.C. Toh, and M.J. Todd (2003) Solving semidefinite-quadratic-linear programs using
SDPT3. Mathematical Programming Ser. B 95:189–217.
Bertsimas et al.: Adaptive Distributionally Robust Optimization
30 Article submitted;
Wiesemann, W., D. Kuhn, M. Sim (2014) Distributionally Robust Convex Optimization. Operations Research
62(6): 1358–1376.
Xin, L., DA. Goldberg, and A. Shapiro (2015) Time (in)consistency of multistage distributionally robust
inventory models with moment constraints. https://arxiv.org/abs/1304.3074.
Xu, H., S. Mannor (2012) Distributionally robust Markov decision processes. Mathematics of Operations
Research 37(2):288–300.
Žáčková, J . (1966) On minimax solution of stochastic linear programming problems. Časopis pro Pěstov́anı́
Matematiky, 91:423–430.
Zhen, J., M. Sim, den Hertog, D. (2016) Adjustable Robust Optimization via Fourier-Motzkin Elimination
Optimization online
e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization ec1
Eletronic Companion
For any feasible solution in (9) and (z̃, ũ) ∼ P ∈ G observe that
Hence, we have
where r ∈ R, s ∈ RL1 , t ∈ RI2 are the dual variables corresponding to the expectation constraints of
the ambiguity set, G .
Under the assumption of relatively complete recourse, Q(x, z) is finite and by strong duality of
linear optimization, we have equivalently
Therefore,
β1 (x) = inf r + s0 µ + t0 σ
0
p0p (b1 − A1 x)
s.t. r ≥ sup .
.
− G0
s
z − t0
u + p0
(b 0
− A 0
x) ∀p ∈ [P ]
.
p
(z,u)∈W̄
I I
p0p (b 1 − A 1 x)
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2 .
(EC.1)
ec2 e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization
Using the explicit formulation of W̄ in (8), by weak conic duality (see, for instance, Ben-Tal and
0
p0p (b1 − A1 x)
sup .
.
− G 0
s
z − t0
u + p 0
(b0
− A 0
x)
.
p
(z,u)∈W̄
I1 I1
0
pp (b − A x)
≤
inf π p 0 h + p0p (b0 − A0 x)
0 1
pp (b − A1 x)
s.t. C 0 π p = ..
− G0 s
.
p0p (bI1 − AI1 x)
D 0 π p = −t
E 0πp = 0
π p K 0
π p ∈ RL 2 ,
where π p ∈ RL2 , ∀p ∈ [P ] are the dual variables associated with the conic constants in W̄ . Note that
Hence, using standard robust counterpart techniques, we substitute the dual formulations in
β2 (x) = inf r + s0 µ + t0 σ
s.t. r ≥ π p 0 h + p0p (b0 − A0 x) ∀p ∈ [P ]
0 1
pp (b − A1 x)
C 0πp =
..
− G0 s ∀p ∈ [P ]
.
p0p (bI1 − AI1 x)
D 0 π p = −t ∀p ∈ [P ] (EC.2)
E 0πp = 0 ∀p ∈ [P ]
π p K 0 ∀p ∈ [P ]
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
π p ∈ RL2 ∀p ∈ [P ].
Note that replacing the dual formulation in (EC.1) leads to a restriction to the minimization
problem, hence β1 (x) ≤ β2 (x). Observe that β(x) ≤ β1 (x) ≤ β2 (x), and our goal is to establish
strong duality by showing β2 (x) ≤ β(x). Then we will next approach Problem (EC.2) by taking
e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization ec3
β3 (x) = sup
X
p0p (b0 − A0 x)αp + ..
. z̄ p
I I
p∈[P ] p0p (b 1 − A 1 x)
X
s.t. αp = 1
p∈[P ]
p≥0
αX ∀p ∈ [P ]
(EC.3)
Gz̄ p = µ
p∈[P ]
X
ūp ≤ σ
p∈[P ]
C z̄ p + D ūp + E v̄ p K αp h ∀p ∈ [P ]
αp ∈ R, z̄ p ∈ RI1 , ∀p ∈ [P ]
ūp ∈ RI2 , v̄ p ∈ RI3 ∀p ∈ [P ],
where αp , z̄, ūp , v̄ p , ∀p ∈ [P ] are the dual variables associated with the specified constraints respec-
tively. Under the Slater’s condition, there exists u† ∈ RI2 and v † ∈ RI3 such that
u† < σ
Cz † + Du† + Ev † ≺K h.
1 z† u† v†
αp = , z̄ p = , ūp = , v̄ p = ,
P P P P
for all p ∈ [P ]. Since Problem (EC.3) is strictly feasible, strong duality holds and β2 (x) = β3 (x).
Moreover, there exists a sequence of interior solutions
such that 0
p0p (b1 − A1 x)
lim
X
p0p (b0 − A0 x)αpk + .. k
. z̄ p
= β3 (x).
k→∞
I I
p∈[P ] p0p (b 1 − A 1 x)
X
Observe that for all k, αpk > 0, αpk = 1 and we can construct a sequence of discrete probability
p∈[P ]
distributions {Pk ∈ P0 (RI1 × RI2 )}k≥0 on random variable (z̃, ũ) ∼ Pk such that
z̄ kp ūkp
Pk (z̃, ũ) = αk , αk = αpk ∀p ∈ [P ].
p p
Note that,
EPk [Gz̃] = µ, EPk [ũ] ≤ σ, Pk [(z̃, ũ) ∈ W̄ ] = 1,
ec4 e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization
β3 (x) = lim
X
p0p (b0 − A0 x)αpk + .. k
. z̄ p
k→∞
I I
p∈[P ] p0p (b 1 − A 1 x)
0 1 0
pp (b − A1 x) k
X 0 0 0 .. z̄ p
= lim αpk p (b − A x) +
.
k→∞ p
α k
p
p∈[P ] p0p (bI1 − AI1 x)
0 1 1
0
p q (b − A x)
k
X
k
0 0 0 . z̄ q
≤ lim αp max pq (b − A x) +
..
k
k→∞ q∈[P ]
0 I I
αq
p∈[P ]
pq (b − A x) 1 1
0 1 1
0
pq (b − A x)
=
0 0
lim EPk max pq (b − A x) + 0 .
. z̃
.
k→∞ q∈[P ]
I I
p0q (b 1 − A 1 x)
≤ sup EP [Q(x, z)]
P∈G
= β(x).
Q ũ) ∼ P
(z̃,
z̃ P ∈ F
G = P ∈ P0 (RI1 × RI1 ) EP [ũ] ≤ σ ,
P[(z̃, ũ) ∈ W̄ ] = 1
where the lifted support set is W̄ = {(z, u) ∈ RI1 × RI1 | u ≥ z, u ≥ −z }. The decision rule under
the new LDR approximation is given by
X X
y(z, u) = y 0 + y 1i zi + y 2i ui .
i∈[I1 ] i∈[I1 ]
To show that Problem (15) is feasible and finite under the new LDR, we consider the following
relaxed lifted ambiguity set
(z̃, ũ) ∼ P
Ḡ = P ∈ P0 (RI1 × RI1 ) EP [ũ] ≤ σ ,
P[(z̃, ũ) ∈ W̄ ] = 1
e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization ec5
Since G ⊆ Ḡ , we have βE (x) ≤ β̄E (x). Consequently, it suffices to show that there exists a solution
y ∈ L̄N2 that is feasible in the following problem,
min r + t0 σ
X
s.t. r + t0 u ≥ d0 y 0 + d0 y 2i ui ∀(z, u) ∈ W̄
i∈[I1 ]
X X X
By + 0
By 1i zi + By 2i ui ≥ b0 − A0 x + (bi − Ai x)zi ∀(z, u) ∈ W̄
(EC.5)
i∈[I1 ] i∈[I1 ] i∈I1
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
y 0 , y 1i , y 2i ∈ RN2 i ∈ [I1 ].
Since B is a matrix that satisfies the condition of complete recourse, there exist ȳ 0 , ȳ 2i i ∈ [I1 ], such
that
B ȳ 0 ≥ b0 − A0 x,
x + y 0 u ≥ a + b0 z ∀(z, u) ∈ W̄ .
r = d0 ȳ 0
ti = max{d0 ȳ 2i , 0} ∀i ∈ [I1 ]
y0 = ȳ 0
y 1i = 0 ∀i ∈ [I1 ]
y 2i = 2
ȳ i ∀i ∈ [I1 ].
Q(x, z) = min dy
s.t. A(z)x + By ≥ b(z) ∀(z, u) ∈ W̄
y ∈ R,
ec6 e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization
is unbounded below whenever dB < 0. Since the second stage decision variable y is unconstrained,
for the problem to be finite, we can assume without loss of generality that B > 0 and d ≥ 0. In
which case,
[b(z) − A(z)x]i
Q(x, z) = d max .
i∈[M ] [B]i
0
P = p ∈ RM
+ : B p=d .
βE (x) = min r + s0 + t0 σ
s.t. r + s0 (Gz) + t0 u ≥ dy(z, u) ∀(z, u) ∈ W̄
A(z)x + By(z, u) ≥ b(z) ∀(z, u) ∈ W̄
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
y ∈ L̄1
or equivalently
βE (x) = min r + s0 + t0 σ
0 0
s.t. r + s0 (Gz) + t0 u ≥ d(y 0 + y 1 z + y 2 u) ∀(z, u) ∈ W̄
(y 0 + y 1 z + y 2 u) ≥ [b(z )−[BA(z )x]i
0 0
]i
∀i ∈ [M ], ∀(z, u) ∈ W̄
(EC.7)
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
y 0 ∈ R, y 1 ∈ RI1 , y 2 ∈ RI2 .
Let (r† , s† , t† ) be a feasible solution of Problem (EC.6). We can construct a feasible solution
(r, s, t, y 0 , y 1 , y 2 ) to Problem (EC.7) by letting
which yields the same objective as Problem (EC.6). Hence, β(x) ≤ βE (x) ≤ β(x).
e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization ec7
for all i ∈ [M ], which are not necessarily feasible optimization problems. Let M ⊆ [M ] denote the
subset of indices in which their corresponding subproblems are feasible, i.e., M = {i ∈ [M ] | fi∗ <
∞}, and M̄ = [M ] \ M. Correspondingly, let (ȳ i , q̄ i ) be the optimal solution of Problem (EC.8) for
all i ∈ M. Here, fi∗ = d0 ȳ i ≥ 0, i ∈ M is assumed or otherwise, Q(x, z) would be unbounded from
below. The solution to deflected linear decision is obtained by solving the following optimization
problem,
X
βD (x) = min sup EP [d0 y(z̃)] + fi∗ sup EP (−qi (z̃))+
P∈F P∈F
i∈M
s.t. A(z)x + By(z) = b(z) + q(z) ∀z ∈ W
qi (z) ≥ 0 ∀i ∈ M̄, ∀z ∈ W (EC.9)
y ∈ LN2
q ∈ LM .
Suppose (y ∗ , q ∗ ) is the optimal solution of Problem (EC.9), the corresponding decision rule under
DLDR approximation is given by
X
y D (z) = y ∗ (z) + ȳ i ((−qi∗ (z))+ ).
i∈M
Chen et al. (2008), Goh and Sim (2010) show that y D (z̃) is a feasible solution of Problem (11).
Moreover,
β(x) ≤ sup EP d0 y D (z̃) ≤ βD (x) ≤ βL (x).
P∈F
Our next result shows that the new LDR approximation may improve the bound provided by the
DLDR approximation.
Theorem EC.1. The bound obtained via the new LDR approximation is no larger than the
bound obtained via DLDR approximation, i.e.,
βE (x) ≤ βD (x).
Proof. Following similar exposition of Theorem 1, we have the equivalent form of βD (x) as follows:
X
βD (x) = min r0 + s00 µ + t00 σ + fi∗ (ri + s0i µ + t0i σ)
i∈M
s.t. r0 + s0 (Gz) + t0 u ≥ d0 y(z)
0 0
∀(z, u) ∈ W̄
ri + s0i (Gz) + t0i u ≥ −qi (z) ∀i ∈ M, ∀(z, u) ∈ W̄
ri + s0i (Gz) + t0i u ≥ 0 ∀i ∈ M, ∀(z, u) ∈ W̄
t≥0 ∀i ∈ {0} ∪ M (EC.10)
A(z)x + By(z) = b(z) + q(z) ∀(z, u) ∈ W̄
qi (z) ≥ 0 ∀i ∈ M̄
ri ∈ R, si ∈ RL1 , ti ∈ RI2 ∀i ∈ {0} ∪ M
y ∈ L N2
q ∈ LM .
βE (x) = min r + s0 µ + t0 σ
s.t. r + s0 (Gz) + t0 u ≥ d0 y(z, u) ∀(z, u) ∈ W̄
A(z)x + By(z, u) ≥ b(z) ∀(z, u) ∈ W̄
(EC.11)
t≥0
r ∈ R, s ∈ RL1 , t ∈ RI2
y ∈ L̄N2 .
Let y † , q † , ri† , s†i , t†i , i ∈ {0} ∪ M be a feasible solution of Problem (EC.10). We will show that there
exists a corresponding feasible solution for Problem (EC.11) with the same objective value. Let
X
r = r0† + dȳ i ri†
i∈M
X
s= s†0 + d0 ȳ i s†i
i∈M
X
t= t†0 + d0 ȳ i t†i ,
i∈M
X 0 0
y(z, u) = y † (z) + ri† + s†i (Gz) + t†i u ȳ i .
i∈M
0 0 X 0 0
r + s0 µ + t0 σ = r0† + s†0 µ + t†0 σ + (ri† + s†i µ + t†i σ)d0 ȳ i
i∈M
0 0 X 0 0
= r0† + s†0 µ + t†0 σ + fi∗ (ri† + s†i µ + t†i σ).
i∈M
e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization ec9
We next check the feasibility of the solution in Problem (EC.11). Note that t ≥ 0 and for all
(z, u) ∈ W̄ ,
!0 !0
X 0
X 0
X 0
r + s0 (Gz) + t0 u = r0† + d ȳ i ri† + s†0 + d ȳ i s†i (Gz) + t†0 + d ȳ i t†i u
i∈M i∈M i∈M
0 0 X 0 0
= r0 + s†0 (Gz) + t†0 u +
†
ri† + s†i (Gz) + t†i u d0 ȳ i
i∈M
X 0 0
0 †
≥ d y (z) + ri† + s†i (Gz) + t†i u d0 ȳ i
i∈M
= d0 y(z, u),
where the inequality follows from the first robust counterpart constraint in Problem (EC.10). We
now show the feasibility of second robust counterpart constraint in Problem (EC.11). Observe that
for all (z, u) ∈ W̄ ,
X 0 0
A(z)x + By(z, u) = A(z)x + By † (z) + ri† + s†i (Gz) + t†i u B ȳ i
i∈M
X 0 0
= b(z) + q (z) + †
ri† + s†i (Gz) + t†i u q̄ i
i∈M
X 0 0
X X
= b(z) + qi† (z)ei + qj† (z)ej + ri† + s†i (Gz) + t†i u q̄ i
i∈M j∈M̄ i∈M
X 0 0
X X
≥ b(z) + qi† (z)ei + qj† (z)ej + ri† + s†i (Gz) + t†i u ei
i∈M j∈M̄ i∈M
X 0 0
X
= b(z) + qj† (z)ej + qi† (z) + ri† + s†i (Gz) + t†i u ei
j∈M̄ i∈M
≥ b(z).
0 0
The first inequality holds because q̄ i ≥ ei and ri† + s†i (Gz) + t†i u ≥ 0 for all i ∈ M, (z, u) ∈ W̄ . The
0 0
second inequality is due to ri† + s†i (Gz) + t†i u ≥ −qi† (z) for all i ∈ M, (z, u) ∈ W̄ and qi† (z) ≥ 0 for
all i ∈ M̄, (z, u) ∈ W̄ .
To show that in general βE (x) 6= βD (x), we consider the following complete recourse problem,
β ∗ = min sup EP [y(z̃)]
P∈F
s.t. y(z) ≥ zi ∀z ∈ R3 , i ∈ {1, 2, 3} (EC.12)
y ∈ R1,1 .
where
z̃ ∼ P
EP [z̃] = 0
F= P ∈ P0 (R3 ) hP i .
3 2
k=1 k ≤ 1
z̃
EP
From Theorem 4, we can obtain the optimal objective using the new LDR approach, which is
βE = β ∗ = 0.8165. On the other hand, under the DLDR approximation, the objective value is
βD = 1.154, which is significantly higher than βE .
ec10 e-companion to Bertsimas et al.: Adaptive Distributionally Robust Optimization
W̄ = {(z, u) ∈ H : z ∈ W} . (EC.15)
with C ∈ RL3 ×I1 , D ∈ RL3 ×I2 , E ∈ RL3 ×I3 , h ∈ RL3 and K ⊆ RL3 being a proper cone. We also
impose the following Slater’s condition:
0
P = p ∈ RM
+ : B p=d .
β(x) = min r + s0 µ + t0 σ
s.t. r ≥ π i 0 h+ pi 0 b0 − pi 0 A0 x ∀i ∈ [P ]
0 1 1
pi (b − A x)
0 ..
C πi = − G0 s ∀i ∈ [P ]
.
p0i (bI1 − AI1 x)
D 0 π i = −t ∀i ∈ [P ] (EC.18)
E 0πi = 0 ∀i ∈ [P ]
π i K∗ 0 ∀i ∈ [P ]
t K0∗ 0
r ∈ R, s ∈ RL1 , t ∈ RL2
π i ∈ RL 3 ∀i ∈ [P ].
Proof. We leave this to the readers as the steps of the proof is identical to the proof of Theorem
1.