Convexity Under Operators: Supremum, Infimum, Marginal Infimum and Infimum Convolution
Convex functions are central in convex optimization, and it’s natural to ask whether convexity is preserved under common operations like taking the supremum, infimum, or partial minimization. This post explores the following progression:
- The supremum over a family of convex functions is convex.
- The infimum over a family of convex functions is not necessarily convex.
- However, if the function is jointly convex, then the partial infimum over one variable does preserve convexity.
- A generalization of the convexity of partial infimum, called infimum convolution between two jointly convex functions is also convex.
We explore these ideas with proofs and remarks that illuminate why they are true.
Section 1. Definitions of convex functions
Before exploring how convexity behaves under operations like supremum, infimum, or marginalization, we recall key definitions from convex analysis.
Convex Function
A function F:Rn→R∪{+∞} is called convex if its domain domF:={x∈Rn∣F(x)<+∞} is convex, and for all x1,x2∈domF and θ∈[0,1], we have
F(θx1+(1−θ)x2)≤θF(x1)+(1−θ)F(x2).
This inequality expresses that the graph of F lies below the line segment connecting F(x1) and F(x2) — a hallmark of convexity.
Jointly Convex Function
Let f:Rn×Rm→R∪{+∞}. We say that f is jointly convex in (x,y) if its domain domf:={(x,y)∣f(x,y)<+∞} is convex, and for all (x1,y1),(x2,y2)∈domf and θ∈[0,1]:
f(θx1+(1−θ)x2,θy1+(1−θ)y2)≤θf(x1,y1)+(1−θ)f(x2,y2).
This means f behaves like a convex function on the product space Rn×Rm — convex with respect to both variables simultaneously, not separately.
Joint convexity is stronger than separate convexity in x and y; that is, x↦f(x,y) and y↦f(x,y) can both be convex without f(x,y) being jointly convex.
Example 1: f(x)=⟨a,x⟩ is convex.
Example 2: f(x,y)=⟨a,x⟩+⟨b,y⟩ is jointly convex.
Example 3: f(x,y)=⟨x,y⟩ is convex in x and in y but not jointly convex in (x,y)
These definitions set the stage for the key results to come, where we examine how the convexity of such functions is preserved under optimization operations like sup, inf, marginalization and convolution.
Section 2. Supremum of Convex Functions Is Convex
Theorem
Let {fy(x)=f(x,y)}y∈Y be a family of convex functions indexed by y∈Y (with Y arbitrary — finite or infinite). Define:
F(x):=y∈Ysupf(x,y).
Then F is convex.
Proof
Let x1,x2∈Rn, and θ∈[0,1]. Define xθ=θx1+(1−θ)x2.
For each y∈Y, since f(x,y) is convex in x:
f(xθ,y)≤θf(x1,y)+(1−θ)f(x2,y).
Taking the supremum over all y∈Y on both sides:
F(xθ)=y∈Ysupf(xθ,y)≤y∈Ysup[θf(x1,y)+(1−θ)f(x2,y)].
Using the fact that sup is sublinear in this context:
F(xθ)≤θy∈Ysupf(x1,y)+(1−θ)y∈Ysupf(x2,y)=θF(x1)+(1−θ)F(x2).
Thus, F is convex.
Section 3. Infimum of Convex Functions May Not Be Convex
While the supremum of convex functions preserves convexity, the infimum does not in general, i.e.
F(x):=y∈Yinff(x,y)
is not convex.
Counterexample: Let
f1(x)=(x−1)2,f2(x)=(x+1)2.
Define:
F(x)=inf{f1(x),f2(x)}.
Then:
F(x)={(x+1)2,(x−1)2,x≤0x≥0
This function is not convex.
But wait!
If infimum does not preserve the convexity, then if we add extra condition on f, can we obtain convexity after applying infimum? The answer is yes!
Let f(x,y) be jointly convex in (x,y), and define the marginal minimum
F(x):=y∈Yinff(x,y).
Then F is convex in x.
Furthermore, there is a more general result! Let f and g are jointly convex, then
h(x,z):=y∈Yinff(x,y)+g(y,z)
is jointly convex.
In the next two sections, we will prove theses.
Section 3. Marginal Minimum of Jointly Convex Function
Theorem
Let f(x,y) be jointly convex on Rn×Rm, and Y⊆Rm be convex. Define the partial minimum function:
F(x):=y∈Yinff(x,y).
Then F is convex.
Proof Sketch
Let x1,x2∈Rn, and θ∈[0,1]. Let xθ=θx1+(1−θ)x2.
For ε>0, choose y1,y2∈Y such that:
- f(x1,y1)≤F(x1)+ε
- f(x2,y2)≤F(x2)+ε
Let yθ=θy1+(1−θ)y2∈Y (by convexity of Y). By joint convexity:
f(xθ,yθ)≤θf(x1,y1)+(1−θ)f(x2,y2).
So:
F(xθ)≤f(xθ,yθ)≤θF(x1)+(1−θ)F(x2)+ε.
Letting ε→0 gives:
F(xθ)≤θF(x1)+(1−θ)F(x2),
proving convexity.
Section 4. Infimum Convolution of Jointly Convex Functions
We now extend the result from Section 3 to a more general setting involving two jointly convex functions.
Theorem
Let f(x,y) be jointly convex in (x,y)∈Rn×Rm and g(y,z) be jointly convex in (y,z)∈Rm×Rk, and let Y⊆Rm be a convex set. Define
h(x,z):=y∈Yinf[f(x,y)+g(y,z)].
Then h(x,z) is jointly convex in (x,z).
Proof
Let (x1,z1),(x2,z2)∈Rn×Rk and let θ∈[0,1]. Define:
(xθ,zθ):=θ(x1,z1)+(1−θ)(x2,z2).
For any ε>0, choose y1,y2∈Y such that:
f(x1,y1)+g(y1,z1)≤h(x1,z1)+ε,f(x2,y2)+g(y2,z2)≤h(x2,z2)+ε.
Let yθ:=θy1+(1−θ)y2∈Y (by convexity of Y). Using joint convexity of f and g:
f(xθ,yθ)≤θf(x1,y1)+(1−θ)f(x2,y2),g(yθ,zθ)≤θg(y1,z1)+(1−θ)g(y2,z2).
Adding both inequalities:
f(xθ,yθ)+g(yθ,zθ)≤θ[f(x1,y1)+g(y1,z1)]+(1−θ)[f(x2,y2)+g(y2,z2)].
Hence:
h(xθ,zθ)≤f(xθ,yθ)+g(yθ,zθ)≤θh(x1,z1)+(1−θ)h(x2,z2)+ε.
Since ε>0 is arbitrary, taking the limit yields:
h(xθ,zθ)≤θh(x1,z1)+(1−θ)h(x2,z2),
which proves that h is jointly convex in (x,z).
Summary
Let’s recap the natural flow of ideas:
-
Supremum of convex functions:
- supy∈Yf(x,y) is convex if each f(x,y) is convex in x.
- Holds for any index set Y — even infinite.
-
Infimum of convex functions:
- infy∈Yf(x,y) is not necessarily convex.
- Can fail even for simple convex quadratic functions.
-
Marginal Infimum of jointly convex function:
- infy∈Yf(x,y) is convex if f(x,y) is jointly convex and Y is convex.
- This is especially useful in partial minimization, parametric optimization, and duality.
-
Infimum convolution of two jointly convex functions:
- infy∈Yf(x,y)+g(y,z) is convex in (x,z) if f and g are jointly convex and Y is convex.
- This family of functions form a semigroup.
Understanding these operations sharpens your ability to reason about convexity in convex optimization.