Wednesday, July 2, 2025

partial-minimization

partial-minimization

Convexity Under Operators: Supremum, Infimum, Marginal Infimum and Infimum Convolution

Convex functions are central in convex optimization, and it’s natural to ask whether convexity is preserved under common operations like taking the supremum, infimum, or partial minimization. This post explores the following progression:

  • The supremum over a family of convex functions is convex.
  • The infimum over a family of convex functions is not necessarily convex.
  • However, if the function is jointly convex, then the partial infimum over one variable does preserve convexity.
  • A generalization of the convexity of partial infimum, called infimum convolution between two jointly convex functions is also convex.

We explore these ideas with proofs and remarks that illuminate why they are true.


Section 1. Definitions of convex functions

Before exploring how convexity behaves under operations like supremum, infimum, or marginalization, we recall key definitions from convex analysis.

Convex Function

A function F:RnR{+}F: \mathbb{R}^n \to \mathbb{R} \cup \{+\infty\} is called convex if its domain domF:={xRnF(x)<+}\operatorname{dom} F := \{x \in \mathbb{R}^n \mid F(x) < +\infty\} is convex, and for all x1,x2domFx_1, x_2 \in \operatorname{dom} F and θ[0,1]\theta \in [0,1], we have

F(θx1+(1θ)x2)θF(x1)+(1θ)F(x2). F(\theta x_1 + (1 - \theta)x_2) \leq \theta F(x_1) + (1 - \theta)F(x_2).

This inequality expresses that the graph of FF lies below the line segment connecting F(x1)F(x_1) and F(x2)F(x_2) — a hallmark of convexity.


Jointly Convex Function

Let f:Rn×RmR{+}f: \mathbb{R}^n \times \mathbb{R}^m \to \mathbb{R} \cup \{+\infty\}. We say that ff is jointly convex in (x,y)(x, y) if its domain domf:={(x,y)f(x,y)<+}\operatorname{dom} f := \{(x, y) \mid f(x, y) < +\infty\} is convex, and for all (x1,y1),(x2,y2)domf(x_1, y_1), (x_2, y_2) \in \operatorname{dom} f and θ[0,1]\theta \in [0, 1]:

f(θx1+(1θ)x2,θy1+(1θ)y2)θf(x1,y1)+(1θ)f(x2,y2). f(\theta x_1 + (1 - \theta)x_2, \theta y_1 + (1 - \theta)y_2) \leq \theta f(x_1, y_1) + (1 - \theta) f(x_2, y_2).

This means ff behaves like a convex function on the product space Rn×Rm\mathbb{R}^n \times \mathbb{R}^m — convex with respect to both variables simultaneously, not separately.

Joint convexity is stronger than separate convexity in xx and yy; that is, xf(x,y)x \mapsto f(x, y) and yf(x,y)y \mapsto f(x, y) can both be convex without f(x,y)f(x, y) being jointly convex.

Example 1: f(x)=a,xf(x) = \langle a, x\rangle is convex.
Example 2: f(x,y)=a,x+b,yf(x, y) = \langle a, x\rangle + \langle b, y\rangle is jointly convex.
Example 3: f(x,y)=x,yf(x, y) = \langle x, y\rangle is convex in xx and in yy but not jointly convex in (x,y)(x, y)

These definitions set the stage for the key results to come, where we examine how the convexity of such functions is preserved under optimization operations like sup\sup, inf\inf, marginalization and convolution.


Section 2. Supremum of Convex Functions Is Convex

Theorem

Let {fy(x)=f(x,y)}yY\{f_y(x) = f(x, y)\}_{y \in Y} be a family of convex functions indexed by yYy \in Y (with YY arbitrary — finite or infinite). Define:

F(x):=supyYf(x,y). F(x) := \sup_{y \in Y} f(x, y).

Then FF is convex.

Proof

Let x1,x2Rnx_1, x_2 \in \mathbb{R}^n, and θ[0,1]\theta \in [0, 1]. Define xθ=θx1+(1θ)x2x_\theta = \theta x_1 + (1 - \theta)x_2.

For each yYy \in Y, since f(x,y)f(x, y) is convex in xx:

f(xθ,y)θf(x1,y)+(1θ)f(x2,y). f(x_\theta, y) \leq \theta f(x_1, y) + (1 - \theta)f(x_2, y).

Taking the supremum over all yYy \in Y on both sides:

F(xθ)=supyYf(xθ,y)supyY[θf(x1,y)+(1θ)f(x2,y)]. F(x_\theta) = \sup_{y \in Y} f(x_\theta, y) \leq \sup_{y \in Y} \left[ \theta f(x_1, y) + (1 - \theta) f(x_2, y) \right].

Using the fact that sup\sup is sublinear in this context:

F(xθ)θsupyYf(x1,y)+(1θ)supyYf(x2,y)=θF(x1)+(1θ)F(x2). F(x_\theta) \leq \theta \sup_{y \in Y} f(x_1, y) + (1 - \theta) \sup_{y \in Y} f(x_2, y) = \theta F(x_1) + (1 - \theta) F(x_2).

Thus, FF is convex.


Section 3. Infimum of Convex Functions May Not Be Convex

While the supremum of convex functions preserves convexity, the infimum does not in general, i.e.
F(x):=infyYf(x,y) F(x) := \inf_{y \in Y} f(x, y)
is not convex.

Counterexample: Let

f1(x)=(x1)2,f2(x)=(x+1)2. f_1(x) = (x - 1)^2, \quad f_2(x) = (x + 1)^2.

Define:

F(x)=inf{f1(x),f2(x)}. F(x) = \inf \{f_1(x), f_2(x)\}.

Then:

F(x)={(x+1)2,x0(x1)2,x0 F(x) = \begin{cases} (x+1)^2, & x \leq 0 \\ (x-1)^2, & x \geq 0 \end{cases}

This function is not convex.


But wait!

If infimum does not preserve the convexity, then if we add extra condition on ff, can we obtain convexity after applying infimum? The answer is yes!

Let f(x,y)f(x, y) be jointly convex in (x,y)(x, y), and define the marginal minimum

F(x):=infyYf(x,y). F(x) := \inf_{y \in Y} f(x, y).

Then FF is convex in xx.

Furthermore, there is a more general result! Let ff and gg are jointly convex, then
h(x,z):=infyYf(x,y)+g(y,z) h(x, z) := \inf_{y \in Y} f(x, y) + g(y, z)
is jointly convex.

In the next two sections, we will prove theses.


Section 3. Marginal Minimum of Jointly Convex Function

Theorem

Let f(x,y)f(x, y) be jointly convex on Rn×Rm\mathbb{R}^n \times \mathbb{R}^m, and YRmY \subseteq \mathbb{R}^m be convex. Define the partial minimum function:

F(x):=infyYf(x,y). F(x) := \inf_{y \in Y} f(x, y).

Then FF is convex.

Proof Sketch

Let x1,x2Rnx_1, x_2 \in \mathbb{R}^n, and θ[0,1]\theta \in [0,1]. Let xθ=θx1+(1θ)x2x_\theta = \theta x_1 + (1 - \theta) x_2.

For ε>0\varepsilon > 0, choose y1,y2Yy_1, y_2 \in Y such that:

  • f(x1,y1)F(x1)+εf(x_1, y_1) \leq F(x_1) + \varepsilon
  • f(x2,y2)F(x2)+εf(x_2, y_2) \leq F(x_2) + \varepsilon

Let yθ=θy1+(1θ)y2Yy_\theta = \theta y_1 + (1 - \theta)y_2 \in Y (by convexity of YY). By joint convexity:

f(xθ,yθ)θf(x1,y1)+(1θ)f(x2,y2). f(x_\theta, y_\theta) \leq \theta f(x_1, y_1) + (1 - \theta) f(x_2, y_2).

So:

F(xθ)f(xθ,yθ)θF(x1)+(1θ)F(x2)+ε. F(x_\theta) \leq f(x_\theta, y_\theta) \leq \theta F(x_1) + (1 - \theta) F(x_2) + \varepsilon.

Letting ε0\varepsilon \to 0 gives:

F(xθ)θF(x1)+(1θ)F(x2), F(x_\theta) \leq \theta F(x_1) + (1 - \theta) F(x_2),

proving convexity.


Section 4. Infimum Convolution of Jointly Convex Functions

We now extend the result from Section 3 to a more general setting involving two jointly convex functions.

Theorem

Let f(x,y)f(x, y) be jointly convex in (x,y)Rn×Rm(x, y) \in \mathbb{R}^n \times \mathbb{R}^m and g(y,z)g(y, z) be jointly convex in (y,z)Rm×Rk(y, z) \in \mathbb{R}^m \times \mathbb{R}^k, and let YRmY \subseteq \mathbb{R}^m be a convex set. Define

h(x,z):=infyY[f(x,y)+g(y,z)]. h(x, z) := \inf_{y \in Y} \left[ f(x, y) + g(y, z) \right].

Then h(x,z)h(x, z) is jointly convex in (x,z)(x, z).

Proof

Let (x1,z1),(x2,z2)Rn×Rk(x_1, z_1), (x_2, z_2) \in \mathbb{R}^n \times \mathbb{R}^k and let θ[0,1]\theta \in [0,1]. Define:

(xθ,zθ):=θ(x1,z1)+(1θ)(x2,z2). (x_\theta, z_\theta) := \theta (x_1, z_1) + (1 - \theta)(x_2, z_2).

For any ε>0\varepsilon > 0, choose y1,y2Yy_1, y_2 \in Y such that:

f(x1,y1)+g(y1,z1)h(x1,z1)+ε,f(x2,y2)+g(y2,z2)h(x2,z2)+ε. f(x_1, y_1) + g(y_1, z_1) \leq h(x_1, z_1) + \varepsilon, \\ f(x_2, y_2) + g(y_2, z_2) \leq h(x_2, z_2) + \varepsilon.

Let yθ:=θy1+(1θ)y2Yy_\theta := \theta y_1 + (1 - \theta)y_2 \in Y (by convexity of YY). Using joint convexity of ff and gg:

f(xθ,yθ)θf(x1,y1)+(1θ)f(x2,y2),g(yθ,zθ)θg(y1,z1)+(1θ)g(y2,z2). f(x_\theta, y_\theta) \leq \theta f(x_1, y_1) + (1 - \theta) f(x_2, y_2), \\ g(y_\theta, z_\theta) \leq \theta g(y_1, z_1) + (1 - \theta) g(y_2, z_2).

Adding both inequalities:

f(xθ,yθ)+g(yθ,zθ)θ[f(x1,y1)+g(y1,z1)]+(1θ)[f(x2,y2)+g(y2,z2)]. f(x_\theta, y_\theta) + g(y_\theta, z_\theta) \leq \theta \left[ f(x_1, y_1) + g(y_1, z_1) \right] + (1 - \theta)\left[ f(x_2, y_2) + g(y_2, z_2) \right].

Hence:

h(xθ,zθ)f(xθ,yθ)+g(yθ,zθ)θh(x1,z1)+(1θ)h(x2,z2)+ε. h(x_\theta, z_\theta) \leq f(x_\theta, y_\theta) + g(y_\theta, z_\theta) \leq \theta h(x_1, z_1) + (1 - \theta) h(x_2, z_2) + \varepsilon.

Since ε>0\varepsilon > 0 is arbitrary, taking the limit yields:

h(xθ,zθ)θh(x1,z1)+(1θ)h(x2,z2), h(x_\theta, z_\theta) \leq \theta h(x_1, z_1) + (1 - \theta) h(x_2, z_2),

which proves that hh is jointly convex in (x,z)(x, z).


Summary

Let’s recap the natural flow of ideas:

  1. Supremum of convex functions:

    • supyYf(x,y)\sup_{y \in Y} f(x, y) is convex if each f(x,y)f(x, y) is convex in xx.
    • Holds for any index set YY — even infinite.
  2. Infimum of convex functions:

    • infyYf(x,y)\inf_{y \in Y} f(x, y) is not necessarily convex.
    • Can fail even for simple convex quadratic functions.
  3. Marginal Infimum of jointly convex function:

    • infyYf(x,y)\inf_{y \in Y} f(x, y) is convex if f(x,y)f(x, y) is jointly convex and YY is convex.
    • This is especially useful in partial minimization, parametric optimization, and duality.
  4. Infimum convolution of two jointly convex functions:

    • infyYf(x,y)+g(y,z)\inf_{y \in Y} f(x, y) + g(y, z) is convex in (x,z)(x, z) if ff and gg are jointly convex and YY is convex.
    • This family of functions form a semigroup.

Understanding these operations sharpens your ability to reason about convexity in convex optimization.

Popular Posts