Min-Max Graph Partitioning and Small Set Expansion
Min-Max Graph Partitioning and Small Set Expansion
DOI. 10.1137/120873996
∗ Received by the editors April 18, 2012; accepted for publication (in revised form) February 13,
2013; published electronically April 29, 2014. An extended abstract of this paper appeared in Pro-
ceedings of 52nd Symposium on Foundations of Computer Science, IEEE Computer Society, Los
Alamitos, CA, 2011.
http://www.siam.org/journals/sicomp/43-2/87399.html
† Eindhoven University of Technology, 5600 MB Eindhoven, The Netherlands ([email protected]).
Work supported in part by the Netherlands Organisation for Scientific Research (NWO) grant
639.022.211.
‡ Weizmann Institute of Science, Rehovot, Israel ([email protected], robert.krauthgamer@
weizmann.ac.il). The second author was supported in part by The Israel Science Foundation (grant
873/08). The third author was supported in part by the Israel Science Foundation (grant 452/08),
the US-Israel BSF (grant 2010418), and by a Minerva grant.
§ Microsoft Research, One Microsoft Way, Redmond, WA 98052 (konstantin.makarychev@
microsoft.com, [email protected]).
¶ IBM T.J. Watson Research Center, P.O. Box 218, Yorktown Heights, NY 10598 (viswanath@
us.ibm.com).
Computer Science Dept., Technion, Haifa, Israel ([email protected]). Supported by ISF
Min-max partitions arise naturally in many settings. Consider the following appli-
cation in the context of cloud computing, which is a special case of the general graph-
mapping problem considered in [5] (and also implicit in other previous works [36, 35,
11]). There are n processes communicating with each other, and there are k ma-
chines, each having a bandwidth capacity C. The goal is to allocate the processes to
machines in a way that balances the load (roughly n/k processes per machine), and
meets the outgoing bandwidth requirement. Viewing the processes as vertices and the
traffic between them as edge weights, we get the min-max k-partitioning problem. In
general, balanced partitioning (either min-sum or min-max) is at the heart of many
heuristics that are used in a wide range of applications, including VLSI layout, circuit
testing and simulation, parallel scientific processing, and sparse linear systems.
Balanced partitioning, particularly in its min-sum version, has been studied exten-
sively during the last two decades, with impressive results and connections to several
fields of mathematics; see, e.g., [25, 13, 26, 4, 3, 20, 22, 9]. The min-max variants,
in contrast, have received much less attention. Kiwi, Spielman, and Teng [18] formu-
lated the min-max k-partitioning problem and designed algorithms achieving absolute
bounds for special classes of graphs such as meshes and planar graphs. Prior to our
work, no approximation algorithm for the min-max k-partitioning problem was given
explicitly for general graphs,
√ and the approximation that follows from known results
is not smaller than O(k log n).1 However, Raghavendra, Steurer, and Tulsiani [31,
Theorem IV.5] recently proved that a constant factor approximation for this problem,
even for k = 2, as well as a bicriteria approximation, is SSE-hard, meaning that it
would refute the so-called SSE hypothesis (which asserts a certain promise version of
the SSE problem defined below does not have a polynomial-time algorithm). A dif-
ferent min-max partitioning problem, min-max multiway cut was studied by Svitkina
and Tardos [34], who designed an O(log2 n) approximation algorithm.
An important tool in our result above is an approximation algorithm for the Small-
Set Expansion (SSE) problem. This problem was suggested recently by Raghavendra
and Steurer [32] (see also [30, 31]) in the context of the unique games conjecture.
Recall that the edge expansion of a subset S ⊆ V with 0 < |S| ≤ 12 |V | is
δ(S)
Φ(S) := .
|S|
The input to the SSE problem is an edge-weighted graph and ρ ∈ (0, 12 ], and the goal
is to compute
Φρ := min Φ(S).
|S|≤ρn
Raghavendra, Steurer, and Tetali [30] designed for the SSE problem in d-regular
graphs
an algorithm that approximates the expansion of small sets within a factor of
O( (d/Φρ ) log(1/ρ)), while violating the bound on |S| by no more than a constant
1 One could reduce the problem to the min-sum version of k-partitioning. The latter admits
√
bicriteria approximation O( log n log k) [20], but the reduction loses another factor of k/2. Another
possibility is to repeatedly remove n/k vertices from the graph, paying again a factor of k/2 on top
of the approximation in a single iteration, which is, say, O(log n) by [29].
general Φρ can be as small as 1/poly(n), in which case this guarantee is quite weak.
Using Räcke’s technique of converting a graph into a tree [29], one can obtain
an O(log n) approximation for SSE for any value of ρ, without √ violating the bound
|S| ≤ ρn.3 A better (bicriteria) approximation factor of O( log n) is achieved by
Arora, Rao, and Vazirani [3] at the price of slightly violating the size constraint.
However, this latter result works only for ρ = Ω(1). In our context of min-max
problems we need a solution to the case ρ =√1/k, where k = k(n) is part of the input.
Therefore, it is desirable to extend the O( log n) bound of [3] to a large range of
values for ρ.
1.1. Main results. Our two main results are bicriteria approximation algo-
rithms for the min-max k-partitioning and the SSE problems; here, the term bicriteria
means that the output generated by our algorithms violates the size constraints by at
most a multiplicative constant, while the cost of the solution is compared to that of
an optimal solution (that does satisfy the size constraints). The notation Oε (t) hides
multiplicative factors depending on ε, i.e., stands for O(f (ε) · t).
Definition 1.1. An (α, β) bicriteria approximate solution for the min-max k-
partitioning problem is a partition S1 , . . . , Sk of the input graph such that
1. max1≤i≤k δ(Si ) ≤ α · OPT (where OPT is the value of an optimal solution)
and
2. max1≤i≤k |Si | ≤ β · nk .
Theorem 1.1. For every positive √ constant ε > 0, min-max k-partitioning admits
a bicriteria approximation of (Oε ( log n log k), 2 + ε).
Theorem 1.1 provides a polynomial-time algorithm that with high probability
outputs
√ a partition S1 , . . . , Sk such that maxi |Si | ≤ (2 + ε) nk and maxi δ(Si ) ≤
O( log n log k)OPT. We note that the guarantee on part size can be improved slightly
to 2 − k1 + ε.
Definition 1.2. An (α, β) bicriteria approximate solution for the SSE problem
is a nonempty cut S ⊆ V such that
1. Φ(S) ≤ α · OPT (where OPT is the value of an optimal solution) and
2. |S| ≤ β · ρn.
Theorem 1.2. For every constant ε > 0, SSE admits a bicriteria approximation
of (Oε ( log n log (1/ρ)), 1 + ε).
Theorem 1.2 provides a polynomial-time algorithm that with high probability
outputs
a cut S ⊆ V of size 0 < |S| ≤ (1 + ε)ρn whose edge expansion is δ(S)/|S| =
O( log n log (1/ρ))OPT. We point out that our algorithm actually handles a more
general version, called weighted SSE, which is required in the proof of Theorem 1.1.
So in a sense, Theorem 1.1 is proved by reducing min-max k-partitioning to weighted
SSE. We defer the precise details to section 2.
1.2. Additional results and extensions.
1.2.1. ρ-unbalanced cut. Closely related to the SSE problem is the ρ-unbal-
anced cut problem. The input is again a graph G = (V, E) with nonnegative edge
weights and a parameter ρ ∈ (0, 12 ], and the goal is to find a subset S ⊆ V of size
2 The algorithm of [30] works also in general (nonregular) graphs for the closely related objective
of conductance.
3 For very small values of ρ, roughly ρn ≤ O(log2 n), a better approximation ratio is known [15].
|S| = ρn that minimizes δ(S). The relationship between this problem and SSE is
similar to the one between balanced cut and sparsest cut.
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
also section 5.
3. max1≤i≤k δ(Si ) ≤ C,
k
4. i=1 δ(Si ) ≤ D.
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
1.3. Techniques. For clarity, we restrict the discussion here mostly to our main
application, min-max k-partitioning. Our approach has two main ingredients. First,
we reduce the problem to a weighted version of SSE, showing that an α (bicriteria)
approximation for the latter can be used to achieve an O(α)
(bicriteria) approximation
for min-max k-partitioning. Second, we design an Oε ( log n log(1/ρ)) (bicriteria)
approximation for weighted SSE (our applications will use ρ = 1/k).
The SSE problem. Let us first examine SSE, and assume for simplicity of presen-
tation that ρ = 1/k. Note that SSE bears obvious similarity to both balanced cut
and min-sum k-partitioning—its solution contains a single cut (vertex subset) with
a size condition, as in balanced cut, but the size of this cut is n/k similarly to the
k pieces in min-sum k-partitioning. Thus, our algorithm is inspired by, but different
from, the approximation algorithms known for these two problems [3, 20]. As in these
two problems, we use a semidefinite programming (SDP) relaxation to compute an
22 metric on the graph vertices. However, new spreading constraints are needed since
SSE is highly asymmetric in its nature—it contains only a single cut of size n/k. We
devise a randomized rounding procedure based on the orthogonal separator technique,
first introduced by Chlamtac, Makarychev, and Makarychev [10] in the context of the
unique games problem. These ideas lead to an algorithm that √ computes a cut S of
expected size |S| ≤ O(n/k) and of expected cost δ(S) ≤ O( log n log k) times the
SDP value. An obvious concern is that both properties occur only in expectation and
might be badly correlated, e.g., the expected edge expansion E[δ(S)/|S|] might be ex-
tremely large. Nevertheless, we prove that with good probability, both |S| = O(n/k)
and δ(S)/|S| is sufficiently small.
tending the notion of orthogonal separators to linear programs (LPs) and designing
such “LP separators” of low-distortion for these special graph families. The proof
uses the probabilistic decompositions of Klein, Plotkin, and Rao [19] and Lee and
Sidiropoulos [23]. We believe that this result may be of independent interest—the LP
formulation for SSE is nontrivial and requires new spreading constraints. We remark
that even on planar graphs, the decomposition of Räcke [29] suffers an Ω(log n) loss
in the approximation guarantee, and thus does not yield an o(log n) ratio for SSE on
this class of graphs.
From SSE to min-max partitioning. Several natural approaches for designing an
approximation algorithm for min-max k-partitioning fail. First, reducing the problem
to trees à la Räcke [29] is not very effective, because there might not be a single tree
in the distribution that preserves all k cuts simultaneously. Standard arguments show
that the loss might be a factor of O(k log n) in the case of k different cuts. Second, one
can try and formulate a relaxation for the problem. However, the natural linear and
semidefinite relaxations both have large integrality gaps. As a case study, consider
for a moment min-max-multiway-cut. The standard linear relaxation of Calinescu,
Karloff, and Rabani [7] was shown by Svitkina and Tardos [34] to have an integrality
gap of k/2. In Appendix A we extend this gap to the semidefinite relaxation that
includes all 22 triangle inequality constraints. A third attempt is a greedy approach
that repeatedly removes from the graph, using SSE, pieces of size Θ(n/k). However,
by removing the “wrong” vertices from the graph, this process might arrive at a
subgraph in which every cut of Θ(n/k) vertices has edge weight greater by a factor
of Θ(k) than the original optimum (see Appendix B for details). Thus, a different
approach is needed.
Our approach. Our approach is to use multiplicative weight updates on top of
the algorithm for weighted SSE.√ This yields a collection S of sets S, each of size |S| =
Θ(n/k) and cost δ(S) ≤ O( log n log k)OPT, that covers every vertex v ∈ V at least
Ω(|S|/k) times. Alternatively, by assigning each set S ∈ S a fractional value of k/|S|,
we can view this as a fractional covering of vertices by valid configurations,
√ where valid
configurations correspond to sets S with |S| = Θ(n/k) and δ(S) = O( log n log k) ·
OPT.
Next, we randomly sample sets S1 , . . . , St from S until V is covered, and derive
a partition given by P1 = S1 , P2 = S2 \ S1 , and in general Pi = Si \ (∪j<i Sj ). This
method of converting a cover into a partition was previously used in [1, Theorem 2],
although it is somewhat counterintuitive, since the sets Pi may have a very large cost
δ(Pi ). We use the fact that S1 , . . . , St are chosen randomly (and in particular are in
random order), to show that the total expected boundary of the partition is not very
large, i.e.,
E δ(Pi ) ≤ O(k log n log k) OPT.
i
number of these sets might be very large, i.e., much larger then k. Thus, the last step
of the rounding is to merge small sets together. We show that this can be done while
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
maintaining simultaneously the constraints on the sizes and on the costs of the sets.
Organization. We first show in section 2 how to approximate weighted SSE in
both general and excluded-minor graphs, and then show (section 2.4) that an ap-
proximation algorithm for weighted SSE also yields an approximation algorithm for
weighted ρ-unbalanced cut. In section 3 we present an approximation algorithm for
min-max k-partitioning that uses the aforementioned algorithm for ρ-unbalanced cut.
The common generalization of both min-max k-partitioning and min-max-multiway-
cut, min–max cut, appears in section 4. Finally, Theorem 1.5, the hardness result for
min-max-multiway-cut, is proved in section 5.
Intuitively, η in the above definition is used to measure the sparsity of the cut;
thus, the objective is to minimize δ(S)/η(S) and it is required that η(S) > 0 (i.e., S
is not an empty cut with respect to η). On the other hand, μ is used to measure the
size of S, hence, it is required that μ(S) is not too large. Note that when both η and
μ are uniform measures, we obtain the usual SSE problem.
Throughout, we restrict attention to cases where Φρ,μ,η is defined (excluding, e.g.,
the case where μ, η are supported on the same single vertex). Part I of the following
theorem immediately implies Theorem 1.2 since we can choose both η and μ to be
the uniform measures.
Theorem 2.1 (approximating SSE). I. For every fixed ε > 0, there is a polynomial-
time algorithm that given as input an edge-weighted graph G = (V, w), two measures
μ and η on V (μ(V ) = η(V ) = 1), and some ρ ∈ (0, 1/2], finds a set S ⊆ V satisfying
η(S) > 0, μ(S) ≤ (1 + ε)ρ, and
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
δ(S)
≤ D · Φρ,μ,η (G),
η(S)
where D = Oε ( log n log(1/ρ)) and n = |V |.
II. Furthermore, when the input contains also a parameter H ∈ (0, 1), there is an
algorithm to find a set S ⊆ V satisfying η(S) ∈ [Ω(H), 2(1 + ε)H], μ(S) ≤ (1 + ε)ρ,
and
δ(S) δ(T )
≤ D · min : η(T ) ∈ [H, 2H], μ(T ) ≤ ρ ,
η(S) η(T )
where D = Oε ( log n log(max{1/ρ, 1/H})).
Remark 2.2. Let us clarify why we need part II of the theorem. In section 3.1, we
repeatedly apply this theorem to generate a family of sets S ∈ S that uniformly covers
all vertices in V . The measure μ will be uniform and ρ = k1 , corresponding to the
requirement that each part in min-max k-partitioning contains n/k vertices. Each set
S ∈ S must satisfy two conditions: μ(S) ≤ ρ and δ(S) ≤ D·OPT. The second measure
η is used to ensure that we uniformly cover all vertices: we dynamically update η
(starting with the uniform measure) so that yet uncovered vertices have larger weight.
We need the lower bound η(S) ≥ Ω(H) to ensure adequate new coverage. We also
need an upper bound on η(S) to ensure that δ(S) is small:
OPT
δ(S) ≤ D η(S) · ≤ 2(1 + ε) D · OPT.
H
We prove part I of the theorem in section 2.1, and part II in section 2.2. These
algorithms require the following notion of m-orthogonal separators due to Chlamtac,
Makarychev, and Makarychev [10].
Definition 2.3 (orthogonal separators). Let X be an 22 metric (i.e., a finite
collection of vectors satisfying 22 triangle inequalities) and m > 0. A distribution
over subsets S ⊆ X is called an m-orthogonal separator of X with distortion D > 0,
probability scale α > 0, and separation threshold 0 < β < 1, if the following conditions
hold:
1. for all u ∈ X we have Pr(u ∈ S) = α u 2;
2. for all u, v ∈ X with u − v 2 ≥ β min{||u 2 , v 2 },
We could constrain all vectors v̄ to have length at most 1, i.e., v̄ 2 ≤ 1, but it turns
out our algorithm never uses this constraint. We require that the vectors {v̄ : v ∈
V } ∪ {0}, i.e., all the vectors corresponding to V and the origin, satisfy the 22 triangle
inequalities. For every u, v, w ∈ V :
2 2
ū − w̄ ≤ ū − v̄ + v̄ − w̄ 2 ,
2 2
ū ≤ ū − v̄ + v̄ 2 ,
2 2
ū − w̄ ≤ ū + w̄ 2 .
Note that any valid integral solution (where each vector is either zero or a fixed unit
vector) satisfies these conditions. Suppose now that we have approximately guessed
the measure H of the optimal solution H ≤ η(S) ≤ 2H (this step is not necessary but
it simplifies the exposition). This can be done since the measure of every set S lies in
the range from η(u) to n · η(u), where u is the heaviest element in S, hence H can be
chosen from the set {2t η(u) : u ∈ V, t = 0, . . . , log2 n } of size O(n log n). Therefore,
we add a constraint:
(1) v̄ 2 η(v) ≥ H.
v∈V
We also ignore all vertices v ∈ V with η(v) > 2H, since they do not participate in the
optimal solution, and thus write the constraint (equivalent to removing them from
the graph)
2
(2) v̄ = 0, whenever η(v) > 2H.
Finally, we introduce new spreading constraints which state that for every u ∈ V ,
(3) μ(v) · min{ ū − v̄ 2 , ū 2 } ≥ (1 − ρ) ū 2 .
v∈V
5 Let u−v2 ≥ βu2 and assume w.l.o.g that u2 ≤ v2 . Using arithmetic one can prove that
u, v ≤ (1 − β/2)v2 . Applying Lemma 4.1 in [10] implies that ϕ(u), ϕ(v) ≤ (1 − β/2). Hence,
ϕ(u) − ϕ(v)2 ≥ β > 0 and using Corollary 4.6 in [10] finishes the proof since ψ(u) − ψ(v) ≥ 2γ =
√
β/4 > 0.
We remark that one could alternatively use a slightly simpler, almost equivalent,
constraint
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
ū, v̄ μ(v) ≤ ρ ū 2 .
v∈V
2
min w(u, v) ū − v̄
(u,v)∈E
2 2 2
s.t. ū − w̄ + w̄ − v̄ ≥ ū − v̄ ∀u, v, w ∈ V,
2 2 2
ū − v̄ ≥ ū − v̄ ∀u, v ∈ V,
2 2
ū + v̄ ≥ ū − v̄ 2 ∀u, v ∈ V,
μ(v) · min{ ū − v̄ 2 , ū 2 } ≥ (1 − ρ) ū 2
∀ u ∈ V,
v∈V
2
v̄ =0 whenever η(v) > 2H,
2
(4) η(v) v̄ ≥ H.
v∈V
Lemma 2.3. Assuming a correct guess of H, the optimal value of SDP (4) is at
most 2H · Φρ,μ,η (G).
Proof. Let S ∗ be an optimal solution to the SSE problem, i.e., Φρ,μ,η (G) =
δ(S )/η(S ∗ ). Construct the following solution to SDP (4). If v ∈ S ∗ set v̄ = 1 for
∗
some fixed unit vector, and if v ∈ S ∗ set v̄ = 0. Assuming that H was guessed correctly
(i.e., H ≤ η(S ∗ ) ≤ 2H), each vertex v ∈ S ∗ has η(v) ≤ 2H, and constraint (1) is also
satisfied. It is clear that all the 22 triangle inequality constraints with the origin are
satisfied. Let us focus now on the spreading constraints. If u ∈ S ∗ , then ū = 1 and
the sum in the spreading constraint equals μ(V \ S) ≥ 1 − ρ. Otherwise, if u ∈ / S∗,
then ū = 0 and both sides of the spreading constraint equal 0. Hence, we can conclude
that the solution we defined for (4) is feasible.
Note that for every edge (u, v) ∈ E, if the edge crosses the cut defined by S ∗
then it contributes its weight w(u, v) to the sum in the objective of (4). Otherwise,
its contribution is 0. Thus, δ(S ∗ ) equals the objective value of this SDP solution and
the lemma follows.
We now describe the approximation algorithm.
Approximation Algorithm I. We first informally describe the main idea behind
the algorithm. The algorithm solves the SDP relaxation and obtains a set of vectors
{ū}u∈V , whose value we denote by SDP . Now it samples an orthogonal separator, a
random set S ⊆ V , and returns it. Assume for the moment that the probability scale
α equals 1. Since Pr(v ∈ S) = v̄ 2 , we get E[η(S)] ≥ H. The expected size of the
cut is at most D · SDP by the third property of orthogonal separators; and thus the
ratio of expectations
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
E[δ(S)] SDP
≤D· ≤ 2D · Φρ,μ,η (G).
E[η(S)] H
We will show that the expected ratio is also bounded by a similar quantity. Moreover,
using the second property of orthogonal separators and the spreading constraints, we
can show that μ(S) ≤ (1 + ε)ρ. We now proceed to the formal argument.
We may assume that ε is sufficiently small, i.e., ε ∈ (0, 1/4). The approximation
algorithm guesses a value H satisfying H ≤ η(S) ≤ 2H. Then, it solves the SDP
while further constraining ū = 0 for every u ∈ V with η(u) > 2H, to obtain a
set of vectors X = {v̄}v∈V . After that, it samples an orthogonal separator S with
m = ε−1 ρ−1 and β = ε (from Theorem 2.2). For convenience, we let S be the set of
vertices corresponding to vectors belonging to the orthogonal separator rather than
the vectors themselves. The algorithm repeats the previous step α−1 n2 times (recall
α is the probabilistic scale of the orthogonal separator). It outputs the set S having
minimum δ(S)/η(S) value among those satisfying 0 < μ(S) < (1 + 10ε)ρ. If no
S satisfies this constraint (which we shall see happens with an exponentially small
probability), the algorithm outputs an arbitrary set satisfying the constraints.
Analysis. We first estimate the probability of the event {u ∈ S and μ(S) < (1 +
10ε)ρ} for a fixed vertex u ∈ V . Let Au = {v : ū − v̄ 2 ≥ β ū 2 } be all the vertices
which are far from u and Bu = {v : ū − v̄ 2 < β ū 2 } all the vertices which are close
to u. Assume for now that u ∈ S, and let us show that only a small fraction of Au
is likely to belong to S, and that the entire set Bu has small measure μ(Bu ). Recall
that u satisfies the spreading constraint,
(5) min{||ū − v̄ 2 , ū 2 } · μ(v) ≥ (1 − ρ) ū 2 .
v∈V
As the event {u ∈ S and μ(S) < (1+10ε)ρ} holds for each vertex u with probability at
least 34 α ū 2 , and by (6) we have E[δ(S)] ≤ αD · SDP , we can bound the expectation
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
as
3α ū 2 η(u) αH αH
E[f (S)] ≥ − ≥ .
4 4 2
u∈V
The second inequality uses the last SDP constraint u∈V ū 2 η(u) ≥ H. The random
variable f (S) is always upper bounded by 2nH, because by the SDP constraints,
ū = 0 whenever η(u) > 2H. We note that the zeroing of all u ∈ V such that
η(u) > 2H is not essential to the algorithm. However, this zeroing enables us to
obtain a better bound on the expected value of f , which affects only the number of
samples of orthogonal separators the algorithm needs to perform.
Thus, Pr[f (S) > 0] ≥ α/(4n). Therefore, after α−1 n2 iterations, with probability
exponentially close to 1, the algorithm will find S with f (S) > 0. The latter implies
that η(S) > 0, μ(S) < (1 + 10ε)ρ, and
δ(S) SDP
≤ 4D · .
η(S) H
This finishes the proof of Theorem 2.1 part I, since SDP/(2H) ≤ Φρ,μ,η (G).
2.2. Approximation Algorithm II: Small-set expansion in general graphs.
We now prove part II of Theorem 2.1.
SDP relaxation. Our SDP relaxation is similar to part I, namely, we use SDP (4)
with the following additional constraints. Recall that Approximation Algorithm II
gets H (an approximate value of η(S) in the optimal solution) as input, and thus
does not need to guess it. First, we add a spreading constraint also on the η-measure
of the solution
(7) min{ ū − v̄ 2 , ū 2 } · η(v) ≥ (1 − 2H) ū 2 ∀u ∈ V.
v∈V
(8) v̄ ≤ 1 ∀v ∈ V,
(9) v̄ 2 μ(v) ≤ ρ.
v∈V
It is not difficult to verify (analogously to Lemma 2.3) that this SDP is indeed a
relaxation of the problem stated in Theorem 2.1 part II.
Approximation Algorithm II. The algorithm solves the SDP relaxation and then
executes several iterations. Each iteration computes a vertex subset S by using a
variant of Approximation Algorithm I from section 2.1 detailed below, and then “re-
moves” this S from the graph G and the SDP solution, by zeroing the weight of edges
incident to S (i.e., discarding these edges) and zeroing the SDP vectors corresponding
to vertices in S (i.e., setting them to the origin 0). The algorithm maintains the ver-
tices removed so far in a set T ⊂ V , by initializing T = ∅, and then at each iteration
updating T = T ∪ S. The iterations repeat until either μ(T ) ≥ ρ/4 or η(T ) ≥ H/4,
at which point the algorithm returns a vertex subset F determined as follows: if
μ(T ) ≤ ρ and η(T ) ≤ H, then F = T ; otherwise, F = S (as computed in the last
iteration).
(Alternatively, one can return an arbitrary feasible S if the number of samples exceeds
a certain threshold.)
Remark 2.4. To handle terminals in the extended version of the problem (see
section 4), we guess a single terminal u that belongs to the optimal solution S (if
any), and set ū = 1 and v̄ = 0 for all other terminals v. Since an orthogonal
separator never contains the zero vector, the returned solution F will always contain
at most one terminal.
Analysis. Let us examine the effect of the algorithm’s changes to the SDP solution
(by zeroing some vectors and removing edges). Notice that the objective value of the
SDP may only decrease, and that constraints (8), (9) and the triangle inequalities
clearly remain satisfied. Constraint (1) may get violated, but not by too much; the
removed vertices have total measure η(T ) ≤ H/4, and due to (8) we still have the
slightly weaker constraint
(1 ) ū 2 η(u) ≥ 3H/4.
u∈V
Let us show that the spreading constraints (3) and (7) are still satisfied. Consider the
constraint (3) for a given vertex u ∈ V :
μ(v) · min{ ū − v̄ 2 , ū 2 } ≥ (1 − ρ) ū 2 .
v∈V
Notice that f has an extra term and an extra condition involving η(S) when compared
to f .
Consider now a single sampling of S from the orthogonal separator. The same
argument as before shows, using constraints (5) and (7), that for every given u ∈ V ,
3
Pr μ(S) ≤ (1 + 10ε)ρ and η(S) ≤ 2(1 + 10ε)H | u ∈ S ≥ .
4
Using the new constraint (9), we get E[μ(S)] ≤ αρ, and recalling the bound (6) on
E[f (S)] ≥ − − ≥ − ≥ ,
4 4 4 4 2 16
u∈V
where the second inequality used (1’). As before, we can bound Pr[f (S) > 0] ≥
Ω(α/n), because with probability 1 we have f (S) ≤ η(S) ≤ 2nH (recall that ū = 0
whenever η(u) > 2H). So after O(n2 /α) samples, with probability that is exponen-
tially close to 1, the algorithm finds a set of vertices S satisfying f (S) > 0.6
At every iteration we have f (S) > 0, which implies that
SDP
(11) δ(S) ≤ 4D · η(S),
H
(12) η(S) ≥ H · μ(S)/(4ρ),
Let T denote the set T at the end of the algorithm, and let T denote the
penultimate set T . Also let S denote the last set S; so T = T ∪ S . Note that
μ(T ) < ρ/4 and η(T ) < H/4 by the iterations’ stopping condition. In addition,
either μ(T ) ≥ ρ/4 or η(T ) ≥ H/4, but in both cases η(T ) ≥ H/16 follows from (13).
We consider the two possible outputs.
• F = T . This implies μ(T ) ≤ ρ and η(T ) ≤ H. And also η(T ) ≥ H/16
from above.
• F = S . This means μ(T ) > ρ or η(T ) > H, which imply, respectively,
μ(S ) > 34 ρ or η(S ) > 34 H; in either case, η(S ) > 3H/16 follows from (12).
Also our algorithm ensures that μ(S ) ≤ (1 + 10ε)ρ and η(S ) ≤ 2(1 + 10ε)H.
We thus see that either output F satisfies μ(F ) ≤ (1 + 10ε)ρ, η(F ) ≤ 2(1 + 10ε)H,
and η(F ) ≥ H/16. Finally, inequality (11) holds for every set S added, and thus
a similar inequality holds for the output F as well. This completes the proof of
Theorem 2.1 part II.
2.3. Small-set expansion in minor-closed graph families. In this subsec-
tion we prove Theorem 1.7. We start by writing the LP relaxation (14). For every
vertex u ∈ V we introduce a variable x(u) taking values in [0, 1]; and for every pair of
vertices u, v ∈ V we introduce a variable z(u, v) = z(v, u) also taking values in [0, 1].
The intended integral solution corresponding to a set S ⊆ V has x(u) = 1 if u ∈ S,
and x(u) = 0 otherwise; z(u, v) = |x(u) − x(v)|. Intuitively, x(u) is the distance to a
fictional vertex o that never belongs to S. The analogous object in the SDP relaxation
(4) is the origin, and indeed, it is instructive to think of x(u) as an analogue of ū 2
and of z(u, v) as an analogue of ū − v̄ 2 .
It is easy to verify that LP (14) below is a relaxation of the SSE problem. The
first three constraints say that z(u, v) is a metric (strictly speaking, a semimetric), by
requiring the triangle inequality, even with the fictional vertex o. The novelty of the
LP is in the fourth constraint, which is a new spreading constraint for ensuring that
the size of S is small.
6 In fact, now f (S) ≤ 2(1 + 10ε)H, thus Pr[f (S) > 0] ≥ Ω(α/H) and we need only O(n/αH)
iterations.
min w(u, v) z(u, v)
(u,v)∈E
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
u,v∈S
z(u, v) if u, v ∈ V ,
(u, v) :=
x(u) if u ∈ V and v = o.
For ease of notation, let the modified graph (V , E ) also be called G, and let dG
denote the shortest-path metric according to these edge lengths. It follows from the
LP’s triangle inequalities that z(u, v) ≤ dG (u, v) for all u, v ∈ V . The following
algorithm outputs a subset S ⊂ V which is an LP separator.
1. Choose δ ∈ [0, 1] uniformly at random, and sample a random partition Pδ
from Theorem 2.4 applied to the metric (V , dG ) and parameter Δ = βδ/2.
2. Let S1 , . . . , St ∈ Pδ denote the clusters that contain at least one vertex u ∈ V
with x(u) ∈ [δ, 2δ]. Clearly t ≤ |V |.
3. Output cluster S = Si with probability 1/|V | for each i ∈ {1, . . . , t}; with the
remaining probability 1 − t/|V | output the empty cluster S = ∅.
Let us show that the output cluster S is an LP separator with α = 1/(2|V |) and
D = O(r2 /β). For the first requirement from an LP separator, we see that for every
u∈V,
1 x(u)
Pr[u ∈ S] ≥ Pr x(u) 2 ≤ δ ≤ x(u) · = = α · x(u).
|V | 2|V |
Observe that conditioned on δ, if the output cluster S = ∅ then there is some
w ∈ S with x(w) ∈ [δ, 2δ] by step 2. Since the diameter of cluster S under the
metric dG is bounded by Δ = βδ/2, for all v ∈ S we have |x(w) − x(v)| ≤ z(w, v) ≤
dG (w, v) ≤ βδ/2, and therefore
x(u) from inequality (15) and (ii) dG (u, v) ≤ Δ = βδ/2 by the diameter bound on
x(u)
S. The former implies δ ≤ 1−β/2 < 2x(u) since β < 1, while the latter implies
2
δ ≥ β · dG (u, v) ≥ 2x(u). We conclude that Pr[u, v ∈ S] = 0.
For the third requirement from an LP separator, we will bound Pr[u ∈ S, v ∈ S]
for an edge (u, v) ∈ E, the other (symmetric) case being identical. Using inequal-
ity (15),
1 x(u) x(u)
|V | if 2+β/2 ≤δ≤ 1−β/2 ,
Pr[ u ∈ S | δ ] ≤
0 otherwise.
Using Bayes’ rule, then the fact that Pδ is a separating decomposition, and then that
dG (u, v) ≤ (u, v) = z(u, v) (because it is an edge, we in fact have equality),
dG (u, v) 2D0 · z(u, v)
Pr v ∈ S | u ∈ S, δ ≤ Pr Pδ (u) = Pδ (v) | δ ≤ D0 ≤ ,
βδ/2 βδ
where D0 = O(r2 ) is from Theorem 2.4. Combining the last two bounds, and recalling
that δ ∈ [0, 1] uniformly at random and β < 1,
2x(u)
1 2D0 · z(u, v) 2 ln 6 · D0 · z(u, v)
Pr[ u ∈ S, v ∈ S ] ≤ dδ = .
x(u)/3 |V | βδ β|V |
We thus have
4 ln 6 · D0 · z(u, v) 8 ln 6 · D0
Pr[IS (u) = IS (v)] ≤ =α z(u, v),
β|V | β
and
Furthermore,
where OPT denotes the optimal value of the min-max k-partitioning instance.
2. for all v ∈ V ,
|{S ∈ S : v ∈ S}| 1
≥ .
|S| (5γk)
√
Above, α = Oε ( log n log k), β = 1 + ε, and γ = O ε (1), for any fixed ε > 0.
Proof. For an iteration t, let us denote Y t := v∈V y t (v). The first assertion of
the theorem is immediate from the following claim.
Claim 3.2. Every iteration t of Algorithm 1 satisfies: δ(S t ) ≤ α · OPT and
|S t | ≤ β · n/k.
Proof. We claim that no matter what the values of the y t ’s are, the optimal
solution to the weighted ρ-unbalanced cut instance G, y t , w, k1 , k1 is at most OPT.
To see this, consider the optimal solution {Si∗ }ki=1 of the original min-max k-partitio-
ning instance. We have |Si∗ | ≤ n/k and w(δ(Si∗ )) ≤ OPT for all i ∈ [k]. Since {Si∗ }ki=1
partitions V , there is some j ∈ [k] with y t (Sj∗ ) ≥ Y t /k. It now follows that Sj∗ is
a feasible solution to the weighted ρ-unbalanced cut instance G, y t , w, k1 , k1 , with
objective value at most OPT, which proves the claim.
We proceed to prove the second assertion of Theorem 3.1. Let denote the
number of iterations of the while loop, for the given min-max k-partitioning instance.
For any v ∈ V , let Nv denote the number of iterations t in which v ∈ S t . Then,
by the updating rule for the y weights we have that y +1 (v) = 1/2Nv . Moreover,
the termination condition implies that y +1 (v) ≤ 1/n (since Y +1 ≤ 1/n). Thus, we
obtain that Nv ≥ log2 n for all v ∈ V . From the approximation guarantee of the
weighted ρ-unbalanced cut algorithm (Corollary 2.6), it follows that y t (S t ) ≥ γ1k · Y t
in every iteration t. Thus,
t+1 t 1 t t 1
Y = Y − · y (S ) ≤ 1 − · Y t.
2 2γ k
However, Y > 1/n since the algorithm performs iterations. Thus, ≤ 1+4γ k·ln n ≤
5γ k · log2 n. This finished the proof since
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
xS ≥ 0 ∀S ∈ Fλ .
The goal is to determine the smallest λ > 0 such that P(λ) ≤ k. Since the config-
uration LP (16) has an exponential number of variables, we formulate its dual and
present a separation oracle for it:
D(λ) = max yv
v∈V
(17) s.t. yv ≤ 1 ∀S ∈ Fλ ,
v∈S
yv ≥ 0 ∀v ∈ V.
One can (approximately) solve the dual formulation since a separation oracle which
uses weighted SSE exists. The details for approximating the configuration LP are
rather technical (since we only have a bicriteria approximation for weighted SSE, as
opposed to a usual multiplicative approximation) and do not offer any advantage over
the multiplicative updates method except for an alternative intuition regarding the
issue. Thus, they are omitted.
3.2. Aggregation. The aggregation process, which might be of independent
interest, transforms a cover of G into a partition, without violating the min-max
objective by much. In particular, we show the following.
Theorem 3.3. There is a randomized polynomial-time algorithm (Algorithm 2
below) that when given a graph G = (V, E), an ε ∈ (0, 1), and a cover S consisting of
subsets of V such that
1. every vertex in V is covered by at least a fraction c/k of sets S ∈ S for some
c ∈ (0, 1], and
2. each S ∈ S satisfies |S| ≤ 2n/k and δ(S) ≤ B,
outputs a partition P of V into at most k sets such that for all P ∈ P we have
|P | ≤ 2(1 + ε)n/k and E[maxP ∈P δ(P )] ≤ 8B/(cε).
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
and E[w(E(Pi , ∪j>i Pj ))] ≤ (1 − c/k)i−1 B. Therefore, ∞the total expected weight of
i
edges
crossing the boundary of the Pi ’s is at most i=0 (1 − c/k) B = kB/c, and
E[ i δ(Pi )] ≤ 2kB/c. This finishes the analysis of Step 1.
Let us now analyze Step 2. Our potential function is i δ(Pi ), whose initial value
was bounded in Step 1. We prove that this potential function reduces quickly over the
iterations of Step 2, thus, Step 2 terminates after a “small” number of steps. After
each iteration of Step 2, the following invariant holds: the collection of sets {Pi } is
a partition of V and Pi ⊂ Si for all i. Particularly, |Pi | ≤ |Si | ≤ 2n/k. The key
observation is that at every iteration of the “while” loop, the sum j δ(Pj ) decreases
The above inequalities use the facts that Pi ⊆ Si for all i and that all the Pj ’s are
disjoint. Hence, the number of iterations of the loop in Step 2 is always polynomially
bounded and after the last iteration E[ i δ(Pi )] ≤ 2kB/c (the expectation is over
random choices at Step 1; Step 2 does not use random bits). Hence, E[B ] ≤ 4B/c;
1
recall that B = max{ k i δ(Pi ), 2B}.
Let us now analyze Step 3. The following analysis holds conditional on any value
of B . After each iteration of Step 3, the following invariant holds: the collection of
sets {Pi } is a partition of V , and for all i, |Pi | ≤ 2(1 + ε)n/k and δ(Pi ) ≤ 2B ε−1 .
Clearly, after Step 2, δ(Pi ) ≤ 2B ≤ B and |Pi | ≤ 2n/k for all i. Moreover, the sum
i δ(Pi ) can only decrease in Step 3.
When the loop terminates, we obtain a partition
of V into sets Pi satisfying
|Pi | ≤ 2(1 + ε)n/k, i |Pi | = n, δ(Pi ) ≤ 2B ε−1 , i δ(Pi ) ≤ kB , such that no two
sets can be merged without violating the above constraints. Hence, by Lemma 3.4
below (with ai = |Pi | and bi = δ(Pi )), the number of nonempty sets is at most
n kB
2 + = (1 + ε)−1 k + (ε/2)k ≤ k.
2(1 + ε)n/k 2B ε−1
Proof.
t By rescaling we assume that A = 1 and B = 1. Moreover, we may assume
t
that i=1 ai < S and i=1 bi < T by slightly decreasing the values of all ai and bi
so that all inequalities still hold.
We write two linear programs:
(LPI ) min xi
i
s.t. xi + xj ≥ 1 ∀(i, j) ∈ {(i, j) : ai + aj ≥ 1},
xi ≥ 0 ∀i,
s.t.
yi ≥ 0 ∀i.
The first LP (LPI ) has variables xi and constraints xi + xj ≥ 1 for all i, j such
that ai + aj ≥ 1. The second LP (LPII ) has variables yi and constraints yi + yj ≥ 1
such that bi + bj ≥ 1. The LP objectives are to minimize i xi and to
for all i, j
minimize i yi . Note, that {ai } is a feasible point for LPI and {bi } is a feasible point
for LPII . Thus, the optimum values of LPI and LPII are strictly less than S and T ,
respectively.
Observe that both LPs are half-integral. Consider optimal solutions x∗i , yj∗ where
xi , yj ∈ {0, 1/2, 1}. Note that for every i, j either x∗i +x∗j ≥ 1 or yi∗ +yj∗ ≥ 1. Consider
∗ ∗
t
several cases. If for all i, x∗i + yi∗ ≥ 1, then t < S + T , since i=1 (x∗i + yi∗ ) < S + T .
If for some j, x∗j + yj∗ = 0 (and hence x∗j = yj∗ = 0), then x∗i + yi∗ ≥ 1 for i = j and,
thus, t < S + T + 1. Finally, assume that for some j, x∗j + yj∗ = 1/2, and w.l.o.g.,
x∗j = 1/2 and yj∗ = 0. The number of i’s with x∗i = 0 is (strictly) bounded by 2S.
For the remaining i’s, x∗i = 0 and hence yi∗ = 1 (because yi∗ = yi∗ + yj∗ ≥ 1), and
thus the number of such i’s is (strictly) bounded by T . This finishes the proof of the
lemma.
4. Further extensions. Both Theorems 1.1 and 1.4 follow from a more general
result for a problem that we call min–max cut. This problem models the aforemen-
tioned cloud computing scenario, where in addition, certain processes are preassigned
to machines (each set Ti maps to machine i ∈ [k]). The goal is to assign the processes
V to machines [k] while respecting the preassignment and machine load constraints,
and minimizing both bandwidth per machine and total volume of communication.
It is clear that in fact Theorem 1.6 generalizes both Theorems 1.1 and 1.4. Let us
now describe modifications to the min-max k-partitioning algorithm used to obtain
Theorem 1.6.
Uniform coverings. First, by the introduction of vertex weights, we can shrink
each preassigned set Ti to a single terminal ti (for i ∈ [k]). Then, feasible vertex sets C
in the covering procedure (section 3.1) consist of those S ⊆ V where weight(S) ≤ ρ n
(balance constraint) and |S ∩ {ti }ki=1 | ≤ 1 (preassignment constraint). The subprob-
lem weighted ρ-unbalanced cut also has the additional |S ∩ {ti }ki=1 | ≤ 1 constraint;
this can be handled in the Approximation Algorithm II from section 2 by guess-
ing which terminal belongs to S (see Remark 2.4). Using Corollary 2.6 we assume
an (α, β, γ) approximation algorithm forthis (modified) weighted ρ-unbalanced cut
problem, where for any ε > 0, α = Oε ( log n log(max{1/ρ, 1/τ })), β = 1 + ε, and
γ = Oε (1).
Algorithm 3 below gives the procedure to obtain a uniform covering S bounding
total edge cost in addition to the conditions in Theorem 3.1.
Theorem 4.1. For any instance of min–max cut, output S of Algorithm 3 sat-
isfies
1. δ(S) ≤ α · C and |S| ≤ β · nk for all S ∈ S;
2. |{S ∈ S : v ∈ S}| ≥ log2 n for all v ∈ V ;
3. |S|
≤ 5γ k · log2 n;
4. S∈S δ(S) ≤ 17α γ log√2 n · D.
Above, for any ε > 0, α = Oε ( log n log k), β = 1 + ε, and γ = Oε (1).
Proof. We start with the following key claim.
while Y t > n1 do
for i = 0, . . . , log2 k + 1 do
Solve the weighted ρ-unbalanced cut instance G, y t , w, 21i , ρ using
the algorithm from Corollary 2.6, to obtain S t (i) ⊆ V .
If δ(S t (i)) ≤ α · min{C, 4D/2i } then S t ← S t (i) and quit this loop.
Add S t to S.
for v ∈ V do
Set y t+1 (v) ← 12 · y t (v) if v ∈ S t , and y t+1 (v) ← y t (v) otherwise.
Set Y t+1 ← v∈V y t+1 (v).
Set t ← t + 1.
return S.
k
Yt Yt
y t (Sj∗ ) ≤ · δ(Sj∗ ) ≤ · δ(Sj∗ ) ≤ Y t /2.
2D 2D j=1
j ∈L j ∈L
t
Since |L| ≤ k, there is some q ∈ L with y t (Sq∗ ) ≥ Y2k . Let i ∈ {1, . . . , log2 k + 1}
be the value such that y t (Sq∗ )/Y t ∈ [ 21i , 2i−1
1
]; note that such an i exists because
t ∗ t 1
y (Sq )/Y ∈ [ 2k , 1]. For this i, consider the weighted ρ-unbalanced cut instance
G, y t , w, 2−i , ρ. Observe that Sq∗ is a feasible solution here since y t (Sq∗ ) ≥ Y t /2i ,
|Sq∗ | ≤ ρn, and Sq∗ contains at most one terminal. Hence, the optimal value of this
instance is at most:
2D 4D
δ(Sq∗ ) ≤ min C, t · y t (Sq∗ ) ≤ min C, i .
Y 2
The first inequality uses the definition of L and that δ(Sq∗ ) ≤ C, and the second
inequality is by the choice of i. It now follows from Corollary 2.6 √ that the solu-
tion S t (i) satisfies the claimed properties. We note that α = Oε ( log n log k) be-
cause each instance of weighted ρ-unbalanced cut has parameters τ = 2−i ≥ 1/k and
ρ ≥ 1/k.
For each iteration t, let it ∈ {1, . . . , log2 k + 1} be the index such that S t = S t (it ).
Claim 4.2 implies that such an index always exists, and so the algorithm is indeed well
defined. Condition 1 of Theorem 4.1 also follows directly. Let denote the number of
iterations of the while loop, for the given min–max cut instance. For any v ∈ V , let
Nv denote the number of iterations t in which v ∈ S t . Then, by the update rule of
the weights y we have y +1 (v) = 1/2Nv . Moreover, the termination condition implies
that y +1 (v) ≤ 1/n (since Y +1 ≤ 1/n). Thus we obtain Nv ≥ log2 n for all v ∈ V ,
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
We need to prove that condition 3 and condition 4 also hold. First, using the fact
that y t (S t ) ≥ Y t /(2γ k) in each iteration, we can deduce that
t+1 1 1
Y = Y − · y t (S t ) ≤
t
1− · Y t.
2 4γ k
However, Y > 1/n since the algorithm performs iterations. Thus, ≤ 1+8γ k·ln n ≤
9γ k · log2 n. This proves condition 3 in Theorem 4.1.
Yt
Second, using y t (S t ) ≥ 4αγ t
D · δ(S ) in each iteration, Y
t+1
= Y t − 12 · y t (S t ) ≤
δ(S t )
(1 − 8αγ D ) · Y t . So,
t −1 t
1 −1 δ(S ) δ(S )
< Y ≤ Πt=1 1− · Y 1 ≤ exp − t=1 ·n
n 8αγ D 8αγ D
This implies:
−1
δ(S t ) ≤ (16αγ ln n) · D.
t=1
parts to k, we merge arbitrarily each part containing a terminal with one nonterminal
part, and output this as the final solution. It is clear that √ each part has at most one
terminal, has size ≤ (3 + ε)ρn, and cut value at most Oε ( log n log k) · C. The bound
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
on total cost (condition 4 in Theorem 1.6) is by the following claim. This proves a
weaker version of Theorem 1.6, with size bound (3 + ε)ρn.
Claim 4.3. Algorithm √ S from Theorem 4.1 outputs parti-
2k applied on collection
tion {Pi }ki=1 satisfying E[ i=1 w(δ(Pi ))] = Oε ( log n log k) D.
Proof. We will show that the random√partition {Pi } at the end of Step 1 in
Algorithm
2 satisfies E [ i w(δ(Pi ))] ≤ Oε ( log n log k) D. This would suffice, since
i w(δ(P i )) does not increase in Steps 2 and 3. For notational convenience, we assume
(by adding empty sets) that |S| = 5γ k · log2 n in Theorem 4.1; note that this does
not affect any of the other conditions.
To bound the cost of the partition {Pi } in Step 1, consider any index i ≤ |S|.
From the proof of Theorem 3.3, we have:
where c = 1/(5γ) is such that each vertex lies in at least a fraction c/k of sets S. We
remove the conditioning,
i−1 i−1
E[w(E(Pi , ∪j>i Pj ))] ≤ (1 − c/k) · E[δ(Si )] = (1 − c/k) · δ(S)/|S| ,
S∈S
where we used that Si is a uniformly random set from S. Therefore, the total edge
cost can be upper bounded as follows:
E δ(Pi ) = 2 · E[w(E(Pi , ∪j>i Pj ))]
i i
⎛ ⎞
≤⎝ i⎠
(1 − c/k) · δ(S)/|S|
i≥0 S∈S
k
= · δ(S)/|S|.
c
S∈S
Using S∈S δ(S) ≤ 17αγ log2 n · D and |S| = 5γ k · log2 n from Theorem 4.1, we can
conclude that
17α
E δ(Pi ) ≤ D = Oε ( log n log k) D
i
5c
√
since α = Oε ( log n log k) and 1/c = Oε (1).
Obtaining size bound of (2 + ε)ρn. We now describe a modified aggre-
gating step (replacing Step 3 of Algorithm 2) that yields the promised guarantee of
Theorem 1.6. Given the uniform cover S from Algorithm 3, run Steps 1 and 2 of
Algorithm 2 (use B = αC) to obtain parts P1 , . . . , P|S| . Then perform the following.
1. Set B := max{ k1 i δ(Pi ), 2B}.
2. While there are Pi , Pj = ∅ (i = j) such that |Pi | + |Pj | ≤ (1 + ε)ρn, δ(Pi ) +
δ(Pj ) ≤ 2B , and Pi ∪ Pj does not contain a terminal: replace Pi ← Pi ∪ Pj
and Pj ← ∅.
These two steps are formalized in the two foregoing Lemmas 5.1 and 5.3. Putting
them together immediately proves Theorem 1.5.
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
B n B
OPT(I) + n · + · (k − 1) · ≤ OPT(I) + 2B,
n k n
the term OPT(I) is due to edges in E, the second term is due to edges at ti , and
the third is due to edges at all other {tj : j = i}. The claim now follows since
B ≤ 2 OPT(I).
Let P denote A’s solution to J (B), and {Pi }ki=1 the partition of V induced
by P . From Claim 5.2, P has objective value (for min-max-multiway-cut) at most
5ρ OPT(I). Note that the boundary (in graph GB ) of ti ’s piece in P costs at least
c(δG (Pi )) + (k − 1) B
n · |Pi |. Thus we obtain
B
c(δG (Pi )) + (k − 1) · |Pi | ≤ 5ρ OPT(I), for all i ∈ [k].
n
Proof. Let A denote the (α, k 1−ε ) bicriteria approximation algorithm. The idea is
to use A recursively to obtain the claimed (α · log log k, γ) approximation; details are
below. Let G denote the input graph with n vertices, and OPT the optimal balanced
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
The second-last inequality uses ki−1 ≥ 32/ε which is true by the choice of t. Now we
can fill pieces of Q with dummy singleton vertices of instance I to obtain a balanced
ki -partition of I.
This completes the proof of Lemma 5.3.
problem:
min λ
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
(18)
s.t. λ≥ ||yu,i − yv,i ||22 ∀i = 1, 2, . . . , k,
(u,v)∈E
k
(19) ||yu,i ||22 = 1 ∀u ∈ V,
i=1
(20) ||yti ,i ||22 = 1 ∀i = 1, 2, . . . , k,
(21) yu,i · yu,j = 0 ∀u ∈ V, ∀i = j,
k
(22) yu,i · yv,j = ||yu,i ||22 ∀u, v ∈ V, ∀i = 1, 2, . . . , k,
j=1
(23) ||yu,i − yv,j ||22 + ||yv,j − yw,r ||22 ≥ ||yu,i − yw,r ||22 ∀u, v, w ∈ V,
∀i, j, r = 1, 2, . . . , k
(24) yu,i · yv,j ≥ 0 ∀u, v ∈ V, ∀i, j = 1, 2, . . . , k,
(25) ||yu,i ||22 ≥ yu,i · yv,j ∀u, v ∈ V, ∀u, v ∈ V,
∀i, j = 1, 2, . . . , k.
In the above relaxation, each vertex u ∈ V is associated with k different vectors:
yu,1 , yu,2 , . . . , yu,k . Each of those corresponds to a different terminal. Notice that
the last two constraints of the relaxation are the 22 triangle inequality constraints
including the origin (the same constraints used by [10]).
As stated, the integrality gap we consider is the star graph which contains a single
vertex u connected by k edges to the k terminals. The value of any integral solution
to this instance is exactly k − 1 since the only choice is to which terminal should u be
assigned. Any choice made results in a min-max objective value of k − 1.
Let us construct the fractional solution to the above relaxation. Fix e to be an
arbitrary unit vector, and x1 , x2 , . . . , xk to be k unit vectors which are all orthogonal
to e and the inner product between any two of them is −1/(k − 1). Set the following
fractional solution:
yti ,i = e ∀i = 1, 2, . . . , k,
yti ,j = 0 ∀i = j,
√
1 k−1
yu,i = e+ xi ∀i = 1, 2, . . . , k.
k k
First, we show that the above fractional solution is feasible. Constraints (19), (20),
and (21) are obviously feasible for all terminals. Vertex u also upholds constraint (19)
since
k $$ √ $$2 k
$$ 1 $$
$$ e + k − 1 xi $$ = 1
+
k−1
= 1.
$$ k k $$ k2 k2
i=1 2 i=1
Focus on constraint (22). It is easy to verify that for all terminals, the sum of all
k
k vectors that are associated with the picked vertex is exactly e. Since i=1 xi = 0,
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
the sum of all vectors associated with u is also e. Therefore, all constraints of type
(22) are satisfied.
Regarding the last three constraints, which are the 22 triangle inequality con-
straints including the origin, one can verify the following:
||e − 0||22 = 1,
$$ √ $$2
$$ $$
$$e − 1 e − k − 1 xi $$ =1−
1
,
$$ k k $$ k
2
$$ √ $$2
$$ $ $
$$0 − 1 e − k − 1 xi $$ =
1
,
$$ k k $$ k
2
$$ √ √ $$2
$$ 1 k−1 1 k−1 $ $ 2
$$ xj $$$$
$$ k e − k
xi −
k
e−
k
=
k
.
2
The last three constraints are all derived from the above calculations. Hence, we
can conclude that the fractional solution defined above is feasible for the semidefinite
relaxation.
The value of this solution is
1 1 k−1
max 1− + (k − 1) · =2 .
1≤i≤k k k k
REFERENCES
[1] N. Alon and B. Klartag, Economical toric spines via Cheeger’s inequality, J. Topol. Anal.,
1 (2009), pp. 101–111.
[2] K. Andreev and H. Racke, Balanced graph partitioning, Theoret. Comput. Systems, 39
(2006), pp. 929–939.
[3] S. Arora, S. Rao, and U. V. Vazirani, Geometry, flows, and graph-partitioning algorithms,
Commun. ACM, 51 (2008), pp. 96–105.
[4] Y. Aumann and Y. Rabani, An O(log k) approximate min-cut max-flow theorem and approx-
imation algorithm, SIAM J. Comput., 27 (1998), pp. 291–301.
[5] N. Bansal, K.-W. Lee, V. Nagarajan, and M. Zafer, Minimum congestion mapping in a
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy
[28] A. Moitra, Approximation algorithms for multicommodity-type problems with guarantees inde-
pendent of the graph size, in in Proceedings of the Symposium on Foundations of Computer
Science, IEEE Computer Society, Los Alamitos, CA, 2009, pp. 3–12.
Downloaded 09/27/24 to 169.235.25.107 . Redistribution subject to SIAM license or copyright; see https://epubs.siam.org/terms-privacy