50
Convergence rate analysis of several splitting schemes ? Damek Davis and Wotao Yin Abstract Operator-splitting schemes are iterative algorithms for solving many types of numerical problems. A lot is known about these methods: they converge, and in many cases we know how quickly they converge. But when they are applied to optimization problems, there is a gap in our understanding: The theoretical speed of operator-splitting schemes is nearly always measured in the ergodic sense, but ergodic operator-splitting schemes are rarely used in practice. In this chapter, we tackle the discrepancy between theory and practice and uncover fundamental lim- its of a class of operator-splitting schemes. Our surprising conclusion is that the relaxed Peaceman-Rachford splitting algorithm, a version of the Alternating Direc- tion Method of Multipliers (ADMM), is nearly as fast as the proximal point al- gorithm in the ergodic sense and nearly as slow as the subgradient method in the nonergodic sense. A large class of operator-splitting schemes extend from the re- laxed Peaceman-Rachford splitting algorithm. Our results show that this class of operator-splitting schemes is also nearly as slow as the subgradient method. The tools we create in this chapter can also be used to prove nonergodic convergence rates of more general splitting schemes, so they are interesting in their own right. Key words: Krasnosel’ski˘ ı-Mann algorithm, Douglas-Rachford Splitting, Peaceman- Rachford Splitting, Alternating Direction Method of Multipliers, nonexpansive op- erator, averaged operator, fixed-point algorithm, little-o convergence Department of Mathematics, University of California, Los Angeles, CA 90025, USA. e-mail: damek / [email protected] ? This work is supported in part by NSF grants DMS-1317602 and ECCS-1462398. 1

Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splittingschemes?

Damek Davis and Wotao Yin

Abstract Operator-splitting schemes are iterative algorithms for solving many typesof numerical problems. A lot is known about these methods: they converge, and inmany cases we know how quickly they converge. But when they are applied tooptimization problems, there is a gap in our understanding: The theoretical speedof operator-splitting schemes is nearly always measured in the ergodic sense, butergodic operator-splitting schemes are rarely used in practice. In this chapter, wetackle the discrepancy between theory and practice and uncover fundamental lim-its of a class of operator-splitting schemes. Our surprising conclusion is that therelaxed Peaceman-Rachford splitting algorithm, a version of the Alternating Direc-tion Method of Multipliers (ADMM), is nearly as fast as the proximal point al-gorithm in the ergodic sense and nearly as slow as the subgradient method in thenonergodic sense. A large class of operator-splitting schemes extend from the re-laxed Peaceman-Rachford splitting algorithm. Our results show that this class ofoperator-splitting schemes is also nearly as slow as the subgradient method. Thetools we create in this chapter can also be used to prove nonergodic convergencerates of more general splitting schemes, so they are interesting in their own right.

Key words: Krasnosel’skiı-Mann algorithm, Douglas-Rachford Splitting, Peaceman-Rachford Splitting, Alternating Direction Method of Multipliers, nonexpansive op-erator, averaged operator, fixed-point algorithm, little-o convergence

Department of Mathematics, University of California, Los Angeles, CA 90025, USA. e-mail:damek / [email protected]

? This work is supported in part by NSF grants DMS-1317602 and ECCS-1462398.

1

Page 2: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

2 D. Davis and W. Yin

1 Introduction

Operator-splitting schemes are iterative algorithms for solving optimization prob-lems (and more generally PDE) [39, 52, 32, 47, 27]. These algorithms are useful forsolving medium to large-scale problems2 in signal processing and machine learning(see [10, 51, 54]). Operator-splitting schemes converge, and in many cases, we knowhow quickly they converge [4, 54, 9, 16, 50, 35, 22, 38, 41, 42, 43, 25, 26]. On thesurface, we seem to have a complete understanding of these algorithms. However,there is a missing piece hidden beneath the contributions in the literature: The theo-retical speed of operator-splitting schemes is nearly always measured in the ergodicsense, but ergodic operator-splitting schemes are rarely used in practice3.

In this chapter, we tackle the discrepancy between theory and practice and un-cover fundamental limits of a class of operator-splitting schemes. Many of themost powerful operator-splitting schemes extend from a single algorithm calledthe relaxed Peaceman-Rachford splitting algorithm (PRS). For most of the chap-ter, we will only study relaxed PRS, but along the way, we develop tools that willhelp us analyze other algorithms.4 These tools are key to analyzing the speed ofoperator-splitting schemes in theory and in practice. We also determine exactlyhow fast the relaxed PRS algorithm can be; these results uncover fundamental lim-its on the speed of the large class operator-splitting schemes that extend relaxedPRS [21, 23, 53, 14, 7, 8, 19, 6, 20, 48, 16].5

Relaxed PRS is an iterative algorithm for solving

minimizex∈H

f (x)+g(x) (1)

where H is a Hilbert space (e.g., Rn for some n ∈ N), and f ,g : H → (−∞,∞] areclosed (i.e., lower semi-continuous), proper, and convex functions. The algorithm iseasy to state: given γ > 0, z0 ∈H , and (λ j) j≥0 ∈ [0,2], define

for all k ∈ N

xk

g = argminx∈H

g(x)+ 1

2γ‖x− zk‖2

;

xkf = argminx∈H

f (x)+ 1

2γ‖x− (2xk

g− zk)‖2

;

zk+1 = zk +λk(xkf − xk

g).

(2)

Although there is no precise mathematical definition of the term operator-splittingscheme, the iteration in (2) is a classic example of such an algorithm because it splitsthe difficult problem (1) into the sequence of simpler xk

g and xkf updates.

2 E.g., with gigabytes to terabytes of data.3 By ergodic, we mean the final approximate solution returned by the algorithm is an average overthe history of all approximate solutions formed throughout the algorithm.4 After the initial release of this chapter, we used these tools to study several more general algo-rithms [25, 26, 27]5 This list is not exhaustive. See the comments after Theorem 8 for more details.

Page 3: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 3

A large class of algorithms [21, 23, 53, 14, 7, 8, 19, 6, 20, 48, 16] extendsrelaxed PRS, and we can demonstrate how the theoretical analysis of such algo-rithm differs from how we use them in practice by measuring the convergencespeed with the objective error6 ( f + g)(xk)− ( f + g)(x∗) evaluated at a sequence(x j) j≥0 ⊆H . Because relaxed PRS is an iterative algorithm, it outputs the finalapproximate solution xk after k ∈ N iterations of the algorithm. In practice, it iscommon to output the nonergodic iterate xk ∈ xk

g,xkf as the final approximate so-

lution. But most theoretical analysis of splitting schemes assumes that the ergodiciterate xk ∈ (1/∑

ki=0 λi)∑

ki=0 xi

g,(1/∑ki=0 λi)∑

ki=0 xi

f is the final approximate so-lution [16, 9, 41, 42, 43]. In many applications, the nonergodic iterates are qualita-tively and quantitatively better approximate solutions to (1) than the ergodic iterates(see [27, Figure 5.3(c)] for an example). The difference in performance is particu-larly large in sparse optimization problems because, unlike the solution to (1), theergodic iterate is often a dense vector. In this chapter, we create some tools to an-alyze the nonergodic convergence rate of the objective error in relaxed PRS. Thesetools can also be used to prove nonergodic convergence rates of more general split-ting schemes [25, 26], so they are interesting in their own right.

Though practical experience suggests that the nonergodic iterates are better ap-proximate solutions than the ergodic iterates, our theoretical analysis actually pre-dicts that they are quantitatively worse. The difference between the theoretical speedof the two iterates is large: we prove that the ergodic iterates have a convergence rateof ( f +g)(xk)−( f +g)(x∗) = O(1/(k+1)), while the best convergence rate that wecan prove for the nonergodic iterates is ( f + g)(xk)− ( f + g)(x∗) = o(1/

√k+1).

Moreover, we provide examples of functions f and g which show that these tworates are tight. This result proves that there are fundamental limits on how quicklythe ergodic and nonergodic iterates of operator-splitting schemes converge, at leastfor the large class of algorithms that extend relaxed PRS. These results comple-ment, but do not follow from, the well-developed lower bounds for (sub)gradientalgorithms [44, 45], which do not address the PRS algorithm.

Our analysis also applies to the Alternating Direction Method of Multipliers(ADMM) algorithm, which is an iterative algorithm for solving:

minimizex∈H1, y∈H2

f (x)+g(y)

subject to Ax+By = b (3)

where H1,H2, and G are Hilbert spaces, b ∈ G , and A : H1→ G and B : H2→ Gare bounded linear operators. We present these results at the end of the chapter.

We prove several other theoretical results that are of independent interest: weprove the fixed-point residual of the Krasnosel’skiı-Mann algorithm (KM) withsummable errors has convergence rate o(1/

√k+1), and we show that the rate is

tight; we prove that relaxed PRS can converge arbitrarily slowly; we prove conver-gence rates and lower bounds for the proximal point algorithm and the forward-

6 x∗ ∈H is a minimizer of (1).

Page 4: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

4 D. Davis and W. Yin

backward splitting algorithm (see Appendix A); and we give several examples ofour results on concrete applications (see Appendix D).

1.1 Notation

The symbols H ,H1,H2,G denote (possibly infinite dimensional) Hilbert spaces.Sequences (λ j) j≥0 ⊂ R+ denote relaxation parameters, and

Λk :=k

∑i=0

λi

denote kth partial sums. The reader may assume that λk ≡ (1/2) and Λk = (k+1)/2in the DRS algorithm or that λk ≡ 1 and Λk = (k+1) in the PRS algorithm. Givena sequence (x j) j≥0 ⊂H , we let xk = (1/Λk)∑

ki=0 λixi denote its kth average with

respect to the sequence (λ j) j≥0.Given a closed, proper, convex function f : H → (−∞,∞], the set ∂ f (x) denotes

its subdifferential at x, and we let

∇ f (x) ∈ ∂ f (x), (4)

denotes an arbitrary subgradient; the actual choice of the subgradient ∇ f (x) willalways be clear from the context. (This notation was used in [5, Eq. (1.10)].)The convex conjugate of a proper, closed, and convex function f is f ∗(y) :=supx∈H 〈y,x〉− f (x) . Let IH denote the identity map. For any x ∈H and scalarγ ∈ R++, we let

proxγ f (x) := argminy∈H

f (y)+

12γ‖y− x‖2 and reflγ f := 2proxγ f − IH ,

be the proximal and reflection operators, and we define the PRS operator:

TPRS := reflγ f reflγg. (5)

1.2 Assumptions

Assumption 1 Every function we consider is closed, proper, and convex.

Unless otherwise stated, a function is not necessarily differentiable.

Assumption 2 (Differentiability) Every differentiable function we consider is Frechetdifferentiable [2, Definition 2.45].

Page 5: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 5

Assumption 3 (Solution existence) Functions f ,g : H → (−∞,∞] satisfy zer(∂ f +∂g) 6= /0.

Note that this last assumption is slightly stronger than the existence of a mini-mizer, because zer(∂ f +∂g) 6= zer(∂ ( f +g)), in general [2, Remark 16.7]. Never-theless, this assumption is standard.

1.3 The Algorithms

In this chapter we study the relaxed PRS algorithm:

Algorithm 1: Relaxed Peaceman-Rachford splitting (relaxed PRS)

input : z0 ∈H , γ > 0, (λ j) j≥0 ⊆ (0,1]for k = 0, 1, . . . do

zk+1 = (1−λk)zk +λkreflγ f reflγg(zk);

The special cases λk ≡ 1/2 and λk ≡ 1 are called the DRS and PRS algorithms,respectively. The relaxed PRS algorithm can be applied to problem (3). To this end,we define the Lagrangian:

Lγ(x,y;w) := f (x)+g(y)−〈w,Ax+By−b〉+ γ

2‖Ax+By−b‖2.

Section 8 presents Algorithm 1 applied to the Lagrange dual of (3), which reducesto the following algorithm:

Algorithm 2: Relaxed alternating direction method of multipliers (relaxedADMM)

input : w−1 ∈H ,x−1 = 0,y−1 = 0,λ−1 = 1/2, γ > 0,(λ j) j≥0 ⊆ (0,1]for k =−1, 0, . . . do

yk+1 = argminy Lγ(xk,y;wk)+ γ(2λk−1)〈By,(Axk +Byk−b)〉;wk+1 = wk− γ(Axk +Byk+1−b)− γ(2λk−1)(Axk +Byk−b);xk+1 = argminx Lγ(x,yk+1;wk+1);

If λk ≡ 1/2, Algorithm 2 recovers the standard ADMM.Each of the above algorithms is a special case of the Krasnosel’skiı-Mann (KM)

iteration [37, 40, 29]. An averaged operator is the average of a nonexpansive oper-ator T : H →H and the identity mapping IH . I.e., for all λ ∈ (0,1), the operator

Tλ := (1−λ )IH +λT (6)

is called λ -averaged and every λ -averaged operator is exactly of the form Tλ forsome nonexpansive map T .

Given a nonexpansive map T , we define the fixed-point iteration of the map Tλ :

Page 6: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

6 D. Davis and W. Yin

Algorithm 3: Krasnosel’skiı-Mann (KM)

input : z0 ∈H ,(λ j) j≥0 ⊆ (0,1]for k = 0, 1, . . . do

zk+1 = Tλk(zk);

1.4 Basic properties of averaged operators

The following properties are included in textbooks such as [2].

Proposition 1. Let H be a Hilbert space; let f ,g : H → (−∞,∞] be closed, proper,and convex functions; and let T : H →H be a nonexpansive operator

1. Let x ∈H . Then x+ = proxγ f (x) if, and only if, (1/γ)(x− x+) ∈ ∂ f (x+).2. The operator reflγ f : H →H is nonexpansive. Therefore,

TPRS := reflγ f reflγg (7)

3. For all λ ∈ (0,1] and (x,y) ∈H ×H , the operator Tλ (see (6)) satisfies

‖Tλ x−Tλ y‖2 ≤ ‖x− y‖2− 1−λ

λ‖(IH −Tλ )x− (IH −Tλ )y‖2.

4. The operator proxγ f : H →H is 12 -averaged.

5. Tλ and T have the same set of fixed points.

2 Summable sequence lemma

Summable sequences of positive numbers always converge to 0. We can even deter-mine how quickly they converge:

Lemma 1 (Summable sequence convergence rates). Let nonnegative scalar se-quences (λ j) j≥0 and (a j) j≥0 satisfy ∑

∞i=0 λiai < ∞. Let Λk := ∑

ki=0 λi for k ≥ 0.

1. Monotonicity: If (a j) j≥0 is monotonically nonincreasing, then

ak ≤1

Λk

(∞

∑i=0

λiai

)and ak = o

(1

Λk−Λdk/2e

). (8)

In particular,

(a) if (λ j) j≥0 is bounded away from 0, then ak = o(1/(k+1));(b) if λk = (k+1)p for some p≥ 0 and all k ≥ 1, then ak = o(1/(k+1)p+1);

Page 7: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 7

(c) as a special case, if λk = (k+1) for all k ≥ 0, then ak = o(1/(k+1)2).

2. Monotonicity up to errors: Let (e j) j≥0 be a sequence of scalars. Suppose thatak+1 ≤ ak + ek for all k ≥ 0 and that ∑

∞i=0 Λiei < ∞. Then

ak ≤1

Λk

(∞

∑i=0

λiai +∞

∑i=0

Λiei

)and ak = o

(1

Λk−Λdk/2e

). (9)

The rates of ak in Parts 1(a), 1(b), and 1(c) continue to hold as long as∑

∞i=0 Λiei <∞. In particular, the rates hold if ek =O(1/(k+1)q) for some q> 2,

q > p+2, and q > 3, in parts 1(a), 1(b), and 1(c), respectively.3. Faster rates: Suppose that (b j) j≥0 and (e j) j≥0 are nonnegative scalar se-

quences, that ∑∞i=0 b j < ∞, and that ∑

∞i=0(i+1)ei < ∞, and that for all k≥ 0 we

have λkak ≤ bk−bk+1 + ek. Then the following sum is finite:

∑i=0

(i+1)λiai ≤∞

∑i=0

bi +∞

∑i=0

(i+1)ei < ∞.

In particular,

(a) if (λ j) j≥0 is bounded away from 0, then ak = o(1/(k+1)2);(b) if λk = (k+1)p for some p≥ 0 and all k ≥ 1, then ak = o(1/(k+1)p+2).

4. No monotonicity: For all k ≥ 0, define the sequence of indices

kbest := argminiai|i = 0, · · · ,k.

Then (a jbest) j≥0 is monotonically nonincreasing and the above bounds continueto hold when ak is replaced with akbest .

Proof. Fix k ≥ 0.Part 1. For all i≤ k, we have ak ≤ ai and λiai ≥ 0 and hence, Λkak ≤∑

ki=0 λiai ≤

∑∞i=0 λiai. This shows the left part of (8). To prove the right part of (8), observe that

(Λk−Λdk/2e)ak =k

∑i=dk/2e+1

λiak ≤k

∑i=dk/2e+1

λiaik→∞→ 0.

Part 1(a). Let λ := inf j≥0 λ j > 0. For every integer k ≥ 2, we have dk/2e ≤(k+ 1)/2. Thus, Λk−Λdk/2e ≥ λ (k−dk/2e) ≥ λ (k− 1)/2 ≥ λ (k+ 1)/6. Hence,ak = o(1/(Λk−Λdk/2e)) = o(1/(k+1)) follows from (8).

Part 1(b). For every integer k ≥ 3, we have dk/2e+1 ≤ (k+3)/2 ≤ 3(k+1)/4and Λk −Λdk/2e = ∑

ki=dk/2e+1 λi = ∑

ki=dk/2e+1(i + 1)p ≥

∫ kdk/2e(t + 1)pdt = (p +

1)−1((k + 1)p+1− (dk/2e+ 1)p+1) ≥ (p+ 1)−1(1− (3/4)p+1)(k + 1)p+1. There-fore, ak = o(1/(Λk−Λdk/2e)) = o(1/(k+1)p+1) follows from (8).

Part 1(c) directly follows from Part 1(b).

Page 8: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

8 D. Davis and W. Yin

Part 2. For every integer 0 ≤ i ≤ k, we have ak ≤ ai +∑k−1j=i e j. Thus, Λkak =

∑ki=0 λiak≤∑

ki=0 λiai+∑

ki=0 λi(∑

k−1j=i e j)=∑

ki=0 λiai+∑

k−1i=0 ei(∑

ij=0 λ j)=∑

ki=0 λiai+

∑k−1i=0 Λiei ≤∑

∞i=0 λiai +∑

∞i=0 Λiei, from which the left part of (9) follows. The proof

for the right part of (9) is similar to Part 1. The condition ek = O(1/(k+1)q) for ap-propriate q is used to ensure that ∑

∞i=0 Λiei < ∞ for each setting of λk in the previous

Parts 1(a), 1(b), and 1(c).Part 3. Note that

λk(k+1)ak ≤ (k+1)bk− (k+1)bk+1 +(k+1)ek

= bk+1 +((k+1)bk− (k+2)bk+1)+(k+1)ek.

Thus, because the upper bound on (k+1)λkak is the sum of a telescoping term anda summable term, we have ∑

∞i=0(i+1)λiai ≤∑

∞i=0 bi +∑

∞i=0(i+1)ei < ∞. Parts 3(a)

and 3(b) are similar to Part 1(b).Part 4 is straightforward, so we omit its proof. ut

Part 1 of Lemma 1 generalizes [36, Theorem 3.3.1] and [28, Lemma 1.2].

3 Iterative fixed-point residual analysis

In this section we determine how quickly the fixed fixed-point residual (FPR) ‖T zk−zk‖2 converges to 0 in Algorithm 3.

Algorithm 3 always converges weakly to a fixed-point of T under mild conditionson (λ j) j≥0 [18, Theorem 3.1] (also see [22, 38]). Because strong convergence ofAlgorithm 3 may fail (when H is infinite dimensional), the quantity ‖zk − z∗‖,where z∗ is a fixed point of T , may not converge to zero. However, the propertylimk→∞ ‖T zk − zk‖ = 0, known as asymptotic regularity [15], holds when a fixedpoint of T exists. Thus, we can always measure the convergence rate of the FPR.

We measure ‖T zk − zk‖2 when we could just as well measure ‖T zk − zk‖. Wechoose to measure the squared norm because it naturally appears in our analysis. Inaddition, it is summable and monotonic, which is analyzable by Lemma 1.

In first-order optimization algorithms, the FPR typically relates to the size of ob-jective gradient. For example, in the unit-step gradient descent algorithm, defined forall k≥ 0 by the recursion zk+1 = zk−∇ f (zk), the FPR is given by ‖∇ f (zk)‖2. In theproximal point algorithm, defined for all k ≥ 0 by the recursion zk+1 = prox f (z

k),the FPR is given by ‖∇ f (zk+1)‖2 where ∇ f (zk+1) := (zk− zk+1) ∈ ∂ f (zk+1) (seePart 1 of Proposition 1). When the objective is the sum of multiple functions, theFPR is a combination of the (sub)gradients of those functions in the objective. In thischapter we will use the subgradient inequality and bounds on the FPR to measurehow quickly f (zk)− f (x∗) converges to 0 for the minimizers x∗ of f .

Page 9: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 9

3.1 o(1/(k+1)) FPR of averaged operators

In the following theorem, the little-o convergence rates in Equation (12) and Part 5are new; the rest of the results can be found in [2, Proof of Proposition 5.14], [22,Proposition 11], and [38].

Theorem 1 (Convergence rate of averaged operators). Let T : H → H be anonexpansive operator, let z∗ be a fixed point of T , let (λ j) j≥0 ⊆ (0,1] be a sequenceof positive numbers, let τk := λk(1−λk), and let z0 ∈H . Suppose that (z j) j≥0⊆His generated by Algorithm 3: for all k ≥ 0, let

zk+1 = Tλk(zk), (10)

where Tλ is defined in (6). Then, the following results hold

1. ‖zk− z∗‖2 is monotonically nonincreasing;2. ‖T zk− zk‖2 is monotonically nonincreasing;3. τk‖T zk− zk‖2 is summable:

∑i=0

τi‖T zi− zi‖2 ≤ ‖z0− z∗‖2; (11)

4. if τk > 0 for all k ≥ 0, then the convergence rates hold:

‖T zk− zk‖2 ≤ ‖z0− z∗‖2

∑ki=0 τi

(12)

and ‖T zk− zk‖2 = o

1

∑ki=d k

2 e+1τi

.

In particular, if (τ j) j≥0 ⊆ (ε,∞) for some ε > 0, then ‖T zk− zk‖2 = o(1/(k+1)).

5. Instead of Iteration (10), for all k ≥ 0, let

zk+1 := Tλk(zk)+λkek (13)

for an error sequence (e j) j≥0 ⊆H that satisfies ∑ki=0 λi‖ei‖< ∞ and ∑

∞i=0(i+

1)λ 2i ‖ei‖2 < ∞. (Note that these bounds hold, for example, when for all k ≥ 0

λk‖ek‖ ≤ ωk for a sequence (ω j) j≥0 that is nonnegative, summable, and mono-tonically nonincreasing.) Then if (τ j) j≥0 ⊆ (ε,∞) for some ε > 0, we continueto have ‖T zk− zk‖2 = o(1/(k+1)).

Proof. As noted before the Theorem, for Parts 1 through 4, we only need toprove the little-o convergence rate. This follows from the monotonicity of (‖T z j−z j‖2) j≥0, Equation (11), and Part 1 of Lemma 1.

Part 5: We first show that the condition involving the sequence (ω j) j≥0 is suf-ficient to guarantee the error bounds. We have ∑

∞i=0 λi‖ei‖ ≤ ∑

∞i=0 ωi < ∞ and

Page 10: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

10 D. Davis and W. Yin

∑∞i=0(i + 1)λ 2

i ‖ei‖2 ≤ ∑∞i=0(i + 1)ω2

i < ∞, where the last inequality is shown asfollows. By Part 1 of Lemma 1, we have ωk = o(1/(k+ 1)). Therefore, there ex-ists a finite K such that (k + 1)ωk < 1 for k > K. Therefore, ∑

∞i=0(i + 1)ω2

i <

∑Ki=0(i+1)ω2

i +∑∞i=K+1 ωi < ∞.

Fix k ≥ 0. For simplicity, introduce pk := T zk− zk, pk+1 := T zk+1− zk+1, andrk := zk+1− zk. Then from (6) and (13), we have pk = 1

λk(rk−λkek). Also introduce

qk := T zk+1−T zk. Then, pk+1− pk = qk− rk.

We will show: (i) ‖pk+1‖2 ≤ ‖pk‖2 +λ 2

kτk‖ek‖2 and (ii) ∑

∞i=0 τi‖pi‖2 < ∞. Then,

applying Part 2 of Lemma 1 (with ak = ‖pk‖2, ek =λ 2

kτk‖ek‖2 , and λk = 1 for which

we have Λk = ∑ki=0 λi = (k+ 1)) and noticing that τk ≥ ε uniformly, we obtain the

rate ‖T zk− zk‖2 = o(1/(k+1)).Part (i): We have

‖pk+1‖2 = ‖pk‖2 +‖pk+1− pk‖2 +2〈pk+1− pk, pk〉

= ‖pk‖2 +‖qk− rk‖2 +2λk〈qk− rk,rk−λkek〉.

By the nonexpansiveness of T , we have ‖qk‖2 ≤ ‖rk‖2 and thus

2〈qk− rk,rk〉= ‖qk‖2−‖rk‖2−‖qk− rk‖2 ≤−‖qk− rk‖2.

Therefore,

‖pk+1‖2 ≤ ‖pk‖2− 1−λk

λk‖qk− rk‖2−2〈qk− rk,ek〉.

= ‖pk‖2− 1−λk

λk

∥∥∥∥qk− rk +λ k

1−λkek∥∥∥∥2

+λ k

1−λk‖ek‖2

≤ ‖pk‖2 +λ 2

kτk‖ek‖2.

Part (ii): First, ‖zk − z∗‖ is uniformly bounded because ‖zk+1 − z∗‖ ≤ (1−λk)‖zk−z∗‖+λk‖T zk−z∗‖+λk‖ek‖≤ ‖zk−z∗‖+λk‖ek‖ by the triangle inequalityand the nonexpansiveness of T . From [2, Corollary 2.14], we have

‖zk+1− z∗‖2 = ‖(1−λk)(zk− z∗)+λk(T zk− z∗+ ek)‖2

= (1−λk)‖zk− z∗‖2 +λk‖T zk− z∗+ ek‖2−λk(1−λk)‖pk + ek‖2

= (1−λk)‖zk− z∗‖2 +λk

(‖T zk− z∗‖2 +2λk〈T zk− z∗,ek〉+λk‖ek‖2

)−λk(1−λk)

(‖pk‖2 +2〈pk,ek〉+‖ek‖2

)≤ ‖zk− z∗‖2− τk‖pk‖2 +λ

2k ‖ek‖2 +2λk‖T zk− z∗‖‖ek‖+2τk‖pk‖‖ek‖︸ ︷︷ ︸

=:ξk

.

Page 11: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 11

Because we have shown (a) ‖T zk − z∗‖ and ‖pk‖ are bounded, (b) ∑∞i=0 τi‖ei‖ ≤

∑∞i=0 λi‖ei‖<∞, and (c) ∑

∞i=0 λ 2

i ‖ei‖2 <∞, we have ∑∞i=0 ξk <∞ and thus ∑

∞i=0 τi‖T zi−

zi‖2 ≤ ‖z0− z∗‖2 +∑∞i=0 ξk < ∞. ut

3.1.1 Notes on Theorem 1

The FPR, ‖T zk− zk‖2, is a normalization of the successive iterate difference zk+1−zk = λk(T zk− zk). Thus, the convergence rates of ‖T zk− zk‖2 naturally imply con-vergence rates of ‖zk+1− zk‖2.

Note that o(1/(k+1)) is the optimal convergence rate for the class of nonexpan-sive operators [12, Remarque 4]. In the special case that T =proxγ f for some closed,proper, and convex function f , the rate of ‖T zk− zk‖2 improves to O(1/(k+ 1)2)[12, Theoreme 9]. See Section 6 for more optimality results. Also, the (error free)little-o convergence rate of the fixed-point residual associated with the resolvent ofa maximal monotone linear operator was shown in [12, Proposition 4]. Finally, wemention the parallel work [24], which proves a similar little-o convergence rate forthe fixed-point residual of the relaxed proximal point algorithm.

In general, it is possible that the nonexpansive operator, T : H →H , is alreadyaveraged, i.e. there exists a nonexpansive operator N : H → H and a positiveconstant α ∈ (0,1] such that T = (1−α)IH +αN, where T and N share the samefixed point set. Thus, we can apply Theorem 1 to N = (1− (1/α))IH +(1/α)T .Furthermore, Nλ = (1−λ/α)IH +(λ/α)T . Thus, when we translate this back toan iteration on T , it enlarges the region of relaxation parameters to λk ∈ (0,1/α)and modifies τk accordingly to τk = λk(1−αλk)/α . The same convergence resultscontinue to hold.

To the best of our knowledge, The little-o rates produced in Theorem 1 havenever been established for the KM iteration. See [22, 38] for similar big-O re-sults. Note that our rate in Part 5 is strictly better than the one shown in [38],and it is given under a much weaker condition on the error because [38] provesan O(1/(k+ 1)) convergence rate only when ∑

∞i=0(i+ 1)‖ei‖ < ∞, which implies

that mini=0,··· ,k‖ei‖= o(1/(k+1)2) by Lemma 1. In contrast, any error sequenceof the form ‖ek‖= O(1/(k+1)α) with α > 1 will satisfy Part 5 of our Theorem 1.Finally, for Banach spaces, we cannot improve big-O rates to little-o [22, Section2.4].

3.2 o(1/(k+1)) FPR of relaxed PRS

In this section, we apply Theorem 1 to the TPRS operator defined in (5). For thespecial case of DRS ((1/2)-averaged PRS), it is straightforward to establish the rateof the FPR

Page 12: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

12 D. Davis and W. Yin

‖(TPRS)1/2zk− zk‖2 = O(

1k+1

)from two existing results: (i) the DRS iteration is a proximal iteration applied toa certain monotone operator [29, Section 4]; (ii) the convergence rate of the FPRfor proximal iterations is O(1/(k+ 1)) [12, Proposition 8] whenever a fixed pointexists. We improve this rate to o(1/(k+1)) for general relaxed PRS operators.

Corollary 1 (Convergence rate of relaxed PRS). Let z∗ be a fixed point of TPRS, let(λ j) j≥0 ⊆ (0,1] be a sequence of positive numbers, let τk := λk(1−λk) for all k≥ 0,and let z0 ∈H . Suppose that (z j) j≥0 ⊆H is generated by Algorithm 1. Then thesequence ‖zk− z∗‖2 is monotonically nonincreasing,and if τ := inf j≥0 τ j > 0, thenthe following convergence rates hold:

‖TPRSzk− zk‖2 ≤ ‖z0− z∗‖2

τ(k+1)and ‖TPRSzk− zk‖2 = o

(1

k+1

). (14)

3.3 O(1/Λ 2k ) ergodic FPR of Fejer monotone sequences

The following definition is often used in the analysis of optimization algorithms[17].

Definition 1. A sequence (z j) j≥0⊆H is Fejer monotone with respect to a nonemptyset C ⊆H if for all z ∈C, we have ‖zk+1− z‖2 ≤ ‖zk− z‖2.

The following fact is trivial, but useful.

Theorem 2. Let (z j) j≥0 be a Fejer monotone sequence with respect to a nonemptyset C ⊆ H . Suppose that for all k ≥ 0, zk+1 − zk = λk(xk − yk) for a sequence((x j,y j)) j≥0 ⊆H 2 and a sequence of positive real numbers (λ j) j≥0. For all k ≥ 0,let zk := (1/Λk)∑

ki=0 λizi, let xk := (1/Λk)∑

ki=0 λixi, and let yk := (1/Λk)∑

ki=0 λiyi.

Then the following bound holds for all z ∈C:

‖xk− yk‖2 ≤ 4‖z0− z‖2

Λ 2k

.

Proof. Λk‖xk− yk‖=∥∥∑

ki=0(zk+1− zk

)∥∥= ∥∥zk+1− z0∥∥≤ 2

∥∥z0− z∥∥ . ut

Part 1 of Theorem 1 shows that any sequence (z j) j≥0 generated by Algorithm 3is Fejer monotone with respect to the set of fixed-points of T . Therefore, Theorem 2applies to iteration Equation (10) with the choice xk = T zk and yk = zk for all k≥ 0.

The interested reader can proceed to Section 6 for several examples that showthe optimality of the rates predicted in this section.

Page 13: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 13

4 Subgradients and fundamental inequalities

This section establishes fundamental inequalities that connect the FPR in Section 3to the objective error of the relaxed PRS algorithm.

In first-order optimization algorithms, we only have access to (sub)gradientsand function values. Consequently, the FPR is usually a linear combination of(sub)gradients. In simple first-order algorithms, like the (sub)gradient method, a(sub)gradient is drawn from a single point at each iteration. In splitting algorithmsfor minimizing sums of convex functions, each function draws at subgradient at adifferent point. There is no natural point at which we can evaluate the entire objec-tive function; this complicates the analysis of the relaxed PRS algorithm.

In the relaxed PRS algorithm, the two operators reflγ f and reflγg are calculatedone after another at different points, neither of which equals zk or zk+1. Conse-quently, the expression zk− zk+1 is more complicated, and the analysis for the stan-dard (sub)gradient iteration does not carry through.

We let x f and xg be the points where subgradients of f and g are drawn,respectively, and use Figure 1 to prove algebraic relations among points z, x f andxg. We use these relations many times. Propositions 2 and 3 use these algebraicrelations to bound the objective error by the FPR. In these bounds, the objectiveerrors of f and g are measured at two points x f and xg such that x f 6= xg. Laterwe will assume that one of the objectives is Lipschitz continuous and evaluate bothfunctions at the same point (See Corollaries 2 and 3).

We conclude this introduction by combining the subgradient notation in Equa-tion (4) and Part 1 of Proposition 1 to arrive at the expressions

proxγ f (x) = x− γ∇ f (proxγ f (x)) (15)

and reflγ f (x) = x−2γ∇ f (proxγ f (x)).

4.1 A subgradient representation of relaxed PRS

In this section we write the relaxed PRS algorithm in terms of subgradients.Lemma 2, Table 1, and Figure 1 summarize a single iteration of relaxed PRS.

Lemma 2. Let z ∈H . Define points xg := proxγg(z) and x f := proxγ f (reflγg(z)).Then the identities hold:

xg = z− γ∇g(xg) and x f = xg− γ∇g(xg)− γ∇ f (x f ).

where ∇g(xg) := (1/γ)(z− xg) ∈ ∂g(xg) and ∇ f (x f ) := (1/γ)(2xg − z− x f ) ∈∂ f (x f ). In addition, each relaxed PRS step has the following representation:

(TPRS)λ (z)− z = 2λ (x f − xg) =−2λγ(∇g(xg)+ ∇ f (x f )). (16)

Page 14: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

14 D. Davis and W. Yin

z TPRS(z)(TPRS)λ (z)

xg x f

−γ∇g(xg)

−γ∇g(xg) −γ∇ f (x f )

−γ∇ f (x f )

2λ (x f − xg)

Fig. 1 A single relaxed PRS iteration, from z to (TPRS)λ (z).

Proof. Figure 1 provides an illustration. Equation (2) follows from reflγg(z) = 2xg−z = xg− γ∇g(xg) and Equation (15). Now, we can compute TPRS(z)− z:

TPRS(z)− z(7)= reflγ f (reflγg(z))− z = 2x f − reflγg(z)− z

= 2x f − (2xg− z)− z = 2(x f − xg).

The subgradient identity in (16) follows from (2). Finally, Equation (16) followsfrom (TPRS)λ (z)− z = (1−λ )z+λTPRS(z)− z = λ (TPRS(z)− z). ut

Point Operator identity Subgradient identity

xsg = proxγg(z

s) = zs− γ∇g(xsg)

xsf = proxγ f (reflγg(zs)) = xs

g− γ(∇g(xsg)+ ∇ f (xs

f ))

(TPRS)λ (zs) = (1−λ )zs +λTPRS(zs) = zs−2γλ (∇g(xsg)+ ∇ f (xs

f ))

Table 1 Overview of the main identities used throughout the chapter. The letter s denotes a super-script (e.g. s = k or s = ∗). See Lemma 2 for a proof.

4.2 Optimality conditions of relaxed PRS

The following lemma characterizes the zeros of ∂ f +∂g in terms of the fixed pointsof the PRS operator.

Lemma 3 (Optimality conditions of TPRS). The following identity holds:

Page 15: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 15

zer(∂ f +∂g) = proxγg(z) | z ∈H ,TPRSz = z. (17)

That is, if z∗ is a fixed point of TPRS, then x∗ = x∗g = x∗f solves Problem 1 and

∇g(x∗) :=1γ(z∗− x∗) ∈ ∂g(x∗). (18)

Proof. See [2, Proposition 25.1] for the proof of Equation (17). Equation (18) fol-lows because x∗ = proxγg(z

∗) if, and only if, z∗− x∗ ∈ γ∂g(x∗). ut

4.3 Fundamental inequalities

We now compute upper and lower bounds of the quantities f (xkf )+g(xk

g)−g(x∗)−f (x∗). Note that xk

f and xkg are not necessarily equal, so this quantity can be negative.

The most important properties of the inequalities we establish below are:

1. The upper fundamental inequality has a telescoping structure in zk and zk+1.2. They can be bounded in terms of ‖zk+1− zk‖2.

Properties 1 and 2 will be used to deduce ergodic and nonergodic rates, respectively.

Proposition 2 (Upper fundamental inequality). Let z ∈H , let z+ := (TPRS)λ (z),and let x f and xg be defined as in Lemma 2. Then for all x ∈ dom( f )∩dom(g)

4γλ ( f (x f )+g(xg)− f (x)−g(x))

≤ ‖z− x‖2−‖z+− x‖2 +

(1− 1

λ

)‖z+− z‖2.

Proof. We use the subgradient inequality and (16) in the following derivation:

4γλ ( f (x f )+g(xg)− f (x)−g(x))

≤ 4λγ

(〈x f − x, ∇ f (x f )〉+ 〈xg− x, ∇g(xg)〉

)= 4λγ

(〈x f − xg, ∇ f (x f )〉+ 〈xg− x, ∇ f (x f )+ ∇g(xg)〉

)= 2

(〈z+− z,γ∇ f (x f )〉+ 〈z+− z,x− xg〉

)(∵ −xg = γ∇g(xg)− z) = 2〈z+− z,x+ γ(∇g(xg)+ ∇ f (x f ))− z〉

= 2〈z+− z,x− 12λ

(z+− z)− z〉

= ‖z− x‖2−‖z+− x‖2 +

(1− 1

λ

)‖z+− z‖2.

Proposition 3 (Lower fundamental inequality). Let z∗ be a fixed point of TPRS andx∗ := proxγg(z

∗). For all x f ∈ dom( f ) and xg ∈ dom(g), the following bound holds:

Page 16: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

16 D. Davis and W. Yin

f (x f )+g(xg)− f (x∗)−g(x∗)≥ 1γ〈xg− x f ,z∗− x∗〉. (19)

Proof. Let ∇g(x∗) = (z∗ − x∗)/γ ∈ ∂g(x∗) and let ∇ f (x∗) = −∇g(x∗) ∈ ∂ f (x∗).Then the result follows by adding f (x f )− f (x∗) ≥ 〈x f − x∗, ∇ f (x∗)〉 and g(xg)−g(x∗)≥ 〈xg− x f , ∇g(x∗)〉+ 〈x f − x∗, ∇g(x∗)〉. ut

5 Objective convergence rates

In this section we will prove ergodic and nonergodic convergence rates of relaxedPRS when f and g are closed, proper, and convex functions that are possibly nons-mooth. For concise notation, we let

h(x,y) := f (x)+g(y)− f (x∗)−g(x∗). (20)

To ease notational memory, the reader may assume that λk = (1/2) for all k ≥ 0,which implies that Λk = (1/2)(k+1), and τk = λk(1−λk) = (1/4) for all k ≥ 0.

Throughout this section the point z∗ denotes an arbitrary fixed point of TPRS, andwe define a minimizer of f +g by the formula (Lemma 3):

x∗ = proxγg(z∗).

The constant (1/γ)‖z∗ − x∗‖ appears in the bounds of this section. This term isindependent of γ: For any fixed point z∗ of TPRS, the point x∗ = proxγg(z

∗) is a

minimizer and z∗−proxγg(z∗) = γ∇g(x∗) ∈ γ∂g(x∗). Conversely, if x∗ ∈ zer(∂ f +

∂g) and ∇g(x∗)∈ (−∂ f (x∗))∩∂g(x∗), then z∗ = x∗+γ∇g(x∗) is a fixed point. Notethat in all of our bounds, we can always replace (1/γ)‖z∗− x∗‖= ‖∇g(x∗)‖ by theinfimum infz∗∈Fix(TPRS)(1/γ)‖z∗−proxγg(z

∗)‖ (the infimum might not be attained).

5.1 Ergodic convergence rates

In this section, we analyze the ergodic convergence of relaxed PRS. The proof fol-lows the telescoping property of the upper and lower fundamental inequalities andan application of Jensen’s inequality.

Theorem 3 (Ergodic convergence of relaxed PRS). For all k ≥ 0, let λk ∈ (0,1].Then we have the following convergence rate

− 2γΛk‖z0− z∗‖‖z∗− x∗‖ ≤ h(xk

f ,xkg)≤

14γΛk

‖z0− x∗‖2.

In addition, the following feasibility bound holds:

Page 17: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 17

‖xkg− xk

f ‖ ≤2

Λk‖z0− z∗‖. (21)

Proof. Equation (21) follows directly from Theorem 2 because (z j) j≥0 is Fejermonotone with respect to Fix(T ) and for all k≥ 0, we have zk+1− zk = λk(xk

f −xkg).

Recall the upper fundamental inequality from Proposition 2 :

4γλkh(xkf ,x

kg)≤ ‖zk− x∗‖2−‖zk+1− x∗‖2 +

(1− 1

λk

)‖zk+1− zk‖2.

Because λk ≤ 1, it follows that (1− (1/λk))≤ 0. Thus, we sum Equation (5.1) fromi = 0 to k, divide by Λk, and apply Jensen’s inequality to get

14γΛk

(‖z0− x∗‖2−‖zk+1− x∗‖2)≥ 1Λk

k

∑i=0

λih(xif ,x

ig)≥ h(xk

f ,xkg).

The lower bound is a consequence of the fundamental lower inequality and (21)

h(xkf ,x

kg)

(19)≥ 1

γ〈xk

g− xkf ,z∗− x∗〉

(21)≥ − 2

γΛk‖z0− z∗‖‖z∗− x∗‖. ut

In general, xkf /∈ dom(g) and xk

g /∈ dom( f ), so we cannot evaluate g at xkf or f at

xkg. But the conclusion of Theorem 3 is improved if f or g is Lipschitz continuous.

The following proposition is a sufficient condition for Lipschitz continuity on a ball:

Proposition 4 (Lipschitz continuity on a ball). Suppose that f : H → (−∞,∞] isproper and convex. Let ρ > 0 and let x0 ∈H . If δ = supx,y∈B(x0,2ρ) | f (x)− f (y)|<∞, then f is (δ/ρ)-Lipschitz on B(x0,ρ).

Proof. See [2, Proposition 8.28]. ut

To use this fact, we need to show that the sequences (x jf ) j≥0, and (x j

g) j≥0 arebounded. Recall that xs

g = proxγg(zs) and xs

f = proxγ f (reflγg(zs)), for s ∈ ∗,k(recall that ∗ is used for quantities associated with a fixed point). Proximal andreflection maps are nonexpansive, so we have the following bound:

max‖xkf − x∗‖,‖xk

g− x∗‖ ≤ ‖zk− z∗‖ ≤ ‖z0− z∗‖.

Thus, (x jf ) j≥0,(x

jg) j≥0 ⊆ B(x∗,‖z0− z∗‖). By the convexity of the closed ball, we

also have (x jf ) j≥0,(x

jg) j≥0 ⊆ B(x∗,‖z0− z∗‖).

Corollary 2 (Ergodic convergence with single Lipschitz function). Let the nota-tion be as in Theorem 3. Suppose that f (respectively g) is L-Lipschitz continuouson B(x∗,‖z0− z∗‖), and let xk = xk

g (respectively xk = xkf ). Then the following con-

vergence rate holds

0≤ h(xk,xk)≤ 14γΛk

‖z0− x∗‖2 +2LΛk‖z0− z∗‖.

Page 18: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

18 D. Davis and W. Yin

Proof. From Equation (21), we have ‖xkg − xk

f ‖ ≤ (2/Λk)‖z0 − z∗‖. In addition,

(x jf ) j≥0,(x

jg) j≥0 ⊆ B(x∗,‖z0− z∗‖). Thus, it follows that

0≤ h(xk,xk)≤ h(xkf ,x

kg)+L‖xk

f − xkg‖

(21)≤ h(xk

f ,xkg)+

2LΛk‖z0− z∗‖.

The upper bound follows from this equation and Theorem 3. ut

5.2 Nonergodic convergence rates

In this section, we prove the nonergodic convergence rate of Algorithm 1 wheneverτ := inf j≥0 τ j > 0. The proof uses Theorem 1 to bound the fundamental inequalitiesin Propositions 2 and 3.

Theorem 4 (Nonergodic convergence of relaxed PRS). For all k ≥ 0, let λk ∈(0,1). Suppose that τ := inf j≥0 λk(1−λk)> 0. Recall that the function h is definedin (20). Then we have the convergence rates:

1. In general, we have the bounds:

−‖z0− z∗‖‖z∗− x∗‖2γ√

τ(k+1)≤ h(xk

f ,xkg)≤

(‖z0− z∗‖+‖z∗− x∗‖)‖z0− z∗‖2γ√

τ(k+1)

and |h(xkf ,x

kg)|= o

(1/√

k+1).

2. If H = R and λk ≡ 1/2, then for all k ≥ 0,

‖z0− z∗‖‖z∗− x∗‖√2γ(k+1)

≤ h(xk+1f ,xk+1

g )≤ (‖z0− z∗‖+‖z∗− x∗‖)‖z0− z∗‖√2γ(k+1)

and |h(xk+1f ,xk+1

g )|= o(1/(k+1)) .

Proof. We prove Part 1 first. For all λ ∈ [0,1], let zλ = (TPRS)λ (zk). Evaluate theupper inequality in Equation (2) at x = x∗ to get

4γλh(xkf ,x

kg)≤ ‖zk− x∗‖2−‖zλ − x∗‖2 +

(1− 1

λ

)‖zλ − zk‖2.

Recall the following identity:

‖zk− x∗‖2−‖zλ − x∗‖2−‖zλ − zk‖2 = 2〈zλ − x∗,zk− zλ 〉.

By the triangle inequality, because ‖zλ −z∗‖≤ ‖zk−z∗‖, and because (‖z j−z∗‖) j≥0is monotonically nonincreasing (Corollary 1), it follows that

‖zλ − x∗‖ ≤ ‖zλ − z∗‖+‖z∗− x∗‖ ≤ ‖z0− z∗‖+‖z∗− x∗‖. (22)

Page 19: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 19

Thus, we have the bound:

h(xkf ,x

kg)≤ inf

λ∈[0,1]

14γλ

(2〈zλ − x∗,zk− zλ 〉+2

(1− 1

)‖zλ − zk‖2

)≤ 1

γ‖z1/2− x∗‖‖zk− z1/2‖

(22)≤ 1

γ

(‖z0− z∗‖+‖z∗− x∗‖

)‖zk− z1/2‖

(14)≤ (‖z0− z∗‖+‖z∗− x∗‖)‖z0− z∗‖

2γ√

τ(k+1).

The lower bound follows from the identity xkg− xk

f = (1/2λk)(zk− zk+1) and thefundamental lower inequality in Equation (19):

h(xkf ,x

kg)≥

12γλk

〈zk− zk+1,z∗− x∗〉 ≥ −‖zk+1− zk‖‖z∗− x∗‖

2γλk

(14)≥ −‖z

0− z∗‖‖z∗− x∗‖2γ√

τ(k+1).

Finally, the o(1/√

k+1) convergence rate follows from Equations (5.2) and (5.2)combined with Corollary 1 because each upper bound is of the form (boundedquantity)×

√FPR, and

√FPR has rate o(1/

√k+1).

Part 2 follows by the same analysis but uses Theorem 12 in Appendix (Page 37)to estimate the FPR convergence rate. ut

Whenever f or g is Lipschitz, we can compute the convergence rate of f + gevaluated at the same point. The next theorem is similar to Corollary 2 in the ergodiccase. The proof is a combination of the nonergodic convergence rate in Theorem 4and the convergence rate of ‖xk

f − xkg‖= (1/λk)‖zk+1− zk‖ shown in Corollary 1.

Corollary 3 (Nonergodic convergence with Lipschitz assumption). Let the nota-tion be as in Theorem 4. Suppose that f (respectively g) is L-Lipschitz continuouson B(x∗,‖z0− z∗‖), and let xk = xk

g (respectively xk = xkf ). Recall that the objective-

error function h is defined in (20). Then we have the convergence rates of the non-negative term:

1. In general, we have the bounds:

0≤ h(xk,xk)≤(‖z0− z∗‖+‖z∗− x∗‖+ γL

)‖z0− z∗‖

2γ√

τ(k+1)

and h(xk,xk) = o(1/√

k+1).

2. If H = R and λk ≡ 1/2, then for all k ≥ 0,

Page 20: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

20 D. Davis and W. Yin

0≤ h(xk+1,xk+1)≤(‖z0− z∗‖+‖z∗− x∗‖+ γL

)‖z0− z∗‖

√2γ(k+1)

and h(xk+1,xk+1) = o(1/(k+1)) .

Proof. We prove Part 1 first. Recall that ‖xkg−xk

f ‖= (1/(2λk))‖zk+1− zk‖. In addi-

tion, (x jf ) j≥0,(x

jg) j≥0 ⊆ B(x∗,‖z0− z∗‖) (See Section 5.1). Thus, it follows that

h(xk,xk)≤ h(xkf ,x

kg)+L‖xk

f − xkg‖= h(xk

f ,xkg)+

L‖zk+1− zk‖2λk

(14)≤ h(xk

f ,xkg)+

L‖z0− z∗‖2√

τ(k+1).

Therefore, the upper bound follows from Theorem 4 and Equation (5.2). In addition,the o(1/

√k+1) bound follows from Theorem 4 combined with Equation (5.2) and

Corollary 1 because each upper bound is of the form (bounded quantity)×√

FPR,and√

FPR has rate o(1/√

k+1).Part 2 follows by the same analysis, but uses Theorem 12 in Appendix (Page 37)

to estimate the FPR convergence rate. ut

6 Optimal FPR rate and arbitrarily slow convergence

In this section, we provide two examples in which the DRS algorithm convergesslowly. Both examples are special cases of the following example, which originallyappeared in [1, Section 7].

Example 1 (DRS applied to two subspaces). Let H = `22(N) = (z j) j≥0 | ∀ j ∈

N,z j ∈ R2,∑∞i=0 ‖z j‖2

R2 < ∞. Let Rθ denote counterclockwise rotation in R2 byθ radians. Let e0 := (1,0) denote the standard unit vector, and let eθ := Rθ e0. Sup-pose that (θ j) j≥0 is a sequence in (0,π/2] and θi → 0 as i→ ∞. We define twosubspaces:

U :=∞⊕

i=0

Re0 and V :=∞⊕

i=0

Reθi

where Re0 = αe0 : α ∈ R and Reθi = αeθi : α ∈ R. Let T := (TPRS)1/2 beapplied to f = ιV and g = ιU . The next identities and properties were shown in [1,Section 7]:

(PU )i =

[1 00 0

]and (PV )i =

[cos2(θi) sin(θi)cos(θi)

sin(θi)cos(θi) sin2(θi)

];

T = c0Rθ0 ⊕ c1Rθ1 ⊕·· · ;

Page 21: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 21

and (z j) j≥0, recursively defined for all k by zk+1 = T zk, converges in norm to z∗ = 0for any initial point z0. ut

6.1 Optimal FPR rates

The following theorem shows that the FPR estimates derived in Corollary 1 aretight.

Theorem 5 (Lower FPR complexity of DRS). There is a Hilbert space H and twoclosed subspaces U and V with zero intersection, U ∩V = 0, such that for everyα > 1/2, there exists z0 ∈H such that if (z j) j≥0 is generated by T = (TPRS)1/2applied to f = ιV and g = ιU , then for all k ≥ 1, we have the bound:

‖T zk− zk‖2 ≥ 1(k+1)2α

.

Proof. We assume the setting of Example 1. For all i ≥ 0 set ci = (i/(i+ 1))1/2.Then for all i≥ 0,

IR2 − cos(θi)Rθi =

[sin2(θi) sin(θi)cos(θi)

−sin(θi)cos(θi) sin2(θi)

]=

[1

i+1

√i

i+1

−√

ii+1

1i+1

]. (23)

Therefore, the point z0 = (√

2αe((1/( j+1)α ,0)) j≥0 ∈H has image

w0 = (I−T )z0 =

(√2αe

(1

( j+1)α+1 ,−√

j( j+1)α+1

))j≥0

.

and for all i≥ 1, we have ‖w0i ‖R2 =

√2αe(i+1)−(1+2α)/2. Thus, for all k ≥ 1,

‖T zk− zk‖2 = ‖T kw0‖2 =∞

∑i=0

c2ki ‖w0

i ‖2R2 ≥

∑i=k

ik

(i+1)k2αe

(i+1)1+2α≥ 1

(k+1)2α. ut

Remark 1. In the proof of Theorem 5, if α = 1/2, then ‖z0‖= ∞.

6.1.1 Notes on Theorem 5

With this new optimality result in hand, we can make the following list of optimalFPR rates, not to be confused with optimal rates in objective error, for a few standardsplitting schemes:

Proximal point algorithm (PPA): For the class of monotone operators, thecounterexample in [12, Remarque 4] shows that there is a maximal monotoneoperator A such that when iteration (10) is applied to the resolvent JγA, the rateo(1/(k + 1)) is tight. In addition, if A = ∂ f for some closed, proper, and convex

Page 22: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

22 D. Davis and W. Yin

function f , then the FPR rate improves to O(1/(k+1)2) [12, Theoreme 9]. We im-prove this result to o(1/(k+1)2) in Theorem 12 in Appendix (Page 37). This resultis new and is optimal by [12, Remarque 6].

Forward backward splitting (FBS): The FBS method reduces to the proximalpoint algorithm when the differentiable (or single valued operator) term is trivial.Thus, for the class of monotone operators, the o(1/(k+1)) FPR rate is optimal by[12, Remarque 4]. We improve this rate to o(1/(k+1)2) in Theorem 12 in Appendix(Page 37). This result is new, and is optimal by [12, Remarque 6].

Douglas-Rachford splitting/ADMM: Theorem 5 shows that the optimal FPRrate is o(1/(k + 1)). Because the DRS iteration is equivalent to a proximal pointalgorithm (PPA) applied to a special monotone operator [29, Section 4], Theo-rem 5 provides an alternative counterexample to [12, Remarque 4]. In particular,Theorem 5 shows that, in general, there is no closed, proper, convex function fsuch that (TPRS)1/2 = proxγ f . In the one dimensional case, we improve the FPR too(1/(k+1)2) in Theorem 13 in Appendix (Page 38).

Miscellaneous methods: By similar arguments we deduce the tight FPR it-eration complexity for the following methods, each of which has rate at leasto(1/(k+1)) by Theorem 1: Standard Gradient descent o(1/(k+1)2): (the rate fol-lows from Theorem 12 in Appendix (Page 37). Optimality follows from the fact thatPPA is equivalent to gradient descent on Moreau envelope [2, Proposition 12.29] and[12, Remarque 4]); Forward-Douglas Rachford splitting [13]: o(1/(k+1)) (choosethe zero cocoercive operator and use Theorem 5); Chambolle and Pock’s primal-dual algorithm [16] o(1/(k+1)): (reduce to DRS (σ = τ = 1) [16, Section 4.2] andapply Theorem 5 using the transformation zk = primalk +dualk [16, Equation (24)]and the lower bound

‖zk+1− zk‖2 ≤ 2‖primalk+1−primalk‖2 +2‖dualk+1−dualk‖2;

Vu/Condat’s primal-dual algorithm [53, 23] o(1/(k+1)): (extends from Chambolleand Pock’s method [16]).

Note that the rate established in Theorem 1 has broad applicability, and this listis hardly extensive. For PPA, FBS, and standard gradient descent, the FPR alwayshas rate that is the square of the objective value convergence rate. We will see thatthe same is true for DRS in Theorem 8.

6.2 Arbitrarily slow convergence

In [1, Section 7], DRS applied to Example 1 is shown to converge in norm, but notlinearly. We improve this result and show that a proper choice of parameters yieldsarbitrarily slow convergence in norm.

The following technical lemma will help us construct a sequence that conver-genes arbitrarily slowly. The proof idea follows from the proof of [30, Theorem

Page 23: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 23

4.2], which shows that the method of alternating projections can converge arbitrar-ily slowly.

Lemma 4. Suppose that h : R+→ (0,1) is a function that is strictly decreasing tozero such that 1/( j+ 1) | j ∈ N\0 ⊆ range(h). Then there exists a monotonicsequence (c j) j≥0 ⊆ (0,1) such that ck→ 1− as k→ ∞ and an increasing sequenceof integers (n j) j≥0 ⊆ N∪0 such that for all k ≥ 0,

ck+1nk

nk +1> h(k+1)e−1.

Proof. Let h2 be the inverse of the strictly increasing function (1/h)− 1, let [x]denote the integer part of x, and for all k ≥ 0, let

ck =h2(k+1)

1+h2(k+1).

Note that because 1/( j+ 1) | j ∈ N\0 ⊆ range(h), ck is well defined. Indeed,k + 1 ∈ dom(h2)∩N if, and only if, there is a y ∈ R+ such that (1/h(y))− 1 =k+1⇐⇒ h(y) = 1/(k+2). It follows that (c j) j≥0 is monotonic and ck→ 1−.

For all x ≥ 0, we have h−12 (x) = 1/h(x)− 1 ≤ [1/h(x)], thus, x ≤ h2([1/h(x)]).

To complete the proof, choose nk ≥ 0 such that nk +1 = [1/h(k+1)] and note that

ck+1nk

nk +1≥ h(k+1)

(k+1

1+(k+1)

)k+1

≥ h(k+1)e−1. ut

Theorem 6 (Arbitrarily slow convergence of DRS). There is a point z0 ∈ `22(N),

such that for every function h : R+→ (0,1) that strictly decreases to zero and sat-isfies 1/( j + 1) | j ∈ N\0 ⊆ range(h), there are two closed subspaces U andV with zero intersection, U ∩V = 0, such that the relaxed PRS sequence (z j) j≥0generated with the functions f = ιV and g = ιU and relaxation parameters λk ≡ 1/2converges in norm but satisfies the bound

‖zk− z∗‖ ≥ e−1h(k).

Proof. We assume the setting of Example 1. Suppose that z0 = (z0j) j≥0, where for

all k≥ 0, z0k ∈R2, and ‖z0

k‖R2 = 1/(k+1). Then it follows that ‖z0‖2 = ∑∞i=0 1/(k+

1)2 < ∞ and so z0 ∈H . Thus, for all k,n≥ 0,

‖T k+1z0‖ ≥ ck+1n ‖z0

n‖R2 =1

n+1ck+1

n .

Therefore, we can achieve arbitrarily slow convergence by picking (c j) j≥0, and asubsequence (n j) j≥0 ⊆ N using Lemma 4. ut

Page 24: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

24 D. Davis and W. Yin

7 Optimal objective rates

In this section we construct four examples that show the nonergodic and ergodicconvergence rates in Corollary 3 and Theorem 3 are optimal up to constant factors.

7.1 Ergodic convergence of minimization problems

In this section, we will construct an example where the ergodic rates of convergencein Section 5.1 are optimal up to constant factors. Our example only converges inthe ergodic sense and diverges otherwise. Throughout this section, we let γ = 1 andλk ≡ 1, we work in the Hilbert space H = R, and we use the following objectivefunctions: for all x ∈ R, let

g(x) = 0 and f (x) = |x|.

Recall that for all x ∈ R

proxg(x) = x and prox f (x) = max(|x|−1,0)sign(x).

The proof of the following lemma is simple so we omit it.

Lemma 5. The unique minimizer of f + g is equal to 0 ∈ R. Furthermore, 0 is theunique fixed point of TPRS.

Because of Lemma 5, we will use the notation:

z∗ = 0 and x∗ = 0.

We are ready to prove our main optimality result.

Proposition 5 (Optimality of ergodic convergence rates). Suppose that z0 = 2−ε

for some ε ∈ (0,1). Then the PRS algorithm applied to f and g with initial pointz0 does not converge. As ε goes to 0, the ergodic objective convergence rate inTheorem 3 is tight, the ergodic objective convergence rate in Corollary 2 is tight upto a factor of 5/2, and the feasibility convergence rate of Theorem 3 is tight up to afactor of 4.

Proof. We compute the sequences (z j) j≥0, (x jg) j≥0, and (x j

f ) j≥0 by induction: Firstx0

g = proxγg(z0) = z0 and x0

f = proxγ f(2x0

g− z0)= max

(|z0|−1,0

)sign(z0) = 1−

ε. Thus, it follows that z1 = z0 +2(x0f −x0

g) = 2−ε +2(1−ε− (2−ε)) = z0 =−ε .Similarly, x1

g = z1 = −ε . Finally, x1f = max(ε−1,0)sign(−ε) = 0 and z2 = z1 +

2(x1f − x1

g) = z1 + 2ε = ε . We only examined the base case, but it is clear that byinduction we have the following identities:

zk = (−1)kε, xk

g = (−1)kε, xk

f = 0, k = 1,2, . . . .

Page 25: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 25

The sequences (z j) j≥0 and (x jg) j≥0 do not converge; they oscillate around 0 ∈

Fix(T ).We will now compute the ergodic iterates:

xkg =

1k+1

k

∑i=0

xig(7.1)=

2−ε

k+1 if k is even;2−2ε

k+1 otherwise.

xkf =

1k+1

k

∑i=0

xif(7.1)=

1− ε

k+1.

Let us use these formulas to compute the objective values:

f (xkf )+g(xk

f )− f (0)−g(0)(7.1)=

1− ε

k+1

f (xkg)+g(xk

g)− f (0)−g(0)(7.1)=

2−ε

k+1 if k is even;2−2ε

k+1 otherwise.

Theorem 3 upper bounds the objective error at xkf by

|z0− x∗|2

4(k+1)=

4−4ε

4(k+1)+

ε2

4(k+1)=

1− ε

k+1+

ε2

4(k+1).

By taking ε to 0, we see that this bound is tight. Because f is 1-Lipschitz continuous,Corollary 2 bounds the objective error at xk

g with

|z0− x∗|2

4(k+1)+

2|z0− z∗|(k+1)

(7.1)=

1− ε

k+1+

ε2

4(k+1)+2

2− ε

k+1=

5−3ε

k+1+

ε2

4(k+1).

As we take ε to 0, we see that this bound it tight up to a factor of 5/2. Finally,consider the feasibility convergence rate:

|xkg− xk

f |(7.1)=

1

k+1 if k is even;1−ε

k+1 otherwise..

Theorem 3 predicts the following upper bound for Equation (7.1):

2|z0− z∗|k+1

= 22− ε

k+1=

4−2ε

k+1.

By taking ε to 0, we see that this bound is tight up to a factor of 4. ut

Page 26: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

26 D. Davis and W. Yin

7.2 Optimal nonergodic objective rates

Our aim in this section is to show that if λk ≡ 1/2, then the non-ergodic convergencerate of o(1/

√k+1) in Corollary 3 is essentially tight. In particular, for every α >

1/2, we provide examples of f and g such that f is 1-Lipschitz and

h(xkg,x

kg) = Ω

(1

(k+1)α

),

where h is defined in (20) and Ω gives a lower bound. Our example uses point-to-setdistance functions.

Proposition 6. Let C be a closed, convex subset of H and let dC : H →H bedefined by dC(·) := miny∈C ‖ ·−y‖. Then dC(x) is 1-Lipschitz and for all x ∈H ,

proxγdC(x) = θPC(x)+(1−θ)x where θ =

γ

dC(x)if γ ≤ dC(x);

1 otherwise.

Proof. Follows from the formula for the subgradient of dC [2, Example 16.49]. ut

Proposition 6 says that proxγdC(x) reduces to a projection map whenever x is

close enough to C. Proposition 7 constructs a family of examples such that if γ ischosen large enough, then DRS does not distinguish between indicator functionsand distance functions.

Proposition 7. Suppose that V and U are linear subspaces of H , U ∩V = 0, andz0 ∈H . If γ ≥ ‖z0‖ and λk = 1/2 for all k ≥ 0, then Algorithm 1 applied to theeither pair of objective functions ( f = ιV ,g = ιU ) and ( f = dV ,g = ιU ) producesthe same sequence (z j) j≥0.

Proof. Let (z j1) j≥0, (x j

g,1) j≥0, and (x jf ,1) j≥0 be sequences generated by the function

pair ( f = ιV ,g = ιU ), and let (z j2) j≥0, (x j

g,2) j≥0, and (x jf ,2) j≥0 be sequences gen-

erated by the function pair ( f = dV ,g = ιU ). Define operators TPRS,1 and TPRS,2likewise. Observe that x∗ := 0 is a minimizer of both functions pairs and z∗ := 0is a fixed point of (TPRS,1)1/2. To show that zk

1 = zk2 for all k ≥ 0, we just need to

show that proxγdV(reflg(zk)) = xk

f ,1 = xkf ,2 = PV (reflg(zk)) for all k ≥ 0. In view of

Proposition 6, the identity will follow if

γ ≥ dV (reflg(zk)) = ‖reflg(zk)−PV (reflg(zk))‖.

However, this is always the case because (reflg(z∗)−PV (reflg(z∗))) = 0 and

‖reflg(zk)−PV (reflg(zk))− (reflg(z∗)−PV (reflg(z∗)))‖2

+‖PV (reflg(zk))−PV (reflg(z∗))‖2

≤ ‖reflg(zk)− reflg(z∗)‖2 ≤ ‖zk− z∗‖2 ≤ ‖z0− z∗‖2 = ‖z0‖2 ≤ γ2

Page 27: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 27

because PV is 12 -averaged. ut

For the rest of this section, we define for all i≥ 0,

θi := cos−1

(√i

i+1

)and ci := cos(θi) =

√i

i+1.

Theorem 7. Assume the notation of Theorem 5. Then for all α > 1/2, there exists apoint z0 ∈H such that if γ ≥ ‖z0‖ and (z j) j≥0 is generated by DRS applied to thefunctions ( f = dV ,g = ιU ), then dV (x∗) = 0 and

dV (xkg) = Ω

(1

(k+1)α

).

Proof. Fix k ≥ 0. Let z0 = ((1/( j+1)α ,0)) j≥0 ∈ H . Now, choose γ ≥ ‖z0‖ =(∑

∞i=0 1/(i+1)2α

)1/2. Define w0 ∈H using Equation (23):

w0 = (I−T )z0 =

(1

( j+1)α

(1

j+1,−√

jj+1

))j≥0

.

Then ‖w0i ‖= 1/(1+ i)(1+2α)/2.

Now we will calculate dV (xkg) = ‖PV xk

g−xkg‖. First, recall that T k =

⊕∞i=0 ck

i Rkθi ,where for all θ ∈ R,

Rθ =

[cos(θ) −sin(θ)sin(θ) cos(θ)

].

Thus,

xkg := PU (zk) =

([1 00 0

]ck

jRkθ

(1

( j+1)α,0))

j≥0

=

([1 00 0

]ck

j1

( j+1)α(cos(kθ j),sin(kθ j))

)j≥0

=

(ck

jcos(kθ j)

( j+1)α(1,0)

)j≥0

.

Furthermore, from the identity

(PV )i =

[cos2(θi) sin(θi)cos(θi)

sin(θi)cos(θi) sin2(θi)

]=

[i

i+1

√i

i+1√i

i+11

i+1

],

we have

PV xkg =

(ck

jcos(kθ j)

( j+1)α

(j

j+1,

√j

j+1

))j≥0

.

Page 28: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

28 D. Davis and W. Yin

Thus, the the difference has the following form:

xkg−PV xk

g =

(ck

jcos(kθ j)

( j+1)α

(1

j+1,−√

jj+1

))j≥0

.

Now we derive the lower bound:

d2V (x

kg) = ‖xk

g−PV xkg‖2 =

∑i=0

c2ki

cos2(kθi)

(i+1)2α+1

=∞

∑i=0

c2ki

cos2(

k cos−1(√

ii+1

))(i+1)2α+1

≥ 1e

∑i=k

cos2(

k cos−1(√

ii+1

))(i+1)2α+1 . (24)

The next three lemmas will focus on estimating the order of the sum in Equa-tion (24). After which, Theorem 7 will follow from Equation (24) and Lemma 8,below. ut

Lemma 6. Let h : R+ → R+ be a continuously differentiable function such thath ∈ L1(R+) and ∑

∞i=1 h(i)< ∞. Then for all positive integers k,∣∣∣∣∣∫

kh(y)dy−

∑i=k

h(i)

∣∣∣∣∣≤ ∞

∑i=k

maxy∈[i,i+1]

∣∣h′(y)∣∣ .Proof. We just apply the Mean Value Theorem:∣∣∣∣∣

∫∞

kh(y)dy−

∑i=k

h(i)

∣∣∣∣∣≤∣∣∣∣∣ ∞

∑i=k

∫ i+1

i(h(y)−h(i))dy

∣∣∣∣∣≤ ∞

∑i=k

∫ i+1

i|h(y)−h(i)|dy

≤∞

∑i=k

maxy∈[i,i+1]

|h′(y)|. ut

The following lemma will quantify the deviation of integral from the sum.

Lemma 7. The following bound holds:∣∣∣∣∣∣∣∞

∑i=k

cos2(

k cos−1(√ i

i+1

))(i+1)2α+1 −

∫∞

k

cos2(

k cos−1(√ y

y+1

))(y+1)2α+1 dy

∣∣∣∣∣∣∣= O

(1

(k+1)2α+ 12

).

(25)

Proof. We will use Lemma 6 with

Page 29: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 29

h(y) =cos2

(k cos−1

(√y

y+1

))(y+1)2α+1 .

to deduce an upper bound on the absolute value. Indeed,

|h′(y)|=

∣∣∣∣∣∣∣k sin

(k cos−1

(√y

y+1

))cos(

k cos−1(√

yy+1

))√

y(y+1)(y+1)2α+1 −cos2

(k cos−1

(√y

y+1

))(y+1)2α+2

∣∣∣∣∣∣∣= O

(k

(y+1)2α+1+3/2 +1

(y+1)2α+2

).

Therefore, we can bound Equation (25) by the following sum:

∑i=k

maxy∈[i,i+1]

|h′(y)|= O(

k(k+1)2α+3/2 +

1(k+1)2α+1

)= O

(1

(k+1)2α+1/2

). ut

In the following lemma, we estimate the order of the oscillatory integral approx-imation to the sum in Equation (24). The proof follows by a change of variables andan integration by parts.

Lemma 8. The following bound holds:

∑i=k

cos2(

k cos−1(√

ii+1

))(i+1)2α+1 dy = Ω

(1

(k+1)2α

). (26)

Proof. Fix k≥ 1. We first perform a change of variables u = cos−1(√

y/(y+1)) onthe integral approximation of the sum:

∫∞

k

cos2(

k cos−1(√

yy+1

))(y+1)2α+1 dy

= 2∫ cos−1

(√k/(k+1)

)0

cos2(ku)cos(u)sin4α−1(u)du. (27)

We will show that the right hand side of Equation (27) is of order Ω(1/(k+1)2α

).

Then Equation (26) will follow by Lemma 7.Let ρ := cos−1(

√k/(k+1)). We have

2∫

ρ

0cos2(ku)cos(u)sin4α−1(u)du =

∫ρ

0(1+ cos(2ku))cos(u)sin4α−1(u)du

= p1 + p2 + p3

where

Page 30: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

30 D. Davis and W. Yin

p1 =∫

ρ

01 · cos(u)sin4α−1(u)du =

14α

sin4α(ρ);

p2 =12k

sin(2kρ)cos(ρ)sin4α−1(ρ);

p3 =−12k

∫ρ

0sin(2ku)d(cos(u)sin4α−1(u));

and we have applied integration by parts for∫ ρ

0 cos(2ku)cos(u)sin4α−1(u)du =p2 + p3.

Because sin(cos−1(x)) =√

1− x2, for all η > 0, we get

sinη(ρ) = sinη cos−1(√

k/(k+1))=

1(k+1)η/2 .

In addition, we have cos(ρ)= coscos−1(√

k/(k+1))=√

k/(k+1) and the trivialbounds |sin(2kρ)| ≤ 1 and |sin(2ku)| ≤ 1.

Therefore, the following bounds hold:

p1 =1

4α(k+1)2αand |p2| ≤

√k/(k+1)

2k(k+1)2α−1/2 = O(

1(k+1)2α+1/2

).

In addition, for p3, we have d(cos(u)sin4α−1(u)) = sin4α−2(u)((4α−1)cos2(u)−sin2(u))du. Furthermore, for u∈ [0,ρ] and α > 1/2, we have sin4α−2(u)∈ [0,1/(k+1)2α−1]and the following lower bound: (4α − 1)cos2(u)− sin2(u) ≥ (4α − 1)cos2(ρ)−sin2(ρ) = (4α−1)(k/(k+1))−1/(k+1)> 0 as long as k≥ 1. Therefore, we havesin4α−2(u)((4α−1)cos2(u)− sin2(u))≥ 0 for all u ∈ [0,ρ], and thus,

|p3| ≤12k

cos(ρ)sin4α−1(ρ) =

√k/(k+1)

2k(k+1)2α−1/2 = O(

1(k+1)2α+1/2

).

Therefore, p1 + p2 + p3 ≥ p1−|p2|− |p3|= Ω((k+1)−2α

). ut

We deduce the following theorem from the sum estimation in Lemma 8:

Theorem 8 (Lower complexity of DRS). There exists closed, proper, and convexfunctions f ,g : H → (−∞,∞] such that f is 1-Lipschitz and for every α > 1/2, thereis a point z0 ∈H and γ ∈R++ such that if (z j) j≥0 is generated by Algorithm 1 withλk = 1/2 for all k ≥ 0, then

h(xkg,x

kg) = Ω

(1

(k+1)α

),

where the objective-error function h is defined in (20).

Proof. Assume the setting of Theorem 7. Then f = dV and g= ιU , and by Lemma 8,we have h(xk

g,xkg) = dV (xk

g) = Ω(1/(k+1)α

). ut

Page 31: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 31

Theorem 8 shows that the DRS algorithm is nearly as slow as the subgradientmethod. We use the word nearly because the subgradient method has complexityO(1/

√k+1), while DRS has complexity o(1/

√k+1). To the best of our knowl-

edge, this is the first lower complexity result for DRS algorithm. Note that Theo-rem 8 implies the same lower bound for the Forward-Douglas-Rachford splittingalgorithm [13] and the many primal-dual operator-splitting schemes [21, 23, 53, 14,7, 8, 19, 6, 20, 16] (this list is not exhaustive) that contain Chambolle and Pock’salgorithm [16] as a special case because the algorithm is known to contain Douglas-Rachford splitting as a special case; see the comments in Section 6.1.1.

8 From relaxed PRS to relaxed ADMM

It is well known that ADMM is equivalent to DRS applied to the Lagrange dual ofProblem (3) [31].7 Thus, if we let d f (w) := f ∗(A∗w) and dg(w) := g∗(B∗w)−〈w,b〉,then relaxed ADMM is equivalent to relaxed PRS applied to the following problem:

minimizew∈G

d f (w)+dg(w).

We make two assumptions regarding d f and dg:

Assumption 4 (Solution existence) Functions f ,g : H → (−∞,∞] satisfy

zer(∂d f +∂dg) 6= /0.

This is a restatement of Assumption 3, which is used in our analysis of the primalcase.

Assumption 5 The following differentiation rule holds:

∂d f (x) = A∗ (∂ f ∗)A and ∂dg(x) = B∗ (∂g∗)B−b.

See [2, Theorem 16.37] for conditions that imply this identity, of which the weakestare 0 ∈ sri(range(A∗)− dom( f ∗)) and 0 ∈ sri(range(B∗)− dom(g∗)), where sri isthe strong relative interior of a convex set. This assumption may seem strong, but itis standard in the analysis of ADMM. because it implies the dual proximal operatoridentities in (33) on Page 40 below.

7 A recent result reported in Chapter 5 [55] of this volume shows the direct (non-duality) equiva-lence between ADMM and DRS when they are both applied to Problem (3).

Page 32: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

32 D. Davis and W. Yin

8.1 Primal objective convergence rates in ADMM

With a little effort (see Appendix C.3), we can show the following convergence ratesfor ADMM.

Theorem 9 (Ergodic primal convergence of ADMM). Define the ergodic primaliterates by the formulas: xk = (1/Λk)∑

ki=0 λixi and yk = (1/Λk)∑

ki=0 λiyi. Then

−2‖w∗‖‖z0− z∗‖γΛk

≤ h(xk,yk)≤ ‖z0− (z∗−w∗)‖2

4γΛk,

where h is defined in (20).

The ergodic rate presented here is stronger and easier to interpret than the onein [34] for the ADMM algorithm (λk ≡ 1/2). Indeed, the rate presented in [34,Theorem 4.1] shows the following bound: for all k ≥ 1 and for any bounded setD ⊆ dom( f )×dom(g)×G , we have the following variational inequality bound

sup(x,y,w)∈D

(h(xk−1,yk)+ 〈wk

dg,Ax+By−b〉−〈Axk−1 +Byk−b,w〉

)≤

sup(x,y,w)∈D ‖(x,y,w)− (x0,y0,w0dg)‖2

2(k+1).

If (x∗,y∗,w∗) ∈ D , then the supremum is positive and bounds the deviation of theprimal objective from the lower fundamental inequality.

Theorem 10 (Nonergodic primal convergence of ADMM). For all k≥ 0, let τk =λk(1− λk). In addition, suppose that τ = inf j≥0 τ j > 0. Recall that the objective-error function h is defined in (20). Then

1. In general, we have the bounds:

−‖z0− z∗‖‖w∗‖2√

τ(k+1)≤ h(xk,yk)≤ ‖z

0− z∗‖(‖z0− z∗‖+‖w∗‖)2γ√

τ(k+1)

and |h(xk,yk)|= o(1/√

k+1).2. If G = R and λk ≡ 1/2, then for all k ≥ 0,

−‖z0− z∗‖‖w∗‖√2(k+1)

≤ h(xk+1,yk+1)≤ ‖z0− z∗‖(‖z0− z∗‖+‖w∗‖)√

2γ(k+1)

and |h(xk+1,yk+1)|= o(1/(k+1)).

The rates presented in Theorem 10 are new and, to the best of our knowledge, theyare the first nonergodic rate results for ADMM primal objective error.

Page 33: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 33

9 Conclusion

In this chapter, we provided a comprehensive convergence rate analysis of the FPRand objective error of several splitting algorithms under general convexity assump-tions. We showed that the convergence rates are essentially optimal in all cases. Allresults follow from some combination of a lemma that deduces convergence ratesof summable monotonic sequences (Lemma 1), a simple diagram (Figure 1), andfundamental inequalities (Propositions 2 and 3) that relate the FPR to the objectiveerror of the relaxed PRS algorithm. The most important open question is whetherand how the rates we derived will improve when we enforce stronger assumptions,such as Lipschitz differentiability and/or strong convexity, on f and g. This will bethe subject of future work.

References

1. Bauschke, H.H., Bello Cruz, J.Y., Nghia, T.T.A., Phan, H.M., Wang, X.: The rate of linearconvergence of the Douglas-Rachford algorithm for subspaces is the cosine of the Friedrichsangle. Journal of Approximation Theory 185(0), 63–79 (2014)

2. Bauschke, H.H., Combettes, P.L.: Convex Analysis and Monotone Operator Theory in HilbertSpaces. Springer (2011)

3. Bauschke, H.H., Deutsch, F., Hundal, H.: Characterizing arbitrarily slow convergence in themethod of alternating projections. International Transactions in Operational Research 16(4),413–425 (2009)

4. Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverseproblems. SIAM Journal on Imaging Sciences 2(1), 183–202 (2009)

5. Bertsekas, D.P.: Incremental gradient, subgradient, and proximal methods for convex opti-mization: A survey. Optimization for Machine Learning pp. 85–120 (2011)

6. Bot, R.I., Csetnek, E.R.: On the convergence rate of a forward-backward type primal-dualsplitting algorithm for convex optimization problems. Optimization 64(1), 5–23 (2015)

7. Bot, R.I., Hendrich, C.: A Douglas–Rachford type primal-dual method for solving inclusionswith mixtures of composite and parallel-sum type monotone operators. SIAM Journal onOptimization 23(4), 2541–2565 (2013)

8. Bot, R.I., Hendrich, C.: Solving monotone inclusions involving parallel sums of linearly com-posed maximally monotone operators. arXiv:1306.3191 [math] (2013)

9. Bot, R.I., Hendrich, C.: Convergence analysis for a primal-dual monotone+ skew splittingalgorithm with applications to total variation minimization. Journal of Mathematical Imagingand Vision pp. 1–18 (2014)

10. Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statisti-cal learning via the alternating direction method of multipliers. Foundations and Trends inMachine Learning 3(1), 1–122 (2011)

11. Bredies, K.: A forward–backward splitting algorithm for the minimization of non-smooth con-vex functionals in Banach space. Inverse Problems 25(1), 015,005 (2009)

12. Brezis, H., Lions, P.L.: Produits infinis de resolvantes. Israel Journal of Mathematics 29(4),329–345 (1978)

13. Briceno-Arias, L.M.: Forward-Douglas-Rachford splitting and forward-partial inverse methodfor solving monotone inclusions. Optimization 64(5), 1239–1261 (2015)

14. Briceno-Arias, L.M., Combettes, P.L.: A monotone+ skew splitting model for compositemonotone inclusions in duality. SIAM Journal on Optimization 21(4), 1230–1250 (2011)

Page 34: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

34 D. Davis and W. Yin

15. Browder, F.E., Petryshyn, W.V.: The solution by iteration of nonlinear functional equations inBanach spaces. Bulletin of the American Mathematical Society 72(3), 571–575 (1966)

16. Chambolle, A., Pock, T.: A first-order primal-dual algorithm for convex problems with appli-cations to imaging. Journal of Mathematical Imaging and Vision 40(1), 120–145 (2011)

17. Combettes, P.L.: Quasi-Fejerian analysis of some optimization algorithms. Studies in Com-putational Mathematics 8, 115–152 (2001)

18. Combettes, P.L.: Solving monotone inclusions via compositions of nonexpansive averagedoperators. Optimization 53(5-6), 475–504 (2004)

19. Combettes, P.L.: Systems of structured monotone inclusions: duality, algorithms, and applica-tions. SIAM Journal on Optimization 23(4), 2420–2447 (2013)

20. Combettes, P.L., Condat, L., Pesquet, J.C., Vu, B.C.: A forward-backward view of someprimal-dual optimization methods in image recovery. In: 2014 IEEE International Confer-ence on Image Processing (ICIP), pp. 4141–4145 (2014)

21. Combettes, P.L., Pesquet, J.C.: Primal-dual splitting algorithm for solving inclusions withmixtures of composite, Lipschitzian, and parallel-sum type monotone operators. Set-Valuedand variational analysis 20(2), 307–330 (2012)

22. Cominetti, R., Soto, J.A., Vaisman, J.: On the rate of convergence of Krasnosel’skiı-Manniterations and their connection with sums of Bernoullis. Israel Journal of Mathematics pp.1–16 (2014)

23. Condat, L.: A primal–dual splitting method for convex optimization involving Lipschitzian,proximable and linear composite terms. Journal of Optimization Theory and Applications158(2), 460–479 (2013)

24. Corman, E., Yuan, X.: A generalized proximal point algorithm and its convergence rate. SIAMJournal on Optimization 24(4), 1614–1638 (2014)

25. Davis, D.: Convergence Rate Analysis of Primal-Dual Splitting Schemes. SIAM Journal onOptimization 25(3), 1912–1943 (2015)

26. Davis, D.: Convergence Rate Analysis of the Forward-Douglas-Rachford Splitting Scheme.SIAM Journal on Optimization 25(3), 1760–1786 (2015)

27. Davis, D., Yin, W.: A three-operator splitting scheme and its optimization applications. arXivpreprint arXiv:1504.01032v1 (2015)

28. Deng, W., Lai, M.J., Yin, W.: On the o(1/k) convergence and parallelization of the alternatingdirection method of multipliers. arXiv preprint arXiv:1312.3040 (2013)

29. Eckstein, J., Bertsekas, D.P.: On the Douglas-Rachford splitting method and the proximalpoint algorithm for maximal monotone operators. Mathematical Programming 55(1-3), 293–318 (1992)

30. Franchetti, C., Light, W.: On the von Neumann alternating algorithm in Hilbert space. Journalof mathematical analysis and applications 114(2), 305–314 (1986)

31. Gabay, D.: Application of the methods of multipliers to variational inequalities. In: M. Fortin,R. Glowinski (eds.) Augmented Lagrangians: Application to the Numerical Solution ofBoundary Value Problems, pp. 299–331. North-Holland, Amsterdam (1983)

32. Glowinski, R., Marrocco, A.: Sur l’approximation, par elements finis d’ordre un, et laresolution, par penalisation-dualite d’une classe de problemes de Dirichlet nonlineaires. Rev.Francaise dAut. Inf. Rech. Oper R-2, 41–76 (1975)

33. Guler, O.: On the Convergence of the Proximal Point Algorithm for Convex Minimization.SIAM Journal on Control and Optimization 29(2), 403–419 (1991)

34. He, B., Yuan, X.: On the O(1/n) convergence rate of the Douglas-Rachford alternating direc-tion method. SIAM Journal on Numerical Analysis 50(2), 700–709 (2012)

35. He, B., Yuan, X.: On non-ergodic convergence rate of Douglas–Rachford alternating directionmethod of multipliers. Numerische Mathematik 130(3), 567–577 (2015)

36. Knopp, K.: Infinite Sequences and Series. Courier Dover Publications (1956)37. Krasnosel’skii, M.A.: Two remarks on the method of successive approximations. Uspekhi

Matematicheskikh Nauk 10(1), 123–127 (1955)38. Liang, J., Fadili, J., Peyre, G.: Convergence rates with inexact nonexpansive operators. arXiv

preprint arXiv:1404.4837 (2014)

Page 35: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 35

39. Lions, P.L., Mercier, B.: Splitting algorithms for the sum of two nonlinear operators. SIAMJournal on Numerical Analysis 16(6), 964–979 (1979)

40. Mann, W.R.: Mean value methods in iteration. Proceedings of the American MathematicalSociety 4(3), 506–510 (1953)

41. Monteiro, R.D., Svaiter, B.F.: Iteration-complexity of block-decomposition algorithms and thealternating direction method of multipliers. SIAM Journal on Optimization 23(1), 475–507(2013)

42. Monteiro, R.D.C., Svaiter, B.F.: On the complexity of the hybrid proximal extragradientmethod for the iterates and the ergodic mean. SIAM Journal on Optimization 20(6), 2755–2787 (2010)

43. Monteiro, R.D.C., Svaiter, B.F.: Complexity of Variants of Tseng’s Modified F-B Splitting andKorpelevich’s Methods for Hemivariational Inequalities with Applications to Saddle-point andConvex Optimization Problems. SIAM Journal on Optimization 21(4), 1688–1720 (2011)

44. Nemirovsky, A.S., Yudin, D.B.: Problem Complexity and Method Efficiency in Optimization.Wiley, New York (1983)

45. Nesterov, Y.: Introductory Lectures on Convex Optimization, vol. 87. Springer US, Boston,MA (2004)

46. Ogura, N., Yamada, I.: Non-strictly convex minimization over the fixed point set of an asymp-totically shrinking nonexpansive mapping. Numerical Functional Analysis and Optimization23(1-2), 113–137 (2002)

47. Passty, G.B.: Ergodic convergence to a zero of the sum of monotone operators in Hilbert space.Journal of Mathematical Analysis and Applications 72(2), 383 – 390 (1979)

48. Pock, T., Cremers, D., Bischof, H., Chambolle, A.: An algorithm for minimizing theMumford-Shah functional. In: Computer Vision, 2009 IEEE 12th International Conferenceon, pp. 1133–1140. IEEE (2009)

49. Schizas, I.D., Ribeiro, A., Giannakis, G.B.: Consensus in ad hoc WSNs with noisy links. PartI: Distributed estimation of deterministic signals. Signal Processing, IEEE Transactions on56(1), 350–364 (2008)

50. Shefi, R., Teboulle, M.: Rate of convergence analysis of decomposition methods based onthe proximal method of multipliers for convex minimization. SIAM Journal on Optimization24(1), 269–297 (2014)

51. Shi, W., Ling, Q., Yuan, K., Wu, G., Yin, W.: On the linear convergence of the ADMM indecentralized consensus optimization. IEEE Transactions on Signal Processing 62(7), 1750–1761 (2014)

52. Tseng, P.: A modified forward-backward splitting method for maximal monotone mappings.SIAM Journal on Control and Optimization 38(2), 431–446 (2000)

53. Vu, B.C.: A splitting algorithm for dual monotone inclusions involving cocoercive operators.Advances in Computational Mathematics 38(3), 667–681 (2013)

54. Wei, E., Ozdaglar, A.: Distributed alternating direction method of multipliers. In: Decisionand Control (CDC), 2012 IEEE 51st Annual Conference on, pp. 5445–5450. IEEE (2012)

55. Yan, M., Yin, W.: Self equivalence of the alternating direction method of multipliers. In:R. Glowinski, S. Osher, W. Yin (eds.) Splitting Methods in Communication and Imaging,Science and Engineering. Springer (2016)

A Further applications of the results of Section 3

A.1 o(1/(k+1)2) FPR of FBS and PPA

In problem (1), let g be a C1 function with Lipschitz derivative. The forward-backward splitting (FBS) algorithm is the iteration:

Page 36: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

36 D. Davis and W. Yin

zk+1 = proxγ f (zk− γ∇g(zk)), k = 0,1, . . . . (28)

The FBS algorithm generalizes and has the following subgradient representation:

zk+1 = zk− γ∇ f (zk+1)− γ∇g(zk) (29)

where ∇ f (zk+1) := (1/γ)(zk− zk+1− γ∇g(zk)) ∈ ∂ f (zk+1), and zk+1 and ∇ f (zk+1)are unique given zk and γ > 0.

In this section, we analyze the convergence rate of the FBS algorithm given inEquations (28) and (29). If g = 0, FBS reduces to the proximal point algorithm(PPA) and β = ∞. If f = 0, FBS reduces to gradient descent. The FBS algorithmcan be written in the following operator form:

TFBS := proxγ f (I− γ∇g).

Because proxγ f is (1/2)-averaged and I− γ∇g is γ/(2β )-averaged [46, Theorem3(b)], it follows that TFBS is αFBS-averaged for

αFBS :=2β

4β − γ∈ (1/2,1)

whenever γ < 2β [2, Proposition 4.32]. Thus, we have TFBS = (1−αFBS)I+αFBSTfor a certain nonexpansive operator T , and TFBS(zk)− zk = αFBS(T zk− zk). In par-ticular, for all γ < 2β the following sum is finite:

∑i=0‖TFBS(zk)− zk‖2

(11)≤ αFBS‖z0− z∗‖2

(1−αFBS).

To analyze the FBS algorithm we need to derive a joint subgradient inequalityfor f + g. First, we recall the following sufficient descent property for Lipschitzdifferentiable functions.

Theorem 11 (Descent theorem [2, Theorem 18.15(iii)]). If g is differentiable and∇g is (1/β )-Lipschitz, then for all x,y ∈H we have the upper bound

g(x)≤ g(y)+ 〈x− y,∇g(y)〉+ 12β‖x− y‖2.

Corollary 4 (Joint descent theorem). If g is differentiable and ∇g is (1/β )-Lipschitz, then for all points x,y ∈ dom( f ) and z ∈H , and subgradients ∇ f (x) ∈∂ f (x), we have

f (x)+g(x)≤ f (y)+g(y)+ 〈x− y,∇g(z)+ ∇ f (x)〉+ 12β‖z− x‖2. (30)

Proof. Inequality (30) follows from adding the upper bound

Page 37: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 37

g(x)−g(y)≤ g(z)−g(y)+ 〈x− z,∇g(z)〉+ 12β‖z− x‖2 ≤ 〈x− y,∇g(z)〉+ 1

2β‖z− x‖2

with the subgradient inequality: f (x)≤ f (y)+ 〈x− y, ∇ f (x)〉. ut

We now improve the O(1/(k+ 1)2) FPR rate for PPA in [12, Theoreme 9] byshowing that the FPR rate of FBS is actually o(1/(k+1)2).

Theorem 12 (Objective and FPR convergence of FBS). Let z0 ∈ dom( f )∩dom(g)and let x∗ be a minimizer of f +g. Suppose that (z j) j≥0 is generated by FBS (itera-tion (28)) where ∇g is (1/β )-Lipschitz and γ < 2β . Then for all k ≥ 0,

h(zk+1,zk+1)≤ ‖z0− x∗‖2

k+1×

12γ

if γ ≤ β ;(12γ+(

12β− 1

)αFBS

(1−αFBS)

)otherwise,

andh(zk+1,zk+1) = o(1/(k+1)),

where the objective-error function h is defined in (20). In addition, for all k ≥ 0, wehave ‖TFBSzk+1− zk+1‖2 = o(1/(k+1)2) and

‖TFBSzk+1− zk+1‖2 ≤ ‖z0− x∗‖2( 1γ− 1

)(k+1)2

×

12γ

if γ ≤ β ;(12γ+( 1

2β− 1

)αFBS

(1−αFBS)

)otherwise.

Proof. Recall that zk− zk+1 = γ∇ f (zk+1)+ γ∇g(zk) ∈ γ∂ f (zk+1)+ γ∇g(zk) for allk ≥ 0. Thus, the joint descent theorem shows that for all x ∈ dom( f ), we have

f (zk+1)+g(zk+1)− f (x)−g(x)(30)≤ 1

γ〈zk+1− x,zk− zk+1〉+ 1

2β‖zk− zk+1‖2

=12γ

(‖zk− x‖2−‖zk+1− x‖2

)+

(1

2β− 1

)‖zk+1− zk‖2.(31)

If we set x = x∗ in Equation (31), we see that (h(z j+1,z j+1)) j≥0 is positive,summable, and

∑i=0

h(z j+1,z j+1)≤

12γ‖z0− x∗‖2 if γ ≤ β ;(12γ+(

12β− 1

)αFBS

(1−αFBS)

)‖z0− x∗‖2 otherwise.

(32)

In addition, if we set x = zk in Equation (31), then we see that (h(z j+1,z j+1)) j≥0 isdecreasing: (

1γ− 1

)‖zk+1− zk‖2 ≤ h(zk,zk)−h(zk+1,zk+1).

Therefore, the rates for h(zk+1,zk+1) follow by Lemma 1 Part (a), with ak =h(zk+1,zk+1) and λk ≡ 1.

Page 38: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

38 D. Davis and W. Yin

Now we prove the rates for ‖TFBSzk+1− zk+1‖2. We apply Part 3 of Lemma 1with ak = (1/γ−1/(2β ))‖zk+2− zk+1‖2, λk ≡ 1, ek = 0, and bk = h(zk+1)−h(x∗)for all k≥ 0, to show that ∑

∞i=0(i+1)ai is less than the sum in Equation (32). Part 2

of Theorem 1 shows that (a j) j≥0 is monotonically nonincreasing. Therefore, theconvergence rate of (a j) j≥0 follows from Part (b) of Lemma 1. ut

When f = 0, the objective error upper bound in Theorem 12 is strictly betterthan the bound provided in [45, Corollary 2.1.2]. In FBS, the objective error rate isthe same as the one derived in [4, Theorem 3.1], when γ ∈ (0,β ], and is the sameas the one given in [11] in the case that γ ∈ (0,2β ). The little-o FPR rate is newin all cases except for the special case of PPA (g ≡ 0) under the condition that thesequence (z j) j≥0 strongly converges to a minimizer [33].

A.2 o(1/(k+1)2) FPR of one dimensional DRS

Whenever the operator (TPRS)1/2 is applied in R, the convergence rate of the FPRimproves to o(1/(k+1)2).

Theorem 13. Suppose that H = R, and suppose that (z j) j≥0 is generated by theDRS algorithm, i.e. Algorithm 1 with λk ≡ 1/2. Then for all k ≥ 0,

|(TPRS)1/2zk+1− zk+1|2 = |z0− z∗|2

2(k+1)2 and |(TPRS)1/2zk+1− zk+1|2 = o(

1(k+1)2

).

Proof. Note that (TPRS)1/2 is (1/2)-averaged, and, hence, it is the resolvent of somemaximal monotone operator on R [2, Corollary 23.8]. Furthermore, every maxi-mal monotone operator on R is the subdifferential operator of a closed, proper, andconvex function [2, Corollary 22.19]. Therefore, DRS is equivalent to the proximalpoint algorithm applied to a certain convex function on R. Thus, the result followsby Theorem 12 applied to this function. ut

B Further lower complexity results

B.1 Ergodic convergence of feasibility problems

Proposition 8. The ergodic feasibility convergence rate in Equation (21) is optimalup to a factor of two.

Proof. Algorithm 1 with λk = 1 for all k ≥ 0 (i.e. PRS) is applied to the functionsf = ι(x1,x2)∈R2|x1=0 and g = ι(x1,x2)∈R2|x2=0 with the initial iterate z0 = (1,1) ∈R2. Because TPRS = −IH , it is easy to see that the only fixed point of TPRS is z∗ =(0,0). In addition, the following identities are satisfied:

Page 39: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 39

xkg =

(1,0) even k;(−1,0) odd k.

zk =

(1,1) even k;(−1,−1) odd k.

xkf =

(0,−1) even k;(0,1) odd k.

Thus, the PRS algorithm oscillates around the solution x∗ = (0,0). However, notethat the averaged iterates satisfy:

xkg =

( 1

k+1 ,0) even k;(0,0) odd k.

and xkf =

(0, −1

k+1 ) even k;(0,0) odd k.

It follows that ‖xkg− xk

f ‖= (1/(k+1))‖(1,−1)‖= (1/(k+1))‖z0− z∗‖, ∀k ≥ 0.ut

B.2 Optimal objective and FPR rates with Lipschitz derivative

The following examples show that the objective and FPR rates derived in Theo-rem 12 are essentially optimal. The setup of the following counterexample alreadyappeared in [12, Remarque 6] but the objective function lower bounds were notshown.

Theorem 14 (Lower complexity of PPA). There exists a Hilbert space H , and aclosed, proper, and convex function f such that for all α > 1/2, there exists z0 ∈Hsuch that if (z j) j≥0 is generated by PPA, then

‖proxγ f (zk)− zk‖2 ≥ γ2

(1+2α)e2γ(k+ γ)1+2α,

f (zk+1)− f (x∗)≥ 14αe2γ(k+1+ γ)2α

.

Proof. Let H = `2(R), and define a linear map A : H →H :

A(z1,z2, · · · ,zn, · · ·) =(

z1,z2

2, · · · , zn

n, · · ·).

For all z ∈H , define f (x) = (1/2)〈Az,z〉. Thus, we have the proximal identity forf and

proxγ f (z) = (I + γA)−1(z) =(

jj+ γ

z j

)j≥1

and (I−proxγ f )(z) =(

γ

j+ γz j

)j≥1

.

Now let z0 = (1/( j+ γ)α) j≥1 ∈H , and set T = proxγ f . Then we get the fol-lowing FPR lower bound:

Page 40: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

40 D. Davis and W. Yin

‖zk+1− zk‖2 = ‖T k(T − I)z0‖2 =∞

∑i=1

(i

i+ γ

)2kγ2

(i+ γ)2+2α

≥∞

∑i=k

(i

i+ γ

)2kγ2

(i+ γ)2+2α

≥ γ2

(1+2α)e2γ(k+ γ)1+2α.

Furthermore, the objective lower bound holds

f (zk+1)− f (x∗) =12〈Azk+1,zk+1〉= 1

2

∑i=1

1i

(i

i+ γ

)2(k+1) 1(i+ γ)2α

≥ 12

∑i=k+1

(i

i+ γ

)2(k+1) 1(i+ γ)1+2α

≥ 14αe2γ(k+1+ γ)2α

. ut

C ADMM convergence rate proofs

Given an initial vector z0 ∈ G , Lemma 2 shows that at each iteration relaxed PRSperforms the following computations:

wkdg= proxγdg(z

k);

wkd f

= proxγd f(2wk

dg− zk);

zk+1 = zk +2λk(wkd f−wk

dg).

In order to apply the relaxed PRS algorithm, we need to compute the proximaloperators of the dual functions d f and dg.

Lemma 9 (Proximity operators on the dual). Let w,v ∈ G . Then the update for-mulas w+ = proxγd f

(w) and v+ = proxγdg(v) are equivalent to the following com-putations

x+ = argminx∈H1f (x)−〈w,Ax〉+ γ

2‖Ax‖2;w+ = w− γAx+;y+ = argminy∈H2

g(y)−〈v,By−b〉+ γ

2‖By−b‖2;v+ = v− γ(By+−b);

(33)

respectively. In addition, the subgradient inclusions hold: A∗w+ ∈ ∂ f (x+) andB∗v+ ∈ ∂g(y+). Finally, w+ and v+ are independent of the choice of x+ and y+,respectively, even if they are not unique solutions to the minimization subproblems.

Page 41: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 41

We can use Lemma 9 to derive the relaxed form of ADMM in Algorithm 2.Note that this form of ADMM eliminates the “hidden variable” sequence (z j) j≥0 inEquation (C). This following derivation is not new, but is included for the sake ofcompleteness. See [31] for the original derivation.

Proposition 9 (Relaxed ADMM). Let z0 ∈ G , and let (z j) j≥0 be generated by therelaxed PRS algorithm applied to the dual formulation in Equation (8). Chooseinitial points w−1

dg= z0,x−1 = 0 and y−1 = 0 and initial relaxation λ−1 = 1/2. Then

we have the following identities starting from k =−1:

yk+1 = argminy∈H2

g(y)−〈wkdg,Axk +By−b〉+

γ

2‖Axk +By−b+(2λk−1)(Axk +Byk−b)‖2

wk+1dg

= wkdg− γ(Axk +Byk+1−b)− γ(2λk−1)(Axk +Byk−b)

xk+1 = argminx∈H1

f (x)−〈wk+1dg

,Ax+Byk+1−b〉+ γ

2‖Ax+Byk+1−b‖2

wk+1d f

= wk+1dg− γ(Axk+1 +Byk+1−b)

Proof. By Equation (C) and Lemma 9, we get the following formulation for the k-thiteration: Given z0 ∈H

yk = argminy∈H2g(y)−〈zk,By−b〉+ γ

2‖By−b‖2;wk

dg= zk− γ(Byk−b);

xk = argminx∈H1f (x)−〈2wk

dg− zk,Ax〉+ γ

2‖Ax‖2;

wkd f

= 2wkdg− zk− γAxk;

zk+1 = zk +2λk(wkd f−wk

dg).

We will use this form to get to the claimed iteration. First,

2wkdg− zk = wk

dg− γ(Byk−b) and wk

d f= wk

dg− γ(Axk +Byk−b). (34)

Furthermore, we can simplify the definition of xk:

xk = argminx∈H1

f (x)−〈2wkdg− zk,Ax〉+ γ

2‖Ax‖2

(34)= argmin

x∈H1

f (x)−〈wkdg− γ(Byk−b),Ax〉+ γ

2‖Ax‖2

= argminx∈H1

f (x)−〈wkdg,Ax+Byk−b〉+ γ

2‖Ax+Byk−b‖2.

Note that the last two lines of Equation (C) differ by terms independent of x.We now eliminate the zk variable from the yk subproblem: because wk

d f+ zk =

2wkdg− γAxk, we have

Page 42: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

42 D. Davis and W. Yin

zk+1 = zk +2λk(wkd f−wk

dg)

(34)= zk +wk

d f−wk

dg+ γ(2λk−1)(Axk +Byk−b)

= wkdg− γAxk− γ(2λk−1)(Axk +Byk−b).

We can simplify the definition of yk+1 by with the identity in Equation (C):

yk+1 = argminy∈H2

g(y)−〈zk+1,By−b〉+ γ

2‖By−b‖2

(C)= argmin

y∈H2

g(y)−〈wkdg− γAxk− γ(2λk−1)(Axk +Byk−b),By−b〉+ γ

2‖By−b‖2

= argminy∈H2

g(y)−〈wkdg,Axk +By−b〉+ γ

2‖Axk +By−b+(2λk−1)(Axk +Byk−b)‖2.

The result then follows from Equations (C), (34), (C), and (C), combined with theinitial conditions listed in the statement of the proposition. In particular, note thatthe updates of x,y,wd f , and wdg do not explicitly depend on z. ut

Remark 2. Proposition 9 proves that wk+1d f

= wk+1dg− γ(Axk+1 +Byk+1− b). Recall

that by Equation (C), zk+1− zk = 2λk(wkd f−wk

dg). Therefore, it follows that

zk+1− zk =−2γλk(Axk +Byk−b). (35)

C.1 Dual feasibility convergence rates

We can apply the results of Section 5 to deduce convergence rates for the dual objec-tive functions. Instead of restating those theorems, we just list the following boundson the feasibility of the primal iterates.

Theorem 15. Suppose that (z j) j≥0 is generated by Algorithm 2, and let (λ j) j≥0 ⊆(0,1]. Then the following convergence rates hold:

1. Ergodic convergence: The feasibility convergence rate holds:

‖Axk +Byk−b‖2 =4‖z0− z∗‖2

γΛ 2k

.

2. Nonergodic convergence: Suppose that τ = inf j≥0 λ j(1−λ j)> 0. Then

‖Axk +Byk−b‖2 ≤ ‖z0− z∗‖2

4γ2τ(k+1)and ‖Axk +Byk−b‖2 = o

(1

k+1

).

Page 43: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 43

Proof. Parts 1 and 2 are straightforward applications of Corollary 1. and the FPR

identity: zk− zk+1 (35)= 2γλk(Axk +Byk−b). ut

C.2 Converting dual inequalities to primal inequalities

The ADMM algorithm generates the following five sequences of iterates:

(z j) j≥0,(wjd f) j≥0, and (w j

dg) j≥0 ⊆ G and (x j) j≥0 ∈H1,(y j) j≥0 ∈H2.

The dual variables do not necessarily have a meaningful interpretation, so it is de-sirable to derive convergence rates involving the primal variables. In this section wewill apply the Fenchel-Young inequality [2, Proposition 16.9] to convert the dualobjective into a primal expression.

The following proposition will help us derive primal fundamental inequalitiesakin to Proposition 2 and 3.

Proposition 10. Suppose that (z j) j≥0 is generated by Algorithm 2. Let z∗ be a fixedpoint of TPRS and let w∗ = proxγd f

(z∗). Then the following identity holds:

4γλk(h(xk,yk)) =−4γλk(d f (wkd f)+dg(wk

dg)−d f (w∗)−dg(w∗))

+

(2(

1− 12λk

)‖zk− zk+1‖2 +2〈zk− zk+1,zk+1〉

). (36)

Proof. We have the following subgradient inclusions from Proposition 9: A∗wkd f∈

∂ f (xk) and B∗wkdg∈ ∂g(yk). From the Fenchel-Young inequality [2, Proposition

16.9] we have the expression for f and g:

d f (wkd f) = 〈A∗wk

d f,xk〉− f (xk) and d f (wk

dg) = 〈B∗wk

dg,yk〉−g(yk)−〈wk

dg,b〉.

Therefore,

−d f (wkd f)−dg(wk

dg) = f (xk)+g(yk)−〈Axk +Byk−b,wk

d f〉−〈wk

dg−wk

d f,Byk−b〉.

Let us simplify this bound with an identity from Proposition 9: from wkd f−wk

dg=

−γ(Axk +Byk−b), it follows that

−d f (wkd f)−dg(wk

dg) = f (xk)+g(yk)+

1γ〈wk

d f−wk

dg,wk

d f+ γ(Byk−b)〉. (37)

Recall that γ(Byk−b) = zk−wkdg

. Therefore

wkd f+γ(Byk−b)= zk+(wk

d f−wk

dg)= zk+

12λk

(zk+1−zk)=1

2λk(2λk−1)(zk−zk+1)+zk+1,

Page 44: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

44 D. Davis and W. Yin

and the inner product term can be simplified as follows:

1γ〈wk

d f−wk

dg,wk

d f+ γ(Byk−b)〉= 1

γ〈 1

2λk(zk+1− zk),

12λk

(2λk−1)(zk− zk+1)〉

+1γ〈 1

2λk(zk+1− zk),zk+1〉

=− 12γλk

(1− 1

2λk

)‖zk+1− zk‖2

− 12γλk

〈zk− zk+1,zk+1〉. (38)

Now we derive an expression for the dual objective at a dual optimal w∗. First,if z∗ is a fixed point of TPRS, then 0 = TPRS(z∗)− z∗ = 2(w∗dg

−w∗d f) = −2γ(Ax∗+

By∗−b). Thus, from Equation (37) with k replaced by ∗, we get

−d f (w∗)−dg(w∗) = f (x∗)+g(y∗)+ 〈Ax∗+Bx∗−b,w∗〉= f (x∗)+g(y∗).(39)

Therefore, Equation (36) follows by subtracting (39) from Equation (37), rear-ranging and using the identity in Equation (38). ut

The following two propositions prove two fundamental inequalities that boundthe primal objective.

Proposition 11 (ADMM primal upper fundamental inequality). Let z∗ be a fixedpoint of TPRS and let w∗ = proxγdg(z

∗). Then for all k ≥ 0, we have the bound:

4γλkh(xk,yk)≤ ‖zk− (z∗−w∗)‖2−‖zk+1− (z∗−w∗)‖2 +

(1− 1

λk

)‖zk− zk+1‖2,

(40)

where the objective-error function h is defined in (20).

Proof. The lower inequality in Proposition 3 applied to d f +dg shows that

−4γλk(d f (wkd f)+dg(wk

dg)−d f (w∗)−dg(w∗))≤ 2〈zk+1− zk,z∗−w∗〉.

The proof then follows from Proposition 10, and the simplification:

2〈zk− zk+1,zk+1− (z∗−w∗)〉+2(

1− 12λk

)‖zk− zk+1‖2

= ‖zk− (z∗−w∗)‖2−‖zk+1− (z∗−w∗)‖2 +

(1− 1

λk

)‖zk− zk+1‖2. ut

Remark 3. Note that Equation (40) is nearly identical to the upper inequality inProposition 2, except that z∗−w∗ appears in the former where x∗ appears in thelatter.

Page 45: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 45

Proposition 12 (ADMM primal lower fundamental inequality). Let z∗ be a fixedpoint of TPRS and let w∗ = proxγdg(z

∗). Then for all x ∈H1 and y ∈H2 we havethe bound:

h(x,y)≥ 〈Ax+By−b,w∗〉, (41)

where the objective-error function h is defined in (20).

Proof. The lower bound follows from the subgradient inequalities:

f (x)− f (x∗)≥ 〈x− x∗,A∗w∗〉 and g(y)−g(y∗)≥ 〈y− y∗,B∗w∗〉.

We sum these inequalities and use Ax∗+By∗ = b to get Equation (41). ut

Remark 4. We use Inequality (41) in two special cases:

h(xk,yk)≥ 1γ〈wk

dg−wk

d f,w∗〉

h(xk,yk)≥ 1γ〈wk

dg−wk

d f,w∗〉.

These bounds are nearly identical to the fundamental lower inequality in Proposi-tion 3, except that w∗ appears in the former where z∗− x∗ appeared in the latter.

C.3 Converting dual convergence rates to primal convergence rates

We can use the inequalities deduced in Section C.2 to derive convergence ratesfor the primal objective values. The structure of the proofs of Theorems 9 and 10are exactly the same as in the primal convergence case in Section 5, except thatwe use the upper and lower inequalities derived in the Section C.2 instead of thefundamental upper and lower inequalities in Propositions 2 and 3. This amountsto replacing the term z∗− x∗ and x∗ by w∗ and z∗−w∗, respectively, in all of theinequalities from Section 5. Thus, we omit the proofs.

D Examples

In this section, we apply relaxed PRS and relaxed ADMM to concrete problems andexplicitly bound the associated objectives and FPR terms with the convergence rateswe derived in the previous sections.

Page 46: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

46 D. Davis and W. Yin

D.1 Feasibility problems

Suppose that C f and Cg are closed convex subsets of H , with nonempty intersec-tion. The goal of the feasibility problem is the find a point in the intersection ofC f and Cg. In this section, we present one way to model this problem using convexoptimization and apply the relaxed PRS algorithm to reach the minimum.

In general, we cannot expect linear convergence of relaxed PRS algorithm for thefeasibility problem. We showed this in Theorem 6 by constructing an example forwhich the DRS iteration converges in norm but does so arbitrarily slow. A similarresult holds for the alternating projection (AP) algorithm [3]. Thus, in this sectionwe focus on the convergence rate of the FPR.

Let ιC f and ιCg be the indicator functions of C f and Cg. Then x ∈ C f ∩Cg, if,and only if, ιC f (x)+ ιCg(x) = 0, and the sum is infinite otherwise. Thus, a point isin the intersection of C f and Cg if, and only if, it is the minimizer of the followingproblem:

minimizex∈H

ιC f (x)+ ιCg(x).

The relaxed PRS algorithm applied to this problem, with f = ιC f and g = ιCg , hasthe following form: Given z0 ∈H , for all k ≥ 0, let

xkg = PCg(z

k);xk

f = PC f (2xkg− zk);

zk+1 = zk +2λk(xkf − xk

g).

Because f = ιC f and g = ιCg only take on the values 0 and ∞, the objective valueconvergence rates derived earlier do not provide meaningful information, other thanxk

f ∈C f and xkg ∈Cg. However, from the FPR identity xk

f −xkg = 1/(2λk)(zk+1− zk),

we find that after k iterations, Corollary 1 produces the bound

maxd2Cg(x

kf ),d

2C f(xk

g) ≤ ‖xkf − xk

g‖2 = o(

1k+1

)whenever (λ j) j≥0 is bounded away from 0 and 1. Theorem 5 showed that this rateis optimal. Furthermore, if we average the iterates over all k, Theorem 3 gives theimproved bound

maxd2Cg(x

kf ),d

2C f(xk

g) ≤ ‖xkf − xk

g‖2 = O(

1Λ 2

k

),

which is optimal by Proposition 8. Note that the averaged iterates satisfy xkf =

(1/Λk)∑ki=0 λixi

f ∈C f and xkg = (1/Λk)∑

ki=0 λixi

g ∈Cg, because C f and Cg are con-vex. Thus, we can state the following proposition:

Page 47: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 47

Proposition 13. After k iterations the relaxed PRS algorithm produces a point ineach set with distance of order O(1/Λk) from each other.

D.2 Parallelized model fitting and classification

The following general scenario appears in [10, Chapter 8]. Consider the followinggeneral convex model fitting problem: Let M : Rn → Rm be a feature matrix, letb∈Rm be the output vector, let l : Rm→ (−∞,∞] be a loss function and let r : Rn→(−∞,∞] be a regularization function. The model fitting problem is formulated as thefollowing minimization:

minimizex∈Rn

l(Mx−b)+ r(x). (42)

The function l is used to enforce the constraint Mx = b+ν up to some noise ν in themeasurement, while r enforces the regularity of x by incorporating prior knowledgeof the form of the solution. The function r can also be used to enforce the uniquenessof the solution of Mx = b in ill-posed problems.

We can solve Equation (42) by a direct application of relaxed PRS and ob-tain O(1/Λk) ergodic convergence and o

(1/√

k+1)

nonergodic convergence rates.Note that these rates do not require differentiability of f or g. In contrast, the FBS al-gorithm requires differentiability of one of the objective functions and a knowledgeof the Lipschitz constant of its gradient. The advantage of FBS is the o(1/(k+ 1))convergence rate shown in Theorem 12. However, we do not necessarily assumethat l is differentiable, so we may need to compute proxγl(M(·)−b), which can besignificantly more difficult than computing proxγl . Thus, in this section we separateM from l by rephrasing Equation (42) in the form of Problem (3).

In this section, we present several different ways to split Equation (42). Eachsplitting gives rise to a different algorithm and can be applied to general convexfunctions l and r. Our results predict convergence rates that hold for primal ob-jectives, dual objectives, and the primal feasibility. Note that in parallelized modelfitting, it is not always desirable to take the time average of all of the iterates. In-deed, when r enforces sparsity, averaging the current r-iterate with old iterates, allof which are sparse, can produce a non-sparse iterate. This will slow down vectoradditions and prolong convergence.

D.2.1 Auxiliary variable

We can split Equation (42) by defining an auxiliary variable for My−b:

minimizex∈Rm,y∈Rn

l (x)+ r(y)

subject to My− x = b. (43)

Page 48: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

48 D. Davis and W. Yin

The constraint in Equation (43) reduces to Ax+By = b where B = M and A =−IRm .If we set f = l and g = r and apply ADMM, the analysis of Section C.3 shows that

|l(xk)+ r(yk)− l(My∗−b)− r(y∗)|= o(

1√k+1

)‖Myk−b− xk‖2 = o

(1

k+1

).

In particular, if l is Lipschitz, |l(xk)− l(Myk−b)|= o(1/√

k+1). Thus, we have

|l(Myk−b)+ r(yk)− l(My∗−b)− r(y∗)|= o(

1√k+1

).

A similar analysis shows that

|l(Myk−b)+ r(yk)− l(My∗−b)− r(y∗)|= O(

1Λk

)‖Myk−b− xk‖2 = O

(1

Λ 2k

).

In the next two splittings, we leave the derivation of convergence rates to the reader.

D.2.2 Splitting across examples

We assume that l is block separable: we have l(Mx−b) = ∑Ri=1 li(Mix−bi) where

M =

M1...

MR

and b =

b1...

bR

.Each Mi ∈ Rmi×n is a submatrix of M, each bi ∈ Rmi is a subvector of b, and∑

Ri=1 mi = m. Therefore, an equivalent form of Equation (42) is given by

minimizex1,··· ,xR,y∈Rn

R

∑i=1

li(Mixi−bi)+ r(y)

subject to xr− y = 0, r = 1, · · · ,R. (44)

We say that Equation (44) is split across examples. Thus, to apply ADMM to thisproblem, we simply stack the vectors xi, i= 1, · · · ,R into a vector x=(x1, · · · ,xR)

T ∈RnR. Then the constraints in Equation (44) reduce to Ax+By = 0 where A = IRnR

and By = (−y, · · · ,−y)T .

Page 49: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

Convergence rate analysis of several splitting schemes 49

D.2.3 Splitting across features

We can also split Equation (42) across features, whenever r is block separable inx, in the sense that there exists C > 0, such that r = ∑

Ci=1 ri(xi), and xi ∈ Rni where

∑Ci=1 ni = n. This splitting corresponds to partitioning the columns of M, i.e. M =[

M1, · · · ,MC], and Mi ∈Rm×ni , for all i = 1, · · · ,C. For all y ∈Rn, My = ∑

Ci=1 Miyi.

With this notation, we can derive an equivalent form of Equation (42) given by

minimizex,y∈Rn

l

(C

∑i=1

xi−b

)+

C

∑i=1

ri(yi)

subject to xi−Miyi = 0, i = 1, · · · ,C. (45)

The constraint in Equation (45) reduces to Ax+By = 0 where A = IRmC and By =−(M1y1, · · · ,MCyC)

T ∈ RnC.

D.3 Distributed ADMM

In this section our goal is to use Algorithm 2 for

minimizex∈H

m

∑i=1

fi(x)

by using the splitting in [49]. Note that we could minimize this function by refor-mulating it in the product space H m as follows:

minimizex∈H m

m

∑i=1

fi(xi)+ ιD(x),

where D = (x, · · · ,x) ∈H m | x ∈H is the diagonal set. Applying relaxed PRSto this problem results in a parallel algorithm where each function performs a localminimization step and then communicates its local variable to a central processor.In this section, we assign each function a local variable but we never communicateit to a central processor. Instead, each function only communicates with neighbors.

Formally, we assume there is a simple, connected, undirected graph G = (V,E)on |V | = m vertices with edges E that describe a connection among the differentfunctions. We introduce a variable xi ∈H for each function fi, and, hence, we setH1 = H m, (see Section 8). We can encode the constraint that each node communi-cates with neighbors by introducing an auxiliary variable for each edge in the graph:

minimizex∈H m,y∈H |E|

m

∑i=1

fi(xi)

subject to xi = yi j,x j = yi j, for all (i, j) ∈ E. (46)

Page 50: Convergence rate analysis of several splitting schemeswotaoyin/splittingbook/ch4-davis-yin.pdf · Convergence rate analysis of several splitting schemes 3 A large class of algorithms

50 D. Davis and W. Yin

The linear constraints in Equation (46) can be written in the form of Ax+By = 0for proper matrices A and B. Thus, we reformulate Equation (46) as

minimizex∈H m,y∈H |E|

m

∑i=1

fi(xi)+g(y)

subject to Ax+By = 0, (47)

where g : H |E|→ R is the zero map.Because we only care about finding the value of the variable x ∈H m, the fol-

lowing simplification can be made to the sequences generated by ADMM applied toEquation (47) with λk = 1/2 for all k ≥ 1 [51]: Let Ni denote the set of neighborsof i ∈V and set x0

i = α0i = 0 for all i ∈V . Then for all k ≥ 0,xk+1

i = argminxi∈H fi(x)+γ|Ni|

2 ‖xi− xki − 1

|Ni| ∑ j∈Ni xkj +

1γ|Ni|αi‖2 + γ|Ni|

2 ‖xi‖2

αk+1i = αk

i + γ

(|Ni|xk+1

i −∑ j∈Ni xk+1j

).

The above iteration is truly distributed because each node i ∈V only requires infor-mation from its local neighbors at each iteration.

In [51], linear convergence is shown for this algorithm provided that fi arestrongly convex and ∇ fi are Lipschitz. For general convex functions, we can de-duce the nonergodic rates from Theorem 10∣∣∣∣∣ m

∑i=1

fi(xki )− f (x∗)

∣∣∣∣∣= o(

1√k+1

)

∑i∈Vj∈Ni

‖xki − zk

i j‖2 + ∑i∈Vi∈N j

‖xkj− zk

i j‖2 = o(

1k+1

),

and the ergodic rates from Theorem 9∣∣∣∣∣ m

∑i=1

fi(xki )− f (x∗)

∣∣∣∣∣= O(

1k+1

)

∑i∈Vj∈Ni

‖xki − zk

i j‖2 + ∑i∈Vi∈N j

‖xkj− zk

i j‖2 = O(

1(k+1)2

).

These convergence rates are new and complement the linear convergence resultsin [51]. In addition, they complement the similar ergodic rate derived in [54] for adifferent distributed splitting.