999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Variants of Alternating Minimization Method with Sublinear Rates of Convergence for Convex Optimization

2020-06-04 06:42:42ChengLizhiZhangHui
數學理論與應用 2020年2期

Cheng Lizhi Zhang Hui

Abstract The alternating minimization (AM) method is a fundamental method for minimizing convex functions whose variables consist of two blocks. How to efficiently solve each subproblem when applying the AM method is the most concerned task. In this paper we investigate this task and design two new variants of the AM method by borrowing proximal linearized techniques. The first variant is suitable for the case where one of the subproblems is hard to solve and the other can be directly computed. The second variant is designed for parallel computation. Theoretically, with the help of proximal operators, we first formulate the AM variants into a unified form, and then show sublinear convergence results under some mild assumptions.

Key words Alternating minimization Sublinear rate of convergence Convex optimization

1 Introduction

The alternating minimization (AM) method is a fundamental algorithm for solving the following optimization problem:

(1)

whereΨ(x,y) is a convex function. Starting with a given initial point (x0,y0), the AM method generates a sequence {(xk,yk)}k∈Nvia the scheme

xk+1∈argmin{H(x,yk)+f(x)},

(2a)

yk+1∈argmin{H(xk+1,y)+g(y)}.

(2b)

In the literature, lots of work existed concerning its convergence with certain assumptions. To obtain stronger convergence results for more general settings, the recent paper [1] proposed an augmented alternating minimization (AAM) method by adding proximal terms, that is

(3a)

(3b)

whereck,dkare positive real numbers. From practical computational perspective, the authors in another recent paper [5] suggested the proximal alternating linearized minimization (PALM) scheme:

(4a)

(4b)

In this paper, we concern both of the computational and theoretical aspects of the AM method. On the computational hand, we follow the proximal linearized technique employed by the PALM method and propose two new variants of the AM method, called AM-variant-I and AM-variant-II. They read as follows respectively:

(5a)

(5b)

and

(6a)

(6b)

AM-variant-I can be viewed as a hybrid of the original AM method and the PLAM method. The proximal linearized technique is only employed to update thex-variable, and the updating ofy-variable is as same as the original AM method. The idea lying in AM-variant-I is mainly motivated by the iteratively reweighted least square (IRLS) method where the subproblem with respective to (w.r.t.) they-variable can be easily computed but the subproblem w.r.t. thex-variable might greatly benefit from proximal linearized techniques. AM-variant-II is very similar to the PALM method. The only difference is that we use ?yH(xk,yk) rather than ?yH(xk+1,yk) when update they-variables. The biggest merit of this scheme is that it is very suitable for parallel computation.

The rest of this paper is organized as follows. In section 2, we list some basic properties and formulate all AM-variants into uniform expressions by using the proximal operator. In section 3, we first list all the assumptions that needed for convergence analysis, and then state the main convergence results with a proof sketch. The proof details are postponed to section 6. In section 4, we introduce two applications to show the motivation and advantages of AM-variant-I for some special convex optimization problem. Future work is briefly discussed in section 5.

2 Mathematical Preliminaries

In this section, we layout some basic properties about the gradient-Lipschitz-continuous functions and the proximal operator, and then formulate all AM-variants into uniform expressions based on the proximal operator.

2.1 Basic properties

Lemma 1([7]) Leth:Rn→R be a continuously differentiable function and assume that its gradient ?his Lipschitz continuous with constantLh<+∞:

Then, it holds that

and

Letσ:Rn→(-∞,+∞] be a proper and lower semicontinuous convex function. For givenx∈Rnandt>0, the proximal operator is defined by:

(7)

The following characterization about proximal operators is very important for convergence analysis and will be frequently used later.

Lemma 2([4]) Letσ:Rn→(-∞,+∞] be a proper and lower semicontinuous convex function. Then

if and only if for anyu∈domσ:

σ(u)≥σ(w)+t.

(8)

The next result was established in [5]; its corresponding result in the convex setting appeared in an earlier paper [3].

Lemma 3([5]) Leth:Rn→R be a continuously differentiable function and assume that its gradient ?his Lipschitz continuous with constantLh<+∞ and letσ:Rn→R be a proper and lower semicontinuous function with infRnσ>-∞. Fix anyt>Lh. Then, for anyu∈domσand anyu+∈Rndefined by

we have

2.2 Uniform AM-variant expressions

In what follows, we express all AM-variants in a uniform way.

AM-variant-I:

(9a)

(9b)

AM-variant-II:

(10a)

(10b)

The AAM method:

(11a)

(11b)

The PALM method:

(12a)

(12b)

These expressions can be easily derived by the first order optimality condition and the following basic fact:

We omit all the deductions here.

3 Main results

For convenience, we letz=(x,y)∈Rn1×Rn2andK(x,y)=f(x)+g(y). With these notations, the objective function in problem (1) equals toK(x,y)+H(x,y) orK(z)+H(z), andzk=(xk,yk).

3.1 Assumptions and convergence results

Before stating main results, we make the following basic assumptions throughout the paper:

Assumption 1The functionsf:Rn1→(-∞,+∞] andg:Rn2→(-∞,+∞] are proper and lower semicontinuous convex function satisfying infRn1f>-∞ and infRn2g>-∞. The functionH(x,y) is a continuously differentiable convex function over domf×domg. The functionΨsatisfies infRn1×Rn2Ψ>-∞.

Assumption 2The minimizer set of (1), denoted byZ*, is nonempty, and the corresponding minimum is denoted byΨ*. The level set

S={z∈domf×domg:Ψ(z)≤Ψ(z0)}

is compact, wherez0is some given initial point.

Besides these two basic assumptions, we need additional assumptions to analyze different AM-variants.

Assumption 3For any fixedy, the gradient ?xH(x,y) is Lipschitz continuous with constantL1(y):

Assumption 4For any fixedx, the gradient ?yH(x,y) is Lipschitz continuous with constantL2(x):

Assumption 5For any fixedy, the gradient ?yH(x,y)w.r.t. the variablesxis Lipschitz continuous with constantL3(y):

Assumption 6For any fixedx, the gradient ?xH(x,y)w.r.t. the variablesyis Lipschitz continuous with constantL4(x):

Assumption 7?H(z) is Lipschitz continuous with constantL5.

(13a)

(13b)

(13c)

Theorem 1(The convergence rate of AM-variant-I) Suppose that Assumptions 3 and 8 hold. Takeck=γ·L1(yk) withγ>1 and let {(xk,yk)}k∈Nbe the sequence generated by AM-variant-I. Then, for allk≥2,

Theorem 3(The convergence rate of the AAM method) Suppose that Assumptions 6 and 8 hold. Letck,dkbe positive real numbers such thatρ1=inf{ck,dk:d∈N}>0 andρ2=sup{ck,dk:k∈N}<+∞. Let {(xk,yk)}k∈Nbe the sequence generated by the AAM method. Then, for allk≥2,

Theorem 4(The convergence rate of the PALM method) Suppose that Assumptions 3, 4, 5, and 8 hold. Takeck=γ·L1(yk) anddk=γ·L2(xk+1) withγ>1. Let {(xk,yk)}k∈Nbe the sequence generated by the PALM method. Then, for allk≥2,

3.2 Sketch of the proof

In the light of [2], we describe a theoretical framework under which all the theorems stated above can be proved. Assume that a generic algorithmAgenerates a sequence {zk}k∈Nfor solving the problem (1). Our aim is to show that

Our proof mainly consists of two steps:

(a) Find a positive constantτ1such that

Ψ(zk)-Ψ(zk+1)≥τ1·d(zk,zk+1)2,k=0,1,2,…

whered(·,·) is some distance function.

(b) Find a positive constantτ2such that

Ψ(zk+1)-Ψ*≤τ2·d(zk,zk+1),k=0,1,2,…

(Ψ(zk)-Ψ*)-(Ψ(zk+1)-Ψ*)≥α·(Ψ(zk+1)-Ψ*)2,k=0,1,2,…

All the theorems directly follow by invoking the following lemma:

Lemma 4([2]) Let {Ak}k≥0be a nonnegative sequence of real numbers satisfying

Then, for anyk≥2,

4 Applications

In this part, we first explain our original motivation of proposing AM-variant-I by studying a recent application of the IRLS method; and then we apply AM-variant-I to solving a composite convex model. We begin with the general problem of minimizing the sum of a continuously differentiable function and sum of norms of affine mappings:

(14)

whereXis a given convex set,Aiandbiare given matrices and vectors, ands(x) is some continuously differentiable convex function. This problem was considered and solved in [2] by applying the IRLS method to its smoothed approximation problem:

(15)

or equivalently, by applying the original AM method to an auxiliary problem

(16)

where for a given setZthe indicator functionδ(x,Z) is defined by

(17)

(18a)

(18b)

In [2], the author first established the sublinear rate of convergence for the AM method and hence the same convergence result for the IRLS method follows. However, in many cases the subproblem of updatingxkis very hard to solve and even prohibitive for large-scale problems. It is just this drawback motivating us to propose AM-variant-I. Now, applying AM-variant-I and with some simple calculation, we at once obtain a linearized scheme of the IRLS method, that is

(19a)

(19b)

wherePXis the projection operator ontoX. IfPXcan be easily computed, then the linearized scheme becomes very simple. In addition, its sublinear rate of convergence can be guaranteed by Theorem 1. Nevertheless, we would like to point out that the scheme (34) can also be obtained by applying the proximal forward-backward (PFB) method [6] to the following problem

(20)

(21)

whereA∈Rm×nand the proximal operators offandgcan be easily computed. In [2], the author applied the AM method to its auxiliary problem

(22)

and obtained the following scheme:

(23a)

(23b)

Because the entries of vectorxare coupled byAx, the updating ofxkis usually very hard for large-scale problems. AM-variant-I fixes this problem and generates the following simple scheme:

(24a)

(24b)

5 Discussion

In this paper, we discussed a group of variants of the AM method and derived the sublinear rates of convergence under very minimal assumptions. Although we restricted our attention onto convex optimization problems, these variants for nonconvex optimization problems might obtain computational advantages over the AM method as well. Because our theory is limited to convex optimization, the convergence of AM-variant-I and AM-variant-II for general cases is unclear at present. In future pursuit, we will analyze the convergence of AM-variant-I and AM-variant-II under the nonconvex setting.

6 Proof details

6.1 Proof of Theorem 1

(25a)

(25b)

On the other hand, sinceyk+1minimizes the objectiveH(xk+1,y)+g(y), it holds that

(26)

By summing up the above two inequalities, we obtain

(27)

Step 2: prove the property (b). By Lemma 1, we have that

(28)

By the convexity ofH(z), it follows thatH(zk)-H(z*)≤?H(zk),zk-z*. Thus,

(29)

f(x*)≥f(xk+1)+ckxk-xk+1,x*-xk+1+

(30)

and

g(y*)≥g(yk)+.

(31)

By summing up the above two inequalities, we obtain

(32)

Combining inequalities (48) and (51) and noticingγ>1, we have that

(33)

From inequalities (25b) and (26), it follows that

(34)

(35)

6.2 Proof of Theorem 2

Step 1: prove the property (a). Sincexk+1andyk+1are the minimizers to the subproblems in (6) respectively, we get that

(36)

and

(37)

By summing up the above two inequalities, we obtain

(38)

By the convexity ofH(z), it follows that

H(zk)-H(zk+1)≥.

(39)

By summing up the above two inequalities, we get

(40)

By Assumption 7, it holds that

(41)

(42)

Now, combining inequalities (40), (41) and (42), we finally get that

(43)

Step 2: prove the property (b). By Assumption 7 and Lemma 1, we have

(44)

The convexity ofH(z) impliesH(zk)-H(z*)≤. Thus, we get

(45)

Applying Lemma 2 to (10), we obtain that

f(x*)≥f(xk+1)+ck+

(46)

and

g(y*)≥g(yk+1)+dk+.

(47)

By summing the above two inequalities, we get

K(z*)≥K(zk+1)+ck+dk
+.

(48)

Combining inequalities (45) and (48), we have

(49)

(50)

(51)

6.3 Proof of Theorem 3

Step 1: prove the property (a). Sincexk+1andyk+1are the minimizers to the subproblems in (18) respectively, we get that

(52)

and

(53)

By summing up the above two inequalities, we obtain

(54)

Sinceρ1=inf{ck,dk:k∈N}>0, we have

(55)

Thus, we finally get

(56)

Step 2: prove the property (b). Applying Lemma 2 to (18), we obtain that

f(x*)≥f(xk+1)+ck+

(57)

and

g(y*)≥g(yk+1)+dk+.

(58)

By summing up the above two inequalities and letting

we obtain

(59)

By the convexity ofH(z), it follows that

H(zk+1)-H(z*)≤.

(60)

Thus, combining (59) and (60) yields

By Assumptions 6 and 8, we deduce that

(61a)

(61b)

(61c)

By the Cauchy-Schwartz inequality and the notationρ2=sup{ck,dk:k∈N}, we have

(62)

(63a)

(63b)

This completes the proof.

6.4 Proof of Theorem 4

Step 1: prove the property (a). The following proof appeared in [5]. For completion, we include it here. Applying Lemma 3 to the scheme (12), we derive that

(64a)

(64b)

and

(65a)

(65b)

By summing up the above two inequalities, we obtain that

(66)

By Assumption 8, we finally get that

(67)

Step 2: prove the property (b). Applying Lemma 2 to (12), we obtain that

f(x*)≥f(xk+1)+ck+

(68)

and

g(y*)≥g(yk+1)+dk+.

(69)

By summing up the above two inequalities and letting

we obtain that

(70)

By Assumption 4 and Lemma 1, we have that

(71a)

(71b)

By Assumption 3 and Lemma 1 and the convexity ofH(z), we derive that

(72a)

(72b)

By summing up inequalities (71) and (72), we get

(73)

Together with (70) and utilizing Assumption 5 and the factγ>1, we derive that

(74a)

(74b)

(74c)

(74d)

(74e)

主站蜘蛛池模板: 国产91av在线| 日韩在线视频网| 草逼视频国产| 国产成人av一区二区三区| 韩国自拍偷自拍亚洲精品| 色综合网址| 日韩欧美在线观看| 国产麻豆精品在线观看| 国产欧美视频在线| A级毛片无码久久精品免费| 一本久道久久综合多人 | 日韩精品高清自在线| 欧美亚洲日韩不卡在线在线观看| 亚洲男人在线天堂| 2020精品极品国产色在线观看| 国产第四页| 色婷婷狠狠干| 久久国产精品波多野结衣| 国产精品亚洲va在线观看| 国产精品视频久| 日韩午夜片| 欧美中文字幕无线码视频| 欧美精品H在线播放| 午夜福利亚洲精品| 无码高清专区| 亚洲最大综合网| 成年人国产视频| 广东一级毛片| 欧美有码在线| 亚洲国产精品人久久电影| 伊伊人成亚洲综合人网7777| 亚洲妓女综合网995久久| 91青青草视频在线观看的| 六月婷婷激情综合| 五月婷婷综合网| 欧美成人一级| 性色生活片在线观看| 国产特一级毛片| 成人国内精品久久久久影院| 日a本亚洲中文在线观看| 国产99视频精品免费视频7 | 欧美综合成人| 日韩精品一区二区三区中文无码| 欧美日韩导航| 日韩黄色大片免费看| 免费无码AV片在线观看中文| 成年人久久黄色网站| 欧美成人a∨视频免费观看| 亚洲一区二区约美女探花| 亚洲人成网站色7799在线播放| 久久免费成人| 国产玖玖视频| 亚洲三级色| 久久 午夜福利 张柏芝| 青青草原国产精品啪啪视频| 亚洲精品无码高潮喷水A| 日韩欧美国产中文| 日韩av无码精品专区| www.91在线播放| 国产麻豆永久视频| 中国成人在线视频| 国产成人久久综合777777麻豆| 久久96热在精品国产高清| 狠狠干欧美| 日本免费福利视频| 人妻丰满熟妇啪啪| 精品久久香蕉国产线看观看gif| 亚洲综合第一区| 亚洲黄网在线| 久久99国产精品成人欧美| 日韩在线视频网| 欧美日韩综合网| 国产一区二区丝袜高跟鞋| 波多野结衣在线se| 一级毛片不卡片免费观看| 欧美亚洲欧美| 在线无码九区| 九九九久久国产精品| 亚洲精品无码在线播放网站| 2021精品国产自在现线看| 亚洲欧美精品一中文字幕| 国产三级国产精品国产普男人|