亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        A Descent Gradient Method and Its Global Convergence

        2014-07-24 15:29:22LIUJinkui

        LIU Jin-kui

        (School of Mathematics and Statistics,Chongqing Three Gorges University,Wanzhou 404100,China)

        A Descent Gradient Method and Its Global Convergence

        LIU Jin-kui

        (School of Mathematics and Statistics,Chongqing Three Gorges University,Wanzhou 404100,China)

        Y Liu and C Storey(1992)proposed the famous LS conjugate gradient method which has good numerical results.However,the LS method has very weak convergence under the Wolfe-type line search.In this paper,we give a new descent gradient method based on the LS method.It can guarantee the suffi cient descent property at each iteration and the global convergence under the strong Wolfe line search.Finally,we also present extensive preliminary numerical experiments to show the effi ciency of the proposed method by comparing with the famous PRP+method.

        unconstrained optimization;conjugate gradient method;strong Wolfe line search;suffi cient descent property;global convergence

        §1. Introduction

        Consider the unconstrained optimization problem

        where f:Rn→ R is continuously differentiable nonlinear function.The conjugate gradient method is one of the effective methods to solve the above problem and its iteration formulas

        where gk=▽f(xk),αk>0 is a step length determined by some line search,dkis the search direction andβkis a scalar.Some well-known conjugate gradient methods include Liu-Storey(LS) method[1]and Polak-Ribi`ere-Polyak(PRP)method[23].The parametersβkof these methods are specified as follows

        where||·||is the Euclidean norm.Obviously,if f is a strictly convex quadratic function,the above methods are equivalent in the case that an exact line search is used.If f is non-convex, their behaviors may be further different.

        In the past few years,the PRP method has been regarded as the most effi cient conjugate gradient method in practical computation.One remarkable property of the PRP method is that it essentially performs a restart if a bad direction occurs(see[4]).However,Powell[5]pointed out that the PRP method has a drawback that it be may not globally convergent when the objective function is non-convex.Powell[6]suggested that the parameterβkis negative in the PRP method and definedGilbert and Nocedal[7]considered Powell’s suggestion,and proved the globalconvergence ofthe modified PRP method(i.e.,PRP+method) for non-convex functions under the appropriate line search.The convergence properties of LS method have also been studied extensively(see[8-10]).However,the convergence of LS method has not been solved completely under the Wolfe-type line search.

        Due to solve the weak convergence of LS method under the Wolfe-type line search,we give a new descent gradient method based on the LS method.And the parameterβksatisfies

        In the next section,we give the corresponding algorithm and prove the suffi cient descent property ofthe new descent method without any line search.In section 3,the globalconvergence of the new method is given under the strong Wolfe line search.In section 4,numerical results are reported.

        §2. The Suffi cient Descent Property

        In this section,we firstly give the specific algorithm.Then the nice property ofthe presented algorithm is proved,which plays a prominent role in the proof of global convergence.

        Algorithm 2.1

        Step 1 Data x1∈Rn,ε≥0,δ=0.01,σ=0.1.Set d1=?g1,if||g1||≤ε,then stop.

        Step 2 Computeαkby the strong Wolfe line search

        Step 3 Let xk+1=xk+αkdk,gk+1=g(xk+1),if||gk+1||≤ε,then stop.

        Step 4 Computeβk+1by(1.3)and generate dk+1by(1.2).

        Step 5 Set k=k+1,go to step 2.

        Lemma 2.1 Let the sequences{gk}and{dk}be generated by Algorithm 2.1,then

        Proof The conclusion can be proved by induction.Since gT1d1=?||g1||2,the conclusion holds for n=1.Now we assume that the conclusion holds for n=k?1,for k≥2 and gk/=0. We need to prove that the result holds for n=k.Multiplying(1.2)by gTk,we have

        From(2.4),we get

        Obviously,the conclusion(2.3)also holds.Therefore,the conclusion(2.3)holds for k.

        §3.Global Convergence

        In this paper,we prove the global convergence of the new method under the following Assumption(H)

        (i)The level set?={x∈Rn|f(x)≤f(x1)}is bounded,where x1is the starting point.

        (ii)In a neighborhood V of?,f is continuously differentiable and its gradient g is Lipschitz continuous,namely,there exists a constant L>0 such that

        From Assumption(H),there exists a positive constantγ,such that

        The conclusion of the following lemma,often called the Zoutendijk condition,is used to prove the global convergence of nonlinear conjugate gradient methods.It was originally given by Zoutendijk[11].Due to that the strong Wolfe line search is a special case of the Wolfe line search,we also get the Zoutendijk condition holds under the general Wolfe line search.

        Lemma 3.1 Suppose that Assumption(H)holds.Let the sequences{gk}and{dk}be generated by Algorithm 2.1,then

        Lemma 3.2 Suppose that Assumption(H)holds.Let the sequences{gk}and{dk}be generated by Algorithm 2.1.If there exists a positive constant r,such that

        Then we have

        Proof According to the given condition,(2.3)holds.Then dk/=0,?k∈N+.

        From(1.2),we have

        According to the definition of uk,we have||uk||=1,then we have that

        By the definitions ofβkandδk,we haveδk≥0.Then,we obtain that

        By(2.3),(3.3)~(3.4)and(3.6),we have that

        Theorem 3.1 Suppose that Assumption(H)holds.Let the sequences{gk}and{dk}be generated by Algorithm 2.1.Then we have

        Proof The conclusion(3.7)willbe proved by contradiction.Suppose that(3.7)does not hold,then(3.4)holds.

        By the triangle inequality,we have

        whereξis the diameter of?,i.e.,ξ=max{||x?y||:?x,y∈?}.By(1.3),(2.3),(3.1)~(3.2) and(3.4),we have

        LetΔbe a large enough positive integer such that

        According to(3.5),the seriesis convergent,then there exits a k0large enough such that

        For?i∈[k+1,k+Δ],by the Cauchy-Schwarz inequality and(3.10),we have

        Combining this inequality and(3.8),we have

        Define Si=2ω2||si||2.By(1.2)and(3.2),for?l≥k0+1,we have

        From the above inequality,we have

        By the above inequality,the product is defined to be one whenever the index range is vacuous.Let us consider a product of consecutive Si,where k≥k0.Combining(3.10)and (3.12),by the arithmetic-geometric mean inequality,we have

        Then,we have that

        From(2.3),(3.3)~(3.4),we have

        which contradicts(3.14).Hence,we have that

        Tabel 1 The Numerical Results of PRP+Method and New Methods

        §4.Numerical Results

        In this section,we test Algorithm 2.1 on the unconstrained problems in[12]and compare its performance to that of the PRP+method under the strong Wolfe line search conditions.We stop the iteration if the iteration number exceeds 9999 or the inequity||gk||≤ 10?6is satisfied.All codes were written in MATLAB 7.0n a PC with 2.0GHz CPU processor and 512 MB memory and Windows XP operation system.

        In Algorithm 2.1,we select the parameter u=0,u=0.2 and u=1 and the corresponding methods are denoted as New1 method,New2 method and New3 method,respectively.

        The numericalresults ofour tests are reported in Table 1.The column“Problem”represents the problem’s name in[12].The detailed numerical results are listed in the form NI/NF/NG, where NI,NF,NG denote the number of iterations,function evaluations and gradient evaluations respectively.“Dim”denotes the dimension of the test problems.“NaN”means the calculation failure.

        In order to rank the average performance of allabove conjugate gradient methods,one can compute the total number of function and gradient evaluation by the formula

        where l is some integer.According to the results on automatic differentiation[13-14],the value of l can be set to 5,i.e.,

        That is to say,one gradient evaluation is equivalent to fi ve function evaluations if automatic differentiation is used.

        We compare new methods with PRP+method as follows:for the i th problem,compute the total numbers of function evaluations and gradient evaluations required by new methods and PRP+method by the above formula and denote them by Ntotal,i(New j)and Ntotal,i(PRP+), j=1,2,3.Then we calculate the ratio

        If the i0th problem is not run by the method,we use a constantλ=max{γi(method)|i∈S1} instead ofγi0(method),where S1denotes the set of the test problems which can be run by the method.

        The geometric mean ofthese ratios for new methods over allthe test problems is defined by

        where S denotes the set of the test problems,and|S|denotes the number of elements in S.One advantage of the above rule is that,the comparison is relative and hence does not be dominated by a few problems for which the method requires a great dealoffunction evaluations and gradient functions.According to the above method,we can obtain Table 2 and the results shows that average performance of new methods is much better than the PRP+method,which means that the effi ciency of the new methods is encouraging.

        Acknowledgments The author wishes to express their heartfelt thanks to the referees for their detailed and helpful suggestions for revising the manuscript.

        Table 2 The Relative Effi ciency of PRP+Method and New Methods

        [1]LIU Yu,STOREY C.Effi cient generalized conjugate gradient algorithms,Part 1:theory[J].Journal of Optimization Theory and Application,1992,69:129-137.

        [2]POLAK E,RIBIRE G.Note convergence de directions conjugees[J].Revue Fran?caise d’Informatique et de Recherche Op′erationnelle,1969,16:35-43.

        [3]POLAK B T.The conjugate gradient method in extreme problems[J].USSR Comput Math Math Phys, 1969,9:94-112.

        [4]HAGER W W,ZHANG Hong-chao.A survey of nonlinear conjugate gradient methods[J].Pacifi c Journal of Optimization,2006,2:35-58.

        [5]POWELL M J D.Nonconvex minimization calculations and the conjugate gradient method[R].Berlin: Springer,Germany,1984:122-141.

        [6]POWELL M J D.Convergence properties of algorithms for nonlinear optimization[J].SIAM Review,1986, 28:487-500.

        [7]GILBERT J C,NOCEDAL J.Global convergence properties of conjugate gradient methods for optimization[J].SIAM Journal on Optimization,1992,2:21-42.

        [8]LI Zheng-feng,CHEN Jing,DENG Nai-yang.A new conjugate gradient method and its global convergence properties[J].Systems Science and Mathematical Sciences,1998,11(1):53-60.

        [9]YU Gao-hang,ZHAO Yan-lin,WEI Zeng-xin.A descent nonlinear conjugate gradient method for large-scale unconstrained optimization[J].Applied Mathematics and Computation,2007,187:636-643.

        [10]LIU Jin-kui,DU Xiang-lin,WANG Kai-rong.Convergence of Descent Methods with Variable Parameters[J]. J ACTA MATHEMATICAE APPLICATAE SINICA,2010,33:222-232.

        [11]ZOUTENDIJK G.Nonlinear Programming,Computational Methods[M].Abaclce:North-Holland,1970: 37-86.

        [12]MORE J J,GARBOW B S,HILLSTROME K E.Testing unconstrained optimization software[J].ACM Trains Math Software 1981,7:17-41.

        [13]DAI Yu-hong,NI Qin.Testing diff erent conjugate gradient methods for large-scale unconstrained optimization[J].Journal of Computational Mathematics,2003,213:11-320.

        [14]GRIEWANK A.On Automatic Differentiation[M].Argonne Illinois:Kluwer Academic Publishers,1989: 84-108.

        tion:65K,90C

        1002–0462(2014)01–0142–09

        Chin.Quart.J.of Math. 2014,29(1):142—150

        date:2013-08-07

        Supported by The Youth Project Foundation of Chongqing Three Gorges University(13QN17);Supported by the Fund of Scientific Research in Southeast University(the Support Project of Fundamental Research)

        Biography:LIU Jin-kui(1982-),male,native of Anyang,Henan,a lecturer of Chongqing Three Gorges University,Master,engages in optimization theory and applications.

        CLC number:O221.2 Document code:A

        少妇高潮无套内谢麻豆传 | 亚洲中国美女精品久久久| 日本一区二区三区不卡在线| 午夜不卡无码中文字幕影院| 免费a级毛片出奶水| 手机看片福利日韩国产| 人妻露脸国语对白字幕| 日韩在线 | 中文| 亚洲欧美日韩人成在线播放| 国产日韩精品一区二区在线观看播放| 久久夜色精品国产三级| 国产成人精品无码免费看| 狼人香蕉香蕉在线28 - 百度| 精品午夜福利1000在线观看| 久久综合一本中文字幕| 国产综合开心激情五月| 国产三区在线成人av| 一级毛片60分钟在线播放| 久久爱91精品国产一区| 自拍偷自拍亚洲一区二区| 亚洲一区 日韩精品 中文字幕 | 国产微拍精品一区二区| 国产成人精品人人做人人爽| 精品中文字幕在线不卡| 国模雨珍浓密毛大尺度150p| 久久精品久久精品中文字幕| 狼人av在线免费观看| 国产99一区二区三区四区| 最近最新中文字幕| 日韩亚洲中文图片小说| 亚洲丝袜美腿精品视频| 日本又色又爽又黄的a片18禁| 无遮无挡三级动态图| 亚洲中文字幕人妻诱惑| 人妻少妇中文字幕久久| 亚洲色在线v中文字幕| 日韩av二区三区一区| 亚洲丝袜美腿精品视频| 大屁股人妻女教师撅着屁股| 欧美中文字幕在线| 蜜桃av福利精品小视频|