D. C. Sorensen: SIAM Journal On Numerical Analysis, Vol. 19, No. 2. (Apr., 1982), Pp. 409-426
D. C. Sorensen: SIAM Journal On Numerical Analysis, Vol. 19, No. 2. (Apr., 1982), Pp. 409-426
D. C. Sorensen: SIAM Journal On Numerical Analysis, Vol. 19, No. 2. (Apr., 1982), Pp. 409-426
D. C. Sorensen
SIAM Journal on Numerical Analysis, Vol. 19, No. 2. (Apr., 1982), pp. 409-426.
Stable URL:
http://links.jstor.org/sici?sici=0036-1429%28198204%2919%3A2%3C409%3ANMWAMT%3E2.0.CO%3B2-F
SIAM Journal on Numerical Analysis is currently published by Society for Industrial and Applied Mathematics.
Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at
http://www.jstor.org/about/terms.html. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained
prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in
the JSTOR archive only for your personal, non-commercial use.
Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained at
http://www.jstor.org/journals/siam.html.
Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed
page of such transmission.
The JSTOR Archive is a trusted digital repository providing for long-term preservation and access to leading academic
journals and scholarly literature from around the world. The Archive is supported by libraries, scholarly societies, publishers,
and foundations. It is an initiative of JSTOR, a not-for-profit organization with a mission to help the scholarly community take
advantage of advances in technology. For more information regarding JSTOR, please contact [email protected].
http://www.jstor.org
Sat Oct 20 12:58:12 2007
SIAM J N U M E R . A N A L . a 1982 S o o e t ) for Industr~ala nd Applted \lathemat~cr
Abstract. A modified Newton method for unconstrained minimization is presented and analyzed. The
modification is based upon the model trust region approach. This report contains a thorough analysis of
the locally constrained quadratic minimizations that arise as subproblems in the modified Newton iteration.
Several promising alternatives are presented for solving these subproblems in ways that overcome certain
theoretical difficulties exposed by this analysis. Very strong convergence results are presented concerning
the minimization algorithm. In particular, the explicit use of second order information is justified by
demonstrating that the iterates converge to a point which satisfies the second order necessary conditions
for minimization. With the exception of very pathological cases this occurs whenever the algorithm is
applied to problems with continuous second partial derivatives.
*Received by the editors November 11, 1980, and in final revised form April 23, 1981. This work
was supported by the Applied Mathematical Sciences Research Program (KC-04-02) of the Office of
Energy Research, U.S. Department of Energy under contract W-31-109-Eng-38.
t Argonne National Laboratory, Argonne, Illinois 60439.
410 D. C . SORENSEN
directions of negative curvature to accomplish various tasks such as escape from saddle
points [ 5 ] ,[ 9 ] ,[13],search along more general paths [13],[19],[22],[23],[31],obtain
convergence to points that satisfy second order necessary conditions [13],[19],[22],
[23] etc. We observe along with Gay [ l o ] ,[ I l l that the method proposed here will
accomplish these things in a very elegant and intuitively appealing way.
It is hoped that this report will present a succinct but thorough analysis of this
method. In particular, we feel it is important to clearly describe the theoretical nature
of the locally constrained quadratic minimization in § 2. The analysis given in § 4 is
made sufficiently general to apply to several possible implementations. These
possibilities are described in 8 5 , where particular attention is paid to overcoming a
practical problem of implementation exposed by the theoretical discussion in 9 2. We
make an effort to offer several alternatives to implementation but shall make no
recommendations until there is numerical evidence to present.
2. Constrained quadratic minimization. An important portion of the uncon-
strained minimization procedure presented in rj 3 will be concerned with the solution
of the following problem:
~ e$ (t w ) = f +g T +;w
~ T ~ ~ .
(2.1)
Find p E R" such that 4 ( p )= min { ( ~ ( w 11)w: 115 A}.
In (2.1),B = B T E R'lxn; W , g E [Wfl ;f , A E [W with A > 0 , and 11. /I throughout is the 2-norm.
There are some important subtleties to this problem. The purpose of this section is
to give a complete discussion of the theoretical aspects of problem (2.1) and to expose
the nature of the computational difficulties that may be present.
Several authors have considered problem (2.1)or related problems. This problem
appears implicitly as a subsidiary calculation in Levenberg-Marquardt type algorithms
for nonlinear least squares [17], [18].The computational aspect of this calculation
was fully discussed by More in [21].A relatively early paper by Forsythe and Golub
[ 8 ]considers a closely related problem concerning minimization of the form
(2.2) min { ( x- b l T B ( x- 6 ) :Ilxll= 1).
While their work gives an extensive study of problem (2.2), it is not fully applicable
to problem (2.1) since g E range ( B ) may not hold, and the interior is not considered.
Gander [9]has also done closely related work, but does not explicitly consider problem
(2.1).Problem (2.1)first appeared as a subsidiary calculation in unconstrained minimiz-
ation in the work of Goldfeld, Quandt and Trotter [ l l ] .Hebden [15] made an
important contribution concerning the practical computation of a solution to (2.1).
More recently the problem has been discussed by Gay [ l o ] .H e gives a different proof
of the results in Lemmas 2.4 and 2.8.
If the method of Lagrange is applied to the equivalent problem
(2.3) min $ ( w ) s.t. w T~ 5 h2,
it is a straightforward conclusion of the first order necessary conditions [4, Chap. 21
that p solves (2.3) and hence (2.1) only if p satisfies an equation of the form
( B + AI)p = -g with A ( w Tw - A2) = 0 , where A 2 0 is the Lagrange multiplier associated
with the constraint w Tw 5 h2 .
LEMMA2.4. If p is a solution to (2.1) then p is a solution to an equation of the form
Proof. It has been noted that p must solve an equation of the form of (2.5) with
A E 0 and A (\vTw- A2)= 0. It remains to show that B + A I is positive semidefinite.
Suppose that p # 0. Since p solves (2.1), it also solves min {*(w): llwll= llpll}. It follows
that *(w) 2 *(p) for all w such that llwll= llpll. This inequality together with (2.5) gives
(2.6) f-PT(~
+AI)W + + W ~ B2
Wf - P T ( ~+ A I ) ~+ + p T ~ p .
Rearranging terms in (2.6) gives
for all w such that llw\l= llpll. Since p #O, it follows readily from (2.7) that B +AI is
positive semidefinite. If p = 0, it follows from (2.5) that g = 0. Therefore p = 0 solves
min.{tw T ~ :w 11 w 11 5 A} and we must conclude that B is positive semidefinite. Since
A 2 0 is necessary, B + A I is positive semidefinite. 0
Lemma 2.4 establishes necessary conditions concerning the pair A, p when p
solves (2.1). Our next result establishes sufficient conditions that will ensure p is a
solution to (2.1). These results are essentially given in [ I l l . However, we wish to
present a statement and proof of these results that is more complete and better suited
to this presentation.
LEMMA 2.8. Let A E R, p E R" satisfy
(2.9) ( B + AI)p = -g with B + A I positive semidefinite.
(i) If A = 0 and llpll5 A then p solves (2.1).
(ii) If llpll= A then p solves
Statements (i), (ii), (iii) are immediate consequences of (2.11). The uniqueness state-
ment follows from (2.10) because the inequality is strict when B + A I is positive definite
and w # p . 0
The solution of problem (2.1) is closely related to solving the nonlinear equation
(2.12) 4(a)= A, where 4 ( a ) = ll(B + a ~ ) - ' ~ l l .
Using the eigensystem of the symmetric matrix B together with the invariance of 11. ( 1
under orthogonal transformations it is easy to show if g # 0 that c$'(cu) is a rational
function with second order poles all belonging to a subset of the eigenvalues of -B.
Since lirn,,,+,4(a) = 0, it follows that (2.12) has a solution whenever A > 0 and g # 0.
We can construct a solution to problem (2.1) using a particular solution of (2.12).
Let A be the smallest eigenvalue of B ; let S I = {q E Rn:Bq = Alq}; let I3 be the largest
root of (2.12) when g # 0 and I3 = 0 when g = 0. If there is any q E S1such that gTq # 0
then I3 > - A l must hold. If ~ E S : then - A l is not a pole of 4. Thus &(-Al) is
well defined when g E S t and this is the only possibility for I3 5 - A t to occur. Put
412 D. C. SORENSEN
where q E S1,llqll= 1, and (t) denotes pseudo-inverse [28]. Note B + A 1 must be positive
semidefinite with this choice of A. Since q T ( +~ A I ) ~= 0 when A = -Al, it is easily
checked that p is a solution to (2.9) and satisfies either condition (i) or (iii) of Lemma
2.8. Thus p solves (2.1) and IIpII = A whenever A1 S O . The solution given by (2.13)
shows that p is not unique whenever g E S1 and 4(-Al) < A due to the arbitrary choice
of sign in defining 8.
This discussion of the theoretical subtleties of solving (2.1) indicates numerical
difficulties may arise when a solution to problem (2.1) is sought. The case g E ST,
A = - A l in (2.13) will give rise to a very sensitive numerical problem. Any computa-
tional technique for solving (2.9) will introduce roundoff error. However, in this
sensitive case, small perturbations in the quantities B, g, A can lead to large perturba-
tions of the solution p due to the fact that B + A 1 will be nearly singular. Apparently
the true nature of the difficulty here is the nonuniqueness of the solution p given by
(2.13).We illustrate this point with a simple example. Let
The difficult case arises for values of 77 such that 1/(1- T ) 2 5 A2. For these values of
77 the solutions to (2.1) are of the form
1 1
PT=-(l-n, 8)
(1-77)
,
with --+ o2 = A2.
used to greater advantage. This will become evident as we present some very strong
convergence results in # 4.
A particular method for obtaining numerical solutions to (2.1) will be suggested
in # 5 . For the moment we assume that a numerical solution p to problem (2.1) can
be obtained which satisfies.
(i) ( B + AI)p = -g + Sg with B + A I positive semidefinite,
(ii)
and
(iii)
I (IpII- A1 5 E ~ Awhen A > 0,
Ilpll d (1 + e2)A when A = 0,
some fixed 0 < e l < e2 in ( 0 , l ) that are consistent with the finite precision arithmetic.
The results of Lemma 2.8 imply that such a p solves the modified problem
where ( - e2)h5 5 (1+ e2)A and = g + Sg with 11Sgll5 elllgl( when g # 0. In our
A
analysis we shall assume e l = e2 = 0. A trivial but tedious modification of the analysis
would apply to a computed step p which satisfies the above criteria. This is primarily
because the crucial inequality of Lemma 2.14 will become
1
(2.16) f - ~ ( p )z~~~~~~
min(A,M, z L ( l e l ) ~ min
11B11 - 2
- ~g~~
ALGORITHM3.3.
(1) Let k = 1, and let 0 < q1 < q 2< 1, 0 < y1 < 1< y2 be prespecified constants;
(2) Let xl E Rn,A1 > 0 be given;
(3) If "convergence" then STOP;
(4) Evaluate fk:= f(xk); Gk :=v2f (xk);
Factor B k:=J ~ T G ~ J ~; Evaluate gk:=J C V ~( xk);
TRUST REGION MODIFICATION OF NEWTON'S METHOD 415
ared
ared
(8) If q ,S -then
pred
ared
2) if -> 772 then Ak:=y2hk;
pred
3) hk+,:=hk; k : = k + I ;
(9) Go to 3;
end.
There are ways to update the value of A at step 7 and step 8.2 which make better
use of the information available at the current iterate xk. For example, the cubic
polynomial that fits @?(a)= &(awk) by interpolating @(O), @?'(O),Q?"(O)and @(l) will
have a minimum & in (0, I ) when the test at step 7 is passed. The region is contracted
by setting yl = G if 4 is not "too close" to 0 or 1. Details of this type of idea appear
in [4], [8], [12], [18]. Similar ideas may be applied at (8.2) to obtain an expansion
factor y2 2 1 that depends upon available information. Other variations involving step
7 include accepting the predicted minimizer if 0 < qo5 aredlpred 5 q l but reducing
the trust region. The analysis we shall perform on Algorithm 3.3 can be adapted to
cover these possibilities in a fairly straightforward way. However, the gain in generality
will result in a substantial loss in clarity of exposition in the analysis so we shall analyze
the simple choices set forth in Algorithm 3.3.
Finally, it should be pointed out that this iteration is well defined because step
7 will produce a sufficiently small Ak to obtain ared/pred> q 1 after a finite number
of steps since the quadratic function t,!tk(w)is defined by the first three terms of the
Taylor series for <bk(w).Our statement of the strategy is slightly different than the
usual description in that xktl is always different from xk. By doing this we avoid having
to distinguish between "successful" and "unsuccessful" iterates in the analysis. With
this exception, the statement of the algorithm and the ensuing analysis are in the spirit
of the paper presented by Powell [26]. Numerical schemes for producing the con-
strained quadratic minimization at step 5 will be presented in § 5 .
4. Convergence of the modified Newton iteration. In this section we shall estab-
lish that some very strong convergence properties are possessed by Algorithm 3.3.
The first result is a slight modification of Powell's result in [26]. Our proof is much
simpler due to the fact that here second order information is explicitly available.
Throughout the analysis the notation for a level set o f f is
THEOREM4.1. Let f : Rn+ R be bounded below and let G(x) = v2f (x) be contirtuous
and satisfy IlG(x)lfa /3 fbr all x E ~ ' ( x o ) .Let {xk]c Rn be the sequence produced by
Algorithm 3.3 applied to f giuen starting ualue XO. Assume llJkll, 1 1 ~ 11~ s' c,k = 0 , 1 , 2 , .
f i r some rr 2 1. Then there is no constant E > 0 such that IlVf (xk)l(a E for all k.
Proof. Assume there is an E > 0 such that l(Df(xk)jl2 E for all k. Since gk = ~ ; f(xk), ~ f
416 D. C . SORENSEN
we have
(4.4)
la- w : ( ~+ A~ ( ~ ) I )+wA~( ~ )Fwk
w
1
5- sup lIBk(@)-Bkll,
050-1
(4.6)
i Y kJ 5 i
I, = max 1 E [ k , kj+,):IlgiI? g, =l .
I
From inequality (4.2)we obtain that
TRUST REGION MODIFICATION OF NEWTON'S METHOD 417
fki -f1,+1=
'i
Z fi -
1 =k,
fl-12
4u
min ( 1 A,, 2PuY
1,
l=ki
---7).
and the right-hand side of (4.9) is forced to zero due to (4.8). The uniform bound on
G ( . )implies the uniform continuity of V f ( x )on ~ ( x o and
) , it follows that
for all j sufficiently large. The assumption that (IVf(xk,)((Z e > 0 has led to a contradiction
and we must conclude that limk+mllVf(xk)ll= 0. 0
This result has established that every limit point of the sequence { k k )satisfies the
first order necessary conditions for a minimum. Now we shall establish results which
give added justification to the use of second order information when it is available.
Several authors [13],[19],[22],[23]have proposed modified Newton methods which
guarantee convergence to a critical point x* with the additional feature that the
Hessian G ( x * ) be positive semidefinite. Thus second order necessary conditions for
a minimum are satisfied by x*. The following series of results show that Algorithm
3.3 shares this property.
LEMMA4.10. Let the hypotheses of Theorem 4.1 be satisfied. If G ( x ) is uniformly
continuous on 2 ( x k o )then
, there is no positive number A > 0 such that A ' ~ 2
' A fork 2 ko.
Proof. If A ' ~ '2 A, then Ilwk((= Ak, due to Lemma 2.4. We conclude from inequality
(4.4) that
ared ( k ) 1
1 -
pred ( k )
I / 5 - sup ( ( ~ ~-(~0 ~) ( 1 ,
A oceci
where B k ( 0 )= J : ( G ~ ( x+~Osk))Jk.Since pred ( k )E A ' k ' ~ 2 k 2A A:, it follows that Ak + 0
because pred ( k )+ 0 . Now the boundedness of (IJk((, (IJkl(1, together with the uniform
continuity of G ( x )on 2 ( x k o ) g, ives
We must conclude, as in the proof of Theorem 4.1, that there exists a K>O such that
Ak 2 AK for some K > 0 . This contradiction establishes the result. 0
Since 5 h i k ) ,which is the smallest eigenvalue of B k , the next theorem follows
easily from the boundedness of llJkll,( ( J k (1l together with Lemma 4.10.
418 D. C . SORENSEN
(1) xk(:,l)+x:;
(2) k ( i , j + l ) > k ( i , j ) , f o r l s i s m ;
(3) {xk}= U 1 A: ;
(4) A, nA , = 0 for i # 1.
ared(k)-pred ( k ) l (sklI2
~ IlGk(0)- Gkll(1- 0 ) do,
0
and we must conclude that there is some K>O such that Ak 2 AK for all k 2 K .
< ahK,so the Newton
Since xk + X * with V f ( x * )= 0 , it follows that I I G ( x k ) - ' ~(xk)(l
f
step is accepted for all k sufficiently large. Hence the tail of the sequence { x k }is the
unmodified Newton iteration which is quadratically convergent to x* since G ( x * ) is
positive definite [24, p. 4211. U
While these results hold little computational meaning in the presence of roundoff
error, it is satisfying to have established such strong results about the iteration. This
is especially true since the method has such an intuitive appeal. Our aim in this section
has been to establish these theoretical results in a framework that is general enough
to encompass many possible implementations. We shall consider some of these
implementations in the next section.
5. Implementation. Numerical performance of the algorithm described in 9 3
and analyzed in 9 4 is obviously going to depend upon a careful implementation of
the locally constrained minimization of the quadratic model. In $ 2 we pointed out
several theoretical facts that indicate great care should be exercised in this computation.
In this section we shall put forth several possible implementations. Each of these will
have certain advantages and disadvantages depending upon the nature of the optimiz-
ation problem at hand. The convergence theory provided in $ 4 was purposely made
sufficiently general to apply to all of the alternative implementations to be presented
here.
Our main concern is to provide an efficient and stable method for the solution
of problem (2.1).To this end we consider factorizations
If the eigensystem of B is easily obtained (i.e., in case (a) or case (c) when B is
block diagonal) then we are able to solve problem (2.1) directly by solving the nonlinear
equation (2.12) for the largest root and then constructing a solution to (2.1) using
formula (2.13). This method of solution has the particular advantage that the case
when g E St is explicitly revealed. (Recall S1 = {q E R": Bq = Alq), where A l is the
smallest eigenvalue of B.)
A disadvantage of using factorization (a) is that it is relatively expensive to
compute. One of the reasons for introducing generality into the model trust region
calculation was to allow use of the Bunch-Parlett factorization [2]. This factorization
is very efficient due to the fact that symmetry is exploited. The matrix B for this
factorization has an eigensystem that is easily computed. Moreover, the matrices J
satisfy the criteria (IJ(1,( l ~ - ' l ( Z
a, so in theory all of the results of § 4 apply. There may
be some cause for concern regarding the effect of the transformation J on the descent
direction, because the triangular coordinate system may be very skewed even though
the matrix J is well conditioned.
Nevertheless, our main concern with either of these factorizations is the efficient
and reliable solution to an equation of the form
for the largest root A. The left-hand side of (5.1) is precisely the form of 4(a)=
l\(B+ a ~ ) ~ ' gin
\ I (2.12) regardless of whether or not B is diagonal. Several authors
[15], [21], [27] discuss the solution of equations that closely resemble (5.1). The key
observation is that Newton's method which is based on a local linear approximation
to 4(a) is not likely to be the best method for solving (5.1) because the rational
structure of d 2 ( a ) is ignored. Instead, an iteration for solving (5.1) can be derived
based upon a local rational approximation to 4 . The iteration is obtained by requiring
& ( a )= y/(a + @ )to satisfy
squares problem [21]. The result of MorC's work is a very elegant robust algorithm
for nonlinear least squares. In [21] careful attention is paid to safeguarding the step
calculation. The safeguarding task is somewhat more difficult in the present setting
due to the fact that B may have negative eigenvalues. The essential difficulty seems
to stem from the fact that without explicit knowledge of the eigensystem it is difficult
to detect the case g E S:. Moreover, it seems to be necessary to have an estimate of
the smallest eigenvalue and a corresponding eigenvector in order to obtain a solution
to (2.1) ih case g E S: (see (2.13)). This was recognized by Hebden but he did not
provide a suitable solution. Gay [lo] suggests obtaining an eigenvector using inverse
iteration if the case g E S: is detected because a factorization of the (nearly) singular
matrix B + A I will be available.
Here we suggest an alternative to the methods which have been proposed
previously. In the following we are considering J to be diaginal nonsingular matrix.
Let us return to the derivation of iteration (5.2). Another way to obtain this iteration
is to apply Newton's method to the problem.
From this observation we can see that iteration (5.2) is closely related to Newton's
method applied to the problem
and this matrix is singular at a solution A, p, of (2.1) in the sensitive case g~ s:,
IIB:gll< A, where A = - A l .
Of course, this situation impairs the local rate of convergence. Moreover, as the
iteration converges to such a solution the method requires solving linear systems which
have increasingly ill-conditioned coefficient matrices.
As an alternative, we suggest removing the explicit dependence of 4(a) on the
variable a in (5.4). Instead of (5.4) we shall apply Newton's method to solve
and this matrix is nonsingular at a solution to (2.1) in the cases that are most likely
to occur. This is important since it follows that Newton's method applied to (5.6) will
usually enjoy a quadratic rate of convergence. A precise statement of when (5.7) is
nonsingular at a solution is given in the following lemma.
LEMMA5.7. Let a = A 2 0, p = pA # 0 be a solution to problem (5.6) with B Apositive
semidefinite. If BA is positive definite or if IIBig((= A and dim (S1)= 1 then the Jacobian
matrix (5.7) is nonsingular.
Proof. Let p = pA, a = A. It is sufficient to show
Then
(i) B A z + p5= 0 and (ii) pTz = 0.
where T is tridiagonal and e: = (0, . . . , 0 , 1). The details of this factorization are
given in Algorithm 1.2 of Stewart's book [28]. Initially this factorization is more
expensive than some alternatives (such as the Bunch-Kaufman [3] factorization).
TRUST REGION MODIFICATION OF NEWTON'S METHOD 423
::jl
X X
X X X
X X X X
X X X x 9
X X X
x x
x x x x x x o
where the x's denote nonzero elements. Gaussian elimination with partial pivoting
preserves this structure if the pivots are taken from the tridiagonal part until the very
last elimination step. The result of this strategy applied to (5.13) will be of the form
m x
+
x x
where m's denote multipliers which have overwritten the original matrix elements,
and the +'s denote possible fill-in due to pivoting.
With this scheme only one expensive factorization is required. The rest of the
iteration is performed under the transformation (5.11) and only after convergence to
a vector j? is obtained do we transform back to get p = QTp^as a solution to problem
(2.1).
Since the factorization given in (5.11) is roughly four times as expensive as a
Cholesky factorization we might wish to consider the following alternative scheme.
The system at step 2.2 of Algorithm 5.10 is equivalent (via symmetric permutation)
to one of the form
424 D. C. SORENSEN
(5.16)
where
as an initial guess. Various schemes for computing a. are possible. See Gay [lo], for
example.
Safeguarding this iteration is possible. At present several schemes are being
considered but none of these are elegant. Therefore we shall postpone discussion of
safeguarding at this time.
The decision to stop the iteration should be based upon the following tests:
Require pk+lto satisfy
(a) Bktl :=B + ( Y ~positive
+ ~ Isemidefinite;
where 6a = ( Y ~ + ~ 8p - ( =YP~~, + ~ - P ~ ;
(c) 1 I ( ~ k + i l l - A1 5 E ~ A .
Note (from step 2.2 of Algorithm 5.10) that (B + a k + ' I ) p k + =
' -g+SaSp. Therefore,
if g # 0 then conditions (a) and (b) together with Lemma 2.8 imply that pk+, solves
m i n { f + g " T ~ + ~( (~~ T
( 1~5 6~1:,
A 6 5 (1+ E ~ ) Aand g" = g + 6g with ((6gl(5
where (1 - E ~ ) 5 ~lllglland I(pk+lll= A. On the
TRUST REGION MODIFICATION OF NEWTON'S METHOD 425
other hand, if g = 0 then pk+, will be an approximate eigenvector for B which satisfies
REFERENCES
[I] J. 0. AASEN,O n the reduction of a symmetric matrix to tridiagonal form, BIT 11 (1971), pp. 233-242.
[2] J. R. BUNCH AND B. N. PARLETT, Direct methods for solving symmetric indefinite systems of linear
equations, this Journal, 8 (1971), pp. 639-655.
[3] J. R. BUNCHAND L. KAUFMAN,Some stable methods for calculating inertia and solving symmetric
linear equations, Math. Comp. 31 (1977), pp. 163-179.
[4] A. V. FIACCOAND G. P. MCCORMICK,Nonlinear Programming: Sequential Unconstrained Minimiz-
ation Techniques, John Wiley, New York, 1968.
[5] R. FLETCHER,A modified Marquardt subroutine for nonlinear least squares, Atomic Energy Research
Establishment report R6799, Harwell, England, 1971.
[6] R. FLETCHERAND T. L. FREEMAN,A modified Newton method for minimization, J. Optim. Theory
Appl., 23 (1977), pp. 357-372.
[7] R. FLETCHER,Practical Methods of Optimization, vol. 1, John Wiley, New York;1980.
[8] G. E. FORSYTHEAND G. H. GOLUB, O n the stationary values of a second-degree polynomial on the
unit sphere, J. Soc. Indust. Appl. Math., 13 (1965), pp. 1050-1068.
[9] W. GANDER,O n the linear least squares problem with a quadratic constraint, Dept. Comp. Sci. Rept.,
STAN-CS-78-697, Stanford University, Stanford, CA, 1978.
[ l o ] D. M. GAY, Computing optimal locally constrained steps, MRC Rept. #2013, Mathematics Research
Center, Univ. of Wisconsin, Madison, October 1979.
[Ill - , O n solving robust and generalized linear regressionsproblems, MRC Rept. # 2000, Mathematics
Research Center, Univ. of Wisconsin, Madison, September 1979.
[12] P. E. GILL AND W. MURRAY, Newton type methods for unconstrained and linearly constrained
optimization, Math. Prog., 7 (1974), pp. 311-350.
[13] D. GOLDFARB,Curvilinear path steplength algorithms for minimization which use directions of negative
curvature, Math. Prog., 18 (1980), pp. 31-40.
[14] S. M. GOLDFELD,R. E. QUANDTAND H. F. TROTTER, Maximization by quadratic hill-climbing,
Econometrica, 34 (1966), pp. 541-551.
[15] M. D. HEBDEN,A n algorithm for minimization usingexactsecond derivatices, Atomic Energy Research
Establishment report TP5 15, Harwell, England, 1973.
[16] S. KANIELA N D A. DAX, A modified Newton's method for unconsfrained minimization, this Journal,
16 (1979), pp. 324-331.
[17] K. LEVENBERG,A method for the solution of certain nonlinear problems in least squares, Quart. Appl.
Math., 2 (19441, pp. 164-168.
[18] D. W. MARQUARDT,A n algorithm for least squares estimation of nonlinear parameters, SIAM J.
Appl. Math., 11 (1963), pp. 431-441.
[19] G. MCCORMICK,A modification of Armijo's step-size rule for negative curvature, Math. Prog., 1 3
(1977), pp. 111-115.
[20] L. MIRSKY,A n Infroduction to Linear Algebra, Oxford University Press (Clarendon), London and
New York, 1955.
[21] J. J . MORE, The Levenberg-Marquardt algorithm: implementation and theory, Lecture Notes in
Mathematics 630, G. A. Watson, ed., Springer-Verlag, Berlin-Heidelberg-New York, 1978, pp.
105-116.
[22] J. J. MORE AND D. C. SORENSEN,O n the use of directions of negative curnature in a modified Newton
method, Math. Prog. 16 (1979), pp. 1-20.
[23] H. MUKAI AND E. POLAK,A second order method for unconstrained optirnization, J. Optim. Theory
Appl., 26 (1978), pp. 501-513.
[24] J. M. ORTEGA AND W. C. RHEINBOLDT,Iteratice Solution of Nonlinear Equations in Seceral
Variables, Academic Press, New York, 1970.
[25] M. J. D. POWELL, A hybrid method for nonlinear equations, in Numerical Methods for Nonlinear
Algebraic Equations, P. Rabinowitz, ed., Gordon and Breach, London, 1970, pp. 87-114.
[261 - , Concergence properties of a class of minimization algorithms, in Nonlinear Programming 2,
0. L. Mangasarian, R. R. Meyer and S. M. Robinson, eds., Academic Press, New York, 1975.
[27] C. H. REINSCH,Smoothing by spline functions. II, Numer. Math. 16 (1971), pp. 451-454.
[28] G. W. STEWART,Introduction to Matrix Computations, Academic Press, New York, 1973.
[29] S. W. THOMAS,Sequential estimation techniques for quasi-Newton algorithms, Ph.D. thesis, Department
of Computer Science, Cornell University, Report TR75-227, January 1975.
[30] J. H. WILKINSON,The Algebraic Eigenvalue Problem, Oxford University Press (Clarendon), London
and New York, 1965.
[31] J.-P. VIAI. AND I. ZANG, Unconstrained optimization by approximation of the gradient path, Math.
Oper. Res., 2 (1977), pp. 253-265.
http://www.jstor.org
LINKED CITATIONS
- Page 1 of 1 -
This article references the following linked citations. If you are trying to access articles from an
off-campus location, you may be required to first logon via your library web site to access JSTOR. Please
visit your library's website or contact a librarian to learn about options for remote access to JSTOR.
References
2
Direct Methods for Solving Symmetric Indefinite Systems of Linear Equations
J. R. Bunch; B. N. Parlett
SIAM Journal on Numerical Analysis, Vol. 8, No. 4. (Dec., 1971), pp. 639-655.
Stable URL:
http://links.jstor.org/sici?sici=0036-1429%28197112%298%3A4%3C639%3ADMFSSI%3E2.0.CO%3B2-9
3
Some Stable Methods for Calculating Inertia and Solving Symmetric Linear Systems
James R. Bunch; Linda Kaufman
Mathematics of Computation, Vol. 31, No. 137. (Jan., 1977), pp. 163-179.
Stable URL:
http://links.jstor.org/sici?sici=0025-5718%28197701%2931%3A137%3C163%3ASSMFCI%3E2.0.CO%3B2-X
14
Maximization by Quadratic Hill-Climbing
Stephen M. Goldfeld; Richard E. Quandt; Hale F. Trotter
Econometrica, Vol. 34, No. 3. (Jul., 1966), pp. 541-551.
Stable URL:
http://links.jstor.org/sici?sici=0012-9682%28196607%2934%3A3%3C541%3AMBQH%3E2.0.CO%3B2-K
16
A Modified Newton's Method for Unconstrained Minimization
Shmuel Kaniel; Achiya Dax
SIAM Journal on Numerical Analysis, Vol. 16, No. 2. (Apr., 1979), pp. 324-331.
Stable URL:
http://links.jstor.org/sici?sici=0036-1429%28197904%2916%3A2%3C324%3AAMNMFU%3E2.0.CO%3B2-P
NOTE: The reference numbering from the original has been maintained in this citation list.