Fixed-Point Iteration Based Algorithm For A Class
Fixed-Point Iteration Based Algorithm For A Class
Fixed-Point Iteration Based Algorithm For A Class
Ashok D. Belegundu
Department of Mechanical and Nuclear Engineering
The Pennsylvania State University, University Park, PA 16802 USA
[email protected]
1. Introduction
tive, it is shown that the fixed-point iteration can be made to converge to a KKT
point. Opposite signs on the derivatives, viz. a positive objective function gradient
with a negative constraint function gradient can be handled via inverse variables.
An active set strategy is used to handle lower and upper bounds. While fixed-point
iteration algorithms can be found in the structural optimization literature [1-8],
these are presented without clearly stating assumptions under which convergence
may be achieved, and in fact, have been observed to not converge in certain situa-
tions without an explanation [5]. On the other hand, these algorithms have built-in
bells and whistles relating to step size control and Lagrange multiplier updates that
render them efficient on a wider variety of problems than targeted herein. Here, the
algorithm targets general functions f and g within a subspace of functions, as op-
posed to problem specific applications in structural optimization. Further, within
this general context, the fixed-point variable update formula is given physical sig-
nificance.
For the subclass of problems considered, the algorithm presented is vastly supe-
rior to descent based NLP methods such as sequential quadratic programming or
a generalized reduced gradient or feasible directions. In the latter category of
methods, a line search needs to be performed along a search direction, which is
computationally expensive as it involves multiple simulations in order to evaluate
the functions. Further, in fixed-point iterative methods, the number of function
evaluations involved in reaching the optimum are very weakly dependent on the
number of variables [7], allowing an efficient solution of problems with a large
number of variables. Fixed-point iterations have been applied extensively in the
area of equation solving and in market equilibrium in economics while hardly at all
in nonlinear programming. The algorithm may be applied to a subclass of problems
in resource allocation and inventory models, in addition to allocation of material in
structures [9, 10].
Fixed-point iterations have been used widely in equation solving. Two recur-
rence formulas that are prevalent will be illustrated in finding the root of a one
variable equation, x = g(x). This will pave the discussion of fixed-point iterations
for solving optimization problems in the next section. Convergence is established
by the following theorem [11, 12]:
Theorem 2.1. Assume α is a solution of x = g(x) and suppose that g(x) is con-
tinuously differentiable in some neighboring interval about α with g ′ (α ) < 1 ,
dg
where g ′ ≡ . Then, provided the starting point x0 is chosen sufficiently close to
dx
α, xk+1 = g(xk), k ≥ 0, with Limit x k = α .
k →∞
Fixed-point iteration based algorithm for a class of nonlinear programming problems 31
(( x ) 10 cos( x ) )
k p −1 k
=
which, for this example, converges for p ≥ 6 provided the starting guess is close
enough. Formulas in Eqs. (1) or (2) will be used for optimization problems as di-
scussed subsequently.
3.1. Assumptions
where xL are lower bounds. Descent based methods like sequential quadratic pro-
gramming and others referred to above update variables as
xk+1 = xk + βk dk (4)
∂f ∂g
+µ =0
∂x j ∂ xj
while increasing the constraint value until it reaches its limit. Thus, with the stated
assumptions, the solution of (3) using optimality conditions yields a KKT point.
The fixed-point iteration is derived as follows. The constraint g(x) ≤ 0 is linear-
n
∂g ∂g k
ized as cTx ≤ c0, where c j = > 0, c0 = − g (x k ) + ∑ xi with the derivatives
∂xj ∂
i =1 xi
evaluated at the current point xk. Working with the linearized constraint, the local
optimization problem may be stated as {min. f (x) : cTx ≤ c0, x ≥ xL}. We define
a Lagrangian function L = f + µ (cT x − c0), and an active set
I = { i : xi = xiL , 1≤i ≤ n } . The optimal point x is obtainable from
x ∈ arg minL L ( x , µ ) . The optimality conditions can be stated as
x≥x
∂L
for j ∈ I , x j = x Lj , ≥0
∂x j
(6)
∂L ∂ f
for j ∉ I , x j > x Lj , = + µ cj = 0
∂x j ∂x j
j∈I: xj = xjL
∂f
− x j ∂x
= max
j
j∉I: c j x kj +1 , c j x Lj (7)
µ
Substituting Eq. (7) into cT x = c0 gives an expression for µ, which when substi-
tuted back into Eq. (7) gives
j∈I: xj = xjL
j∉I: c j x kj +1 = max E j (c0 − ∑ cr xrL ) , c j x Lj (8)
r ∈I
where
∂f
xj −
∂x
Ej = n
j
(9)
∂f
∑ xi −
i =1
i∉ I
∂xi
34 A.D. Belegundu
A starting point x0 > xL is used which may be feasible or infeasible with respect
to the constraint. We use the fixed-point recurrence relation or ‘re-sizing’ formula:
j ∈ I : xj = xjL
j ∉ I : c j y = max E j (c0 − ∑ cr xrL ) , c j x Lj (10)
r ∈I
(11b)
Basic steps are given in the algorithm below.
Input Data:
x0, xL, w (or p), gmax, move limit ∆ max , Iteration limit, tolabs, tolrel
Set initial active set I0 = {null set}
Typical Iteration:
∂ f ∂g
i. Evaluate derivatives , and then evaluate c, c0, Ej.
∂x j ∂x j
ii. Compute xtrial ≡ xk+1 from Eq. (11a) or (11b).
iii. Reset xk+1 to satisfy move limits based on xtrial
j − x kj ≤ ∆ max x kj as well as to
satisfy the bounds.
iv. Based on xk+1, update active set from Ik to Ik+1. A free variable that has reached
x k +1
a bound is included in the active set based on whether 1 − j L ≥ − g max for
xj
each j ∉ I k . Also, variables in Ik are tested if they should become free.
v. Stopping criteria is based on (i) iteration limit (= number of function calls), (ii)
based on whether
f k − f k −1 ≤ tolrel f k −1 + tolabs
and g ( x k ) ≤ g max
sents the available resource after accounting for variables at their bounds. If a vari-
able xi ≡ θ represents an angle and crosses zero as 0 ≤ θ ≤ 2π , it can be substituted
by 2π ≤ θ ≤ 4π .
Importantly, the fixed-point iteration variable update in Eq. (10) can be physi-
∂f ∆f
cally interpreted as follows. Firstly, the derivative = lim has often been
∂x j ∆ x j → 0 ∆x j
termed a ‘sensitivity coefficient’ since it represents, to within a linear approxima-
tion, the change in f due to a unit change in xj. Similarly, the term
∂f ∆f
xj = lim represents the change in f due to a unit percentage change in
∂x j ∆ x j → 0 ∆ x j
xj
xj within a linear approximation. Thus, as x → x*, the contribution of the jth term
(cj xj) in the constraint
c1 x1 + c2 x2 + ... + ( cj xj ) + ... + cn xn = c0
36 A.D. Belegundu
bles at their bounds. Thus, the variable updates for xj only compete for the resource
that is available. The resizing of variables can thus be expressed as “re-allocate
term (ci xi) = (fractional % reduction in objective) x (available resource)”. Interest-
∂f
ingly, the term x j has only the units of f and is independent of the units of xj.
∂x j
Also, the iterations are driven by sensitivity only whereas in NLP based descent
methods, the function f is also monitored during the line search phase.
4. Applications
4.1. Searching for a missing vessel
a j , b j , b > 0, x ∈ R+n
n
−bj xj −b j x j
Defining f =−∑aj (1− e ) for minimization, we have df / dx j = − a j bj e < 0,
j=1
lation Mj with the number of samples chosen xj. The total sample is to be allocated
to the strata to obtain a minimum variance of the global estimate [9]:
n ( M j − x j )σ 2j
minimize ∑ω
j =1
2
j
( M j −1) x j
n
subject to ∑x
i =1
i = b,
x j ≥ 1, j =1, … , n
Mj
where ω j = , σ 2 is an appropriate estimate of the variance .in each strata, b is
M
the total sample size. We have df / dx j = − ω 2j σ 2j ( M j −1) M j < 0 and derivative of
the constraint function g ≤ 0 is 1. Further, variables are positive. The fixed-point
algorithm can thus be applied.
εj
Ej = n
(13)
∑ε
i =1
i
i∉I
38 A.D. Belegundu
That is, variable update or re-sizing is governed by the energy in the jth element
as a fraction of the total energy. Thus, many references exist in structural optimiza-
tion to energy based optimality criteria methods, which are shown here as special
cases stemming from a general formula.
5. Numerical examples
Examples are presented to show that the algorithm is far superior to a general
nonlinear programming method for problems that satisfy Assumptions 3.1 and 3.2.
The fixed-point algorithm is compared to MATLAB’s constrained optimizers
(sequential quadratic programming and interior point optimizers under ‘fmincon’).
Consider the simple problem of allocating material to minimize the tip deflec-
tion of a cantilever beam with two length segments, with certain assumed problem
data:
32 1
minimize f = +
x12 x22
subject to 2 x1 + x2 ≤ 1
x>0
64 / x12 2 / x22
From Eq. (9), E1 = , E = . The recurrence
( 64 / x12 + 2 / x22 ) 2 ( 64 / x12 + 2 / x22 )
formula in Eq. (11a) becomes
c0
x kj +1 = w E j + (1 − w) xkj
cj
Fixed-point iteration based algorithm for a class of nonlinear programming problems 39
Table 1
Searching for a missing vessel
tk n
a1 i j
subject to g1 (x) ≡ ∑C ∏x
i =1
1i
j =1
j ≤1
0 < x L ≤ x ≤ xU
with the restriction that all coefficients Ck i > 0, a0 i j < 0 , a1 i j > 0 . Note that
Ck i = coefficient of the kth function, ith term, while ak i j = exponent corresponding
to the kth function, ith term, and jth variable. Derivatives are calculated analytical-
ly.
Data: xL = 10‒6, xU = 1, C0 i ∈ (0,1) , C1 i ∈ (0.2,1) , a0 ij ∈ (−1, 0) , ak i j ∈ (0 ,1) ,
t0 = tk = 8, x0 = 0.5. To ensure that the constraint is active, ∑C 1i >1.
i
Table 2 shows results for various random generated trials. The fixed-point itera-
tion method far outperforms the NLP routine for the class of problems considered.
Total computation in NLP methods increase with n while fixed-point iteration met-
hods are insensitive to it, as also noted in the structural optimization literature.
40 A.D. Belegundu
Table 2
Randomized test problem
4
n NFE as per NFE as per f *_fixed-point /
Fixed-point algorithm1 NLP algorithm2 f *_NLP 2
10 50 363 1.027
20 50 1,131 1.030
3
40 50 3,008 1.003
1
fixed value
2
averaged over 5 random trials
3
larger values of n cannot be solved in reasonable time by the NLP routine
4
NFE = Number of Function Evaluations
6. Conclusions
References
[1] Venkayya V.B., Design of optimum structures, Computers and Structures 1971, 1, 265-309.
[2] Berke L., An efficient approach to the minimum weight design of deflection limited structures,
AFFD1-TM-70-4-FDTR, Flight Dynamics Laboratory, Wright Patterson AFB, OH 1970.
[3] Khot N.S., Berke L., Venkayya V.B., Comparison of optimality criteria algorithms for minimum
weight design of structures, AIAA Journal 1979, 17(2), 182-190.
[4] Dobbs M.W., Nelson R.B., Application of optimality criteria to automated structural design,
AIAA Journal 1975, 14(10), 1436-1443.
[5] Khan M.R., Willmert K.D., Thornton W.A., Optimality criterion method for large-scale struc-
tures, AIAA Journal 1979, 17(7), 753-761.
[6] McGee O.G., Phan K.F., A robust optimality criteria procedure for cross-sectional optimization
of frame structures with multiple frequency limits, Computers and Structures 1991, 38(5/6), 485-
-500.
[7] Yin L., Yang W., Optimality criteria method for topology optimization under multiple con-
straints, Computers and Structures 2001, 79, 1839-1850.
[8] Belegundu A.D., A general optimality criteria algorithm for a class of engineering optimization
problems, Engineering Optimization 2015, 47(5), 674-688.
[9] Patriksson M., A survey on the continuous nonlinear resource allocation problem, European
Journal of Operational Research 2008, 185(1), 1-46.
[10] Jose J.A., Klein C.A., A note on multi-item inventory systems with limited capacity, Operations
Research Letters 1988, 7(2), 71-75.
[11] Atkinson K.E., An Introduction to Numerical Analysis, John Wiley, 1978.
[12] Bryant V., Metric Spaces: Iteration and Application, Cambridge University Press, 1985.
[13] Belegundu A.D., Chandrupatla T.R., Optimization Concepts and Applications in Engineering,
2nd edition, Cambridge University Press, 2011.