1. Introduction
Many problems arising from various applications such as optimization, differential equations, variational inequalities problems and so on, can be converted into nonlinear system of equations. Hence the study of iterative algorithms for solving nonlinear equations is of paramount importance especially when analytical method is not feasible or difficult to implement.
Let
be a monotone mapping and
be a subset of
. We wish to find a point
such that
The feasible set
is assumed to nonempty closed and convex. We call problem (
1) system of nonlinear monotone equations with convex constraints. This problem appears as a subproblem in generalized proximal algorithms with Bregman distance [
1]. In addition, some monotone variational inequality problems of finding
for which
can be converted into systems of monotone equations [
2]. Furthermore,
norm regularized optimization problems can be reformulated as monotone nonlinear equations [
3].
Consider the following unconstrained optimization problem
where
is assumed to be continuous, bounded below and its gradient, denoted by
exists. Fermat’s extremum theorem suggests that if a point
is the local minimizer of the unconstrained optimization problem (
2) then problem (
1) holds. In addition, suppose
is the minimizer of problem (
2), then problem (
1) is the first order necessary condition for the unconstrained optimization problem (
2). This also underlines the importance of problem (
1).
Starting from a given initial point
popular iterative methods, such as Newton’s method, quasi-Newton method, conjugate gradient method, for solving (
2) use an updating rule defined as follows
where
and
denote stepsize and search direction respectively.
The search direction in (
3) is usually defined as
where
is either the exact Hessian matrix
in the case of Newton’s method or the approximation of the Hessian matrix in the case of quasi-Newton method. The approximation of the Hessian matrix,
is required to satisfy the following secant equation
and
The Quasi-Newton method was developed to overcome one of the major shortcomings associated with the famous Newton’s method which is the need to compute second derivative of the objective function in every iteration. However, it inherits the problem of storing
matrices throughout the iteration process which makes it unsuitable for large scale problems. One of the crucial approaches developed to overcome the storage problem of the quasi Newton method is the matrix-free method proposed by Barzilai and Borwein (BB) [
4]. The BB method uses (
3) to generate the next iterate with the search direction given by
and the stepsize taken as diagonal matrix
which is supposed to satisfy the secant Equation (
4). However, since
produces diagonal matrices with identical diagonal elements, it is usually very difficult to find
for which
satisfies (
4) when the dimension is greater than one. Consequently, Barzilai and Borwein required
to approximately satisfies (
4) by finding
which minimizes the following least square problems
and
The solutions of the minimization problems (
5) and (
6) are respectively given as
By Cauchy Schwarz inequality, we see that the stepsize produced by
is always greater than or equal to the one produced by
whenever
Barzilai and Borwein proved that the iterative scheme (
3) with
and
converges with R-superlinear rate for two-dimensional strictly convex quadratic problems.
One disadvantage of the BB method, however, is that the stepsizes
and
may become negative if the objective function is not convex. Thus, Dai et al. [
5] proposed and analyzed the following positive stepsize
The stepsize (
8) is the geometric mean of
and
They showed that the iterative scheme (
3) with
has the same rate of convergence with the stepsize
under certain conditions for two-dimensional strictly convex quadratic functions. Recently, Dai et al. [
6] proposed a family of gradient methods whose stepsize is a convex combination of
and
The stepsize is obtained by solving the following problem
It was shown that if
and
then
has a unique solution in the closed interval
,
They proved that their method is R-superlinearly and R-linearly convergent for two- dimensional strictly convex quadratics and any finite dimensional case respectively. Convergence analysis of the BB stepsizes has been explored and interested reader may refer to the following References [
7,
8,
9,
10,
11,
12].
On the other hand, the BB method with the stepsize
has been extended to solve unconstrained nonlinear equations by La Cruz and Raydan [
13]. Their algorithm is built on the strategy of nonmonotone line search technique which guarantees the global convergence of the method. Numerical experiments presented reveal their method competes with some well-established existing methods. However, their algorithm requires descent directions with respect to the squared norm of the residual. This means computation of a directional derivative, or its good approximation is needed at every iteration. Consequently, La Cruz et al. [
14] proposed another BB method with a different nonmonotone line-search technique for solving unconstrained nonlinear equations. Their approach has advantage because unlike the former, the computations of directional derivatives are completely avoided. Based on the projection technique of Solodov and Svaiter [
15], Zhang and Zhou [
16] proposed an interesting projection spectral method which can be viewed as a modification of the method given in References [
13,
14]. They proposed a new line search strategy which does not require any merit function and takes the monotonicity of
F into account. They established the global convergence of the method under some suitable assumptions and present some numerical experiments to demonstrate its computational advantage. In Reference [
17], Yu et al. extended the method given by Zhang and Zhou [
16] to solve monotone system of nonlinear equations with convex constraints. Their method is globally convergent under some conditions and preliminary numerical results show that the method works well and is more suitable compared to the projection method in Reference [
18]. Recently, Mohammad and Abubakar [
19] proposed a positive spectral method for unconstrained monotone nonlinear equations based on the projection technique in Reference [
15]. The spectral parameter proposed is a convex combination of modified
and
Their method works well and was extended to solve monotone nonlinear equations with convex constraints in Reference [
20] as well as signal and image restoration in Reference [
21].
Inspired by above contributions, we propose a two step iterative scheme based on the projection technique for solving system of monotone nonlinear equations with convex constraints. We define two search directions using Barzilai and Borwein (BB1 and BB2) spectral parameters with modifications. In addition, we investigate the efficiency of the propose algorithm in restoring blurred images. The symbols
and
denote inner product and Euclidean norm respectively. The remaining part of this paper is organized as follows. In
Section 2, we describe the proposed method and its global convergence. We report numerical experiments to show the efficiency of the algorithm in
Section 3. We describe the application of the proposed algorithm in
Section 4 and give some conclusions as well as possible future research perspective in
Section 5.
2. Two Step Iterative Scheme and Its Convergence Analysis
We begin this section with the following definition.
Definition 1. Let a mapping is said to be
- (i)
- (ii)
Lipschitzian continuous if there exists such that
From the discussions in the preceding section, we observe that all the methods use the one-step formula (
3) to update their respective sequence of iterates. Let
I be an identity map in
if we set
then formula (
3) is closely related to the well-known Mann iterative scheme [
22]
where
Mann iteration has been applied to solve different kind of nonlinear problems successfully. However, its convergence speed is relatively slow. Different studies have shown that the famous two-step Ishikawa iterative scheme [
23]
where
converges faster than the one-step Mann iteration.
Let
and
then Ishikawa iterative scheme can be rewritten as follows
Based on the fact that the two step Ishikawa iterative scheme has faster convergence speed than the one-step Mann iterative scheme, in this paper, we propose a new two-step iterative scheme incorporating nonnegative BB parameters with projection strategy to solve monotone nonlinear equation with convex constraints. Given a starting point
and
we define the updating formula for the proposed two-step scheme as follows
where
is a projection operator defined below and
For simplicity we denote
and
. The parameters
and
are modifications of the BB parameters (
7) given as follows
where
Assumption 1. Throughout this paper, we assume the following
- (i)
The solution set of problem (1) is nonempty. - (ii)
The mapping satisfies (10)–(11). - (iii)
The sequence is in such that
The following Lemma shows that the spectral parameters (
17) are well-defined and bounded.
Lemma 1. Suppose that Assumption 1 holds and , then we haveandwhere and Proof of Lemma 1. The monotonicity of
F gives
. Therefore, by the definition of
and
we have
On the other hand, by (
11) and Cauchy Schwarz inequality, we have
Also since
, from (
23) we can have
Therefore, by (
19) and (
21) we have
and from (
20), (
22) and (
24) we have
□
Remark 1. We give the following remarks
- (i)
From Lemma 1, it is not difficult to see that the two search directions and satisfy the descent condition. That is, - (ii)
The two search directions and satisfy the following inequalities
Next, we describe the projection operator in (
15) which is usually used in iterative algorithms for solving problems such as fixed point problem, variational inequality problem, and so on. Let
and define an operator
by
The operator
is called a projection onto the feasible set
and it enjoys the nonexpansive property, that is,
If
then
and therefore, we have
We now state the steps of the proposed algorithm which we call two-step spectral gradient method.
Remark 2. We quickly note the following remarks
- (i)
We claim that there exists a step-size satisfying the line search (1) for any Suppose on the contrary that there exists some such that for any the line search (1) is not satisfied, that is Since F is continuous and is bounded for all k, letting yields It is clear that the inequality (29) cannot hold. Hence the line search (1) is well-defined. - (ii)
The line search defined by (1) is more general than that of Reference [24]. - (iii)
It follows from (15) and Assumption 1 that
The next Lemma is very crucial to the convergence of Algorithm 1.
Algorithm 1: Two-Step Spectral Gradient Projection Method (TSSP) |
|
Lemma 2. Let the Assumption 1 holds, then the sequences and generated by Algorithm 1 are bounded. In addition, there exist some positive constants and such that Proof of Lemma 2. Let
be a solution of problem (
1), then by monotonicity of
we have
By the definition of
and (
33) we have
This implies that
for all
and therefore the sequence
is bounded and
exists. Let
be a positive constant such that
since
F is Lipschitzian continuous, we have
It follows from (
26), that
and
It further follows from (
15) that
is bounded. By Lipschitzian continuity of
there exists
such that
Since
is bounded, it follows from the definition
that
is also bounded. By Lipschitzian continuity of
there exists some constant
for which
Since the stepsize
in Step 4 of Algorithm 1 satisfies
then from (1), we have
Combining with (
34) gives
By (
36) and (
37), we have
This together with the definition of
in Step 5 of Algorithm 1 yields
By the property of projection (
27), we have
□
Theorem 1. Let be the sequence generated by Algorithm 1. Suppose that Assumption 1 holds, then the sequence converges to a point which satisfies
Proof of Theorem 1. Suppose on the contrary that (
42) does not hold, then there exists
for which
If
since Algorithm 1 uses a backtracking process to compute
starting from
then
does not satisfy (1), that is,
Consequently, we have from Remark 1 (i),
where
is bounded above by a positive constant
This means
Taking limit on both sides as
we have
This contradicts (
39). Hence (
42) must hold. Now, since
F is continuous and the sequence
is bounded, then there is some accumulation point of
say
for which
. By boundedness of
we can find subsequence
of
for which
From the proof of Lemma 2, we know that
exists. Therefore, we can conclude that
and the proof is complete. □
3. Numerical Results and Comparison
Attention is now turn to numerical experiments. The experiment is divided into two parts. The first experiment aims to explore the role of the parameter c in the definition of the line search (1). On the other hand, the second experiment discusses the computational advantage of the proposed method in comparison with two existing methods. The two existing methods are:
- (i)
Spectral gradient projection method for monotone nonlinear equations with convex constraints proposed by Yu et al. [
17].
- (ii)
Two spectral gradient projection methods for constrained equations and their linear convergence rate proposed by Liu and Duan [
25]. This method has two algorithms i.e., Algorithm 2.1 and Algorithm 2.2. We only compare our proposed method with Algorithm 2.1 since Algorithm 2.2 is similar with that Yu et al. [
17].
These two methods were chosen because their search directions are defined based on the BB parameters. For convenience, we respectively denote the two methods by SGPM and TSGP. Algorithm 1 TSSP is implemented using the following parameters
and
The parameters used for the SGPM and TSGP methods were taken respectively from References [
17] and [
25]. The metrics used for the comparison are: number of iterations (ITER), number of function evaluations (FVAL) and CPU time (TIME). In the course of the experiments, we solved six benchmark test problems using six (6) different starting points (see
Table 1) by varying the number of dimension. The test problems are denoted by
Since the proposed algorithm is derivative-free, the test problems include two nonsmooth problems. The three solvers were coded in MATLAB R2017a and run on a PC with intel Core(TM) i5-8250u processor with 4 GB of RAM and CPU 1.60 GHZ. The MATLAB code for the TSSP algorithm is available in
https://github.com/aliyumagsu/TSSP_Algorithm. The iteration process is terminated whenever the inequality
or
is satisfied and failure is declared whenever the number of iterations exceeds 1000 and the terminating criterion mentioned above has not been satisfied.
First experiment. This experiment discusses the role of the parameter
c in the definition of the line search (1) with regards to the performance of the TSSP algorithm. We solved all the test Problems 1–6 with dimension
using all the given initial guesses in
Table 1 by varying the values of
c. That is,
The comparison is based on ITER, FVAL and norm of the objective function, (NORM), where the experimental results are presented in
Table 2. CPU time results are omitted in
Table 2 because virtually all are less than 1 s. The results obtained reveal that the parameter
c slightly affected the performance of TSSP algorithm when solving Problems 2 and 6. For problem 2, Algorithm 1 TSSP recorded least ITER and FVAL when
and 5 while different ITER and FVAL values recorded for different values of
c may be associated with the random starting points chosen independently by MATLAB. However, extensive numerical experiment is needed to investigate the role of the parameter
c in the performance of the TSSP algorithm.
Second experiment. This experiment presents the computational advantage of the proposed method in comparison with the two existing methods mentioned above based on ITER, FVAL and TIME. All the test problems
were solved using the starting points in
Table 1 with three (3) different dimensions
50,000 and 100,000. In this experiment, we take
The results obtained by each solver are reported in
Table 3 and
Table 4. The NORM results presented in
Table 3 and
Table 4 show that each solver successfully obtained solutions of all the test Problems 1–6. However, it is clear that the TSSP algorithm obtained the solutions of virtually all the test problems with least ITER, FVAL and TIME. These information are summarized in
Figure 1,
Figure 2 and
Figure 3 based on the Dolan and Mor
performance profile [
26]. This performance profile tells the percentage win by each solver. In all the experiments, we see from
Figure 1,
Figure 2 and
Figure 3 that the proposed TSSP algorithm performs better with higher percentage win based on ITER, FVAL and TIME for solving all the test problems. In fact, the TSSP algorithm recorded 100 percent least FVAL for all the test problems.
We use the following test problems where
Problem 2 ([
28])
.where . Problem 5 ([
29])
.where and . Problem 6 ([
31])
.where . 4. Applications in Image Deblurring
In this section, we apply the proposed Algorithm 1 to solve problems arising from compressive sensing, particularly image deblurring. Consider the following least square problem with
regularization term
where
is the underlying images,
is the observed images,
(
), linear operator, is an
blurring matrix, and the parameter
Problem (
46) is of great importance because it appears in many areas of applications arising from compressive sensing. Recently, problem (
46) has been investigated by many researchers and different kinds of iterative algorithms have been proposed in the literature [
3,
32,
33,
34,
35]. Many algorithms for solving (
46) fall into two categories namely: algorithms that required differentiability assumption and algorithms that are derivative free. Since
norm is a nonsmooth function, algorithms that require the assumption of differentiabilty are not suitable for problem (
46) in its original form. Consequently, either
is approximated with some smooth function or problem (
46) is reformulated into an equivalent problem. For instance, Figueiredo et al. [
3] translate problem (
46) into convex quadratic program as follows. For any
we can find some vectors, say
such that
where
,
for all
Thus, we can write
, where
is an
dimensional vector with all elements one. Therefore, we can rewrite problem (
46) as
Furthermore, if we let
then from Reference [
3], we can write (
47) as the following
where
,
,
. It is not difficult to see that
G is a positive semi-definite matrix.
In Reference [
36], the resulted constrained minimization problem (
48) is further translated into the following linear variable inequality problem
Since the feasible region of
q is
problem (
49) is equivalent to the following linear complementary problem
We can see that the point
is a solution of the above linear complementary problem (
50) if and only if it satisfies the following system of nonlinear equation
The mapping
F is a vector-valued and the “min” operator denotes the componentwise minimum of two vectors. Interestingly, Lemma 3 of Reference [
37] and Lemma 2.2 of Reference [
36] show that the mapping
F satisfies Assumption 1 (ii) i.e., is Lipschitzian continuity and monotonicity. Therefore our proposed TSSP algorithm can be applied to solve it.
Image Deblurring Experiment
We tested the performance of the two-step TSSP algorithm in restoring some blurred images in comparison with the one-step spectral gradient method for
problems in compressed sensing (SGCS) [
36]. The images used for the experiment are the well-known gray test images namely: Lena, House, Pepper, Camera man and Barbara where the size of each image is given in
Table 5. The following metrics are used to assess the performance and quality of restoration by each algorithm tested: number of iteration (ITER), CPU time in seconds (TIME), signal-to-noise-ratio (SNR) which is defined as
and the structural similarity (SSIM) index that measure the similarity between the original image and the restored image [
38] for each of the two experiments. The MATLAB implementation of the SSIM index can be obtained at
http://www.cns.nyu.edu/~lcv/ssim/. To achieve fairness in comparison, each code was run from same initial point
and terminate when
where
is the merit function evaluation at
, with
. The parameters used for both TSSP and SGCS in this experiment come from Reference [
36] except for
in the line search (1) and
The original, blurred and restored images by each algorithm are given in
Figure 4. The two tested algorithms restored the blurred images successfully with different speed and level of quality. The results of the restoration by each algorithm are reported in
Table 5. We see from
Table 5 that TSSP restored all the five images with less ITER. Taking TIME into consideration, we see that though the SGCS is faster in restoring two of the images (i.e., Camera man and Barbara), TSSP is faster in restoring the remaining three images (i.e., Lena, House and Pepper). In addition, the SNR and SSIM values recorded by each algorithm revealed that TSSP restored the five blurred images with slightly better quality than SGCS except for Camera man. Taking everything together, this experiment shows that the two-step TSSP can deal with the
regularization problems effectively and can be a favourable alternative for image reconstruction.