1. Introduction
To maintain the requirements of the expected scale of the increasing data traffic and mobile terminals, the fifth generation (5G) wireless network [
1] faces some challenges in terms of system capacity, energy consumption, and so on. The cloud radio access network (C-RAN) [
2,
3], which has emerged as a promising solution in reducing both the capital and operating expenditures, is expected to be an effective approach to fulfil these requirements. In C-RAN, a central unit (CU) or baseband unit (BBU) pool connects all the deployed low-power base stations (BSs) using the finite-capacity fronthaul links that allows joint signal processing and transmission. Despite various attractive advantages brought by C-RANs [
4,
5,
6], such as joint beamforming, and centralized encoding and decoding, the performance bottleneck for large-scale implementation comes with the high capacity requirements of the fronthaul links. Therefore, in practical C-RAN, the data sharing [
7,
8,
9,
10,
11,
12,
13] and fronthaul compression designs [
14,
15,
16,
17,
18,
19,
20,
21,
22] are recognized as two promising approaches to overcome the significant impact of the constrained fronthaul on spectral efficiency (SE) and energy efficiency (EE) (bit-per-Joule) [
23,
24,
25]. The data sharing strategy [
8,
9,
10,
11,
12,
13] reduces the fronthaul consumption through limiting the data transfer among BSs (one BS serves a small number of the total MUs). For the latter one, the CU computes the precoded signals intended to be transmitted to each BS, and then the signals are quantized and sent to the BSs through the capacity limited fronthaul links. The compression process is usually modeled as a test channel [
14,
15,
16,
17,
18,
19,
20,
21,
22], where the uncompressed and compressed signals respectively represent the input and output, and the quantization noise are modeled as an independent additive Gaussian random variable [
13,
20,
21]. The authors in [
13] proved that the fronthaul compression achieves a better performance than the data sharing strategy, so that this paper considers the fronthaul compression.
In general, the fronthaul compression has been extensively studied in both the uplink [
16,
17,
18] and downlink C-RANs [
13,
19,
20,
21,
22]. The majority research of the existing literature on fronthaul compression in C-RAN is focused on the design of beamforming and quantization noises to maximize the achievable sum rate from the information-theoretic perspective [
11,
13,
16,
17,
18,
19,
20,
21,
26]. A joint adaptive decompression and detection algorithm was proposed in [
16] to improve the information-theoretic capacity for the uplink C-RAN. In [
18], the authors developed a distributed compression for the sum rate maximization (SEmax) problem in the uplink C-RAN accounting for both perfect and imperfect channel state information (CSI). For the downlink C-RAN, a joint precoding and multivariate compression scheme has been studied in [
19], where an iterative majorization-minimization approach was proved to achieve a stationary point solution of the SEmax problem. In [
13], a comparison between the data sharing and fronthaul compression strategy was investigated for the power minimization problem with finite-capacity fronthaul links in downlink C-RAN. Then, a hybrid compression and data sharing strategy was designed in [
11] for optimizing the achievable sum rate. However, maximizing the EE in accordance with fronthaul compression [
22] is technically far more challenging compared to SRmax, and the closed-form solution remains unexplored.
This paper investigates the EEmax problem in downlink C-RANs with the consideration of the fronthaul compression. This optimization problem is formulated as a non-convex fractional programming problem with respective to the beamforming and quantization noises under a limited power budget and capacity-finite fronthaul links, which is NP-hard and is difficult to solve. To deal with the fraction objective function of the original problem, some work has been done to firstly transform the fractional objective function into an equivalent subtractive-form optimization problem via exploiting the fractional programming [
27,
28,
29,
30]. Although the power allocation schemes for orthogonal frequency division multiple access (OFDMA) system [
28,
29] were obtained using the dual decomposition approach, such approaches cannot be applied to solve our problem due to the joint consideration of fronthaul compression and beamforming in our problem. Moreover, the Lagrangian based decomposition algorithm [
30] for the multicell system is still not applicable for decentralized implementation because the considered problem is more complex and the beamformings among BSs are coupled. The first-order Taylor expansion is adopted to linearize the non-convex data rate and fronthaul capacity constraint [
20], however, the optimal solution cannot be obtained. By exploiting the relationship between the achievable data rate and the mean square error (MSE) in [
31], Ref. [
11,
12] compared the sum rate performance of the data sharing and compression strategies, but they are limited to the SRmax problem. The authors in [
22] considered a joint design of beamforming, multivariate compression and BS-MU link selection to maximize the EE. By using the epigraph form of the original EEmax problem, the authors [
22] proposed a difference of convex (DC) function based algorithm. However, the optimal solution of [
22] is still unknown, and the computational complexity is typically high since a series of semi-definite programming (SDP) or second-order cone programming (SOCP) problems are solved. More importantly, since the beamforming and quantization noises are computed centrally at the CU, the algorithms proposed in the literature [
11,
12,
13,
19,
20,
22] can be computation intensive for large-scale C-RANs. Therefore, in this paper, we first propose an optimal algorithm for the EEmax problem to provide a performance benchmark, and then develop a low-complexity decentralized algorithm with closed-form solution. For practical usage in downlink C-RANs, the impact of CSI errors on the EE performance is further investigated. This paper not only offers an optimal solution for the EEmax problem with limited fronthaul capacity, but also lays a sound foundation for decentralized implementation with closed-form solutions in this field. Thus, the methods derived in this paper are significant in advancing this field.
The main contributions of this paper are summarized as follows.
Firstly, we formulate the EEmax problem of joint beamforming and quantization noises design under a limited power budget and capacity-limited fronthaul links as a non-convex fractional programming problem. We first derive an optimal solution method based on branch-and-bound (BnB) technique [
32,
33] to solve the EEmax problem globally. Specifically, the BnB algorithm computes the upper and lower bounds, and deletes the regions that do not contain the optimal solution. The algorithm terminates when the difference between the upper and lower bounds is smaller than a predefined accuracy.
Secondly, to reduce the computational complexity of the optimal algorithm and facilitate decentralized implementation, we propose to transform the problem into a parametric subtractive form, and further proposed a two-layer decentralized (TLD) algorithm to solve the equivalent subtractive problem. Specifically, an one-dimension search approach is used to find the EE in the outer layer, and a decentralized algorithm based on alternating direction method of multipliers (ADMM) is proposed to solve a subproblem in the inner layer. The proposed algorithm achieves closed-form solution in parallel manner with convergence guaranteed.
Thirdly, considering the imperfection of the obtained CSI in practical C-RANs, the robust optimal and TLD algorithms for the considered EEmax problem are also proposed to characterize the performance degradation of the CSI errors. In particular, the robust optimal can also achieve a performance benchmark, and the robust TLD algorithm also has closed-form solution in a parallel manner.
Finally, we validate the effectiveness of the proposed algorithms through extensive simulations. The results demonstrate that both the optimal and TLD algorithms are convergent, and the TLD algorithm can achieve near optimal solution which is much higher than the SEmax one. Numerical analysis also show that the EE performance is susceptible to the channel errors, and a smaller channel error reaches a higher EE.
The remainder of the paper is organized as follows. In
Section 2, the system model and problem formulation are presented.
Section 3 describes the proposed optimal algorithm with perfect CSI.
Section 4 presents TLD algorithm with perfect CSI. For imperfect CSI case, the robust optimal and TLD algorithms are also presented in
Section 5. The simulation results are given in
Section 6. Finally, we conclude this paper in
Section 7.
Notations: We use to denote the set of complex numbers, and to denote the set of all matrices with complex entries. We use boldface capital and lower case letters are respectively used to denote matrices and vectors. , and represent the matrix inverse, Hermitian transport and the trace, respectively. represents the Euclidean norm. is the expectation operator, and represents a diagonal matrix with diagonal elements given by . For a complex number x, is the mode of x. “”stands for “subject to”.
3. Optimal Algorithm with Perfect CSI
In this section, we will propose a global optimal algorithm, which based on Branch-and-Bound method [
32], to solve problem
. The essential idea of the proposed algorithm is based on the following equivalent transformation
where
are the introduced variables, (8b) is the transformation of
. The equivalence between problems
and
is that the constraints (8b) and (8c) hold with equality at optimum.
Although
is more tractable compared to
, it is still hard to solve due to the coupled variables of
and
. It is observed that if one increase each
in the feasible set of
, a better objective value can be obtained. This motivates us to use the monotonic optimization in [
32], i.e., the optimal BnB algorithm, to solve problem
.
BnB Algorithm
To solve , we first denote the feasible set for variables by , i.e., . Denote and by the aggregated lower and upper bound of . The interval indicates that each element of is bounded by its lower and upper bounds. The objective function in is monotonically increasing in the interval . In particular, is upper bounded by ignoring the interferences, i.e., , and the lower bound of is for . Similarly, we can constrain by , where and . It is obvious that the feasible set in must be contained by .
For a given
, problem
reduces to the feasibility problem given by
Obviously, when , problem is infeasible. That is because the sum rate should not equal to zero with maximum transmit power. In this paper, we will customize the BnB algorithm to solve problem globally. The BnB algorithm divides the box into smaller ones, and cuts off boxes that do not contain an optimal solution. The algorithm will converge to the global optimal solution after finite iterations.
Since
is non-convex due to the SINR constraint, in the following, we first recast them as convex ones. Let
, (8b) is equivalently rewritten as [
7,
33]
In the above formulation, we note that (10a) is a second-order cone (SOC) constraint. The constraint (10b) is without loss of generality due to the fact that a phase rotation of the beamformers does not effect the objective of the problem [
25,
33].
Moreover, (8c) is easily rewritten as
Then, problem
becomes a SOCP feasibility problem which can solved efficiently. In the proposed BnB algorithm, the bounding function can be formally expressed as
where
and
are the upper and lower bound respectively,
is defined as
where
and
denote the end points of the
kth edge of
,
and
.
We denote by the collection of all created boxes at iteration i. Then, the work flow of the BnB algorithm to obtain the global optimal solution is presented in Algorithm 1.
Algorithm 1 Proposed Optimal Algorithm. |
0: Initialization: given tolerance . Set , , , . |
1: Check the feasibility problem with given . If it is infeasible, exit; Otherwise go to step 2. |
2: repeat |
3: Set where satisfies . |
4: Branch into two smaller boxes and using the bisection subdivision along the longest edge of . |
5: Let . |
6: Update . |
7: Delete boxes that do not contain an optimal solution where . |
8: Update . |
9: Set . |
10: Until . |
Remark 1 (H)
. According to [34], the convergence of Algorithm 1 is guaranteed due to the monotonic property of . The main step of Algorithm 1 is to delete the boxes that do not contain the optimal solution. This step is referred to as pruning, and a smaller box that contains the optimal solution is obtained. Therefore, step 7 confirms the convergence of Algorithm 1. The corresponding optimal EE is , the optimal achieved data rate is , and the network power consumption is . This algorithm gives an optimal solution to problem (equivalently to problem ) when the tolerance is small enough. Algorithm 1 provides a performance benchmark for any other suboptimal algorithms. However, the computational complexity of Algorithm 1 is very high in general. Therefore, an improved box reduction approach approach was proposed in [33,34] to reduce the searching time, but we use the basic BnB approach in this paper for simplicity. 4. Decentralized Algorithm with Perfect CSI
In this section, we first transform the original problem into an equivalent subtractive-form using the Dinkelbach’s method. By exploiting the equivalence between the achievable data rate and its MSE, an ADMM algorithm is proposed to solve a QCQP subproblem with closed-form solution in a parallel manner.
4.1. Equivalent Optimization Problem
It is noted that
is a nonlinear fractional programming problem and can be transformed using the Dinkelbach’s method [
27]. Defining the optimal EE of problem
by
, we have
where
is the optimal total power consumption,
is the optimum of
R, and
is the sum rate.
According to [
29], the optimal EE
can be achieved if and only if
where
and
is the feasible region of problem
.
Thus, based on the theoretical results in [
25,
30,
33], problem
is transformed as the following parametric programming problem
It is noted that if problem is optimally solved with , problem can be solved optimally. However, if problem cannot be optimally solved, we can still solve problem through solving a sequence of problem . Unfortunately, the optimal solution of problem is not guaranteed in this case. We will provide detailed analysis in the next subsection.
The function of
is a monotonically decreasing function over
. Therefore, a bisection method, which is demonstrated in Algorithm 2, should perform well enough to find
[
25,
30].
Algorithm 2 Outer Layer Algorithm. |
1: Initialize the minimum and maximum as and respectively, and a small threshold value . |
2: repeat |
3: Set , solve problem with . |
4: If , . Otherwise, . |
5: Until or the maximum iteration number is reached. |
It is important to initialize in reducing the search time of Algorithm 2. Here, we initialize the interval that is bounded by its lower and upper bounds. Intuitively, is lower bounded by when the the sum rate equals to zero. For , it is upper bounded by ignoring the interference and using maximum transmit power in , and ignoring the transmit power and the quantization noises in . Specifically, , and . Therefore, , and .
4.2. Decentralized Algorithm for Subproblem
The key step for finding the quantized noises and the beamformings in Algorithm 1 lies in solving the subproblem
. The main difficulty arises from the non-convex
in the objective function (16a). Fortunately, by extending the equivalence between the SRmax problem and MMSE problem [
31,
35],
in problem
can be reformulated into a tractable form.
where
is a scalar variable associated with MU
k,
is the MSE for MU
k, given by
The proof of the equality in (17) is based on the first-order optimality condition [
31], which is omitted here for brevity. Then, problem
can be recast as
where
and
.
It is worth noting that problem
is not jointly convex in
, but it is convex with respect to
or
by given
or
, respectively. Thus, with fixed
and
, the optimal weight
is
where
is the optimal MSE for MU
k. Then, the optimal receive beamforming coefficient
under fixed
,
and
is a MMSE receiver [
31]
With fixed
and
, the optimal
and
can be obtained by solving the following quadratic constraint quadratic programming (QCQP) problem.
where
is a quadratic objective function, and
and
, and
. It is observed that this problem is a convex optimization problem with respect to
and
, which can be solved centrally by standard mathematical tools, i.e., CVX [
36]. It is noted that by simply replacing the constraints of the SINR and the maximum transmit power per BS as in [
20], such an alternative optimization can also be adopted to solve
with multivariate compression. That because the replacement does not affect the convexity of the subproblems. However, such an interior-point method solves
with high computational complexity and it does not reveal the structure of the solution. Meanwhile, it is implementation intensive for large-scale C-RANs due to the centralized computation of the beamformings and quantization noises at the CU. Unlike the beamforming design problem in multicell system [
30], the beamformings are coupled among BSs in our problem, making the Lagrangian based decomposition algorithm invalid in solving
. Towards this end, we propose a novel approach using ADMM method to solve
with closed-form solution optimally in a parallel manner.
In particular, in
, the two constraints (7b) and (7c), respectively, provide an upper and lower bound on
. Then, the constraint
should be satisfied. By rearranging this constraint, we have
Since the objective function of
is monotonically decreasing over
, we can replace the inequality constraint (7c) with equality, i.e.,
. We denote a new beamforming vector for BS
l as
, and then we have
. As a result, problem
is equivalent to the following problem in only a single set of variables
.
where
.
The objective function in (23a) contains two parts, and they are functions of different variables, i.e.,
and
, rather than the same variable. Therefore, problem
is not a standard group lest absolute shrinkage and selection operator (LASSO) problem [
37]. Hence, the existing algorithms for the group LASSO problems are not directly applicable. This fact motivates us to find new algorithm to solve problem
. Fortunately, it is observed that problem
has a special structure that can be solved by developing the famous ADMM algorithm. To account for the difference between
and
in problem
, we first introduce a copy
for
, and define
. Problem
can be equivalently expressed as
The partial augmented Lagrangian function of problem
is
where
with
is the vector of Lagrangian dual variables for the equality constraint (24c), and
is some constant.
The idea of the ADMM is to update the local variables when fixing the other variables. Specifically, the variables updating procedure of the ADMM algorithm is detailed as follows.
By fixing
and
at the
-th iteration,
at the
-th iteration is updated by solving the following convex problem
We show that problem (26) can be solved in a parallel manner. Specifically, we first decompose problem (26) into
L subproblems
that can be solved independently at the CU.
The Karush-Kuhn-Tucker (KKT) conditions of problem (27) are
where
is the optimal Lagrangian multiplier associated with the power constraint,
is the optimum of
, and
is defined as
.
If
, we have
under the condition of
(equivalent to
). If
,
. Then,
when
. Specially, when
,
. Therefore,
is updated as
To update
with fixed
, the following problem is solved
Due to the relationship between
and
, we have
. Problem (32) is equivalent to
Then, problem (33) can be decomposed into the following
K subproblems, and can be solved in a parallel manner with each MU.
By differentiating (34) with respect to
and set to zero, we obtain the optimal
with closed-form expression in the
-th iteration, given by
Using the relationship between
and
,
is obtained. Then, with obtained
and
, the multipliers
are updated
Therefore, the decentralized algorithm for solving is summarized in Algorithm 3. The convergence of Algorithm 3 is guaranteed by Theorem 1.
Algorithm 3 Decentralized Algorithm for Subproblem . |
1: Initialization: choose initial value for and , set iteration index , and choose the initial value for and . |
2: Repeat |
3: Update by (20), and obtain . |
4: Calculate according to (18) with . |
5: Update with , and obtain . |
6: Let iteration index and . |
7: while do |
8: Update by (31); |
9: Update by (35); |
10: Update the multipliers by (36). |
11: . |
12: end while |
13: . |
14: Update . |
15: Update by with . |
16: Until convergence or the maximum iteration number is reached. |
Theorem 1. Algorithm 3 generates a sequence that converge to a stationary point of problem .
Proof. The proof is based on the convergence of the alternative optimization method and ADMM algorithm. With initialized
, the inner loop of Algorithm 3 from steps 7 to 12 converges to an optimal solution of
due to the convergence of ADMM algorithm, and the proof can be found in [
38]. On the other hand, the outer loop of Algorithm 3 converges to a stationary point of subproblem
due to the convergence property of block coordinate decent algorithm [
31,
35]. According to [
25,
30], for an arbitrary
, the objective (16a) in problem
is shown to be non-decreasing during each iteration of the outer loop of Algorithm 3. Therefore, Algorithm 3 is guaranteed to converge to a stationary point of problem
, and the proof is completed. ☐
Denote by the actually optimal solution of problem , and the obtained suboptimal solution returned by the Algorithm 2 when using Algorithm 3 to solve problem . Since Algorithm 3 converges to a stationary point of problem , the suboptimal objective of equals to zero which equivalently equals to . Moreover, considering the fact that is monotonically decreasing, the actually optimal solution that satisfies must be no smaller than , i.e., . Therefore, the obtained solution returned by Algorithm 2 when using Algorithm 3 is no larger than the optimal solution of problem . Simulation results will demonstrate that the obtained solution is very close to the optimal one that verifies the effectiveness of the proposed algorithm numerically.
Combining Algorithms 2 and 3, it is concluded that problem can be solved efficiently by the proposed TLD algorithm. Since the limits of the upper and lower bounds are updated iteratively, the bisection procedure will stop any way. However, we may not have .
4.2.1. Parallelized Implementation
Since problem (26) is decomposed into L subproblems, at step 8 in Algorithm 3 is updated in a parallel manner with closed-form solution. Similarly, the beamformings and multipliers are also updated using (35) and (36) in a parallel manner with closed-form solution. Therefore, we derive closed-form expressions for the optimal beamformings, the optimal receiver filters and the auxiliary variables in Algorithm 3, that provide some insights on the EEmax problem. For example, if (no transmission power consumed by BS l), BS l can be switched off to save static power and improve EE.
4.2.2. Complexity Analysis
According to the algorithm flow, the computational complexity of Algorithm 3 contains two parts. In Algorithms 3, the computational complexity of computing (20) arises from the matrix inversion of the receive beamforming, i.e.,
. When the interior-point method is adopted to solve
, the computational complexity is
[
39]. In this case, to serve
K MUs, the overall computational complexity is in the order of
. On the other hand, since the ADMM algorithm is applied to solve
, the main computational complexity is the matrix inversion in step 9 of Algorithm 3 where the transmit beamformings are computed by (35). Thus, the overall computational complexity of Algorithm 3 is in the order of
. For the outer layer algorithm, simulation results will show that it converges rapidly (about 5 iterations). Therefore, we can deduce that the proposed TLD algorithm is computation efficient.
4.2.3. Generalization to the Multi-Antenna System
Although a single-antenna is equipped at each BS and each MU in the above discussion, we claim that proposed TLD algorithm can be generalized to the multi-antennas BSs and the single-antenna MUs scenarios. This is because deployed multiple antennas (each with
N antennas) at BSs, one only needs to replace the corresponding channel coefficient
from BS
l to MU
k with
. Similarly, the transmit beamforming coefficient
and receive beamforming coefficient
from BS
l to MU
k are replaced with vectors
and
, respectively. While
is replaced by
, and the corresponding covariance matrice is
. If
, one only need to replace
in (7b) and (7c) with
. Therefore, the proposed TLD algorithm can be extended to solve the EEmax problem with multi-antenna BSs and single-antenna MUs but requires additional efforts. Moreover, for the multi-antenna BSs and multi-antenna MUs C-RAN, we point out that the weighted minimum MSE algorithm in [
31] might provide some insights on how to apply the proposed algorithm.
5. Robust Algorithms with Imperfect CSI
In practical C-RAN, due to the limited feedback [
40], partial CSI [
41] or estimated error, the obtained channel is not perfect. According to [
42,
43,
44], the imperfection in CSI has significantly impact on the system performance. Therefore, we will extend the proposed algorithms in the previous two sections to solve the robust EEmax problem in the presence of imperfect CSI.
The same system model is considered as in
Section 2. Since the path-loss fading and the log-normal fading can be estimated accurately, the imperfection usually comes from the uncertain small-scale fading. Thus, different from the worst-case design [
42,
43], the Gaussian distribution in [
44] is adopted to model the channel imperfection. The real channel from BS
l to MU
k is expressed as
where
is the channel gain consisting of the antenna gain
G, the path-loss fading
at distance
in km and the log-normal random fading
.
and
are respectively the estimated channel and channel uncertainty from BS
l to MU
k.
is assumed to be independent identically distributed (i.i.d) zero mean circularly symmetrical complex Gaussian (ZMCSCG) random variables with variance
. The channel from all the BSs to MU
k is
where
, and
and
are the aggregated collection of
and
, respectively. Let
,
and
. The received SINR at MU
k is expressed as
where
denotes the interference caused by uncertainty part of CSI. The first term of
contains the interference caused by the CSI error of the intended signal and the signals of other MUs.
The robust EEmax problem has the same form as problem by simply using the imperfect channel, i.e., replacing by (28). The details of the extended algorithms to solve the robust problem with the consideration of imperfect CSI are given in the following two subsections, respectively.
5.1. Robust Optimal Algorithm
To apply the optimal algorithm under imperfect CSI, we only need to reformulate (8b) because the CSI is only involved in (8b). In particular, (8b) is recast as
Follow the same procedure as demonstrated in Algorithm 1, one only need to replace constraint (10a) in problem with (39a). In this case, the computational complexity is the same as Algorithm 1 (both of them are very high but provide optimal solutions).
5.2. Robust TLD Algorithm
Since the CSI is involved only in
, the TLD algorithm performs well to solve the robust EEmax problem but requires some transformations. In particular, the out layer iteration of the robust TLD algorithm follows the procedure as Algorithm 2. To solve the robust subproblem
in the inner layer, Algorithm 3 cannot be applied directly. Fortunately, with the same transformations as
Section 4.2, the corresponding MMSE receiver can be expressed as
and the MSE of the
k-th MU is
Then, with fixed
and
, the robust
with respect to
and
becomes
We further reformulate problem
as
where
,
, and
Since problems
and
have the same form, the ADMM algorithm presented in the inner layer of Algorithm 3 performs well in solving
. Particularly,
is updated by (31), and
is updated using the following closed-form expression
where
and
are shown in (44) and (45), respectively.
The corresponding multipliers are updated by (36). Similar to the TLD algorithm, the robust TLD algorithm procedure is omitted here for brevity. It is easily verified that the robust TLD algorithm is convergent, and it has the same computational complexity as the non-robust TLD algorithm, and it achieves a suboptimal solution with closed-form expression in a parallel manner as well.
6. Simulation Results and Discussions
In this section, we evaluate the performance of the proposed algorithms via Monte-Carlo simulation. We consider a downlink C-RAN with
single-antenna BSs and
single-antenna MUs, where one BS locates at the circle centre, and the other four BSs are located in a circle region at equal distances apart with radius
km, as shown in
Figure 1. The four single-antenna MUs are randomly deployed in the circle with uniform distribution. Each BS and MU are equipped with a single antenna, and we set the maximum transmit power
and the fronthaul capacity
for all BSs. The convergence errors are set as
for all the proposed algorithms. Unless specified, other simulation parameters are listed in
Table 1, and all the simulation results are averaged over 50 times independent MU locations, each with a single random channel realization. In order to verify the effectiveness of the proposed algorithms, we consider two baseline algorithms for comparison:
SRmax algorithm: The proposed TLD algorithm is adopted to solve the SRmax problem by simply setting .
DC algorithm: We also modify the algorithm proposed in [
22] for comparison. Specifically, to arrive at a tractable formulation, one can use the epigraph form of the original problem. Based on the transformation, one can transform the constraints into convex ones by using the first-order Taylor expansion. Then, the approximated problem is iteratively solved, and the solution converges to a KKT point [
22]. The computational complexity of the DC algorithm using the interior-point method is
(
is the maximum of iteration number), which is much higher than the TLD algorithm.
6.1. Non-Robust Performance
We first investigate the convergence behavior of the proposed algorithms over a typical random channel realization for
= 2 W,
= 30 dBm, and
C = 5 bits/s/Hz, which are presented in
Figure 2.
It is observed that the outer layer (Algorithm 2) of TLD in
Figure 2a converges to a near optimum very fast (about 5 iterations). Meanwhile,
Figure 2b illustrates that the objective function of problem (16) of Algorithm 3 at the 2nd iteration of Algorithm 2 in
Figure 2a increases and converges to a stationary point in less than 30 iterations. Thus, the proposed TLD algorithm converges rapidly. It is also found that the convergence rate of the optimal algorithm is much slower that the TLD algorithm. Due to a large number of infeasible boxes being removed, the gap between upper and lower bounds of the optimal algorithm reduces rapidly during the first iterations.
Then, we explore the effect of fronthaul capacity on the performance in terms of EE, sum rate and transmit power, shown in
Figure 3,
Figure 4 and
Figure 5, respectively.
The TLD algorithm achieves a comparable performance compared to the DC one in terms of EE.
Figure 3 shows the EE comparison among the four algorithms under different fronthaul capacities. As an be seen from
Figure 3 that the TLD and DC algorithms achieve an approximate EE performance as the optimal one, and both of them outperform the SRmax one by about three times in the middle-high fronthaul capacity region. Since the network power is penalized in (7a), and the sum rate and power are jointly optimized, the EE is improved compared to SRmax. This can be explained from
Figure 4 and
Figure 5 that the saved ratio of the transmit power of TLD (save for about 95% compared with SRmax) is much higher than its sum rate reduction ratio (decrease by about 50% compared with SRmax), resulting in a higher EE. It is also observed from
Figure 3 that the EE performance increases with the growing fronthaul capacity in the low fronthaul capacity region, and then gradually converge to a constant in the middle to high fronthaul capacity region, and a similar trend can be found in
Figure 4. According to the expression of quantization noises (QN), i.e.,
, it reduces exponentially with the increasing fronthaul capacity (shown in
Figure 5). As a result, at low fronthaul capacity region, the SINR is limited by the high QN, leading to a low sum rate in this region. Moreover, the corresponding data transmit power (DTP) in
Figure 5 increases very slowly from low to high fronthaul capacity region. That because with the increase of the fronthaul capacity, more powers of DTP and QN are consumed to improve the sum rate. However, more network power is consumed which makes the increment of EE gradually. It should be noted that the optimal algorithm cannot provide DTP and QN but only the total network power, as a result only the transmit power (sum of DTP and QN) of the optimal algorithm is given in
Figure 5. In summary,
Figure 3,
Figure 4 and
Figure 5 illustrate that the TLD and optimal algorithms can achieve the balance between the sum rate and power consumption, implying a much higher EE than the SRmax one.
Figure 6 compares the EE performance with respect to the maximum transmit power
for a given fronthaul capacity
bits/s/Hz. It is observed that TLD and DC significantly outperform the SRmax in terms of EE in the middle to high transmit power region (≥18 dBm). This is because by jointly optimizing the sum rate and network power, the increased sum rate is slightly larger than the increased power consumption ratio, resulting in a gradually increase of EE. While for the SRmax algorithm, because of the interference between MUs, its sum rate gain cannot compensate for the negative impact of the network power, making its EE worse than the TLD and DC algorithms in the event that
≥ 18 dBm. It is also observed that the optimal, and TLD and DC algorithms have comparable EE performance as the SRmax one at low maximum transmit power (≤18 dBm), which suggests that at this region, transmitting with the maximum transmit power provides comparable sum rate gain for the optimal, EEmax and SRmax algorithms.
To investigate further, we also test the impact of static power consumption on the EE performance of the proposed algorithms, shown in
Figure 7. It depicts that, at low to middle static power (
W), the EE of the optimal, DC and TLD algorithms decreases significantly with the growing static power
. This can be explained that the static power dominates the network power at this region and the transmit power is optimized as well, leading to a rapid decrease of EE. Whereas, since the maximum transmit power is usually adopted by the SRmax algorithm for transmission and it accounts for a large amount of the network power for
ranging from 2 to 10 W, the decrease EE of the SRmax algorithm becomes gentle. These facts bring a higher EE performance for the TLD algorithm than the SRmax at different static powers.
The effect of MU number on the EE performance is also investigated with
bits/s/Hz,
dBm and
W, which is shown in
Figure 8. We learn from
Figure 8 that, with more MUs to be served, the EE of the TLD and DC algorithms increase significantly and exhibits obvious advantage over the SRmax one, e.g., higher than SRmax by more than 300%. Intuitively, the data rate of each MU decreases with the growing number of served MUs due to the increased interference between MUs. The obtained sum rate gain is very large when the serving MU number is small, while the sum rate increases much more slowly if the served MU number is large. It is attributed to the rapidly increase of interference among MUs by serving large number of MUs, implying a data rate reduction for each MU and a slightly increase on the sum rate of all MUs. Besides, more transmit power is needed to achieve the sum rate gain when serving more MUs, which also perform negative for boosting EE performance. It should also be noted that by further increasing the number of MUs, the EE of all the four algorithms will firstly become stable and then show a decrease trend. This is because, due to the increased interference caused by increasing the MU number, the decreased sum data rate of MUs is much larger than the increased data rate brought by the newly added MUs.
6.2. Robust Performance
The EE performance of the proposed robust TLD algorithm is investigated over the channels with different channel errors, and the results are shown in
Figure 9. Similar to the non-robust design, each point in
Figure 9 is averaged over 100 channel realizations. In the simulation,
means perfect CSI is known at the CU. It was observed in the previous subsection that the SRmax algorithm has a worse EE performance than that of the TLD (TLD has near optimal performance), and thus we only plot the EE curve of robust TLD algorithm under different CSI errors. Due to the high complexity imposed by the DC algorithm, we do not compare it with the TLD in this section.
We can see from
Figure 9 that by increasing the channel errors from
to
, the EE performance decreases significantly especially in the middle to high fronthaul capacity. In the low fronthaul capacity region, according to (38), the received SINR is slightly influenced by the channel errors since the increased interference is very small. While a worse sum rate is obtained in the middle to high fronthaul capacity due to the impairment of channel error on the received SINR, and more amount of transmit power is consumed in order to achieve the same sum rate for a larger channel errors. As a result, the sum rate gain cannot compensate the network power reduction that caused by the channel error, resulting in lower EE. In summary, the EE performance is susceptible to the channel error especially for a larger channel error.