Next Article in Journal
PDMAC: A Priority-Based Enhanced TDMA Protocol for Warning Message Dissemination in VANETs
Previous Article in Journal
Convolutional-Neural Network-Based Image Crowd Counting: Review, Categorization, Analysis, and Performance Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning (RL)-Based Energy Efficient Resource Allocation for Energy Harvesting-Powered Wireless Body Area Network

1
College of Information Science and Technology, Nanjing Forestry University, Nanjing 210037, China
2
School of Electrical Engineering and Telecommunications, University of New South Wales, Sydney 2052, Australia
3
School of Information Science and Technology, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(1), 44; https://doi.org/10.3390/s20010044
Submission received: 20 November 2019 / Revised: 16 December 2019 / Accepted: 17 December 2019 / Published: 19 December 2019
(This article belongs to the Section Sensor Networks)

Abstract

:
Wireless body area networks (WBANs) have attracted great attention from both industry and academia as a promising technology for continuous monitoring of physiological signals of the human body. As the sensors in WBANs are typically battery-driven and inconvenient to recharge, an energy efficient resource allocation scheme is essential to prolong the lifetime of the networks, while guaranteeing the rigid requirements of quality of service (QoS) of the WBANs in nature. As a possible alternative solution to address the energy efficiency problem, energy harvesting (EH) technology with the capability of harvesting energy from ambient sources can potentially reduce the dependence on the battery supply. Consequently, in this paper, we investigate the resource allocation problem for EH-powered WBANs (EH-WBANs). Our goal is to maximize the energy efficiency of the EH-WBANs with the joint consideration of transmission mode, relay selection, allocated time slot, transmission power, and the energy constraint of each sensor. In view of the characteristic of the EH-WBANs, we formulate the energy efficiency problem as a discrete-time and finite-state Markov decision process (DFMDP), in which allocation strategy decisions are made by a hub that does not have complete and global network information. Owing to the complexity of the problem, we propose a modified Q-learning (QL) algorithm to obtain the optimal allocation strategy. The numerical results validate the effectiveness of the proposed scheme as well as the low computation complexity of the proposed modified Q-learning (QL) algorithm.

1. Introduction

Recent advances in sensors and wireless communication technology have resulted in a promising development of wireless body area networks (WBANs) [1]. The specific application scenario of WBANs is to continuously monitor the vital physiological signals of the human body and transmit the real-time sensory data to the users and doctors without any interruptions in their daily lifestyle to realize smart healthcare in the framework of smart cities [2,3]. Different from the conventional complex and wired healthcare devices, WBANs typically consist of a number of battery-driven, invasive, and/or non-invasive body sensors and one hub (mobile phone or personal digital assistant (PDA)) with the communication function in the form of wireless. The hub normally has rich resources, such as energy supply, processing capability, and buffer storage. In contrast, body sensors are energy-limited owing to the small size. Meanwhile, it is difficult or inconvenient to replace the battery as these body sensors may be implanted in the human body. Therefore, designing an energy efficient resource allocation scheme has great significance to WBANs. A comprehensive survey on the major characteristics, research issues, and challenges in WBANs for patient monitoring from a practical design and implementation perspective was provided in the works of [4,5]. Additionally, several previous works in the literature were proposed to investigate the energy-saving technologies from the aspects of the media access control (MAC) protocol design, power control, and cross-layer resource allocation strategies to make efforts to prolong the lifetime of WBANs [6,7,8,9,10,11,12]. In the work of [7], the authors presented a time division multiple access (TDMA)-based technique to improve WBANs’ reliability and energy efficiency by adaptively synchronizing nodes while tackling channel and buffer status. The simulation results show that the method can improve packet loss and energy consumption. In the work of [9], a reinforcement learning-based sensor access control scheme in WBANs was investigated. The authors considered the signal-to-interference plus noise ratio, transmission priority, battery level, and transmission delay to make a decision on the access time and transmit power. In the work of [10], a multi-hop routing protocol and routing decision strategies in WBANs for medical application were proposed. The authors formulated the optimization problem into Markov decision process (MDP) subjects to various conditions such as battery level, event occurrence, packet transmission rate, and link quality. However, these schemes cannot ensure the WBANs achieve the ultimate goal of ‘uninterrupted work’. Once these batteries are drained, the sensors will be dead. In addition to further improving energy efficiency, energy harvesting is an appealing solution [13,14,15]. Energy harvesting is a technology that enables devices to collect energy from ambient sources [16]. Various types of energy sources can be exploited as energy supplies, for instance, thermal, locomotion, and electromagnetic wave [17,18]. Meanwhile, in the existing literature, several energy harvesting models have also been investigated in the literature [19,20,21]. Therefore, energy harvesting (EH)-powered WBANs (EH-WBANs) have the potential ability to achieve the infinite lifetime and perpetual operation, which is called energy neutral operation (ENO) [22]. However, owing to the fluctuation of ambient energy source and the immaturity of energy conversion technology, the available energy of each body sensor will become a vital factor in the design of resource allocation schemes in EH-WBANs. A comprehensive survey on energy scavengers and their potential utilization in WBANs was given in the work of [23]. In another paper [24], a wireless charged wearable network for monitoring physiological signals of patients was investigated. The authors proposed to cluster the wearable devices to improve the reliability and lifetime of the network. In the work of [25], two scenarios of a point-to-point communication system in WBANs were studied, in which two protocols, called the power splitting protocol and time switching protocol, were proposed to maximize the network throughput.
To the best of our knowledge, the research on resource allocation for EH-WBANs is in its infancy, despite having some pioneering studies [26,27,28,29,30,31,32,33,34,35]. Mohammadi et al. [26] proposed a link adaption mechanism to maximize energy efficiency in Institute of Electrical and Electronics Engineers (IEEE) 802.15.6 impulse radio ultra-wideband (IR-UWB) WBANs. In the work of He et al. [27], the transmission power and source rate of sensors are jointly optimized to provide quality of service (QoS) requirements for WBANs. However, the authors only considered single hop transmission between the sensor and coordinator. Liu et al. [28] proposed a transmission rate allocation scheme to efficiently adjust the transmission rate at each sensor to guarantee the packet loss ratio requirement. Jung et al. [29] proposed a novel contention-based MAC protocol in WBANs; the performance is evaluated by formulating the problem to be a discrete-time Markov chain model. Despite that the aforementioned studies made contributions to the development of WBANs, the lifetime of WBANs is still limited by the battery longevity. As an emergency solution to break the battery limitation in wireless networks, EH technology has attracted great attention from both industry and academia. In the work of Qiu et al. [30], a transmission power control scheme was proposed to improve the lifetime of the wind-powered wireless sensor networks (WSNs) by jointly considering the residual energy level and the amount of energy harvested. An optimal energy management strategy for solar-powered WSNs was proposed in the work of Niyato et al. [31]; the authors concentrated on the sleep and wakeup scheduling for energy conservation. However, different from EH-powered WSNs, EH-WBANs mainly harvest energy from human body bio-energy sources [32]. These human body bio-energy sources can be categorized into bio-chemical and bio-mechanical energy sources. The bio-chemical energy sources convert electrochemical to electricity for invasive body sensors, while bio-mechanical energy can be obtained from the locomotion of the human body [33]. In the work of Quwaider et al. [34], the weighted sum of the outage probabilities was the objective function to be minimized. The harvested energy was known as a priori to the scheduler, and an optimal offline algorithm was proposed to get the optimal solution. A joint power–QoS control scheme was proposed in the work of Ibarra et al. [35], namely powered by energy harvesting–quality of service (PEH–QoS). The PEH–QoS scheme combines three interconnected modules: the power–EH aware management module, the data queue aware control module, and the packet aggregator system. The core idea of PEH–QoS is to use the amount of power available for transmission and the amount of data stored in the queue to determine the maximum number of packets that can be transmitted in each data communication process. The simulation results show that the energy efficiency of the body sensor can be improved.
Looking at these previous works, the energy efficiency issue is rarely considered; even if some works tried to investigate EH-WBANs, they covered only limited aspects such as the sum-rate, transmission power, and the tradeoff between different objectives. In order to fill this gap, we investigate the energy efficient resource allocation scheme in EH-WBANs with the goal of maximizing the energy efficiency in this paper. This is also the motivation behind this work.
The main contributions of this paper are as follows:
  • We consider a resource allocation problem for EH-WBANs with the goal of maximizing the average energy efficiency of body sensors. The resource allocation problem jointly considers the transmission mode, relay selection, allocated time slots, transmission power, and energy status to make the optimal allocation decision;
  • We formulate the energy efficiency problem to be a discrete-time and finite-state Markov decision process (DFMDP) and a modified Q-learning algorithm, which reduces the state-action space in the original Q-learning algorithm, is proposed to solve the modeled problem;
  • From the numerical analysis, we show that the proposed scheme can obtain the best energy efficiency and with the more rapid convergence speed by eliminating the irrelevant exploration space in the Q-table as compared with the classical Q-learning algorithm.
The remainder of this paper is organized as follows. In Section 2, the network model considered in this paper is presented. After that, the corresponding energy efficiency maximization problem is formulated and the proposed modified Q-learning algorithm is elaborated in Section 3. The simulation results are discussed in Section 4. Finally, we conclude the paper in Section 5.

2. Network Model Descriptions

In this section, we first depict the system model of the EH-WBAN, which is then followed by the details on the data transmission model, energy harvesting model, and energy efficiency model in EH-WBANs. Table 1 summarizes the different symbols and notations used throughout this paper.

2.1. Network Model

In this treatise, we consider a single EH-WBAN with one hub and multiple EH-powered body sensors, as illustrated in Figure 1. The hub is placed on the belt, while various body sensors, for example, the electrocardiogram sensor (ECG), the electromyography sensor (EMG), the electroencephalography sensor (EEG), the glucose sensor, and motion sensor, are placed in different positions of the body according to different detection purposes. For simplicity, we only consider the uplink transmission from body sensors to the hub and only body sensors are equipped with the EH function. Here, we denote the hub as H and body sensors as S n ,   n 1 ,   2 ,   ,   N . In this work, we assume that both direct transmission and cooperative transmission modes are supported by the network layer, as recommended by IEEE 802.15.6 standard [36]. In cooperative transmission mode, only two-hop transmission is permitted. We define a binary parameter α S n 0 ,   1   n 1 ,   2 ,   ,   N to indicate the transmission mode that is utilized recently by the n-th body sensor. α S n = 1 denotes that the n-th body sensor transmits data to the hub directly, while α S n = 0 indicates that the n-th body sensor is in cooperative transmission mode. In the MAC layer, TDM technology is employed to prevent the mutual interference. As the slotted system model is adopted, each transmission frame can be divided into K number of time slots and the time slot set is denoted as ψ = 1 ,   2 ,   ,   K . We set t 0 = 0 and t K = T . The duration of each slot is denoted as τ k = t k t k 1   k ψ . In this model, these time slots can be assigned to the body sensors, whether they operate in direct transmission or cooperative transmission modes. Meanwhile, it should be noted that the transmission mode of each body sensor is determined by the resource allocation strategy in the proposed scheme. For example, after the proposed scheme finds the optimal resource allocation strategy, in which a specific body sensor is determined to transmit data in cooperative mode, then the body sensor will use the certain time slot (which is also determined by the resource allocation strategy) to transmit data in cooperative mode.
In case of direct transmission mode, we define a binary parameter β S n k 0 ,   1 ,   n 1 ,   2 ,   ,   N ,   k ψ to indicate which time slot is assigned to a specific body sensor. β S n k = 1 denotes that the k-th time slot is assigned to the n-th body sensor for direct transmission, while β S n k = 0 means the k-th time slot is not assigned to the n-th body sensor for direct transmission. More specifically, another two reasonable assumptions are made in this model: (1) the hub can only receive data from one body sensor at each time slot; (2) in each time frame, each body sensor only be assigned at most to one time slot for direct transmission. Thus, we can derive two constraints as Equations (1) and (2):
n = 1 N β S n k 1 ,   k ψ ,
k = 1 K β S n k 1 ,   n 1 ,   2 ,   ,   N .
In the case of cooperative transmission mode, we assume that the K time slots in a time frame are allocated to both source-relay and relay-hub links. This assumption is mainly to be used to guarantee the fairness between direct transmission and cooperative transmission, to obtain the optimal resource allocation strategy. Similarly, we define a parameter δ S n S m k 0 ,   1 ,   n , m 1 ,   2 ,   ,   N ,   k ψ as an indicator that the k-th time slot is allocated to n-th body sensor for transmitting data to the m-th body sensor, which is selected as the relay node of the n-th body sensor. Meanwhile, δ S m H k 0 ,   1 n , m 1 ,   2 ,   ,   N ,   k ψ is denoted as the indicator that the m-th body sensor forwards the data from the n-th body sensor to the hub using the k-th time slot. In this mode, we assume that each source sensor can select one relay sensor and each relay sensor can only forward data from one source sensor at any time slot. Thus, we can obtain two constraints as Equations (3) and (4):
m = 1 , m n N δ S n S m k 1 ,         n = 1 , n m N δ S n S m k 1 ,
n = 1 , n m N δ S m H k 1 ,           m = 1 , m n N δ S m H k 1 .
Furthermore, because each link can only be assigned at most to one time slot, we can obtain the constraint as Equation (5):
k = 1 K δ S n S m k ,   k = 1 K δ S m H k 1     n m .
Another aspect to note is that the data transmission from the source sensor to relay sensor should be prior to the transmission from the relay sensor to hub. Therefore, we can obtain Equation (6):
k = 1 x δ S n S m k k = x + 1 K δ S m H k 0 ,   x 1 ,   2 ,   ,   K 1 .

2.2. Data Transmission Model

In WBANs, different body sensors with different monitoring purposes may have the heterogeneity of data rate requirement. In the proposed data transmission model, we suppose that all the body sensors can be served as the source node in direct transmission mode or relay node in cooperative mode. The selection of mode is determined by the resource allocation scheme. We denote R n as the data rate of the n-th body sensor, and it can be expressed as Equation (7):
R n = α S n · R n d + 1 α S n · R n c ,   n 1 ,   2 ,   ,   N ,
where R n d is the data rate of the n-th body sensor in direct transmission mode, and R n c denotes the data rate of the n-th body sensor in cooperative transmission mode. On the basis of the above analysis, we can derive the instantaneous signal to interference plus noise ratio (SINR) of direct transmission and cooperative transmission. Equations (8)–(10) give the instantaneous SINR of direct link, source-relay link, and relay-hub link in the k-th time slot, respectively.
S I N R n , k d = p n , k d · g S n H n 1 = 1 , n 1 n N m = 1 ,   m     n , n 1 N δ S n 1 S m k · p n 1 , m , k s · g S n 1 S m + n 0 ,
where p n , k d denotes the instantaneous transmission power of the n-th body sensor in the k-th time slot when transmitting data to the hub; g S n H is the transmission gain between the n-th body sensor and the hub; p n 1 , m , k s denotes the instantaneous transmission power of n1-th body sensor in the k-th time slot when transmitting data to m-th body sensor, which is selected as the relay sensor of the n1-th body sensor; g S n 1 S m is the transmission gain between the n1-th body sensor and m-th body sensor; and n 0 is the noise power.
S I N R n , m , k s r = p n , m , k s r · g S n S m I n , m , k s r + n 0 ,
I n , m , k s r = n 1 = 1 n 1 n , m N m 1 = 1 m 1 n , n 1 , m N δ S n 1 S m 1 k · p n 1 , m 1 , k s r · g S n 1 S m , + n 1 = 1 n 1 n , m N β S n 1 k · p n 1 , k d · g S n 1 S m , + n 1 = 1 n 1 n , m N m 1 = 1 m 1 n , n 1 , m N δ S m 1 H k · p n 1 , m 1 , k r H · g S m 1 S m ,
where p n , m , k s r denotes the instantaneous transmission power of the n-th body sensor when transmitting data to the m-th body sensor, which is selected as its relay sensor in the k-th time slot; p n , m , k r H denotes the instantaneous transmission power of m-th body sensor in the k-th time slot when forwarding data from n-th body sensor to the hub; and I n , m , k s r is the total instantaneous interference of the source-relay link in the k-th time slot. The expression of I n , m , k s r includes three items; the first item indicates the interference from other source-relay links, the second item is the interference from direct transmission between the source body sensor and hub, and the three item represents the interference from relay-hub links.
S I N R n , m , k r H = p n , m , k r H · g S m H , I n , m , k r H + n 0 ,
I n , m , k r H = n 1 = 1 n 1 m N m 1 = 1 m 1 m , n 1 N δ S n 1 S m 1 k · p n 1 , m 1 , k s r · g S n 1 S m ,
where I n , m , k r H is the total instantaneous interference between the relay sensor and hub when the m-th body sensor is selected as the relay of the n-th body sensor in the k-th time slot.
Additionally, channel fading between body sensors and the hub is affected by many factors such as clothing and obstructions due to different body segments [37], thus the dynamic link characteristics should be taken into full consideration. In this paper, the channel fading takes into account both large-scale fading and small-scale fading. The channel gain g can be represented as Equation (11):
g = α · h ,
where α denotes the large-scale fading, which includes path loss and shadowing. It can be modeled as α = P L · β d φ . P L is the path loss constant; β denotes the log-normal shadowing random component; d is the distance between transmitter and receiver in a communication link; φ is the power decay exponent; and h is the small-scale fading, which is assumed as Rayleigh small-scale fading with unit mean.
According to Shannon’s theorem, we can obtain the transmission rate of direct mode as given in Equation (12):
R n d = k = 1 K β S n k · B · l o g 2 1 + S I N R n , k d ,
when B denotes the bandwidth of the channel.
The transmission rate of the cooperative mode R n c can be divided into two parts: one is the transmission rate of source-relay link R n c ,   s r and another is the transmission rate of relay-hub link R n c ,   r H , as shown in Equations (13) and (14):
R n c ,     s r = m = 1 m n N k = 1 K δ S n S m k · B · l o g 2 1 + S I N R n , m , k s r ,
R n c ,   r H = m = 1 m n N k = 1 K δ S m H k · B · l o g 2 1 + S I N R n , m , k r H .
However, in cooperative transmission mode, the transmission rate of the path between the source body sensor and hub is limited by the smaller transmission rate of the source-relay link and relay-hub link. Hence, the transmission rate of the cooperative mode is R n c = m i n   R n c ,   s r ,   R n c ,   r H .

2.3. Data Serving Model

In this scenario, we make the assumption that the data are stored in the form of packets in the buffer of the device. The arrived data at each body sensor follow an independently and identically distributed (i.i.d.) sequence with an average rate of λ d [38]. Practically, we assume that the buffer of the device is finite and served in first in first out fashion. We denoted D Q S n k as the instantaneous data queue length at the n-th body sensor in time slot k. The maximum traffic queue length of body sensors is represented by   D Q S n m a x . Accordingly, we can obtain the update function of the instantaneous data queue length as Equation (15):
D Q S n k = m i n   D Q S n m a x ,   D Q S n k 1 m i n α S n · R n d + 1 α S n · R n c P S d a t a ,   D Q S n k 1 + A S n k 1
where P S d a t a is the traffic packet size, α S n · R n d + 1 α S n · R n c P S t r a f f i c is the instantaneous service rate of transmission link of n-th body sensor in k − 1-th time slot, and A S n k 1 is the arriving traffic packets of the n-th body sensor in the k − 1-th time slot.

2.4. Energy Harvesting Model

We denoted E n ,   k as the energy harvested by the n-th body sensor in the k-th time slot. { E n ,   1 ,   E n ,   2 ,   ,   E n ,   t ,   ,   E n ,   K } is the time sequence of energy harvested in a transmission frame. It is also i.i.d. sequence with average rate of λ e [38]. We denote E Q S n k as the instantaneous energy queue length of the n-th body sensor in the k-th time slot. The maximum energy queue length of body sensors is represented by E Q S n m a x . Therefore, we can obtain the update function of the instantaneous energy queue length as Equation (16):
E Q S n k = m i n E Q S n m a x ,   Q S n k 1 m i n p n , k 1 P S e n e r g y ,   E Q S n k 1 + E n ,   k 1 ,
where P S e n e r g y is the energy packet size with the unit of Joules/packet. p n , k 1 denotes the transmission power of the body sensor in the k − 1-th time slot. According on the transmission mode, p n , k 1 can be set to one of   p n , k 1 d , p n , m , k 1 s r , and p n , m , k 1 r H .
It is worth noting that, because the capacity of the energy storage device is finite, two constraints can be derived from Equation (15), as expressed in Equations (17) and (18):
k = 1 K p n , k 1 P S e n e r g y k = 1 K E Q S n k ,   K 1 , 2 , ,
k = 1 K E Q S n k k = 1 K p n , k 1 P S e n e r g y E Q S n m a x ,   K 1 , 2 , .
Equation (16) depicts that the current available energy cannot exceed the total energy in the battery. Equation (17) expresses that the total energy stored in the battery cannot exceed the maximum battery capacity.

2.5. Energy Efficiency Model

In this paper, we define the energy efficiency (EE) of WBANs as the ratio of the transmission rate to the consumed transmission power. Equation (19) gives the energy efficiency of the n-th body sensor in the k-th time slot.
E E S n k = α S n · R n d + 1 α S n · R n c p n , k     n 1 ,   2 ,   ,   N ,   k ψ
Therefore, the average energy efficiency of the overall WBANs is presented as follows:
E E = 1 N · k = 1 K N = 1 N     E E S n k .
The corresponding EE optimization problem can be formulated as follows:
m a x i m i z e α S n ,   β S n k , δ S n k ,   p n , k ,                       E E ,
subject to:
n = 1 N β S n k 1 ,   k   ψ , k = 1 K β S n k 1 ,   n 1 ,   2 ,   ,   N , m = 1 , m     n N δ S n S m k 1 ,   n = 1 , n m N δ S n S m k 1 , n = 1 , n m N δ S m H k 1 , m = 1 , m n N δ S m H k 1 , k = 1 K δ S n S m k , k = 1 K δ S m H k 1   n m , k = 1 x δ S n S m k k = x + 1 K δ S m H k 0 ,   x 1 ,   2 ,   ,   K 1 , k = 1 K p n , k 1 P S e n e r g y k = 1 K E Q S n k ,   K 1 , 2 , , k = 1 K E Q S n k k = 1 K p n , k 1 P S e n e r g y   E Q S n m a x ,   K 1 , 2 , , p n , k d p n m a x   n 1 ,   2 ,   ,   N ,   k ψ , p n , m , k s r p n m a x   n , m 1 ,   2 ,   ,   N ,   n m , k ψ , p n , m , k r H p n m a x n , m 1 ,   2 ,   ,   N ,   n m , k ψ .

3. Problem Formulation and Optimization Algorithm

From the energy efficiency maximization problem, we can see that it is a long-term multi-objective optimization problem. Simultaneously, because the variables p n , k are continuous, while α S n ,   β S n k ,   and   δ S n k are binary, problem (21) is a mixed integer nonlinear programming problem, which cannot be directly solved by convex optimization methods. Even if we can transform the original problem into a tractable convex optimization problem, the problem still requires the prior network information such as channel state information (CSI) to achieve optimal performance. However, WBANs normally work in a dynamic channel characteristic owing to the posture and environment variation [39,40]. Furthermore, from Equation (15), we found that the current consumed energy packets are only related to current arrivals and the previous remainders in the energy queue. Thus, we can formulate problem (21) as the discrete-time and finite-state Markov decision process (DFMDP) [41]. More specifically, in this work, we formulate our scenario into a centralized DFMDP. Therefore, the hub should acquire all information about both the network and users to make the optimal decision. The reasons for formulating the centralized DFMDP are as follows: (1) the hub has more abundant resources compared with the body sensor; (2) the centralized DFMDP will reduce the network signaling overhead and redundancy as compared with distributed DFMDP. In distributed DFMDP, each body sensor should make the decision without the complete knowledge and global network information that will increase the computation complexity and consume more energy. Meanwhile, owing to the high computation complexity, we propose to utilize a modified Q-learning algorithm to solve the optimization problem.

3.1. DFMDP Model

DFMDP is a discrete time stochastic control process that provides a mathematical framework for modeling decision-making problems in uncertain and stochastic environments [42]. Typically, a DFMDP is defined by a tuple (S, A, p, r), where S is a finite set of states, A is a finite set of actions, p is a transition probability from state s to state s’ ( s S , s S ) after action a a A is performed, and r is the immediate reward obtained after a   a A is executed. We denote π as a policy that is a mapping from a state to an action. Our goal is to find the optimal policy denoted as π to maximize the reward function over a finite time horizon in the DFMDP. Therefore, the detailed tuple in our proposed model is designed as follows:
  • The state of each body sensor S n in the k-th time slot can be denoted as S t a S n k S . In this model, S t a S n k contains two parts: D Q S n k and E Q S n k . They are the data and energy queue lengths of the n-th body sensor at the beginning of the k-th time slot, respectively. To ensure the completeness of the exploration of state space, D Q S n k and E Q S n k are specified to be an integer and take the values of 0 ,   1 ,   ,   D Q S n m a x and 0 ,   1 ,   ,   E Q S n m a x , respectively.
  • The action a   a A in this scenario should be the resource allocation variables, which include transmission mode α S n , time slot allocation β S n k , relay selection δ S n k , and power allocation p n , k . To make sure the integrity of the exploration of action space, p n , k d , p n , m , k s r , and p n , m , k r H should be subject to the maximum transmission power p n m a x .
  • Obviously, the reward r is the immediate reward corresponding to current state–action pair, which is given by Equation (20).
However, the traditional value-based algorithms such as Monte Carlo [43] and temporal difference (TD) [44] algorithms have some shortcomings in practical applications, for instance, they cannot handle the tasks in continuous action space efficiently and the final solution may not be globally optimal. Therefore, we intend to adopt a policy-based algorithm in this paper.
In order to address the formulated DFMDP problem, the Q-learning algorithm is an effective tool [45]. The core idea behind the Q-learning algorithm is to first define the value function V π s d j k r that represents the expected value gotten by policy π from each state s d j k S . The value function V π for policy π quantifies the goodness of the policy via an infinite horizon and discounted MDP, which can be represented as Equation (22):
V π s = E π [ k = 0 γ · r k s k , a k | s 0 = s ] = E π r k s k , a k + γ · V π s k + 1 | s 0 = s .
Additionally, it should be noted that, to solve the MDP problem with discrete state and action spaces, the Q-learning algorithm is capable of obtaining the optimal policy [42]. Because we aim to find the optimal policy π , the optimal action at each state can be found by means of the optimal value function, as in Equation (23):
V s = m a x a k E π r k s k , a k + γ · V π s k + 1 .
If we denoted Q s , a r k s k , a k + γ · E π V π s k + 1 as the optimal Q-function for all state–action pairs, the optimal value function can be rewritten by V s = m a x a Q s , a . The Q s , a can be obtain through the iterative process according to the Equation (24):
Q k + 1 s k , a k = Q k + α r k s k , a k + γ m a x Q k s k , a k + 1 Q k s k , a k ,
where α is the learning rate to determine the impact of new information to the existing Q-value, and γ 0 , 1 is the discount factor. Algorithm 1 gives the pseudo-code of the Q-learning algorithm.
Algorithm 1 The Q-learning based resource allocation algorithm
  • initialize the table entry Q s , a arbitrarily for each state-action pair s , a
  • observe the current state s , initialize the value of α and γ
  • forepisode = 1 to Mdo
  •   from the current state-action pair s , a , execute action a and obtain
      the immediate reward r and a new state s
  •   select an action a based on the state s and update the table entry for
       Q s , a as expressed in Equation (18)
  •   replace s s
  • end for
  • Output: π s = a r g m a x a Q s , a
Initially, the resource allocation scheme randomly selects a transmission mode, relay node (if necessary), time slot, and transmission power without the consideration of the data and energy queue status in each body sensor to get an initial state–action pair. Meanwhile, the algorithm initializes the value of α and γ . After the first iteration, from current state–action pair, we can obtain an immediate reward and a new state (the data and energy queue lengths of each body sensor). Then, the algorithm updates the state–action pair, as expressed by Equation (19). Once either all Q-values or a certain number of iterations is reached, the algorithm will terminate. The optimal policy indicating an action (resource allocation scheme) to be taken at each state is maximized for all states. However, the convergence speed of the classical Q-learning algorithm may not be able to find the optimal policy within the acceptable time, especially in the practical and complicated model. Thus, one aspect that needs to be further considered is the necessity of cut down of the original state–action space of the Q-learning algorithm, as the convergence speed is sensitive to the size of the state–action space [46]. Therefore, we proposed a modified Q-learning algorithm, which aims to improve the convergence speed by cutting down the irrelevant state–action pairs.

3.2. The Proposed Modified Q-Learning Algorithm

In order to ensure the integrity of the exploration space, the values of state and action are set to an integer from 0 to the maximum value. However, this assumption will result in unnecessary exploring of state-action pairs in the original space. The proposed modified Q-learning algorithm intends to cut down these irrelevant pairs from both the irrelevant state and irrelevant action. In the irrelevant state aspect, we define the valid state space that should be explored to simultaneously achieve two requirements: available energy and serviceable data. Table 2 demonstrates the valid state space that should be explored. In Table 2, * represents the valid state and 0 indicates the invalid state. Nevertheless, in the view of the valid state, all the actions should be explored. Moreover, we investigate the irrelevant action in this work.
Similarly, in the irrelevant action aspect, according to exploration space integrity, the value of one action in the Q-table is as follows: p n , k ranges from 0 to p n m a x . However, the value of p n m a x may be larger than the current available energy p n , k . In this case, the range of action value can be cut down from 0 to p n , k . In this way, we can further reduce the state–action space.
To better illustrate the computation complexity, we compare the state–action space size of the proposed modified Q-learning based algorithm with the classical Q-learning algorithm. We assume that the maximum available energy and serviceable data in each sensor is x packets and the maximum power control parameter in each sensor is y packets. On the basis of the definition of the state–action place, Table 3 gives the computation complexity of the proposed modified Q-learning based algorithm and classical Q-learning algorithm.
From Table 3, we can see that the computation complexity of the classical Q-learning algorithm increases exponentially with the number of sensors. However, the state–action space size of the proposed modified Q-learning based algorithm is completely free from the influence of sensor numbers.
Meanwhile, the proposed modified Q-learning scheme also contributes to the balance of energy consumption among body sensors. This is because the proposed modified Q-learning scheme considered the amount of harvested energy of each body sensor while allocating resources. In such a situation, if a specific body sensor harvests less energy from the environment, it will not be selected as relay. Moreover, the standard deviation of the consumed energy is less, which indicates that the consumed energy is distributed among body sensors and the lifetime of the overall WBANs can be extended.

4. Simulation Results and Analysis

In this section, we compare the proposed algorithm with other three schemes: (1) a random power allocation scheme; (2) the classical Q-learning resource allocation scheme; and (3) a joint power–QoS control scheme proposed in the literature [35]. To verify the effectiveness of the proposed algorithm, we evaluate the performance in terms of energy efficiency and convergence speed.

4.1. Simulation Setting

In simulations, we consider a WBAN scenario in which multiple heterogeneous body sensors and one hub are deployed with different positions for various detection purposes. Five typical body sensors with their initial energy are considered. They are as follows: ECG with initial energy of 20 mJ, EMG with initial energy of 12 mJ, EEG with initial energy of 16 mJ, glucose sensor with initial energy of 12 mJ, and motion sensor with initial energy of 18 mJ. For simplicity, we assume that the current energy harvesting technology is able to provide the required conversion efficiency. The hub is placed at the center of this topology with a communication range of 10 m, and it knows all the position information of the body sensors. Each body sensor is randomly placed in the topology with the communication range of 2–5 m [47]. Simultaneously, we suppose that only body sensors are equipped with the energy harvesting function, and the energy harvesting process is Poisson-distributed with a rate λ e at arrival instants t k . The data arriving process is also Poisson-distributed with a rate λ d at arrival instants t k . Moreover, the proposed modified Q-learning algorithm has no prior knowledge about them. Meanwhile, because the scenario contains lots of instability, we set 200 time instants for one episode, and the energy efficiency will be averaged to reduce the instability. For each configuration, we generate 100 independent runs and average the performance of energy efficiency. All of the detailed simulation variables used in this paper are summarized in Table 4.

4.2. Results and Analysis

The influence of learning rate α and discount factor γ on energy efficiency
In order to avoid other factors influencing the performance, we first evaluate the influence of learning rate α and discount factor γ on energy efficiency. We implement two scenarios in which one body sensor in direct transmission mode and two body sensors in cooperative transmission mode are deployed, respectively. The energy harvesting rate λ e is set to 3 packet/s and the data arriving rate λ t is set to 5 packet/s. Figure 2 and Figure 3 show the average energy efficiency under different values of α and γ . From the results of both scenarios, we can see that either the decrease of learning rate α or the increase of discount factor γ will cause the instability of energy efficiency in the proposed resource allocation algorithm. These two cases are depicted as brown and green marks in Figure 2 and Figure 3. This is because a smaller α leads to less exploration; in this case, the proposed algorithm increasingly concentrates on the greedy action, which has a more immediate effect in increasing the users’ utility. Contrarily, a larger γ causes less foresight in the policy updating, which will reduce the average utility in the long term [48]. From the blue marks in Figure 2 and Figure 3, we can observe that, while the α is set to 0.9 and γ is set to 0.1, the average energy efficiency is more stable, which means that the algorithm has the highest convergence speed. Furthermore, we also tried some more complex scenarios in which more sensors are deployed, but the influences of learning rate α and discount factor γ are similar. For simplicity and ease of understanding, we only demonstrate this scenario and we can obtain a vivid result that the proposed algorithm performs better in the case of a higher α and lower γ . Consequently, we set α = 0.9 and γ = 0.1 , respectively, in the following simulations.

4.2.1. Comparison between the Proposed Algorithm and Classical Q-Learning Algorithm

Figure 4 illustrates the optimization processes for energy efficiency of the proposed modified Q-learning algorithm and classical Q-learning algorithm. The simulation result gives two observations. First, after 30 episodes, the proposed modified algorithm trends to convergence rather than 80 episodes of classical Q-learning algorithm. This is because the proposed modified algorithm eliminates the irrelevant state and action spaces that reduce the exploration space. Hence, the convergence speed is accelerated. Second, as the episodes increase, the performance of the classical Q-learning algorithm trends to stable. However, the proposed algorithm outperforms the classical Q-learning algorithm over approximately 20%. This is because of the lower computation complexity and signaling overhead in the proposed algorithm.

4.2.2. The Influence of the Number of Body Sensors Deployed

Figure 5 presents the average energy efficiency for a different number of body sensors deployed in WBAN. From the result, it can be observed that the proposed algorithm has the highest energy efficiency among the classical Q-learning algorithm, the PEH–QoS algorithm proposed in the literature [35], and the random power allocation algorithm. For the Q-learning based algorithm and the PEH–QoS algorithm, as the number increases, the average energy efficiency is increased. This is because of the fact that more body sensors will lead to a higher data rate. However, it also can be observed that, as the number increases to 9, the average energy efficiency tends to stable. This is because these three algorithms all take available energy into consideration when allocating transmission power. For the random power allocation algorithm, the energy efficiency goes up when the number of body sensors is less than 7. As more body sensors are deployed, the energy efficiency is reduced, while it reaches a saturation point for most algorithms (except the random power allocation algorithm). This is because more body sensors deployed will involve more mutual interference, which further increases the transmission power.

4.2.3. The Standard Deviation of Consumed Energy of Each Body Sensor

Figure 6 presents the standard deviation of consumed energy versus the different number of body sensors deployed. The proposed modified Q-learning scheme gives the best performance with an average standard deviation of 10.33, compared with 15.65 using the classical Q-learning scheme, 19.98 using the PEH–QoS scheme from the literature [35] and 26.57 using the random power allocation scheme. This is because of the fact that the proposed modified Q-learning scheme is able to maintain the fairness of each body sensor by means of selecting the optimal transmission mode and relay body sensors to distribute the energy consumption to the relay sensors with more residual energy. It can also be observed that, at the beginning, there is an increase in the standard deviation as the number of body sensors increases. This is because, at the initial time, the number of body sensors is small and the consumed energy in each body sensor is unbalanced. After a further increase in the number of body sensors, the standard deviation tends to decrease and becomes stable. Another interesting finding is that the standard deviation of energy consumed of the random power allocation scheme increased sharply with the number of body sensors. This is because this scheme allocates transmission power randomly and does not consider the available energy residual in each body sensor.

4.2.4. The Influence of Energy Harvesting Rate λ e and Data Arrival Rate λ d

Figure 7 and Figure 8 present the energy efficiency with different energy harvesting rates λ e . The data arrival rate λ d is set to 5 and 8, respectively. From the results in Figure 7 and Figure 8, it is clear that the proposed modified Q-learning algorithm and classical Q-learning algorithm can achieve higher energy efficiency as compared with the PEH–QoS scheme from the literature [35] and the random power allocation scheme. With the increase of λ e , the energy efficiency is improved sharply. This is because more energy can be harvested in each time slot with the higher λ e , and the Q-learning-based algorithms are able to obtain an optimal correlation between energy harvesting time, transmission mode, relay selection, and power allocation. The PEH–QoS scheme proposed in the literature [35] gives a slightly better performance than the random power allocation scheme. However, another interesting finding is that, when the λ e is less than 4, the random power allocation scheme has the best energy efficiency. This is because the random power allocation scheme does not take into account the available energy when allocating transmission power.
Figure 9 and Figure 10 plot the energy efficiency with the different data arrival rate λ d . In this scenario, the energy harvesting rate λ e is set to 3 and 5, respectively. As shown in Figure 9 and Figure 10, the proposed algorithm still obtains the highest energy efficiency in the comparison of the four algorithms along with the λ d . The reason is that more data arrival will increase the transmission rate; in the meantime, the proposed algorithm can acheive the optimal coupling relationship between the transmission rate and energy consumption, thus improving the performance of energy efficiency. From the results of Figure 9 and Figure 10, we also observe that, as λ d achieves 7, the energy efficiency trends to stable; this is because more data to be served requires more transmission power p n , k , but the λ e is constantly set to 3 and 5 in this simulation. Hence, the energy efficiency is going to stable. However, although the PEH–QoS algorithm from the literature [35] considered energy harvesting in resource allocation strategy, and the authors have also shown the performance with different transmission rates up to 1 Mbps, it did not take into consideration the data arrival rate of each body sensor. Thus, the results of the PEH–QoS algorithm in Figure 9 and Figure 10 are similar. Meanwhile, because the random power allocation scheme allocates transmission power randomly, the energy efficiency does not change with λ e and λ d .
In addition, because the WBANs concentrate mainly on the stringent reliability requirement for the safety-critical information. In this simulation, we also evaluate the reliability of the proposed scheme. The reliability is represented by the delivery probability. Specifically, we define the delivery probability as the probability of successfully delivering the sensory data of each body sensor with the size of P bits within an acceptable time T. Hence, the delivery probability can be given as P r b α S n · R n d + 1 α S n · R n c P T . Figure 11 gives the average delivery probability of the WBANs, while different numbers of body sensors are deployed. We set P for each body sensor with the constant size of 1 Mb. From the result, we can find that, as the number of body sensors increases, the average delivery probabilities decrease for all schemes. This is because more body sensors being deployed will cause more mutual interference. However, the proposed scheme still gives better performance compared with the other two schemes throughout the tested cases. Remarkably, the proposed scheme is capable of guaranteeing the average delivery probability above 90% even in the worst case. Moreover, in conjunction with the result from Figure 5, we can validate the practicability of the proposed scheme.

5. Conclusions

The main motivation of this paper is to study the resource allocation scheme for EH-WBANs. Unlike the traditional WBANs, the available energy will be another vital issue that should be considered in the resource allocation scheme. Specifically, with the goal of maximizing the average energy efficiency, we formulate the resource allocation problem to be a DFMDP, in which the transmission mode, relay selection, allocated time slot, power allocation, and energy constraint of each body sensor are considered. Owing to the high complexity of the problem, we solve the maximization problem using a modified Q-learning algorithm. Through extensive simulations, it is shown that the proposed scheme enhances the energy efficiency significantly for different network settings. Additionally, with the conjunction of transmission reliability, we validate the practicability of the proposed scheme in EH-WBANs.

Author Contributions

Conceptualization, Y.-H.X. and J.-W.X.; Methodology, Y.-H.X.; Software, J.-W.X.; Validation, Y.-G.Z., M.H. and W.Z.; Formal analysis, Y.-H.X.; Investigation, Y.-H.X.; Resources, W.Z.; Data curation, M.H.; Writing—original draft preparation, Y.-H.X.; Writing—review and editing, M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by China Scholarship Council, Introduction of high-level talents and overseas returnee’s scientific fund in Nanjing Forestry University (No. GXL015), and National natural science foundation of China under Grant of 61801225.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, M.; Gonzalez, S.; Vasilakos, A.; Cao, H.; Leung, V.C.M. Body area networks: A survey. Mob. Netw. Appl. 2011, 16, 171–193. [Google Scholar] [CrossRef]
  2. Marwa, S.; Ahmed, A.D.; Imed, R. Wireless Body Area Network (WBAN): A survey on reliability, fault tolerance, and technologies coexistence. ACM Comput. Surv. 2017, 50, 3. [Google Scholar]
  3. Dagdeviren, C.; Li, Z.; Wang, Z.L. Energy harvesting from the animal/human body for self-powered electronics. Annu. Rev. Biomed. Eng. 2017, 19, 85–108. [Google Scholar] [CrossRef]
  4. Chakraborty, C.; Gupta, B.; Ghosh, S.K. A review on telemedicine-based WBAN framework for patient monitoring. Telemed. J. E-Health 2013, 19, 619–626. [Google Scholar] [CrossRef] [PubMed]
  5. Elli, K.; Aris, S.L.; Angelos, A.; Stefano, T.; Marco, D.R.; Luis, A.; Christos, V. A survey on M2M systems for mHealth: A wireless communications perspective. Sensors 2014, 14, 18009–18052. [Google Scholar]
  6. Esteves, V.; Angelos, A.; Elli, K.; Manel, P.V.; Pere, M.C.; Christos, V. Cooperative energy-harvesting-adaptive MAC protocol for WBANs. Sensors 2015, 15, 12635–12650. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Salayma, M.; Ahmed, A.D.; Romdhani, I.; Nasser, Y. Reliability and energy efficiency enhancement for emergency-aware wireless body area networks (WBANs). IEEE Trans. Green Commun. Netw. 2018, 2, 804–816. [Google Scholar] [CrossRef]
  8. Liu, Z.; Liu, B.; Chen, C.W. Transmission-rate-adaption assisted energy-efficient resource allocation with QoS support in WBANS. IEEE Sens. J. 2017, 17, 5767–5780. [Google Scholar] [CrossRef]
  9. Chen, G.H.; Zhan, Y.J.; Sheng, G.Y.; Xiao, L.; Wang, Y.H. Reinforcement learning-based sensor access control for WBANs. IEEE Access 2018, 7, 8483–8494. [Google Scholar] [CrossRef]
  10. Roy, M.; Chowdhury, C.; Aslam, N. Designing transmission strategies for enhancing communications in medical IoT using Markov decision process. Sensors 2018, 18, 4450. [Google Scholar] [CrossRef] [Green Version]
  11. Ibarra, E.; Angelos, A.; Elli, K.; Christos, V. HEH-BMAC: Hybrid polling MAC protocol for wireless body area networks operated by human energy harvesting. Telecommun. Syst. 2015, 58, 111–124. [Google Scholar] [CrossRef] [Green Version]
  12. Min, M.H.; Wan, X.Y.; Xiao, L.; Chen, Y.; Xia, M.H.; Wu, D.; Dai, H.Y. Learning-based privacy-aware offloading for healthcare IoT with energy harvesting. IEEE Int. Things J. 2019, 6, 4307–4316. [Google Scholar] [CrossRef]
  13. Ahmed, I.; Ikhlef, A.; Schober, R.; Mallik, R.K. Power allocation for conventional and buffer-aided link adaptive relaying systems with energy harvesting nodes. IEEE Trans. Wirel. Commun. 2014, 13, 1182–1195. [Google Scholar] [CrossRef]
  14. Wang, H.Q.; Chi, X.F.; Zhao, L.L. Energy adaptive MAC protocol for IEEE 802.15.7 with energy harvesting. Optoelectr. Lett. 2016, 12, 370–374. [Google Scholar] [CrossRef]
  15. Harb, A. Energy harvesting: State-of-the-art. Renew. Energy 2011, 36, 2641–2654. [Google Scholar] [CrossRef]
  16. Paradiso, J.A.; Starner, T. Energy scavenging for mobile and wireless electronics. IEEE Pervasive Comput. 2005, 4, 18–27. [Google Scholar] [CrossRef]
  17. Ku, M.L.; Wei, L.; Yan, C.; Liu, K.J.R. Advances in energy harvesting communications: Past, present, and future challenges. IEEE Commun. Surv. Tutor. 2016, 18, 1384–1412. [Google Scholar] [CrossRef]
  18. Gao, H.H.; Ejaz, W.; Jo, M. Cooperative wireless energy harvesting and spectrum sharing in 5G networks. IEEE Access 2016, 4, 3647–3658. [Google Scholar] [CrossRef]
  19. Altinel, D.; Kurt, G.K. Modeling of hybrid energy harvesting communication systems. IEEE Trans. Green Commun. Netw. 2019, 3, 523–534. [Google Scholar] [CrossRef]
  20. Azmat, F.; Chen, Y.; Stocks, N. Predictive modelling of RF energy for wireless powered communications. IEEE Commun. Lett. 2015, 20, 173–176. [Google Scholar] [CrossRef]
  21. Fan, D.; Lopez Ruiz, L.; Gong, J.; Lach, J. Ehdc: An energy harvesting modeling and profiling platform for body sensor networks. IEEE J. Biomed. Health Inf. 2018, 22, 33–39. [Google Scholar] [CrossRef]
  22. Kansal, A.; Hsu, J.; Zahedi, S.; Srivastava, M.B. Power management in energy harvesting sensor networks. ACM Trans. Embed. Comput. Syst. 2007, 6, 32. [Google Scholar] [CrossRef]
  23. Demir, S.M.; Fadi, A.T.; Ali, M. Energy scavenging methods for WBAN applications: A review. IEEE Sens. J. 2018, 18, 6477–6488. [Google Scholar] [CrossRef]
  24. Mekikis, P.; Angelos, A.; Elli, K.; Nikos, P.; Luis, A.; Christos, V. Stochastic modeling of wireless charged wearables for reliable health monitoring in hospital environments. In Proceedings of the IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017. [Google Scholar]
  25. Ling, Z.; Hu, F.Y.; Wang, L.H.; Yu, J.C.; Liu, X.L. Point-to-point wireless information and power transfer in WBAN with energy harvesting. IEEE Access 2017, 5, 8620–8628. [Google Scholar] [CrossRef]
  26. Mohammadi, M.S.; Zhang, Q.; Dutkiewicz, E.; Huang, X. Optimal frame length to maximize energy efficiency in IEEE 802.15.6 uwb body area networks. IEEE Wirel. Commun. Lett. 2014, 3, 397–400. [Google Scholar] [CrossRef]
  27. He, Y.; Zhu, W.; Guan, L. Optimal resource allocation for pervasive health monitoring systems with body sensor networks. IEEE Trans. Mob. Comput. 2011, 10, 1558–1575. [Google Scholar] [CrossRef] [Green Version]
  28. Liu, Z.; Liu, B.; Chen, C.; Chen, C.W. Energy-efficient resource allocation with QoS support in wireless body area networks. In Proceedings of the IEEE Global Communications Conference, San Diego, CA, USA, 6–10 December 2015. [Google Scholar]
  29. Jung, B.H.; Akbar, R.U.; Sung, D.K. Throughput, energy consumption, and energy efficiency of IEEE 802.15.6 body area network (BAN) MAC protocol. In Proceedings of the IEEE International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Sydney, NSW, Australia, 9–12 September 2012. [Google Scholar]
  30. Qiu, J.; Lin, B.; Liu, P.; Zhang, S.; Dai, G. Energy level based transmission power control scheme for energy harvesting WSNs. In Proceedings of the IEEE Global Communications Conference, Houston, TX, USA, 5–9 December 2011. [Google Scholar]
  31. Niyato, D.; Hossain, E.; Rashid, M.M.; Bhargava, V.K. Wireless sensor networks with energy harvesting technologies: A game-theoretic approach to optimal energy management. IEEE Wirel. Commun. 2007, 14, 90–96. [Google Scholar] [CrossRef]
  32. Leng, S.; Yener, A. Resource allocation in body area networks for energy harvesting healthcare monitoring. In Handbook of Large-Scale Distributed Computing in Smart Healthcare; Springer: Cham, Switzerland, 2017; pp. 553–587. [Google Scholar]
  33. Akhtar, F.; Rehmani, M.H. Energy harvesting for self-sustainable wireless body area networks. IT Prof. 2017, 19, 32–40. [Google Scholar] [CrossRef]
  34. Wei, S.; Guan, W.; Liu, K.J.R. Power scheduling for energy harvesting wireless communications with battery capacity constrain. IEEE Trans. Wirel. Commun. 2015, 14, 4640–4653. [Google Scholar] [CrossRef]
  35. Ibarra, E.; Antonopoulos, A.; Kartsakli, E.; Rodrigues, J.J.P.C.; Verikoukis, C. Qos-aware energy management in body sensor nodes powered by human energy harvesting. IEEE Sens. J. 2015, 16, 542–549. [Google Scholar] [CrossRef] [Green Version]
  36. IEEE Standard for Local and Metropolitan Area Networks—Part 15.6: Wireless Body Area Networks; IEEE: New York, NY, USA, 2012.
  37. Quwaider, M.; Rao, J.; Biswas, S. Body-posture-based dynamic link power control in wearable sensor networks. IEEE Commun. Mag. 2010, 48, 134–142. [Google Scholar] [CrossRef]
  38. Mitran, P. On optimal online policies in energy harvesting systems for compound poisson energy arrivals. In Proceedings of the IEEE International Symposium on Information Theory, Cambridge, MA, USA, 1–6 July 2012. [Google Scholar]
  39. Reusens, E.; Joseph, W.; Braem, B.; Tanghe, E.; Martens, L.; Moerman, I.; Blondia, C. Characterization of on-body communication channel and energy efficient topology design for wireless body area networks. IEEE Trans. Inf. Technol. Biomed. 2009, 3, 933–945. [Google Scholar] [CrossRef] [PubMed]
  40. D’Errico, R.; Ouvry, L. A statistical model for on-body dynamic channels. Int. J. Wirel. Inf. Netw. 2010, 17, 92–104. [Google Scholar] [CrossRef]
  41. Baxter, L. Markov decision processes: Discrete stochastic dynamic programming. Technometrics 1995, 37, 353. [Google Scholar] [CrossRef]
  42. Luong, C.N.; Hoang, D.T.; Gong, S.; Niyato, D.; Wang, P.; Liang, C.Y.; Kim, D.I. Applications of deep reinforcement learning in communications and networking: A survey. IEEE Commun. Surv. Tutor. 2019, 21, 3133–3174. [Google Scholar] [CrossRef] [Green Version]
  43. Quah, H.K.; Quek, C. MCES: A novel Monte Carlo evaluative selection approach for objective feature selections. IEEE Trans. Neural Netw. 2007, 18, 431–448. [Google Scholar] [CrossRef]
  44. Caarls, W.; Schuitema, E. Parallel online temporal difference learning for motor control. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1457–1468. [Google Scholar] [CrossRef]
  45. Cai, X.; Zheng, J.; Zhang, Y. A Graph-coloring based resource allocation algorithm for D2D communication in cellular networks. In Proceedings of the IEEE International Conference on Communications (ICC), London, UK, 8–12 June 2015. [Google Scholar]
  46. Fedrizzi, R.; Goratti, L.; Sithamparanathan, K.; Rasheed, T. A Heuristic Approach to Mobility Robustness in 4G LTE Public Safety Networks. In Proceedings of the IEEE Wireless Communications and Networking Conference, Doha, Qatar, 3–6 April 2016. [Google Scholar]
  47. Sandhu, M.M.; Javaid, N.; Akbar, M.; Najeeb, F.; Qasim, U.; Khan, Z.A. FEEL: forwarding data energy efficiently with load balancing in wireless body area networks. In Proceedings of the IEEE International Conference on Advanced Information Networking and Applications, Victoria, BC, Canada, 13–16 May 2014. [Google Scholar]
  48. Zhang, Y.; Fu, F.; Van der Schaar, M. On-line learning and optimization for wireless video transmission. IEEE Trans. Signal Process. 2010, 58, 3108–3124. [Google Scholar] [CrossRef]
Figure 1. Scenario of the wireless body area network (WBAN). ECG: electrocardiogram sensor; EMG: electromyography sensor; EEG: electroencephalography sensor.
Figure 1. Scenario of the wireless body area network (WBAN). ECG: electrocardiogram sensor; EMG: electromyography sensor; EEG: electroencephalography sensor.
Sensors 20 00044 g001
Figure 2. Influence of α and γ on energy efficiency in direct transmission mode.
Figure 2. Influence of α and γ on energy efficiency in direct transmission mode.
Sensors 20 00044 g002
Figure 3. Influence of α and γ on energy efficiency in cooperative transmission mode.
Figure 3. Influence of α and γ on energy efficiency in cooperative transmission mode.
Sensors 20 00044 g003
Figure 4. The optimization process for energy efficiency.
Figure 4. The optimization process for energy efficiency.
Sensors 20 00044 g004
Figure 5. Energy efficiency versus different numbers of body sensor. PEH: powered by energy harvesting; QoS: quality of service.
Figure 5. Energy efficiency versus different numbers of body sensor. PEH: powered by energy harvesting; QoS: quality of service.
Sensors 20 00044 g005
Figure 6. Standard deviation of consumed energy versus different numbers of body sensors.
Figure 6. Standard deviation of consumed energy versus different numbers of body sensors.
Sensors 20 00044 g006
Figure 7. Energy efficiency versus energy harvesting rate λ e with constant λ d = 5 .
Figure 7. Energy efficiency versus energy harvesting rate λ e with constant λ d = 5 .
Sensors 20 00044 g007
Figure 8. Energy efficiency versus energy harvesting rate λ e with constant λ d = 8 .
Figure 8. Energy efficiency versus energy harvesting rate λ e with constant λ d = 8 .
Sensors 20 00044 g008
Figure 9. Energy efficiency versus traffic arrival rate λ t with constant λ d = 3 .
Figure 9. Energy efficiency versus traffic arrival rate λ t with constant λ d = 3 .
Sensors 20 00044 g009
Figure 10. Energy efficiency versus traffic arrival rate λ t with constant λ d = 5 .
Figure 10. Energy efficiency versus traffic arrival rate λ t with constant λ d = 5 .
Sensors 20 00044 g010
Figure 11. Average delivery probability of WBANs versus different numbers of body sensor.
Figure 11. Average delivery probability of WBANs versus different numbers of body sensor.
Sensors 20 00044 g011
Table 1. Table of notations. SINR: signal to interference plus noise ratio.
Table 1. Table of notations. SINR: signal to interference plus noise ratio.
SymbolDefinition
HHub
S n n-th body sensor n 1 ,   2 ,   ,   N
ψ Time slot ψ = 1 ,   2 ,   ,   K
τ k k-th time slot
α S n Transmission mode of n-th body sensor α S n 0 ,   1
β S n k k-th time slot is assigned to the n-th body sensor for direct transmission β S n k 0 ,   1 ,   n 1 ,   2 ,   ,   N ,   k ψ
δ S n S m k k-th time slot is allocated to n-th body sensor for transmitting data to m-th body sensor. δ S n S m k 0 ,   1 ,   n , m 1 ,   2 ,   ,   N ,   k ψ
δ S m H k m-th body sensor forwards the data from n-th body sensor to the hub at the k-th time slot. δ S m H k 0 ,   1 n , m 1 ,   2 ,   ,   N ,   k ψ
R n Data rate of the n-th body sensor
R n d Data rate of the n-th body sensor in direct transmission mode
R n c Data rate of the n-th body sensor in cooperative transmission mode
S I N R n , k d SINR of n-th body sensor in k-th time slot in direct transmission mode
S I N R n , m , k s r SINR of the source-relay link in k-th time slot in cooperative transmission mode
S I N R n , m , k r H SINR of the relay-hub link in k-th time slot in cooperative transmission mode
p n , k d Transmission power of the n-th body sensor in the k-th time slot in direct transmission mode
g S n H Transmission gain between the n-th body sensor and hub
p n , m , k s r Transmission power of n-th body sensor in the k-th time slot to m-th body sensor in cooperative transmission mode
g S n S m Transmission gain between the n-th body sensor and m-th body sensor in cooperative transmission mode
p n , m , k r H Transmission power of m-th body sensor in the k-th time slot to hub in cooperative transmission mode
g S m H Transmission gain between the m-th body sensor and hub in cooperative transmission mode
n 0 Noise power
R n c , s r Date rate of source-relay link in in cooperative transmission mode
R n c , r H Date rate of relay-hub link in in cooperative transmission mode
D Q S n k Data queue length at the n-th body sensor in time slot k
D Q S n m a x Maximum traffic queue length of body sensors
A S n k 1 Arriving traffic packets of n-th body sensor in time slot k − 1
E Q S n k Energy queue length at the n-th body sensor in time slot k
E Q S n m a x Maximum energy queue length of body sensors
E n , k 1 Amount of energy harvested by n-th body sensor in time slot k − 1
P S d a t a Date packet size
P S e n e r g y Energy packet size
p n m a x Maximum transmission power of body sensors
E E S n k Energy efficiency of n-th body sensor in time slot k
Table 2. The irrelevant state mapping table.
Table 2. The irrelevant state mapping table.
State   Space :   { D Q S n k ,   E Q S n k } The State Space If Needs to Be Explored
{0, 0}No
{0, *}No
{*, 0}No
{*, *}Yes
*: valid state; 0: invalid state.
Table 3. Computation complexity comparison between the modified and classical Q-learning algorithms.
Table 3. Computation complexity comparison between the modified and classical Q-learning algorithms.
Modified Q-Learning AlgorithmClassical Q-Learning Algorithm
xyx3y2
xyx5y3
--
xyx(2n+1)y(n+1)
Table 4. Simulation parameters setting.
Table 4. Simulation parameters setting.
ParametersValue
R10 m
Distance of each body sensorRandom distributed in (2, 5) m
S n (1:1:10)
B1 MHz
n 0 −94 dBm/Hz
p n m a x 10 dBm
λ t (1:1:8) packet/time slot
λ e (1:1:8) packet/time slot
ψ 200
τ k 0.5 ms
P S t r a f f i c 8 bits/packet
P S e n e r g y 0.0002 J/packet
D Q D U s m a x 50 packets
E Q D U s m a x 50 packets

Share and Cite

MDPI and ACS Style

Xu, Y.-H.; Xie, J.-W.; Zhang, Y.-G.; Hua, M.; Zhou, W. Reinforcement Learning (RL)-Based Energy Efficient Resource Allocation for Energy Harvesting-Powered Wireless Body Area Network. Sensors 2020, 20, 44. https://doi.org/10.3390/s20010044

AMA Style

Xu Y-H, Xie J-W, Zhang Y-G, Hua M, Zhou W. Reinforcement Learning (RL)-Based Energy Efficient Resource Allocation for Energy Harvesting-Powered Wireless Body Area Network. Sensors. 2020; 20(1):44. https://doi.org/10.3390/s20010044

Chicago/Turabian Style

Xu, Yi-Han, Jing-Wei Xie, Yang-Gang Zhang, Min Hua, and Wen Zhou. 2020. "Reinforcement Learning (RL)-Based Energy Efficient Resource Allocation for Energy Harvesting-Powered Wireless Body Area Network" Sensors 20, no. 1: 44. https://doi.org/10.3390/s20010044

APA Style

Xu, Y. -H., Xie, J. -W., Zhang, Y. -G., Hua, M., & Zhou, W. (2020). Reinforcement Learning (RL)-Based Energy Efficient Resource Allocation for Energy Harvesting-Powered Wireless Body Area Network. Sensors, 20(1), 44. https://doi.org/10.3390/s20010044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop