Next Article in Journal
Complex Deep Neural Networks from Large Scale Virtual IMU Data for Effective Human Activity Recognition Using Wearables
Next Article in Special Issue
Ultra-Wideband (UWB) Systems in Biomedical Sensing
Previous Article in Journal
Rail Corrugation Detection and Characterization Using Computer Vision
Previous Article in Special Issue
Validation of a New Contactless and Continuous Respiratory Rate Monitoring Device Based on Ultra-Wideband Radar Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Respiration Based Non-Invasive Approach for Emotion Recognition Using Impulse Radio Ultra Wide Band Radar and Machine Learning

1
Department of Computer Science, Khwaja Fareed University of Engineering and Information Technology, Rahim Yar Khan 64200, Pakistan
2
Management and Information Technology, Jubail Industrial College, Al Jubail 35718, Saudi Arabia
3
Department of Computer Science, Broward College, Broward County, FL 33301, USA
4
Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Korea
5
School of Engineering, London South Bank University, London SE1 0AA, UK
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(24), 8336; https://doi.org/10.3390/s21248336
Submission received: 2 November 2021 / Revised: 9 December 2021 / Accepted: 9 December 2021 / Published: 13 December 2021
(This article belongs to the Special Issue Ultra Wideband (UWB) Systems in Biomedical Sensing)

Abstract

:
Emotion recognition gained increasingly prominent attraction from a multitude of fields recently due to their wide use in human-computer interaction interface, therapy, and advanced robotics, etc. Human speech, gestures, facial expressions, and physiological signals can be used to recognize different emotions. Despite the discriminating properties to recognize emotions, the first three methods have been regarded as ineffective as the probability of human’s voluntary and involuntary concealing the real emotions can not be ignored. Physiological signals, on the other hand, are capable of providing more objective, and reliable emotion recognition. Based on physiological signals, several methods have been introduced for emotion recognition, yet, predominantly such approaches are invasive involving the placement of on-body sensors. The efficacy and accuracy of these approaches are hindered by the sensor malfunctioning and erroneous data due to human limbs movement. This study presents a non-invasive approach where machine learning complements the impulse radio ultra-wideband (IR-UWB) signals for emotion recognition. First, the feasibility of using IR-UWB for emotion recognition is analyzed followed by determining the state of emotions into happiness, disgust, and fear. These emotions are triggered using carefully selected video clips to human subjects involving both males and females. The convincing evidence that different breathing patterns are linked with different emotions has been leveraged to discriminate between different emotions. Chest movement of thirty-five subjects is obtained using IR-UWB radar while watching the video clips in solitude. Extensive signal processing is applied to the obtained chest movement signals to estimate respiration rate per minute (RPM). The RPM estimated by the algorithm is validated by repeated measurements by a commercially available Pulse Oximeter. A dataset is maintained comprising gender, RPM, age, and associated emotions which are further used with several machine learning algorithms for automatic recognition of human emotions. Experiments reveal that IR-UWB possesses the potential to differentiate between different human emotions with a decent accuracy of 76% without placing any on-body sensors. Separate analysis for male and female participants reveals that males experience high arousal for happiness while females experience intense fear emotions. For disgust emotion, no large difference is found for male and female participants. To the best of the authors’ knowledge, this study presents the first non-invasive approach using the IR-UWB radar for emotion recognition.

1. Introduction

Emotions are an integral part of our everyday life that represent conscious and/or unconscious mental reactions to events, objects, and situations. Emotions are a combined form of feelings, thoughts, and behavior and show people’s psychophysiological reactions. Emotions affect the way people think, comply, and feel about people, things, and events. The most widely used emotion classification models are Ekman and pleasure, arousal, dominance (PAD) models [1]. The PAD emotional state model was developed by Albert Mehrabian and James A. Russell to define and measure emotional states. The PAD model describes continuous emotion in three dimensions of pleasure, arousal, and dominance [2,3,4]. Initially, it was used in an environmental psychology theory with the central assumption that environments influence individuals through their emotional effect [2]. PAD is a psychological model for estimating all the emotional states of humans with respect to pleasure, arousal, and dominance. Paul Ekman and his colleagues concluded that there are six basic emotions including disgust, fear, anger, happiness, sadness, and surprise [5], where each emotion has particular characteristics that allow them to be expressed to varying degrees [6]. Since emotions are basically genetically determined, distinct emotions are perceived similarly throughout most cultures or nations [7]. Many works employ dimensional representations to analyze human emotions in many emotional dimensions. Emotions are composed of five coordinated activities: mental situation assessment, clinical findings (central and autonomous nervous system response), actions, facial expressions, and thoughts [8]. The usage and role of emotion recognition are indispensable in a multitude of domains. For example, emotion detection plays an essential role in the field of medicine specifically for patients with psycho-neural disorder or patients with learning disabilities and autism. Children with autism spectrum disorder (ASD) typically find it difficult to understand, communicate and regulate feelings [9]. For advanced human-computer interface (HCI) designs, the interaction between robots and humans can be more realistic and dynamic if the emotional state of humans can be determined accurately. In this way, it can strengthen human-machine engagement by using emotional information during conversation.
The importance and diverse use of emotions in artificial intelligence, psychology, cognitive neuroscience, and advanced robotics, etc., has made emotions recognition extremely significant. Consequently, several methods have been developed for emotion recognition over the past years. Predominantly, such methods are based on individual physical signs that include different expressions, speech, body movement, and gestures, etc. [10]. So, facial expressions, speech, behavior, and physiological signals can be used for emotion recognition [10,11,12]. Depending upon the nature of the data used for emotion recognition, these methods fall under subjective and objective categories. The first three methods are subjective as it is easier for people to hide their genuine emotions by deliberately changing their voice, manipulating their facial expression, and altering their behavior [13]. Physiological signals-based methods are objective and more reliable where the signals generated by central nervous systems such as electroencephalogram (EEG) are used for emotion recognition. Objective methods are less susceptible to manipulation and show better performance [14]. Thus, emotion recognition mapping with physiological signals intuitively seems more reasonable. Several methods record physiological signals such as Galvanic skin response (GSR), EEG, Electrocardiogram (ECG), and Electromyography (EMG) for the electric activity of the heart, skin, muscles, and brain, respectively. Understanding emotional response using the physiological signals is promising because they show unconscious representations and are not consciously manipulated by humans [15].
Emotion recognition methods using physiological signals involve invasive technologies or on-body sensors for signals measurement which make them prone to error. For example, the electroencephalogram cap used in EEG is comprised of electrodes, amplifiers, and analog to digital converters for recording human brain activity and it is to be placed on the human head for this purpose [16]. EEG signals analysis is challenging due to being non-stationary and the influence of complex environmental factors. The EEG signals, for example, are noisy and highly susceptible to environmental interference due to their low amplitude (i.e., 50 μ V to 100 μ V) [14]. Similarly, GSR records skin conductance that occurs due to sweat gland activity and involves skin electrodes mounted on hand and foot regions. Research shows that the data and the accuracy of GSR-based methods are sensitive to several factors such as inappropriately worn devices, unrestricted movements of participants and gender, etc. [17,18]. Additionally, many humans do not feel comfortable with the on-body sensors, and their movements introduce noise and error in the collection that affects the performance of emotion recognition [19]. Keeping in view the challenges and limitations associated with invasive emotion recognition approaches, a non-invasive emotion recognition method is a compelling necessity. This study utilizes impulse radio ultra-wideband (IR-UWB) to propose a non-invasive emotion recognition method and makes the following contributions:
  • A non-invasive emotion recognition method is proposed using IR-UWB radar. Emotion recognition is carried by measuring the chest movement of the subjects without involving on-body sensors and invasive technology.
  • A dataset of IR-UWB data is maintained involving 35 participants in total, including both males and females.
  • An approach is presented to measure respiration per minute (RPM) from the measured chest to IR-UWB distance data. Results of the proposed approach are verified by a commercial pulse oximeter.
  • For identifying the emotions, the IR-UWB data is complemented with the machine learning approach where ensemble voting is utilized including both hard voting and soft voting.
  • Analysis has been carried out for male and female participants separately to present the differences with respect to gender. The emotion recognition performance of the ensemble models is compared with other machine learning models.
The rest of the paper is structured as follows. Section 3 discusses important research works related to the current study. The process of data collection using IR-UWB proposed research methodology, and its related contents are described in Section 4. Section 5 provides the analysis and discussion of results. In the end, the conclusion is given in Section 6.

2. Background on Suitability of Respiration and IR-UWB for Emotion Recognition

2.1. Respiration Patterns during Emotions

Physiological signals record the response of various human organs such as the brain, heart, and sweat glands, etc. when facing situations involving fear, anger, love, and hatred, etc. For recording physiological signals, several devices are used; for example, EEG for brain activity, GSR for skin conductance and photoplethysmogram (PPG) for measuring the volume of blood flow [18]. Research shows patterns from such and similar other physiological signals can be subsequently used for human emotion recognition [20].
Of the physiological signals recorded during emotion arousal, respiration is more apparent and prevalent. The respiration patterns show a high correlation with human emotions, e.g., fast breathing may be caused excitement due to happiness, anger, or anxiety [21]. Happiness and other positive emotions have a substantial impact on respiratory changes [22,23]. The high frequency of heart rate variability is substantially influenced by respiration due to heart rate increase during inspiration and heart rate decrease during expiration, a phenomenon called respiratory sinus arrhythmia (RSA) [24,25]. Happiness has a variable effect on breathing rate depending on how arousal one is, whilst arousing one increases the respiration rate [26]. Similarly, humans tend to suppress their breaths when having emotions of disgust [22,23]. Research also shows [22,23,27] that humans demonstrate shallower and faster breathing when facing fear. In the light of these findings, this study adopts the respiration rate measurement for predicting various emotions using the IR UWB radar.

2.2. Suitability of IR-UWB for Emotion Recognition

Currently, different devices are deployed for physiological signals measurement. For example, EEG is used for recording the electrical activity of the human brain when facing certain emotions [28]. The heart rate is monitored using electrocardiography (ECG) involving the famous 12-lead ECG technique where nine sensors are placed on arms, legs, and the chest. Heart signals show different patterns during different emotions. GSR is used to record continuous skin conductance of the human skin. GSR shows the changes in sweat reaction caused by the change in the sympathetic nervous system [29]. Similarly, skin temperature measurement, electromyogram (EMG), and electrooculography (EOG), etc. are among the commonly used devices for recording physiological signals. For respiration rate monitoring, the use of radar has been increased over the past few years. Consequently, several types of radars have been leveraged for recording respiration patterns such as continuous wave Doppler radar, ultra-wideband radar, frequency modulated continuous wave (FMCW) radar and stepped-frequency continuous wave radar [30]. Radar can provide larger detection areas as compared to other approaches, for example, video and thermal cameras, and can be used to monitor multiple subjects. Provided the proper body position of the subject, higher accuracy of up to 97% is achievable for respiration rate monitoring using FMCW radar [31]. Radar shows high accuracy for monitoring the tiny movements of the chest wall and can provide accurate respiration-related movements.

3. Related Work

On account of large interest in emotion recognition during the past few years and the importance of respiration rate association with emotions, several studies have presented methods and systems to recognize emotions based on respiration rate.
Augsburg’s dataset of physiological signals has been used for the classification of emotions by [32]. Augsburg dataset consists of twenty-five records for four emotions including joy, sadness, pleasure, and anger where these emotions were triggered by the musical induction process. Four forms of physiological signals have been obtained during this process: ECG, EMG, respiration, and skin conductivity (SC). The ensemble empirical mode decomposition (EEMD) approach is used to extract time-domain non-linear, time-frequency, and intrinsic mode (IMF) features. The C4.5 decision tree (DT) is used to limit the number of features to five optimal features with a major contribution to classification. The correct classification rate (CRR) is used to measure the output and results show a CRR of 88% using the selected features.
The study [33] uses a dataset for emotion analysis using physiological signals (DEAP) dataset containing ECG, GSR, blood pressure, breathing, skin temperature (ST), EMG, and Electrooculogram signals of thirty-two participants (16 of each gender) with ages from 19 to 37 years. Data are recorded when the subjects are watching forty-one-minute music video clips with a rating of 1-9 which is negative/low to positive/high using arousal, valence, and liking. The study uses only two ECG and respiration signals for emotion recognition. Thirteen features are obtained at a sampling rate of 512 Hz. Respiration rate (RR) interval, low frequency (LF), heart rate (HR), and high frequency (HF), RSA power, RSA frequency, and RSA amplitude are among the thirteen cardiac features. Breathing frequency and amplitude, RSA amplitude ratio to respiratory oscillation, respiratory and RSA frequency difference, the phase difference of respiration and RSA, the slope of phase difference, and standard deviation are calculated. For emotion classification, an SVM classifier with a multilayer perceptron kernel is used. Low/high liking, positive/negative valence, and low/high arousal are performed for the classification of ECG and respiration signals. Using the ECG signals, accuracy scores of 74%, 71%, and 72% are obtained for liking, arousal, and valence, respectively. Respiration rate shows an accuracy of 73% for liking, 72% for arousal, and 70% for valence. On the other hand, with a combination of HR and RR, the classification accuracy is increased to 76%, 74%, and 74% for liking, arousal, and valence, respectively.
Along the same lines, ref. [34] uses various physiological signals such as ECG, ST, GSR, EMG, HR, respiration rate, blood oxygen level, systolic blood pressure (SBP), diastolic blood pressure (DBP), and blood volume pulse (BVP) to distinguish anger, pleasure and neutral emotions. A total of three stable male subjects aged 18 to 19 years participated in the data collection process. Zephyr BioPatch chest brace is used for the collection of ECG and respiration signals while E4 wrist-band is used to capture GSR, ST, and (BVP). The SBP and DBP are calculated using CONTEC which is an off-the-shelf Bluetooth-enabled blood pressure device. Blood oxygen level is measured by a pulse oximeter. The Zephyr chest strap is worn under the pectoral muscle for data collection. Blood glucose readings are taken twice at the beginning and the end of the trial. Baseline signals are captured while the participant is sitting comfortably, followed by inducing the joy emotion joy using video clips. For inducing anger, cognitive techniques are applied. The physiological signals of each subject are recorded twice in the state of the induced emotion. Four physiological signs SBP, DBP, EMG, and blood oxygen levels are omitted because they do not contribute much to the classification of joy and anger emotions. Bandpass filtering is used for ECG, respiration, GSR, ST, HR, and BVP signals. Because the dataset is limited (two instances for each emotion class), predictive and statistical analysis of the data is carried out. The emotion is classified as happiness when the signal has high GSR, HR, and moderate respiration value while anger has low GSR, ST, high respiration, and moderate HR values.
ECG, RR, blood pressure, and respiration inhalation, and exhalation temperature are used for emotion classification in [35]. Several sensors are used to capture four physiological signals from a single subject at the time of the induction of emotion through a three-minute video clip. The movie clip is played at a one-meter distance at the laptop and the signals are recorded for one minute. By applying the Low-Pass filter to the raw ECG and respiratory signal, the noise is reduced. Afterward, nineteen statistical, temporal, and spectral features are extracted for emotion recognition. An artificial neural network (ANN) with two hundred hidden layers is used for classification. A total of six emotions happy, sad, fear, disgust, anger, and surprise are classified with an average accuracy of 80%.
In [36] MAHNOB-HCI physiological signal dataset is used for the classification of emotions in the arousal valence model. ECG, respiration, skin temperature, and galvanic skin reaction are used for emotion classification. The MAHNOB-HCI multimodal dataset contains data for twenty-four subjects using twenty video samples. The signal of the first and last thirty seconds is omitted because of neutral emotion. Butterworth filter is applied on GSR, ECG, and respiration signal with a cutoff frequency of 0.3 Hz, 0.7 Hz, and 1 Hz, respectively. Heart rate variability (HRV) is calculated from the ECG signal and respiration rate from respiration amplitude. A total of one hundred sixty-nine features are extracted from these signals. An SVM with different kernels is used for classifying the samples into high and low arousal, negative and positive in valence. SVM with RBF kernel shows a better accuracy of 68.5% for arousal and 68.75% for valence class.
A physiological signal interpretation framework, Emo-CSI, is presented for emotional classification in [37] which uses heart rate, respiration pattern, skin humidity, and strength to recognize emotions of pleasure, displeasure, calm, neutral, and excited emotions. Twenty-three subjects with ages from 20 to 27 are included in the data collection process seconds comprising ten males and thirteen females. Emotions are induced using pictures and matching sounds. A total of thirty-two features including average, maximum, minimum, and standard deviation, etc. are extracted from physiological signals to support prediction using SVM, DT, and artificial neural network (ANN). SVM outperforms the other two classification models with an accuracy of 55.45% and 59% for valence and arousal classes, respectively.
Respiration data are used in [38] where the emotions are classified using that are Fast Fourier Transform (FFT) and machine learning models separately. Twenty-five males and females with ages from 18 to 25 participated in the experiments where seven movie clips are shown to subjects for six emotions inducing happiness, sadness, surprise, anger, anxiety, and disgust. The study uses the BIOPAC instrument, airflow sensor, and a mouthpiece for collecting respiration, and breathing patterns. FFT-based classification achieves an 80% accuracy while LR obtains an 80% recall. A system using HRV signal based on respiration rate for emotion classification is proposed in [39]. Twenty-five subjects (12 males & 13 females) with ages from 18 to 35 are involved in experiments for anger, fear, joy, and sadness emotions. ECG and respiration signals are recorded using a BIOPAC device. The largest peak from the bandpass filtered respiratory signal is used to calculate respiratory frequency. The area under the receiver operating characteristic curve (AUC) is calculated to find the capability of features in emotion classification. Features having AUC greater than or equal to 0.70 are considered for further process. Accuracy of two-class classification (relax vs. joy, joy vs. sad, and joy vs. anger) is 79.2%, 77.8%, and 77.3%, respectively.
The literature study has several important findings. First, despite the higher classification accuracy, predominantly, the used methods involve invasive or on-body sensors. Invasive sensors can be used in virtual or controlled environments but for real-life situations their practicability is limited. Often causing attention or inconvenience, the data collection process becomes erroneous. Secondly, respiration rate has been employed largely in recent studies and is a potential candidate for accurate emotion detection. Thirdly, UWB radar has not been studied extensively and requires further research efforts to explore its full potential for emotion recognition. Keeping in view these points, this study presents a non-invasive emotion recognition approach by employing IR-UWB for recording physiological signals.

4. Materials and Methods

4.1. Impulse Radio-Ultra Wide Band Radar

Time domain and extreme spectrum commercialized IR-UWB radar appeared in the late 1990s [40,41]. It is a developing technology that was originally used by the United States (US) army in the 1970s. The Federal Communications Commission (FCC) of the US has allocated a bandwidth of 7.5 GHz for UWB signals. A signal is called UWB if its bandwidth exceeds 500 MHz and its frequency range falls in the ranges from 3.1 GHz to 10.6 GHz [40]. Despite the high data rates, UWB signals have a limited transmission capacity. IR-UWB radar creates high bandwidth signals by transmitting very short-duration pulses in the range of nanoseconds. The IR-UWB radar has no privacy concerns and is unaffected by environmental variables. Because of the IR-UWB radar’s extremely low emission power, it has no adverse effect on the human body and passes the safety standards. Because of its non-ionizing, non-intrusive, non-tackling, no blind spot angles, and ability to penetrate a range of materials or barriers, the IR-UWB radar offers benefits over other existing tools for recording the physiological signals [42,43,44].

4.2. Proposed Research Methodology

Figure 1 shows the architecture of the research methodology followed in the current study. Initially, the objectives of the study are defined including the emotion prediction, types of emotions, and the non-invasive prediction procedure. The second step is the selection of appropriate sensors for which a sensor is to be decided regarding its ease of use, cost, robustness, and ease of deployment for practical applications. It is followed by the data collection set up including the decisions regarding the number, age, and gender of the participants and the data collection procedure such as static or dynamic. Data cleaning and transformation procedures are defined to remove the noise and increase data quality. Then the experimental design is made with the selection of appropriate tools to be used for data collection routines and sensors placement for experiments, etc. In the end, prediction is carried out using the selected machine learning and deep learning models which are selected based on their reported results and suitability with respect to the task at hand.

4.2.1. Research Questions and Objectives

In the first phase of this research, objectives are defined. The primary objective of this research is to devise a non-invasive method for detecting human emotions. Initially, three emotions are considered to validate the feasibility of non-invasive emotion detection. In addition, the study also aims to obtain higher accuracy using single sensor data which means that data preparation and transformation should also be studied.

4.2.2. Sensors Selection

Appropriate data is the fundamental element for obtaining accurate human emotions and data quality is directly linked with the data collection sensor. Besides data quality, other important aspects are the cost, availability, and ease of deployment of the data collection sensor. Currently, GSR, ECG, EEG, EMG, respiration, and SC physiological signals are used for emotion detection, each with its own merits and demerits. Keeping in view the results reported from different research studies [35,38,39], respiration rate has been selected for emotion detection. IR UWB radar’s tolerance to environmental noise, human eye safety, reduced cost, and ease of deployment are convincing traits for using it for respiration data gathering.

4.2.3. Data Collection Setup

For collecting the physiological signals, this study uses an X4m300 IR-UWB radar, as shown in Figure 2. It has a configurable frame size with a detection time of 1.5 to 3.5 s while the detection zone is programmable to up to 9.4 m [45,46].
The IR-UWB radar runs on the X4 chip framework (system on chip (SoC)) with unlicensed core frequencies of 7.29 GHz or 8.748 GHz at 1.4/1.5 GHz (−10 dB) bandwidth. It is operated by a very sensitive XeThru X4 UWB chip that senses smaller motion and has the best signal-to-noise ratio. It has built-in antennas and has a very high-resolution rate due to short pulse transmission at nanoseconds [47]. The built-in firmware creates a baseband signal that covers 9.4 m beginning at 0.18 m when the board’s default settings are used. This distance is split into 181 bins with 0.0514 m bin lengths. For the current study, the effective range is between 0.2 to 1.6 m. A baseband data frame counter is given with a frame size of 0.2 to 1.6 m. The corresponding bins for this effective range start with the 2nd bin at 0.282 m and end at the 28th bin at 1.569 m. A baseband data frame counter is given with a frame size of 2 32 . For each radar frame, i.e., the X4 UWB radar SoC output, the frame counter is increased by one. The frame counter does not reset when a profile is stopped or started. When the limit is hit, the frame counter resets to 0. The frame counter will be reset if the X4 UWB radar SoC is reset or the sensor module’s power is toggled [48].
The IR-UWB radar is used to record the respiration process where the chest moves due to the lungs’ respiratory operations. During the respiration process, the inhalation fills the human lungs with oxygen and expands the upper portion of the body and the space between the radar and the chest of the human body reduces, whereas the exhalation does exactly the opposite. A plot of the resulting distance is given in Figure 3 where the impact of inhaling and exhaling is shown. The inhalation increases the distance of the chest from the radar while the exhaling decreases this distance. To count the number of expands and contracts of the chest due to respiration the area under the curve is considered.
For data collection, the frequency of radar is set to 20 Hz. Subsequently, 1200 (20 × 60) areas under the curve are estimated in one minute. The data are recorded for 5 min and contain 6000 (20 × 60 × 5) area values under the curve. A total of 35 subjects participated in the data collection process including 14 males and 21 female participants of age ranges between 18 to 30 years and the average age of 24 years. The subjects are wearing seasonal clothes during the experiment and no special clothes are worn for the data collection. Research shows that the UWB radar is not influenced by the clothes and if slight variations are caused, they can be moved during the data preprocessing using the low pass filter [49,50]. An ethical approval statement is designed and approved by the Khwaja Fareed University of Engineering and Information Technology (KFUEIT) ethical committee and a consent form is signed by each subject. The testbed is set up in the Computer Science Department lab in KFUEIT. The nature and process of the whole experiment are explained and demonstrated to the subjects before data collection.
A total of nine videos have been selected to induce emotions and three emotions are considered for this study including happiness, fear, and disgust. Each video lasts for five minutes. Happiness is triggered by showing comedy movie clips while fear by showing clips of horror movies. Disgust emotion is induced by showing movie clips of persons eating distasteful food. Subjects are asked to enter the room alone and sit on the chair facing the IR-UWB radar. A distance of one meter from the IR-UWB is maintained for the person. This distance is selected based on the best view of the video clips from the sitting place. Figure 4 shows the experiment setup for data collection.

4.2.4. Data Transformation

The collected data for area values under the curve comprises respiration, pulse, and noise including heartbeat, belly movement, eye-blinking, eyeball movement, and other ambient motions. However, the proposed method requires only the respiration patterns for which respiration signals are obtained from the collected data. Steps involved in data cleaning and obtaining the respiration data are portrayed in Figure 5.
Using the Fourier transform, the frequency spectrum of the gathered signal is obtained. The maximum frequency of an adult’s respiration rate is 0.4 Hz [44,51,52,53,54]. To get the respiration signal, a filter with a cut-off frequency of 0.4 Hz is required. For the normalized frequency in this study, the cut-off frequency of 0.4 Hz shifts to 0.04. Thus, a respiration signal is extracted by applying a tenth-order low-pass Butterworth filter with a cut-off frequency of 0.04 to remove the higher frequency noise. The Butterworth filter is a digital filter with a very smooth passband frequency response curve. The square amplitude response function of the filter is
| H ( j w ) | 2 = 1 1 + ( w w C ) 2 N
where N denotes the filter’s order, which is a positive integer, and W c is the low-pass filter’s cutoff frequency. In this study, N = 10 and W c = 0.04 .

4.2.5. Experimental Design

For emotion recognition, this study uses the respiration signals acquired from IR-UWB radar. However, the raw signals are not used for emotion classification. Instead, respiration per minute (RPM) is leveraged to this end. For obtaining the RPM, initially, the data from IR-UWB radar are collected for the inhaling and exhaling process.
Approximation of area under the curve is performed using the Trapezoidal rule. For this purpose, the area under the curve for each frame is found. The trapezoidal rule evaluates the area under the curve by splitting the area into trapezoids, unlike the Reimann sums which follow a rectangular approach [55]. Let f ( x ) be continuous signal on [ a , b ] , the interval [ a , b ] can be partitioned into n equal subintervals where the width of each is
Δ x = b a n , a = x 0 < x 1 < x 2 . . . < x n = b
Trapezoidal rule to approximate a b f ( x ) d x is given as
a b f ( x ) d x T n = Δ x 2 [ f ( x 0 ) + 2 f ( x x ) + . . . + 2 f ( x n 1 ) + f ( x n ) ] ,
where Δ x is given in Equation (2) and x i = a + i Δ x .
The Trapezoidal rule on 60,000 frames, provides 60,000 area values of the curve present in each frame. These values correspond to lungs inhaling and exhaling, the area under the curve increases during the inhaling process and vice versa. Subsequently, Fast Fourier Transform (FFT) is applied to the values obtained from the Trapezoidal rule. FFT can turn a time-domain signal f ( t ) into a frequency domain signal F ( j w ) . The Formulae (4) and (5) are for Discrete Fourier Transform (DFT) and the inverse transform.
X ( k ) = n = 0 N 1 x ( n ) e j k ( 2 π N ) , k = 0 , 1 , 2 , 3 , . . . , N 1
The discrete Fourier transform converts the time-domain sequence x ( n ) into the discrete frequency domain signal X ( k ) . The Inverse Discrete Fourier Transform (IDFT) formula is as follows
x ( n ) = 1 N n = 0 N 1 X ( k ) e j k ( 2 π N ) n , k = 0 , 1 , 2 , 3 , . . . , N 1
FFT is an enhanced form of DFT that greatly speeds up the calculation time of DFT. The FFT is used in the proposed work for the curve x ( n )
x ( n ) , n = 0 , 1 , 2 , . . . , N 1
Obtaining RPM from IR-UWB data involves processing the data through several steps. The trapezoidal rule applied data is transformed into the frequency domain using FFT. Subsequently, the Butterworth filter is applied to extract the data related to respiration only. The Butterworth filtered data is used to find the peaks which represent the inhale process. For peaks, high movement locations in the data are to be found. Each round of inhaling and exhaling is regarded as one respiration. Finally, RPM can be obtained using
R P M = n p T
where n p refers to peaks (data magnitude α ) while T shows the time in minutes.
For emotion recognition, machine learning algorithms are trained and tested using the obtained RPM information for each emotion.

4.2.6. Prediction

Several supervised machine learning models are applied for emotion recognition on IR-UWB obtained data. These models are selected based on the results reported in other research works and include K nearest neighbor (KNN), extra tree classifier (ETC), AdaBoost classifier (ADB), gradient boosting machine (GBM). To obtain high classification accuracy, several hyperparameters have been fine-tuned for these models and a complete list is provided in Table 1.
Besides using the machine learning models, two deep learning models have been employed as well for emotion recognition problems. For this purpose, multi-layered perceptron (MLP) [56] and convolutional neural network (CNN) [57] have been used so that a performance comparison can be made between the machine and deep learning approaches for the task at hand. The architecture of the models is customized in terms of the number of layers and neurons, and several other parameters such as optimization model, learning rates, and activation functions. A complete list of models’ parameters and architecture is provided in Figure 6. Each model is compiled with ‘categorical_crossentropy’ loss because of multi-class data and ‘adam’ optimizer is used for optimization. Deep learning models are fitted with a batch size of 16 and 100 epochs are used to train the models. For avoiding the overfitting problem, dropout layers with different rates are used both in CNN and MLP networks.

4.3. Proposed Hard Voting and Soft Voting Models

Besides the machine learning models, this study proposes two ensembles of XGB, ADB, GBM, and KNN which use hard voting (HV) and soft voting (SV) criteria to make the final prediction. The architectures of both models are shown in Figure 7. SV works based on probability of each class as predicted by each model and these probabilities pass through average criteria. In the end, the argmax function is used to average the probabilities to compute the final prediction [58]. In SV model, X 1 , X 2 , and X 3 are the probabilities by XGB for class 1 ( C 1 ), class 2 ( C 2 ), and class 3 ( C 3 ). For HV model, P 1 , P 2 , P 3 , and P 4 are the predictions by each model and class with more votes will be final class [59]. The C 1 p , C 2 p , and C 3 p are the average probabilities for C 1 , C 2 , and C 3 . S V p is the final prediction by using soft voting criteria and H V p is the final prediction using hard voting criteria.
Algorithm 1 shows the steps followed by the ensemble model based on the hard voting criteria. The T X G B , T A D A , T G B M , and T K N N represent the trained XGB, ADA, GBM, and KNN models on the feature vector of different emotions. Each model predicts a given sample with respect to one of three emotions. The prediction from each of these models has one vote and the final prediction H V P r e d is based on the majority of the given models’ prediction for a particular emotion.
Algorithm 1 HV Algorithm for Emotion Prediction
Input: Emotion RPM
Output: Happiness, Fear, or Disgust
 1:
T X G B T r a i n e d X G B
 2:
T A D A T r a i n e d A D A
 3:
T G B M T r a i n e d G B M
 4:
T K N N T r a i n e d K N N
 5:
for i in D a t a s e t do
 6:
     X G B P r e d i c t i o n T X G B ( i )
 7:
     A D A P r e d i c t i o n T A D A ( i )
 8:
     G B M P r e d i c t i o n T G B M ( i )
 9:
     K N N P r e d i c t i o n T K N N ( i )
10:
     H V P r e d a r g m a x { X G B P r e d i c t i o n , A D A P r e d i c t i o n , G B M P r e d i c t i o n , K N N P r e d i c t i o n }
11:
end for
12:
Output : H a p p i n e s s | F e a r | D i s g u s t H V P r e d i c t i o n
Algorithm 2 shows the working steps of the decision made under the soft voting criteria. Similar to the hard voting scheme, T X G B , T A D A , T G B M , and T K N N are the trained XGB, ADA, GBM, and KNN models. A probability score of each emotion is predicted by each trained model. For example, H a p p y P o b X G B represents the probability score for happy emotion from the trained XGB model. Similarly, XGB provides the probability score for disgust and fear emotions. For making the final prediction, H a p p y P r o b , the average probability score for happy emotion, is calculated from each trained model and the highest probability score gives the final prediction S V P r e d .
Algorithm 2 SV Algorithm for Emotion Prediction
Input: Emotion RPM
Output: Happiness, Fear, or Disgust
 1:
T X G B T r a i n e d X G B
 2:
T A D A T r a i n e d A D A
 3:
T G B M T r a i n e d G B M
 4:
T K N N T r a i n e d K N N
 5:
for i in D a t a s e t do
 6:
      H a p p y P o b X G B T X G B ( i )
 7:
      F e a r P o b X G B T X G B ( i )
 8:
      D i s g u s t P o b X G B T X G B ( i )
 9:
      H a p p y P o b A D A T A D A ( i )
10:
     F e a r P o b A D A T A D A ( i )
11:
     D i s g u s t P o b A D A T A D A ( i )
12:
     H a p p y P o b G B M T G B M ( i )
13:
     F e a r P o b G B M T G B M ( i )
14:
     D i s g u s t P o b G B M T G B M ( i )
15:
     H a p p y P o b K N N T K N N ( i )
16:
     F e a r P o b K N N T K N N ( i )
17:
     D i s g u s t P o b K N N T K N N ( i )
18:
     H a p p y P r o b ( H a p y p P o b X G B + H a p p y P o b A D A + H a p p y P o b G B M + H a p p y P o b K N N ) / 4
19:
     F e a r P r o b ( F e a r P o b X G B + F e a r P o b A D A + F e a r P o b G B M + F e a r P o b K N N ) / 4
20:
     D i s g u s t P r o b ( D i s g u s t P o b X G B + D i s g u s t P o b A D A + D i s g u s t P o b G B M + D i s g u s t P o b K N N ) / 4
21:
     S V P r e d a r g m a x H a p p y P r o b , F e a r P r o b , D i s g u s t P r o b
22:
end for
23:
Output : H a p p i n e s s | F e a r | D i s g u s t S V P r e d i c t i o n

5. Results and Discussion

Figure 8 shows the steps followed to perform experiments. It involves data collection, data cleaning to obtain the respiration signal, followed by RPM gathering, data split, and training and testing.

5.1. Results of RPM Using Machine Learning Models

Since the training and prediction of the emotions are based on RPM, so accurate RPM estimation is very critical. For this purpose, the raw respiration data are processed to clean the noise. Figure 9a,b shows the noisy and clean data, respectively. It can be seen that the processed data provides smooth peaks as compared to the noisy data and peak estimation is easy in the cleaned data. The cleaned data are used to detect and count the peaks using the defined α threshold, as shown in Figure 10. RPM is counted using Equation (3).
For validating the performance and extent of accuracy for RPM, experiments are performed using different participants. The procedure was replicated several times against various participants. The obtained RPM is validated against a commercial Pulse oximeter used in [44] and shown in Figure 11.
RPM validation experiments are performed for both static and dynamic environments where the subject sits on a chair in the static movement while the subject’s measurements are taken for the dynamic environment when reading a book. Ten people are chosen for the validation experiment including five males and five females between the age of 25 to 30 years. The person operating radar signals informs the subjects and triggers the radar and in the meantime, the subject triggers the wearing pulse oximeter. The hand movement of the subject while he triggers the pulse oximeter does not affect the RPM obtained from the chest movement signal recorded by the UWB radar. Individuals are told to sit comfortably in a chair facing the radar, with the pulse oximeter attached to their left index finger. Each subject’s chest movements are recorded twice for 1 min each time. RPM using IR-UWB data is calculated following the above-described procedure.
Table 2 displays the results of calculated RPM and commercially available pulse oximeter. Results show that the RPM calculated by the proposed method is in complete agreement with the pulse oximeter results validating the accuracy of the proposed method.
To ensure that the movement of participants does not affect the performance of the proposed system, experiments are carried out while subjects are reading a book. A pulse oximeter is attached to the left index finger, and radar is placed in front of the subject. The experiment included eight male individuals, and data was collected using a pulse oximeter and a radar at the same time. Experimental results for the dynamic environment are provided in Table 3. Results suggest that the method is robust and can perform well even in dynamic environments with slight variation in the measured RPM. For example, for participants 1, 2, 4, and 5 a difference of 1 RPM is found between the pulse oximeter RPM and the calculated RPM, however, it is not significant to influence the emotion detection process.
For corroborating the hypothesis that there is no statistically significant difference between the pulse oximeter and calculated RPMs, a T-test is performed. Table 4 shows the value of the statistical T-test to evaluate the RPM for static and dynamic movement of the participant. Similarly, the RPM statistics of males and females are also analyzed. In both cases, the RPM data is statically equal indicating that there is no significant difference in static body RPM and dynamic body RPM and there is also no difference between female and male RPMs.

5.2. Results of RPM for Male and Female Participants

To analyze the difference of RPM in different gender, a separate set of experiments has been carried out. Such experiments are performed with a two-fold purpose. First, the difference in male and female participants can be analyzed using the obtained RPM. Second, the RPM difference is investigated for each gender with respect to the three emotions studied in this study. The average RPM of males and females when different emotions are induced is presented in Table 5. Two important findings are obtained during experiments. First, there is an observable difference in RPM for males and females. Second, the RPM of males and females is affected differently when faced with different emotions. For happiness, male participants show a higher RPM indicating that they are aroused highly as compared to female participants. On the other hand, fear emotion is more experienced by female participants. Regarding the disgust emotion, the RPM is almost similar for both males and females. We performed statistical analysis on female and male RPM using statistical t-test and results of t-test show that male and female RPMs are not statistically different regarding each emotion. Both RPMs show the same statistics as t-test findings given in Table 4 indicates.

5.3. Results for Emotion Recognition

For capturing the data during different induced emotions, videos are played on the laptop remotely controlled using the team-viewer. The Chest movement of each subject is acquired three times by showing three videos of each emotion for three consecutive days (one emotion per day). The RPM is estimated from the chest movement of subjects using the proposed approach. All classifiers are trained and evaluated to compare their performance and select the best fit model. A structured dataset is maintained comprising of RPM, gender (’0’ for ’female’ and ’1’ for ’male’), and age of the subject along with the labels ’0’ for ’happy’, ’1’ for ’disgust’ and ’2’ for ’fear’. A few sample records from the dataset are shown in Table 6.
The use of the RPM feature is validated through feature importance. Figure 12 shows the feature importance using the random forest (RF) algorithm. It shows that RPM has significant importance for emotion classification as compared to age and gender. While gender feature score is the lowest which also indicates that the female and male do not impact the classification accuracy.
The dataset contains 315 tuples in total with 105 for each emotion. The dataset is divided into two set for training and testing in different ratios. The ML models are fed with a non-standardized feature vector including gender, RPM (average respiration rate of 5 m), age, and emotion as input. Following that classifiers are evaluated using previously split test data. Table 7 shows the classification results using 90:10 train-test split. GBM obtains the highest accuracy of 0.79 following by HV and SV with 0.77 and 0.74, respectively.
Reducing the train data degrades models’ performance as expected. The highest accuracy is now 0.78 by GBM, as shown in Table 8. HV obtains the 2nd highest accuracy with 0.75 followed by 0.73 each from SV and XGB. Despite the reported accuracy scores, the F1 score is preferred over accuracy as it incorporates both precision and recall and has been regarded as a more balanced measure to evaluate the performance of a model. The proposed ensemble SV obtains the highest F1 score of 0.76. Similarly, XGB shows an F1 score of 0.74 which is also significant considering that the problem is a multiclass classification.
The performance of the models is further reduced once the ratio of the train-test split is changed to 70:30, as shown in Table 9. Although the performance is degraded in terms of accuracy, yet, the F1 score is improved on average with six out of seven classifiers having 0.70 plus F1 score which is not the case for 90:10 or 80:20 train-test split.
For further analysis of the models’ performance, experiments are performed using 60:40 and 50:50 train-test splits and results are shown in Table 10 and Table 11, respectively. Results indicate that both the accuracy, as well as, F1 score has been impacted on account of the decrease in the training data. One important point to be mentioned is the consistent performance of GBM and SV which shows better performance for all train-test splits and be among the significantly performing models.
The performance of all models is more significant with 70:30 and 80:20 train-test split ratios, especially, in terms of F1-Score. With enough data for models’ training, the results are significantly better. The used dataset is multiclass and also small in size, so considering the model’s overfitting problem F1 Score can be more appropriate for a fair comparison. That is the reason for selecting the F1 score as the primary performance evaluation measure. Results indicate that the proposed SV is significant with the highest F1 score of 0.76 with an 80:20 splitting ratio. The significant performance of SV is because of probability-based ensemble architecture. Other tree-based models, such as ETC, GBM, ADAB, and XGB also show good results.

5.4. Results of K-Fold Cross Validation

For validating the performance of machine learning models, 10-fold cross-validation has been performed and the results are given in Table 12. The proposed SV achieves the highest accuracy of 0.66 with a ±0.33 standard deviation (SD). Proposed HV and XGB are just behind the ESV with 0.65, and 0.64 accuracy scores and ±0.29, ± 0.40 SD, respectively.

5.5. Experimental Results Using Deep Learning Models

Results of deep learning models are shown in Table 13. Results indicate that the performance of deep learning models is not significant in comparison to machine learning models. CNN obtains the best F1 score of 0.54 with an 80:20 ratio while MLP achieves a 0.61 F1 score with a 50:50 ratio. On average, MLP shows better performance as compared to the CNN model. The maximum accuracy of 0.72 is obtained by MLP. Primarily, deep learning models tend to show better performance when fed with a large amount of training data. For the current study, the small amount of the collection may be a probable reason for the poor performance of the deep learning models. CNN required a large feature set for better training and in this study dataset features set is small which leads to a poor fit of the CNN model. While MLP is somehow better in comparison with CNN because of MLP’s simple architecture as compared to CNN. MLP is a basic model which can be good even on a small dataset because of its simple architecture.

5.6. Comparison with Recent Research Studies

The performance of the proposed approach is compared to other research works that utilize respiration rates for emotion classification. Table 14 indicates that [35,38] shows slightly better accuracy than the proposed system. However, ref. [35] uses both ECG and respiration data for emotion classification, and data from multiple sensors tend to show better results. Electrodes were used to record ECG signals while respiration rate was calculated from the inhale and exhale of gases through the oxygen mask subject wore during the data collection. In contrast, the proposed approach obtains very similar results using only the respiration rate. A Biopac device and mouthpiece were used by [38] to collect respiration rate and respiration patterns and the proposed approach is an invasive method, contrary to the non-invasive approach presented in this study. The devices used by [35,38] for data collection need to be worn on the body that requires correct placement and makes the subject responsive and overly reacting resulting in inaccurate data. The proposed approach obtains respiration rates from chest movements using UWB radar. UWB radar is a non-invasive technology that can detect chest movements across a range of 0.2 to 1.6 m. Other movements such as blinking, head movements, and so on, can be identified and filtered.

6. Conclusions and Future Work

Emotions are conscious and/or unconscious mental reactions from humans to various events, objects, and situations and are part and parcel of human life. Although emotion recognition is important for humans to respond appropriately, the past few years are marked with increased research in emotion recognition for human therapy, the latest human-computer interface, and advanced humanoids. Several kinds of physiological signals can be utilized for the task at hand with EEG, ECG, GSR, and respiration-based methods as the most famous. Predominantly, such methods involve placing devices on the head and chest or attaching sensors to various limbs. Consequently, such methods introduce inconvenience for humans which leads to noisy and erroneous data and wrong emotion recognition.
This study presents a novel non-invasive emotion recognition approach based on human respiration patterns. The respiration data are obtained using the novel use of IR-UWB radar for three emotions including happiness, fear, and disgust. These emotions are induced using movie clips while the chest movements of thirty-five participants are collected including both males and females. RPM calculated by the proposed method is validated by a commercial Pulse Oximeter which shows a 100% accuracy. RPM and other features are fed into the machine and deep learning models for training and testing. Additionally, hard and soft voting-based ensemble models are proposed for emotion recognition as well. Extensive experiments have been carried out involving different train-test splits. Results indicate that using the proposed novel non-invasive approach, a 76% F1 score can be obtained from the proposed SV with an 80:20 train-test split. Besides the basic emotion recognition, separate experiments are performed based on gender which provides insight on gender-based behavior during different emotions. Male participants show high arousal in happiness while fear emotion is more prevalent and intense in female participants. The emotion intensity for disgust is almost similar for both males and females. This study lays the foundation for IR-UWB based non-invasive emotion recognition, yet, only three emotions are studied at the moment. In the future, the dataset will be extended to add further emotions, as well as, improve the accuracy of machine learning, and deep learning models by incorporating additional features.

Author Contributions

Conceptualization, H.U.R.S. and H.F.S.; Data curation, H.U.R.S., H.F.S. and A.A.S.; Formal analysis, H.U.R.S. and A.A.S.; Funding acquisition, E.L.; Investigation, A.B.K.K., E.L. and S.D.; Methodology, H.U.R.S., F.R. and I.A.; Project administration, A.B.K.K.; Resources, E.L. and S.D.; Software, H.F.S. and F.R.; Supervision, H.U.R.S., I.A. and S.D.; Validation, A.A.S. and A.B.K.K.; Visualization, F.R.; Writing—original draft, H.U.R.S.; Writing—review & editing, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Florida Center for Advanced Analytics and Data Science funded by Ernesto.Net (under the Algorithms for Good Grant).

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Acknowledgments

This research was supported by the Florida Center for Advanced Analytics and Data Science funded by Ernesto.Net (under the Algorithms for Good Grant).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Landowska, A. Towards New Mappings between Emotion Representation Models. Appl. Sci. 2018, 8, 274. [Google Scholar] [CrossRef] [Green Version]
  2. Mehrabian, A.; Russell, J.A. An Approach to Environmental Psychology; The MIT Press: Cambridge, MA, USA, 1974. [Google Scholar]
  3. Mehrabian, A. Basic Dimensions for a General Psychological Theory Implications for Personality, Social, Environmental, and Developmental Studies. 1980. Available online: https://philpapers.org/rec/MEHBDF (accessed on 25 September 2021).
  4. Bales, R.F. Social Interaction Systems: Theory and Measurement; Routledge: London, UK, 2017. [Google Scholar]
  5. Ekman, P. Facial Expressions of Emotion: New Findings, New Questions. 1992. Available online: https://psycnet.apa.org/record/1992-26206-001 (accessed on 25 September 2021).
  6. Ekman, P. An argument for basic emotions. Cogn. Emot. 1992, 6, 169–200. [Google Scholar] [CrossRef]
  7. Ekman, P.; Friesen, W.V.; O’sullivan, M.; Chan, A.; Diacoyanni-Tarlatzis, I.; Heider, K.; Krause, R.; LeCompte, W.A.; Pitcairn, T.; Ricci-Bitti, P.E.; et al. Universals and cultural differences in the judgments of facial expressions of emotion. J. Personal. Soc. Psychol. 1987, 53, 712. [Google Scholar] [CrossRef]
  8. Scherer, K.R. What are emotions? And how can they be measured? Soc. Sci. Inf. 2005, 44, 695–729. [Google Scholar] [CrossRef]
  9. Chanel, G.; Kierkels, J.J.; Soleymani, M.; Pun, T. Short-term emotion assessment in a recall paradigm. Int. J.-Hum.-Comput. Stud. 2009, 67, 607–627. [Google Scholar] [CrossRef]
  10. Chenchah, F.; Lachiri, Z. Acoustic emotion recognition using linear and nonlinear cepstral coefficients. Int. J. Adv. Comput. Sci. Appl. 2015, 6, 1–4. [Google Scholar] [CrossRef] [Green Version]
  11. Suja, P.; Tripathi, S. Real-time emotion recognition from facial images using Raspberry Pi II. In Proceedings of the 2016 3rd International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 11–12 February 2016; pp. 666–670. [Google Scholar]
  12. Chanthaphan, N.; Uchimura, K.; Satonaka, T.; Makioka, T. Facial emotion recognition based on facial motion stream generated by kinect. In Proceedings of the 2015 11th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Bangkok, Thailand, 23–27 November 2015; pp. 117–124. [Google Scholar]
  13. Wiem, M.B.H.; Lachiri, Z. Emotion sensing from physiological signals using three defined areas in arousal-valence model. In Proceedings of the 2017 International Conference on Control, Automation and Diagnosis (ICCAD), Hammamet, Tunisia, 19–21 January 2017; pp. 219–223. [Google Scholar]
  14. Zhang, J.; Yin, Z.; Chen, P.; Nichele, S. Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review. Inf. Fusion 2020, 59, 103–126. [Google Scholar] [CrossRef]
  15. Picard, R.W.; Vyzas, E.; Healey, J. Toward machine emotional intelligence: Analysis of affective physiological state. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 1175–1191. [Google Scholar] [CrossRef] [Green Version]
  16. Kumar, J.S.; Bhuvaneswari, P. Analysis of Electroencephalography (EEG) signals and its categorization—A study. Procedia Eng. 2012, 38, 2525–2536. [Google Scholar] [CrossRef] [Green Version]
  17. Kyriakou, K.; Resch, B.; Sagl, G.; Petutschnig, A.; Werner, C.; Niederseer, D.; Liedlgruber, M.; Wilhelm, F.H.; Osborne, T.; Pykett, J. Detecting moments of stress from measurements of wearable physiological sensors. Sensors 2019, 19, 3805. [Google Scholar] [CrossRef] [Green Version]
  18. Goshvarpour, A.; Goshvarpour, A. The potential of photoplethysmogram and galvanic skin response in emotion recognition using nonlinear features. Phys. Eng. Sci. Med. 2020, 43, 119–134. [Google Scholar] [CrossRef]
  19. Goshvarpour, A.; Goshvarpour, A. EEG spectral powers and source localization in depressing, sad, and fun music videos focusing on gender differences. Cogn. Neurodyn. 2019, 13, 161–173. [Google Scholar] [CrossRef]
  20. Dzedzickis, A.; Kaklauskas, A.; Bucinskas, V. Human emotion recognition: Review of sensors and methods. Sensors 2020, 20, 592. [Google Scholar] [CrossRef] [Green Version]
  21. Chunawale, A.; Bedekar, D. Human Emotion Recognition using Physiological Signals: A Survey. 2020. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3645402 (accessed on 25 September 2021).
  22. Boiten, F.A. The effects of emotional behaviour on components of the respiratory cycle. Biol. Psychol. 1998, 49, 29–51. [Google Scholar] [CrossRef]
  23. Jerath, R.; Beveridge, C. Respiratory Rhythm, Autonomic Modulation, and the Spectrum of Emotions: The Future of Emotion Recognition and Modulation. Front. Psychol. 2020, 11, 1980. [Google Scholar] [CrossRef] [PubMed]
  24. Yasuma, F.; Hayano, J.I. Respiratory sinus arrhythmia: Why does the heartbeat synchronize with respiratory rhythm? Chest 2004, 125, 683–690. [Google Scholar] [CrossRef]
  25. Valderas, M.T.; Bolea, J.; Laguna, P.; Bailón, R.; Vallverdú, M. Mutual information between heart rate variability and respiration for emotion characterization. Physiol. Meas. 2019, 40, 084001. [Google Scholar] [CrossRef] [PubMed]
  26. Kreibig, S.D. Autonomic nervous system activity in emotion: A review. Biol. Psychol. 2010, 84, 394–421. [Google Scholar] [CrossRef]
  27. Masaoka, Y.; Sugiyama, H.; Katayama, A.; Kashiwagi, M.; Homma, I. Remembering the past with slow breathing associated with activity in the parahippocampus and amygdala. Neurosci. Lett. 2012, 521, 98–103. [Google Scholar] [CrossRef]
  28. Louis, E.K.S.; Frey, L.; Britton, J.; Hopp, J.; Korb, P.; Koubeissi, M.; Lievens, W.; Pestana-Knight, E. Electroencephalography (EEG): An Introductory Text and Atlas of Normal and Abnormal Findings in Adults, Children, and Infants; American Epilepsy Society: Chicago, IL, USA, 2016. [Google Scholar]
  29. Wu, G.; Liu, G.; Hao, M. The analysis of emotion recognition from GSR based on PSO. In Proceedings of the 2010 International Symposium on Intelligence Information Processing and Trusted Computing, Huanggang, China, 28–29 October 2010; pp. 360–363. [Google Scholar]
  30. Liu, H.; Allen, J.; Zheng, D.; Chen, F. Recent development of respiratory rate measurement technologies. Physiol. Meas. 2019, 40, 07TR01. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Adib, F.; Mao, H.; Kabelac, Z.; Katabi, D.; Miller, R.C. Smart homes that monitor breathing and heart rate. In CHI ’15: Proceedings of the 33rd annual ACM Conference on Human Factors in Computing Systems, Seoul, Korea, 18–23 April 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 837–846. [Google Scholar]
  32. Gong, P.; Ma, H.T.; Wang, Y. Emotion recognition based on the multiple physiological signals. In Proceedings of the 2016 IEEE International Conference on Real-Time Computing and Robotics (RCAR), Angkor Wat, Cambodia, 6–10 June 2016; pp. 140–143. [Google Scholar]
  33. Mirmohamadsadeghi, L.; Yazdani, A.; Vesin, J.M. Using cardio-respiratory signals to recognize emotions elicited by watching music video clips. In Proceedings of the 2016 IEEE 18th International Workshop on Multimedia Signal Processing (MMSP), Montreal, QC, Canada, 21–23 September 2016; pp. 1–5. [Google Scholar]
  34. Hassani, S.; Bafadel, I.; Bekhatro, A.; Al Blooshi, E.; Ahmed, S.; Alahmad, M. Physiological signal-based emotion recognition system. In Proceedings of the 2017 4th IEEE International Conference on Engineering Technologies and Applied Sciences (ICETAS), Salmabad, Bahrain, 29 November–1 December 2017; pp. 1–5. [Google Scholar]
  35. Kumar, C.N.; Shivakumar, G. A Real Time Human Emotion Recognition System Using Respiration Parameters and ECG. In International Conference on Intelligent Human Computer Interaction; Springer: Berlin/Heidelberg, Germany, 2018; pp. 36–45. [Google Scholar]
  36. Wiem, M.B.H.; Lachiri, Z. Emotion recognition system based on physiological signals with Raspberry Pi III implementation. In Proceedings of the 2017 3rd International Conference on Frontiers of Signal Processing (ICFSP), Paris, France, 6–8 September 2017; pp. 20–24. [Google Scholar]
  37. Rattanadoung, K.; Champrasert, P.; Aramkul, S. The emotional state classification using physiological signal interpretation framework. In Proceedings of the 2018 International Conference on Signals and Systems (ICSigSys), Bali, Indonesia, 1–3 May 2018; pp. 79–85. [Google Scholar]
  38. Hameed, R.A.; Sabir, M.K.; Fadhel, M.A.; Al-Shamma, O.; Alzubaidi, L. Human emotion classification based on respiration signal. In Proceedings of the International Conference on Information and Communication Technology, Baghdad, Iraq, 15–16 April 2019; pp. 239–245. [Google Scholar]
  39. Yamuza, M.T.V.; Bolea, J.; Orini, M.; Laguna, P.; Orrite, C.; Vallverdú, M.; Bailón, R. Human emotion characterization by heart rate variability analysis guided by respiration. IEEE J. Biomed. Health Inform. 2019, 23, 2446–2454. [Google Scholar] [CrossRef]
  40. Brown, R.; Ghavami, N.; Adjrad, M.; Ghavami, M.; Dudley, S. Occupancy based household energy disaggregation using ultra wideband radar and electrical signature profiles. Energy Build. 2017, 141, 134–141. [Google Scholar] [CrossRef]
  41. Ghavami, M.; Michael, L.; Kohno, R. Ultra Wideband Signals and Systems in Communication Engineering; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  42. Rana, S.P.; Dey, M.; Siddiqui, H.U.; Tiberi, G.; Ghavami, M.; Dudley, S. UWB localization employing supervised learning method. In Proceedings of the 2017 IEEE 17th International Conference on Ubiquitous Wireless Broadband (ICUWB), Salamanca, Spain, 12–15 September 2017; pp. 1–5. [Google Scholar]
  43. Dudley, S.; Rana, S.; Dey, M.; Brown, R.; Siddiqui, H. Remote Vital Sign Recognition Through Machine Learning Augmented UWB. In Proceedings of the European Conference on Antennas and Propagation, London, UK, 9–13 April 2018; London South Bank University: London, UK, 2018. [Google Scholar]
  44. Siddiqui, H.U.R.; Saleem, A.A.; Brown, R.; Bademci, B.; Lee, E.; Rustam, F.; Dudley, S. Non-invasive driver drowsiness detection system. Sensors 2021, 21, 4833. [Google Scholar] [CrossRef]
  45. Novelda. Novelda X4. Available online: https://novelda.com/x4-soc.html (accessed on 25 September 2021).
  46. Laonuri. X4M300 Datasheet. Available online: http://laonuri.techyneeti.com/wp-content/uploads/2019/02/X4M300_DATASHEET.pdf (accessed on 25 September 2021).
  47. Kim, D.H. Lane detection method with impulse radio ultra-wideband radar and metal lane reflectors. Sensors 2020, 20, 324. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Novelda. X4-Datasheet. Available online: https://novelda.com/content/wp-content/uploads/2021/01/NOVELDA-x4-datasheet-revF.pdf (accessed on 25 September 2021).
  49. Lee, Y.; Park, J.Y.; Choi, Y.W.; Park, H.K.; Cho, S.H.; Cho, S.H.; Lim, Y.H. A novel non-contact heart rate monitor using impulse-radio ultra-wideband (IR-UWB) radar technology. Sci. Rep. 2018, 8, 1–10. [Google Scholar] [CrossRef]
  50. Jeger-Madiot, N.; Gateau, J.; Fink, M.; Ing, R.K. Non-contact and through-clothing measurement of the heart rate using ultrasound vibrocardiography. Med. Eng. Phys. 2017, 50, 96–102. [Google Scholar] [CrossRef] [PubMed]
  51. Quintana, D.S.; Elstad, M.; Kaufmann, T.; Brandt, C.L.; Haatveit, B.; Haram, M.; Nerhus, M.; Westlye, L.T.; Andreassen, O.A. Resting-state high-frequency heart rate variability is related to respiratory frequency in individuals with severe mental illness but not healthy controls. Sci. Rep. 2016, 6, 1–8. [Google Scholar]
  52. Ahmed, A.; Harness, J.; Mearns, A. Respiratory control of heart rate. Eur. J. Appl. Physiol. Occup. Physiol. 1982, 50, 95–104. [Google Scholar] [CrossRef]
  53. Tiinanen, S.; Kiviniemi, A.; Tulppo, M.; Seppänen, T. RSA component extraction from cardiovascular signals by combining adaptive filtering and PCA derived respiration. In Proceedings of the 2010 Computing in Cardiology, Belfast, UK, 26–29 September 2010; pp. 73–76. [Google Scholar]
  54. Kircher, M.; Lenis, G.; Dössel, O. Separating the effect of respiration from the heart rate variability for cases of constant harmonic breathing. Curr. Dir. Biomed. Eng. 2015, 1, 46–49. [Google Scholar] [CrossRef]
  55. Yeh, S.T. Using trapezoidal rule for the area under a curve calculation. In Proceedings of the 27th Annual SAS® User Group International (SUGI’02), Orlando, FL, USA; 2002. [Google Scholar]
  56. Rustam, F.; Reshi, A.A.; Ashraf, I.; Mehmood, A.; Ullah, S.; Khan, D.M.; Choi, G.S. Sensor-based human activity recognition using deep stacked multilayered perceptron model. IEEE Access 2020, 8, 218898–218910. [Google Scholar] [CrossRef]
  57. Reshi, A.A.; Rustam, F.; Mehmood, A.; Alhossan, A.; Alrabiah, Z.; Ahmad, A.; Alsuwailem, H.; Choi, G.S. An Efficient CNN Model for COVID-19 Disease Detection Based on X-Ray Image Classification. Complexity 2021, 2021, 6621607. [Google Scholar] [CrossRef]
  58. Rupapara, V.; Rustam, F.; Shahzad, H.F.; Mehmood, A.; Ashraf, I.; Choi, G.S. Impact of SMOTE on Imbalanced Text Features for Toxic Comments Classification using RVVC Model. IEEE Access 2021, 9, 78621–78634. [Google Scholar] [CrossRef]
  59. Rustam, F.; Mehmood, A.; Ullah, S.; Ahmad, M.; Khan, D.M.; Choi, G.S.; On, B.W. Predicting pulsar stars using a random tree boosting voting classifier (RTB-VC). Astron. Comput. 2020, 32, 100404. [Google Scholar] [CrossRef]
Figure 1. Steps carried out in the research.
Figure 1. Steps carried out in the research.
Sensors 21 08336 g001
Figure 2. The X4m300 IR-UWB radar used for data collection.
Figure 2. The X4m300 IR-UWB radar used for data collection.
Sensors 21 08336 g002
Figure 3. UWB radar signals during chest movement.
Figure 3. UWB radar signals during chest movement.
Sensors 21 08336 g003
Figure 4. Subject sitting in front of radar while watching videos.
Figure 4. Subject sitting in front of radar while watching videos.
Sensors 21 08336 g004
Figure 5. Steps involved in data cleaning and acquiring respiration data.
Figure 5. Steps involved in data cleaning and acquiring respiration data.
Sensors 21 08336 g005
Figure 6. Architecture of the deep learning models used for experiments.
Figure 6. Architecture of the deep learning models used for experiments.
Sensors 21 08336 g006
Figure 7. The architecture of ensemble models, (a) Soft voting, and (b) Hard voting.
Figure 7. The architecture of ensemble models, (a) Soft voting, and (b) Hard voting.
Sensors 21 08336 g007
Figure 8. Architecture of the proposed methodology.
Figure 8. Architecture of the proposed methodology.
Sensors 21 08336 g008
Figure 9. Data processed for RPM estimation, (a) Noisy data, and (b) Clean data.
Figure 9. Data processed for RPM estimation, (a) Noisy data, and (b) Clean data.
Sensors 21 08336 g009
Figure 10. Detected peaks from the cleaned data.
Figure 10. Detected peaks from the cleaned data.
Sensors 21 08336 g010
Figure 11. Pulse oximeter used for measuring RPM.
Figure 11. Pulse oximeter used for measuring RPM.
Sensors 21 08336 g011
Figure 12. Feature importance of dataset attributes using RF.
Figure 12. Feature importance of dataset attributes using RF.
Sensors 21 08336 g012
Table 1. List of hyperparameters used for experiments.
Table 1. List of hyperparameters used for experiments.
ClassifierHyperparameters
ETCn_estimators = 200, random_state = 100, max_depth = 200, min_samples_split = 80
ADBn_estimators = 50, random_state = 100, learning_rate = 1.0
GBMmax_depth = 100, n_estimators = 100, random_state = 42, min_samples_split = 90, min_samples_leaf = 10
KNNn_neighbors = 7, leaf_size = 1
XGBn_estimators = 50, max_depth = 100, learning_rate = 0.8
HVBase learners = XGB, ADA, GBM, KNN, voting = hard/majority
SVBase learners = XGB, ADA, GBM, KNN, voting = soft
Table 2. Results for validation experiments for RPM.
Table 2. Results for validation experiments for RPM.
SubjectRespiration Rate
Pulse OximeterProposed Method
Participant 11616
Participant 11919
Participant 22121
Participant 22222
Participant 31515
Participant 31212
Participant 41515
Participant 41717
Participant 51819
Participant 51616
Participant 61818
Participant 62020
Participant 71717
Participant 71919
Participant 81717
Participant 81515
Participant 91212
Participant 91414
Participant 101818
Participant 101717
Table 3. Results for validation experiments for RPM in dynamic environment.
Table 3. Results for validation experiments for RPM in dynamic environment.
SubjectRespiration Rate
Pulse OximeterProposed Method
Participant 11514
Participant 11716
Participant 21616
Participant 21415
Participant 31717
Participant 32019
Participant 41616
Participant 41817
Participant 51415
Participant 52019
Participant 61718
Participant 61617
Participant 71817
Participant 71719
Participant 81617
Participant 81414
Table 4. Statistical T test to validate the RPM results.
Table 4. Statistical T test to validate the RPM results.
Statistical T TestStatic/DynamicMale/Female
df3030
cv1.6971.697
p-value0.9200.945
t-statistic0.9200.069
alpha0.050.05
Table 5. Average RPM of males and females During different emotions.
Table 5. Average RPM of males and females During different emotions.
GenderAverage RPM
HappinessDisgustFear
Male19.5619.3519.85
Female18.4719.3820.54
Table 6. Sample records from the collected dataset.
Table 6. Sample records from the collected dataset.
SubjectRPMGenderAgeEmotion
Participant 1201260
Participant 2201270
Participant 3210270
Participant 4221241
Participant 5230261
Participant 6201301
.....
.....
.....
Table 7. Performance metrics with 90:10 train and test size.
Table 7. Performance metrics with 90:10 train and test size.
ClassifierAccuracyPrecisionRecallF1 Score
ETC0.700.670.680.66
ADB0.700.690.700.69
GBM0.790.690.700.69
KNN0.730.590.610.59
XGB0.690.690.690.69
EHV0.770.660.660.66
ESV0.740.720.730.72
Table 8. Performance metrics with 80:20 train and test size.
Table 8. Performance metrics with 80:20 train and test size.
ClassifierAccuracyPrecisionRecallF1 Score
ETC0.680.680.680.68
ADB0.680.680.710.68
GBM0.780.710.730.71
KNN0.720.680.680.68
XGB0.730.750.730.74
EHV0.750.710.730.71
ESV0.730.760.770.76
Table 9. Performance metrics with 70:30 train and test size.
Table 9. Performance metrics with 70:30 train and test size.
ClassifierAccuracyPrecisionRecallF1 Score
ETC0.670.670.690.68
ADB0.700.720.730.72
GBM0.780.720.720.72
KNN0.700.720.710.72
XGB0.720.720.720.72
EHV0.730.730.730.73
ESV0.740.740.740.74
Table 10. Performance metrics with 60:40 train and test size.
Table 10. Performance metrics with 60:40 train and test size.
ClassifierAccuracyPrecisionRecallF1 Score
ETC0.670.670.710.67
ADB0.680.680.700.68
GBM0.780.730.740.73
KNN0.710.630.630.63
XGB0.680.690.680.68
EHV0.710.710.740.71
ESV0.750.710.710.71
Table 11. Performance Metrices with 50:50 train and test size.
Table 11. Performance Metrices with 50:50 train and test size.
ClassifierAccuracyPrecisionRecallF1 Score
ETC0.630.650.630.63
ADB0.680.680.680.67
GBM0.700.700.700.70
KNN0.530.550.530.52
XGB0.670.670.670.67
EHV0.720.620.620.61
ESV0.720.710.720.71
Table 12. Performance of machine learning models with 10-fold cross validation.
Table 12. Performance of machine learning models with 10-fold cross validation.
ClassifierAccuracy (±Std. Dev.)
ETC0.64 (±0.35)
ADB0.61 (±0.37)
GBM0.65 (±0.36)
KNN0.57 (±0.22)
XGB0.64 (±0.40)
EHV0.65 (±0.29)
ESV0.66 (±0.33)
Table 13. Performance of deep learning models with different train-test split ratios.
Table 13. Performance of deep learning models with different train-test split ratios.
RatioClassifierAccuracyPrecisionRecallF1 Score
90:10MLP0.620.650.620.60
CNN0.400.510.400.42
80:20MLP0.670.670.670.66
CNN0.550.620.550.54
70:30MLP0.620.620.620.61
CNN0.400.410.400.40
60:40MLP0.630.640.630.61
CNN0.380.530.380.40
50:50MLP0.720.620.620.61
CNN0.400.510.400.42
Table 14. Performance comparison with different emotion detection approaches.
Table 14. Performance comparison with different emotion detection approaches.
ReferenceFeature VectorAccuracy
[33]Respiration rate interval, low frequency, heart rate, high frequency, RSA power, RSA frequency, and RSA amplitude breathing frequency, breathing amplitude, RSA amplitude, ratio to respiratory oscillation, respiratory and RSA frequency difference, the phase difference of respiration and RSA, the slope of phase difference, and standard deviation.73% for liking, 72% for arousal, and 70% for valence
[35]Root-mean-square, intrinsic mode functions, ‘mean’, ‘max’, from respiration and ECG signal Respiration rate and heart rate80%
[36]Heart rate variability from ECG signal, respiration rate and amplitude from respiration signal.68.5% for arousal, 68.75% for valence
[37]Statistical features average, maximum, minimum, and standard deviation etc.55.45% for arousal, 59% for valence
[38]Air flow rate and volume80%
[39]Different statistical and time domain features79.2% for relax vs. joy, 77.8% for joy vs. sad, and 77.3% for joy vs. anger
Current studyRPM, Age and, Gender79%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Siddiqui, H.U.R.; Shahzad, H.F.; Saleem, A.A.; Khan Khakwani, A.B.; Rustam, F.; Lee, E.; Ashraf, I.; Dudley, S. Respiration Based Non-Invasive Approach for Emotion Recognition Using Impulse Radio Ultra Wide Band Radar and Machine Learning. Sensors 2021, 21, 8336. https://doi.org/10.3390/s21248336

AMA Style

Siddiqui HUR, Shahzad HF, Saleem AA, Khan Khakwani AB, Rustam F, Lee E, Ashraf I, Dudley S. Respiration Based Non-Invasive Approach for Emotion Recognition Using Impulse Radio Ultra Wide Band Radar and Machine Learning. Sensors. 2021; 21(24):8336. https://doi.org/10.3390/s21248336

Chicago/Turabian Style

Siddiqui, Hafeez Ur Rehman, Hina Fatima Shahzad, Adil Ali Saleem, Abdul Baqi Khan Khakwani, Furqan Rustam, Ernesto Lee, Imran Ashraf, and Sandra Dudley. 2021. "Respiration Based Non-Invasive Approach for Emotion Recognition Using Impulse Radio Ultra Wide Band Radar and Machine Learning" Sensors 21, no. 24: 8336. https://doi.org/10.3390/s21248336

APA Style

Siddiqui, H. U. R., Shahzad, H. F., Saleem, A. A., Khan Khakwani, A. B., Rustam, F., Lee, E., Ashraf, I., & Dudley, S. (2021). Respiration Based Non-Invasive Approach for Emotion Recognition Using Impulse Radio Ultra Wide Band Radar and Machine Learning. Sensors, 21(24), 8336. https://doi.org/10.3390/s21248336

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop