Introduction
Related Works
Stress Prediction and Heart Rate Variability
Artificial Intelligence Algorithms
Rule-Based Approaches
Shallow Machine Learning Approaches
Deep Machine Learning Approaches
Summary of AI Algorithms
Type | Pros | Cons | Description |
---|---|---|---|
FL | • FL models are flexible. • handle distorted, imprecise or vague data. | • The accuracy is not optimal on imprecise inputs. | Incorporates uncertainty in real-world problems. Boolean logic recognises only two states, whereas fuzzy logic can replicate human thinking considering more than two states, defined as membership degrees |
FNN | • Combines the strengths of both NN and fuzzy Logic. • Allows integration of expert knowledge into the system. | • The structure of a fuzzy NN is not completely interpretable. • Not a concurrent system. | Blends fuzzy control systems’ high-level, IF-THEN rule logic with neural networks’ low-level learning and computing capabilities [71]. This algorithm resolves the deficit of flexibility in fuzzy number decision-making and the selection of fuzzy shapes which are able to represent expert experiences correctly [72] |
NB | • Does not require that much data be trained. • Accepts both discrete and continuous-valued datasets. | • Assumes independent predictors, which is impossible in real-life data. • Has zero frequency issue in case of categorical variables. | |
LR | • Can classify new data from the continuous and discrete dataset. • Simple to implement and train. | • Cannot resolve nonlinear problems. • Only maps result in 0-1 Value. | Capable of using any actual value number and mapping it into the 0-1 value |
DT | • Consider all possible outcomes. • Can solve both classification and regression-based problems. | • Small changes lead to structural changes. • Comparatively less accurate than other tree-based algorithms. | Constructs a model that predicts the target variable. The problem is solved using a tree representation, in which the leaf node corresponds to a class label and the core node of the tree expresses features [75] |
KNN | • Can solve both regression and classification issues. • Does not require linear separability between classes. | • Suspectable to inaccurate feature variables. • Does not perform well in big data. | |
SVM | • Can conduct nonlinear classification. • Transfers input to high dimensional feature space. | • Has large computational cost. • Training time is high. | Seeks to identify an ideally separating hyperplane between classes using the structural risk minimisation principle. Determines the optimal decision boundary for categorising objects from the training data set [74] |
KMC | • Can easily adapt to new data. • Better in solving classification tasks. | • Requires exact values of clusters. • Data grouping can be unstable. | An iterative clustering approach that classifies unlabelled datasets into different clusters by splitting from set data into cluster K. The initial value of the cluster’s centre point influences cluster identification |
RF | • Lower risk of overfitting. • Work on both regression and classification problems. | • Slower in computing. • Not for real-time problem-solving. | |
RNN | • Works in a feedback loop. • Can perform speech recognition and video classification. | • Comparatively unstable. • Not possible to stack up. • Training time is slower. | Conventionally constructed by defining a transition and output function [79]. By assigning identical weights and biases to all of the layers, creates dependent activations |
DNN | • Can do multi-level processing. • Can work with complicated feature representation. | • Slow learning process. • Requires a large amount of data. | DNN has multiple hidden layers in between the input and output layers. It has the ability to extract features from data without any feature extraction algorithms. DNNs mimic the multilevel processing mechanism of vision by Cortical areas of the brain [80] |
CNN | • Little reliance on pre-processing. • Provide more accurate results. | • Slower in training than DNNs. • Training set needs to be much larger for successful training. |
AI for Stress Prediction
Rule-Based Approach
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[85] | Fuzzy theoretic nonparametric deep model | - | PPG | R-R Features |
[83] | Mamdani fuzzy | Downsampling | Ohmeda 2300 Finapress, Gazepoint GP3 eye tracker | HRV2, mPD |
[86] | Sugeno neuro fuzzy | Wavelet transformation, noise removal | ECG | Time-frequency features |
[87] | Sugeno fuzzy clustering | Continuous wavelet transformation | Polar S810i | HR, Mw, 1/a, p1,p2,p3 |
[84] | Fuzzy ARTMAP | Dimensionality reduction, normalisation | ECG, microtiter plate spectrophotometer | Alpha amylase, cortisol, R-R intervals |
Shallow Machine Learning Approaches
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[88] | SVM, KNN | - | Kinect 3D sensor, ECG sensor | RMSSD, AVNN, SDANN, SDNN, NN50, PNN50, LF, HF, LF/HF |
[89] | SVM, KNN, NB, LR | - | ECG | NN.Mean, PNN50, rMSSD, TP, LF, HF, VLF, LF/HF |
[90] | k-NN, SVM, DT, NB | SG-filter, BWF, ANC algorithm | GSR, BVP, ST, 3-D ACC, HR | Time-domain, frequency-domain, and distribution features of BVP, ACC, GSR, and HR |
[75] | KNN, SVM, NB, DT and DNN | - | Biometric Sensors | Statistical Features of RR |
[91] | SVM | FIR filters, IIR filters, EMD and DWT | ECG v.12 devices, SX230 surface electrode, Skintact F-55 electrode | Statistical and time-domain features of RR |
[92] | SVM | Canny’s edge detection algorithm, LPFr | Pulse oximeter, ST, ECG, and eye tracker | HRV, HR, PS, temperature, SpO2 |
[93] | SVM, C4.5 DT | 3-sigma rule | ECG, Polar H7 chest strap | Statistical features of RR; frequency-domain features of RR |
[94] | RF, SVM | time series polynomial fit and bandpass filtered | ECG | Time-domain features of RR, HRV triangular index, ECG envelope, frequency-domain features of RR |
[95] | NB, SVM, MLP, AB, C4.5 DT | QRS detector, PhysioNet’s WAVE | ECG | Statistical features of RR, absolute power, frequency-domain features of RR, SampEn, D2, fa1, dfa2, ShanEn |
[96] | BN, SVM, k-NN, C4.5 DT | Lomb-Scargle algorithm, LPF, CDA | Zephyr BioHarness34, Shimmer3 GSR + Development Kit5 | Statistical features of RR, tot spectrum power, frequency-domain features of RR, Amps, ISCR, mean SCL, mean EDA, max EDA deflection |
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[97] | SVM; LR; RF | PPG-PD; HR; RR-ISF; HRV; EDA | DA wrist, PPG, Spare, TVOC, CO2, TEMP | Statistical features of RR, SCR, SCL; RiseT |
[98] | SVM; KNN; EnL | BCG processing | EMFi sensor | Statistical features |
[99] | NB; DT | Normalization and transformation | Fitbit Tracker | Number of calls, duration, no. of SMS, the app usage information |
[100] | PCA, SVM, KNN, LR, RF, MLP | Artefact-(interpolation/removal) | 3D accelerometer (ACC), PPG, EDA | Statistical and frequency-domain features of RR |
[101] | RF | Filtering using BWBPF | Wristband device, Polar H10 | SD1, SD2, RMSSD, SDNN, MHR, MRRI, TP, VLP, LF, HF |
[102] | NB, SVM, KNN, Bayes Net, RF; DT; MLP | - | ECG and skin conductance | Facial expressions, head orientation action units, emotion, body postures, joint angles, HR, HRV, skin conductance |
[103] | KNN | Artefacts and noise removal | HR, GSR and BT | RR, GSR, BT |
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[104] | KNN, SVM | HR, accelerometer, EDA, Resp. Rate, SpO2 | Statistical features of HR and RR, mean SpO2 | |
[105] | SVM, DT | RR-SE and Corr | ECG | Statistical features of NN, dfa1, dfa2, RPlmean, RPlmax, REC, RPadet, ShanEn |
[106] | NN, KNN, DA, NB | Corr | PPG | mHR, RR, SDHR, SDRR, CVRR, RMSSD, pRR20, and pRR50. ULF, VLF, LF, HF |
[107] | SVM | BPF (5–25 Hz) | BioHarness 3, Zephyr | Time-domain and frequency-domain features of HRV |
[108] | RF classifier | Normalising, BWF | GSR, ECG, Resp | Time-domain and frequency-domain features of GSR respiration: the mean and variance time-domain and frequency-domain features of HRV |
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[109] | RF, KNN, SVM, and XGB | LS-method | ECG, EMG, PPG | Time-domain and frequency-domain features of HRV |
[110] | KNN, SVM, MLP, RF, and GB | WQRS tool PhysioNet HRV toolkit, pyhrv, data normalization | Apple watch | HR, AVNN, SDNN, RMSSD, pNN50, TP, and VLF |
[111] | SVM | Downsampling, noise removal | BioNomadix module from Biopac, model BN-PPGED | ppgt, ppgaut, HRV t, EDA t |
[112] | SVM, KNN | Filtering, derivative, SW integration, QRS detection | ECG sensor | Time-domain and frequency-domain features of RR |
Deep Machine Learning Approaches
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[113] | LVQ | R-PD, IBI outlier removal, and SCRG method | ECG, GSR, RSP | Time-domain, frequency-domain, and distribution features of ECG, GSR, RSP |
[114] | LSTM-RNN | All features are normalised into a range of [-1,1] | A Tizen component on smart-watch | HR avg, HRV arg, HR min, HR max, SDHR, SDHRV, hr diff avg, hr diff var |
[115] | CNN-LSTM | A Butterworth band-pass filter (5–15 Hz), R-peaks are extracted, Pan-Tompkins algorithm | ECG | Mean, standard deviation, mean of the first difference of HRV, average normal-to-normal (NN) and intervals, SDNN, RMSSD, PNN50 |
[116] | FFNN | DWT, Pan-Tompkins algorithm, down sampling | CVDiMo wearable sensor | Statistical and frequency-domain features of RR |
[117] | LSTM | Noise filtering | The bio beam band | Statistical and frequency-domain features of RR |
[118] | CNN | Bandpass filter | ECG | HR, LH, SDNN, SD2, pQ |
Ref. | Model | Pre-processing | Sensors | Features |
---|---|---|---|---|
[119] | NN | - | Pulse oximeter | MEAN, SDNN, SDANN, RMSSD, TF, VLF, LF, HF, LF/HF |
[120] | BPNN, SVM, KNN | WD and HP, LP, and RMS filtering | ECG, EDA, EMG, respiration sensors | AVHR, LF/HF, Yrm, MF, SC mean, respiration |
[121] | ANN | WT | ECG probe, BIOPAC ECG100C | Time-domain and frequency-domain features of NN |
[122] | LSTM, DNN | Transfer function, biosppy and pyhrv libraries | Polar H10, Actigraph wGT3X-BT | HR and HRV features |
[123] | NP, SVC, KNN | HT algorithm, BWF | ECG, GSR | Time-domain and frequency-domain features of HR, time-domain features of GSR |
[124] | LR, NB, NN, SVM, RF, KNN | - | PPG | Mean RR, Min RR, Max RR, Median RR, SDNN, RMSSD, pNN50 |
Performance Analysis and Discussion
Rule-Based Approaches
Ref. | Models | Dataset | Evaluation metrics | Performance |
---|---|---|---|---|
[85] | Fuzzy theoretic nonparametric deep model | Private/100 subjects | RMSE | PCC: 0.8162 (old dataset) vs 0.6809 (new dataset), RMSE: 6.8382 (old dataset) vs 9.4872 (new dataset) |
[83] | Mamdani fuzzy model | Private/3 subjects for training and 2 subjects for testing | EI | EI = 2.9412 (Subject 1), 1 (Subject 2) |
[86] | Sugeno neuro-fuzzy model | Private/20 subjects | - | Established a direct functional relationship between heart rate variability and mental stress |
[87] | Sugeno fuzzy clustering | Private/26 males, 12 females, aged 18–29 years | - | Minimised the worst-case influence of uncertainty on fuzzy parameter identification performance |
[84] | Fuzzy ARTMAP | Private/22 subjects | Acc | Acc = 80% (ECG characteristics),75% (salivary alpha-amylase) |
Shallow Machine Learning Approaches
Ref. | Models | Dataset | Evaluation metrics | Performance |
---|---|---|---|---|
[88] | SVM and KNN | SWELL-KW | Acc | Acc of 0.9275 |
[89] | SVM, KNN, NB, and LR | Private/35 participant (mean age of 23 ± 4 years and a male-to-female ratio of 1:1.3) | AUC | Acc = 0.755 AUC = 0.74 |
[90] | NB, J48, RF and bagging | Private/8 participants | Acc | prediction Acc = 0.857 |
[75] | KNN, SVM, DT, NB | Private/34 participants | Acc | Acc = 0.991 |
[91] | SVM | Private/34 students (23 females and 11 males, aged 20–37 years) | Acc | The accuracies two level = 1.0, three level = 0.976, and four levels were 0.962 |
[92] | SVM | Private/50 participants (33 men & 17 women) | AUC, variable collection cost | AUC = 0.994, Vcost = 16 |
[93] | SVM, C5 DT | Private | Pre, Rec f1, Acc | 88% Acc, all |
[94] | RF, SVM | Private/24 participants ageing 47.3±9.3 years | Acc | Acc = 0.844 |
[95] | NB, SVM, MLP, AB, Dt C4.5 | Private/42 participants | Sen, Spe and Acc | Sen=0.78, Spe=0.80 and Acc rate=0.79 |
[96] | BN, SVM, k-NN, C4.5 DT, AB | Private/9 older adults | Acc; Pre; Rec; AUPRC | RF (Acc =87.0%; Pre =92.4%; Rec=88.2%; AUPRC =0.97) and AB (Acc=88.2%; Pre =92.3%; Rec=92.0%; AUPRC =0.92) |
Ref. | Models | Dataset | Evaluation metrics | Performance |
---|---|---|---|---|
[97] | SVM, LR, RF | Private | Acc | 80% accuracy for HRV features in baseline and about 77% for HRV and EDA simultaneous features |
[98] | SVM, KNN, and EnL | 15 office workers (five female, five males, age: 31 ± 5.3) | Acc | Accuracies of up to 91% |
[99] | NB, DT | 35 young adults | Sen, Spe, Acc, Pearson’s correlation | NB classifier has 72% accuracy |
[100] | PCA, SVM, KNN, LR, RF, MLP | 21 participants (18 males and 3 females with an average age of 20) | Acc, f-Measure, Pre, Rec | Obtained 92.15% accuracy maximum three-level classification |
[101] | RF | 6 healthy participants ages 21–40 years old | Acc | 10-fold accuracy of stress state is 98%, and F1-score reaches 80% |
[102] | NB, SVM, KNN, BN, RF, DT, MLP | 25 participants (8 female, average age 25, stdv 3.25) | Acc | Best results were obtained with an SVM (RBF kernel): 90.0298% |
[103] | KNN | Private/10 young subjects (mean age 24; 5 female) | Acc, Error | Acc = 0.845, misclassification Error = 0.26 |
[104] | KNN, SVM | Private | Acc | The overall classification accuracy of this system is 96% |
Ref. | Models | Dataset | Evaluation metrics | Performance |
---|---|---|---|---|
[105] | SVM, DT | ECG from 42 healthy subjects (19 female, 23 male) | AUC, Sen, Spe, Acc | Achieved good performance accuracy above 88% |
[106] | NN, KNN, DT, NB | 41 students | AUC, ROC, Acc, confusion matrix | - |
[107] | SVM | Data was collected from 27 (6 females) police | BAC | In HRV segments BAC =0.579 for 300 s window duration |
[108] | RFr | Private | Acc | Acc = 94% |
[109] | RF, KNN, SVM, XGB | RML, WESAD | Acc., precision, recall, F1-Score | Acc. = 66.6 (RML dataset), Acc. = 72.7 (WESAD dataset) |
[110] | KNN, SVM, MLP, RF, GB | PhysioNet | AUROC | MLP, RF and GB yielded an AUROC of 83%, 85%, and 85%, respectively |
[111] | SVM | Private/5 participants aged 18 to 39 | Acc, Pre | Best Acc and Pre for P3:83.08(Acc) & 83.87(Pre) |
[112] | SVM-RBF, KNN | DRIVEDB | Acc | Acc = 83% |
Deep Machine Learning Approaches
Ref. | Models | Dataset | Evaluation metrics | Performance |
---|---|---|---|---|
[113] | LVQ | Private/61 (20 male, 41 female) | Acc | RSP = 0.71, ECG = 0.834, Acc = 0.877 |
[115] | LSTM-RNN | Private/a group of Vietnamese students | Framework proposed | - |
[114] | CNN-LSTM | Private/27 participants aged 21–40 years (55% male) | Acc Sen Spe: Pre | Acc: 0.928, Sen: 0.9413, Spe: 0.9737 and Pre: 0.95 |
[116] | FFNN | CVDiMo/conducted with 30 volunteers | AUC, Acc | AUC = 0.978, Acc= 0.92 |
[117] | LSTM | Private/for trial-1: 91 participants (62% female, 38% male); for trial-2: 600 (72% female, 28% male) | ACC | Acc = 0.85 |
[118] | CNN | Private/20 healthy subjects, aged from 18 to 35 | ER, FAR, and FRR | CNN ER=17.3 FAR= 0.01 FRR = 32.1 |
[119] | NN | 56 samples | Acc | 93.75% accuracy |
[120] | BPNN, SVM, KNN | 18 right-handed, healthy individuals, 20.1 ± 0.94 years | Acc, Rec, Pre | Accuracy can reach 96.4% and 78.3% |
[121] | ANN | Private/57 participants, and all are above 65 years | acc | Acc = 90.83% |
[122] | LSTM, DNN | Private/15 gamers Age of 10 to 22 | acc | Acc = 64% (DNN) vs 92% (LSTM) |
[123] | LR, SVM, KNN | Drivedb, WESAD | Avg precision, AUC | WESAD: 0.957 average precision, same-participant vs 0.780 other-participant, drivedb: 0.804 vs 0.757 |
[124] | LR, NN, NB, SVM, RF, and KNN | Private/63 (76.8%) were female, and 19 (23.2%) were male aged 17 to 38 years | Sen, Spe | For Model 1, Sen = 0.752 and Spe = 0.779. For Model 2, Sen = 0.742 and Spe = 0.781 |