Abstract
A wearable system for the personalized EEG-based detection of engagement in learning 4.0 is proposed. In particular, the effectiveness of the proposed solution is assessed by means of the classification accuracy in predicting engagement. The system can be used to make an automated teaching platform adaptable to the user, by managing eventual drops in the cognitive and emotional engagement. The effectiveness of the learning process mainly depends on the engagement level of the learner. In case of distraction, lack of interest or superficial participation, the teaching strategy could be personalized by an automatic modulation of contents and communication strategies. The system is validated by an experimental case study on twenty-one students. The experimental task was to learn how a specific human-machine interface works. Both the cognitive and motor skills of participants were involved. De facto standard stimuli, namely (1) cognitive task (Continuous Performance Test), (2) music background (Music Emotion Recognition—MER database), and (3) social feedback (Hermans and De Houwer database), were employed to guarantee a metrologically founded reference. In within-subject approach, the proposed signal processing pipeline (Filter bank, Common Spatial Pattern, and Support Vector Machine), reaches almost 77% average accuracy, in detecting both cognitive and emotional engagement.
Subject terms: Attention, Electrical and electronic engineering
Introduction
Man’s relationship with knowledge is increasingly mediated by technology. Since the second half of the last century, the digital era, namely the period of the pervasive use of information and communication technologies in every area of life, has had a major impact on the human learning1. Currently, the ongoing Fourth Industrial Revolution (Industry 4.0) further expands the role of technology in learning processes: automated teaching platforms can adapt in real-time to the user skills and the new generation interfaces allow multi-sensorial interactions with virtual contents2–4. In the pedagogical domain, the concept of “Learning 4.0” is emerging and it is not just a marketing gimmick5. The 4.0 technologies are strongly impacting on the creation, the conservation, and the transmission of knowledge6. In particular, the new immersive eXtended Reality (XR) solutions make possible to achieve embodied learning by enhancing the catalytic learning role of bodily activities7. Furthermore, wearable transducers and embedded Artificial Intelligence (AI) increase real-time adaptivity in Human-Machine Interaction8. In detail, in the Learning 4.0 context, the adaptation between humans and machines is reciprocal: the subject learns to use the human-machine interface, but also the machine adapts to human by learning from her/him9.
Traditionally, learning how to use a new technology interface was a once-in-a-lifetime effort conducted at a young age. For many people this has occurred with learning to read and write. Recently, the rapidity of technological evolution has been entailing the need to learn how to use several interfaces. The joy-pad, icon, touch/multi-touch screen, speech and gesture recognition are examples of the evolution of new interfaces (hardware and software components).
More specifically, learning to use an interface is a hard task which requires complex cognitive-motor skills. When human beings learned to use the mouse/touchscreen, as well as when they learned to write, read or speak, their minds learned complex cognitive-body patterns10,11. Regarding the human-machine interfaces of older generations, the user was autonomously required to explore the different available resources and learn their use. Currently, the Interfaces 4.0 can adapt in real time to the use, by supporting the learning process2,3,12. Proper adaptive strategies could be aimed to improve the learner engagement. Indeed, according to the literature, the effectiveness learning process depends on the engagement level of the learner. In a study on the role of learning engagement in technology-mediated learning13 and its effects on learning effectiveness, 212 university students were observed during learning Adobe Photoshop. It was reported how learning materials can negatively affect learning engagement, which in turn reduces the perceived learning effectiveness and satisfaction. The role of e-learning on learning engagement and its effectiveness was evaluated in a study on 181 students reporting higher academic results in case of learning engagement14. Therefore, the engagement monitoring is a fundamental aspect allowing the machine to adapt to the user.
In this context, Engagement stands for concentrated attention, commitment, and active involvement, in contrast to apathy, lack of interest or superficial participation15,16. In the learning context, Newman defines engagement as: “the student’s psychological investment in and effort directed toward learning, under-standing, or mastering the knowledge, skills, or crafts that academic work is intended to promote”17,18. Moreover, Frederiks defines the student engagement as a meta-construct including behavioral, emotional, and cognitive engagement19.
As concerns the engagement measurability, evaluation grids and self-assessment questionnaires (to be filled out by the observer or by the learner autonomously) are traditionally the most used methods for the behavioral, cognitive, and emotional engagement detection20. In recent years, measures based on biosignals are spreading very rapidly. Furthermore, the use of physiological sensors allows the adoption of real-time machine adaptive strategies, by detecting cognitive and emotional engagement. Among the different physiological biosignals, the EEG appears to be one of the most promising technology thanks to its low cost, low invasiveness, and high temporal resolution21–23. Moreover, the EEG contains a broader range of information about the mental state of a subject with respect to others biosignals24. In25 an engagement index was proposed to decide when to use the autopilot and when to switch to the manual control during a fly simulator session. The engagement index was where and are the EEG frequency bands. This index was used as engagement estimator also in learning contexts20,26,27. However, the proposed index does not take into account the different engagement types (i.e., cognitive, emotional and behavioural) proposed by the theories previously reported.
In this study, a method for EEG-based cognitive and emotional engagement detection during learning activities is proposed. High wearability is guaranteed by a low number of dry electrodes. This property allows the cognitive and emotional learning engagement detection in daily life applications. Furthermore, the proposed method can be used also in traditional school contexts. For example, acquiring cognitive and emotional engagement data during the lessons can provide (1) real-time feedbacks to the teacher, for maximizing class engagement, and (2) student engagement trends over the time that can be used for academic program adaptation to individuals or to the whole class.
This work is organized as follows: in “Background” Section, a background on the engagement in the learning context is reported. In “Proposal” Section, the basic ideas and the proposed solution are described. Then, in “Methods” Section, the methods are presented. Finally, the experimental results are discussed in “Experimental Results” Section.
Background
In general, learning a new interface can be traced back to a classic learning problem. In the constructivism framework, learning consists in the construction of the schemes: units of knowledge, each relating to different aspect of the world, including actions, objects, and abstract concepts28. When a subject learns a specific pattern, the neuroplasticity process is activated modifying the neural brain structure29. Once the process is learned, the brain builds a myelinated axon connection system to automate that. The adjacent neurons fire in unison, and more the experience or operation is repeated, more the synaptic link between neurons becomes strong30. The automated use of all mental processes as well as the understanding and use of new technologies occurs through the creation of synaptic pathways31–33. For instance, Markham et al.31 in their work stated: “Histological examination of the brains of animals exposed to either a complex (‘enriched’) environment or learning paradigm, compared with appropriate controls, has illuminated the nature of experience-induced morphological plasticity in the brain [...] that changes in synapse number and morphology are associated with learning and are stable, in that they persist well beyond the period of exposure to the learning experience.” Kennedy et al.32 affirmed: “Learning and memory require the formation of new neural networks in the brain. A key mechanism underlying this process is synaptic plasticity at excitatory synapses, which connect neurons into networks.” During life, humans learn new skills or modify the already learned ones by enriching the existing synaptic pathways. Therefore, the introduction of increasingly innovative technologies requires a continuous brain re-adaptation to new interfaces34. This effort is more effective when the learner is engaged. An engaged user actuates learning in an optimal way, avoiding distractions, and increasing the mental performance35,36.
In37, three different types of engagement are proposed: behavioural, emotional, and cognitive engagements. Behavioral engagement focuses on the observable actions during the learning process38,39. Emotional engagement regards the impact of emotions on the cognitive process effectiveness and the effort sustainability for the users40. Cognitive engagement refers to the amount of cognitive resources spent by the user in a specific activity39,41,42.
Different methods for learning engagement detection are proposed in literature27. For the behavioral engagement assessment, observation grids (used to support direct observations or video analysis) were proposed43,44. For the cognitive and emotional engagement assessment, self-assessment questionnaires and surveys (compiled autonomously by the user) were developed45,46. In recent years alternative engagement assessment methods based on physiological sensors have established: heart-rate variability, galvanic skin response, and EEG47–49. Among these biosignal, the most promising for engagement assessment is the EEG. As already described, the learning is based on a neurological changes set, and the EEG presents the possibility of studying these neural modifications20,50–53. The EEG system is non-invasive, and provides information on brain activity within milliseconds. Recently low-cost solution appeared on the market (i.e. Emotiv epoc or Muse54,55).
It is now commonly used in many applications56,57 including the cognitive and emotion engagement assessment as well as the detection of the underlying elements: emotions recognition and cognitive load activity assessment respectively58–64.
To achieve a correct metrological reference of the EEG-based cognitive and emotional engagement constructs, a reproducibility problem arises. From emotional point of view, when eliciting a specific emotion, the same stimulus does not often induce the same emotion in different subjects. The effectiveness of the induction can be verified by means of self-assessment questionnaires or scales. The combined use of standardized stimuli and subject’s self-assessment ratings can be an effective way to build a metrological reference for a reliable EEG-based emotional engagement detection65. From the cognitive point of view, when the subject is learning, the working memory identifies the incoming information and the long-term memory constructs and stores new schemes on the basis of the past ones. While the already built schemes decrease in the working memory load, the construction of new schemes entails its increase24,66. Therefore, increasing difficulty levels allows to induce different cognitive states; the cognitive engagement level grows up according to the difficulty of the proposed exercise increases67–70.
Proposal
This study proposes an EEG-based cognitive and emotional engagement detection method during a learning task. In this section the Basic ideas, the Architecture, and the adopted Processing framework are outlined.
Basic ideas
The proposed method is based on the following key concepts:
EEG-based subject-adaptative system In the context of learning 4.0, the adaptability of Intelligent Teaching Systems is improved by means of new input channels (EEG).
Cognitive and emotional learning engagement detection the assessment of student engagement is realized considering both cognitive and emotional aspects, according to the Frederiks theory19.
Within and cross-subject designs both the approaches are experimentally validated in order to pursue accuracy maximization or calibration-time minimization, respectively.
Domain Adaptation procedure in cross-subject case a Transfer Component Analysis (TCA)71 allows to use knowledge acquired on other subjects to simplify the system calibration on a new subject.
Wearable system an ultralight wireless EEG device with few dry electrodes maximizes the wearability.
Multi-factorial metrological reference the system is calibrated by using (1) standardized strategies for inducing different levels of cognitive load, and (2) a public acoustic stimuli dataset to elicit emotions. Moreover, the metrological reference of emotional engagement was confirmed by statistical analysis on the outputs of self-assessment questionnaires.
Narrow EEG frequency intervals the EEG features resolution is improved by a 12-band Filter-Bank, obtained by sub-dividing the traditional EEG five bands (delta, theta, alpha, beta, and gamma).
Architecture
The architecture of the proposed system is depicted in Fig. 1. The eight Active Dry Electrodes acquire the EEG signals directly from the scalp. Each channel is differential with respect to AFz (REF), and referred to Fpz (GND), according to the 10/20 international system. After transduction, analog signals are conditioned by the Analog Front End. Next, they are digitized by the Analog Digital Converter (ADC), and submit an Artifact removal block performed by an ICA based algorithm. Then, the signals are sent by the wireless Bluetooth transmission to the Data Processing stage. Here, the suitable feature are extracted by a 12-component Filter Bank. The two Support Vector Machine (SVM) classifiers receive the features array from two trained Common Spatial Pattern (CSP) algorithms for detecting the Cognitive and the Emotional Engagement respectively. Only in the cross-subject case, a baseline removal followed by a TCA procedure is provided during the training stage of the classifier.
Processing framework
In this section, (1) the “Feature extraction and selection” section, the (2) “Baseline removal and Domain Adaptation” section, and (3) the “Classification” section are detailed.
Feature extraction and selection
In this work, a novel Filter Bank version57 is adopted. EEG signals are acquired by an eight channels device with programmable sample rate.
The feature extraction pipeline is based on a filter bank and a Common Spatial Pattern. The filter bank is composed of 12 infinite impulse response (IIR) band-pass Chebyshev type 2 filters with 4 Hz amplitude, equally spaced from 0.5 to 48.5 Hz. The Common Spatial Pattern (CSP)24 implements a mathematical data transformation that improves the data separability. The adoption of a pipeline based on Filter-bank and Common Spatial Pattern allows to combine two different goals. Firstly, the EEG frequency spectrum [0.5–48.5] Hz can be investigated, as literature suggests in case of mental state detection72. Secondly, by adopting a bank based on 12 filters the resolution of the frequency intervals increases with respect to the five typical bands used in EEG analysis (alpha, beta, delta, gamma, theta). Furthermore, Common Spatial Pattern is widely used in EEG-based motor imagery feature extraction57,73,74. Recently, it was demonstrated effective in cognitive57 and emotional75 mental state detection.
Baseline removal and domain adaptation
A cross-subject approach has several advantages with respect to a within-subject one, such as the reduction of time for the initial calibration procedure. Unfortunately, the non-stationarity nature of the EEG signal leads to a greater data variability between subjects. This is a well-known problem in the literature, which makes the cross-subject approach a very challenging task75. Currently, the Domain Adaptation methods76 are obtaining a great attention from the scientific community. In this work, the Transfer Component Analisys (TCA)71 is adopted. TCA is a well-established technique of Domain Adaptation already used in the EEG signal classification literature with promising results75.
Classification
For the classification stage, Support Vector Machines (SVMs)77 are implemented. Considering inputs as points in a vector space, SVM is a binary classifier which discriminates data according to a decision hyperplane. Differently from other hyperplane-based classifiers, an SVM finds the hyperplane maximizing the separation between the classes, i.e. the hyperplane having the largest distance from the margins of the classes.
Methods
In this section the EEG instrumentation, the data acquisition protocol, the data labelling, and the data processing are presented.
EEG instrumentation
The AB-Medica Helmate system Class IIA (certified according to the Regulation on medical devices (EU) 2017/745) is used for the EEG signal measurements78 (Fig. 2a). The device provides 10 dry electrodes disposed according to the International Positioning System 10/20: Fp1, Fp2, Fz, Cz, C3, C4, O1, O2, AFz (ref), and Fpz (Ground). The signals are differentially acquired with respect to the Fpz electrode and grounded to the AFz electrode. The Electrodes (made of a conductive rubber ending with Ag/AgCl coating) are of three different shapes to minimize the contact impedance in each scalp area (Fig. 2b). The Helm8 AB-Medica Software Manager78 allows to (1) verify the contact impedance level, and (2) apply several digital filters for a real-time signal visual analysis. The EEG signals are acquired with a 512 Sa/s sampling rate and sent via Bluetooth to a computation device.
Data acquisition protocol
Twenty-one school age subjects (9 males and 13 females, 23.7 ± 4.1 years) participated in the experiment. The experimental sample was extracted from the population of college students in order to soft the impact of age and educational attainment on performance. The ethical committee of the University of Naples Federico II approved the experimental protocol. All methods were performed in accordance with the relevant guidelines and regulations. Before the experiment, each subject read and signed the informed consent. All volunteers have no neurological diseases. Each subject was seated in a comfortable chair at a distance of 1 m from the computer screen. The location was sanitized before and after of each acquisition, as indicated in the COVID-19 academic protocols. Each subject was equipped with a mouse to carry out the experimental test. After putting the EEG-cap on, the contact impedance was assessed to guarantee optimal signal-acquisition conditions. Each subject underwent an experimental session composed of 8 trials. Various stimuli to induce high and low levels of emotive and cognitive engagements were equally distributed among the trials. Continuous Performance Test (CPT)79 was used to modulate the cognitive engagement. In particular, a CPT version based on a learning by doing activity on how an interface works was adopted. Whereas, proper background music and social feedback was used to modulate the emotive engagement level. More in detail, the three different stimuli are described as follows:
Revised CPT: a red cross and a black circle on the computer screen were presented to the subject. The red cross tends to run out from the circle on the screen in random directions. The subject was asked to keep the cross inside the circle by using the mouse. For each trial, a different difficulty level was set by the experimenter changing the cross speed. The percentage of the time spent by the red cross inside the black circle with respect to the total time was reported to the subject at the end of the trial (Fig. 3).
Background music: for each trial, a particular emotive engagement level was favored by proper background music. The music tracks were randomly selected from the MER80 database where songs are organized according to the 4 quadrants of the emotion Russell’s circumplex model81. The songs associated with the Q1 and Q4 quadrants (cheerful music) were employed in high emotional engagement trials, Q2 and Q3 for the low ones (sad music).
- Social feedbacks: during each trial, the experimenters gave proper social feedbacks according to the emotive engagement levels under the experimental protocol. The positive and negative social feedbacks consisted of encouraging and disheartening comments respectively, given to subject on his/her ongoing performance. The positive and negative social feedbacks were administrated using sentences composed of words extracted from a validated database proposed by Hermans and De Houwer82 (e.g. intelligent, game, fast, rule, surprise, applause, good humour, strong, tenacious, skilful, damn, attentive, careless, talented, energetic, music, careless, weak, naive, silly, confused, inexperienced, clumsy, inhibited, great, etc.). For example, subjects were encouraged and discouraged through comments such as:
- “Applause to you. You did great, you achieved a very impressive score in this game. You deserve a round of applause. You are a real talent, what a nice surprise.”
- “Damn. You didn’t do very well. You were careless. Shall we try again?”
A well-founded metrological reference is ensured by two assessment procedures validating the stimuli effectiveness:
Performance index: an empirical threshold was used to confirm that an appropriate CPT stimuli response was given by the participant. The threshold changed according to the trial difficulty level.
Self Assessment Manikin questionnaire (SAM): the emotional engagement level was assessed by a 9-level version of the SAM. The lower emotional engagement level was associated to the SAM score 1, while the greater one to 9.
The experimental session started with the administration of the SAM to get information about the initial emotional condition of the subject. Then, a preliminary CPT training phase to uniform all the participants starting levels was realized. After this preliminary phase, each trial was implemented by a succession of a CPT stage followed by a SAM administration.
Data labelling
45 s acquisition EEG signals were labeled according to two parameters: (1) high or low emotional engagement, and (2) high or low cognitive engagement. More in detail, regarding the cognitive engagement, the trials were labeled according to the CPT speed66,83, since the higher was the speed the more the cognitive engagement increased24,66. Vesga et al. correlate directly the cognitive engagement with the cognitive workload42. Many studies show how the changes in game difficulty are correlated with cognitive engagement and cognitive load68–71,84,85. In86, the concept of desirable difficulties is presented in terms of “varying the conditions of learning rather than keeping conditions constant and predictable”. This concept is particularly interesting because it connects together the difficulty of the task, the level of involvement and the effectiveness of learning. In this study, the greater difficulty of the task is supposed to induce an increase in the cognitive resources employed by the participant, only if the performances remain compatible. The percentage of the time spent by the red cross inside the black circle with respect to the total time is the performance index used in this study. In detail, for each trial the performance index was analyzed and the subject was assessed as engaged if the final score was within 20% variation with respect to the baseline. Otherwise, the trial was not included in the dataset. The trials having speed lower than 150 pixels/s were labeled as low, whereas high were assigned to the trials having speed higher than 300 pixels/s.
As concern the emotional engagement, the trials characterized by cheerful/sad music and positive/negative social feedback were labelled as high/low. For each trial, the SAM results (normalized to the initial pre-session values) were consistent with the proposed stimuli. In fact, a one-tailed t-student analysis revealed a 0.02 P-value in the worst case.
Data processing
An artifact removal stage preceded the feature extraction and the classification stages. Independent Component Analysis (ICA) was used to filter out the artifacts from the EEG signals using the Runica module of the EEGLab tool85. Then, data were normalized by subtracting their mean and dividing by their standard deviation.
Feature extraction
EEG data were divided in epochs of 3 s, overlapping by 1.5 s. Owing to the sampling rate of 512 Sa/s, for each subject 232 epochs of 1536 samples per channel were extracted.
Five different strategies were compared:
Engagement Index: to make a comparison with the classical literature approach, the engagement index proposed in25 was extracted. Although the Engagement Index was not defined for a particular engagement type, given the experimental setup proposed in25, it can be assumed compatible with the cognitive engagement proposed in this work.
Butterworth-Principal Component Analysis (BPCA): data were filtered by a fourth-order bandpass Butterworth filter [0.5–45] Hz; then, relevant features were extracted using Principal Component Analysis (PCA)87 selecting the components explaining the 95% of the total variance.
Butterworth-CSP (BCSP): data were filtered using a fourth-order bandpass Butterworth filter [0.5 - 45] Hz followed by a CSP projection stage; In a binary problem, CSP works by computing the covariance matrices related to the two classes, simultaneously diagonalized such that the eigenvalues of two covariance matrices sum up to 1. Afterwards, a matrix is computed to project the input into a space where the differences between the class variances are maximized. More precisely, in a binary problem, the projected components are sorted by variances in a decreasing or ascending order: the former, when the projection matrix is applied to inputs belonging to the first class, while the latter when inputs belong to the second class88.
Filter Bank - CSP (FBCSP): data were filtered through a 12 IIR bandpass Chebyshev filter type 2 filter bank with a 4 Hz bandwidth equally spaced from 0.5 to 48.5 Hz, followed by a CSP projection stage.
- Domain adaptation (TCA): only in the cross-subject approach, a baseline removal and a TCA were adopted. In a nutshell, TCA searches for a common latent space between data sampled from two different (but related) data distributions by preserving data properties. More in detail, TCA searches for a data projection that minimizes the Maximum Mean Discrepancy (MMD) between the two distributions, that is:
where and are the numbers of points in the first (source) and the second (target) domain set respectively, while and are the th point (epoch) in the two different sets. The data projected in the new latent space are then used as input for the classification pipeline. Domain adaptation (TCA) with For-subject average removal: in general, TCA works with only two different domains, differently from a multiple-subject environment, which can lead to a domain composed of several sub-domains generated by the different subjects or sessions. In75, TCA was tested by considering for the first domain a subset of samples from subjects, where N is the total number of subjects, and with the data of the remaining subject for the other domain. However, this approach does not take into consideration the fact that different subjects may belong to very different domains, leading to poor results. A simple solution consists in subtracting to each subject a baseline signal recorded from the user, for example, in rest condition. However, this last point requires new subject acquisition. Instead, in this work, an average of the signals for each subject is used as baseline, thus avoiding the need for new signal acquisitions.
Classification
The output of the classification stage can be “high” or “low” both for cognitive and emotional engagement. Since we are dealing with a binary classification problem, the theoretical chance level for prediction is 50%. For each feature selection strategy shown in the previous subsection, several classifiers were compared with the adopted SVM: Linear Discriminant Analysis (LDA)89, k-Nearest Neighbour (k-NN)89, shallow Artificial Neural Networks (ANN), Deep Neural Networks (DNN)90, and Convolutional Neural Networks (CNN)91,92. LDA searches for a linear projection of the data in a lower dimensional space, while keeping preserved the discriminatory information between the data classes. k-NN is a model that, given a set P of non-labelled points to classify, a distance measure d (such as the Euclidean distance), a positive integer k, and a set D of labelled points, assigns to each point the most frequent class, according to the measure d, among its k neighbours in D. ANN is a model consisting of a set of basic elements (called neurons), arranged in several full-connected layers. Each neuron computes the linear combination of its inputs, that is subsequently given as input of an activation function. The number of neurons, the number of layers and the activation functions are a priori hyperparameters, while the coefficients of each linear combination are learned during a training stage. According to the number of layers, in this work ANNs are referred as shallow when they are made by a single layer, otherwise they are referred as deep (DNN). CNNs are deep networks inspired by the functioning of the visual cortex of the brain in processing and recognizing images. Differently from classical deep neural networks, CNNs extract features from the input using the mathematical convolution operator.
Each combination of feature selection strategies and classifiers were used on both emotional and cognitive engagement.
The best model was selected by a stratified leave-2-trials out technique in order to maintain a balancing among the classes in each fold. A Grid search strategy was adopted as approach for hyperparameters tuning for each classifier (Table 1).
Table 1.
Classifier | Optimized Hyperparameter | Variation Range |
---|---|---|
k-Nearest neighbour (k-NN) | Algorithms | {Ball tree, KD Tree, Brute force} |
Distance Weight | {equal, inverse} | |
Num Neighbors | [1, 10], step: 1 | |
Support vector machine (SVM) | C Regularization | {0.01, 0.1, 1, 5, 10} |
Kernel Function | {radial basis, polynomial} | |
Polynomial Order | {2, 3} | |
Linear discriminant analysis (LDA) | Solver | {Singular value decomposition, least squares} |
Shrinkage | {None, Ledoit-Wolf lemma} | |
shallow artificial neural network (ANN) | Activation function | {ReLU, sigmoid} |
Number of neurons | [5, 50], step: 25 | |
Number of layers | {1} | |
Learning rate | {0.001, 0.01} | |
Deep neural network (DNN) | Activation function | {ReLU, sigmoid} |
Number of neurons per layer | [5, 100], step: 25 | |
Number of layers | {2,3,4} | |
Learning rate | {0.00001, 0.0001, 0.001} | |
Convolutional neural network (CNN) | Activation function | {ReLU, sigmoid} |
Number of layers | {1,2,3} | |
Number of filters | {13,32,64} | |
Kernel size | {3, 5} | |
Stride | {1, 2} | |
Dense layer size | {50, 100} | |
Learning rate | {0.001, 0.01} |
Experimental results
In this section, the experimental results obtained in within- and cross-subject cases are reported.
Within-subjects
Firstly, to make a comparison with the classical literature approach, the engagement index proposed in25 was used as feature for a classification of the cognitive engagement. Unfortunately, as highlighted by the results reported in Table 2, accuracy performances were not optimal. In fact, this feature is mainly used in non-predictive applications (e.g.,27).
Table 2.
Method | Cognitive engagement |
---|---|
SVM | 54.8 ± 4.9 |
k-NN | 53.7 ± 5.7 |
ANN | 53.1 ± 5.4 |
LDA | 50.7 ± 6.2 |
DNN | 53.4 ± 3.8 |
Instead, the best results both on cognitive and emotional engagements (Fig. 4) were achieved using features extracted by Filter-Bank and CSP.
Quantitative results related to the use of Filter Bank and CSP for each classifier can be observed in Table 3: among the different classifiers, SVM stands out with a better performance than the other ones, reaching its best mean accuracies of on cognitive engagement classification and of on emotional engagement. Results are computed as the average accuracy over all the subjects. Current results suggest that SVMs can be optimal to address the proposed classification problem. A possible explanation is that the kernel spaces induced by the Support Vector Machines resulted particularly suitable for the acquired data size in junction with the features transformation adopted (FBCSP). As shown in Fig. 5, where the Filter Bank effects are represented using t-SNE, FBCSP improves the data separability between the classes and simplifies the classification problem. Therefore, the task can be dealt with a classifier having a low number of parameters identifiable even from datasets not necessarily large.
Table 3.
Method | Cognitive engagement (proposed) | Emotional engagement (proposed) |
---|---|---|
SVM | 76.9 ± 10.2 | 76.7 ± 10.0 |
k-NN | 73.0 ± 9.7 | 74.2 ± 10.3 |
ANN | 74.0 ± 9.2 | 73.9 ± 9.1 |
LDA | 72.1 ± 11.4 | 71.6 ± 9.3 |
DNN | 73.7 ± 8.9 | 73.4 ± 9.6 |
CNN | 74.1 ± 10.1 | 74.4 ± 9.4 |
The best performance average values are highlighted in bold.
The results reported in Fig. 2b show that the Filter Bank improves the classification performance by a significant proportion. This can be due to the use of several sub-bands which highlight the signal main characteristics, allowing the CSP computation to project the subject data in a more discriminative common space. In Fig. 5, BCSP and FBCSP are compared through t-SNE93 on the subjects data transformed using the two different methods. The figure shows that, for several subjects, CSP applied after FB projects the data in a space where they are easily separable with respect to the BCSP case.
Cross-subject approach
A t-SNE plot of the data first and after removing the average value of each subject is shown in Fig. 6. The data without for-subject average removal (Fig. 6a) are disposed in several clusters over the t-SNE space, exhibiting a fragmentation tendency. Instead, after the for-subject average removal (Fig. 6b), the data result more homogeneous, enhancing the model generalizability. A comparison using TCA with and without the for-subject average removal is made and the resulting performances are reported in Table 4. The results show that removing the for-subject average from each subject boosts the performance with respect to using TCA alone (more than 3% of improvement in almost all classifiers, especially in Cognitive Engagement case).
Table 4.
Method | With for-subject average removal | Without for-subject average removal | ||
---|---|---|---|---|
Cognitive engagement | Emotional engagement | Cognitive engagement | Emotional engagement | |
SVM | 72.8 ± 10.9 | 66.2 ± 13.8 | 64.0 ± 10.8 | 61.7 ± 10.2 |
k-NN | 69.6 ± 11.2 | 61.9 ± 9.1 | 57.1 ± 8.9 | 56.9 ± 10.6 |
ANN | 72.6 ± 11.8 | 65.7 ± 14.1 | 69.7 ± 12.3 | 65.8 ± 14.8 |
LDA | 69.5 ± 12.2 | 65.3 ± 13.8 | 69.6 ± 12.9 | 64.6 ± 13.2 |
DNN | 72.1 ± 11.6 | 66.0 ± 14.0 | 70.4 ± 13.4 | 65.8 ± 14.6 |
CNN | 69.9 ± 10.8 | 64.5 ± 11.6 | 65.9 ± 12.8 | 63.8 ± 14.5 |
The best performance values are highlighted in bold.
Conclusion
In this work, a wearable system for personalized EEG-based cognitive and emotional engagement detection is proposed. The system can be used in the context of Learning 4.0 as a new input channel of an adaptive automated teaching platform to improve the learning effectiveness. The wearability is guaranteed by a wireless cap with dry electrodes and 8 data acquisition channels.
The system is validated on students during a training stage involving cognitive and motor skills and aimed to learn how to use a human-machine interface. Standard stimuli, performance indicator, and self assessment questionnaires were employed to guarantee a well founded metrologically reference. The proposed method, based on Filter Bank, CSP and SVM, experimentally showed the best performance. In particular, in the cross-subject case, an average accuracy of 72.8% and 66.2% was reached for the cognitive engagement and emotional engagement respectively by using TCA and for-subject average removal. Instead, in the within-subject case, an accuracy of 76.9% and 76.7% was reached for the cognitive engagement and emotional engagement, respectively. This study was conducted in laboratory, therefore a prototype demonstration in operational environment still lacks. In future works, the proposed solution will be tested in real educational situations (e.g. a real lesson) and validated by means of standardized engagement assessment procedures (e.g. self-reports).
Appendix
The Self-Assessment Manikin (SAM) is an iconic assessment questionnaire that directly measures the valence, arousal, and dominance associated with a person’s emotional response94. In Fig. 7, the 9-Point version for valence assessment is reported. Specifically, a single-item scale measures valence/pleasure of the response from positive to negative. SAM is freely available questionnaire and it is very intuitive. It requires little explanation and participants generally have no difficulty in completing it. SAM resulted reliable after comparison analysis with other standard assessment tools (i.e., Semantic Differential Scale95). It was also shown to be employable in different cultural contexts and with participants of different ages96.
Author contributions
All authors conceived and designed the experiments, performed the experiments, analysed the results, wrote and reviewed the manuscript.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Battro AM, Fischer KW. Mind, brain, and education in the digital era. Mind Brain Educ. 2012;6(1):49. [Google Scholar]
- 2.Barrett R, Gandhi HA, Naganathan A, Daniels D, Zhang Y, Onwunaka C, Luehmann A, White AD. Social and tactile mixed reality increases student engagement in undergraduate lab activities. J. Chem. Educ. 2018;95(10):1755. [Google Scholar]
- 3.Gan HS, Tee NYK, Bin Mamtaz MR, Xiao K, Cheong BHP, Liew OW, Ng TW. Augmented reality experimentation on oxygen gas generation from hydrogen peroxide and bleach reaction. Biochem. Mol. Biol. Educ. 2018;46(3):245. doi: 10.1002/bmb.21117. [DOI] [PubMed] [Google Scholar]
- 4.Yoon SA, Elinich K, Wang J, Steinmeier C, Tucker S. Using augmented reality and knowledge-building scaffolds to improve learning in a science museum. Int. J. Comput.-Support. Collab. Learn. 2012;7(4):519. [Google Scholar]
- 5.Klopp, M. & Abke, J. In 2018 IEEE International Conference on Teaching, Assessment, and Learning for Engineering (TALE) (IEEE, 2018), pp. 871–876.
- 6.Mardiana H, Daniels HK. Technological determinism, new literacies and learning process and the impact towards future learning. Online Submiss. 2019;5(3):219. [Google Scholar]
- 7.Lindgren R, Johnson-Glenberg M. Emboldened by embodiment: Six precepts for research on embodied learning and mixed reality. Educ. Res. 2013;42(8):445. [Google Scholar]
- 8.Anjarichert LP, Gross K, Schuster K, Jeschke S. Learning 4.0: Virtual immersive engineering education. Digit. Univ. 2016;2:51. [Google Scholar]
- 9.Janssen D, Tummel C, Richert A, Isenhardt I. Virtual environments in higher education-immersion as a key construct for learning 4.0. IJAC. 2016;9(2):20. [Google Scholar]
- 10.Willingham DB. A neuropsychological theory of motor skill learning. Psychol. Rev. 1998;105(3):558. doi: 10.1037/0033-295x.105.3.558. [DOI] [PubMed] [Google Scholar]
- 11.Sailer U, Flanagan JR, Johansson RS. Eye-hand coordination during learning of a novel visuomotor task. J. Neurosci. 2005;25(39):8833. doi: 10.1523/JNEUROSCI.2658-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Bagustari, B. & Santoso, H. in J. Phys. Conf. Ser., vol. 1235 (IOP Publishing, 2019) 1235: 012033
- 13.Hu PJH, Hui W. Examining the role of learning engagement in technology-mediated learning and its effects on learning effectiveness and satisfaction. Decis. Support Syst. 2012;53(4):782. [Google Scholar]
- 14.Rodgers T. Student engagement in the e-learning process and the impact on their grades. Int. J. Cyber Soc. Educ. 2008;1(2):143. [Google Scholar]
- 15.Park SY. Student engagement and classroom variables in improving mathematics achievement. Asia Pac. Educ. Rev. 2005;6(1):87. [Google Scholar]
- 16.Lamborn, S., Newmann, F. & Wehlage, G. The significance and sources of student engagement, Student engagement and achievement in American secondary schools pp. 11–39 (1992)
- 17.Lutz A, Slagter HA, Dunne JD, Davidson RJ. Attention regulation and monitoring in meditation. Trends Cogn. Sci. 2008;12(4):163. doi: 10.1016/j.tics.2008.01.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Connell, J.P. & Wellborn, J.G. in Cultural processes in child development: The Minnesota symposia on child psychology, (Psychology Press, 1991), vol. 23, pp. 43–78
- 19.Fredricks JA, Blumenfeld PC, Paris AH. School engagement: Potential of the concept, state of the evidence. Rev. Educ. Res. 2004;74(1):59. [Google Scholar]
- 20.Khedher AB, Jraidi I, Frasson C, et al. Tracking students’ mental engagement using eeg signals during an interaction with a virtual learning environment. J. Intell. Learn. Syst. Appl. 2019;11(01):1. [Google Scholar]
- 21.Debener S, Minow F, Emkes R, Gandras K, De Vos M. How about taking a low-cost, small, and wireless eeg for a walk? Psychophysiology. 2012;49(11):1617. doi: 10.1111/j.1469-8986.2012.01471.x. [DOI] [PubMed] [Google Scholar]
- 22.Galway, L., McCullagh, P., Lightbody, G., Brennan, C. & Trainor, D. in 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing (IEEE, 2015), pp. 1554–1559
- 23.LaRocco, J., Le, M.D. & Paeng, D.G. A systemic review of available low-cost eeg headsets used for drowsiness detection, Frontiers in neuroinformatics 14 (2020) [DOI] [PMC free article] [PubMed]
- 24.Andreessen LM, Gerjets P, Meurers D, Zander TO. Toward neuroadaptive support technologies for improving digital reading: A passive bci-based assessment of mental workload imposed by text difficulty and presentation speed during reading. User Model. User-Adap. Inter. 2021;31(1):75. [Google Scholar]
- 25.Pope AT, Bogart EH, Bartolome DS. Biocybernetic system evaluates indices of operator engagement in automated task. Biol. Psychol. 1995;40(1–2):187. doi: 10.1016/0301-0511(95)05116-3. [DOI] [PubMed] [Google Scholar]
- 26.Eldenfria A, Al-Samarraie H. Towards an online continuous adaptation mechanism (ocam) for enhanced engagement: An eeg study. Int. J. Hum. Comput. Interact. 2019;35(20):1960. [Google Scholar]
- 27.Kosmyna N, Maes P. Attentivu: An eeg-based closed-loop biofeedback system for real-time monitoring and improvement of engagement for personalized learning. Sensors. 2019;19(23):5200. doi: 10.3390/s19235200. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.McLeod, S. Jean piaget’s theory of cognitive development, Simply Psychology pp. 1–9 (2018)
- 29.Garland EL, Howard MO. Neuroplasticity, psychosocial genomics, and the biopsychosocial paradigm in the 21st century. Health & Social Work. 2009;34(3):191. doi: 10.1093/hsw/34.3.191. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Kleim, J.A. & Jones, T.A. Principles of experience-dependent neural plasticity: implications for rehabilitation after brain damage (2008) [DOI] [PubMed]
- 31.Markham JA, Greenough WT. Experience-driven brain plasticity: Beyond the synapse. Neuron Glia Biol. 2004;1(4):351. doi: 10.1017/s1740925x05000219. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Kennedy MB. Synaptic signaling in learning and memory. Cold Spring Harb. Perspect. Biol. 2016;8(2):a016824. doi: 10.1101/cshperspect.a016824. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Owens MT, Tanner KD. Teaching as brain changing: Exploring connections between neuroscience and innovative teaching. CBE-Life Sci. Educ. 2017;16(2):fe2. doi: 10.1187/cbe.17-01-0005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Doidge, N. The brain that changes itself: Stories of personal triumph from the frontiers of brain science (Penguin, 2007)
- 35.VanDeWeghe R. Engaged Learning. USA: Corwin Press; 2009. [Google Scholar]
- 36.Jang H, Reeve J, Deci EL. Engaging students in learning activities: It is not autonomy support or structure but autonomy support and structure. J. Educ. Psychol. 2010;102(3):588. [Google Scholar]
- 37.Alrashidi O, Phan HP, Ngu BH. Academic engagement: An overview of its definitions, dimensions, and major conceptualisations. Int. Educ. Stud. 2016;9(12):41. [Google Scholar]
- 38.Cappella E, Kim HY, Neal JW, Jackson DR. Classroom peer relationships and behavioral engagement in elementary school: The role of social network equity. Am. J. Commun. Psychol. 2013;52(3–4):367. doi: 10.1007/s10464-013-9603-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Pilotti M, Anderson S, Hardy P, Murphy P, Vincent P. Factors related to cognitive, emotional, and behavioral engagement in the online asynchronous classroom. Int. J. Teach. Learn. Higher Educ. 2017;29(1):145. [Google Scholar]
- 40.Silva, A. & Simoes, R. Handbook of Research on Trends in Product Design and Development: Technological and Organizational Perspectives: Technological and Organizational Perspectives (IGI Global, 2010)
- 41.Vesga, J. B., Xu, X. & He, H. in 2021 IEEE Virtual Reality and 3D User Interfaces (VR) (IEEE, 2021), pp. 645–652
- 42.Rotgans JI, Schmidt HG. Cognitive engagement in the problem-based learning classroom. Adv. Health Sci. Educ. 2011;16(4):465. doi: 10.1007/s10459-011-9272-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Wigfield A, Guthrie JT, Perencevich KC, Taboada A, Klauda SL, McRae A, Barbosa P. Role of reading engagement in mediating effects of reading comprehension instruction on reading outcomes. Psychol. Sch. 2008;45(5):432. [Google Scholar]
- 44.Helme S, Clarke D. Identifying cognitive engagement in the mathematics classroom. Math. Educ. Res. J. 2001;13(2):133. [Google Scholar]
- 45.Chen PSD, Lambert AD, Guidry KR. Engaging online learners: The impact of web-based learning technology on college student engagement. Comput. Educ. 2010;54(4):1222. [Google Scholar]
- 46.Jaafar, S., Awaludin, N. S. & Bakar, N. S. in E-proceeding of the Conference on Management and Muamalah (2014), pp. 128–135
- 47.Darnell DK, Krieg PA. Student engagement, assessed using heart rate, shows no reset following active learning sessions in lectures. PLoS ONE. 2019;14(12):e0225709. doi: 10.1371/journal.pone.0225709. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Monkaresi H, Bosch N, Calvo RA, D’Mello SK. Automated detection of engagement using video-based estimation of facial expressions and heart rate. IEEE Trans. Affect. Comput. 2016;8(1):15. [Google Scholar]
- 49.Peacock J, Purvis S, Hazlett RL. Which broadcast medium better drives engagement?: measuring the powers of radio and television with electromyography and skin-conductance measurements. J. Advert. Res. 2011;51(4):578. [Google Scholar]
- 50.Berka C, Levendowski DJ, Lumicao MN, Yau A, Davis G, Zivkovic VT, Olmstead RE, Tremoulet PD, Craven PL. Eeg correlates of task engagement and mental workload in vigilance, learning, and memory tasks. Aviat. Space Environ. Med. 2007;78(5):B231. [PubMed] [Google Scholar]
- 51.Kumar, N. & Michmizos, K. P. in 2020 8th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics (BioRob) (IEEE, 2020), pp. 521–526
- 52.Park W, Kwon GH, Kim DH, Kim YH, Kim SP, Kim L. Assessment of cognitive engagement in stroke patients from single-trial eeg during motor rehabilitation. IEEE Trans. Neural Syst. Rehabil. Eng. 2014;23(3):351. doi: 10.1109/TNSRE.2014.2356472. [DOI] [PubMed] [Google Scholar]
- 53.Khedher, A. B., Jraidi, I. & Frasson, C. in EdMedia+ Innovate Learning (Association for the Advancement of Computing in Education (AACE), 2018), pp. 394–401
- 54.Emotiv.https://www.emotiv.com/epoc/
- 55.Muse.https://choosemuse.com/
- 56.Angrisani, L., Arpaia, P., Donnarumma, F., Esposito, A., Frosolone, M., Improta, G., Moccaldi, N., Natalizio, A. & Parvis, M. in 2020 IEEE International Instrumentation and Measurement Technology Conference (I2MTC) (IEEE, 2020), pp. 1–6
- 57.Apicella A, Arpaia P, Frosolone M, Moccaldi N. High-wearable eeg-based distraction detection in motor rehabilitation. Sci. Rep. 2021;11(1):1. doi: 10.1038/s41598-021-84447-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Benlamine, M. S., Dufresne, A., Beauchamp, M. H. & Frasson, C. Bargain: behavioral affective rule-based games adaptation interface–towards emotionally intelligent games: Application on a virtual reality environment for socio-moral development, User Modeling and User-Adapted Interaction pp. 1–35 (2021)
- 59.Wang XW, Nie D, Lu BL. Emotional state classification from eeg data using machine learning approach. Neurocomputing. 2014;129:94. [Google Scholar]
- 60.Soleymani M, Asghari-Esfeden S, Fu Y, Pantic M. Analysis of eeg signals and facial expressions for continuous emotion detection. IEEE Trans. Affect. Comput. 2015;7(1):17. [Google Scholar]
- 61.Zhuang, N., Zeng, Y., Tong, L., Zhang, C., Zhang, H. & Yan, B. Emotion recognition from eeg signals using multidimensional information in emd domain, BioMed research international 2017 (2017) [DOI] [PMC free article] [PubMed]
- 62.Jraidi, I., Chaouachi, M. & Frasson, C. A hierarchical probabilistic framework for recognizing learners’ interaction experience trends and emotions, Advances in Human-Computer Interaction 2014 (2014)
- 63.Aricò P, Borghini G, Di Flumeri G, Colosimo A, Pozzi S, Babiloni F. A passive brain-computer interface application for the mental workload assessment on professional air traffic controllers during realistic air traffic control tasks. Prog. Brain Res. 2016;228:295. doi: 10.1016/bs.pbr.2016.04.021. [DOI] [PubMed] [Google Scholar]
- 64.Wang S, Gwizdka J, Chaovalitwongse WA. Using wireless eeg signals to assess memory workload in the -back task. IEEE Trans. Hum. Mach. Syst. 2015;46(3):424. [Google Scholar]
- 65.Jenke R, Peer A, Buss M. Feature extraction and selection for emotion recognition from eeg. IEEE Trans. Affect. Comput. 2014;5(3):327. [Google Scholar]
- 66.Paas F, Tuovinen JE, Tabbers H, Van Gerven PW. Cognitive load measurement as a means to advance cognitive load theory. Educ. Psychol. 2003;38(1):63. [Google Scholar]
- 67.Zhang X, Lyu Y, Hu X, Hu Z, Shi Y, Yin H. Evaluating photoplethysmogram as a real-time cognitive load assessment during game playing. Int. J. Hum. Comput. Interact. 2018;34(8):695. [Google Scholar]
- 68.Das, S., Ghosh, L. & Saha, S. in 2020 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT) (IEEE, 2020), pp. 1–6
- 69.Mandinach EB, Corno L. Cognitive engagement variations among students of different ability level and sex in a computer problem solving game. Sex Roles. 1985;13(3):241. [Google Scholar]
- 70.Ke F, Xie K, Xie Y. Game-based learning engagement: A theory-and data-driven exploration. Br. J. Edu. Technol. 2016;47(6):1183. [Google Scholar]
- 71.Pan SJ, Tsang IW, Kwok JT, Yang Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Networks. 2010;22(2):199. doi: 10.1109/TNN.2010.2091281. [DOI] [PubMed] [Google Scholar]
- 72.Angrisani L, Arpaia P, Esposito A, Moccaldi N. A wearable brain–computer interface instrument for augmented reality-based inspection in industry 4.0. IEEE Trans. Instrum. Meas. 2019;69(4):1530. [Google Scholar]
- 73.Kumar, S., Sharma, R., Sharma, A. & Tsunoda, T. in 2016 international joint conference on neural networks (IJCNN) (IEEE, 2016), pp. 2090–2095
- 74.Bentlemsan, M., Zemouri, E. T., Bouchaffra, D., Yahya-Zoubir, B. & Ferroudji, K. in 2014 5th International conference on intelligent systems, modelling and simulation (IEEE, 2014), pp. 235–238
- 75.Zheng, W. L., Zhang, Y. Q., Zhu, J. Y. & Lu, B. L. in 2015 international conference on affective computing and intelligent interaction (ACII) (IEEE, 2015), pp. 917–922
- 76.Pan SJ, Yang Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009;22(10):1345. [Google Scholar]
- 77.Noble WS. What is a support vector machine? Nat. Biotechnol. 2006;24(12):1565. doi: 10.1038/nbt1206-1565. [DOI] [PubMed] [Google Scholar]
- 78.Ab-medica s.p.a.https://www.abmedica.it/ (2020)
- 79.Gaume A, Dreyfus G, Vialatte FB. A cognitive brain-computer interface monitoring sustained attentional variations during a continuous task. Cogn. Neurodyn. 2019;13(3):257. doi: 10.1007/s11571-019-09521-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Panda R, Malheiro R, Paiva RP. Novel audio features for music emotion recognition. IEEE Trans. Affect. Comput. 2018;11(4):614. [Google Scholar]
- 81.Russell JA. A circumplex model of affect. J. Pers. Soc. Psychol. 1980;39(6):1161. [Google Scholar]
- 82.Vanderhasselt MA, Remue J, Ng KK, Mueller SC, De Raedt R. The regulation of positive and negative social feedback: A psychophysiological study. Cognitive Affect. Behav. Neurosci. 2015;15(3):553. doi: 10.3758/s13415-015-0345-8. [DOI] [PubMed] [Google Scholar]
- 83.Antonenko P, Paas F, Grabner R, Van Gog T. Using electroencephalography to measure cognitive load. Educ. Psychol. Rev. 2010;22(4):425. [Google Scholar]
- 84.Benzing V, Heinks T, Eggenberger N, Schmidt M. Acute cognitively engaging exergame-based physical activity enhances executive functions in adolescents. PLoS ONE. 2016;11(12):e0167501. doi: 10.1371/journal.pone.0167501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Delorme A, Makeig S. Eeglab: an open source toolbox for analysis of single-trial eeg dynamics including independent component analysis. J. Neurosci. Methods. 2004;134(1):9. doi: 10.1016/j.jneumeth.2003.10.009. [DOI] [PubMed] [Google Scholar]
- 86.Bjork RA, Kroll JF. Desirable difficulties in vocabulary learning. Am. J. Psychol. 2015;128(2):241. doi: 10.5406/amerjpsyc.128.2.0241. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Abdi H, Williams LJ. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010;2(4):433. [Google Scholar]
- 88.Asensio-Cubero J, Gan JQ, Palaniappan R. Multiresolution analysis over graphs for a motor imagery based online bci game. Comput. Biol. Med. 2016;68:21. doi: 10.1016/j.compbiomed.2015.10.016. [DOI] [PubMed] [Google Scholar]
- 89.Friedman, J., Hastie, T. & Tibshirani, R. et al. The elements of statistical learning, vol. 1 (Springer series in statistics New York, 2001)
- 90.Bishop CM. Pattern Recognition and Machine Learning. USA: Springer; 2006. [Google Scholar]
- 91.Albawi, S., Mohammed, T. A. & Al-Zawi, S. in 2017 International Conference on Engineering and Technology (ICET) (Ieee, 2017), pp. 1–6
- 92.Kim P. MATLAB Deep Learning. USA: Springer; 2017. pp. 121–147. [Google Scholar]
- 93.Van der Maaten, L. & Hinton, G. Visualizing data using t-sne. J. Mach. Learn. Res.9(11) (2008)
- 94.Bynion, T. M. & Feldner, M. T. Self-assessment manikin. Encyclopedia of personality and individual differences pp. 4654–4656 (2020)
- 95.Bradley MM, Lang PJ. Measuring emotion: The self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry. 1994;25(1):49. doi: 10.1016/0005-7916(94)90063-9. [DOI] [PubMed] [Google Scholar]
- 96.Backs RW, da Silva SP, Han K. A comparison of younger and older adults’ self-assessment manikin ratings of affective pictures. Exp. Aging Res. 2005;31(4):421. doi: 10.1080/03610730500206808. [DOI] [PubMed] [Google Scholar]