Skip to main content
ACS AuthorChoice logoLink to ACS AuthorChoice
. 2023 Feb 8;5(2):877–886. doi: 10.1021/acsaelm.2c01436

Soft Wireless Headband Bioelectronics and Electrooculography for Persistent Human–Machine Interfaces

Seunghyeb Ban †,, Yoon Jae Lee ‡,§, Shinjae Kwon ‡,, Yun-Soung Kim , Jae Won Chang #,*, Jong-Hoon Kim †,¶,*, Woon-Hong Yeo ‡,∥,∇,○,*
PMCID: PMC9979786  PMID: 36873262

Abstract

graphic file with name el2c01436_0008.jpg

Recent advances in wearable technologies have enabled ways for people to interact with external devices, known as human–machine interfaces (HMIs). Among them, electrooculography (EOG), measured by wearable devices, is used for eye movement-enabled HMI. Most prior studies have utilized conventional gel electrodes for EOG recording. However, the gel is problematic due to skin irritation, while separate bulky electronics cause motion artifacts. Here, we introduce a low-profile, headband-type, soft wearable electronic system with embedded stretchable electrodes, and a flexible wireless circuit to detect EOG signals for persistent HMIs. The headband with dry electrodes is printed with flexible thermoplastic polyurethane. Nanomembrane electrodes are prepared by thin-film deposition and laser cutting techniques. A set of signal processing data from dry electrodes demonstrate successful real-time classification of eye motions, including blink, up, down, left, and right. Our study shows that the convolutional neural network performs exceptionally well compared to other machine learning methods, showing 98.3% accuracy with six classes: the highest performance till date in EOG classification with only four electrodes. Collectively, the real-time demonstration of continuous wireless control of a two-wheeled radio-controlled car captures the potential of the bioelectronic system and the algorithm for targeting various HMI and virtual reality applications.

Keywords: soft materials, flexible headband, wireless bioelectronics, electrooculography, deep learning, real-time classification, human−machine interface

Introduction

The need for human–machine interface (HMI) technology connecting healthcare applications is increasing rapidly. For example, a touch screen and joystick are HMI, a user interface connecting a person to a machine. The global HMI market is expected to generate more than $8 billion in revenue from 2017 to 2023.1 Among various HMI fields, the field of healthcare is receiving a lot of attention. Recent studies have shown that the wheelchair based on HMI was developed to aid disabled people in their daily activities.25 Typical input signals for HMI are body motions such as hand or finger motion and biopotential. Healthcare applications require an ergonomic approach and high precision.6 In this case, biopotential signals are attractive candidates since biopotential is non-invasive, requires minimal hardware, and contains user movement information. That is why physiological biopotentials and human activities that wearable devices can measure have been suggested for HMI, such as controllers for various healthcare applications. Physiological biopotentials, such as electromyography (EMG), electroencephalography (EEG), and electrooculography (EOG), can be the control commands. For example, EMG signals from muscle movements with a fast response have proved possible to connect with HMI.7 However, muscle weakness due to disabilities cannot produce the required stimulus for the detection of EMG.8 EEG can be another way, which exploits neural information as input control for HMI. However, prior studies further indicated that noninvasive EEG features do not contain sufficient information about small movements.9 High-fidelity EEG is also difficult to acquire and not feasible for real-time and accurate HMI applications. When measured from the scalp, an EEG signal has an amplitude between about 10 and 100 μV. However, EOG amplitude, which has an amplitude between about 0.05 and 3 mV, is quite larger than EEG amplitude.1013 Due to frail grip strength and issues with controlling their bodies for existing motorized wheelchair users,1418 there are restrictions on the use of EMG and EEG. As one of the technologies for tracking eye movements by measuring the potential via the positively charged cornea and negatively charged retina, EOG has gained interest in HMI.19

In several papers, wearable EOG devices in the form of glasses have been used because of their easy and fast wear. For example, the conventional wearable EOG eyeglasses named JINS MEME have one electrode on the bridge of the nose and one on each of the nose pads of the eyeglasses.5,20 Some studies also manufactured 3D printed glasses-type wearable EOG devices.21 However, a glasses-type device is inconvenient to people who are already wearing glasses. These devices are restricted when the electrode is secured to the skin or when there is movement. In addition, glasses-type platforms can be challenging to wear for people with a variety of head sizes because glasses-type platforms are made with a fixed frame width and temple length. Since the glasses-type platform was not size-adjustable, it should be used by users who fit the prescribed size of glasses. Also, wearing it in an inappropriate head size can cause the glasses-type platform to come off during active movements. From the perspective of electrodes, prior studies using the gel electrode show high-fidelity recording, the existing electrodes still have limitations, such as poor breathability, skin irritation, and loss of performance during long-term monitoring due to drying. The conventional gel electrodes dehydrate and reduce the electrode performance over time.22 For the aforementioned reasons, the gel electrodes should be changed periodically. Constant changing of electrodes is not convenient in healthcare applications and is inefficient.23 On the other hand, prior work demonstrated HMI using eye-tracking capability within wearable devices by integrating infrared cameras.24 This HMI using eye-tracking has several problems. There should always be a camera that blocks that person’s view. This system also requires clear pupil and eye images of the user. Still, eyelashes and eyelids can hinder the successful detection of the pupil and bright light can also interfere with pupil detection.25

In this work, we introduce a soft material-based, all-in-one headband EOG device integrating a flexible wireless circuit and an array of fractal gold electrodes that compensate for the limitations of the pre-existing devices mentioned above. The headband platform is advantageous over glasses-type counterparts such as a size-adjustable and stable adhesion. In the case of a glasses-type platform, the part that supports the face is narrow, but the headband type platform has a wider electrode-skin contact area, so multiple electrodes can be secured to the face. To address the gel issues such as skin irritation and short-term durability, we introduce ultrathin, dry electrodes. In recent papers, with its well-studied biocompatibility and processibility, mesh-patterned gold electrodes have been widely used to measure biopotentials.26,27 Specifically, the ultrathin, fractal-designed gold electrode can help the electrode accommodate dynamic skin deformation for a high-fidelity recording of EOG and causes fewer skin irritations compared to the existing gel electrodes. Also, the wearable EOG device can acquire EOG data and classify eye directions in real-time. Compared to prior articles within the scope of real-time classification of eye movements based on wearable EOG devices, our device shows the highest accuracy in classifying six different classes with only four electrodes. Overall, the presented system can meet requirements such as an ergonomic approach and high-precision interfaces. The wearable EOG device with this system allows people to acquire EOG signals stably and control various healthcare applications.

Results and Discussion

Overview of a Wireless, Portable Wearable EOG Device

Figure 1 summarizes the overview of an integrated bioelectronic system for detecting eye movements and persistent HMI. A portable and wearable EOG system enables real-time, continuous, and long-term recording of EOG signals to classify eye movements. Figure 1A shows a subject wearing the headband-type EOG device, integrated with the flexible circuit and fractal gold electrodes. Based on the prior studies,2830 we selected the electrode locations to fit the headband-type platform. Two electrodes were positioned 1 cm above each eye. One electrode was placed 1 cm below the left lower eyelid for vertical eye movement. A common grounding electrode was placed on the middle of the forehead.29 The 3D-printed wearable EOG device is composed of a tension string for securing electrodes to the subject’s face. To accommodate various head sizes, thermoplastic polyurethane (TPU), a flexible rubber-like material, makes the headset platform. Figure S1 shows the flexibility of the headband platform. A dry nanomembrane electrode has a stretchable fractal pattern (Figure 1B). The main contribution of this design is to offer maximized stretchability and bending capability without mechanical fracture. The graph in Figure 1C shows EOG signals for left and right eye movements, recorded by two types of electrodes. Then, the calculated signal-to-noise ratio (SNR) compares the performance of the dry gold electrode with the conventional gel electrode.31 In the experiment, two electrodes detected changes in EOG amplitudes according to angles of eye direction. The electrodes were positioned 1 cm away from each eye for concurrent comparison. Sensitivity measurements are performed by tracing a series of marked targets, located 60 cm away from the eyes (Figure S2).14 The gold electrode’s sensitivity is 12.3 ± 0.5 μV/°, and the conventional gel electrode’s sensitivity is 11.7 ± 0.9 μV/°. The result in Figure 1C shows that the gold electrode (SNR: 22.1 ± 1.7 dB) has a slightly higher SNR than the commercial electrode (SNR: 19.2 ± 2.2 dB), capturing the performance of the dry electrode for high-quality EOG detection. As shown in Figure S3, we compared the skin-electrode contact impedance between a conventional gel electrode and our dry electrode, showing comparable values in the impedance density. In addition, compared to the gel electrode causing skin irritation, the dry gold electrode shows excellent skin compatibility while having intimate contact with the skin (Figure 1D). The overall process that uses EOG signals from eye movements for various applications is well described in Figure 1E. With two electrode channels, we measure EOG data that are preprocessed, filtered, and classified. Figure 1E also shows an example of a radio-controlled (RC) car via EOG, as demonstrated in this work.

Figure 1.

Figure 1

Overview of an integrated bioelectronic system for detecting eye movements and persistent HMI. (A) Photographs of an all-in-one wireless wearable system for EOG-based HMI. (B) Photographs of a skin-like nanomembrane electrode with a zoom-in view (top) and skin-mounted view (bottom) around the eye. (C) Comparison of EOG signals detected by a conventional gel electrode and a dry gold electrode, capturing the higher performance of the dry one. (D) Skin rash after removal of the gel electrode from the skin (top-left) compared to no adverse event for the skin with the dry electrode (bottom-left), continuous mounting of the dry one on the skin for multiple hours (right). (E) Overview of signal processing steps from EOG detection to classification for an example of EOG-based RC car control, as demonstrated in this work.

Fabrication and Characterization of a Wearable EOG Device System

Recent wearable devices use hard–soft material integration, nanomanufacturing, and chip packaging technologies.3234 In this work, we combine thin-film metallization, laser manufacturing, 3D printing, and system integration to develop a fully integrated all-in-one wearable EOG platform. The base structure uses TPU made by 3D printing, which includes a set of nanomembrane electrodes and a flexible wireless circuit (Figure 2A). A subject can easily wear the headband device with a size-adjustable mechanism (Figure 2B). For wireless signal detection, the system includes a low-profile, flexible circuit having a Bluetooth-low-energy chip and other chip components (Figure 2C; details are given in Figure S4 and Table S2). Time-varying EOG signals are captured by the fractal gold electrodes at 250 Hz and transmitted to the front analog-to-digital converter (ADS1292). Next, to receive sensor data and regulate circuit operation with a built-in microprocessor, the multiprotocol system-on-chip module (nRF52832), which can process and transmit data over 2.4 GHz, is used via the built-in microprocessor. For multiple uses of the wearable device, the flexible circuit contains a rechargeable lithium-polymer battery, charging magnets, and a switch (Figure S5). In addition, the headband system includes a set of fractal gold electrodes that are transfer-printed to the adhesive side of the medical patch (9907T) using a water-soluble tape (Figure 2D).22,35 A flexible thin-film cable connects the electrodes with the circuit (Figure 2D). The gold electrodes used in this study were fabricated by following multiple manufacturing steps using a coating of a polymer (polyimide) on a soft PDMS substrate, metallization of Cr and Au, and laser micromachining to create stretchable patterns (Figure 2E).

Figure 2.

Figure 2

Materials, designs, and fabrication processes of the headband wearable system. (A) Photographs of the fabricated headband EOG system, integrating a flexible wireless circuit and an array of nanomembrane electrodes. (B) Illustrations of a subject who wears the wearable EOG device, showing the exact location of the two-channel electrodes. (C) Photograph of the flexible circuit with integrated chips for EOG signal detection and wireless transmission via Bluetooth. (D) Photographs of the fractal-patterned gold electrode on a soft fabric, connected to the circuit via a thin-film cable. (E) Fabrication processes of the gold electrode using polymer coating, thin-film deposition, and laser cutting.

Characterization of Mechanical Behavior and Compatibility of the Membrane Electrodes

The mechanical reliability of stretchable electrodes is critical to maintaining the skin-contact quality during real-time continuous EOG detection. Therefore, we conducted a set of computational studies using finite element analysis (FEA), considering cyclic stretching and bending situations when an electrode is mounted on the skin. Figure 3A shows the FEA results of an electrode, showing that the maximum principal strain applied to Au is less than 1% under the tensile and bending strain. The fractal-patterned design was used to manufacture electrodes, and we validated the mechanical reliability (Figure 3B). A microscopic investigation observes mechanical fractures before and after stretching and bending tests, showing no visual damage. In this test, the maximum tensile strain was applied up to 30%, and the bending angle was 180° with a radius of curvature: 6 mm. We chose 30% strain based on prior studies showing that a human’s exterior epithelial tissue can be stretched up to 20% without damage,36 and normal skin deformations do not exceed the selected bending curvature.37 The visual observation of mechanical fracture was further validated by measuring electrical resistance. Figure 3C shows the negligible resistance changes during the electrode stretching and bending. Furthermore, we investigated the skin biocompatibility of a gel electrode and a dry gold electrode using infrared thermography (Figure 3D). While the dry electrode shows no side effects after 8 h of wearing, the gel electrode causes skin irritation and temperature elevation after 4 h (Figure S6). When using the gel electrode, adhesive pads mounting the rigid electrode to the skin remove dead skin cells from the epidermis, causing skin rashes.

Figure 3.

Figure 3

Characterization of mechanical behavior and compatibility of the membrane electrodes. (A) Computational study results of mechanical behavior estimation of an electrode with stretching and bending. (B) Experimental validation of the electrode’s reliability with stretching and bending. (C) Resistance measurements of the electrode to quantify the reliability, showing negligible changes during stretching and bending. (D) Comparison of skin biocompatibility of a gel electrode (left) and a gold dry electrode (right), showing that the dry electrode has no side effects while the gel electrode causes skin irritation and elevates temperature.

Optimization of Real-Time Classification via Signal Processing and Feature Extraction

A flow chart in Figure 4A shows a step-by-step sequence of processing of measured EOG signals from the wearable device; four corresponding graphs on the right show examples of processed signals after each step, including bandpass filter, DC offset, detrend, and classification. In this process, the EOG raw data are received by a Python program through Bluetooth. Since EOG data mainly contain low frequencies (sampling rate: 250 Hz), a third-order Butterworth finite impulse response filter (FIR) is used to remove noise.38 FIR is a bandpass filter widely used in digital signal processing, showing an excellent linear phase character.39 To remove the DC offset, the first offset value is removed from others. As a result, measured signals can show trends that are not intrinsic to the data. To eliminate this trend, detrend function is used. Lastly, filtered data from three different steps are classified by evaluating the magnitude. The classified data are converted into a signal with a size of 1, and the direction of the eye is classified according to the code. A set of representative EOG signals in Figure 4B show raw data from four different eye movements. Among them, the horizontal direction of the eyes is channel 1, and the vertical direction of the eyes is channel 2. After signal processing, these signals are classified as left, right, up, and down motions (Figure 4C). A real-time demonstration in Video S1 shows how this process works.

Figure 4.

Figure 4

Optimization of signal processing and feature extraction for real-time data classification. (A) Step-by-step sequence of processing of measured EOG signals from the wearable device; four different graphs on the right show examples of processed signals after each step, including bandpass filter, remove DC offset, detrend, and classification. (B) Representative examples of raw EOG signals and corresponding eye movements, including left, right, up, and down motions. (C) Set of classified signals after filtering and feature extraction.

Development and Comparison of Machine Learning Algorithms for Data Classification

Prior studies show the limitation of signal processing when detecting more than five classes;28 with six classes, the accuracy was only 91.25%. According to other studies, a kNN algorithm is more efficient when classifying EOG signals than decision tree and support vector machine methods.40,41 The kNN classification uses the nearest distance metric and the neighbor’s number k value. When one of the parameters is varying, another parameter is fixed.41,42 In this kNN algorithm, testing data are classified by finding the greatest number, with the closest relative distance to neighbors; each neighbor belongs to a specific class. Figure 5A shows an example of a kNN classification where the test candidate is classified as either blue squares or red circles. If k = 3, the candidate is assigned to the red circles (2 red circles > 1 blue square). If k = 6, the candidate is again assigned to the red circles (5 red circles > 1 blue square). To compare the performance of machine learning algorithms, we also developed a CNN classifier. The details of the overall CNN classification processes are shown in Figure 5B. The CNN model featuring layers of one-dimensional convolutions consists of two kinds of modules. Then, this model is followed by filters of flattening and a dense–softmax output. In this study, EOG data collected by the wearable device are split into the training set (75%) and the test set (25%). The preprocessed data are transferred to either the kNN or the CNN classifier. Then, the test data set is analyzed by comparing the training data set. Each model predicts the test results and shows the results through the confusion matrixes. Figure 5C summarizes and compares the performance of signal processing and two machine learning methods. There are multiple eye movements, including up (U), down (D), left (L), right (R), blink (B), and null (N), used in this study. The signal processing method with six classes shows an accuracy of 95.5%. Compared to that, kNN and CNN methods with six classes show higher accuracies, 96.9 and 98.3%, respectively. Confusion matrixes from signal processing, kNN, and CNN algorithms including accuracies of each class are shown in Figure S7. Overall, the CNN classification result shows the highest accuracy among reported articles that detect EOG signals. Also, it takes less than a second from pre-processing to real-time classification. Table 1 captures the advantages of our wearable system and superior classification performance compared to prior studies. Table 1 also shows that our wearable system is compact and flexible by comparing previous EOG devices based on the size and type of circuits.

Figure 5.

Figure 5

Development and comparison of machine learning algorithms for data classification. (A) Illustration of an example of a kNN classifier. (B) Flowchart showing a spatial CNN model with filters of decreasing size, flattening, and followed by the dense–softmax output. (C) Comparison of data classification results, showing three different confusion matrices, including the signal processing outcome (left), kNN algorithm (middle), and CNN algorithm (right); the highest accuracy is achieved by the CNN method. In each confusion matrix, D means down, L means left, N means null, R means right, U means up, and B means blink.

Table 1. Comparison of Wearable EOG Devices for Classification of Eye Movements and HMI Applications.

references circuit type form factor (cm × cm) integrated single device electrode type no. of electrodes no. of classes classification accuracy (%) platform HMI application
this work flexible 2 × 3 yes dry, membrane (gold) 4 6 98.3 headband with a flexible circuit two-wheeled RC car
(14) rigid, bulky 12 × 15 no dry (gold) 5 4 94 wire-connection with a circuit wheel-chair
(12) rigid, bulky 6 × 6 no dry (gold) 5 5 92 wire-connection with a circuit drone helicopter
(43) rigid, bulky 4 × 6 no dry, (graphene) 3 5 98 headband with rigid circuit RC car
(23) rigid, bulky 5.5 × 8 no dry, (graphene) 3 2   headband with rigid circuit LED array
(44) rigid, bulky 6 × 6 no dry (Ag/AgCl) 5 4   eyeglass  
(45) rigid, bulky 5 × 5 no dry (Ag/AgCl) 4 1   headband with rigid circuit  
(46) rigid, bulky 3 × 20 no dry (silver) 3 2   headband with rigid circuit  
(47) rigid, bulky 5 × 7 no gel (Ag/AgCl) 5 5 75.5 eyeglass omnidirectional-robot
(48) rigid, bulky 4 × 8 no gel (Ag/AgCl) 5 5 97 armband game
(49) rigid, bulky 11 × 16 no gel (Ag/AgCl) 5 5 89 helmet keyboard
(28) rigid, bulky 4 × 8 no gel (Ag/AgCl) 4 6 91.3 rigid headband wheelchair + keyboard

Demonstration of Wireless Real-Time Control of a RC Car with the Wearable Device

In this work, we demonstrate an example of persistent wireless HMI using the headband wearable device and EOG signals (Figure 6). Multiple eye movements, detected by sensors, could successfully control a two-wheeled RC car by accurately following the designated pathway and avoiding an obstacle. Figure 6A captures a photograph showing a subject who wears the sensor-integrated headband, a tablet capturing the real-time EOG signals, and a two-wheeled RC car to control. In Figure 6B, the top photograph shows a control track with an obstacle that the car follows, while the bottom photograph captures the zoomed-in view of an Android app for displaying EOG signals and real-time classification outcomes. In this demonstration, we use five different control commands, including up, down, blink, left (CCW; counter-clockwise), and right (CW; clockwise) motions (Figure 6C). Considering an emergency case during operation, the blink command immediately stops the car when unintended eye movements are classified. The two-wheeled RC car follows eye movements of a subject who wears the device, which moves the car from the starting position to the parking location. Seven consecutive commands are delivered to the car (Figure 6D), including (1) go forward, (2) CCW rotation and go forward, (3) CW rotation, (4) go forward, (5) CW rotation and go forward, (6) CW rotation, and (7) go reverse to park. The real-time control of this car using eye movements appears in Video S2.

Figure 6.

Figure 6

Demonstration of wireless real-time control of a two-wheeled RC car with the wearable device. (a) Photograph showing a subject who wears the wearable EOG device, a tablet capturing the real-time EOG signals, and a two-wheeled RC car to control. (b) Top photograph showing a control track with an obstacle that the car follows, while the bottom photograph showing the zoomed-in view of an Android app for displaying EOG signals and real-time classification outcomes. (c) Five different control commands used for the car, including up, down, blink, left (CCW; counter-clockwise), and right (CW; clockwise) motions. (d) Top-view photograph capturing the pathway that the car follows while avoiding the obstacle; in this demonstration, seven commands are used from starting to parking.

Conclusions

This paper reports a comprehensive set of studies that develop a soft headband bioelectronic system and persistent HMI using EOG signals. The wearable headband platform offers a firm contact of stretchable electrodes with the skin, which also can be worn by different users with various head sizes. A simplified manufacturing process, including metal deposition and laser cutting, fabricates an array of thin-film dry electrodes without needing conductive gels for high-quality EOG recording. The highly stretchable and flexible electrode shows reliability in cyclic mechanical tests while demonstrating excellent skin compatibility over 8 h. The fractal-patterned gold electrode could be repeatedly used throughout this study, but quantification of the reusability of the electrodes will be included in future work. Measured EOG signals are filtered and classified by a signal processing method and kNN and CNN algorithms. Among them, the CNN-based classification shows the highest accuracy of 98.3% with six classes. Demonstration of wireless real-time control of a two-wheeled RC car captures the performance of the wearable device for persistent HMI. In this study, seven commands using eye movements could successfully control a car on a confined track while avoiding an obstacle. Future studies will address limitations, such as crosstalk between vertical and horizontal channels or EEG and EMG signals.

Experimental Section

Fabrication of the Integrated Wearable System

The wearable EOG device consists of a fractal gold electrode, headband-type platform, and flexible circuit. PDMS (Sylgard 184, Dow) was spin-coated on a clean glass slide. An 8.47 μm thick polyimide sheet (Kapton film, DuPont) was laminated onto the PDMS-coated glass slide first, followed by a 5 nm thick Cr layer and 200 nm thick Au layer that was deposited using an electron beam deposition tool (Denton Explorer), respectively. We studied open-mesh structured fractal patterns (a bending radius of 0.39 mm and a trace width of 0.16 mm). The fractal pattern was cut by a femtosecond IR laser micromachining tool (WS-Flex, Optec), which is a multi-purpose, high-precision processing tool for various materials. The cut fractal pattern was transferred using the water-soluble tape (ASWT-2, Aquasol) from the PDMS. The wearable 3D headband platform was designed by SolidWorks and printed by a 3D printer (Cubicon Single Plus 3DP-310F) with TPU filaments (Cubicon TPU Filament). TPU filaments are flexible with superior strength. We designed the headband-type platform that can be resized according to the head size through the tension string and the auxiliary equipment (Figure S8). Chip components on the flexible circuit were soldered to the plate with a solder paste (SMDLTLFP10T5, Chip Quik) and then heated at 100 °C. The set temperature increased by 10 °C every minute to a final temperature of 150 °C. A small lithium polymer battery (capacity: 40 mA h, Digi-Key) was modified to allow for easy charging by connecting two charging magnets and a switch to the battery. The circuit with a 40 mA h battery lasted 5.1 h, which is around 8 mA power consumption.31 The flexible circuit was attached to the back of the headband-type platform. The fractal gold electrodes were connected to the circuit via encapsulated ACF wires. Lastly, the electrodes were attached to the tension string.

Finite Element Analysis

This work includes the results of studied FEA to investigate a fractal gold electrode’s mechanical behaviors using commercial software (ANSYS). This analysis focused on the mechanical fracture of the electrode upon cyclic bending and stretching. The modeling analyzed the maximum principal strain in the electrode consisting of three layers: an 8.47 μm thick polyimide sheet, 5 nm thick Cr layers, and 200 nm thick Au layers (Figure S9). Table S1 shows the details of the material properties (Young’s modulus and Poisson’s ratio). One side of the substrate is fixed as a support fix, and the other side is moved using the displacement function. The boundary conditions were applied to the Ecoflex substrate.

Experimental Study of Mechanical Behavior

A customized stretcher conducted the axial stretching test. Two clamps held the sample. The strains were determined by controlling the distance from 0 to 30%. The bending test was conducted manually by a rigid circular cylinder. The bendability from 0 to 180° of the fractal gold electrodes was assessed manually with a bending radius of 6 mm (details are given in Figure S10). A digital multimeter is used to measure and record a resistance change on the fractal gold electrode.

Data Acquisition and Training

To detect EOG signals, two electrodes were positioned 1 cm above each eye. One electrode was placed 1 cm below the left lower eyelid for vertical eye movement. A common grounding electrode was placed on the middle of the forehead (Figure S11). Before obtaining the data, the skin (electrode position) was optionally wiped with alcoholic cotton to remove foreign matter. The eyes were moved in six eye movements (left, right, up, down, blink, and null), and the gaze was within 1 s. The raw EOG signals from the wearable EOG device were measured and recorded by an Android tablet via Bluetooth. The custom Android application simultaneously transmits and exports data for channels 1 and 2. As shown in Figure S12, MATLAB labeled the acquired EOG data to train the CNN classifier. Then, the EOG data were trained and modeled by a machine learning algorithm and TensorFlow platform. The modeled file classifies the subject’s EOG signals in real-time through the machine learning interface and TensorFlow platform in the Android application.

Analysis of the SNR

The experiment was conducted by looking left and right three times at regular intervals for 5 s in this recording. The raw data were recorded into five s segments (five total). This analysis involves the measurement of EOG signal size and removal of the average value of the EOG signal using the following equation: SNR (dB) = 10 Log10(rms_signal/rms_noise). The results and the standard error were calculated as an average over the number of recordings.

Human Subject Study

The human pilot study involved multiple healthy volunteers; the study followed the approved IRB protocol from the Georgia Institute of Technology (no. H20226). All participants agreed and signed the consent form to allow the experiment procedure.

Acknowledgments

We acknowledge the support of the IEN Center Grant from the Georgia Tech Institute for Electronics and Nanotechnology and the support of the Korea Medical Device Development Fund (Ministry of Science and ICT, Ministry of Trade, Industry and Energy, Ministry of Health & Welfare, Ministry of Food and Drug Safety; project number: 1711138229, KMDF_PR_20200901_0124). Electronic devices in this work were fabricated at the Institute for Electronics and Nanotechnology, a member of the National Nanotechnology Coordinated Infrastructure, which is supported by the National Science Foundation (grant ECCS-2025462). J.-H.K. acknowledges the partial support of the National Science Foundation (CBET-1707056).

Supporting Information Available

The Supporting Information is available free of charge at https://pubs.acs.org/doi/10.1021/acsaelm.2c01436.

  • Device bending test; experimental setup for EOG sensitivity; impedance measurement with electrodes; functional chips used in the flexible wireless circuit; battery components; skin rash test; comparison of the detailed confusion matrix; wearable EOG device assembly; FEA simulation modeling; tensile and bending tests; locations of the dry gold electrodes; and raw EOG data and classified result (PDF)

  • Monitoring of EOG (MP4)

  • Demonstration of a mini-drone car (MP4)

Author Contributions

S.B. and Y.J.L. contributed equally to this work. J.W.C. and W.-H.Y. conceived and designed the research; S.B., Y.J.L., S.K., and Y.-S.K. performed the experiment and analyzed the data; and S.B., Y.J.L., J.W.C., J.-H.K., and W.-H.Y. wrote the paper.

The authors declare the following competing financial interest(s): Georgia Tech has a US patent pending technology.

Supplementary Material

el2c01436_si_001.pdf (2.2MB, pdf)
el2c01436_si_002.mp4 (10.7MB, mp4)
el2c01436_si_003.mp4 (15.4MB, mp4)

References

  1. Kartsch V.; Guermandi M.; Benatti S.; Montagna F.; Benini L.. An Energy-Efficient IoT node for HMI applications based on an ultra-low power Multicore Processor. IEEE Sensors Applications Symposium (SAS), 2019; IEEE, 2019; pp 1–6.
  2. Kaur A. Wheelchair control for disabled patients using EMG/EOG based human machine interface: a review. J. Med. Eng. Technol. 2021, 45, 61–74. 10.1080/03091902.2020.1853838. [DOI] [PubMed] [Google Scholar]
  3. Bulling A.; Roggen D.; Tröster G. Wearable EOG goggles: Seamless sensing and context-awareness in everyday environments. J. Ambient Intell. Smart Environ. 2009, 1, 157–171. 10.3233/ais-2009-0020. [DOI] [Google Scholar]
  4. Mala S.; Latha K. Feature selection in classification of eye movements using electrooculography for activity recognition. Comput. Math. Methods Med. 2014, 2014, 713818. 10.1155/2014/713818. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bulling A.; Ward J. A.; Gellersen H.; Tröster G. Eye movement analysis for activity recognition using electrooculography. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 741–753. 10.1109/TPAMI.2010.86. [DOI] [PubMed] [Google Scholar]
  6. Poon C. C.; Leung E. Y.; Lau K. C.; Leung B. H.; Zheng Y. L.; Chiu P. W.; Yam Y.. A novel user-specific wearable controller for surgical robots. International Conference of Design, User Experience, and Usability; Springer, 2015; pp 693–701.
  7. Gray V.; Rice C. L.; Garland S. J. Factors that influence muscle weakness following stroke and their clinical implications: a critical review. Physiother. Can. 2012, 64, 415–426. 10.3138/ptc.2011-03. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Lum P. S.; Godfrey S. B.; Brokaw E. B.; Holley R. J.; Nichols D. Robotic approaches for rehabilitation of hand function after stroke. Am. J. Phys. Med. Rehabil. 2012, 91, S242–S254. 10.1097/phm.0b013e31826bcedb. [DOI] [PubMed] [Google Scholar]
  9. Xiao R.; Ding L. Evaluation of EEG features in decoding individual finger movements from one hand. Comput. Math. Methods Med. 2013, 2013, 243257. 10.1155/2013/243257. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Dey N.Classification and Clustering in Biomedical Signal Processing; IGI global, 2016. [Google Scholar]
  11. Siddiqui U.; Shaikh A. N. An overview of electrooculography. Int. J. Adv. Res. Comput. Commun. Eng. 2013, 2, 4328–4330. [Google Scholar]
  12. Ameri S. K.; Kim M.; Kuang I. A.; Perera W. K.; Alshiekh M.; Jeong H.; Topcu U.; Akinwande D.; Lu N. Imperceptible electrooculography graphene sensor system for human–robot interface. npj 2D Mater. Appl. 2018, 2, 19. 10.1038/s41699-018-0064-4. [DOI] [Google Scholar]
  13. Park S.; Kim H.; Kim J.-H.; Yeo W.-H. Advanced nanomaterials, printing processes, and applications for flexible hybrid electronics. Materials 2020, 13, 3587. 10.3390/ma13163587. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Mishra S.; Norton J. J. S.; Lee Y.; Lee D. S.; Agee N.; Chen Y.; Chun Y.; Yeo W. H. Soft, conformal bioelectronics for a wireless human-wheelchair interface. Biosens. Bioelectron. 2017, 91, 796–803. 10.1016/j.bios.2017.01.044. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Aziz F.; Arof H.; Mokhtar N.; Mubin M. HMM based automated wheelchair navigation using EOG traces in EEG. J. Neural. Eng. 2014, 11, 056018. 10.1088/1741-2560/11/5/056018. [DOI] [PubMed] [Google Scholar]
  16. Barea R.; Boquete L.; Mazo M.; Lopez E. System for assisted mobility using eye movements based on electrooculography. IEEE Trans. Neural Syst. Rehabil. Eng. 2002, 10, 209–218. 10.1109/TNSRE.2002.806829. [DOI] [PubMed] [Google Scholar]
  17. Belkacem A. N.; Shin D.; Kambara H.; Yoshimura N.; Koike Y. Online classification algorithm for eye-movement-based communication systems using two temporal EEG sensors. Biomed. Signal Process. Control 2015, 16, 40–47. 10.1016/j.bspc.2014.10.005. [DOI] [Google Scholar]
  18. Wu S.-L.; Liao L.-D.; Lu S.-W.; Jiang W.-L.; Chen S.-A.; Lin C.-T. Controlling a human–computer interface system with a novel classification method that uses electrooculography signals. IEEE Trans. Biomed. Eng. 2013, 60, 2133. 10.1109/TBME.2013.2248154. [DOI] [PubMed] [Google Scholar]
  19. Ban S.; Lee Y. J.; Kim K. R.; Kim J.-H.; Yeo W.-H. Advances in Materials, Sensors, and Integrated Systems for Monitoring Eye Movements. Biosensors 2022, 12, 1039. 10.3390/bios12111039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Dhuliawala M.; Lee J.; Shimizu J.; Bulling A.; Kunze K.; Starner T.; Woo W.. Smooth eye movement interaction using EOG glasses. Proceedings of the 18th ACM International Conference on Multimodal Interaction, 2016; pp 307–311.
  21. Lee J. H.; Kim H.; Hwang J.-Y.; Chung J.; Jang T.-M.; Seo D. G.; Gao Y.; Lee J.; Park H.; Lee S.; Moon H. C.; Cheng H.; Lee S.-H.; Hwang S.-W. 3D printed, customizable, and multifunctional smart electronic eyeglasses for wearable healthcare systems and human–machine Interfaces. ACS Appl. Mater. Interfaces 2020, 12, 21424–21432. 10.1021/acsami.0c03110. [DOI] [PubMed] [Google Scholar]
  22. Kim Y. S.; Mahmood M.; Lee Y.; Kim N. K.; Kwon S.; Herbert R.; Kim D.; Cho H. C.; Yeo W. H. All-in-one, wireless, stretchable hybrid electronics for smart, connected, and ambulatory physiological monitoring. Adv. Sci. 2019, 6, 1900939. 10.1002/advs.201900939. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Golparvar A. J.; Yapici M. K. Graphene smart textile-based wearable eye movement sensor for electro-ocular control and interaction with objects. J. Electrochem. Soc. 2019, 166, B3184. 10.1149/2.0241907jes. [DOI] [Google Scholar]
  24. Yaramothu C.; Vito d’Antonio- Bertagnolli J.; Santos E. M.; Crincoli P. C.; Rajah J. V.; Scheiman M.; Alvarez T. L. Proceedings no. 37: Virtual Eye Rotation Vision Exercises (VERVE): A Virtual Reality Vision Therapy Platform with Eye Tracking. Brain Stimul. 2019, 12, e107–e108. 10.1016/j.brs.2018.12.206. [DOI] [Google Scholar]
  25. Mishra S.; Kim Y.-S.; Intarasirisawat J.; Kwon Y.-T.; Lee Y.; Mahmood M.; Lim H.-R.; Herbert R.; Yu K. J.; Ang C. S.; Yeo W. H. Soft, wireless periocular wearable electronics for real-time detection of eye vergence in a virtual reality toward mobile eye therapies. Sci. Adv. 2020, 6, eaay1729 10.1126/sciadv.aay1729. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Yu K. J.; Kim T.; Shin Y.; Kang K.; Kim K.; Kim G.; Byeon Y.; Kim H.; Gao Y.; Kim J.; et al. Ultra-thin crystalline silicon-based strain gauges with deep learning algorithms for silent speech interfaces. Nat. Commun. 2022, 13, 5815. 10.1038/s41467-022-33457-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Sang M.; Kang K.; Zhang Y.; Zhang H.; Kim K.; Cho M.; Shin J.; Hong J. H.; Kim T.; Lee S. K.; Yeo W. H.; Lee J. W.; Lee T.; Xu B.; Yu K. J. Ultrahigh Sensitive Au-Doped Silicon Nanomembrane Based Wearable Sensor Arrays for Continuous Skin Temperature Monitoring with High Precision. Adv. Mater. 2022, 34, 2105865. 10.1002/adma.202105865. [DOI] [PubMed] [Google Scholar]
  28. Heo J.; Yoon H.; Park K. S. A novel wearable forehead EOG measurement system for human computer interfaces. Sensors 2017, 17, 1485. 10.3390/s17071485. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Lopez A.; Ferrero F. J.; Valledor M.; Campo J. C.; Postolache O.. A study on electrode placement in EOG systems for medical applications. IEEE International symposium on medical measurements and applications (MeMeA), 2016; IEEE, 2016; pp 1–5.
  30. Zhang Y.-F.; Gao X.-Y.; Zhu J.-Y.; Zheng W.-L.; Lu B.-L.. A novel approach to driving fatigue detection using forehead EOG. 7th International IEEE/EMBS Conference on Neural Engineering (NER), 2015; IEEE, 2015; pp 707–710.
  31. Mahmood M.; Kwon S.; Berkmen G. K.; Kim Y.-S.; Scorr L.; Jinnah H.; Yeo W.-H. Soft nanomembrane sensors and flexible hybrid bioelectronics for wireless quantification of blepharospasm. IEEE Trans. Biomed. Eng. 2020, 67, 3094–3100. 10.1109/tbme.2020.2975773. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kim J.; Salvatore G. A.; Araki H.; Chiarelli A. M.; Xie Z.; Banks A.; Sheng X.; Liu Y.; Lee J. W.; Jang K.-I.; Heo S. Y.; Cho K.; Luo H.; Zimmerman B.; Kim J.; Yan L.; Feng X.; Xu S.; Fabiani M.; Gratton G.; Huang Y.; Paik U.; Rogers J. A. Battery-free, stretchable optoelectronic systems for wireless optical characterization of the skin. Sci. Adv. 2016, 2, e1600418 10.1126/sciadv.1600418. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Kim Y. S.; Lu J.; Shih B.; Gharibans A.; Zou Z.; Matsuno K.; Aguilera R.; Han Y.; Meek A.; Xiao J.; Tolley M. T.; Coleman T. P. Scalable manufacturing of solderable and stretchable physiologic sensing systems. Adv. Mater. 2017, 29, 1701312. 10.1002/adma.201701312. [DOI] [PubMed] [Google Scholar]
  34. Lee S. P.; Ha G.; Wright D. E.; Ma Y.; Sen-Gupta E.; Haubrich N. R.; Branche P. C.; Li W.; Huppert G. L.; Johnson M.; Mutlu H. B.; Li K.; Sheth N.; Wright J. A.; Huang Y.; Mansour M.; Rogers J. A.; Ghaffari R. Highly flexible, wearable, and disposable cardiac biosensors for remote and ambulatory monitoring. NPJ Digit. Med. 2018, 1, 2–8. 10.1038/s41746-017-0009-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Kim Y.-S.; Basir A.; Herbert R.; Kim J.; Yoo H.; Yeo W.-H. Soft materials, stretchable mechanics, and optimized designs for body-wearable compliant antennas. ACS Appl. Mater. Interfaces 2019, 12, 3059–3067. 10.1021/acsami.9b20233. [DOI] [PubMed] [Google Scholar]
  36. Yeo W. H.; Kim Y. S.; Lee J.; Ameen A.; Shi L.; Li M.; Wang S.; Ma R.; Jin S. H.; Kang Z.; Huang Y.; Rogers J. A. Multifunctional epidermal electronics printed directly onto the skin. Adv. Mater. 2013, 25, 2773–2778. 10.1002/adma.201204426. [DOI] [PubMed] [Google Scholar]
  37. Hattori Y.; Falgout L.; Lee W.; Jung S. Y.; Poon E.; Lee J. W.; Na I.; Geisler A.; Sadhwani D.; Zhang Y.; Su Y.; Wang X.; Liu Z.; Xia J.; Cheng H.; Webb R. C.; Bonifas A. P.; Won P.; Jeong J.-W.; Jang K.-I.; Song Y. M.; Nardone B.; Nodzenski M.; Fan J. A.; Huang Y.; West D. P.; Paller A. S.; Alam M.; Yeo W.-H.; Rogers J. A. Multifunctional skin-like electronics for quantitative, clinical monitoring of cutaneous wound healing. Adv. Healthcare Mater. 2014, 3, 1597–1607. 10.1002/adhm.201400073. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Merino M.; Rivera O.; Gómez I.; Molina A.; Dorronzoro E.. A method of EOG signal processing to detect the direction of eye movements. First International Conference on Sensor Device Technologies and Applications, 2010; IEEE, 2010; pp 100–105.
  39. Qiu H.; Guo Z.; Zhang X.. The design of FIR band-pass filter with improved distributed algorithm based on FPGA. International Conference on Multimedia Technology, 2010; IEEE, 2010; pp 1–4.
  40. Hayawi A. A.; Waleed J.. Driver’s drowsiness monitoring and alarming auto-system based on EOG signals. 2nd International Conference on Engineering Technology and its Applications (IICETA), 2019; IEEE, 2019; pp 214–218.
  41. Syal P.; Kumari P.. Comparative Analysis of KNN, SVM, DT for EOG based Human Computer Interface. International Conference on Current Trends in Computer, Electrical, Electronics and Communication (CTCEEC), 2017; IEEE, 2017; pp 1023–1028.
  42. Mustafa M.; Nasir Taib M.; Hj. Murat Z.; Sulaiman N. Comparison between KNN and ANN classification in brain balancing application via spectrogram image. J. Comput. Sci. Comput. Math. 2012, 2, 17–22. 10.20967/jcscm.2012.04.004. [DOI] [Google Scholar]
  43. Golparvar A. J.; Yapici M. K. Toward graphene textiles in wearable eye tracking systems for human–machine interaction. Beilstein J. Nanotechnol. 2021, 12, 180–189. 10.3762/bjnano.12.14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Vourvopoulos A.; Niforatos E.; Giannakos M.. EEGlass: An EEG-eyeware prototype for ubiquitous brain-computer interaction. Adjunct proceedings of the 2019 ACM international joint conference on pervasive and ubiquitous computing and proceedings of the 2019 ACM international symposium on wearable computers, 2019; pp 647–652.
  45. Tabal K. M.; Cruz J. D.. Development of low-cost embedded-based electrooculogram blink pulse classifier for drowsiness detection system. IEEE 13th International Colloquium on Signal Processing & its Applications (CSPA), 2017; IEEE, 2017; pp 29–34.
  46. Vehkaoja A. T.; Verho J. A.; Puurtinen M. M.; Nojd N. M.; Lekkala J. O.; Hyttinen J. A.. Wireless head cap for EOG and facial EMG measurements. IEEE Engineering in Medicine and Biology 27th Annual Conference, 2006; IEEE, 2005; pp 5865–5868. [DOI] [PubMed]
  47. Pérez-Reynoso F. D.; Rodríguez-Guerrero L.; Salgado-Ramírez J. C.; Ortega-Palacios R. Human–Machine Interface: Multiclass Classification by Machine Learning on 1D EOG Signals for the Control of an Omnidirectional Robot. Sensors 2021, 21, 5882. 10.3390/s21175882. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. López A.; Fernández M.; Rodríguez H.; Ferrero F.; Postolache O. Development of an EOG-based system to control a serious game. Measurement 2018, 127, 481–488. 10.1016/j.measurement.2018.06.017. [DOI] [Google Scholar]
  49. O’Bard B.; Larson A.; Herrera J.; Nega D.; George K.. Electrooculography based iOS controller for individuals with quadriplegia or neurodegenerative disease. IEEE International Conference on Healthcare Informatics (ICHI), 2017; IEEE, 2017; pp 101–106.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

el2c01436_si_001.pdf (2.2MB, pdf)
el2c01436_si_002.mp4 (10.7MB, mp4)
el2c01436_si_003.mp4 (15.4MB, mp4)

Articles from ACS Applied Electronic Materials are provided here courtesy of American Chemical Society

RESOURCES