Abstract
Historically, individuals with hearing impairments have faced neglect, lacking the necessary tools to facilitate effective communication. However, advancements in modern technology have paved the way for the development of various tools and software aimed at improving the quality of life for hearing-disabled individuals. This research paper presents a comprehensive study employing five distinct deep learning models to recognize hand gestures for the American Sign Language (ASL) alphabet. The primary objective of this study was to leverage contemporary technology to bridge the communication gap between hearing-impaired individuals and individuals with no hearing impairment. The models utilized in this research include AlexNet, ConvNeXt, EfficientNet, ResNet-50, and VisionTransformer were trained and tested using an extensive dataset comprising over 87,000 images of the ASL alphabet hand gestures. Numerous experiments were conducted, involving modifications to the architectural design parameters of the models to obtain maximum recognition accuracy. The experimental results of our study revealed that ResNet-50 achieved an exceptional accuracy rate of 99.98%, the highest among all models. EfficientNet attained an accuracy rate of 99.95%, ConvNeXt achieved 99.51% accuracy, AlexNet attained 99.50% accuracy, while VisionTransformer yielded the lowest accuracy of 88.59%.
Keywords: image-based, American sign language, deep learning, transfer learning, AlexNet, ConvNeXt, EfficientNet, ResNet-50, VisionTransformer
1. Introduction
Throughout history, humans have employed a variety of communication techniques including gesturing, sounds, drawing, writing, and speaking. However, for people with deafness or hearing impairments, sign language is the primary means of communication and interaction with others, breaking down all barriers of hearing loss condition which severely limits their verbal communication. Because of communication barriers, people with these disabilities have fewer opportunities for development. Sign language is a spontaneous non-verbal language expressed by using manual gestures, facial expressions, and body language to convey messages and meaning. These signs may vary from one country to another, although they have some similarities in using sign language. Unfortunately, there is no universal sign language that can be used for all people with hearing impairments around the world [1]. According to the World Health Organization’s (WHO) most recent research, 5% of the world’s population—432 million adults and 34 million children—have disabling hearing loss, not to mention more than 1 billion people are susceptible to hearing loss due to extended and excessive to loud sounds [2]. In fact, people who lose their hearing sense under any circumstances will lose their ability to speak. The enormous number of deafness and hearing loss conditions has garnered the attention of many researchers and developers in the field of speech recognition and other multidisciplinary fields to conduct their study to assist people with hearing impairments. Their goal is to facilitate the daily life of people with hearing disabilities for communication and social interaction with other individuals. Consequently, with the rapidly growing deaf community, building a sign language recognition system using deep learning technology plays a vital role in interpreting sign language to ordinary individuals and the reverse. This system would ease the process of communication between deaf and normal people. As a result, people with hearing impairments will have the opportunity to become more engaged in society, developing social interaction and relationships [3,4]. Nowadays, the sole means for people with hearing impairments to communicate with other ordinary people is through interpreters. However, it is very costly to hire interpreters who have expertise in interpreting for the deaf, because of the limited number of such interpreters. Nevertheless, there are several obstacles in implementing a sign language recognition system to support the deaf and hearing loss community that should be discussed. Firstly, not all hearing-impaired individuals use sign language as a method of communication, which may give them a sense of isolation and depression. Secondly, there are more than 200 different sign languages and dialects from different countries which may delay the process of implementing a sign language recognition system which would be applicable in various countries [4,5]. Lastly, not everyone is proficient in using today’s modern technology due to a lack of education and development, which can be neglected. Numerous studies and research should be oriented to address and comprehend the obstacles that deaf and hard-of-hearing people encounter, which hinder their societal engagement [6]. The pivotal contribution of our research paper can be illustrated as follows:
We design a scheme based on deep learning technology to enhance the classification of American Sign Language Alphabet hand gestures.
We fit five deep learning models to classify and recognize hand gestures with subtle disparities in shape. These models included AlexNet, ConvNeXt, EfficientNet, ResNet50, and VisionTransformer.
We evaluate the performance of our scheme in terms of accuracy, precision, recall, and F1-score.
Our scheme outperformed the recent studies utilizing the same dataset.
This research paper is organized as follows: Section 2 provides background information about machine and deep learning techniques. Section 3 highlights relevant research and various methods of deep learning used in sign language recognition systems. Section 4 expounds upon the methodology employed in this study, elucidating the approach and techniques utilized. Section 5 is devoted to the results and discussion of using five different deep learning models. Finally, we present a general conclusion in Section 6.
2. Background
The Automated Sign Language Recognition System (SLRS) garnered significant attention from researchers and developers in recent years. This collaborative research area facilitates the construction of sign language recognition systems aimed at supporting individuals with hearing impairments in overcoming communication barriers. In the realm of sign recognition systems, numerous machine and deep learning algorithms have been developed to enable effective communication between individuals with hearing impairments and others. To provide readers with a comprehensive understanding of these algorithms, it is imperative to present a concise introduction to the principles and technologies underlying machine and deep learning. By exploring the foundations of these technologies, readers will gain the necessary background knowledge to comprehend the subsequent discussions on sign language recognition systems.
2.1. Machine Learning
Artificial intelligence (AI) is the process of making machines as intelligent as the human brain to carry out a variety of advanced functions. Machine learning (ML), a subset of AI, focuses on developing computer algorithms capable of learning from data and improving performance through experience. Deep learning, on the other hand, represents a specialized field within ML that utilizes neural networks with multiple layers to extract high-level representations and features. This hierarchical architecture enables models to achieve exceptional performance in various domains. Figure 1 illustrates the relationship between machine learning and deep learning.
The idea of “machine learning” encompasses a variety of stochastic algorithms that can be applied to make intelligent predictions and engage in problem solving based on data. These predictions and the ability to resolve issues automatically improve as the machine learns and generates knowledge from experience. ML is an area in which experts are developing computer algorithms that can access data and learn on their own without being explicit to human intervention [7]. Dealing with massive amounts of data to analyze and extract valuable information requires extraordinary human effort, but with ML, it only takes a few seconds to analyze these data and obtain accurate outcomes [8]. ML algorithms are widely utilized in numerous applications in different sectors including healthcare, transportation, education, energy, agriculture, industry, and many others. Each of these areas has had transformative growth and remarkable development after implementing ML. For instance, in healthcare, ML shows incredible capabilities in predicting diseases based on medical data sources [9,10].
There are three major types of machine learning methods as represented in Figure 2.
2.1.1. Supervised Learning
Supervised learning generates a function that maps a set of input variables x to an output variable y for desired outcomes.
(1) |
These mapping functions can assist in classification and regression. In classification, the model is attempting to predict class labels based on given input data. The input dataset is divided into a training dataset for training the model and a test dataset for testing the model. Regression is related to predicting the numerical value output based on unobservable data. The supervised learning approach is heavily reliant on labeled datasets that can be utilized to train algorithms to accurately classify data or predict outcomes [11,12,13]. The objective of the ML algorithm is to identify patterns and build mathematical models. For instance, by using the supervised learning technique, we may anticipate house prices for the next five years based on already available data, such as x = the size of the house, number of rooms, zip code, age of the house, and current price. y = Predict the price of the house based on the data of x. Many features can be labeled to help algorithms gain knowledge based on previous and present data [14].
2.1.2. Unsupervised Learning
Unsupervised learning is a technique in which users do not have to observe the model. It relies on analyzing and clustering unlabeled data without knowing any information about class labels for input data. Furthermore, it might be challenging to evaluate the outcomes of unsupervised learning techniques. Unsupervised learning algorithms can locate hidden patterns or data groupings, which facilitate the identification of data similarities and differences [11,12]. These abilities to identify hidden and distinct patterns make unsupervised learning an ideal solution for customer segmentation, image recognition, customer persona investigation, anomaly detection, inventory management, and many other applications. Compared to other ML techniques, unsupervised learning can handle more complex tasks.
2.1.3. Reinforcement Learning
Reinforcement learning is a more sophisticated and challenging ML technique. It acquires knowledge via interaction with a set of environments and feedback. It also learns from mistakes [11,12]. RL does not require labeled input/output pairs or examples and sample data due to its dependency on making decisions. Therefore, RL labels should be assigned to each dependent decision [11,12]. RL frequently works with applications requiring fast reactions and adaptation to change under operational conditions. These applications include vehicular traffic management, inventory control, recommender systems, cloud computing, and robotics [15,16,17].
2.2. Application and Practices of Machine Learning
ML is a technique that employs cognitive abilities to allow computers and other technologies to imitate the human brain’s ability to make decisions, solve issues, learn from mistakes, and perform complicated tasks. ML techniques have recently made substantial advancements and offer a promising future in many aspects of our lives. See Figure 3 for a list of existing ML approaches and applications that we use regularly.
2.3. Deep Learning
Deep learning is a subset of ML. It is primarily focused on creating large artificial neural network models to process large and complex amounts of data compared with ML techniques [18]. It operates more efficiently with unstructured data. Deep learning can resolve perceptual issues such as problems with images, speech recognition, facial recognition, and handwritten character recognition that computers were formerly unable to handle in the past. These capabilities result from deep learning relying on multiple processing layers for pattern recognition [18,19]. Each layer derives knowledge from the data—the higher the layer level, the more knowledge we can obtain to improve accuracy [8]. The key role of deep learning is to classify, recognize, and describe objects within data. Deep learning automatically extracts features without relying on previous data processing. For the sign language recognition system, deep learning is an excellent method for recognizing and describing hand gestures. Nowadays, deep learning technology is becoming more efficient with real-time applications.
3. Related Work
This section explores an influx of related publications on sign language recognition techniques. According to [20,21,22,23], the implementation of a sign language recognition system can be carried out either by using a sensor-based approach, an image-based approach, or both approaches (hybrid), as can be seen in Figure 4.
In the sensor-based system technique, the user wears a specialized glove equipped with multiple sensors and wires. These sensors assist the system in tracking and recording the movements of hands and fingers. The information transmitted to a computer includes data on finger bending, movements, orientation, rotation, and hand position for interpretation. This interaction between the smart glove and computer is a clear example of human–computer interaction. There are two types of sensor-based methods used in sign language acquisition: sensors that can only detect finger bending and sensors that detect hand motion and orientation [23,24]. For more detailed information, we suggest reading the comprehensive survey paper “Systems-based sensory gloves for sign language recognition” [23].
In an image-based approach, there is no need to wear a glove overloaded with wires, sensors, and other materials. The idea behind image-based systems is to use image processing techniques and algorithms to perceive sign gestures [25,26,27]. Image-based sign language recognition systems can be developed using smart devices since most smart devices have high-resolution cameras that allow natural movements and easy availability. Sensor-based systems are accurate and reliable because they simulate hand gestures. Nevertheless, sensor-based techniques have significant drawbacks, such as the user’s heavy glove size making it uncomfortable to wear [20,23,28]. In addition, the glove has several wires connected to a computer, which limits the user’s mobility and its usage of real-time applications [23].
In [29], the authors developed an Arabic sign language (ArSL) recognition system based on a CNN. A CNN is a sort of artificial neural network (ANN) used in deep learning for image processing, recognition, and classification. The system’s implementation recognizes and translates hand gestures into text to bridge the communication gap between deaf and non-deaf people. They used a dataset consisting of 40 Arabic signs, with each sign having 700 different images, which is a principal factor for training systems to have multiple samples per sign. They employed various hand sizes, lighting, skin tones, and backgrounds to increase the system’s dependability. The result showed an accuracy of 97.69% for training data and 99.47% for testing data. The system was successfully implemented in both mobile and desktop applications. In the same context, Ref. [30] introduced an offline ArSL recognition system based on a deep convolutional neural networks model that can automatically recognize letters and numbers from one to ten. They utilized a real dataset composed of 7869 RGB images. The proposed system achieved an accuracy of 90.02% by training 80% of dataset images. The research introduced in [31] aims to translate the hand gestures of two-dimensional images into text using a faster region-based convolutional neural network (R-CNN). Their system mapped the position of the hand gestures and recognized the letters. They used a dataset of more than 15,360 images with divergent backgrounds that were captured using a phone camera. The result shows a recognition rate of 93% for the collected ArSL images dataset. The goal of this proposed study by [32] is to create a system that can translate static sign gestures into words. They utilized a vision-based method to obtain data from a 1080 full-HD web camera of the signer. The camera will capture only the hands to feed into the system. The dataset will be built through continuous capturing. CNN is applied as a recognition method for their system. After training the model and testing it, the system acquired an average of 90.04% accuracy for recognizing the American Sign Language (ASL) alphabet, 93.44% for numbers (from 1 to 10), and 97.52% for static word recognition. In [33], the authors presented a vision-based gesture recognition system that uses complicated backgrounds. They designed a method for adapting to the skin color of different users and lighting conditions. Three types of features were combined: principal component analysis (PCA), linear discriminant analysis (LDA), and support vector machine (SVM) to describe the hand gestures. The dataset used contains 7800 images for the ASL alphabet. The overall accuracy achieved is 94%. The authors in this work [34] utilized a supervised ML technique to recognize hand-gesturing in ArSL using two sensors: Microsoft’s Kinect with a Leap Motion Controller in a real-time manner. The proposed system matched 224 cases of the Arabic alphabet letter signed by four participants, each of whom performed over 56 gestures. The work carried out by [35] presents a visual sign language recognition system that automatically converts solitary Arabic word signs into text. The proposed system has four main stages: hand segmentation, hand tracking, hand feature extraction, and hand classification. The use of the hand segmentation technique is performed to utilize dynamic skin detectors. Then, the segmented skin blobs are used to track and identify the hands. This proposal uses a dataset of 30 isolated words frequently used by hearing-impaired students daily in school. The result shows that the system has a recognition rate of 97%. In [36], the authors created a dataset and a CNN sign language recognition system to interpret the American sign gesture alphabet and translate it to our natural language. Three datasets were used to compare the results and accuracy of each. The first dataset, which belongs to the authors, has 104,000 images for 26 letters of ASL; the second dataset of ASL has 52,000 images; and the third dataset contains 62,400 images. The datasets were split into 70% for the training sets and 15% each for the validation and testing sets. The overall accuracy for all three datasets based on the CNN model is 99% with a slight difference in the decimal values. For another proposed sign language recognition system, ref. [28] trained a CNN deep learning model to recognize 87,000 ASL images and translate them into text. They were able to achieve an accuracy of 78.50%. For another ASL classification task, ref. [37] developed EfficientNet model to recognize ASL alphabet hand gestures. Their dataset size was 5400 images. They achieved an accuracy of 94.30%. In the same context, [38] used the same dataset of 87,000 images for classification. They used (AlexNet and GooLeNet) models, and their overall training results were 99.39% for AlexNet and 95.52% for GoogLeNet. In [39], the authors evaluated ASL alphabet recognition utilizing two different neural network architectures, AlexNet and ResNet-50, using the same dataset that we used. The results showed that AlexNet achieved an accuracy of 94.74%, while ResNet-50 outperformed it significantly with an accuracy of 98.88%.
Table 1 demonstrates list of proposed systems, methods, number of images used in each dataset, and the accuracy rate.
Table 1.
Reference | Proposed System | Method | Datasets (Number of Images) |
Accuracy |
---|---|---|---|---|
[29] | Smart recognition system for Saudi sign language translate hand gestures into text |
CNN | 27,301 | 97.69% |
[30] | Recognition System of Arab sign numbers and letters |
SVM | 7869 | 90.02% |
[31] | Sign language recognition system for Arabic alphabets translate hand gestures into text |
ResNet-50 | 15,360 | 93.40% |
[33] | American sign language recognition system for alphabets |
SVM PCA LDA |
7800 | 94% |
[36] | American sign language recognition system alphabet signs |
CNN | 104,000 | 99.38% |
[37] | American sign language recognition system alphabet signs |
EfficientNet | 87,000 | 94.30% |
[38] | American sign language recognition system alphabet signs |
AlexNet GoogLeNet |
87,000 | 99.39% 95.52% |
[39] | American sign language alphabet recognition |
AlexNet ResNet-50 |
87,000 | 94.74% 98.88% |
Proposed work | American sign language alphabet recognition |
ResNet-50 EfficientNet ConvNeXt AlexNet VisionTransformer |
87,000 | 99.98% 99.95% 99.51% 99.50% 88.59% |
4. Methodology
This section explores the five distinct deep learning models for identifying American alphabet gestures. Moreover, it provides comprehensive details on image-based techniques and a brief overview of the dataset employed in ASL recognition systems.
4.1. Image-Based Method
Image-based approaches for ASL recognition systems can be categorized into three types: image-based American alphabet signs recognition, American isolated signs recognition, and continuous American signs recognition [28,40].
4.1.1. Image-Based American Alphabet Signs Recognition
In the alphabet signs recognition system, each letter of the American alphabet is signed separately by the signer using one hand. The deaf community uses alphabet letters to spell people’s names, places, and other words. The semantic meaning of the hand gesture of the alphabet letters comes from its shape, as is the case with the representation of the letters “C”, “D”, “L”, “M”, “N”, “O”, “V”, “W”, and “Z”. In addition, there are some explanations for how other letters are represented that require further investigation. In representing the letters J and Z, the signer must use motion to mimic the shape of each letter. Every image-based recognition system is impacted by visual descriptors, which play a vital role in image processing [41,42]. In fact, there are some similarities between ASL alphabet signs, such as A, E, M, N, and S, which makes it difficult to find a simple deep learning model that can distinguish between hand gestures for classification. The goal of this study was to find a visual descriptor that makes it possible to distinguish between various ASL alphabet gestures. See Figure 5.
4.1.2. Image-Based American Isolated Word Signs Recognition
Isolated word recognition frequently requires a sequence of input images of the entire sign, which is in contrast to alphabet sign recognition. This system works only with letters or words but not complete sentences [28,40]. Image-based isolated words can only handle one word at a time [43].
4.1.3. Image-Based Continuous American Sign Language Recognition
Continuous sign recognition is considerably more complicated compared to the two previous techniques. The primary challenges with this approach are dealing with hand tracking, motion detection, feature extraction, and vocabulary size in a real-time manner [28,40,44]. Many studies have concentrated on developing the most effective features and classification techniques for recognizing and distinguishing between a particular sign from a set of possible ones to accomplish a high accuracy rate.
4.2. American Sign Language Dataset
Machine and deep learning, in general, heavily depend on data, as they use algorithms to analyze data and make intelligent predictions. In fact, the availability of sign language databases is limited, which is one of the most significant issues facing sign language recognition and translation systems. Finding a dataset that has manual and non-manual gestures at the same time is challenging [24,40]. Researchers in this field must create a reasonably sized database from scratch to implement and examine their sign language recognition system. Creating a fingerspelling dataset is easy, and it can be performed using non-expert signers to assist in capturing and collecting sign images for the American alphabet with the use of a typical camera. Most letters are depicted in a static posture, and the lexicon is limited to 26 letters. In our proposed system, we only use manual gestures, which represent the American alphabet in sign language. In the finger spelling database, images only display the signer’s hands without any motion; hence, the dataset found in “IEEE Dataport” is suitable for our system. The IEEE Dataport dataset comprises 87,000 images, categorized into 29 distinct classes. Each class encompasses 3000 images, with 26 classes corresponding to the 26 American sign language alphabets, and other classes allocated for space, deletion, and nothing. Dataset images are in RBG format with 200 × 3200 pixels dimensions and different variations [38].
The AlexNet, ConvNeXt, EfficientNet, and ResNet-50 models were trained using 200 × 200 × 3 pixels dimensions, and for VisionTransformer, we resized the dataset to 224 × 224 × 3 pixels. The workflow of splitting the dataset is shown in Figure 6. Then, 80% of the dataset was split into training and validation and split with 5-fold cross-validation approach to train the models. The remaining 20% of the dataset is used for the testing set.
4.3. American Alphabet Sign Language Recognition System
The most ordinary form of communication relies on alphabetic expression through speech, writing, or sign language. There is a constant need for a sign language recognition system, as it could reduce the communication gap between those with and without hearing impairments. In this proposed system, we utilized five different deep learning models to produce more effective classification results.
4.3.1. Transfer Learning
In deep learning, the model requires a large amount of data in the training phase to gain more knowledge and skills. However, deep transfer learning is the process of training on a new problem using deep learning models that have already been pre-trained. The core function of transfer learning is to find shared information that can be transferred between various domains. Moreover, it designs suitable algorithms to transfer common knowledge [45]. Transfer learning comprises instance-based transfer, feature-based transfer, and shared parameter-based transfer. See [46] for more detailed information about each approach. In our proposed classification system, we used pre-trained models due to the large amount of data that requires a high amount of computational power for training. Using pre-trained models will accelerate the learning process and save some time [45]. The schematic architecture of a typical sign language recognition system is split into four separate phases:
Images or video (input data) acquisition;
Images or video preprocessing;
Features extraction;
Classification and recognition of alphabet letters.
Using a pre-trained model allows us to exclude some of these required phases. Fine-tuning is applied to the following deep learning models to transfer knowledge to our new tasks. The architecture of the transfer learning model is shown in Figure 7.
A convolutional neural network can be scaled into three key factors when performing image classification:
The depth indicates the number of layers in the network. Although increasing the depth can help the network learn more intricate features and representations, it can also increase the risk of overfitting and high computational cost.
The width of a network indicates the number of neurons in each layer. By increasing the width, the representation ability can be improved to recognize fine-grained features.
The resolution of the input image. By increasing the resolution, the network will be able to capture finer patterns. The drawback of increasing the resolution is that it requires huge memory usage.
4.3.2. AlexNet
The AlexNet model was designed by Alex Krizhevsky, Ilya Sutskever, and Geoffery Hinton. They trained their model on the ImageNet dataset, which contains more than 15 million high-resolution labeled images and 22,000 classes. Their model shows an incredible ability to accurately and efficiently classify more than 1.2 million images. AlexNet uses computing technology called a graphics processing unit to improve image classification performance. The neural network of AlexNet consists of eight layers, including five convolutional layers and three fully connected layers. The number of parameters is 60 million and 650,000 neurons. See Figure 8 for the AlexNet architecture.
Overfitting always happens with larger datasets to avoid or reduce overfitting, and they utilized a developed regularization method known as “dropout” along with rectified linear units (ReLUs), overlapping pooling, and data augmentation [47]. These features of the AlexNet model made it the winner of the 2012 ImageNet Large-Scale Visual Recognition Competition (ILSVRC-2012), an annual image classification competition [48]. Overall, AlexNet played a significant role in advancing the field of deep learning and demonstrated the power of CNNs for image classification tasks [48]. By using the AlexNet deep learning model to recognize ASL alphabet gestures, we were able to train the model and obtain a high accuracy of 99.50% for a dataset of size 200 × 200 × 3. See Figure 9 for the confusion matrix of the result.
4.3.3. Convnext
ConvNeXt is a type of deep neural network architecture created to achieve a state-of-the-art CNN performance on several tasks, including image classification, object detection, and semantic segmentation. ConvNeXt works by processing an input image through a series of convolutional layers and pooling layers, followed by several fully connected layers to classify and recognize, in our case, hand gestures [49]. ConvNeXt helps in recognizing more diverse and complementary features, which leads to better accuracy across a range of image classification tasks. For the ConvNeXt pre-trained model, the accuracy of recognizing the ASL alphabet is 99.51%. The size of the dataset is 200 × 200 × 3 and there is no substantial difference in using AlexNet and ConvNext in terms of accuracy. Figure 10 illustrates the confusion matrix of ConvNeXt.
4.3.4. EfficientNet
The EfficientNet model is a series of deep neural networks created to use fewer parameters by combining convolutions, bottleneck blocks, depthwise separable, and squeeze-and-excitation modules. EfficientNet uses a set of fixed scaling coefficients to equally scale all dimensions of depth/width/resolution. The rationale behind the compound scaling approach is that larger input images require more layers to expand the network’s receptive area and more channels to identify more fine-grained patterns on the larger image [3]. The models are available in varied sizes and are labeled as EfficieNet-B0, EfficieNet-B1, EfficieNet-B2, etc. The architecture of EfficientNet is shown in Figure 11.
Compared with other CNN architecture models, EfficientNet uses less computation time, which leads to less computational cost. An impressive result of 99.95% was obtained using the EfficientNet model to identify hand gestures for the ASL alphabet, and the size of the dataset used is 200 × 200 × 3. Figure 12 illustrates the confusion matrix of the EfficientNet.
4.3.5. Resnet
ResNet is a deep learning model which stands for residual network. It is a convolutional neural network (CNN) that was first introduced in the ImageNet Large-Scale Visual Recognition Challenge 2015 (ILSVRC2015) to address the problem of vanishing gradients. ResNet won first place for the LSVRC2015 image classification challenge. The ResNet architecture was designed to support thousands of convolutional layers to avoid vanishing gradients. This is unlike other CNN architectures which are only capable of supporting a few layers, negatively impacting the performance [50]. ResNet’s fundamental concept is to employ “skip connections” to create shortcuts between network layers. These skip connections allow the gradient to flow directly from one layer to another without passing through any non-linear activation functions. ResNet performs with the highest accuracy in recognizing the ASL alphabet 99.98% out of the five deep learning models we examined for image classification tasks. The architecture of the model and the confusion matrix results are shown in Figure 13 and Figure 14, respectively.
4.3.6. VisionTransformer
VisionTransformer (ViT) adopts a different strategy from the conventional convolutional neural network (CNN) architecture in image classification. The ViT architecture is composed of two primary components: a patch embedding layer and a Transformer-based encode layer. The patch embedding layer is responsible for converting the input image into a sequence of flattened patches, which are then processed by the encoder. The patches are typically non-overlapping and have fixed sizes of 16 × 16 or 32 × 32 pixels. The encoder is composed of a series of self-attention layers and feed-forward neural networks (FFNs). The self-attention layers allow the model to attend to distinct parts of the input sequence, allowing it to capture long-range dependencies between patches [45]. The FFNs are used to apply non-linear transformations to the output of the self-attention layers. ViT also includes several additional components, such as layer normalization and dropout, to improve the model’s performance and prevent overfitting. ViT is designed to process images in a more flexible and adaptive way than traditional convolutional neural networks by using self-attention [45]. In comparison to CNNs, ViT has a high computational cost, which makes it less useful for some real-time applications see Figure 15 for Vit architecture.
For our task of classifying the ASL alphabet gestures, ViT achieved an accuracy of 88.59%, which is the lowest compared to other deep learning models. This may be related to resizing the dataset to 224 × 224 × 3 to match the model input size requirements. The confusion matrix results are shown in Figure 16.
5. Results and Discussion
In the assessment of our scheme, we employed evaluation criteria including accuracy, precision, recall, and F1-score. These metrics are delineated below, based on the factors of true positives (TPs), false positives (FPs), true negatives (TNs), and false negatives (FNs).
(2) |
(3) |
(4) |
(5) |
Learning rate is an optimization parameter employed to enhance the performance of a deep learning model. This approach involves adjusting the coefficient, which is responsible for updating the network parameters in response to the error generated during the learning process. In cases where the learning rate is too low, the network parameters are gradually updated, leading to a slower learning process. Conversely, a high learning rate can cause the network to miss the optimal point that minimizes the error. Therefore, optimizing the learning rate is crucial for achieving an optimal performance in the transfer learning models [51,52]. In our image classification task, we employed five different deep learning models that were based on transfer learning and utilized the Adam optimizer. The objective was to train these models to accurately predict American sign language letters and overcome the challenges of distinguishing between similar hand gesture letters. All models were run with the following hyperparameters: a learning rate of 0.001, eight batches, two epoches, Adamax optimizer, and a stochastic gradient descent momentum of 0.9. The five-fold cross-validation was used to measure the performance of the models. Table 2 presents the results obtained from evaluating the performances of the five models. It can be noticed that, in general, ResNet, EfficientNet, AlexNet and Convexnet, respectively, produced the best results in terms of accuracy.
Table 2.
Architecture | L.Rate | Optimizer | Accuracy | Precision | Recall | F1 |
---|---|---|---|---|---|---|
AlexNet | 0.001 | Adamax | 99.50 % | 99.51% | 99.50% | 99.50% |
ConvNext | 0.001 | Adamax | 99.51 % | 99.51% | 99.50% | 99.51% |
EfficientNet | 0.001 | Adamax | 99.95 % | 99.90% | 99.94% | 99.92% |
ResNet | 0.001 | Adamax | 99.98 % | 99.95% | 99.60% | 99.98% |
VisionTransformer | 0.001 | Adamax | 88.59 % | 89.55% | 88.59% | 88.54% |
However, the vision transform model exhibited poorer accuracy in comparison to the other models. This can be attributed to the resizing requirement imposed on the input data to conform to the fixed input size of the original model. In order to meet this requirement, the dataset was resized to dimensions of 224 × 3224 × 33. In the resizing process, the potential of losing valuable information exists and that may affect the performance. On the other hand, the other models that attained higher accuracy were trained on the original input size of 200 × 200 × 3.
Applying data augmentation techniques to a vision transformer model, such as random crops, flips, rotations, or adjustments in brightness and contrast, can introduce additional variations to the training data and potentially improve the model’s performance. However, in our specific case, we chose to compare the performance of five models without utilizing data augmentation as a technique across all models. However, to ensure a fair and unbiased comparison between the five models in our study, we refrained from utilizing data augmentation techniques in any of the models. This approach allows us to assess the inherent capabilities and performance differences among the models based on their architectural design and training process without the influence of additional data variations.
Table 3 shows a comparison between the proposed scheme and recent studies that utilized the same dataset [31,37,38,39]. In terms of evaluation metrics for this dataset, our models demonstrate a promising performance. Consequently, the experimental results affirm that the proposed scheme effectively classifies and recognizes hand gestures, even when there are subtle disparities in shape.
Table 3.
Citation | Year | Models | Accuracy | Precision | Recall | F1 |
---|---|---|---|---|---|---|
[38] | 2019 | AlexNet | 99.39% | Not reported | Not reported | Not reported |
GoogLeNet | 95.52 | |||||
[31] | 2021 | ResNet-18 | 93.40% | 93.30% | 94.30% | 93.70% |
VGG-16 | 93.20% | 93.60% | 93.50% | 93.50% | ||
[37] | 2022 | EfficientNet | 94.30% | 94.30% | 94.46% | 94.13% |
[39] | 2022 | AlexNet | 93.64% | 88.46% | 87.88% | 87.92% |
ResNet-50 | 97.41% | 94.01% | 93.56% | 93.88% | ||
Our work | 2023 | ResNet-50 | 99.98% | 99.95% | 99.60% | 99.98% |
EfficientNet | 99.95% | 99.90% | 99.94% | 99.92% | ||
AlexNet | 99.50% | 99.51% | 99.50% | 99.50% | ||
ConvNext | 99.51% | 99.51% | 99.50% | 99.51% |
Several limitations are encountered when conducting research into deep learning for American sign language (ASL) alphabet classification tasks:
Dataset availability: The number of datasets containing images for ASL hand gestures is limited due to the need for experts to collect and label data.
Data diversity and size: Limited availability of diverse and sizable ASL datasets with various backgrounds can hinder the training and generalization of deep learning models, which represented major obstacles when implementing ASL classification task using deep learning models.
6. Conclusions
In this paper, we presented a study employing transfer learning techniques utilizing five deep learning models for the effective classification of hand gestures representing the American sign language (ASL) alphabet. The obtained results exhibited remarkable achievements, with the ResNet-50 model outperforming other studies in image classification tasks for ASL recognition systems, achieving an outstanding accuracy rate of 99.988%. Notably, the EfficientNet model demonstrated the second-highest accuracy rate, surpassing 99.95%. Similarly, AlexNet and ConvNext models exhibited commendable accuracy levels of 99.51%. Conversely, the VisionTransformer model exhibited a comparatively lower accuracy rate of 88.59%. This reduced accuracy could potentially be attributed to the preprocessing step involving the resizing of the dataset. In our future work, we plan to convert our trained image classification models for American sign language alphabet gestures into a real-time system and evaluate their performance. We aim to gain insights into their suitability for real-world applications. This will allow us to identify any potential challenges or areas for improvement.
Author Contributions
Conceptualization, B.A. and M.I.; methodology, B.A.; software, B.A., A.S.A., A.A. and E.A.; validation, B.A. and M.I.; formal analysis, B.A. and M.I.; investigation, B.A. and M.I.; writing—original draft preparation, B.A.; writing—review and editing, M.I.; visualization, B.A., A.S.A., A.A. and E.A.; supervision, M.I. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data available upon request.
Conflicts of Interest
The authors declare that they have no conflict of interest.
Funding Statement
This research received no external funding.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Zwitserlood I., Verlinden M., Ros J., Van Der Schoot S., Netherlands T. Synthetic signing for the deaf: Esign; Proceedings of the Conference and Workshop on Assistive Technologies for Vision and Hearing Impairment, CVHI; Granada, Spain. 29 June–2 July 2004. [Google Scholar]
- 2.World Health Organisation Deafness and Hearing Loss. 2021. [(accessed on 8 November 2022)]. Available online: http://https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss.
- 3.Alsaadi Z., Alshamani E., Alrehaili M., Alrashdi A.A.D., Albelwi S., Elfaki A.O. A real time Arabic sign language alphabets (ArSLA) recognition model using deep learning architecture. Computers. 2022;11:78. doi: 10.3390/computers11050078. [DOI] [Google Scholar]
- 4.Alsharif B., Ilyas M. Internet of Things Technologies in Healthcare for People with Hearing Impairments; Proceedings of the IoT and Big Data Technologies for Health Care: Third EAI International Conference, IoTCare 2022; Virtual Event. 12–13 December 2022; Berlin/Heidelberg, Germany: Springer; 2023. pp. 299–308. Proceedings. [Google Scholar]
- 5.Farooq U., Rahim M.S.M., Sabir N., Hussain A., Abid A. Advances in machine translation for sign language: Approaches, limitations, and challenges. Neural Comput. Appl. 2021;33:14357–14399. doi: 10.1007/s00521-021-06079-3. [DOI] [Google Scholar]
- 6.Latif G., Mohammad N., AlKhalaf R., AlKhalaf R., Alghazo J., Khan M. An automatic Arabic sign language recognition system based on deep CNN: An assistive system for the deaf and hard of hearing. Int. J. Comput. Digit. Syst. 2020;9:715–724. doi: 10.12785/ijcds/090418. [DOI] [Google Scholar]
- 7.Bell J. Machine Learning and the City: Applications in Architecture and Urban Design. John Wiley & Sons Ltd.; Hoboken, NJ, USA: 2022. What is machine learning? pp. 207–216. [Google Scholar]
- 8.Najafabadi M.M., Villanustre F., Khoshgoftaar T.M., Seliya N., Wald R., Muharemagic E. Deep learning applications and challenges in big data analytics. J. Big Data. 2015;2:1–21. doi: 10.1186/s40537-014-0007-7. [DOI] [Google Scholar]
- 9.Fouladi S., Safaei A.A., Arshad N.I., Ebadi M., Ahmadian A. The use of artificial neural networks to diagnose Alzheimer’s disease from brain images. Multimed. Tools Appl. 2022;81:37681–37721. doi: 10.1007/s11042-022-13506-7. [DOI] [Google Scholar]
- 10.Yousefpanah K., Ebadi M.J. Review of artificial intelligence-assisted COVID-19 detection solutions using radiological images. J. Electron. Imaging. 2022;32:021405. doi: 10.1117/1.JEI.32.2.021405. [DOI] [Google Scholar]
- 11.Saravanan R., Sujatha P. A state of art techniques on machine learning algorithms: A perspective of supervised learning approaches in data classification; Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS); Madurai, India. 14–15 June 2018; Piscataway, NJ, USA: IEEE; 2018. pp. 945–949. [Google Scholar]
- 12.Mahesh B. Machine learning algorithms—A review. Int. J. Sci. Res. (IJSR) 2020;9:381–386. [Google Scholar]
- 13.Alkanjr B., Mahgoub I. A Novel Deception-Based Scheme to Secure the Location Information for IoBT Entities. IEEE Access. 2023;11:15540–15554. doi: 10.1109/ACCESS.2023.3244138. [DOI] [Google Scholar]
- 14.Alkanjr B., Mahgoub I. Location Privacy-Preserving Scheme in IoBT Networks Using Deception-Based Techniques. Sensors. 2023;23:3142. doi: 10.3390/s23063142. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Padakandla S. A survey of reinforcement learning algorithms for dynamically varying environments. ACM Comput. Surv. (CSUR) 2021;54:1–25. doi: 10.1145/3459991. [DOI] [Google Scholar]
- 16.Alsayegh M., Dutta A., Vanegas P., Bobadilla L. Lightweight multi-robot communication protocols for information synchronization; Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Las Vegas, NV, USA. 25–29 October 2020; Piscataway, NJ, USA: IEEE; 2020. pp. 11831–11837. [Google Scholar]
- 17.Newaz A.A.R., Alsayegh M., Alam T., Bobadilla L. Decentralized Multi-Robot Information Gathering From Unknown Spatial Fields. IEEE Robot. Autom. Lett. 2023;8:3070–3077. doi: 10.1109/LRA.2023.3264720. [DOI] [Google Scholar]
- 18.Shinde P.P., Shah S. A review of machine learning and deep learning applications; Proceedings of the 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA); Pune, India. 16–18 August 2018; Piscataway, NJ, USA: IEEE; 2018. pp. 1–6. [Google Scholar]
- 19.Altaher A., Salekshahrezaee Z., Abdollah Zadeh A., Rafieipour H., Altaher A. Using multi-inception CNN for face emotion recognition. J. Bioeng. Res. 2020;3:1–12. [Google Scholar]
- 20.Mummadi C.K., Philips Peter Leo F., Deep Verma K., Kasireddy S., Scholl P.M., Kempfle J., Van Laerhoven K. Real-time and embedded detection of hand gestures with an IMU-based glove. Informatics. 2018;5:28. doi: 10.3390/informatics5020028. [DOI] [Google Scholar]
- 21.Elons A., Ahmed M., Shedid H., Tolba M. Arabic sign language recognition using leap motion sensor; Proceedings of the 2014 9th International Conference on Computer Engineering & Systems (ICCES); Cairo, Egypt. 21–23 December 2014; Piscataway, NJ, USA: IEEE; 2014. pp. 368–373. [Google Scholar]
- 22.Kammoun S., Darwish D., Althubeany H., Alfull R. Proceedings of the Universal Access in Human–Computer Interaction. Applications and Practice: 14th International Conference, UAHCI 2020, Held as Part of the 22nd HCI International Conference, HCII 2020, Copenhagen, Denmark, 19–24 July 2020. Springer; Berlin/Heidelberg, Germany: 2020. ArSign: Toward a mobile based Arabic sign language translator using LMC; pp. 92–101. Proceedings, Part II 22. [Google Scholar]
- 23.Ahmed M.A., Zaidan B.B., Zaidan A.A., Salih M.M., Lakulu M.M.B. A review on systems-based sensory gloves for sign language recognition state of the art between 2007 and 2017. Sensors. 2018;18:2208. doi: 10.3390/s18072208. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.El-Alfy E.S.M., Luqman H. A comprehensive survey and taxonomy of sign language research. Eng. Appl. Artif. Intell. 2022;114:105198. [Google Scholar]
- 25.Alzohairi R., Alghonaim R., Alshehri W., Aloqeely S. Image based Arabic sign language recognition system. Int. J. Adv. Comput. Sci. Appl. 2018;9:3. doi: 10.14569/IJACSA.2018.090327. [DOI] [Google Scholar]
- 26.Al Ani L.A., Al Tahir H.S. Classification Performance of TM Satellite Images. Al-Nahrain J. Sci. 2020;23:62–68. doi: 10.22401/ANJS.23.1.09. [DOI] [Google Scholar]
- 27.Nazari K., Ebadi M.J., Berahmand K. Diagnosis of alternaria disease and leafminer pest on tomato leaves using image processing techniques. J. Sci. Food Agric. 2022;102:6907–6920. doi: 10.1002/jsfa.12052. [DOI] [PubMed] [Google Scholar]
- 28.Mohandes M., Liu J., Deriche M. A survey of image-based arabic sign language recognition; Proceedings of the 2014 IEEE 11th International Multi-Conference on Systems, Signals & Devices (SSD14); Barcelona, Spain. 11–14 February 2014; pp. 1–4. [Google Scholar]
- 29.Al-Obodi A.H., Al-Hanine A.M., Al-Harbi K.N., Al-Dawas M.S., Al-Shargabi A.A. A Saudi Sign Language recognition system based on convolutional neural networks. Build. Serv. Eng. Res. Technol. 2020;13:3328–3334. doi: 10.37624/IJERT/13.11.2020.3328-3334. [DOI] [Google Scholar]
- 30.Hayani S., Benaddy M., El Meslouhi O., Kardouchi M. Arab sign language recognition with convolutional neural networks; Proceedings of the 2019 International Conference of Computer Science and Renewable Energies (ICCSRE); Agadir, Morocco. 22–24 July 2019; Piscataway, NJ, USA: IEEE; 2019. pp. 1–4. [Google Scholar]
- 31.Alawwad R.A., Bchir O., Ismail M.M.B. Arabic sign language recognition using Faster R-CNN. Int. J. Adv. Comput. Sci. Appl. 2021;12:3. doi: 10.14569/IJACSA.2021.0120380. [DOI] [Google Scholar]
- 32.Vanaja S., Preetha R., Sudha S. Hand Gesture Recognition for Deaf and Dumb Using CNN Technique; Proceedings of the 2021 6th International Conference on Communication and Electronics Systems (ICCES); Coimbatore, India. 8–10 July 2021; Piscataway, NJ, USA: IEEE; 2021. pp. 1–4. [Google Scholar]
- 33.Pan T.Y., Lo L.Y., Yeh C.W., Li J.W., Liu H.T., Hu M.C. Real-time sign language recognition in complex background scene based on a hierarchical clustering classification method; Proceedings of the 2016 IEEE Second International Conference on Multimedia Big Data (BigMM); Taipei, Taiwan. 20–22 April 2016; Piscataway, NJ, USA: IEEE; 2016. pp. 64–67. [Google Scholar]
- 34.Almasre M.A., Al-Nuaim H. A real-time letter recognition model for Arabic sign language using kinect and leap motion controller v2. Int. J. Adv. Eng. Manag. Sci. 2016;2:239469. [Google Scholar]
- 35.Ibrahim N.B., Selim M.M., Zayed H.H. An automatic Arabic sign language recognition system (ArSLRS) J. King Saud Univ.-Comput. Inf. Sci. 2018;30:470–477. [Google Scholar]
- 36.Kasapbaşi A., ELBUSHRA A.E.A., Omar A.H., Yilmaz A. DeepASLR: A CNN based human computer interface for American Sign Language recognition for hearing-impaired individuals. Comput. Methods Programs Biomed. Update. 2022;2:100048 [Google Scholar]
- 37.AlKhuraym B.Y., Ismail M.M.B., Bchir O. Arabic sign language recognition using lightweight cnn-based architecture. Int. J. Adv. Comput. Sci. Appl. 2022;13:4. doi: 10.14569/IJACSA.2022.0130438. [DOI] [Google Scholar]
- 38.Cayamcela M.E.M., Lim W. Fine-tuning a pre-trained convolutional neural network model to translate American sign language in real-time; Proceedings of the 2019 International Conference on Computing, Networking and Communications (ICNC); Honolulu, HI, USA. 18–21 February 2019; Piscataway, NJ, USA: IEEE; 2019. pp. 100–104. [Google Scholar]
- 39.Ma Y., Xu T., Kim K. Two-Stream Mixed Convolutional Neural Network for American Sign Language Recognition. Sensors. 2022;22:5959. doi: 10.3390/s22165959. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Hassan M., Assaleh K., Shanableh T. Multiple proposals for continuous arabic sign language recognition. Sens. Imaging. 2019;20:4. [Google Scholar]
- 41.Jan M.T., Moshfeghi S., Conniff J.W., Jang J., Yang K., Zhai J., Rosselli M., Newman D., Tappen R., Furht B. Methods and Tools for Monitoring Driver’s Behavior. arXiv. 2023 doi: 10.1109/csci58124.2022.00228.2301.12269 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Jan M.T., Hashemi A., Jang J., Yang K., Zhai J., Newman D., Tappen R., Furht B. Non-intrusive Drowsiness Detection Techniques and Their Application in Detecting Early Dementia in Older Drivers; Proceedings of the Future Technologies Conference (FTC) 2022; Vancouver, BC, Canada. 20–21 October 2022; Berlin/Heidelberg, Germany: Springer; 2022. pp. 776–796. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Alsaidi M., Altaher A.S., Jan M.T., Altaher A., Salekshahrezaee Z. COVID-19 Classification Using Deep Learning Two-Stage Approach. arXiv. 20222211.15817 [Google Scholar]
- 44.Altaher A.S., Zhuang H., Ibrahim A.K., Ali A.M., Altaher A., Locascio J., McCallister M.P., Ajemian M.J., Chérubin L.M. Detection and localization of Goliath grouper using their low-frequency pulse sounds. J. Acoust. Soc. Am. 2023;153:2190–2202. doi: 10.1121/10.0017804. [DOI] [PubMed] [Google Scholar]
- 45.Bengio Y. Deep learning of representations for unsupervised and transfer learning; Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, JMLR Workshop and Conference Proceedings; Washington, DC, USA. 2 July 2011; pp. 17–36. [Google Scholar]
- 46.Jiang X., Hu B., Chandra Satapathy S., Wang S.H., Zhang Y.D. Fingerspelling identification for Chinese sign language via AlexNet-based transfer learning and Adam optimizer. Sci. Program. 2020;2020:3291426. [Google Scholar]
- 47.Alanazi M., Aldahr R.S., Ilyas M. Effectiveness of Machine Learning on Human Activity Recognition Using Accelerometer and Gyroscope Sensors: A Survey; Proceedings of the 26th World Multi-Conference on Systemics, Cybernetics and Informatics (WMSCI 2022); Online. 12–15 July 2022. [Google Scholar]
- 48.He K., Zhang X., Ren S., Sun J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015;37:1904–1916. doi: 10.1109/TPAMI.2015.2389824. [DOI] [PubMed] [Google Scholar]
- 49.Kim D., Wang K., Sclaroff S., Saenko K. A broad study of pre-training for domain generalization and adaptation; Proceedings of the Computer Vision–ECCV 2022: 17th European Conference; Tel Aviv, Israel. 23–27 October 2022; Berlin/Heidelberg, Germany: Springer; 2022. pp. 621–638. Proceedings, Part XXXIII. [Google Scholar]
- 50.He K., Zhang X., Ren S., Sun J. Deep residual learning for image recognition; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Las Vegas, NV, USA. 27–30 June 2016; pp. 770–778. [Google Scholar]
- 51.Khalid M., Wu J., M. Ali T., Ameen T., Moustafa A.A., Zhu Q., Xiong R. Cortico-hippocampal computational modeling using quantum neural networks to simulate classical conditioning paradigms. Brain Sci. 2020;10:431. doi: 10.3390/brainsci10070431. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Alsayegh M., Vanegas P., Newaz A.A.R., Bobadilla L., Shell D.A. Privacy-Preserving Multi-Robot Task Allocation via Secure Multi-Party Computation; Proceedings of the 2022 European Control Conference (ECC); London, UK. 11–14 July 2022; pp. 1274–1281. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data available upon request.