Skip to main content
Journal of Genetic Engineering & Biotechnology logoLink to Journal of Genetic Engineering & Biotechnology
. 2024 Feb 21;22(1):100359. doi: 10.1016/j.jgeb.2024.100359

Human exons and introns classification using pre-trained Resnet-50 and GoogleNet models and 13-layers CNN model

Feriel Ben Nasr Barber 1,2,, Afef Elloumi Oueslati 1,2,1
PMCID: PMC10903757  PMID: 38494268

Abstract

Background:

Examining functions and characteristics of DNA sequences is a highly challenging task. When it comes to the human genome, which is made up of exons and introns, this task is more challenging. Human exons and introns contain millions to billions of nucleotides, which contributes to the complexity observed in this sequences. Considering how complicated the subject of genomics is, it is obvious that using signal processing techniques and deep learning tools to build a strong predictive model can be very helpful for the development of the research of the human genome.

Results:

After representing human exons and introns with color images using Frequency Chaos Game Representation, two pre-trained convolutional neural network models (Resnet-50 and GoogleNet) and a proposed CNN model having 13 hidden layers were used to classify our obtained images. We have reached a value of 92% for the accuracy rate for Resnet-50 model in about 7 h for the execution time, a value of 91.5% for the accuracy rate for the GoogleNet model in 2 h and a half for the execution time. For our proposed CNN model, we have reached 91.6% for the accuracy rate in 2 h and 37 min.

Conclusions:

Our proposed CNN model is faster than the Resnet-50 model in terms of execution time. It was able to slightly exceed the GoogleNet model for the accuracy rate value.

Keywords: Exons, Introns, Restnet-50, GoogleNet, CNN, Classification

1. Background

The study of all living organisms’ DNA sequences is known as genomics, which is a branch of biology.1 A cell’s genome serves as a representation of every genetic piece the cell contains. Genes linked to diseases can be found thanks to genomics, which also helps us better understand the variety of living things.2 It examines the organization and structure of genomes (or genes), as well as how they operate.

The human cell’s genome is made up of the mitochondrial DNA and the 23 pairs of chromosomes that constitute the nucleus.3 The four nucleotides (or letters) that represent human DNA are A for adenine, C for cytosine, G for guanine, and T for thymine.4 A segment of this DNA called a gene is a unit of genetic information that codes for a single protein. Coding sections (also known as exons), non-coding regions (also known as introns), intergenic regions, repetitive sequences, and others compose human genes.5

Coding sequences or exons are DNA sequences that are translated into mRNA and subsequently transformed by the proper mRNA molecules into a polypeptide chain. A codon, often known as a set of three nucleotides, in a coding sequence codes for one amino acid in a polypeptide chain. Only about 2% of DNA is composed of coding genes; the other 98 percent is made up of non-coding regions.6

According to research published within the context of the ENCODE project,7 which aims to identify the activities of every component of our genome, non-coding DNA or introns, which was wrongly referred to in the past as junk DNA, is important in the regulation of gene activity and the emergence of disorders.8

Genome mapping, finding new genes, and examining how genes are expressed and controlled are just a few of the analyses that form genomics.9 Computers are essential in these analyses because specialized software makes it easier to classify genes based on how similar their sequences are.10 Given the vast volume of data, genomics stores and analyzes data using bioinformatics and deep learning methods.11 It is in this context that our work is situated.

In this work, we will represent human exons and introns with color images. Then, we will use two pre-trained convolutional neural network (CNN) models (resnet50 and googleNet) and we will propose a simplified CNN model for the classification of images representing coding and non-coding regions of human genes. In Section 2, we will introduce deep learning and convolutional neural network models and describe the architecture of the used pre-trained CNN models. Then, in Section 3, we will describe our methodology on representing coding and non-coding regions of human genes with RGB images and classifying them using 3 CNN models. In Section 4, we will apply ResNet50 model, googleNet model and our proposed CNN model to our constructed images and expose obtained classification results. Finally, in Section 5, we will conclude our work.

2. Deep learning and convolutional neural networks

Deep learning algorithms build a model based on the features they extract from big datasets, such as a collection of photos or genomes. Algorithms can use the model once it has been created to analyze different data sets.12

Deep learning is applied in a variety of fields, including bioinformatics,13 genomics,14 robotics,15 language processing,16 image recognition,17 voice recognition,18 and technology for diagnostic support in the medical field.19.

2.1. Convolutional neural network (CNN)

Convolutional neural networks (CNNs) are a subset of artificial neural networks used in image processing and recognition. A multilayered perceptron-like system designed for low processing demands is used by CNNs. Several convolutional layers, grouping layers, fully connected layers, and normalization layers make up the input, output, and hidden layers of a CNN. By removing constraints and enhancing image processing efficiency, a system for image processing and natural language processing is created that is significantly specialized, more efficient and easier to train.20

2.1.1. Convolution layer

The first layer of a convolutional neural network is always the convolution layer. The main objective of this layer is to recognize a set of features in incoming images that are represented (Fig.1). Convolution filtering is employed in this process; the goal is to form a window across the image designating the feature and compute the convolution product between the feature and each segment of the scanned image,21 as shown in the following equation.

Yjl=fiIj(Yjl-1*Kijl+bjl) (1)

Where:

  • Ij represents the input image set.

  • Yjl represents the jth feature map of the l layer.

  • The operator ”*” represents the operation of convolution.

  • Yil-1 is the ith feature map of the l-1 layer.

  • Kijl represents the filter connecting the jth feature map of the l layer and the ith feature map of the l-1 layer.

  • bjl is the bias.

  • f[.] represents the nonlinear activation function capable of resolving issues that linear models are unable to.

Fig. 1.

Fig. 1

Convolution layer.

Common activation methods include sigmoid, tanh, ReLU, etc. The corresponding calculation equations are as follows:

Sigmoid:f(x)=11+e-x (2)
Tanh:f(x)=ex-e-xex+e-x (3)
ReLU:f(x)=max(0,x) (4)

Convolution layer volume is sized using three hyper parameters: layer depth, pitch, which regulates receptive field overlap, and zero padding.

2.1.2. Pooling layer

Usually placed between two convolution layers, this type of layer performs the pooling function on each feature map it receives as input.22 In the pooling process, images are reduced in size while retaining their essential elements (Fig.2).

Fig. 2.

Fig. 2

Pooling layer.

The pooling layer can be used to minimize the amount of parameters and calculations in the network. By doing this, overfitting is avoided and the network’s effectiveness is increased.

2.1.3. ReLU layer

A layer that will apply an activation function to the output signals is introduced between the processing layers to speed up processing. Rectified Linear Unit (ReLU), a nonlinear activation layer used in convolution maps, converts the negative integers in the filtered images to zeroes.23

2.1.4. Fully connected layer

A neural network’s top layer is always the fully connected layer, which accepts a vector as input and produces another vector. This is accomplished by applying an activation function and may be a linear combination to the input data (Fig.3).

Fig. 3.

Fig. 3

Fully connected layers.

2.2. Pre-trained Convolutional Neural Networks

A pre-trained model is a model or a saved network developed by another party and trained on a sizable dataset to solve a related problem. Instead of creating a model from scratch, data scientists can employ a pre-trained model as a starting point. There are many pre-trained CNN models such as VGG16, VGG19, Alexnet, EfficientNet, GoogleNet (or inceptionv3) and resnet50.24

2.2.1. Resnet50 architecture

Resnet, short for residual network, is a particular variety of convolutional neural network that He.K. et al. first introduced in 2015. 25 resnet50 is a 50-layer CNN, it builds networks by stacking residual blocks.

The resnet architecture follows two fundamental design principles. First, regardless of the size of the output feature map, there is the same number of filters in each layer. Second, to maintain the time complexity of each layer when the size of the feature map is halved, it has twice as many filters. The building block for resnet employs a bottleneck design, which lowers the amount of parameters and matrix multiplications. This makes training each layer considerably faster. Instead of using two layers, it employs a stack of three layers (Fig.4.

Fig. 4.

Fig. 4

The Resnet-50 architecture.

The pretrained network is capable of classifying photos into 1,000 different object categories, including various animals, a keyboard, a mouse, and a pencil. Hence, the network has picked up detailed feature representations for a variety of images. The network can accept images up to 224 by 224 pixels.

2.2.2. GoogleNet (Inception V1) architecture

Google Research suggested googlenet in 2014 (with the assistance of several institutions).26 This architecture employs methods like global mean pooling and 1×1 convolutions in the middle of the architecture. There are 22 layers in the entire architecture. It can be used on individual devices even with limited processing capabilities because it was built with computational economy in consideration (Fig.5.

Fig. 5.

Fig. 5

The GoogleNet architecture.

This architecture uses an image with RGB color channels that is 224×224 in size. This architecture uses rectified linear units as the activation function for all convolutions. The network was trained on two different image datasets: imageNet database and Places365 database.

3. Methods

3.1. Methodology

Our methodology consists of five steps; first, to construct our images dataset, we convert nucleotide sequences of exons and introns into images using frequency chaos game representation. Second, we classify obtained images using Resnet-50 model, then we classify them using GooleNet model. After that, we propose our simplified CNN model and execute it to classify our images. Finally, we compare the results of the three applied CNN models (Fig.6).

Fig. 6.

Fig. 6

Methodology Steps.

3.2. Dataset Construction

To construct our dataset, we have randomly downloaded 110 biological sequences of human genes, from the National Center for Biotechnology Information (NCBI) database [https://www.ncbi.nlm.nih.gov/]. Knowing the positions of exons in each gene, we have divided human genes to two classes: 1860 exons and 1631 introns and we have chosen only 1600 for each class for the image representation and classification (Table 1), then we have represented those exons and introns by RGB images.

Table 1.

Human exons and introns dataset details.

Genes’ number 110
Exons’ number 1860
Exons’ medium length 500 bp
Introns’ number 1631
Introns’ medium length 2000 bp

We have decided to represent our exons and introns by the Frequency Chaos Game Representation (FCGR) since the lengths of exons and introns are different. This representation allows us to visually group the images of words of the same suffix in dials. We note that by adding one more letter, the points of the images of any word are found in the same quarter of the figure (Fig. 7).

Fig. 7.

Fig. 7

Frequency Chaos Game Rrepresentation by dials.

Our constructed images are respectively the 4th, 5th and 6th orders of Frequency Chaos Game Representation (FCGR) used in our previous work27 shown in Fig. 8.

Fig. 8.

Fig. 8

Constructed images, (a): FCG Representation of the first exon of ACE human gene, (b): FCG Representation of the first intron of ACE human gene.

Images in Fig. 8 represent human gene ACE. (a): represents the 4th, 5th and 6th orders of the first exon. (b): represents the 4th, 5th and 6th orders of the first intron.

3.3. Dataset

Our final dataset contains three sets of images; each set is split into 2 classes: exons and introns.

Set A: this set holds images representing the 4th order of the FCGR. There are 3200 images divided into 1600 images representing exons and 1600 images representing introns. The size of each image is 224×224×3.

Set B: this set includes images representing the 5th order of the FCGR. There are 3200 images divided into 1600 images representing exons and 1600 images representing introns. The size of each image is 224×224×3.

Set C: this set comprises images representing the 6th order of the FCGR. There are 3200 images divided into 1600 images representing exons and 1600 images representing introns. The size of each image is 224×224×3.

4. Results

In this section, we will expose classification results for the Resnet-50 model, the GoogleNet model and our proposed CNN model.

4.1. Classification results for Resnet-50 model

We remind that the Resnet-50 model is a pre-trained convolutional neural network with 50 layers depth and can classify up to 1000 class (object image). To adapt this model to our classification problem, we have frozen the weights of the ten first layers and we have added one fully connected layer at the end, having 2 neurons since we have a 2-class classification problem. For the classification, we have divided each set into 70% for the training phase and 30% for the test phase and we have trained the Resnet-50 model for only 10 epochs.

In Table 2, we illustrate training rate, accuracy rate values and the execution time of the three sets classified by Resnet-50 model:

Table 2.

Classification results of Resnet-50 model.

Training rate Accuracy Execution Time
Set A 90.21% 90.9% 7 h 07 min 24 s
Set B 90.42% 92% 6 h 54 min 13 s
Set C 89.58% 89.4% 6 h 55 min 03 s

When we classified our images with Resnet-50 model, we have obtained: for set A (the 4th FCGR order), a value of 90.21% for the training rate, a value of 90.9% for the accuracy rate and 7 h 07 min 24 s for the execution time. For set B (the 5th FCGR order),a value of 90.42% for the training rate, a value of 92% for the accuracy rate and 6 h 54 min 13 s for the execution time. For set C (the 6th FCGR order),a value of 89.58% for the training rate, a value of 89.4% for the accuracy rate and 6 h 55 min 03 s for the execution time.

When we visualize Fig. 9, we can notice that the Resnet-50 model, when it was applied on set A, it was capable of predicting 87.3% of exons and 94.6% of introns, for set B, 90% of exons were correctly identified and 94% of introns were correctly identified and for set C 84.8% of exons were correctly recognized and 94% of introns were identified correctly.

Fig. 9.

Fig. 9

Confusion matrices for Resnet-50 model. (a): set A, (b): set B, (c): set C.

4.2. Classification results for GoogleNet model

We remind that the GoogleNet model is a pre-trained convolutional neural network with 22 layers depth and can classify up to 1000 class (object image). To adapt this model to our classification problem, we have frozen the weights of the ten first layers and we have added one fully connected layer at the end, having 2 neurons since we have a 2-class classification problem. For the classification, we have divided each set to 70% for the training phase and 30% for the test phase and we have trained the GoogleNet model for only 10 epochs.

In Table 3, we illustrate training rate, accuracy rate values and the execution time of the three sets classified by GoogleNet model:

Table 3.

Classification results of GoogleNet model.

Training rate Accuracy Execution Time
Set A 91.46% 91.5% 2 h 28 min 23 s
Set B 89.06% 89.1% 2 h 23 min 37 s
Set C 89.27% 89.3% 2 h 26 min 59 s

When we classified our images using GoogleNet model, we have reached: for set A (the 4th FCGR order), a value of 91.46% for the training rate, a value of 91.5% for the accuracy rate and 2 h 28 min 23 s for the execution time. For set B (the 5th FCGR order),a value of 89.06% for the training rate, a value of 89.1% for the accuracy rate and 2 h 23 min 37 s for the execution time. For set C (the 6th FCGR order),a value of 89.27% for the training rate, a value of 89.3% for the accuracy rate and 2 h 26 min 59 s for the execution time.

When we observe Fig. 10, we can notice that the GoogleNet model, when it was applied on set A, it was able to predict 86.7% of exons and 96.3% of introns, for set B, 82.7% of exons were correctly identified and 95.4% of introns were correctly identified and for set C 90.4% of exons were correctly recognized and 88.1% of introns were identified correctly.

Fig. 10.

Fig. 10

Confusion matrices for GoogleNet model. (a): set A, (b): set B, (c): set C.

4.3. Proposed model

With the aim of improving classification rates, we have tested many CNN architectures and we ended up with this model. Our proposed CNN model is a 13 layers depth model having 2 convolution layers, 2 MaxPooling layers, 2 batch Normalization layers, 1 relu layer, 4 fully connected layers, 1 softMax layer and 1 classification layer as shown in Table 4:

Table 4.

Our proposed CNN model architecture.

Input Layer
1stConvolution Filters number 32
Layer Filters size 7×7
Padding 7
1stMaxPooling Pool size 7×7
Layer Padding 7
2ndConvolution Filters number 16
Layer Filters size 5×5
Padding 5
2ndMaxPooling Pool size 5×5
Layer Padding 5
Fully 1st layer neurons’ number 512
Connected 2nd layer neurons’ number 256
Layers 3rd layer neurons’ number 218
4th layer neurons’ number 2
Classification Layer
Epochs Number 10
Training Dataset 70%
Test Dataset 30%

We have chosen to do classification without overlap, we have divided our images into 70% for the training phase and 30% for the test phase and we have trained our model during 10 epochs.

4.4. Classification results for our proposed CNN model

In Table 5, we represent training rate, accuracy rate values and the execution time of the three sets classified by our proposed CNN model:

Table 5.

Classification results of our proposed CNN model.

Training rate Accuracy Execution Time
Set A 91.46% 91.6% 2 h 37 min 55 s
Set B 87.81% 89.7% 2 h 35 min 53 s
Set C 89.06% 89.3% 2 h 38 min 19 s

When we used our proposed CNN model to classify our images, we have reached: for set A (the 4th FCGR order), a value of 91.46% for the training rate, a value of 91.6% for the accuracy rate and 2 h 37 min 55 s for the execution time. For set B (the 5th FCGR order),a value of 87.81% for the training rate, a value of 89.7% for the accuracy rate and 2 h 35 min 53 s for the execution time. For set C (the 6th FCGR order),a value of 89.06% for the training rate, a value of 89.3% for the accuracy rate and 2 h 38 min 19 s for the execution time.

When we observe Fig. 11, we can notice that our proposed CNN model with only 13 layers depth, when it was applied on set A, it was able to predict 90.6% of exons and 92.5% of introns, for set B, 94.8% of exons were correctly identified and 84.6% of introns were correctly identified and for set C 87.3% of exons were correctly recognized and 91.3% of introns were identified correctly. (see Fig. 11).

Fig. 11.

Fig. 11

Confusion matrices for our proposed CNN model. (a): set A, (b): set B, (c): set C.

5. Discussion

When classifying our images with Resnet-50 model, we have obtained a maximum training rate for the set B with a value of 90.42% and an accuracy value of 92% for the same set and the minimum execution time is about 7 hours and 30 minutes.

This Resnet-50 model was able to correctly classify 94% of introns for the three sets of images and 90% of exons for set B.

When classifying our images with GoogleNet model, we have obtained a maximum training rate for the set A with a value of 91.46% and an accuracy value of 91.5% for the same set and for the minimum execution time, it is about 2 hours 38 minutes and 37 seconds for the set B.

The GoogleNet model was capble of predicting correctly 96.3% of introns for set A and 90.4% of exons for set C.

When classifying our images with our proposed CNN model, we have obtained a maximum training rate for the set A with a value of 91.46% and an accuracy value of 91.6% for the same set and for the minimum execution time, it is about 2 hours 58 minutes and 53 seconds for the set B.

Our 13 layers depth proposed CNN model was able to predict 92.5% of introns for set A and 94.8% of exons for set B.

Our obtained results are comparable with results in28 where, based on a simplified version of ResNet, they suggested a novel model (S-ResNet). S-ResNet differs from Resnet in that it places a shortcut link at each layer of the convolution process rather than after a block made up of two convolution layers. They did classification of acceptor splice sites and donor splice sites using S-Resnet and compared it with other CNN models such as: AlexNet, VGG-Net and other CNN models. They reached an accuracy rate value of 93.5% for their proposed S-Resnet model. The results are also comparable with results in29 where they preprocessed their exon and intron dataset using clustering technique and they proposed ANFIS model (a combination between an artificial neural network and a fuzzy inference system) for the classification process. They obtained a value of 88.88% for the accuracy rate on the test phase.

As shown in Table 6, our proposed CNN model has a high value of specificity, a value of 94.2%, this means that our model is capable of classifying introns correctly. While, the GoogleNet model was able to correctly classify exons for set A with a value of 95.9% of sensitivity. The S-Resnet model outperformed the other four models with a value of 93.5% of accuracy rate.

Table 6.

Comparison of results.

Training rate Accuracy Execution Time
Model Acc. rate Sens. Spec.
Resnet-50 92% 94.2% 90.4%
GoogleNet 91.5% 95.9% 90.2%
13-layers CNN 91.6% 92.4% 94.2%
S-Resnet28 93.5% 92.9% 93.8%
ANFIS29 88.88% 75% 100%

6. Conclusion

In this work, we have used CNN models to classify color images representing human exons and introns. We have constructed 3 datasets, each set contains 3200 images divided equally into 2 classes. Then, we have used two pre-trained models: Resnet-50 and GoogleNet to classify those images, we have obtained 92% for the accuracy rate for the resnet-50 model and an accuracy rate of 91.5% for the GoogleNet model. After that, we have proposed our CNN model with 13 layers, we have reached promising results, we have obtained a value of 91.6% for the accuracy rate.

On a future work, we want to enlarge our dataset and we want to apply other pre-trained models to classify human exons and introns.

7. Declarations

List of abbreviations

DNA: Deoxyribonucleic Acid.

FCGR: Frequency Chaos Game Representation.

CNN: Convolutional Neural Network.

Ethics approval and consent to participate

All human genes are provided by the National Center of Biotechnology Information.

Consent for publication

Our manuscript does not any individual personal information.

Availability of data and materials

Our data will not be shared because it’s just the application of coding technique on gene sequences. But human gene sequences are available on https://www.ncbi.nlm.nih.gov/

Funding

There is no source funding for this research paper.

Authors’ contributions

All authors have equally contributed to this research paper

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Contributor Information

Feriel Ben Nasr Barber, Email: feriel.bennasr@enit.utm.tn.

Afef Elloumi Oueslati, Email: afef.elloumi@enit.utm.tn.

References

  • 1.Lesk A.M. Oxford University Press; 2017. Introduction to genomics. ??? [Google Scholar]
  • 2.Kerimov N., Hayhurst J.D., Peikova K., Manning J.R., Walter P., Kolberg L., Samoviča M., Sakthivel M.P., Kuzmin I., Trevanion S.J., et al. A compendium of uniformly processed human gene expression and splicing quantitative trait loci. Nat Genet. 2021;53(9):1290–1299. doi: 10.1038/s41588-021-00924-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Sinden R.R. Gulf Professional Publishing; 1994. DNA structure and function. ??? [Google Scholar]
  • 4.Gunasekaran H., Ramalakshmi K., Rex Macedo Arokiaraj A., Deepa Kanmani S., Venkatesan C., Suresh Gnana Dhas C. Analysis of dna sequence classification using cnn and hybrid models. Comput Math Methods Med 2021. 2021 doi: 10.1155/2021/1835056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Bernardi G. The human genome: organization and evolutionary history. Annu Rev Genet. 1995;29(1):445–476. doi: 10.1146/annurev.ge.29.120195.002305. [DOI] [PubMed] [Google Scholar]
  • 6.Venter J.C., Adams M.D., Myers E.W., Li P.W., Mural R.J., Sutton G.G., Smith H.O., Yandell M., Evans C.A., Holt R.A., et al. The sequence of the human genome. Science. 2001;291(5507):1304–1351. doi: 10.1126/science.1058040. [DOI] [PubMed] [Google Scholar]
  • 7.Souza N. The encode project. Nat Methods. 2012;9(11) doi: 10.1038/nmeth.2238. 1046–1046. [DOI] [PubMed] [Google Scholar]
  • 8.Leypold N.A., Speicher M.R. Evolutionary conservation in noncoding genomic regions. Trends Genet. 2021;37(10):903–918. doi: 10.1016/j.tig.2021.06.007. [DOI] [PubMed] [Google Scholar]
  • 9.Tanisawa K., Wang G., Seto J., Verdouka I., Twycross-Lewis R., Karanikolou A., Tanaka M., Borjesson M., Di Luigi L., Dohi M., et al. Sport and exercise genomics: the fims 2019 consensus statement update. Brit J Sports Med. 2020;54(16):969–975. doi: 10.1136/bjsports-2019-101532. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Banerjee K., Bali V. Design and development of bioinformatics feature based dna sequence data compression algorithm. EAI Endorsed Trans Pervasive Health Technol. 2019;5(20) doi: 10.4108/eai.13-7-2018.164097. [DOI] [Google Scholar]
  • 11.Jiang M., Bu C., Zeng J., Du Z., Xiao J. Applications and challenges of high performance computing in genomics. CCF Trans High Perform Comput. 2021;1–9 [Google Scholar]
  • 12.Ranganathan G., et al. A study to find facts behind preprocessing on deep learning algorithms. J Innovat Image Process (JIIP) 2021;3(01):66–74. [Google Scholar]
  • 13.Berrar D., Dubitzky W. Oxford University Press; 2021. Deep learning in bioinformatics and biomedicine. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Gülocak S.N., Bihter D. The effect of numerical mapping techniques on performance in genomic research. Sakarya Univ J Comput Inform Sci. 2022;5(3):315–340. [Google Scholar]
  • 15.Mouha R.A., et al. Deep learning for robotics. J Data Anal Inform Process. 2021;9(02):63. [Google Scholar]
  • 16.Wu, L., Chen, Y., Ji, H., Liu, B.: Deep learning on graphs for natural language processing. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2651–2653 (2021).
  • 17.Sungheetha A., Sharma R. 3d image processing using machine learning based input processing for man-machine interaction. J Innovat Image Process (JIIP) 2021;3(01):1–6. [Google Scholar]
  • 18.Lee W., Seong J.J., Ozlu B., Shim B.S., Marakhimov A., Lee S. Biosignal sensors and deep learning-based speech recognition: A review. Sensors. 2021;21(4):1399. doi: 10.3390/s21041399. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Shaheen M.Y. ScienceOpen Preprints; 2021. Adoption of machine learning for medical diagnosis. [Google Scholar]
  • 20.Lindsay G.W. Convolutional neural networks as a model of the visual system: Past, present, and future. J Cognit Neurosci. 2021;33(10):2017–2031. doi: 10.1162/jocn_a_01544. [DOI] [PubMed] [Google Scholar]
  • 21.Ghaderizadeh S., Abbasi-Moghadam D., Sharifi A., Zhao N., Tariq A. Hyperspectral image classification using a hybrid 3d–2d convolutional neural networks. IEEE J Select Top Appl Earth Observ Remote Sens. 2021;14:7570–7588. [Google Scholar]
  • 22.Gholamalinezhad H, Khosravi H. Pooling methods in deep neural networks, a review. arXiv preprint arXiv:2009.07485; 2020.
  • 23.Agarap AF. Deep learning using rectified linear units (relu). arXiv preprint arXiv:1803.08375; 2018.
  • 24.Gupta J, Pathak S, Kumar G. Deep learning (cnn) and transfer learning: A review. In: Journal of Physics: Conference Series, vol. 2273, p. 012029; 2022. IOP Publishing.
  • 25.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778; 2016.
  • 26.Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2015. p. 1–9.
  • 27.Nasr F.B., Oueslati A.E. In: 2022 6th International conference on advanced technologies for signal and image processing (ATSIP) IEEE; 2022. A new automatic method for human coding and non-coding zones characterization and classification based on fcgr coding and cnn classifier; pp. 1–6. [Google Scholar]
  • 28.Elsousy R., Kathiresan N., Boughorbel S. On the depth of deep learning models for splice site identification. bioRxiv. 2018;380667 [Google Scholar]
  • 29.Akalin F., Yumuşak N. Classification of exon and intron regions on dna sequences with hybrid use of sbert and anfis approaches. Politeknik Dergisi. 2023;1–1 [Google Scholar]

Articles from Journal of Genetic Engineering & Biotechnology are provided here courtesy of Academy of Scientific Research and Technology, Egypt

RESOURCES