Skip to main content
Animals : an Open Access Journal from MDPI logoLink to Animals : an Open Access Journal from MDPI
. 2023 Sep 8;13(18):2856. doi: 10.3390/ani13182856

Identification of Individual Hanwoo Cattle by Muzzle Pattern Images through Deep Learning

Taejun Lee 1, Youngjun Na 1,2, Beob Gyun Kim 1, Sangrak Lee 1, Yongjun Choi 1,*
Editor: Emma Fàbrega i Romans
PMCID: PMC10525771  PMID: 37760256

Abstract

Simple Summary

Cattle identification is necessary for precision feeding and management. Cattle muzzles have unique patterns, which can be used as a biometric identification key. This study aimed to identify cattle via a deep learning model based on muzzle images. Muzzle patterns were cropped from images using the YOLO v8-based image cropping model. Various artificial intelligence models based on neural networks were studied through transfer learning cropped images for cattle recognition with four optimizers. Several models showed a high prediction accuracy of over 97 percent, implicating the possibility for real farm usage.

Abstract

The objective of this study was to identify Hanwoo cattle via a deep-learning model using muzzle images. A total of 9230 images from 336 Hanwoo were used. Images of the same individuals were taken at four different times to avoid overfitted models. Muzzle images were cropped by the YOLO v8-based model trained with 150 images with manual annotation. Data blocks were composed of image and national livestock traceability numbers and were randomly selected and stored as train, validation test data. Transfer learning was performed with the tiny, small and medium versions of Efficientnet v2 models with SGD, RMSProp, Adam and Lion optimizers. The small version using Lion showed the best validation accuracy of 0.981 in 36 epochs within 12 transfer-learned models. The top five models achieved the best validation accuracy and were evaluated with the training data for practical usage. The small version using Adam showed the best test accuracy of 0.970, but the small version using RMSProp showed the lowest repeated error. Results with high accuracy prediction in this study demonstrated the potential of muzzle patterns as an identification key for individual cattle.

Keywords: cattle identification, deep learning, transfer learning, Efficientnet, Hanwoo, muzzle pattern

1. Introduction

Hanwoo (Korean native beef cattle, Bos taurus coreanae) is the dominant meat cattle species in Korea; Hanwoo cattle are reared in restricted-size pastures and provided with cattle feed due to limited land availability for husbandry purposes. Consequently, the meat prices are inevitably higher compared to those of products from overseas, which are produced in more extensive grazing areas. To maintain the competitiveness of Hanwoo cattle, the aim is to achieve premium meat production through specific precision breeding. Precision feeding tailored to individual requirements reduces feed wastage, economizes feed expenses and mitigates methane emissions resulting from excessive intake [1]. However, achieving precision specifications demands precise individual recognition methods.

Currently, tagging is the most popular identification method for cattle. Without exception, each individual cattle is managed with a Korean national traceability number containing 12 digits [2]. According to the study of Fosgate, the retention period of ear tags is 272 days on average, which was relatively shorter than expected [3]. Fosgate suggested relying solely on ear tags for long-term identification might not be sufficient, and close resemblance of the appearance among Hanwoo cattle might increase the risk of failure in individual recognition due to tag losses (Figure 1). To address the issue of ear-tag losses, employing supplementary identification methods can aid in ensuring recognition even in cases of ear-tag loss. While RFID chips attached to ear tags are a cheap and convenient technology used to identify cattle, a dedicated reader is needed. Furthermore, a Korean slaughter center runs a system in the base of RFID chips, and some errors are made during grading assessments by reading the wrong chip.

Figure 1.

Figure 1

Resemblance of the appearance among Hanwoo cattle.

Before the research, a probe was set to find supplementary methods utilizing inherent physical characteristics resistant to alteration, replication or counterfeiting, without requiring specialized identification devices. Biometrics are the identification method used for human individuals based on biological characteristics including face traits, iris patterns, and vein images [4]. Biometrics are highly reliable, consistent with their utilization in banks and airports. Within biometric traits, fingerprints have achieved a level of mainstream adoption, such as mobile phones and door locks. Cattle have unique muzzle patterns on their nose with similar ridge-valley characteristics and individuality to the human fingerprints providing immunity to alteration, replication, or counterfeiting [5].

Computer vision is a technology used for object detection, image classification and other tasks by analyzing images [6]. Presently, computer vision is one of the most rapidly advancing fields closely intertwined with the realm of artificial intelligence, due to the evolution of computer hardware and the proliferation of imagery accessible through the widespread use of mobile phones equipped with cameras [7]. Machine learning algorithms are algorithm improving automatically through experience from given image data [8]. If machine learning algorithms provide inaccurate predictions, engineers need to intervene and make adjustments. When using deep learning models, the algorithm can assess the accuracy of predictions through its own neural network, eliminating the need for human assistance.

Several studies have been conducted to use muzzle patterns to identify cattle and have provided results with various species of cattle [9,10]. Hanwoo cattle have no pattern or color differences on the muzzle, but previous research did not investigate breeds characterized by monochromaticity. Therefore, the objective of this study was to identify individual Hanwoo cattle through images of muzzle patterns through deep learning.

2. Materials and Methods

2.1. Training Dataset

A total of 9230 muzzle images from 336 individual Hanwoo cattle were constructed from two beef farms (Table 1). No direct contact was made while collecting the image data, and images were collected using cell phone cameras (iPhone Xs, Apple, Cupertino, CA, USA, and Galaxy s 20, Samsung, Suwon, South Korea) from 24 August 2021 to 12 November 2021 (Figure 2).

Table 1.

Location, number of Hanwoo cattle and number of images enrolled in dataset.

Region Location
(Latitude, Longitude)
Number of Animals Images
Jeongeup-si 35.62934, 126.87748 235 6160
Wonju-si 37.20481, 127.5141 101 3070
Sum 336

Figure 2.

Figure 2

Samples of image data with national livestock traceability numbers. The national livestock traceability number is a unique identifier assigned to cattle at birth by the nation in order to identify individuals within the cattle population.

The distance between cattle and camera was pinned at 30 cm with the muzzle pattern centered in the image. To capture whole muzzle patterns, capturing whole bottom lips were the essential requirements. For each individual cattle, approximately 20 images were captured, and five pictures were taken four times each to avoid overfitting the model. Some images containing foreign substances were added.

Manual annotations of the muzzle pattern area were added to 150 images using a commercial annotation tool (roboflow; https://roboflow.com (accessed on 17 July 2023)), and the results were saved in the form of bounding box coordinates in xml files. Guidelines for this annotation were: 1. Configure the endline setting to include the whole muzzle area. 2. Include the line of the bottom lip. A model detecting the muzzle pattern area was developed with 150 images and bounding box coordinates using the YOLO v8 algorithm [11]. Consequently, the identified bounding box coordinates of muzzle area were detected for the remaining data. Muzzle areas in images were cropped with xml files using the Python Imaging Library in order to make the dataset used to train the cropping model (Figure 3). The entire image dataset was divided into three categories: training, validation and testing, with split ratios of 75 percent, 15 percent and ten percent, respectively, according to the guidelines of Yadav and Shukla [8]. Training and validation datasets were shown through the model for the transfer learning, and test datasets were not shown when evaluating the practical application.

Figure 3.

Figure 3

Example of cropping muzzle in the image with YOLO v8-based crop model.

2.2. Data Transformation

Data transformation is useful when modifying data in order to reduce bias [12]. Resize, rotation, flip, warp in frame, adjustment of brightness and contrast can be accomplished in order to decrease the similarity of interference from surrounding factors in the image. Images in each dataset were resized to 224 × 224 pixels, randomly flipped in vertical and horizontal axis and rotated randomly within the range of 20 degrees. After the transformation, image data were normalized by subtracting the mean value for each RGB channel (0.485, 0.456, 0.406 in order) and dividing by standard deviations (0.229, 0.224, 0.225) (Table 2) [13]. The normalization process maintains the transformed image data with a mean close to zero and a standard deviation close to one.

Table 2.

Parameters used through data transformation for train, validation and test dataset.

Parameters Train Data Validation Data Test Data
Resize 224 × 224 pixels 224 × 224 pixels 224 × 224 pixels
Horizontal flip Random Random None
Vertical flip Random Random None
Rotation 0–20 degrees 0–20 degrees None
RGB mean value
RGB standard deviation
[0.485, 0.456, 0.406], [0.229, 0.224, 0.225] [0.485, 0.456, 0.406], [0.229, 0.224, 0.225] [0.485, 0.456, 0.406], [0.229, 0.224, 0.225]

2.3. Data Loader

Data blocks with images and labels were combined to make an image-based classifying model. In Korea, each individual cattle is managed with a twelve-digit number code called the livestock traceability number. The livestock traceability number is given at birth and used until slaughter and distribution. Each muzzle-cropped image and matching national livestock traceability numbers were combined in a data block. Data blocks of train and validation sets were sent to each GPU device separately, by batch sizes set to 32. The term ‘epoch’ refers to a time during which the entire dataset undergoes one complete calculation. Training data were used to train the model through the epoch, and validation data were used to validate the model at the end of the epoch.

2.4. Transfer Learning

In classification models, a loss serves as a metric indicating the disparity between the model’s predictions and the actual targets, aiming to minimize this discrepancy during the training [14]. The two-by-two confusion matrix is a commonly used method for interpretation, composed of actual negative and positive as rows and predicted negative and positive as columns. By adopting the confusion matrix, all predictions can be sorted into four classes based on the relationship between predicted and actual values (Table 3) [15]. Accuracy mentioned in this paper refers to the sum of true positive and negative divided by the total number of predictions:

Accuray=True positive+True negativeTotal number of prediction (1)

Table 3.

Concept of two-by-two confusion matrix as classification interpretation.

Predicted Negative Predicted Positive
Actual negative True negative False positive
Actual positive False negative True positive

The parameters used to train were model, loss function, optimizer algorithm, learning rate management scheduler and the number of epochs. Each epoch was segmented into the train phase and the validation phase, and recorded the loss and accuracy. At the beginning of each epoch, inputs for each data batch along with labels were received and the gradients of the parameters were initialized through the optimizer. During the training phase, model calculations through forward propagation on the inputs were tracked to generate outputs and passed to the loss function. The loss gradients for each parameter received from the forward propagation were computed by the loss function in the backpropagation stage. Optimizer updated parameters in directions with loss gradients to minimize the loss. If the performance of the epoch’s training results exceeded the previous best during the validation phase, the model’s weights were deemed optimal and saved.

2.4.1. Model

Efficientnet v2 models were opted because of the availability of random weight versions designed for custom datasets [16]. EfficientNet v2 represented an advancement in neural network architecture aiming to enhance efficiency and performance beyond conventional convolutional methods. The essence lies in co-optimizing network depth, width and resolution through compound scaling. Unlike previous models, compound scaling balances dimensions for minimized resource usage and maintained accuracy. EfficientNet v2 refines the original architecture, offering variants tailored to balance computation and performance. Noteworthy features include effective high-resolution image handling, compact model preservation, and a range of variants for trade-offs between size and performance.

The architecture of EfficientNet v2 comprises three key components: the backbone, fused-mobile vision convolutional network (MBConv) layer, and scaling and kernel size adjustment [17]. Backbone of EfficientNet v2 was designed through Neural Architecture Search (NAS) to strike a balance between computation and parameter efficiency, offering a balance between computational complexity and performance, with the ability to scale the model to various sizes [18]. Fused-MBConv layers enhances operational efficiency by combining convolutional layer and depth-wise convolution layer (DWConv) at the early stage from image input [16]. EfficientNet v2 employs a scaling strategy to expand the model’s size while maintaining computational efficiency, and Figure 4 demonstrated the overall schematic of the small model (EfficientNet v2-S) (Figure 4).

Figure 4.

Figure 4

Efficientnet v2-s Architecture—(a) represents the schematic of the model from image input to logit output. The numbers indicated on the right side of the boxes represent the layer numbers, while the numbers displayed below the boxes indicate the corresponding stride and channels. (be) are the schematic of each specific algorithm in (a); Fused-MBConv1, Fused-MBConv4 (k3 × 3), MBConv4 (k3 × 3), and MBConv6 (k3 × 3), respectively. The abbreviations are expanded as follows: MBConv as mobile vision convolutional network, SE as squeeze and excitation, FC as full connect, DW as depth wise, BN as batch normalization, H as height, W as width and F as number of channels.

Three variations of EfficientNet v2 models (tiny, small and medium) were used. While these three models perform similar roles, they differ in terms of the number of parameters, Giga multiply accumulates (GMACs), and number of activations (Table 4). Parameters in the model refer to the weights and biases for learning. GMACs serves as a metric indicating the number of operations performed by a model, representing its computational complexity [19]. Activations are the total number of intermediate values generated between each layer within the model. Each layer transforms input data through weights and biases, applying the results to activation functions before passing them to the next layer.

Table 4.

Comparison of Efficientnet v2 models in detail of parameters, Giga Multiply Accumulates (GMACs) and number of activations.

Efficientnet v2
Tiny
Efficientnet v2
Small
Efficientnet v2
Medium
Parameters (million) 13.6 23.9 53.2
GMACs (giga) 1.9 4.9 12.7
Activation (million) 9.9 21.4 47.1

2.4.2. Loss Function

The criterion used for the loss function was cross entropy loss applied across all models. Cross entropy loss is commonly used loss function in classification. The difference between the actual class and the predicted class is measured to guide the learning process of the model. Cross entropy loss is calculated as

Loss=1Nn=1Nlog(iqy~nip^y*=ixn) (2)

where N is the number of training samples, y~ represents predicted class, y* represents the actual class and x represents the input image [20].

2.4.3. Optimizer Algorithms

Stochastic gradient descent (SGD), root-mean-square propagation (RMSProp), adaptive moment estimation (Adam) and evolved sign momentum (Lion) were used as optimizers.

SGD algorithm estimates the gradient on randomly picked batch zt:

wt+1=wtγtQ(zt,wt) (3)

where w is the weight, γ is learning rate and Q(zt,wt) is the loss function [21]. With this equation, weights are updated in each direction to minimize the loss.

RMSProp algorithm works similar to SGD but apply different learning rates to each parameter by moving average of the squared gradients, whereas SGD uses a fixed learning rate [22].

Adam algorithm tracks the decaying average of the past gradients, term of momentum, and corrects bias occurred from moving the average of the squared gradients in the RMSProp algorithm [23].

The Lion optimizer is a trending algorithm which is more memory-efficient than the Adam algorithm and with a similar structure; it can be utilized by adjusting the learning rate and weight decay simultaneously [24].

2.4.4. Learning Rate Schedular

The learning rate is a scaled value applied to the gradient to update the weights. A lower learning rate requires more epochs for effective learning, whereas a higher learning rate could potentially hinder the reduction in loss [25]. The learning rate was set to the basic level of 0.001 for all optimizers: SGD, MRS Prop, Adam and Lion. The learning rate was adjusted every seven epochs with a reduction ratio of 0.1, consistently applied across all models [26].

2.4.5. Epochs

Epochs for learning were pinned to 100 for the whole learning process.

2.5. Computing Resources

The experiments were conducted on a computing cluster with the following specifications:

  • CPU: Intel(R) Xeon(R) w5-3433 1.99 GHz;

  • RAM: 256 GB of DDR4 RAM;

  • GPU: NVIDIA GeForce RTX 4090.

The deep learning models were implemented using the following software configurations:

  • Operating system: Microsoft Windows 10.0.19045.3324 version 22H2;

  • CUDA: CUDA Version: 11.8;

  • Python: 3.11.4;

  • PyTorch: 2.0.0.

3. Results

3.1. Loss and Accuracy Metrics through Transfer Learning through Train and Validation Data

Training loss, training accuracy, validation loss and validation accuracy were tracked throughout the entire transfer learning for all models. The quantity of all the metrics is vast since there are 100 epochs from 12 models. The flow of each of the metrics is visualized on each single graph in Figure 5 and Figure 6, and specific numeric metrics of each model from the best validation epoch accuracy are presented in Table 5.

Figure 5.

Figure 5

Figure 5

Training loss (a) and training accuracy (b) visualized through transfer learning across all models.

Figure 6.

Figure 6

Figure 6

Validation loss (a) and validation accuracy (b) visualized through transfer learning across all models.

Table 5.

Training loss, training accuracy, validation loss and validation accuracy derived from the epoch with the highest validation accuracy achieved through transfer learning across all models.

Model Best Epoch Epoch Time (s) Train_loss Train_acc Val_loss Val_acc
Tiny-SGD 97 59 2.638 0.489 2.381 0.539
Tiny-RMSprop 97 62 0.310 0.917 0.261 0.932
Tiny-Adam 44 58.5 0.133 0.968 0.104 0.976
Tiny-Lion 55 54.9 0.136 0.966 0.098 0.976
Small-SGD 38 62.2 5.791 0.008 5.793 0.014
Small-RMSprop 26 59.1 0.141 0.962 0.091 0.977
Small-Adam 67 58.5 0.111 0.972 0.090 0.978
Small-Lion 36 59 0.090 0.976 0.077 0.981
Medium-SGD 89 60.5 5.803 0.007 5.804 0.011
Medium-RMSprop 97 60.2 0.265 0.937 0.371 0.907
Medium-Adam 51 65.3 0.138 0.968 0.231 0.944
Medium-Lion 22 65.1 0.131 0.968 0.256 0.942

Train_loss—training loss; Train_acc—training accuracy; Val_loss—vValidation loss; Val_acc—validation accuracy.

3.2. Prediction Result on Test Dataset

The top five models showing high validation accuracies with a low time taken for each epoch were evaluated through predictions made on test data. Accuracy, testing time, overall number of errors and repeated errors on the same objects of prediction on the test dataset are presented in Table 6.

Table 6.

Accuracy of prediction from top five models on validation accuracy and epoch time on test data with time elapsed with numbers of entire errors and multiple error prediction on same subject (repeated error).

Model Test Accuracy Testing Time (s) Error Repeated Error
Tiny-Adam 0.967 18.6 92 14
Tiny-Lion 0.967 18.2 92 12
Small-RMSprop 0.968 18.3 91 3
Small-Adam 0.970 17.9 95 9
Small-Lion 0.965 18.2 100 18

4. Discussion

The deep learning models were investigated for their use in identifying individual cattle based on whole muzzle images at the beginning of the study. However, each image contained an average of 4 MB of information, and the database consisted of 17,325 images, resulting in a large dataset size of approximately 69.3 GB. Processing a significant amount of data led to lengthy analysis and model response times [27]. To make the analysis more practical for real-world farm use, the idea of cropping the muzzle area was experimented with.

To build a high-quality model with accurate predictions, the selection of an appropriate model is crucial. The artificial neural network converts image data into one dimension, finds characteristics, and then laminates them to highlight the features [28]. Within the models varying in parameter sizes, the small model showed better validation accuracy in comparison to the tiny and medium models for the same optimizer. In terms of optimizer, SGD algorithms were not fit for the Efficientnet V2 model and images in dataset showing very low metrics including validation accuracy. Considering the optimizer as a focal point, overall validation accuracies were higher in Lion and Adam algorithms with similar metrics than RMSProp, but small Efficientnet V2 model with RMSProp showed the best performance in testing practical use.

Despite the successful development of a deep learning model for individual Hanwoo cattle identification using muzzle patterns, there are certain limitations in this study. One limitation is the reliance on image data captured through cell phone cameras, which could potentially lead to variations in image quality and consistency. Additionally, the model’s performance might be influenced by changes in angles and other environmental factors during image capture.

Additionally, the model’s accuracy could be impacted by the challenges of recognizing cattle in real-world scenarios. Factors including changes in cattle appearance due to weather conditions, or health status could affect the model’s reliability in practical farming environments. Furthermore, the transition from research to real-world applications may require addressing issues related to computational resources, deployment and integration with existing farm management systems. Finally, while the achieved predictive accuracy is promising, the model’s performance should be further validated and refined using larger and more diverse datasets. Expanding the dataset to include a wider range of cattle breeds and environmental conditions will help assess the model’s generalization capabilities and identify potential biases.

Inscription and fingerprints share similarities, but there are notable differences in their measurement methods. Fingerprints exhibit one-dimensional characteristics as they are obtained by pressing the finger onto a surface, which captures the unique patterns through pressure [29]. On the other hand, inscriptions are relatively more complex images with three-dimensional characteristics. They require data to be obtained through camera images, capturing the intricate details and depth of the inscribed patterns. The complex architecture should be capable of capturing and processing the detailed depth information inherent in the inscriptions, enabling more precise predictions to be made. It seems that these complexities indeed necessitate a more complex architecture in order to make accurate predictions.

Hanwoo cattle have light brown fur with no specific patterns or colors on its face or muzzle (Figure 1), while Holstein cattle have black and white fur with a variety of patterns. Furthermore, the muzzle area of Holstein cattle can range from pink to black or brown in color. Prior to initiating the current study, the ResNet architecture, commonly utilized in image classification, was used to train models to explore the possibilities of muzzle pattern identification for both Hanwoo and Holstein. The Holstein-identifying model showed better performance than the Hanwoo-identifying model. The difference in performance between the Hanwoo and Holstein models prompted us to consider that the model might have learned more of the colors of the muzzle area or the overall face, rather than solely capturing the unique patterns of ridges and valleys.

While the high accuracy of the Holstein model leads to the perception of its strong performance, but if the model relies on colors and skin pattern, colors and patterns might become a hurdle for recognizing ridges and valley patterns of the muzzle across various breeds in the future. Therefore, this study aimed to ensure that the model focuses primarily on the patterns and not just the color variations. After research on the Hanwoo breed, an identification model for Holstein cattle was developed through the same methods, with 7835 images of 297 Holstein cattle. The small Efficientnet V2 model, presented as the best performing model for real usage in this research, showed 0.953 predictive accuracy on test data, presenting subtle variance compared with the Hanwoo model, which even showed better performance.

The object of this study was to identify Hanwoo cattle using the deep learning model trained with muzzle images, as mentioned in the Introduction. For this study to hold significance, the developed models must be applicable in practical industry. The high predictive accuracy on unseen data implies the possibility of the model being applied to precision feeding at the farm level without dedicated devices. However, this system still requires a high level of computing resources on a large scale. If slaughterhouses set up a server for the region, farms could identify individual cattle with no computer resources, and slaughterhouses could further strengthen the system for grading. Expanding the scope, data from cattle, including genetic traits, can be easily accumulated on a national scale.

Deep learning is a very convenient technology used to reduce human resources including labor and concentration because the computer trains using the traits of image data without human supervision. With the current technology, while deep learning offers convenience in automating the training process, it remains challenging to determine the key traits for accurately classifying cattle in images [30]. The inner workings of the model and how it achieves classification are not fully understood. However, the development of an accurate classification model for cattle individual based on their muzzle patterns, as demonstrated in this study, holds significant meaning. It suggests the potential for non-contact individual identification through image analysis. Future studies will focus on reducing the complexity of the main data to eliminate unnecessary variables and identify the key traits for classification, as well as trying to make generalized models for various species.

5. Conclusions

This study aimed to develop a deep learning model based on muzzle images, with a focus on Hanwoo cattle. The small Efficientnet V2 model presented a high level of predictive accuracy. While showing subtly better performance compared with same model learned from Holstein, the performance of the Hanwoo model implies the potential for real-farm usage in terms of precision feeding and management strategies, as well as facilitating traceability on a national scale. Furthermore, the proposed approach introduces this perspective by utilizing inherent physical characteristics of cattle for identification purposes, reducing the reliance on ear tags.

As the field of deep learning continues to evolve with developments in computer hardware and the tendency for such hardware to become cheaper, further research could investigate refining the accuracy of the model, exploring other traits and extending its applicability. Improved studies will offer improved precision breeding and more efficient management, bringing advantages to both farmers and the industry.

Acknowledgments

This work was supported by Korea Institute of Planning and Evaluation for Technology in Food, Agriculture and Forestry (IPET) and Korea Smart Farm R&D Foundation (KosFarm) through Smart Farm Innovation Technology Development Program, funded by Ministry of Agriculture, Food and Rural Affairs (MAFRA) and Ministry of Science and ICT (MSIT), Rural Development Administration (RDA) (421014-04) and this paper was supported by the KU Research Professor Program of Konkuk University.

Author Contributions

Conceptualization, S.L. and Y.C.; methodology, Y.N. and T.L.; formal analysis, T.L.; investigation, T.L.; data curation, T.L. and Y.N.; writing—original draft preparation, T.L.; writing—review and editing, T.L., B.G.K. and Y.C.; visualization, T.L.; supervision, Y.C. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This work was funded by the Ministry of Agriculture, Food and Rural Affairs (MAFRA) and Ministry of Science and ICT (MSIT), Rural Development Administration (RDA) of South Korea.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Koch R.M., Swiger L.A., Chambers D., Gregory K.E. Efficiency of Feed Use in Beef Cattle. J. Anim. Sci. 1963;22:486–494. doi: 10.2527/jas1963.222486x. [DOI] [Google Scholar]
  • 2.Chung K.Y., Lee S.H., Cho S.H., Kwon E.G., Lee J.H. Current Situation and Future Prospects for Beef Production in South Korea—A Review. Asian-Australas. J. Anim. Sci. 2018;31:951–960. doi: 10.5713/ajas.18.0187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Fosgate G.T., Adesiyun A.A., Hird D.W. Ear-Tag Retention and Identification Methods for Extensively Managed Water Buffalo (Bubalus Bubalis) in Trinidad. Prev. Vet. Med. 2006;73:287–296. doi: 10.1016/j.prevetmed.2005.09.006. [DOI] [PubMed] [Google Scholar]
  • 4.Jain A., Bolle R., Pankanti S. Introduction to Biometrics. In: Jain A.K., Bolle R., Pankanti S., editors. Biometrics: Personal Identification in Networked Society. Springer; Boston, MA, USA: 1996. pp. 1–41. [Google Scholar]
  • 5.El Hadad H.M., Mahmoud H.A., Mousa F.A. Bovines Muzzle Classification Based on Machine Learning Techniques. Procedia Comput. Sci. 2015;65:864–871. doi: 10.1016/j.procs.2015.09.044. [DOI] [Google Scholar]
  • 6.Brownlee J. Deep Learning for Computer Vision: Image Classification, Object Detection, and Face Recognition in Python. Machine Learning Mastery; Vermont, Australia: 2019. [Google Scholar]
  • 7.Deng J., Berg A.C., Li K., Li F.-F. What Does Classifying More Than 10,000 Image Categories Tell Us? In: Daniilidis K., Maragos P., Paragios N., editors. Proceedings of the Computer Vision—ECCV 2010. Springer; Berlin/Heidelberg, Germany: 2010. pp. 71–84. [Google Scholar]
  • 8.Mitchell T.M. Machine Learning. 1997. [(accessed on 17 July 2023)]. Available online: https://www.cin.ufpe.br/~cavmj/Machine%20-%20Learning%20-%20Tom%20Mitchell.pdf.
  • 9.Bello R.W., Olubummo D.A., Seiyaboh Z., Enuma O.C., Talib A.Z., Mohamed A.S.A. Cattle Identification: The History of Nose Prints Approach in Brief. IOP Conf. Ser. Earth Environ. Sci. 2020;594:012026. doi: 10.1088/1755-1315/594/1/012026. [DOI] [Google Scholar]
  • 10.Kusakunniran W., Wiratsudakul A., Chuachan U., Imaromkul T., Kanchanapreechakorn S., Suksriupatham N., Thongkanchorn K. Analysing muzzle pattern images as a biometric for cattle identification. Int. J. Biom. 2021;13:367–384. doi: 10.1504/IJBM.2021.117852. [DOI] [Google Scholar]
  • 11.Jocher G., Chaurasia A., Qiu J. YOLO by Ultralytics 2023. [(accessed on 17 July 2023)]. Available online: https://github.com/ultralytics/ultralytics.
  • 12.Jiang Y., Cukic B., Menzies T. Can Data Transformation Help in the Detection of Fault-Prone Modules?; Proceedings of the 2008 Workshop on Defects in Large Software Systems; Seattle, WA, USA. 20 July 2008; New York, NY, USA: Association for Computing Machinery; 2008. pp. 16–20. [Google Scholar]
  • 13.Finlayson G.D., Schiele B., Crowley J.L. Comprehensive Colour Image Normalization. In: Burkhardt H., Neumann B., editors. Proceedings of the Computer Vision—ECCV’98; Freiburg, Germany. 2–6 June 1998; Berlin/Heidelberg, Germany: Springer; 1998. pp. 475–490. [Google Scholar]
  • 14.Ma X., Huang H., Wang Y., Romano S., Erfani S., Bailey J. Normalized Loss Functions for Deep Learning with Noisy Labels; Proceedings of the 37th International Conference on Machine Learning; Virtual. 13–18 July 2020; pp. 6543–6553. [Google Scholar]
  • 15.Giese A., Seitzer J. Using a Genetic Algorithm to Evolve a D* Search Heuristic. 2011. [(accessed on 17 July 2023)]. p. 72. Available online: https://ceur-ws.org/Vol-710/paper31.pdf.
  • 16.Tan M., Le Q.V. EfficientNetV2: Smaller Models and Faster Training; Proceedings of the 38th International Conference on Machine Learning; Virtual. 18–24 July 2021. [Google Scholar]
  • 17.Tan M., Chen B., Pang R., Vasudevan V., Sandler M., Howard A., Le Q.V. MnasNet: Platform-Aware Neural Architecture Search for Mobile; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); Long Beach, CA, USA. 15–20 June 2019; pp. 2820–2828. [Google Scholar]
  • 18.Tan M., Le Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks; Proceedings of the 36th International Conference on Machine Learning; Long Beach, CA, USA. 24 May 2019; pp. 6105–6114. [Google Scholar]
  • 19.Garland J., Gregg D. Low Complexity Multiply-Accumulate Units for Convolutional Neural Networks with Weight-Sharing. ACM Trans. Archit. Code Optim. 2018;15:1–24. doi: 10.1145/3233300. [DOI] [Google Scholar]
  • 20.Sukhbaatar S., Bruna J., Paluri M., Bourdev L., Fergus R. Training Convolutional Networks with Noisy Labels. arXiv. 20151406.2080 [Google Scholar]
  • 21.Bottou L. Large-Scale Machine Learning with Stochastic Gradient Descent. In: Lechevallier Y., Saporta G., editors. Proceedings of the COMPSTAT’2010; Paris France. 22–27 August 2010; Heidelberg, Germany: Physica-Verlag HD; 2010. pp. 177–186. [Google Scholar]
  • 22.Kurbiel T., Khaleghian S. Training of Deep Neural Networks Based on Distance Measures Using RMSProp. arXiv. 20171708.01911 [Google Scholar]
  • 23.Kingma D.P., Ba J. Adam: A Method for Stochastic Optimization. arXiv. 20171412.6980 [Google Scholar]
  • 24.Chen X., Liang C., Huang D., Real E., Wang K., Liu Y., Pham H., Dong X., Luong T., Hsieh C.-J., et al. Symbolic Discovery of Optimization Algorithms. arXiv. 20232302.06675 [Google Scholar]
  • 25.Smith L.N. A Disciplined Approach to Neural Network Hyper-Parameters: Part 1—Learning Rate, Batch Size, Momentum, and Weight Decay. arXiv. 20181803.09820 [Google Scholar]
  • 26.Smith L.N. Cyclical Learning Rates for Training Neural Networks; Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV); Santa Rosa, CA, USA. 24–31 March 2017; pp. 464–472. [Google Scholar]
  • 27.Agarwal S., Mozafari B., Panda A., Milner H., Madden S., Stoica I. BlinkDB: Queries with Bounded Errors and Bounded Response Times on Very Large Data; Proceedings of the 8th ACM European Conference on Computer Systems; Prague, Czech Republic. 15–17 April 2013; New York, NY, USA: Association for Computing Machinery; 2013. pp. 29–42. [Google Scholar]
  • 28.Maind S.B., Wankar P. Research Paper on Basic of Artificial Neural Network. Int. J. Recent Innov. Trends Comput. Commun. 2014;2:96–100. [Google Scholar]
  • 29.Komarinski P.D. Cold Case Homicides. CRC Press; Boca Raton, FL, USA: 2017. Automated Fingerprint Identification Systems. [Google Scholar]
  • 30.Reimers C., Runge J., Denzler J. Determining the Relevance of Features for Deep Neural Networks. In: Vedaldi A., Bischof H., Brox T., Frahm J.-M., editors. Proceedings of the Computer Vision—ECCV 2020; Glasgow, UK. 23–28 August 2020; Cham, Switzerland: Springer International Publishing; 2020. pp. 330–346. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data sharing is not applicable to this article.


Articles from Animals : an Open Access Journal from MDPI are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES