Skip to main content
Animals : an Open Access Journal from MDPI logoLink to Animals : an Open Access Journal from MDPI
. 2024 Oct 18;14(20):3014. doi: 10.3390/ani14203014

Machine Learning for the Genomic Prediction of Growth Traits in a Composite Beef Cattle Population

El Hamidi Hay 1
Editor: Sang Hong Lee1
PMCID: PMC11505319  PMID: 39457945

Abstract

Simple Summary

Genomic selection is commonly used in many livestock species to predict important traits. However, the current methods used to make these predictions are not perfect. New approaches, like machine learning, could make these predictions more accurate because they can handle complex relationships in the data. In this study, we tested machine learning methods, Random Forest, Support Vector Machine, Multi-Layer Perceptron, and Convolutional Neural Networks to predict birth weight, weaning weight, and yearling weight in a beef cattle population. We compared these methods with three traditional ones—GBLUP, BayesA, and BayesB. The GBLUP method was the most accurate for predicting birth and yearling weights, while the Random Forest method was better at predicting weaning weight. Additionally, GBLUP provided a closer match to the actual data. Overall, GBLUP gave better predictions and fit the data better than the machine learning methods we tested.

Abstract

The adoption of genomic selection is prevalent across various plant and livestock species, yet existing models for predicting genomic breeding values often remain suboptimal. Machine learning models present a promising avenue to enhance prediction accuracy due to their ability to accommodate both linear and non-linear relationships. In this study, we evaluated four machine learning models—Random Forest, Support Vector Machine, Convolutional Neural Networks, and Multi-Layer Perceptrons—for predicting genomic values related to birth weight (BW), weaning weight (WW), and yearling weight (YW), and compared them with other conventional models—GBLUP (Genomic Best Linear Unbiased Prediction), Bayes A, and Bayes B. The results demonstrated that the GBLUP model achieved the highest prediction accuracy for both BW and YW, whereas the Random Forest model exhibited a superior prediction accuracy for WW. Furthermore, GBLUP outperformed the other models in terms of model fit, as evidenced by the lower mean square error values and regression coefficients of the corrected phenotypes on predicted values. Overall, the GBLUP model delivered a superior prediction accuracy and model fit compared to the machine learning models tested.

Keywords: machine learning, genomic prediction, beef cattle

1. Introduction

The advent of high-throughput technology and its decreasing cost have allowed for the adoption of genomic selection in most livestock species [1,2,3]. The accuracy of genomic selection is influenced by several factors such as the density of the single nucleotide polymorphism (SNP) panel used, the size of the training population, the genetic relatedness of the validation and training populations, the heritability, and the prediction method employed [4,5]. Currently, the methods predominantly used include Genomic Best Linear Unbiased Predictor (GBLUP); single-step GBLUP (ssGBLUP), which combines genotyped and non-genotyped animals [6]; and Bayesian regression models [7,8]. The primary differences among these approaches generally arise from the assumed distribution of SNP marker effects [9]. These models fail to capture complex non-linear interactions such as dominance and epistasis. Therefore, exploring other prediction models is warranted.

In recent years, machine learning (ML) models have garnered the interests of fields including plant and animal breeding [10,11]. Several studies using ML models showed some improvement in accuracy compared to GBLUP and other Bayesian alphabet models [12,13]. However, their superiority did not extend to all traits and species [14,15]. Therefore, the objective of this study is to compare the performance of ML models with existing genomic prediction approaches using the growth traits of a closed composite beef cattle population.

2. Materials and Methods

2.1. Data

Data used in this study consisted of 4680 Composite Gene Combination (CGC) animals (½ Red Angus, ¼ Charolais, ¼ Tarentaise; [16]) born between 2002 and 2019 at USDA-ARS, Fort Keogh Livestock and Range Research Laboratory, Miles City, MT. The Pedigree consisted of 9903 animals.

Phenotypes consisted of birth weight (BW), weaning weight (WW), and yearling weight (YW); for further details on the phenotypes and management refer to [17]. Weaning weights were adjusted to 205 days, yearling weights were adjusted to 365 days, and outliers were removed. A summary description of the data is presented in Table 1.

Table 1.

Estimated heritabilities and summary statistics of the growth traits analyzed.

Trait n h2 (SE) Mean, kg SD, kg
BW (kg) 4660 0.38 (0.04) 32.73 4.13
WW (kg) 4651 0.34 (0.09) 204.69 27.70
YW (kg) 4563 0.28 (0.07) 266.87 35.70

Animals were genotyped using a mixture of low-density SNP 3k panel and high-density Illumina Bovine50k (Illumina, San Diego, CA, USA). Genotypes were called in the Illumina Genome Studio software V2.0.5. Quality control was performed, which consisted of excluding SNP markers with minor allele frequencies of less than 0.05, SNP with a call rate (CRSNP) < 0.90, and Fisher’s exact test p-values for Hardy–Weinberg equilibrium < 1 × 10−5. After quality control, the number of SNP genotypes consisted of 40,533 SNP markers. Missing genotypes for animals with low-density SNP panel were imputed with FImpute V3.0 software using population and pedigree information [18]. The average allelic R2 was 0.94, which indicates the high imputation accuracy of the missing genotypes.

2.2. Statistical Models

In this study, genomic prediction analysis was carried out using corrected phenotypes as the response variable. The fixed effects to correct the phenotypes included sex and contemporary group effect (year of calving and age of dam subclasses). Correction was performed to remove any bias due to the double counting of phenotypic and pedigree information. BLUPF90 software was used [19].

The classical pedigree-based Best Linear Unbiased Predictor (BLUP) using a single-trait model was performed to separately estimate the variance components and breeding values for each trait. The model is as follows:

y=Xb+Z1a+Z2m+e, (1)

in which y, b, a, m, and e are the vectors of phenotypes, fixed effects, random additive genetic effects, random maternal genetic effects, and residual effects, respectively; X, Z1, and Z2 are the incidence matrices relating fixed effects, random additive genetic effects, and random maternal genetic effects, respectively, to the observations. For YW, the maternal effect was not included.

2.2.1. GBLUP

The Genomic Best Linear Unbiased Predictor (GBLUP) model is described in [20] and is as follows:

y=Za+e, (2)

where y is the vector of corrected phenotypes, a is the vector of random animal additive effects, e is the vector of random errors, and Z is an incidence matrix allocating observations to a. The random additive effects were distributed as g~N(0, Gσg2), where G is the genomic relationship matrix, and σg2 and σe2 were the additive genetic variance. Random errors were distributed as e~N (0, Iσg2), where σg2 is the residual variance. The model was implemented using the BLUPF90 package [19].

2.2.2. Bayes A

The Bayes A model used to estimate SNP effects is as follows:

y=μ+j=1nzjαj+e, (3)

where y is the vector of corrected phenotypes, µ is the overall mean, n is the number of SNP, zj is the genotype covariate of the jth SNP coded according to the additive model (0, 1, and 2), αj is the allelic substitution effect of SNPj, and e is the vector of random residuals. A total of 50,000 MCMC iterations with 10,000 burn-in cycles were used. Convergence testing was performed for all parameters following the methods of J Geweke [21] and P Heidelberger and PD Welch [22], and a visual analysis of trace plots was also performed using Bayesian Output Analysis program in R software 4.4.1 (R Development Core Team, Vienna, Austria).

2.2.3. Bayes B

For Bayes B, the model is similar to the Bayes A model in Equation (2), except the SNP effects are modeled as j=1nzjαjδj, where zj is the genotype of the jth marker, coded according to the additive model (0, 1, and 2); αj is the effect of SNP marker j; and δj is an indicator variable that is equal to 1 if the jth marker has a non-zero effect on the trait and 0 otherwise. A binominal distribution with known probability π = 0.01 was assumed for δj. For both the Bayes A and Bayes B models, GenSel software (Version 2.14) was used.

2.2.4. Multi-Layer Perceptron

Multi-Layer Perceptron (MLP) is a type of artificial neural network (NN) composed of multiple layers of nodes [23,24,25]. Each node, or neuron, in a layer uses a non-linear activation function, enabling the network to model complex relationships in the data. The architecture typically includes an input layer, one or more hidden layers, and an output layer, making it powerful for both classification and regression tasks, as is the case in this study. MLPs are trained using a supervised learning technique called backpropagation, which adjusts the weights of the connections to minimize the error in the predictions. In this model, two hidden layers were used. The non-linear activation function for the first hidden layer was the rectifier linear, and the soft rectifier linear was applied for the second layer. For more details on the model applied to genomic prediction, please refer to [26,27]. Python V3.10 was used with library TensorFlow and its high-level API, Keras [28,29]. Additionally, a grid search approach was used to find the optimum hyperparameters.

2.2.5. Convolutional Neural Networks

Convolutional Neural Networks (CNNs) are a class of deep learning models. CNNs are designed to recognize patterns in data. They consist of multiple layers, including convolutional, pooling, and fully connected layers, to learn patterns [30]. In this model, we used one input layer (genotypic matrix), one convolutional layer, and one pooling layer. Similarly to MLP, the hyperparameters used were determined through a grid search. Additionally, the Tanh (hyperbolic tangent) activation function was used for the convolutional layer, and the soft rectifier linear unit function was then used in the connected layers.

Similarly to MLP, Python V3.10 was used with library TensorFlow and its high-level API, Keras [28,29], and a grid search approach was used to find the optimum hyperparameters.

2.2.6. Random Forest

Random Forest (RF) is a machine learning tool that constructs multiple decision trees during training [31]. Each tree is built by selecting random samples of the data and making splits based on input features that minimize the mean squares error for regression [32]. The RF model in this study was implemented using Scikit-learn version 1.4 in Python V3.10 [33] and the RandomForestRegressor option, since the growth traits were continuous variables. The RF regression model is as follows:

y=1Mm=1Mtm(φmy:X), (4)

where y represents the response from the Random Forest regression model. Each tm(φmy:X) denotes an individual regression tree, and M is the total number of trees within the forest. The prediction process involves traversing each tree with the predictor variables and using the estimated value found at each tree’s terminal node as its prediction. To determine the final prediction for the validation data, the predictions from all trees in the Random Forest were averaged. A grid search was utilized to find the optimal hyperparameter and the maximum depth allowable for the trees. Additionally, a 5-fold cross-validation was conducted for hyperparameter tuning.

2.2.7. Support Vector Machine

Support Vector Machine (SVM) is another machine learning model [34,35]. The objective of the model is to achieve a low prediction error on validation data, given the training data. It can handle both linear and non-linear data, allowing it to manage complex relationships. The SVM model in this study is as follows:

y=b0+g(x)Tb+e, (5)

where b is the vector of regression weights, g(x)T is the mapping of the genotypes, b0 is the vector of bias, and e is the vector of random errors.

In this model, the following cost function was minimized:

minb0,b12|b|2+Ci=1nV(yigxTbb0),  (6)

in which

Vεr=0,ifr<εrε, otherwise

where Vε r is the ε insensitive loss and C is the cost parameter that influences how strictly the model should fit the training data. A higher value of C enforces the classification rules, minimizing the error in the training but risking overfitting and making the model more sensitive to noise.

Since the optimal selection of C is crucial in the prediction accuracy, a grid search method used for the optimum hyperparameters using a 5-fold cross validation. For more details on this model, we followed a similar approach to that in the work of Long et al. [36]. The SVM model was implemented using Scikit-learn version 1.4 in Python V3.10 [33].

2.2.8. Accuracy

The accuracy of genomic prediction was evaluated using a fivefold cross-validation (5-fold CV) approach, in which 2/3 of the data were randomly split into five groups. For each CV, four of the five groups were defined as the reference population, and the last group was treated as the validation group. First, we calculated the Pearson correlation between the corrected phenotypes and the predicted values. Additionally, prediction unbiasedness was evaluated through the regression of corrected phenotypes on predicted values in the validation dataset. The 5-fold cross-validation scheme was performed 10 times, and the overall prediction accuracy and unbiasedness were derived from the averages of these 10 repetitions.

Furthermore, the mean square error (MSE) was also computed, which offers an assessment of both prediction accuracy and bias; a lower MSE indicates a more accurate model.

3. Results and Discussion

The heritability estimates derived from the traditional BLUP model are summarized in Table 1. The heritabilities calculated were 0.38 for birth weight, 0.34 for weaning weight, and 0.28 for yearling weight. Our estimates for birth weight are slightly lower than those reported in [37], where a heritability of 0.41 was found in Angus cattle. However, for weaning weight, our estimate is higher; the authors of [37] estimated a heritability of 0.20, suggesting possible differences in data size and structure, as well as the environmental factors influencing this trait across populations. In terms of yearling weight, our results closely match the heritability of 0.25 reported in [38] in polled Hereford cattle. Such differences in heritability estimates might arise from variations in population structure, sample size, or even methodological approaches across various studies.

Table 2 shows the optimum hyperparameters detected using a grid search for the machine learning models; Table 3 outlines the prediction accuracies for each trait using different models. For birth weight, prediction accuracies ranged from 0.21 to 0.44, with the highest accuracy achieved by the GBLUP model, and the lowest by the Bayes B model. Among the machine learning models, CNNs performed the best, resulting in an accuracy of 0.43. This aligns with prior research indicating the robustness of GBLUP in genomic predictions. For instance, Srivastava et al. [39] demonstrated GBLUP’s superior predictive power over machine learning techniques such as SVM, RF, and extreme gradient boosting in predicting traits like backfat thickness and eye muscle area in Hanwoo cattle. Additionally, two studies [26,40] comparing machine learning models with traditional genomic prediction models showed the superiority of GBLIUP over CNN and MLP.

Table 2.

Optimum hyperparameters for SVM and RF models through grid search approach.

Model Optimum Hyperparameters 1
SVM kernel = ‘rbf’; C = 10; gamma = 0.01
RF n_estimators = 300; max_depth = none; optimal number of features = 1/3 of total features
MLP Optimization algorithm = stochastic gradient descent; epochs = 50; learning rate = 0.01; neurons for the first hidden layer = 96; neurons for the second hidden layer = 64
CNN Optimization algorithm = stochastic gradient descent; epochs = 50; learning rate = 0.01; neurons for the first hidden layer = 32; neurons for the second hidden layer = 16

1 The optimal hyperparameters detected through grid search.

Table 3.

Prediction accuracies of different models for all three growth traits.

Model * BW WW YW
GBLUP 0.447 ± 0.023 0.373 ± 0.019 0.321 ± 0.022
Bayes A 0.426 ± 0.018 0.375 ± 0.012 0.303 ± 0.013
Bayes B 0.21 ± 0.010 0.26 ± 0.022 0.22 ± 0.015
RF_default 0.350 ± 0.025 0.405 ± 0.003 0.261 ± 0.001
SVM_default 0.334 ± 0.014 0.360 ± 0.021 0.280 ± 0.027
RF_optimized 0.421 ± 0.039 0.451 ± 0.011 0.306 ± 0.028
SVM_optimized 0.406 ± 0.034 0.380 ± 0.018 0.311 ± 0.004
MLP 0.408 ± 0.021 0.402 ± 0.027 0.304 ± 0.014
CNN 0.433 ± 0.034 0.419 ± 0.024 0.303 ± 0.011

* Prediction accuracy of the model is calculated as Pearson’s correlation between corrected phenotypes and direct genomic values in the validation dataset.

In the case of weaning weight, prediction accuracies varied between 0.26 and 0.45. Notably, the optimized RF model showed the highest accuracy, outperforming all other models. The optimal RF model featured 300 decision trees, as delineated in Table 2, which exceeds the default 100 trees set by Scikit-learn. Although increasing the number of estimators generally enhances model accuracy, as noted by L Breiman [31], this also incurs greater computational demands. Similar trends were observed for yearling weight, where GBLUP again resulted in the highest prediction accuracy.

Machine learning models demonstrated a noticeable increase in prediction accuracy following hyperparameter tuning (Table 2), highlighting the significance of employing methods like grid search to identify ideal hyperparameters for optimal model performance (Table 3). For the deep learning models, the CNNs performed better than MLP for all growth traits. This is consistent with the literature, in Holstein cattle, where the genomic prediction accuracy for sire conception rate was 12% higher for CNNs compared to MLP [26].

Generally, machine learning models achieved prediction accuracies that are competitive with traditional models, except in the case of weaning weight, where the RF model excelled. Such findings are corroborated by the authors of [41] in their evaluation of machine learning models for genomic prediction in pig datasets, revealing a comparable performance across different methodologies. However, divergences exist, as illustrated by the authors of [13], who reported an improved prediction accuracy for reproductive traits with the use of machine learning models. Model performance was also assessed using MSE as a goodness-of-fit metric (Table 4). Typically, a lower MSE indicates a better model fit. In our study, the RF and SVM models resulted in a higher MSE for both birth and yearling weights, while GBLUP consistently presented the lowest MSE. Interestingly, for weaning weight, the RF model achieved the lowest MSE. These results align with the study in [39],which showed GBLUP to have the lowest MSE in genomic predictions. Meanwhile, the authors of [13] found that both SVM and RF offered the lowest MSE for predicting reproductive traits in pigs.

Table 4.

Mean square error and regression slope coefficients of different models for all growth traits.

Model BW WW YW
β1 1 MSE 2 β1 1 MSE 2 β MSE 2
GBLUP 0.971 ± 0.122 0.821 0.991 ± 0.107 2.892 0.954 ± 0.138 10.26
Bayes A 0.917 ± 0.130 0.843 1.101 ± 0.125 2.936 0.923 ± 0.137 11.27
Bayes B 0.902 ± 0.109 0.925 0.927 ± 0.117 3.034 0.906 ± 0.122 12.98
RF_default 1.055 ± 0.115 0.911 1.175 ± 0.129 3.071 1.209 ± 0.104 13.12
SVM_default 1.080 ± 0.136 0.906 1.202 ± 0.122 3.280 1.192 ± 0.117 12.15
RF_optimized 1.112 ± 0.127 0.860 1.209 ± 0.130 2.331 1.291 ± 0.102 11.18
SVM_optimized 1.204 ± 0.132 0.872 1.226 ± 0.142 2.714 1.146 ± 0.096 10.34
MLP 1.115 ± 0.132 0.901 1.188 ± 0.119 3.516 1.123 ± 0.116 12.09
CNN 1.102 ± 0.104 0.856 1.031 ± 0.107 3.104 1.101 ± 0.120 11.27

1 Regression slope of corrected phenotypes on predicted values for animals in the validation dataset to measure the inflation of genomic prediction. 2 Mean squared error of predicted genomic values and corrected phenotypes in the validation dataset as a measure of the fit of the model.

Finally, assessing the inflation or deflation of predicted values through a regression of corrected phenotypes on predicted values within the validation dataset (Table 4), GBLUP emerged as the best model for birth weight, weaning weight, and yearling weight.

4. Conclusions

In this study, machine learning models proved to be beneficial in improving prediction accuracy for weaning weight. However, this improvement did not extend to other traits, and the GBLUP model performed better. This limitation is potentially associated with the genetic architecture and the complexity of the trait. Therefore, future efforts are warranted to refine machine learning models for genomic prediction.

Acknowledgments

Special thanks to Lindsey Cook from the USDA Agricultural Research Service, Fort Keogh Livestock and Range Research Laboratory, Miles City, MT for collecting the pheno-typic and genomic data of this population.

Author Contributions

Conceptualization, E.H.H., methodology E.H.H., data analysis, E.H.H.; data curation, E.H.H.; writing—original draft preparation E.H.H. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Animal Care and Use Committee approval was not obtained for this study because data were from an existing database.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting the findings of this study are available upon request from the author El Hamidi Hay: elhamidi.hay@usda.gov and with permission from the USDA Agricultural Research Service.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.De Roos A., Schrooten C., Veerkamp R., Van Arendonk J. Effects of genomic selection on genetic improvement, inbreeding, and merit of young versus proven bulls. J. Dairy Sci. 2011;94:1559–1567. doi: 10.3168/jds.2010-3354. [DOI] [PubMed] [Google Scholar]
  • 2.Lourenco D., Legarra A., Tsuruta S., Masuda Y., Aguilar I., Misztal I. Single-step genomic evaluations from theory to practice: Using SNP chips and sequence data in BLUPF90. Genes. 2020;11:790. doi: 10.3390/genes11070790. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Toosi A., Fernando R., Dekkers J. Genomic selection in admixed and crossbred populations. J. Anim. Sci. 2010;88:32–46. doi: 10.2527/jas.2009-1975. [DOI] [PubMed] [Google Scholar]
  • 4.Daetwyler H., Kemper K., Van Der Werf J., Hayes B.J. Components of the accuracy of genomic prediction in a multi-breed sheep population. J. Anim. Sci. 2012;90:3375–3384. doi: 10.2527/jas.2011-4557. [DOI] [PubMed] [Google Scholar]
  • 5.Brøndum R.F., Rius-Vilarrasa E., Strandén I., Su G., Guldbrandtsen B., Fikse W., Lund M.S. Reliabilities of genomic prediction using combined reference data of the Nordic Red dairy cattle populations. J. Dairy Sci. 2011;94:4700–4707. doi: 10.3168/jds.2010-3765. [DOI] [PubMed] [Google Scholar]
  • 6.Aguilar I., Misztal I., Johnson D., Legarra A., Tsuruta S., Lawlor T. Hot topic: A unified approach to utilize phenotypic, full pedigree, and genomic information for genetic evaluation of Holstein final score. J. Dairy Sci. 2010;93:743–752. doi: 10.3168/jds.2009-2730. [DOI] [PubMed] [Google Scholar]
  • 7.Habier D., Fernando R.L., Kizilkaya K., Garrick D.J. Extension of the Bayesian alphabet for genomic selection. BMC Bioinform. 2011;12:186. doi: 10.1186/1471-2105-12-186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Meuwissen T.H., Hayes B.J., Goddard M. Prediction of total genetic value using genome-wide dense marker maps. Genetics. 2001;157:1819–1829. doi: 10.1093/genetics/157.4.1819. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Gianola D., de Los Campos G., Hill W.G., Manfredi E., Fernando R. Additive genetic variability and the Bayesian alphabet. Genetics. 2009;183:347–363. doi: 10.1534/genetics.109.103952. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Montesinos-López O.A., Martín-Vallejo J., Crossa J., Gianola D., Hernández-Suárez C.M., Montesinos-López A., Juliana P., Singh R. A benchmarking between deep learning, support vector machine and Bayesian threshold best linear unbiased prediction for predicting ordinal traits in plant breeding. G3 Genes Genomes Genet. 2019;9:601–618. doi: 10.1534/g3.118.200998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Piles M., Bergsma R., Gianola D., Gilbert H., Tusell L. Feature selection stability and accuracy of prediction models for genomic prediction of residual feed intake in pigs using machine learning. Front. Genet. 2021;12:611506. doi: 10.3389/fgene.2021.611506. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Azodi C.B., Bolger E., McCarren A., Roantree M., de Los Campos G., Shiu S.-H. Benchmarking parametric and machine learning models for genomic prediction of complex traits. G3 Genes Genomes Genet. 2019;9:3691–3702. doi: 10.1534/g3.119.400498. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Wang X., Shi S., Wang G., Luo W., Wei X., Qiu A., Luo F., Ding X. Using machine learning to improve the accuracy of genomic prediction of reproduction traits in pigs. J. Anim. Sci. Biotechnol. 2022;13:60. doi: 10.1186/s40104-022-00708-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Montesinos-López O.A., Montesinos-López A., Pérez-Rodríguez P., Barrón-López J.A., Martini J.W., Fajardo-Flores S.B., Gaytan-Lugo L.S., Santana-Mancilla P.C., Crossa J. A review of deep learning applications for genomic selection. BMC Genom. 2021;22:19. doi: 10.1186/s12864-020-07319-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Alves A.A.C., Espigolan R., Bresolin T., Costa R.M.d., Fernandes Júnior G.A., Ventura R.V., Carvalheiro R., Albuquerque L.G.d. Genome-enabled prediction of reproductive traits in Nellore cattle using parametric models and machine learning methods. Anim. Genet. 2021;52:32–46. doi: 10.1111/age.13021. [DOI] [PubMed] [Google Scholar]
  • 16.Newman S., MacNeil M., Reynolds W., Knapp B., Urick J. Fixed effects in the formation of a composite line of beef cattle: I. Experimental design and reproductive performance. J. Anim. Sci. 1993;71:2026–2032. doi: 10.2527/1993.7182026x. [DOI] [PubMed] [Google Scholar]
  • 17.Roberts A., Funston R., Grings E., Petersen M. TRIENNIAL REPRODUCTION SYMPOSIUM: Beef heifer development and lifetime productivity in rangeland-based production systems. J. Anim. Sci. 2016;94:2705–2715. doi: 10.2527/jas.2016-0435. [DOI] [PubMed] [Google Scholar]
  • 18.Sargolzaei M., Chesnais J., Schenkel F. FImpute-An efficient imputation algorithm for dairy cattle populations. J. Dairy Sci. 2011;94:421. [Google Scholar]
  • 19.Misztal I., Tsuruta S., Strabel T., Auvray B., Druet T., Lee D. BLUPF90 and related programs (BGF90); Proceedings of the 7th World Congress on Genetics Applied to Livestock Production; Montpelier, France. 19–23 August 2002; p. 743. [Google Scholar]
  • 20.Montesinos López O.A., Montesinos López A., Crossa J. Multivariate Statistical Machine Learning Methods for Genomic Prediction. Springer Nature; Berlin, Germany: 2022. [PubMed] [Google Scholar]
  • 21.Geweke J. Evaluating the Accuracy of Sampling-Based Approaches to the Calculation of Posterior Moments. Federal Reserve Bank of Minneapolis; Minneapolis, MN, USA: 1991. [Google Scholar]
  • 22.Heidelberger P., Welch P.D. Simulation run length control in the presence of an initial transient. Oper. Res. 1983;31:1109–1144. doi: 10.1287/opre.31.6.1109. [DOI] [Google Scholar]
  • 23.Delashmit W.H., Manry M.T. Recent developments in multilayer perceptron neural networks; Proceedings of the Seventh Annual Memphis Area Engineering and Science Conference, MAESC; Memphis, TN, USA. 11 May 2005; 2005. [(accessed on 5 March 2024)]. p. 33. Available online: https://www.semanticscholar.org/paper/Recent-Developments-in-Multilayer-Perceptron-Neural-Delashmit-Missiles/8657cb338897d912bc417fe3cee7b3ca43a83609. [Google Scholar]
  • 24.Popescu M.-C., Balas V.E., Perescu-Popescu L., Mastorakis N. Multilayer perceptron and neural networks. WSEAS Trans. Circuits Syst. 2009;8:579–588. [Google Scholar]
  • 25.Goodfellow I. Deep Learning. MIT Press; Cambridge, MA, USA: 2016. [Google Scholar]
  • 26.Abdollahi-Arpanahi R., Gianola D., Peñagaricano F. Deep learning versus parametric and ensemble methods for genomic prediction of complex phenotypes. Genet. Sel. Evol. 2020;52:12. doi: 10.1186/s12711-020-00531-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Pérez-Enciso M., Zingaretti L. A guide for using deep learning for complex trait genomic prediction. Genes. 2019;10:553. doi: 10.3390/genes10070553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Abadi M., Barham P., Chen J., Chen Z., Davis A., Dean J., Devin M., Ghemawat S., Irving G., Isard M. {TensorFlow}: A system for {Large-Scale} machine learning; Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16); Savannah, GA, USA. 2–4 November 2016; pp. 265–283. [Google Scholar]
  • 29.Chollet F. Keras: Deep Learning Library for Theano and Tensorflow. 2015. [(accessed on 5 March 2024)]. Available online: https://keras.io.
  • 30.LeCun Y., Bengio Y. Convolutional networks for images, speech, and time series. Handb. Brain Theory Neural Netw. 1995;3361:1995. [Google Scholar]
  • 31.Breiman L. Random forests. Mach. Learn. 2001;45:5–32. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
  • 32.González-Camacho J.M., Ornella L., Pérez-Rodríguez P., Gianola D., Dreisigacker S., Crossa J. Applications of machine learning methods to genomic selection in breeding wheat for rust resistance. Plant Genome. 2018;11:170104. doi: 10.3835/plantgenome2017.11.0104. [DOI] [PubMed] [Google Scholar]
  • 33.Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Prettenhofer P., Weiss R., Dubourg V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011;12:2825–2830. [Google Scholar]
  • 34.Vapnik V. The Nature of Statistical Learning Theory. Springer Science & Business Media; Berlin/Heidelberg, Germany: 2013. [Google Scholar]
  • 35.Burges C.J. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998;2:121–167. doi: 10.1023/A:1009715923555. [DOI] [Google Scholar]
  • 36.Long N., Gianola D., Rosa G.J., Weigel K.A. Application of support vector regression to genome-assisted prediction of quantitative traits. Theor. Appl. Genet. 2011;123:1065–1074. doi: 10.1007/s00122-011-1648-y. [DOI] [PubMed] [Google Scholar]
  • 37.Lourenco D., Tsuruta S., Fragomeni B., Masuda Y., Aguilar I., Legarra A., Bertrand J., Amen T., Wang L., Moser D. Genetic evaluation using single-step genomic best linear unbiased predictor in American Angus. J. Anim. Sci. 2015;93:2653–2662. doi: 10.2527/jas.2014-8836. [DOI] [PubMed] [Google Scholar]
  • 38.Glaze J., Schalles R. Heritabilities and genetic correlations for birth weight, weaning weight, and yearling weight in polled Hereford cattle (1994) Kans. Agric. Exp. Stn. Res. Rep. 1994;1:119–120. doi: 10.4148/2378-5977.2079. [DOI] [Google Scholar]
  • 39.Srivastava S., Lopez B.I., Kumar H., Jang M., Chai H.-H., Park W., Park J.-E., Lim D. Prediction of Hanwoo cattle phenotypes from genotypes using machine learning methods. Animals. 2021;11:2066. doi: 10.3390/ani11072066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Pedrosa V.B., Chen S.-Y., Gloria L.S., Doucette J.S., Boerman J.P., Rosa G.J., Brito L.F. Machine learning methods for genomic prediction of cow behavioral traits measured by automatic milking systems in North American Holstein cattle. J. Dairy Sci. 2024;107:4758–4771. doi: 10.3168/jds.2023-24082. [DOI] [PubMed] [Google Scholar]
  • 41.Zhao W., Lai X., Liu D., Zhang Z., Ma P., Wang Q., Zhang Z., Pan Y. Applications of support vector machine in genomic prediction in pig and maize populations. Front. Genet. 2020;11:598318. doi: 10.3389/fgene.2020.598318. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data supporting the findings of this study are available upon request from the author El Hamidi Hay: elhamidi.hay@usda.gov and with permission from the USDA Agricultural Research Service.


Articles from Animals : an Open Access Journal from MDPI are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES