Next Article in Journal
The Oral Microbiome in Queensland Free-Ranging Koalas (Phascolarctos cinereus) and Its Association with Age and Periodontal Disease
Previous Article in Journal
Dietary Synbiotic Attenuated the Intestinal Inflammation in Weaned Piglets Challenged with Escherichia coli Lipopolysaccharide
Previous Article in Special Issue
Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Approaches for the Prediction of Displaced Abomasum in Dairy Cows Using a Highly Imbalanced Dataset

by
Zeinab Asgari
1,
Ali Sadeghi-Sefidmazgi
2,
Abbas Pakdel
1 and
Saleh Shahinfar
3,*
1
Department of Animal Sciences, College of Agriculture, Isfahan University of Technology, Isfahan 84156-83111, Iran
2
Department of Animal Science, University of Tehran, Karaj 3158711167-4111, Iran
3
Department of Health and Aged Care, Canberra, ACT 2606, Australia
*
Author to whom correspondence should be addressed.
Animals 2025, 15(13), 1833; https://doi.org/10.3390/ani15131833
Submission received: 14 April 2025 / Revised: 1 June 2025 / Accepted: 4 June 2025 / Published: 20 June 2025

Simple Summary

Displaced abomasum (DA) is one of the costliest health problems as well as a welfare concern in dairy cows. Therefore, the predictive potential of DA-susceptible cases is of great importance to reduce economic losses. Hence, this study aimed for early prediction of DA by using five machine learning algorithms, namely Logistic Regression (LR), Naïve Bayes (NB), Decision Tree, Random Forest (RF) and Gradient Boosting Machines (GBM), to predict cases of DA. Model performance metrics indicated that among the algorithms considered in this study, RF and GBM were significantly better in terms of F2 (0.32) and TPR (0.75 and 0.70, respectively). Therefore, given the highly unbalanced data and that DA is a complex feature, machine learning (ML) methods were shown to be promising for predicting cases susceptible to DA at the herd level. This prediction tool can aid dairy farmers in making preventative management decisions by identifying cows susceptible to DA.

Abstract

Displaced abomasum (DA) is a digestive disorder that causes severe economic losses through the reduction in milk yield and early culling of cows. The predictive potential of DA-susceptible cases is of great importance to reduce economic losses. This study aimed for early prediction of DA. However, identifying cows at risk of DA can be difficult because DA is a complex trait and its incidence is low. For this purpose, in this study, the ability of five machine learning algorithms, namely Logistic Regression (LR), Naïve Bayes (NB), Decision Tree, Random Forest (RF) and Gradient Boosting Machines (GBM), to predict cases of DA was investigated. For these predictions, 20 herd–cow-specific features and sire genetic information from 7 Holstein dairy herds that calved between 2010 and 2020 were available. Model performance metrics indicated that GBM and RF algorithms outperformed the others in predicting DA with F2 measures of 0.32. The true positive rate in the RF was the highest compared to other methods at 0.75, followed by GBM at 0.70. Given the highly imbalanced data, this study showed the potential in forecasting cases susceptible to DA. This prediction tool can aid dairy farmers in making preventative management decisions by identifying cows susceptible to DA.

1. Introduction

Displaced abomasum is defined as abomasum mislocated from its normal position to either the left or right side [1] and is considered the most common surgical gastrointestinal disorder in dairy cows [2]. Approximately 30–50% of dairy cows suffer metabolic and infectious diseases after calving [3] and the annual incidence of DA has been reported to range between 0.05% and 6% [4]. DA is an important digestive condition that leads to great concern in terms of animal economics [5] and has a detrimental impact on animal welfare due to disease occurrence, health outcomes and the experience of pain [6], which in turn can influence the profitability in the herd [7]. The direct economic impact of DA includes the costs of treatment, early culling, and the decrease in milk production [8], an increase in the interval of calving to first service [9,10]. Although ketosis is a well-recognized risk factor for left displacement abomasum (LDA), in some studies, the occurrence of some disorders such as ketosis following DA has been reported [11,12,13]. In this regard, it should be noted that sometimes DA may have occurred but had not yet been diagnosed. At this time, the abomasum is periodically filled with gas and fluid, and then emptied. Therefore, in cows with uncomplicated displacements, where the cow may go through periods of apparent well-being, ketosis can develop due to the reduction in feed consumption [14]. Direct and indirect cost per case of DA in the United States have been reported to be more than 700 USD [8]. Hence, the ability to predict cases susceptible to DA could allow for earlier intervention strategies.
DA multifactorial disease occurs mostly in the early postpartum period [10] and depends on the coordinated action of a set of predisposing factors [15,16]. According to previous reports, more than 50% of DA cases occur in the first postpartum period [9,17]. Some earlier studies have reported factors that are associated with the likelihood of DA occurrence [4,13,17,18]. The known non-genetic factors contributing to the incidence of DA include herd, year and season of calving, parity, reproductive and other metabolic disorders [17,18,19,20]. Also, DA is largely affected by nutrition management and concentration of various metabolites [21,22]. In addition to cow- and herd-level factors, numerous studies suggest that DA is a heritable trait [23,24], with an estimated heritability between 0.12 and 0.32 [23]. This means that up to 30% of the phenotypic variation in the DA at the population level can be explained by additive genetic variations. Also, a positive genetic correlation (0.45 ± 0.16) is reported for DA with other metabolic disorders such as ketosis [25].
To monitor the metabolic status of cows, measuring plasma metabolites and metabolic hormones has been suggested [3], but these metabolites and hormones are difficult to measure in the farm [26]. In recent years, animal studies have used machine learning methods to predict traits such as assessing disease impact in dairy cattle [7], assessing the impact of technology integration on cow health [27], the incidence of lameness [28], the incidence of neosporosis in cattle [29], ketosis [30], the incidence of retained placenta [31], and the metabolic status of dairy cows [26]. Machine learning algorithms can deal with inter-correlations [32] and analyze large sets of data, regardless of complexity, quickly and accurately [33]. Machine learning methods have previously been used to predict complex traits [28]; hence, according to the fact that DA is a multifactor trait, the present study was conducted with the aim to investigate machine learning method capabilities in predicting whether a cow is at high risk of DA during the lactation period.

2. Material and Methods

2.1. Data

The data used in this study consisted of health, reproduction, and production records extracted from backup files of Modiran (version 8.21.02.18), a Livestock Management Software, from 7 Holstein dairy farms located at 3 provinces of Iran (Esfahan, Chahar Mahaal en Bakhtiari, and Khorasan). Herds were selected based on their use of an accurate recording system (Livestock Management Software, Modiran 8.21.02.18), availability of data for DA, and other features required for analysis. All herds had comparably similar management routines in terms of veterinary services, vaccination and use of artificial insemination. The feed was as a balanced total mixed ration (similar ratio of forage to concentrate) and the free stall housing system. All herds on average contained ≥500 milking cows that were milked three times per day. Data extracted from the farm databases for each individual cow were: herd and cow identification, sire, lactation period, calving date, management factors, information on health and metabolic disorders and records of milk and milk compositions. Data were edited using R (version 4.1.2; [34]). Records with parity > 6, pregnancy period outside 240 to 298 days, age at first calving outside 578 to 1153 days and dry period length > 160 days were excluded from the analysis. Also records exceeding three standard deviations from the mean of milk yield, fat and protein and calf weight at birth were excluded. The number of excluded records and reasons for excluding them are presented in Table 1. After data cleaning from 93,273 available records, a total of 85,031 records were available following editing for statistical analysis; this included 56,368 multiparous cows and 28,663 primiparous cows.
Cases of DA, which included “all cases of DA” (i.e., left cases and right cases, 90% and 10%, respectively), are usually recorded as a binary trait (1 if a cow was affected by DA and 0 otherwise). Milk production traits including the monthly milk yield, fat and protein percentage (i.e., recorded on one day per month) before occurrence of DA for affected cows were considered (2841 cows). For cows with DA that had no prior milk record, the last record of the previous lactation was used (1394 cows). For unaffected cows, the first lactation record after calving was included in the dataset. Cows with DA and without a monthly record of milk were excluded from the analysis (582 cows). Other features used for predicting DA include parity (1, 2, …, 6), herd (h1, h2, …, h7), pregnancy length, calving season (spring, summer, fall and winter), dystocia and twinning (0 was normal partus, and 1 was abnormal partus), mastitis, ketosis, retained placenta and metritis during the current lactation and before the occurrence of DA, (0 for healthy cows and 1 for affected cows). It is necessary to explain that a positive test for the detection of the ketone bodies in blood (measurement of beta-hydroxybutyric acid (BHBA) in serum or plasma in the lab or using a ketone meter) combined with reduced feed intake was used to diagnose a cow with ketosis. The BHBA threshold used to diagnose subclinical ketosis was 1.2–2.9 mmol/L and for clinical ketosis was ≥3 mmol/L.
In addition to the herd–cow factors, sire EBVs (sire of the cow) obtained from the Alta Genetics website (https://bullsearch.altagenetics.com/us/BS/List, accessed on 1 February 2023) were also included in the dataset. This information includes the daughter pregnancy rate (PTA-DPR), productive life (PTA-PL), sire calving ease (PTA-SCE), sire stillbirth (PTA-SSB), daughter calving ease (PTA-DCE), daughter stillbirth (PTA-DSB), PTA of DA (PTA-DA), PTA of ketosis, PTA of somatic cell score (PTA-SCS), PTA of milk protein percentage and PTA of milk fat percentage. Therefore, they were considered as candidate predictor features for ML algorithms. A general description of predictor feature set available for predicting DA is provided in Table 2. Also, the frequency of each level of nominal features is shown the Table 3.

2.2. Machine Learning Methodology

In this study, five machine learning algorithms were used to predict the likelihood of DA incidence: Logistic Regression, Naïve Bayes, Decision Tree, Random Forest, and Gradient Boosting Machines. These models are described in detail in Shahinfar et al. [28,35].
After quality control and cleaning, the dataset contained 85,031 records (each record is a cow in a parity) of 47,686 Holstein cows calving between 2010 and 2020 and was divided into two sets with an 80:20 ratio as the training set and the hold-out testing set, respectively. Feature selection was performed to select the most informative features [36] only on the training dataset. After balancing the training set, 80% of the total training set was selected as the training set i, i∈{1:10}. Furthermore, 70% of the ith training set was randomly selected to be used for hyperparameter tuning via a nested grid search while internally tested on the remaining 30%. The best-performing hyperparameters in this step were used to train each ML algorithm within their corresponding ith training sets and were tested on the hold-out testing set at the end. The training algorithm is explained in detail in the next section.

2.2.1. Feature Selection

Feature selection was performed to discard uninformative features from the dataset [36]. Feature selection is extremely important in ML algorithms because it reduces computational costs, avoids overfitting [30], leads to the reduction in the distractive effect of noisy features [35], and by choosing the optimal subset of predictors, it promotes the efficiency of the model [37]. In the current study, Random Forest was performed to select the most effective set of features only on the training set. For this purpose, the mean reduction in the Gini index and the mean decrease in accuracy were used. Random Forest was performed with the number of trees to grow (500:1000) and the number of candidate variables randomly sampled in each split (4:5) in each iteration. The top 15 features for the mean decrease accuracy and the mean decrease in Gini index were selected and saved. Finally, the training process was conducted with the union of two feature sets selected. According to the results of trait importance and based on the intersection of both criteria considered in this study, milk yield, PTA_DA, dry period length, parity, milk fat yield, herd, AFC and PTA_DPR are the most important features in predicting DA (Figure 1). The selected features in the feature selection stage were highlighted in Table 2.

2.2.2. Hyperparameter Tuning

The prediction models were developed using 10 repetitions as described in the next section. In each iteration, 80% of the training set was randomly selected and hyperparameter tuning was performed via grid search. Tuned hyperparameters were the number of trees to grow (100:500) and the number of candidate variables randomly sampled in each split (4:10) for RF. The Laplace smoothing coefficient (1:10) and the smallest allowed node size (10:100) were tuned for Naïve Bayes and Decision Tree, respectively. For GBM, the variable interaction levels (1:3) and the learning rate of 0.01, the number of trees (1000:1500), the fraction of the training set observations for the next tree (0.80 and 1), and the minimum number of observations in the terminal nodes of a tree (20 and 25) were tuned.

2.2.3. Analysis

The average incidence rate of DA was approximately 5.0% among the herds under study, which indicates a highly class-imbalanced data (Figure 2). Classification algorithms that aim to minimize the error rate do not perform well when dealing with imbalanced data. One of the approaches to deal with this problem is the use of sampling techniques, including down-sampling and up-sampling [38]. In the present study, down-sampling techniques were used to produce a balanced training set. Machine learning algorithms were tuned, trained and tested based on the algorithm explained below. First, the raw dataset (after preprocessing and data wrangling) was divided into two sets. For this division, first the raw dataset in each herd divided into two classes, sick (majority class) and healthy (minority class), and then, each dataset was separately divided into two datasets with a ratio of 80:20. Twenty percent of the data from each class were randomly selected and merged together; this dataset was set aside as the holding test set. The remaining 80% of each class after merging was considered as the training data. Then in the second step, after feature selection by using the training set (described above), sampling techniques were used to generate 10 balanced training data. Therefore, to keep the total number of cows with DA constant, the aforementioned 80% training dataset split (to cows with DA and healthy), and down-sampling was only performed in the majority class within each herd to stratify the data within the herd structure and ensure that each herd contributes to the training set equally. Hence, ten random samples were taken from unaffected cows of the training set without replacing to match the total number of cows with DA. In the next step, 80% of each training set was selected at random (these ten samples were used for model training). Furthermore, 70% of each balanced training set was selected for training (For hyperparameter tuning) and 30% for testing (to evaluate performance during hyperparameter tuning) purposes of hyperparameter tuning through a nested grid search. Finally, the prediction models were developed using the best-performing hyperparameters on each of the 10 training sets, and was validated on the imbalanced holed-out testing set. Figure 3 represents an overview of the whole process. Model performance metrics (MPMs) were aggregated over the 10 iterations of training and testing. This approach was used to control the prediction variance and enhance model generalizability [35]. All analysis and models were implemented in R (version 4.1.2; [34]).
To evaluate the performance of the models, common MPMs were used. These MPMs include overall accuracy (ACC = (TP + TN)/(TP + FP + TN + FN)), balanced accuracy (BACC = (TPR + TNR)/2), the true positive rate (TPR = TP/(TP + FN)), the true negative rate (TNR = TN/(TN + FP)), precision (PPV = TP/(TP + FP)), the false negative rate (FNR = FN/(FN + TP)) and the false positive rate (FPR = FP/(FP + TN)), where TP is true positive, TN is true negative, FP is false positive, and FN is false negative. In data with highly imbalanced classes, the TPR is more appropriate for measuring the success of any classifier, because the TPR focuses on performance measurement in the positive or minority class, which provides more relevant and appropriate information [39]. The true positive rate (recall) measures the percentage of the positive examples that the model predicted correctly [40]. Because, the TPR only uses values of one column of the confusion matrix, it cancels the changes in the data class distribution, which is the ratio between the positive and negative samples [41]. Model performance metrics that are estimated based on both columns of the confusion matrix will be sensitive to the imbalanced data [42] and they will change as data distributions change [43].
A scalar metric alone might not be an appropriate parameter to measure the predictive performance of a model, especially for nonparametric models [44]. Hence, to measure and evaluate any classifier in this study, in addition to the TPR, the focus was also on F2-measure. Due to the characteristics of imbalanced data, the F-measure was chosen because its emphasis is on the TPR and PPV to calculate an integrated score (i.e., identification of minority class). F2-measure is calculated as the harmonic mean of precision and recall (TPR) and gives a weight to each one [45]:
F2-measure =     1 + β 2 ×   TP 1 + β 2 ×   TP   +   β 2 ×   FN   + FP  
where β, which indicates that recall is more important than precision, is a weight control parameter and gives more weight to recall. In the F2-measure, a β of 2 is selected to emphasize minority class learning by assigning more weight to the TPR in the performance of the algorithms [39].

3. Results

DA is a complex trait and many predisposing factors are involved in its occurrence. This disease usually occurs early in the postpartum period. In the current study, 63% of cases of DA occurred in the first month after calving. The incidence rate of DA is low and leads to a severe imbalance in the data. Hence, it is difficult to achieve accurate prediction. In the present study, down-sampling techniques were applied for the training dataset used to train DA case prediction models.
The performance of the different machine learning algorithms used to predict cases of DA using a highly imbalanced dataset (hold-out testing dataset) is shown in Table 4 and Figure 4. Also, the calibration, which indicates the degree to which the predicted risk matches the actual risk, was shown using calibration plots (Figure 5). The classification accuracy of predictive models ranged from 62% to 69%. In terms of ACC, LR and GBM outperformed other machine learning algorithms at 0.68 and 0.69, respectively. The greatest value of F2-measure was achieved with GBM and RF, whereas NB had the lowest (0.33 vs. 0.27). In terms of the TPR, RF outperformed other machine learning algorithms at 0.75, whereas LR was the lowest. Considering the PPV, it was GBM that outperformed other methods at 0.10. PPVs were low for all algorithms and varied between 0.07 (NB) and 0.10 (GBM) across algorithms. The greatest value of AUPRC and AUROC was achieved with GBM (0.17) and RF (0.76), respectively.

4. Discussion

The results of the present study demonstrate that considering the PPV, it was GBM that outperformed other methods at 0.10. Implemented models’ low PPVs were expected, because the PPV is a function of the trait incidence rate, and a lower incidence leads to a lower PPV. When predicting a trait with a low incidence rate, a model with a good TPR and TNR can have a low PPV. Although the PPV is calculated directly from TP and FP, it is a function of the incidence rate of the trait under study [46]. In other studies, PPVs for mastitis and lameness are 0.01 and 0.10, respectively [47,48], which is lower than the value reported here. This represents a direct relationship between the PPV and the frequency of trait under investigation [49]. Although GBM and LR (0.69 and 0.68, respectively) performed well in terms of accuracy, accuracy is not a good measure to evaluate the performance of prediction models when dealing with imbalanced data, because it assigns more weight for the class with a high frequency than the rare classes [50]. For this reason, balanced accuracy was also estimated, which weighs the two classes equally and is believed to be more plausible than using accuracy alone in situations of using highly unbalanced data [51]. The difference between the two criteria was very small that indicates that a classifier performs equally well on either class [52]. Table 5 provides a sample of recent studies that attempted to predict health traits in dairy cattle using ML.
In terms of AUROC, RF and GBM also outperformed other machine learning algorithms at 0.76 and 0.75, respectively; AUROC is the appropriate measure of a classifier’s ability and reflects the power of positive prediction probability ranking. Of course, in low incidence rate cases, especially when the frequency of the minority class is lower than 5%, the Area Under the Precision–Recall Curve (AUPRC) is preferred because it has a better agreement with the PPV. In this respect, GBM and RF outperformed other machine learning algorithms by 0.17.
According to the characteristics of the imbalanced data, F2-measure and TPR were used as the main metrics to evaluate the predictive performance of models. In this regard, the predictive performance of the ML algorithms used in the present study, based on F2-measure, varied from 0.27 (NB) to 0.32 (RF and GBM). RF and GBM had the highest F2-measure at 0.32 followed by LR (0.30). Considering the TPR, it was RF that outperformed other methods (0.75). GBM had the highest TPR (0.70) after RF, whereas LR had the lowest performance (0.65). A high TPR, especially when analyzing data related to a disease, is very important for a classifier because in medical data, information is stored in the minority class data and the TPR also measures the performance of a classifier on the minority class [39].
These findings showed the ability of ML algorithms to predict cases of DA. Considering all metrics, the results show that both GBM and RF can have success in predicting DA cases. Other models considered in this study showed similar MPMs. The better performance of GBM and RF can be attributed to the ability of the ensemble approach to form a high-performance predictor based on the training set of collaborative predictors [57]. Previous studies have shown that correlation between predictors can be handled well by these models [58,59]. Random selection of a subset of features to generate any of those classifiers can eliminate the correlation between features. Of course, the method of construction and internal evaluation of both of these algorithms are different, although both are tree-based methods [60].
Although better performance metrics of RF and GBM were statistically significant, results indicate a limited improvement. So that, a large proportion of cows are still misclassified (i.e., low PPV). This will prove the fact that prediction of DA is extremely difficult and even the best algorithm in the current study still provides modest accuracy. The low performance of DA prediction models used in this study was expected because in addition to having a highly imbalanced dataset, information on the mechanism(s) of why DA occurs [61,62] and the coordinated action of many predisposing factors involved in its occurrence [15,16] are unknown to us. Among these factors, weather [20,21], environment and management factors such as nutrition and metabolism [63,64], body score condition [18], energy metabolism markers such as β-hydroxybutyric [3,65] and minerals such as calcium levels in serum [65] can be mentioned but were not available to us in the current study. Having a complete set of factors affecting the occurrence of disease can lead to a better performance of prediction models.
However, to the best of the authors’ knowledge, this is the first study that used predictive models, phenotypic data, and sires’ EBVs to identify cows prone to DA and proved the ability of machine learning models in predicting whether a cow is at high risk of DA during its lactation period. Therefore, it can be considered as a decision support tool to help farm managers in monitoring susceptible cows. Identifying cows prone to metabolic disorders such as DA can lead to better cow health management, and provides farmers with the opportunity to modify diet and farm management strategies to prevent disorders a priori [66]. Susceptible cows usually have a poor metabolic status [12]; a metabolic status can be improved by feeding diets with a higher energy content [67] or by reducing the dry period length. If a cow is prone to metabolic disorders (such as DA) but has a good body condition score, limiting energy supply can lead to reduced body fat storage and a reduced risk of metabolic disorders in the next lactation period [68]. So, utilizing the approach introduced in the current study can aid farmers as an alerting system in order to take preventative actions on cows susceptible to these abnormalities.
According to previous reports, displaced abomasum is a costly disease in all parities [69]. The cost per case of DA has been reported from 450 USD per case [70] to more than 700 USD [8]. As an example of how this model can work effectively in practice, suppose the incidence of DA in a herd of 1000 cows is 5.5%, so 55 cows per year will be affected with DA. Using the approach introduced in this study, it is possible to correctly identify 10% of cows at high risk of DA a priori (PPV 10%, Table 4), which is equivalent to six cows in this hypothetical scenario. As a result, assuming an effective preventive intervention practice is available on time, using this algorithm, 2700 to 4200 USD of future financial losses can be prevented, which translates to higher profits for farmers. This approach can be adapted for other dairy cattle health and welfare issues as well to enhance profitability.

5. Conclusions

In the present study, machine learning algorithms were used for initial DA prediction. Among the algorithms considered in this study, RF and GBM were significantly better in terms of F2-measure (0.32) and TPR (0.75 and 0.70, respectively). Other models considered in this study showed similar MPMs. Although, prediction of DA is quite challenging as DA is a complex trait and its low occurrence rate causes severe data imbalance. The primary strength of this study lies in demonstrating the effectiveness of our custom-designed under-sampling and cross-validation algorithm in predicting highly imbalanced outcomes. Suboptimal performance metrics could also be due to the fact that some important DA risk factors were not available to us as predictors, which is expected to be improved with a comprehensive set of DA risk factors as predictor features in the dataset and more advanced predictive models. It is worth mentioning that this is the first study to use predictive models, phenotypic data, and sires’ EBVs to identify cows prone to DA and proves the ability of machine learning models in predicting whether a cow is at high risk of DA during its lactation period.

Author Contributions

Conceptualization, S.S.; methodology, S.S.; software, S.S. and Z.A.; validation, S.S. and Z.A.; formal analysis, Z.A. and S.S.; investigation, Z.A.; resources, A.S.-S. and A.P.; data curation, Z.A.; writing—original draft preparation, Z.A.; writing—review and editing, S.S., A.S.-S. and A.P.; visualization, S.S. and Z.A.; supervision, S.S.; project administration, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Animal Care and Use Committee approval was not required for this study since the data used were obtained from an existing database.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this article are not readily available due to restrictions (commercial restrictions, ethical reasons and privacy). General information about the datasets presented in this study is included in the article. Further inquiries can be directed to the corresponding author (shahinfar@uwalumni.com).

Acknowledgments

The authors would like to acknowledge the herd managers for collaborating and generously providing the information needed.

Conflicts of Interest

The authors declare no competing interests.

References

  1. Caixeta, L.S.; Herman, J.A.; Johnson, G.W.; McArt, J.A. Herd-level monitoring and prevention of displaced abomasum in dairy cattle. Vet. Clin. N. Am. Food Anim. 2018, 34, 83–99. [Google Scholar] [CrossRef] [PubMed]
  2. González-Martín, J.V.; Revilla-Ruiz, Á.; Astiz, S. Abomasum Displacement in Bovine (Calves, Young Stock, and Cows). In Encyclopedia of Livestock Medicine for Large Animal and Poultry Production; Springer: Berlin/Heidelberg, Germany, 2025; pp. 1–9. [Google Scholar]
  3. LeBlanc, S. Monitoring metabolic health of dairy cattle in the transition period. J. Reprod. Dev. 2010, 56, 29–35. [Google Scholar] [CrossRef] [PubMed]
  4. Mueller, K. Diagnosis, treatment and control of left displaced abomasum in cattle. In Pract. 2011, 33, 470–481. [Google Scholar] [CrossRef]
  5. Basoglu, A.; Baspinar, N.; Tenori, L.; Licari, C.; Gulersoy, E. Nuclear magnetic resonance (NMR)-based metabolome profile evaluation in dairy cows with and without displaced abomasum. Vet. Q. 2020, 40, 1–15. [Google Scholar] [CrossRef]
  6. Temple, D.; Manteca, X. Animal welfare in extensive production systems is still an area of concern. Front. Sustain. Food Syst. 2020, 4, 545902. [Google Scholar] [CrossRef]
  7. Marzougui, A.; Mcconnel, C.S.; Adams-Progar, A.; Biggs, T.D.; Ficklin, S.P.; Sankaran, S. Machine learning-derived cumulative health measure for assessing disease impact in dairy cattle. Front. Anim. Sci. 2020, 6, 1532385. [Google Scholar] [CrossRef]
  8. McArt, J.; Nydam, D.; Overton, M. Hyperketonemia in early lactation dairy cattle: A deterministic estimate of component and total cost per case. J. Dairy Sci. 2015, 98, 2043–2054. [Google Scholar] [CrossRef]
  9. Asgari, Z.; Sadeghi-Sefidmazgi, A.; Pakdel, A.; Shahinfar, S. Effect of displaced abomasum on milk production, reproductive performance, and culling of Holstein dairy cattle in Iran. Can. J. Anim. Sci. 2024, 105, 1–9. [Google Scholar] [CrossRef]
  10. Raizman, E.; Santos, J. The effect of left displacement of abomasum corrected by toggle-pin suture on lactation, reproduction, and health of Holstein dairy cows. J. Dairy Sci. 2002, 85, 1157–1164. [Google Scholar] [CrossRef]
  11. Leblanc, S.; Duffield, T.; Leslie, K.; Bateman, K.; Keefe, G.P.; Walton, J.; Johnson, W. Defining and diagnosing postpartum clinical endometritis and its impact on reproductive performance in dairy cows. J. Dairy Sci. 2002, 85, 2223–2236. [Google Scholar] [CrossRef]
  12. Mcart, J.A.; Nydam, D.V.; Oetzel, G.R.; Overton, T.R.; Ospina, P.A. Elevated non-esterified fatty acids and β-hydroxybutyrate and their association with transition dairy cow performance. Vet. J. 2013, 198, 560–570. [Google Scholar] [CrossRef] [PubMed]
  13. Shaver, R.D. Prevention of displaced abomasum. In Proceedings of the Tri-State Dairy Nutrition Conference, Fort Wayne, IN, USA, 17–18 April 2001. [Google Scholar]
  14. Credille, B.C.; Fubini, S. Left Displacement of the Abomasum. In Comparative Veterinary Anatomy; Academic Press: New York, NY, USA, 2022; pp. 1063–1068. ISBN 9780323910156. [Google Scholar] [CrossRef]
  15. Martin, W. Left abomasal displacement: An epidemiological study. Can. Vet. J. 1972, 13, 61. [Google Scholar] [PubMed]
  16. Coppock, C. Displaced abomasum in dairy cattle: Etiological factors. J. Dairy Sci. 1974, 57, 926–933. [Google Scholar] [CrossRef] [PubMed]
  17. Asgari, Z.; Pakdel, A.; Sadeghi-Sefidmazgi, A. Risk factors and population attributable fractions for displaced abomasum in Iranian dairy cattle: A retrospective analysis of field data. Trop. Anim. Health Prod. 2024, 56, 283. [Google Scholar] [CrossRef]
  18. Behluli, B.; Musliu, A.; Youngs, C.; Sherifi, K.; Rexhepi, A. Risk factors for occurrence of displaced abomasum and their relation to nutritional management of Holstein dairy cattle. Vet. Arh. 2017, 87, 419–430. [Google Scholar] [CrossRef]
  19. Sen, I.; Wittek, T.; Guzelbektes, H. Metabolic indicators and risk factors of left displaced abomasum in dairy cattle. Eurasian J. Vet. Sci. 2015, 31, 63–69. [Google Scholar] [CrossRef]
  20. Constable, P.; Miller, G.; Hoffsis, G.; Hull, B.; Rings, D. Risk factors for abomasal volvulus and left abomasal displacement in cattle. Am. J. Vet. Res. 1992, 53, 1184–1192. [Google Scholar] [CrossRef]
  21. Cameron, R.; Dyk, P.; Herdt, T.; Kaneene, J.; Miller, R.; Bucholtz, H.; Liesman, J.; Vandehaar, M.; Emery, R. Dry cow diet, management, and energy balance as risk factors for displaced abomasum in high producing dairy herds. J. Dairy Sci. 1998, 81, 132–139. [Google Scholar] [CrossRef]
  22. Simões, J.; Teixeira, V.; Silva, S.; Gomes, A.; Ventura, A. Relationship between dietary particle size and the incidence of displaced abomasum on Holstein-Friesian dairy farms that feed diets high in maize silage. Livest. Sci. 2013, 157, 478–481. [Google Scholar] [CrossRef]
  23. Gaddis, K.P.; Cole, J.; Clay, J.; Maltecca, C. Genomic selection for producer-recorded health event data in US dairy cattle. J. Dairy Sci. 2014, 97, 3190–3199. [Google Scholar] [CrossRef]
  24. Zerbin, I.; Lehner, S.; Distl, O. Genetics of bovine abomasal displacement. Vet. J. 2015, 204, 17–22. [Google Scholar] [CrossRef] [PubMed]
  25. Zwald, N.; Weigel, K.; Chang, Y.; Welper, R.; Clay, J. Genetic selection for health traits using producer-recorded data. II. Genetic correlations, disease probabilities, and relationships with existing traits. J. Dairy Sci. 2004, 87, 4295–4302. [Google Scholar] [CrossRef] [PubMed]
  26. Xu, W.; van Knegsel, A.T.; Vervoort, J.J.; Bruckmaier, R.M.; van Hoeij, R.J.; Kemp, B.; Saccenti, E. Prediction of metabolic status of dairy cows in early lactation with on-farm cow data and machine learning algorithms. J. Dairy Sci. 2019, 102, 10186–10201. [Google Scholar] [CrossRef] [PubMed]
  27. Matzhold, C.; Schodl, K.; Klimek, P.; Steininger, F.; Egger-Danner, C. A key-feature-based clustering approach to assess the impact of technology integration on cow health in Austrian dairy farms. Front. Anim. Sci. 2024, 5, 1421299. [Google Scholar] [CrossRef]
  28. Shahinfar, S.; Khansefid, M.; Haile-Mariam, M.; Pryce, J. Machine learning approaches for the prediction of lameness in dairy cows. Animal 2021, 15, 100391. [Google Scholar] [CrossRef]
  29. Ballesteros-Ricaurte, J.A.; Fabregat, R.; Carrillo-Ramos, A.; Parra, C.; Moreno, A. Artificial neural networks to predict the presence of Neosporosis in cattle. MBE 2025, 22, 1140–1158. [Google Scholar] [CrossRef]
  30. Ushikubo, S.; Kubota, C.; Ohwada, H. The early detection of subclinical ketosis in dairy cows using machine learning methods. In Proceedings of the 9th International Conference on Machine Learning and Computing, Singapore, 24–26 February 2017. [Google Scholar]
  31. Hosseinabadi, M.R.; Mahdavi, A.H.; Mahnani, A.; Asgari, Z.; Shahinfar, S. Machine learning approaches for the prediction of retained placenta in dairy cows. Theriogenology 2025, 244, 117484. [Google Scholar] [CrossRef]
  32. Dormann, C.F.; Elith, J.; Bacher, S.; Buchmann, C.; Carl, G.; Carré, G.; Marquéz, J.R.G.; Gruber, B.; Lafourcade, B.; Leitão, P.J. Collinearity: A review of methods to deal with it and a simulation study evaluating their performance. Ecography 2013, 36, 27–46. [Google Scholar] [CrossRef]
  33. Patrício, D.I.; Rieder, R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef]
  34. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2021. [Google Scholar]
  35. Shahinfar, S.; Kelman, K.; Kahn, L. Prediction of Sheep Carcass Traits from Early-life Records Using Machine Learning. Comput. Electron. Agric. 2019, 156, 159–177. [Google Scholar] [CrossRef]
  36. Hempstalk, K.; McParland, S.; Berry, D.P. Machine learning algorithms for the prediction of conception success to a given insemination in lactating dairy cows. J. Dairy Sci. 2015, 98, 5262–5273. [Google Scholar] [CrossRef] [PubMed]
  37. Valletta, J.J.; Torney, C.; Kings, M.; Thornton, A.; Madden, J. Applications of machine learning in animal behaviour studies. Anim. Behav. 2017, 124, 203–220. [Google Scholar] [CrossRef]
  38. Chen, C.; Liaw, A.; Breiman, L. Using random forest to learn imbalanced data. Cal. UCB 2004, 110, 24. [Google Scholar]
  39. Devarriya, D.; Gulati, C.; Mansharamani, V.; Sakalle, A.; Bhardwaj, A. Unbalanced breast cancer data classification using novel fitness functions in genetic programming. Expert Syst. Appl. 2020, 140, 112866. [Google Scholar] [CrossRef]
  40. Johnson, J.M.; Khoshgoftaar, T.M. Survey on deep learning with class imbalance. J. Big Data 2019, 6, 27. [Google Scholar] [CrossRef]
  41. Tharwat, A. Classification assessment methods. Appl. Comput. Inform. 2020, 17, 168–192. [Google Scholar] [CrossRef]
  42. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  43. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  44. Hamel, L. Model Assessment with ROC Curves. In Encyclopedia of Data Warehousing and Mining, 2nd ed.; IGI Global: Hershey, PA, USA, 2009; pp. 1316–1323. [Google Scholar]
  45. Dascalu, A.; David, E. Skin cancer detection by deep learning and sound analysis algorithms: A prospective clinical study of an elementary dermoscope. eBioMedicine 2019, 43, 107–113. [Google Scholar] [CrossRef]
  46. Cox, C.R.; Moscardini, E.H.; Cohen, A.S.; Tucker, R.P. Machine learning for suicidology: A practical review of exploratory and hypothesis-driven approaches. Clin. Psychol. Rev. 2020, 82, 101940. [Google Scholar] [CrossRef]
  47. Miekley, B.; Traulsen, I.; Krieter, J. Detection of mastitis and lameness in dairy cows using wavelet analysis. Livest. Sci. 2012, 148, 227–236. [Google Scholar] [CrossRef]
  48. Steeneveld, W.; van der Gaag, L.C.; Ouweltjes, W.; Mollenhorst, H.; Hogeveen, H. Discriminating between true-positive and false-positive clinical mastitis alerts from automatic milking systems. J. Dairy Sci. 2010, 93, 2559–2568. [Google Scholar] [CrossRef] [PubMed]
  49. Post, C.; Rietz, C.; Büscher, W.; Müller, U. The importance of low daily risk for the prediction of treatment events of individual dairy cows with sensor systems. Sensors 2021, 21, 1389. [Google Scholar] [CrossRef] [PubMed]
  50. Bekkar, M.; Djemaa, H.K.; Alitouche, T.A. Evaluation measures for models assessment over imbalanced data sets. J. Inf. Sci. Eng. 2013, 3, 27–38. [Google Scholar]
  51. Powers, R.; Goldszmidt, M.; Cohen, I. Short term performance forecasting in enterprise systems. In Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery in Data Mining, Chicago, IL, USA, 21–24 August 2005. [Google Scholar]
  52. Akosa, J.S. Predictive Accuracy: A Misleading Performance Measure for Highly Imbalanced Data. In Proceedings of the SAS Global Forum, Cary, NC, USA, 2–5 April 2017. [Google Scholar]
  53. Satoła, A.; Bauer, E.A. Predicting subclinical ketosis in dairy cows using machine learning techniques. Animals 2021, 11, 2131. [Google Scholar] [CrossRef]
  54. Zecconi, A.; Zaghen, F.; Meroni, G.; Sommariva, F.; Ferrari, S.; Sora, V. Machine Learning Approach for Early Lactation Mastitis Diagnosis Using Total and Differential Somatic Cell Counts. Animals 2025, 15, 1125. [Google Scholar] [CrossRef]
  55. Contla Hernández, B.; Lopez-Villalobos, N.; Vignes, M. Identifying health status in grazing dairy cows from milk mid-infrared spectroscopy by using machine learning methods. Animals 2021, 11, 2154. [Google Scholar] [CrossRef]
  56. Ebrahimi, M.; Mohammadi-Dehcheshmeh, M.; Ebrahimie, E.; Petrovski, K.R. Comprehensive analysis of machine learning models for prediction of sub-clinical mastitis: Deep Learning and Gradient-Boosted Trees outperform other models. Comput. Biol. Med. 2019, 114, 103456. [Google Scholar] [CrossRef]
  57. Shahinfar, S.; Kelman, K.; Kahn, L. Machine Learning approach for early prediction of adult wool growth and quality in Australian Merion sheep. Comput. Electron. Agric. 2018, 148, 72–81. [Google Scholar] [CrossRef]
  58. Ding, C.; Wang, D.; Ma, X.; Li, H. Predicting short-term subway ridership and prioritizing its influential factors using gradient boosting decision trees. Sustainability 2016, 8, 1100. [Google Scholar] [CrossRef]
  59. Matsuki, K.; Kuperman, V.; Van Dyke, J.A. The Random Forests statistical technique: An examination of its value for the study of reading. Sci. Stud. Read. 2016, 20, 20–33. [Google Scholar] [CrossRef] [PubMed]
  60. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  61. Van Winden, S.; Kuiper, R. Left displacement of the abomasum in dairy cattle: Recent developments in epidemiological and etiological aspects. Vet. Res. 2003, 34, 47–56. [Google Scholar] [CrossRef] [PubMed]
  62. Niederberger, M.D.; Hirsbrunner, G.; Steiner, A.; Brechbühl, M.; Meylan, M. In vitro effects of bethanechol on abomasal and duodenal smooth muscle preparations from dairy cows with left displacement of the abomasum and from healthy dairy cows. Vet. J. 2010, 184, 88–94. [Google Scholar] [CrossRef]
  63. Jacobsen, K.L. Displaced abomasa and thin cows in a component-fed dairy herd. Compendium. Food Anim. 1995, 17, 21–27. [Google Scholar]
  64. Shaver, R. Nutritional risk factors in the etiology of left displaced abomasum in dairy cows: A review. J. Dairy Sci. 1997, 80, 2449–2453. [Google Scholar] [CrossRef]
  65. Song, Y.; Loor, J.J.; Zhao, C.; Huang, D.; Du, X.; Li, X.; Wang, Z.; Liu, G.; Li, X. Potential hemo-biological identification markers to the left displaced abomasum in dairy cows. BMC Vet. Res. 2020, 16, 470. [Google Scholar] [CrossRef]
  66. Rutten, C.; Velthuis, A.; Steeneveld, W.; Hogeveen, H. Invited review: Sensors to support health management on dairy farms. J. Dairy Sci. 2013, 96, 1928–1952. [Google Scholar] [CrossRef]
  67. Grummer, R.R. Etiology of lipid-related metabolic disorders in periparturient dairy cows. J. Dairy Sci. 1993, 76, 3882–3896. [Google Scholar] [CrossRef]
  68. Dewhurst, R.J.; Moorby, J.; Dhanoa, M.; Evans, R.T.; Fisher, W.J. Effects of altering energy and protein supply to dairy cows during the dry period. 1. Intake, body condition, and milk production. J. Dairy Sci. 2000, 83, 1782–1794. [Google Scholar] [CrossRef]
  69. Liang, D.; Arnold, L.; Stowe, C.; Harmon, R.; Bewleys, J. Estimating US dairy clinical disease costs with a stochastic simulation model. J. Dairy Sci. 2017, 100, 1472–1486. [Google Scholar] [CrossRef]
  70. Patelli, T.H.C.; Souza, F.A.A.; Rosa, F.; Barreiros, T.R.R.; Marques, L.C. Economic impact of displaced abomasum on a dairy farm in Carambeí, Paraná. Ars Vet. 2012, 29, 8–12. [Google Scholar] [CrossRef]
Figure 1. Variable importance plot of the top 10 predictor variables, according to the mean reduction in the Gini index and mean decrease in accuracy.
Figure 1. Variable importance plot of the top 10 predictor variables, according to the mean reduction in the Gini index and mean decrease in accuracy.
Animals 15 01833 g001
Figure 2. Distribution of displaced abomasum cases of dairy cattle across herds in the training set (before and after under-sampling) and testing set.
Figure 2. Distribution of displaced abomasum cases of dairy cattle across herds in the training set (before and after under-sampling) and testing set.
Animals 15 01833 g002
Figure 3. Flowchart of the resampling of the original dataset to generate the balanced set and data used for model tuning. Under-sampling performed via ten random samples from unaffected cows of each herd to the total number of cows with DA without replacement.
Figure 3. Flowchart of the resampling of the original dataset to generate the balanced set and data used for model tuning. Under-sampling performed via ten random samples from unaffected cows of each herd to the total number of cows with DA without replacement.
Animals 15 01833 g003
Figure 4. Boxplots of model performance metrics showing prediction variability in different algorithms on the testing set in 10 iterations; DT = Decision Tree, GB = Gradient Boosting (Machines), LR = Logistic Regression, NB = Naïve Bayes, RF = Random Forest, ACC = overall accuracy, BACC = balanced ACC, TPR = true positive rate, TNR = true negative rate, AUROC = Area Under the ROC, PPV = precision, F2 = F2-measure and AUPRC = Area Under the Precision–Recall Curve.
Figure 4. Boxplots of model performance metrics showing prediction variability in different algorithms on the testing set in 10 iterations; DT = Decision Tree, GB = Gradient Boosting (Machines), LR = Logistic Regression, NB = Naïve Bayes, RF = Random Forest, ACC = overall accuracy, BACC = balanced ACC, TPR = true positive rate, TNR = true negative rate, AUROC = Area Under the ROC, PPV = precision, F2 = F2-measure and AUPRC = Area Under the Precision–Recall Curve.
Animals 15 01833 g004
Figure 5. Calibration plots of different machine learning models: Decision Tree (DT), Gradient Boosting (Machine) (GB), Logistic Regression (LR), Naïve Bayes (NB) and Random Forest (RF).
Figure 5. Calibration plots of different machine learning models: Decision Tree (DT), Gradient Boosting (Machine) (GB), Logistic Regression (LR), Naïve Bayes (NB) and Random Forest (RF).
Animals 15 01833 g005
Table 1. Number of excluded records and reasons for excluding them for each parity.
Table 1. Number of excluded records and reasons for excluding them for each parity.
Service Sire PTAs 1Parity > 6AFCPL Milk Protein Fat DPL Calf Weight No Milk Record 2
197820552441171241290503533816582
1 Removal due to lack of Service Sire PTAs; 2 primiparous cows with DA and without a monthly record of milk.
Table 2. Description of features (explanatory variables) used for predicting displaced abomasum in dairy cows.
Table 2. Description of features (explanatory variables) used for predicting displaced abomasum in dairy cows.
No.FeaturesTypeLevelMinimum (%) 3Maximum (%) 4MeanSD
Herd and cow information 1
1HerdNominal71 (15%)7 (7%)--
2Calving yearsNominal112010 (3%)2020 (16%)--
3ParityNominal61 (34%)6 (5%)--
4month of the milk recordNominal121 (8.6%)12 (9.4%)--
5Calving seasonNominal41 (22%)4 (24%)--
6Milk yieldNumeric-3.078.040.912.3
7Milk fat yieldNumeric-0.297.23.71.1
8Milk protein yieldNumeric-1.15.13.00.4
9Age at first calvingNumeric-578115372055
10PregnancyNumeric-240.0298.0275.05.1
11Calf weight at birthNumeric-205340.06.2
12Dry period lengthNumeric-016041.133.2
13KetosisBinary20 (98%)1 (2%)--
14MastitisBinary20 (83%)1 (17%)--
15MetritisBinary20 (86%)1 (14%)--
16Retained placentaBinary20 (92%)1 (8%)--
17Milk feverBinary20 (98%)1 (2%)--
18DystociaBinary21 (96%)2 (4%)--
19TwinningBinary21 (97%)2 (3%)--
20Calf sexBinary21 (51%)2 (49%)--
Service Sire PTAs 2
21PTA-DANumeric-−2.91.40.020.5
22PTA-SSBNumeric-3.810.26.10.8
23PTA-DSBNumeric-2.616.76.51.7
24PTA-SCENumeric-1.06.62.20.5
25PTA-DCENumeric-1.05.52.60.6
26PTA-DPRNumeric-−7.56.9−0.72.1
27PTA-PLNumeric-−7.66.3−0.20.2
28PTA-KetNumeric-−3.73.0−0.12.4
29PTA-fat%Numeric-−0.30.4−0.070.1
30PTA-SCSNumeric-2.43.63.01.0
31PTA-protein%Numeric-−0.20.2−0.30.04
1 Parity consisted of 1 (34%), 2 (26%), 3 (18%), 4 (11%), 5 (6%) and 6 (5%). For all primiparous cows, the length of the dry period was considered zero. For all records related to disease and reproduction, 0 was no and 1 was yes. For calf gender, 1 was considered a female calf and 2 was considered a male calf. 2 PTA of displaced abomasum, sire stillbirth, daughter stillbirth, sire calving ease, daughter calving ease, the daughter pregnancy rate and productive life, somatic cell score, Ketosis, milk protein percentage and milk fat percentage. 3 and 4 Proportions and frequencies for each level of categorical and binary features were shown in parentheses. Features selected during feature selection stage are shown in bold face.
Table 3. Frequency of different levels of nominal features used in the forecasting.
Table 3. Frequency of different levels of nominal features used in the forecasting.
Level123456789101112
Features
Herd15%20%10%15%18%15%7%-----
Calving year3%2%3%4%5%7%12%14%15%17%16%-
month of the milk record8.6%7.3%7.7%7.5%7.5%7.0%8.0%10.0%9.0%9.0%8.0%9.4%
Calving season22%27%27%24%--------
Parity34%26%18%11%6%5%------
Table 4. Model performance metrics for algorithms used in prediction of the incidence of DA in dairy cows for the testing set.
Table 4. Model performance metrics for algorithms used in prediction of the incidence of DA in dairy cows for the testing set.
Algorithm 1ACC 2Balanced ACCPPV 3TPR 4TNR 5F2AUPRCAUROC
Testing dataset
  LR0.68 (±0.00) a 0.66 (±0.00) b0.09 (±0.00) b0.65 (±0.01) b0.68 (±0.00) b0.30 (±0.00) b0.15 (±0.00) b0.73 (±0.00) c
  NB0.62 (±0.02) b0.64 (±0.00) c0.07 (±0.00) c0.66 (±0.01) b0.62 (±0.02) b0.27 (±0.00) c0.10 (±0.00) c0.69 (±0.00) d
  DT0.63 (±0.05) b0.64 (±0.00) c0.08 (±0.00) c0.66 (±0.07) b0.63 (±0.06) b0.28 (±0.00) c0.10 (±0.00) c0.70 (±0.00) d
  RF0.64 (±0.01) b0.69 (±0.00) a0.09 (±0.00) b0.75 (±0.01) a0.63 (±0.01) a0.32 (±0.00) a0.17 (±0.00) a0.76 (±0.00) a
  GBM0.69 (±0.01) a0.68 (±0.00) b0.10 (±0.00) a0.70 (±0.00) b0.67 (±0.01) a0.32 (±0.00) a0.17 (±0.00) a0.75 (±0.00) b
Training dataset
  LR0.67 (±0.00) 0.67 (±0.00) 0.68 (±0.00) 0.67 (±0.00) 0.68 (±0.00) 0.67 (±0.00) 0.73 (±0.00) 0.74 (±0.00)
  NB0.64 (±0.00) 0.64 (±0.00) 0.63 (±0.01) 0.67 (±0.01) 0.61 (±0.02) 0.66 (±0.01) 0.67 (±0.00) 0.69 (±0.00)
  DT0.68 (±0.01) 0.68 (±0.01) 0.67 (±0.01) 0.70 (±0.06) 0.66 (±0.05) 0.69 (±0.05) 0.72 (±0.03) 0.74 (±0.01)
  RF0.68 (±0.00) 0.68 (±0.00) 0.66 (±0.00) 0.74 (±0.01) 0.62 (±0.01) 0.72 (±0.01) 0.73 (±0.00) 0.75 (±0.00)
  GBM0.72 (±0.01) 0.72 (±0.01) 0.72 (±0.00) 0.74 (±0.00) 0.71 (±0.01) 0.73 (±0.00) 0.80 (±0.00) 0.80 (±0.01)
1 LR: Logistic Regression, NB: Naïve Bayes, DT: Decision Tree, RF: Random Forest, GBM: Gradient Boosting Machines; 2 accuracy; 3 precision or positive predictive value (PPV), 4 true positive rate (TPR); 5 true negative rate (TNR). According to the Tukey-HSD multiple comparison test, the values in each column are significantly different (p < 0.05), which are also shown with different superscript. The bold values in the table indicate the best performance.
Table 5. The accuracy of recent studies in predicting health traits in dairy cattle.
Table 5. The accuracy of recent studies in predicting health traits in dairy cattle.
Number of Cows Predicted Trait Accuracy Algorithm with Best Performance Country Year
2535Lameness0.83Naïve BayesAustralia2021 [28]
1000Neosporosis0.82Neural NetworkColombia2025 [29]
882Ketosis0.72Logistic RegressionPoland2021 [53]
363,945Retained Placenta0.78XGBoost and Random ForestIran2025 [31]
14,755Udder Health Status≥0.75Neural Network and Random ForestItaly2021 [54]
1909Health Status0.95Neural NetworkNew Zealand2021 [55]
297,004Subclinical Mastitis≥0.83GBM and Deep LearningNew Zealand2019 [56]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Asgari, Z.; Sadeghi-Sefidmazgi, A.; Pakdel, A.; Shahinfar, S. Machine Learning Approaches for the Prediction of Displaced Abomasum in Dairy Cows Using a Highly Imbalanced Dataset. Animals 2025, 15, 1833. https://doi.org/10.3390/ani15131833

AMA Style

Asgari Z, Sadeghi-Sefidmazgi A, Pakdel A, Shahinfar S. Machine Learning Approaches for the Prediction of Displaced Abomasum in Dairy Cows Using a Highly Imbalanced Dataset. Animals. 2025; 15(13):1833. https://doi.org/10.3390/ani15131833

Chicago/Turabian Style

Asgari, Zeinab, Ali Sadeghi-Sefidmazgi, Abbas Pakdel, and Saleh Shahinfar. 2025. "Machine Learning Approaches for the Prediction of Displaced Abomasum in Dairy Cows Using a Highly Imbalanced Dataset" Animals 15, no. 13: 1833. https://doi.org/10.3390/ani15131833

APA Style

Asgari, Z., Sadeghi-Sefidmazgi, A., Pakdel, A., & Shahinfar, S. (2025). Machine Learning Approaches for the Prediction of Displaced Abomasum in Dairy Cows Using a Highly Imbalanced Dataset. Animals, 15(13), 1833. https://doi.org/10.3390/ani15131833

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop