Next Article in Journal
Fungal Flora in Asymptomatic Pet Guinea Pigs and Rabbits
Previous Article in Journal
Comparative Genome Analysis Reveals the Genomic Basis of Semi-Aquatic Adaptation in American Mink (Neovison vison)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Applying Machine Learning Algorithms for the Classification of Mink Infected with Aleutian Disease Using Different Data Sources

Department of Animal Science and Aquaculture, Dalhousie University, Truro, NS B2N 5E3, Canada
*
Author to whom correspondence should be addressed.
Animals 2022, 12(18), 2386; https://doi.org/10.3390/ani12182386
Submission received: 13 June 2022 / Revised: 8 September 2022 / Accepted: 10 September 2022 / Published: 13 September 2022
(This article belongs to the Section Veterinary Clinical Studies)

Abstract

:

Simple Summary

Aleutian disease (AD) is a major infectious disease found in mink farms, and it causes financial losses to the mink industry. Controlling AD often requires a counterimmunoelectrophoresis (CIEP) method, which is relatively expensive for mink farmers. Therefore, predicting AD infected mink without using CIEP records will be important for controlling AD in mink farms. In the current study, we applied nine machine learning algorithms to classify AD-infected mink. We indicated that the random forest could be used to classify AD-infected mink (accuracy of 0.962) accurately. This result could be used for implementing machine learning in controlling AD in the mink farms.

Abstract

American mink (Neogale vison) is one of the major sources of fur for the fur industries worldwide, whereas Aleutian disease (AD) is causing severe financial losses to the mink industry. A counterimmunoelectrophoresis (CIEP) method is commonly employed in a test-and-remove strategy and has been considered a gold standard for AD tests. Although machine learning is widely used in livestock species, little has been implemented in the mink industry. Therefore, predicting AD without using CIEP records will be important for controlling AD in mink farms. This research presented the assessments of the CIEP classification using machine learning algorithms. The Aleutian disease was tested on 1157 individuals using CIEP in an AD-positive mink farm (Nova Scotia, Canada). The comprehensive data collection of 33 different features was used for the classification of AD-infected mink. The specificity, sensitivity, accuracy, and F1 measure of nine machine learning algorithms were evaluated for the classification of AD-infected mink. The nine models were artificial neural networks, decision tree, extreme gradient boosting, gradient boosting method, K-nearest neighbors, linear discriminant analysis, support vector machines, naive bayes, and random forest. Among the 33 tested features, the Aleutian mink disease virus capsid protein-based enzyme-linked immunosorbent assay was found to be the most important feature for classifying AD-infected mink. Overall, random forest was the best-performing algorithm for the current dataset with a mean sensitivity of 0.938 ± 0.003, specificity of 0.986 ± 0.005, accuracy of 0.962 ± 0.002, and F1 value of 0.961 ± 0.088, and across tenfold of the cross-validation. Our work demonstrated that it is possible to use the random forest algorithm to classify AD-infected mink accurately. It is recommended that further model tests in other farms need to be performed and the genomic information needs to be used to optimize the model for implementing machine learning methods for AD detection.

1. Introduction

Mink is the major source of the fur industry worldwide [1], and Aleutian disease (AD), which is caused by the Aleutian mink disease virus (AMDV), brings tremendous financial losses to the mink industry [2]. The AD is associated with some important traits of farmed mink, including reproductive performance [3,4], body size [5], feed intake [6,7], and pelt quality [8]. There are several methods to test for AD, including the AMDV antigen-based enzyme-linked immunosorbent assay (ELISA-G), the AMDV capsid protein-based ELISA (ELISA-P), counterimmunoelectrophoresis (CIEP), and the iodine agglutination test (IAT). Among them, CIEP has been considered a gold standard [9]. Since controlling AD has not succeeded because no effective vaccine or medicine has been created, culling mink that have tested positive using CIEP has been applied as the primary method to control AD [10,11,12,13]. However, implementing the culling method following a positive diagnosis using CIEP requires screening a large number of animals, consequently causing the demand for an additional labor force and problems in the availability of the CIEP antigen [14]. Therefore, predicting AD without using CIEP will provide an effective way for controlling AD in mink farms.
Machine learning (ML) algorithms have proven successful in diagnosing and predicting diseases [15,16] and are becoming more popular in health care. They are also widely used methods in agriculture [17,18,19], such as for high throughput phenotyping [20] or predicting plant diseases [21]. In animals, the ML algorithms have been used for monitoring the health status [22,23], product quality [24,25], and prediction of diseases [26,27,28,29,30]. Selection of the ML algorithms for studies on farm animals depends on the traits and data, and their performance also varies among the studies [17,28,31]. For instance, different ML algorithms, including artificial neural networks [29], support vector machines [31], or random forest [28], had an accuracy of 85–90% for the classification of mastitis in dairy cows. Overall, ML is appropriate for large data or data with many predictors, missing values, and abnormal distributions. Data from a mink farm often has many predictors and missing values that might benefit from the application of ML. For instance, ML can be used for predicting some phenotypes such as daily feed consumption and reproductive performance or for recognizing the missing labels in the animal IDs in the mink farms. Therefore, we examined the performance of nine ML algorithms including artificial neural networks, decision tree, extreme gradient boosting, gradient boosting method, K-nearest neighbors, linear discriminant analysis, linear support vector machines, naive bayes, and random forest for the classification of AD in American mink. We used 33 different features obtained from different data sources in an AD positive mink farm for the classification of AD-infected mink. We used four different matrices for assessing the performance of the ML algorithms. Finally, we performed the Wilcoxon test to examine if these ML algorithms are significantly different.

2. Materials and Methods

The animals used in this study were raised according to the Code of Practice for the Care and Handling of Farmed Mink guidelines published by the Canada Mink Breeders Association (https:/www.nfacc.ca/pdfs/codes/mink_code_of_practice.pdf, accessed on 20 October 2019). The animal care was followed by protocols approved by the Dalhousie University Animal Care and Use Committee (certification# 2018-009 and 2019-012).

2.1. Animals and the Phenotypic Records

The mink used in this study were raised under standard farming conditions at the Canadian Centre for Fur Animal Research (CCFAR) at Dalhousie University, Faculty of Agriculture (Truro, NS, Canada) from 2013 to 2019. All mink had ad libitum access to feed and water. The feeds were adjusted based on the animals’ needs in each production period, with a higher dry matter added during the growing and furring periods [32]. Three AD tests were conducted using the established protocols described by Hu et al. [33]. In brief, the blood samples of the mink (1157 individuals) were collected using the toenail clipping approach. Both the ELISA-G and ELISA-P systems were employed to quantify the anti-AMDV antibodies in the serum. The ELISA-G scores (0–7) and ELISA-P scores (0–8) were conducted at Middleton Veterinary Services (Middleton, NS, Canada), and Nederlandse Federatie van Edelpelsdierenhouders (Wijchen, The Netherlands), respectively. The CIEP tests were conducted at the Animal Health Laboratory at the University of Guelph (Guelph, ON, Canada) to detect the existence of anti-AMDV antibodies in the blood samples, and the results were recorded as negative or positive. The IATs were conducted at the CCFAR to measure the serum gamma globulin level in the serum, and the results were scored into four categories from 0 (low) to 4 (high). All bodyweight (BW), growth parameters, and feed intake data were collected using the established protocols described by Do et al. [34] and Davoudi et al. [35], respectively. The mink were housed individually in single cages, and the feed was distributed to each pen every morning. The amount of allocated feed was regulated based on the leftover records one day before in order to avoid unnecessary feed waste and to meet the mink’s appetite. The daily feed intake (DFI) was obtained by subtracting the amount of leftover from the quantity of feed supplied. The average daily feed intake (ADFI) was calculated by averaging the DFI records obtained during the test period. The average daily gain (ADG), feed conversion ratio (FCR), Kleiber ratio (KR), residual feed intake (RFI), residual gain (RG), and residual intake and gain (RIG) were derived from the body weight and daily feed intake data [35]. The growth curve parameters, including asymptotic weight (α), growth rate at mature (k), shape parameter (m), weight at the inflection point (WIP), and age at the inflection point (AIP), were obtained from the body weights of mink using the Richard growth model [34]. A total of 33 features were examined for the development of the ML algorithms to classify animals for AD. The number of CIEP positive and negative mink as well as the mean values for these features are given in Table 1. A simplifized workflow of the current study is shown in Figure 1.

2.2. Algorithm Selection and the Data Preparation

The classification of AD-infected mink was constructed using the following algorithms: artificial neural networks, decision tree, extreme gradient boosting, gradient boosting, K-nearest neighbors, linear discriminant analysis, support vector machines (linear form), naive bayes, and random forest. These algorithms were selected as they have been widely used for the diagnosis of human diseases, e.g., cancers [36,37,38], as well as for predicting phenotypes in livestock [17,19,39]. All calculations were performed in R using the caret package [40], and CIEP was used as the response variable in the models. Since the CCFAR was infected with AD in 2012, more animals were positive (954) than negative (203) using CIEP in the current dataset. The missing data of features (Table 1) were input using the mice R package [41]. If the imbalance ratio is high, the decision function favors the majority group (positive CIEP group). For non-probabilistic classifiers such as logistic regression, neural networks, and support vector machines algorithms, an imbalanced data structure can affect their parameters [39]. We used the over-sampling function in the Rose package [42] to create the balanced data. In this function, the minority group (negative CIEP, n = 203) was oversampled from 203 to 954 in order to balance a sample size as the majority group (positive CIEP, n = 954). The preProcess function from the caret package [40] was used to scale and center the variables in the training dataset. The relative importance of the features and feature selection were examined using the Boruta package [43]. The Boruta algorithm is a wrapper approach that is built based on the random forest. This algorithm creates shadow features as a replica of actual features, and then randomly shuffles to remove any correlation with the response variable. In the next step, a random forest classifier is run, and the Z-score is computed by dividing the average loss by its standard deviation. The maximum Z-score of randomized shadow features is used to set a threshold for the selection of important features [43]. If the Z-score computed for an actual feature is significantly more than the Z-score of the shadow feature, then it is considered as an important feature [43].

2.3. Model Training and Performance Assessment

Following the oversampling, the data was randomly divided into 80% for training and 20% for testing datasets. We created ten different sets of training-testing dataset using the createDataPartition function. The models were built in each training dataset and evaluated in each test data. In the training dataset, we used the trainControl function to select the hyperparameters for the model building. The repeated cross-validation methods implemented in the trainControl function were used. In this method, for each one of the ten iterations, the hyperparameters were selected using a search within the 10-fold cross-validation structure on a random 70% subset of the training dataset. Each algorithm was run separately using the default initial hypermeters and the train function of the caret package. The confusionMatrix function was used to evaluate the model performance from the best built model for each training dataset and the corresponding testing dataset.
The model fit and ranking of the models were assessed using several scores that were computed using the number of true positive (TP), true negative (TN), false positive (FP), and false-negative (FN). The following formulas were used for the calculation of accuracy (Equation (1)), specificity (Precision; Equation (2)), which is the fraction of correct predictions, sensitivity (Recall; Equation (3)), which measures a fraction of the correct predictions per true number of samples, and F-Measure (F1; Equation (4)), which is a goodness of fit assessment for a classification analysis that balances precision and recall:
Accuracy = TP   +   TN TP   +   TN   +   FP   +   FN
Specificity   Precision = TP TP   +   FP
Sensitivity   Recall = TP TP   +   FN
F 1   =   2 × Precision   ×   Recall Precision   +   Recall
A receiver operating characteristic (ROC) curve was used to depict the sensitivity against 1-specificity over all possible decision thresholds ranging for classifying the predicted AD-infected mink and was characterized using the pROC package [44]. The accuracy of the models was assessed by calculating the area under the curve (AUC). The values of AUC were interpreted as non-accurate (AUC = 0.5), less accurate (0.5 < AUC ≤ 0.7), moderately accurate (0.7 < AUC ≤ 0.9), highly accurate (0.9 < AUC < 1) and perfectly (AUC = 1) [45]. Moreover, the pairwise differences in the accuracy of the models were compared using the Wilcoxon test.

3. Results and Discussion

3.1. Feature Importance and the Model Performance

The descriptive statistics of all numerical features are shown in Table 1. A total of 33 different features were collected and used as input in the Boruta package. The relative importance of the features based on the random forest from the Boruta package is shown in Figure 2.
The ELISA-P was identified as the most important feature for the classification of AD-infected mink. Other important features based on the ranking by the Boruta package for the AD-infected mink classification were those related to bodyweight measures, growth curve parameters, and DFI. Sex, birth year, and color were the less important features for the classification of AD-infected mink. The importance of the ELISA is expected as the ELISA systems are also alternative methods for the diagnosis of AD [46]. Previously, we also reported that the ELISA tests had significant phenotypic and genetic correlations with CIEP [33]. The age at sampling might be an important feature for CIEP since, if animals stayed on the farm for a longer period of time, they might have a higher chance of being infected by the AMDV. Bodyweights, growth curve parameters, and DFI were important traits for the growth of animals. Since AD harms the animal’s health and growth [47], it was expected that these features would be important for classifying AD-infected animals. Interestingly, the variation in feed intake was important for the CIEP classification, which might be because of the inconsistency in the diet of infected mink. Sex and color type were less important for the CIEP classification, which was also supported by our previous study that these effects were not significantly affecting the CIEP [33].

3.2. Performance Assessment

Table 2 presents the sensitivity, specificity, F1, and accuracy of the nine ML algorithms using four different sub-sampling procedures. Overall, the sensitivity, specificity, F1, and accuracy were varied between the algorithms. The specificity ranged from 0.588 (K-nearest neighbors) to 0.938 (random forest), while the sensitivity ranged from 0.841 (naive bayes) to 0.987 (extreme gradient boosting). All algorithms obtained higher specificity values than sensitivity values. All algorithms had the F1 and accuracy values of more than 0.7, thereby indicating that they could be used for the classification of AD-infected mink with an acceptable accuracy. The random forest algorithm had an excellent performance considering both the F1 measure and accuracy (>0.95).
Table 3 shows the confusion matrix obtained from the random forest algorithm. The random forest could correctly classify 186 out of 190 CIEP positive mink and 184 out of 190 CIEP negative mink. The Friedman test indicated the significant differences in the accuracies obtained from the different algorithms according to the subsampling procedures (p-value < 2.2 × 10−16). The paired samples used in the Wilcoxon tests for the differences indicated that all algorithms had significant differences in their accuracies, except for the K-nearest neighbors with the linear discriminant analysis (p = 0.18) and the naive bayes (p = 0.51) (Table 4).
The average values of the AUC (Figure 3) indicated that the random forest and the extreme gradient boosting were the best algorithms with the highest AUC as their AUC values were >0.90. Other algorithms had a moderate AUC with the AUC values ranging from 0.75 (naive bayes and K-nearest neighbors) to 0.88 (gradient boosting).
All of the tested algorithms in this study have been used for the diagnosis of diseases [17,39,45]. The decision tree, gradient boosting, random forest, and extreme gradient boosting are all tree-based models, and their performances will be influenced by the class imbalances via the leaf impurity. To solve the problem of class imbalance, the oversampling method was chosen for handling the imbalance classifiers as it does not lead to any information loss. In the current study, the random forest outperformed the other methods, which was consistent with the previous study [39] that implemented the random forest to predict the leg weakness in pigs. However, the random forest was not the best method as observed by Shao et al. [48], who showed that the support vector machines outperformed the neural networks, random forest, and linear regression to predict the corrected inventory decision for the market using China’s hog inventory data. The random forest was also less accurate compared with the support vector machines, kernel ridge regression, and Adaboost.R2 in the prediction of the reproductive performance traits in pigs using genomic data [49]. The random forest approach is known to be fairly stable in the presence of outliers and noise and can handle the correlations between the predictors [49,50].
Extreme gradient boosting was the second-best method for the classification of AD-infected mink, which might be because this method could perform implicit variable selections and could capture the non-linear relationships [51,52]. Both the random forest and the extreme gradient boosting showed great potential for the classification of CIEP in the current study with high accuracies, F1 values, and AUCs. Especially, these algorithms were close to perfect (sensitivity > 0.99) in classifying AD-infected mink. Both the random forest and extreme gradient boosting succeeded in detecting the posture and behavior in dairy cows [51] with the accuracy obtained from the extreme gradient boosting algorithm in predicting the posture was 0.99, and the random forest had the highest overall accuracy in predicting the behavior (0.76). The NNET performed fairly well, but was not the best, and that could be due to the limitation of the fine tuning in hyper parameters for the NNET or the limitation of the small sample size in the current study. Nevertheless, the performance of the ML algorithms depends on the data, and therefore, it is necessary to test different algorithms to find the most suitable one.
The current study had some limitations. Although CIEP has been adapted by the mink farmers for controlling AD, it is important to mention that the CIEP test results are binary outcomes while AD is a chronic disease. The CIEP measure is sensitive to the status of the disease; therefore, the results of the current study might be limited by the lack of repeated measurements of CIEP. More frequent measures of CIEP are required to confirm the AD status better and to consequently apply the ML algorithms to the classification of AD-infected mink. Additionally, even though the random forest reached a very high accuracy, specificity, and sensitivity, some individuals were still wrongly classified. Therefore, larger sample sizes with more features or better hyper-parameterizing are required for correctly classifying these individuals. The results of the current study were also limited to use in the AD positive mink farms. However, the majority of mink farms are infected with AD; thus, the results are still beneficial for most mink farmers. In the meantime, these results could be helpful for the farmers who want to cull animals based on the classified AD-infected mink obtained by the ML algorithms.
Finally, although being considered as a gold standard for the AD test, CIEP is a relatively expensive test and requires a large labor force, due to the many steps in the CIEP test that are performed manually. Moreover, the CIEP results are prone to false-positive results as the accuracies of the CIEP results are dependent on the experience of the readers in visualizing the bands. These drawbacks of CIEP limit its application in large mink farms. Alternatively, the ELISA test can be used for high-throughput assays, and the ELISA results can be used in the ML approaches (e.g., random forest or extreme gradient boosting) to accurately classify the AD-infected mink. Therefore, the mink farmers might not need to perform the CIEP test, but use the information from the ELISA test to predict AD risks and to decide which animals are needed to be culled for the control of AD.

4. Conclusions

In summary, among the nine ML algorithms, the random forest was the best method for the classification of AD-infected mink in the current dataset. This study indicated that it is possible to classify AD-infected mink with a high accuracy, specificity, and sensitivity using the random forest algorithms. Therefore, it is suggested that the random forest algorithm might be used for classifying the AD-infected mink in other AD-positive farms. Given the fact that the current study used the data from only one AD-positive farm and the performance of the ML algorithms were sensitive to the data input, it is recommended that further model tests in other AD-positive farms be performed. Since AD is a chronic disease, it is also recommended to collect disease records more frequently for better disease monitoring. Finally, it is also recommended to combine the genomic information to optimize the model for the implementation of machine learning methods in controlling AD.

Author Contributions

Conceptualization, D.N.D. and Y.M.; data preparation: D.N.D., G.H., P.D. and A.S., data analyses: D.N.D.; interpreting of results: D.N.D., G.M. and Y.M., writing—original draft preparation, D.N.D.; writing—review and editing, Y.M., G.M., G.H., P.D., A.S. and D.N.D.; project administration and funding acquisition, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by financial support from the Natural Sciences and Engineering Research Council (NSERC) of Canada, Mitacs, Canada Mink Breeders Association, Nova Scotia Mink Breeders Association, Joint Mink Research Committee and Mink Veterinary Consulting Research and Services.

Institutional Review Board Statement

The animals used in this study were raised according to the Code of Practice for the Care and Handling of Farmed Mink guidelines published by the Canada Mink Breeders Association (https:/www.nfacc.ca/pdfs/codes/mink_code_of_practice.pdf, accessed on 20 October 2019). The animal care was followed by the protocols approved by the Dalhousie University Animal Care and Use Committee (certification# 2018-009, and 2019-012).

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this work are available from the corresponding author upon academic request.

Acknowledgments

The authors gratefully acknowledge financial support from the Natural Sciences and Engineering Research Council (NSERC) of Canada, Canada Mink Breeders Association, Nova Scotia Mink Breeders Association, Nova Scotia Department of Agriculture, and Mink Veterinary Consulting Research and Services. We also thank the CCFAR staff for collecting and providing the data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Manakhov, A.D.; Andreeva, T.V.; Trapezov, O.V.; Kolchanov, N.A.; Rogaev, E.I. Genome analysis identifies the mutant genes for common industrial Silverblue and Hedlund white coat colours in American mink. Sci. Rep. 2019, 9, 1–8. [Google Scholar]
  2. Hu, G.; Do, D.N.; Gray, J.; Miar, Y. Selection for favorable health traits: A potential approach to cope with diseases in farm animals. Animals 2020, 10, 1717. [Google Scholar] [CrossRef] [PubMed]
  3. Henson, J.; Gorham, J.; Leader, R. Field test for Aleutian disease. Natl. Fur News 1962, 34, 8–9. [Google Scholar]
  4. Reichert, M.; Kostro, K. Effect of persistent infection of mink with Aleutian mink disease virus on reproductive failure. Bull. Vet. Inst. Pulawy. 2014, 58, 369–373. [Google Scholar] [CrossRef]
  5. Kowalczyk, M.; Gąsiorek, B.; Kostro, K.; Borzym, E.; Jakubczak, A. Breeding parameters on a mink farm infected with Aleutian mink disease virus following the use of methisoprinol. Arch. Virol. 2019, 164, 2691–2698. [Google Scholar] [CrossRef]
  6. Eklund, C.; Hadlow, W.; Kennedy, R.; Boyle, C.; Jackson, T. Aleutian disease of mink: Properties of the etiologic agent and the host responses. J. Infect. Dis. 1968, 118, 510–526. [Google Scholar] [CrossRef]
  7. Jensen, T.H.; Chriél, M.; Hansen, M.S. Progression of experimental chronic Aleutian mink disease virus infection. Acta Vet. Scand. 2016, 58, 35. [Google Scholar] [CrossRef]
  8. Farid, A.; Ferns, L. Aleutian mink disease virus infection may cause hair depigmentation. Scientifur 2011, 35, 55–59. [Google Scholar]
  9. Andersson, A.-M.; Wallgren, P. Evaluation of two enzyme-linked immunosorbent assays for serodiagnosis of Aleutian mink disease virus infection in mink. Acta Vet. Scand. 2013, 55, 1–6. [Google Scholar] [CrossRef]
  10. Christensen, L.S.; Gram-Hansen, L.; Chriél, M.; Jensen, T.H. Diversity and stability of Aleutian mink disease virus during bottleneck transitions resulting from eradication in domestic mink in Denmark. Vet. Microbiol. 2011, 149, 64–71. [Google Scholar] [CrossRef]
  11. Farid, A.H.; Zillig, M.L.; Finley, G.G.; Smith, G.C. Prevalence of the Aleutian mink disease virus infection in Nova Scotia, Canada. Prev. Vet. Med. 2012, 106, 332–338. [Google Scholar] [CrossRef] [PubMed]
  12. Gunnarsson, E. Documenting Freedom From Disease And Re-Establishing a Free Status After a Breakdown Aleutian Disease (Plasmacytosis) in Farmed Mink in Iceland. Acta Vet. Scand. 2001, 42, S87. [Google Scholar] [CrossRef]
  13. Themudo, G.E.; Østergaard, J.; Ersbøll, A.K. Persistent spatial clusters of plasmacytosis among Danish mink farms. Prev. Vet. Med. 2011, 102, 75–82. [Google Scholar] [CrossRef] [PubMed]
  14. Knuuttila, A.; Aronen, P.; Eerola, M.; Gardner, I.A.; Virtala, A.-M.K.; Vapalahti, O. Validation of an automated ELISA system for detection of antibodies to Aleutian mink disease virus using blood samples collected in filter paper strips. Virol. J. 2014, 11, 141. [Google Scholar] [CrossRef]
  15. Sajda, P. Machine learning for detection and diagnosis of disease. Annu. Rev. Biomed. Eng. 2006, 8, 537–565. [Google Scholar] [CrossRef]
  16. Schaefer, J.; Lehne, M.; Schepers, J.; Prasser, F.; Thun, S. The use of machine learning in rare diseases: A scoping review. Orphanet J. Rare Dis. 2020, 15, 145. [Google Scholar] [CrossRef]
  17. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine learning in agriculture: A review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef]
  18. Neethirajan, S. The role of sensors, big data and machine learning in modern animal farming. Sens. Bio-Sens. Res. 2020, 29, 100367. [Google Scholar] [CrossRef]
  19. Morota, G.; Ventura, R.V.; Silva, F.F.; Koyama, M.; Fernando, S.C. Big data analytics and precision animal agriculture symposium: Machine learning and data mining advance predictive big data analysis in precision animal agriculture. J. Anim. Sci. 2018, 96, 1540–1550. [Google Scholar] [CrossRef]
  20. Singh, A.; Ganapathysubramanian, B.; Singh, A.K.; Sarkar, S. Machine learning for high-throughput stress phenotyping in plants. Trends Plant Sci. 2016, 21, 110–124. [Google Scholar] [CrossRef]
  21. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef] [PubMed]
  22. Lopez-Villalobos, N.; Vignes, M. Identifying Health Status in Grazing Dairy Cows from Milk Mid-Infrared Spectroscopy by Using Machine Learning Methods. Animals 2021, 11, 2154. [Google Scholar]
  23. Cairo, F.C.; Pereira, L.G.R.; Campos, M.M.; Tomich, T.R.; Coelho, S.G.; Lage, C.F.A.; Fonseca, A.P.; Borges, A.M.; Alves, B.R.C.; Dorea, J.R.R. Applying machine learning techniques on feeding behavior data for early estrus detection in dairy heifers. Comput. Electron. Agric. 2020, 179, 105855. [Google Scholar] [CrossRef]
  24. Alaiz-Rodríguez, R.; Parnell, A.C. A machine learning approach for lamb meat quality assessment using FTIR spectra. IEEE Access 2020, 8, 52385–52394. [Google Scholar] [CrossRef]
  25. Chen, D.; Wu, P.; Wang, K.; Wang, S.; Ji, X.; Shen, Q.; Yu, Y.; Qiu, X.; Xu, X.; Liu, Y. Combining computer vision score and conventional meat quality traits to estimate the intramuscular fat content using machine learning in pigs. Meat Sci. 2021, 185, 108727. [Google Scholar] [CrossRef]
  26. Taneja, M.; Byabazaire, J.; Jalodia, N.; Davy, A.; Olariu, C.; Malone, P. Machine learning based fog computing assisted data-driven approach for early lameness detection in dairy cattle. Comput. Electron. Agric. 2020, 171, 105286. [Google Scholar] [CrossRef]
  27. Kaler, J.; Mitsch, J.; Vázquez-Diosdado, J.A.; Bollard, N.; Dottorini, T.; Ellis, K.A. Automated detection of lameness in sheep using machine learning approaches: Novel insights into behavioural differences among lame and non-lame sheep. R. Soc. Open Sci. 2020, 7, 190824. [Google Scholar] [CrossRef] [PubMed]
  28. Fadul-Pacheco, L.; Delgado, H.; Cabrera, V.E. Exploring machine learning algorithms for early prediction of clinical mastitis. Int. Dairy J. 2021, 119, 105051. [Google Scholar] [CrossRef]
  29. Sun, Z.; Samarasinghe, S.; Jago, J. Detection of mastitis and its stage of progression by automatic milking systems using artificial neural networks. J. Dairy Res. 2010, 77, 168–175. [Google Scholar] [CrossRef]
  30. de Alencar Nääs, I.; da Silva Lima, N.D.; Gonçalves, R.F.; de Lima, L.A.; Ungaro, H.; Abe, J.M. Lameness prediction in broiler chicken using a machine learning technique. Inf. Process. Agric. 2021, 8, 409–418. [Google Scholar] [CrossRef]
  31. Mammadova, N.; Keskin, I. Application of the support vector machine to predict subclinical mastitis in dairy cattle. Sci. World J. 2013, 2013, 603897. [Google Scholar] [CrossRef] [PubMed]
  32. Do, D.N.; Miar, Y. Evaluation of growth curve models for body weight in American mink. Animals 2019, 10, 22. [Google Scholar]
  33. Hu, G.; Do, D.N.; Karimi, K.; Miar, Y. Genetic and phenotypic parameters for Aleutian disease tests and their correlations with pelt quality, reproductive performance, packed-cell volume, and harvest length in mink. J. Anim. Sci. 2021, 99, skab216. [Google Scholar] [CrossRef] [PubMed]
  34. Do, D.N.; Hu, G.; Salek Ardestani, S.; Miar, Y. Genetic and phenotypic parameters for body weights, harvest length, and growth curve parameters in American mink. J. Anim. Sci. 2021, 99, skab049. [Google Scholar] [CrossRef] [PubMed]
  35. Davoudi, P.; Do, D.N.; Colombo, S.; Rathgeber, B.; Hu, G.; Sargolzaei, M.; Wang, Z.; Plastow, G.; Miar, Y. Genetic and phenotypic parameters for feed efficiency and component traits in American mink. J. Anim. Sci. 2022, 100, skac216. [Google Scholar] [CrossRef] [PubMed]
  36. Bayrak, E.A.; Kırcı, P.; Ensari, T. Comparison of machine learning methods for breast cancer diagnosis. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019; pp. 1–3. [Google Scholar]
  37. Kourou, K.; Exarchos, T.P.; Exarchos, K.P.; Karamouzis, M.V.; Fotiadis, D.I. Machine learning applications in cancer prognosis and prediction. Comput. Struct. Biotechnol. J. 2015, 13, 8–17. [Google Scholar] [CrossRef] [PubMed]
  38. Cruz, J.A.; Wishart, D.S. Applications of machine learning in cancer prediction and prognosis. Cancer Inform. 2006, 2, 117693510600200030. [Google Scholar] [CrossRef]
  39. Bakoev, S.; Getmantseva, L.; Kolosova, M.; Kostyunina, O.; Chartier, D.R.; Tatarinova, T.V. PigLeg: Prediction of swine phenotype using machine learning. PeerJ 2020, 8, e8764. [Google Scholar] [CrossRef]
  40. Kuhn, M. Building predictive models in R using the caret package. J. Stat. Softw. 2008, 28, 1–26. [Google Scholar] [CrossRef]
  41. van Buuren, S.; Groothuis-Oudshoorn, K. Mice: Multivariate Imputation by Chained Equations in R. J. Stat. Softw. 2011, 45, 1–67. [Google Scholar] [CrossRef]
  42. Lunardon, N.; Menardi, G.; Torelli, N. ROSE: A package for binary imbalanced learning. R J. 2014, 6, 79–89. [Google Scholar] [CrossRef]
  43. Kursa, M.B.; Rudnicki, W.R. Feature selection with the Boruta package. J. Stat. Softw. 2010, 36, 1–13. [Google Scholar] [CrossRef]
  44. Robin, X.; Turck, N.; Hainard, A.; Tiberti, N.; Lisacek, F.; Sanchez, J.-C.; Müller, M. pROC: An open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinform. 2011, 12, 1–8. [Google Scholar] [CrossRef] [PubMed]
  45. Greiner, M.; Pfeiffer, D.; Smith, R. Principles and practical application of the receiver-operating characteristic analysis for diagnostic tests. Prev. Vet. Med. 2000, 45, 23–41. [Google Scholar] [CrossRef]
  46. Knuuttila, A.; Aronen, P.; Saarinen, A.; Vapalahti, O. Development and Evaluation of an Enzyme-Linked Immunosorbent Assay Based on Recombinant VP2 Capsids for the Detection of Antibodies to Aleutian Mink Disease Virus. Clin. Vaccine Immunol. 2009, 16, 1360–1365. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Kashtanov, S.; Salnikova, L. Aleutian mink disease: Epidemiological and genetic aspects. Biol. Bull. Rev. 2018, 8, 104–113. [Google Scholar] [CrossRef]
  48. Shao, Y.; Xiong, T.; Li, M.; Hayes, D.; Zhang, W.; Xie, W. China’s Missing Pigs: Correcting China’s Hog Inventory Data Using a Machine Learning Approach. Am. J. Agric. Econ. 2021, 103, 1082–1098. [Google Scholar] [CrossRef]
  49. Wang, X.; Shi, S.; Wang, G.; Luo, W.; Wei, X.; Qiu, A.; Luo, F.; Ding, X. Using Machine Learning To Improve the Accuracy of Genomic Prediction on Reproduction Traits in Pigs. J. Anim. Sci. Biotechnol. 2021, 13, 1–12. [Google Scholar]
  50. Chen, C.; Liaw, A.; Breiman, L. Using random forest to learn imbalanced data. Univ. Calif. Berkeley 2004, 110, 24. [Google Scholar]
  51. Balasso, P.; Marchesini, G.; Ughelini, N.; Serva, L.; Andrighetto, I. Machine learning to detect posture and behavior in dairy cows: Information from an accelerometer on the animal’s left flank. Animals 2021, 11, 2972. [Google Scholar] [CrossRef]
  52. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
Figure 1. A simplified workflow representation of the current study.
Figure 1. A simplified workflow representation of the current study.
Animals 12 02386 g001
Figure 2. Feature importance based on the Boruta package. Blue boxplots correspond to the minimal, average, and maximum Z scores of a shadow attribute. Green boxplots represent the Z scores of the confirmed attributes. ° in the plot indicated the outliners.
Figure 2. Feature importance based on the Boruta package. Blue boxplots correspond to the minimal, average, and maximum Z scores of a shadow attribute. Green boxplots represent the Z scores of the confirmed attributes. ° in the plot indicated the outliners.
Animals 12 02386 g002
Figure 3. The boxplot for the area under the ROC curve (AUC) for the nine machine learning algorithms*. The values indicated the mean of AUC obtained from tenfold cross-validation. * Decision tree (DEC), gradient boosting (GBM), K-nearest neighbors (KNN), linear discriminant analysis (LDA), naive bayes (NB), artificial neural networks (NNET), random forest (RF), linear support vector machines (SVM), and extreme gradient boosting (Xgboost). The red diamond shapes indicated the mean of AUC; the black dots indicated the outliners.
Figure 3. The boxplot for the area under the ROC curve (AUC) for the nine machine learning algorithms*. The values indicated the mean of AUC obtained from tenfold cross-validation. * Decision tree (DEC), gradient boosting (GBM), K-nearest neighbors (KNN), linear discriminant analysis (LDA), naive bayes (NB), artificial neural networks (NNET), random forest (RF), linear support vector machines (SVM), and extreme gradient boosting (Xgboost). The red diamond shapes indicated the mean of AUC; the black dots indicated the outliners.
Animals 12 02386 g003
Table 1. A summary of features used for the classification of CIEP.
Table 1. A summary of features used for the classification of CIEP.
CIEP NegativeCIEP Positive
FeaturesNameNMeanSDNMeanSD
General features
Sample_Age (day)Age at the collection of blood for AD testing201197.363.1902234.6136.3
RowIDyearRow number where the mink were kept each year2006.552.808535.203.09
AgeFEAge of the mink (in days) when the feeding measure started200198.253.50852197.609.73
DamageThe damage score in the fur2021.420.638811.340.57
Aleutian disease and health-related tests
Elisa_PIn vitro cultured Aleutian mink disease virus antigen-based enzyme-linked immunosorbent assay test2030.330.979241.672.29
Elisa_GCapsid protein of Aleutian mink disease virus-based enzyme-linked immunosorbent assay test2030.640.989222.462.19
PCVPacked cell volume20158.112.9491956.674.03
IATIodine agglutination test2000.420.609180.771.05
Feed intake and efficiency
DFIDaily feed intake200221.3656.24853227.2157.45
ADGAverage Daily Gain1997.453.408218.603.86
FCRFood Conversion Ratio19933.5911.1781330.4211.40
RFIResidual feed intake1987.4534.69820−1.1436.96
RGResidual intake and gain1990.111.12817−0.011.44
RIGResidual gain198−0.141.378190.011.56
KRKleiber ratio1995.191.408215.511.58
OfffeeddaysProportion of off-feed days based on feed intake1990.050.098420.060.08
FeedvariationDay-to-day variation in feed intake19948.2211.5584249.1518.02
Body weight and growth parameters
BW13Body weight at week 131981.150.268371.270.30
BW16Body weight at week 161981.420.368341.580.41
BW19Body weight at week 191981.620.468291.820.53
BW22Body weight at week 221981.820.548242.040.61
BW25Body weight at week 251981.880.568102.130.64
BW28Body weight at week 281971.920.598052.170.67
AlphaWeight at maturity [34]1981.990.638072.250.71
kMaturation rate [34]1980.240.108030.240.11
mInflection parameter [34]1980.680.907980.640.83
AIPAge at the inflection point [34]19510.791.8679910.971.83
WIPWeight at the inflection point [34]1960.890.338041.010.35
SD: Standard Deviation.
Table 2. The mean (SE *) of the sensitivity, specificity, F1, and accuracy of the classification of Aleutian disease using nine machine learning algorithms.
Table 2. The mean (SE *) of the sensitivity, specificity, F1, and accuracy of the classification of Aleutian disease using nine machine learning algorithms.
AlgorithmsSensitivitySpecificityF1Accuracy
Artificial Neural Networks0.805 ± 0.0080.877 ± 0.0160.836 ± 0.0960.841 ± 0.007
Decision tree0.634 ± 0.0070.894 ± 0.0140.726 ± 0.010.764 ± 0.007
Extreme Gradient Boosting0.905 ± 0.0020.987 ± 0.0050.944 ± 0.0020.946 ± 0.002
Gradient Boosting0.831 ± 0.0050.924 ± 0.010.871 ± 0.0350.877 ± 0.004
K-Nearest Neighbors0.588 ± 0.0060.909 ± 0.0110.700 ± 0.0000.749 ± 0.005
Linear Discriminant Analysis0.672 ± 0.0050.865 ± 0.0090.743 ± 0.0010.768 ± 0.004
Naive Bayes0.666 ± 0.0070.841 ± 0.0140.73 ± 0.0020.753 ± 0.007
Random Forest0.938 ± 0.0030.986 ± 0.0050.961 ± 0.0880.962 ± 0.002
Support Vector Machines0.687 ± 0.0050.872 ± 0.010.757 ± 0.0010.779 ± 0.004
* SE: Standard errors.
Table 3. A confusion matrix obtained from the random forest algorithm for the classification of CIEP in mink.
Table 3. A confusion matrix obtained from the random forest algorithm for the classification of CIEP in mink.
Actual Data Accuracy = 186   +   184 186   +   4   +   6   +   184 = 0.974
F 1 = 2 × 0.979   ×   0.968 0.979   +   0.968 = 0.973
PositiveNegative
Predicted dataPositive1844
Negative6186
Total190190
Sensitivity = 184 190 = 0.968 Specificity = 186 190 = 0.979
Table 4. The paired samples used in the Wilcoxon tests for the differences in the accuracies obtained from the nine machine learning algorithms.
Table 4. The paired samples used in the Wilcoxon tests for the differences in the accuracies obtained from the nine machine learning algorithms.
ModelA *ModelBFDR **ModelAModelBFDR
GBMDEC7.90 × 10−18RFKNN7.90 × 10−18
KNNDEC2.36 × 10−3SVMKNN1.18 × 10−7
LDADEC5.15 × 10−3XGBOOSTKNN7.90 × 10−18
NBDEC8.90 × 10−6NBLDA9.33 × 10−3
NNETDEC1.91 × 10−17NNETLDA1.10 × 10−17
RFDEC7.90 × 10−18RFLDA7.90 × 10−18
SVMDEC1.18 × 10−4SVMLDA1.51 × 10−12
XGBOOSTDEC7.90 × 10−18XGBOOSTLDA7.90 × 10−18
KNNGBM7.90 × 10−18NNETNB1.10 × 10−17
LDAGBM7.90 × 10−18RFNB7.90 × 10−18
NBGBM7.90 × 10−18SVMNB8.91 × 10−12
NNETGBM9.37 × 10−13XGBOOSTNB7.90 × 10−18
RFGBM7.90 × 10−18RFNNET7.90 × 10−18
SVMGBM7.90 × 10−18SVMNNET2.23 × 10−17
XGBOOSTGBM7.90 × 10−18XGBOOSTNNET7.90 × 10−18
LDAKNN0.18SVMRF7.90 × 10−18
NBKNN0.51XGBOOSTRF3.77 × 10−8
NNETKNN1.08 × 10−17XGBOOSTSVM7.90 × 10−18
* Decision tree (DEC), gradient boosting (GBM), K-nearest neighbors (KNN), linear discriminant analysis (LDA), naive bayes (NB), artificial neural networks (NNET), random forest (RF), support vector machines (SVM), and extreme gradient boosting (Xgboost). ** FDR: The p-value adjusted for false discovery rate.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Do, D.N.; Hu, G.; Davoudi, P.; Shirzadifar, A.; Manafiazar, G.; Miar, Y. Applying Machine Learning Algorithms for the Classification of Mink Infected with Aleutian Disease Using Different Data Sources. Animals 2022, 12, 2386. https://doi.org/10.3390/ani12182386

AMA Style

Do DN, Hu G, Davoudi P, Shirzadifar A, Manafiazar G, Miar Y. Applying Machine Learning Algorithms for the Classification of Mink Infected with Aleutian Disease Using Different Data Sources. Animals. 2022; 12(18):2386. https://doi.org/10.3390/ani12182386

Chicago/Turabian Style

Do, Duy Ngoc, Guoyu Hu, Pourya Davoudi, Alimohammad Shirzadifar, Ghader Manafiazar, and Younes Miar. 2022. "Applying Machine Learning Algorithms for the Classification of Mink Infected with Aleutian Disease Using Different Data Sources" Animals 12, no. 18: 2386. https://doi.org/10.3390/ani12182386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop