Next Article in Journal
Multi-Omics Revealed the Effects of Different Feeding Systems on Rumen Microorganisms, Cellulose Degradation, and Metabolites in Mongolian Cattle
Previous Article in Journal
Impact of Biosecurity on Production Performance and Antimicrobial Usage in Broiler Farms in Cameroon
Previous Article in Special Issue
Prediction of Weight and Body Condition Score of Dairy Goats Using Random Forest Algorithm and Digital Imaging Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction

1
State Key Laboratory of Swine and Poultry Breeding Industry, College of Animal Science and Technology, Sichuan Agricultural University, Chengdu 611130, China
2
Key Laboratory of Livestock and Poultry Multi-Omics, Ministry of Agriculture and Rural Affairs, College of Animal Science and Technology, Sichuan Agricultural University, Chengdu 611130, China
3
Farm Animal Genetic Resources Exploration and Innovation Key Laboratory of Sichuan Province, Sichuan Agricultural University, Chengdu 611130, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work and share first authorship.
These authors also contributed equally to this work.
Animals 2025, 15(12), 1773; https://doi.org/10.3390/ani15121773
Submission received: 15 April 2025 / Revised: 4 June 2025 / Accepted: 13 June 2025 / Published: 16 June 2025

Simple Summary

In livestock farming, accurately predicting the feed conversion ratio (FCR) is crucial for improving production efficiency and reducing costs. This study aimed to explore whether machine learning could use short-term FCR data to predict long-term FCR. Using 438,552 feed samples from pigs in Sichuan, the researchers tested 19 machine learning algorithms. The results showed that the Gradient Boosting model performed best. When the testing interval exceeded 40 kg, this model achieved high accuracy, with a coefficient of determination of 0.72 and a correlation of 0.85 between predicted and actual values. The most ideal feeding stage for FCR testing was found to be 50–90 kg. These findings provide an effective way to predict feed efficiency, helping farmers optimize feeding strategies, allocate resources better, and manage livestock more proactively, thus contributing to more sustainable and efficient animal husbandry.

Abstract

Feed conversion ratio (FCR) is a critical indicator of production efficiency in livestock husbandry. Improving FCR is essential for optimizing resource utilization and enhancing productivity. Traditional methods for FCR optimization rely on experience and long-term data collection, which are time-consuming and inefficient. This study explores the feasibility of predicting long-term FCR using short-term FCR data based on machine learning techniques. We employed nineteen machine learning algorithms, including Linear Regression, support vector machines (SVMs), and Gradient Boosting, using historical datasets to train and validate the models. The results show that the Gradient Boosting model demonstrated superior performance, achieving a coefficient of determination (R2) of 0.72 and a correlation of 0.85 between predicted and actual values when the testing interval exceeded 40 kg. Therefore, we recommend a minimum feeding measurement interval of 40 kg. Furthermore, when the testing interval was set to 40 kg and further refined to the range of 50–90 kg, the model achieved an R2 of 0.81 and a correlation of 0.90 for FCR prediction in the 30–105 kg range. Among the 19 machine learning algorithms tested, Gradient Boosting, LightGBM, and CatBoost showed superior performance, with Gradient Boosting achieving the best results. Considering practical production requirements, the 50–90 kg feeding stage proved to be the most ideal for FCR testing. This study provides a more effective method for predicting feed efficiency and offers robust data support for precision livestock farming.

1. Introduction

With the continuous growth of the global population and increasing resource scarcity, precision agriculture technology has become a key tool for improving agricultural production efficiency [1]. Yamsani et al. [2] developed a systematic framework for using historical farm data to predict future productivity patterns, which integrates IoT sensor networks and machine learning algorithms to monitor livestock behaviors, analyze real-time physiological data, and establish predictive models for optimizing resource allocation and breeding strategies in smart agriculture.
In animal husbandry, the feed conversion ratio (FCR) is a key indicator of production efficiency. Prakash et al. [3] performed an extensive genetic analysis of FCR and related growth parameters in broiler chickens, establishing the hereditary basis for feed efficiency. Davison et al. [4] built upon this by developing machine learning models that predict FCR using feeding behavior data, while France et al. [5] established advanced methods for evaluating animal feed quality and its impact on FCR.
The complexity of FCR optimization has been well documented. Morota et al. [6] created a comprehensive framework for collecting and analyzing feed efficiency data, while Sharma et al. [7] conducted comparative analyses of various machine learning models for predicting livestock productivity. Mahmud [8] explored the challenges and opportunities in applying machine learning to livestock farming, particularly focusing on data quality and model selection issues. Zhang fan et al. [9] demonstrated practical applications by successfully implementing neural networks for FCR prediction in dairy cows.
Recent technological advances have enabled more sophisticated approaches to FCR prediction. Gaillard et al. [10] summarized the techniques and models of precision sow feeding. Zhou and Hooker [11] developed innovative decision tree models with varying coefficients, significantly improving prediction accuracy. Li and Ma [12] showcased the practical implementation of AI techniques in poultry farming, achieving notable improvements in FCR optimization.
The theoretical foundation for these advances was strengthened by Montgomery et al. [13], who provided the statistical framework for regression analysis in agricultural applications. Sharma [7] contributed valuable insights into random forest applications in agricultural settings, while Cortes and Vapnik [14] established the fundamental principles of support vector machines that are now widely used in FCR prediction.
Fu et al. [15] developed a comprehensive data collection and modeling framework specifically for feed efficiency prediction. LeCun [16] and Piles et al. [17] established the theoretical foundations for deep learning applications in agricultural settings. Shirzadifar [18] conducted a systematic review of machine learning applications in livestock farming, identifying key challenges and opportunities.
The integration of advanced technologies has further enhanced our capabilities. Johansen et al. [19] demonstrated the successful application of neural networks in chicken FCR prediction, Le Boucher et al. [20] were among the first to apply Bayesian methods for predicting the feed intake of seabass. Ellis [21] developed specialized deep learning architectures for precision farming applications.
Recent methodological advances include Monizet al.’s [22] work on Gradient Boosting for mixed models and Fahrurrozi’s [23] development of interpretable machine learning frameworks. Mishra et al. [24] made significant contributions by integrating IoT data with machine learning models, while Cravero et al. [25] combined genetic algorithms with machine learning for improved FCR prediction.
The field has seen rapid advancement in analytical capabilities. Davison et al. [4] optimized feed management through advanced predictive analytics, while Koppe et al. [26] expanded the applications of data mining in agricultural modeling.
Comprehensive reviews and methodological advances have shaped current understanding. Technical innovations continue to emerge, with Griesbach et al. [27] optimizing Gradient Boosting parameters for time series forecasting, and Konstantinov et al. [28] proposed a local and global black box model interpretation method based on a generalized additive model. It can be seen as an extension or modification of the algorithm using neural addition models. Cheng et al. [29] contributed to the methodological framework with their work on supervised learning and random forest applications.
Recent advances in modeling complex systems, demonstrated by Koppe et al. [26] and Kriegeskorte [30], have provided new tools for understanding FCR dynamics. Goodfellow et al. [31], Sutton and Barto [32], and Gu et al. [33] established fundamental principles for advanced machine learning applications. Benos [34] provided a comprehensive review of current applications, while Welper [35] contributed important insights into neural network training optimization. Certainly, the transformation of wearable devices also provides more comprehensive data collection methods for animal breeding and animal management [36,37].
In animal breeding programs, determining shorter feed intake measurement intervals and estimating their reliability is crucial for enhancing selection accuracy for target traits and reducing costs. This is particularly significant for finishing-stage pigs, as shorter feed intake measurement intervals require less single-space feeder time. With a fixed number of feeders, operations can test more batches of pigs. This study aims to evaluate the predictive reliability of individual feed conversion ratios (FCRs) using large-scale commercial feed intake records, specifically assessing the impact of different test intervals and starting points. Of particular interest are intervals where the starting body weight reaches at least 40 kg—a key transition phase from the early growing period to the rapid growth phase of finishing. Understanding pig growth phases is crucial for interpreting our findings. Pigs typically undergo distinct growth stages: early growth (weaning to ~40 kg), rapid growth phase (~40–90 kg), and finishing phase (~90 kg to market weight). The 40 kg threshold identified in our study corresponds to the transition from the early to the rapid growth phase, when feed conversion patterns become more stable and predictable.
This study aims to build upon these foundations by developing an integrated approach to FCR prediction. Using 438,552 feed samples from two core farms in Sichuan Province, we utilize nineteen machine learning algorithms to explore the feasibility of predicting long-term FCRs from short-term data. Our work synthesizes the methodological advances described above while addressing the specific challenges of FCR prediction in practical farming environments.

2. Materials and Methods

The basic experimental process is shown in Figure 1 below:

2.1. Animal and Data Collection

FCR data were collected from Osborne feed testing stations installed at two core breeding farms in Sichuan Province, China. Feeding data were collected using the same Osborne system model across both farms to ensure uniformity in methodology and contribute to consistent rearing conditions. Upon entry of each pig into the feed testing station, the system initiated the collection of data. The dataset comprises 438,552 records of individual feeding visits, encompassing data from 622 individual pigs. Each record includes animal ID, body weight at the time of visit, and feed intake during the visit. Pig body weights ranged from 20 to 180 kg.

2.2. Data Preprocessing

Prior to analysis, we performed a series of preprocessing steps to ensure data quality:
Step 1—Data cleaning and 5 kg division: The core steps of feed data aggregation were carried out using the daily time dimension while considering the continuity of individual pigs, which specifically included the following: (1) Data preprocessing and outlier removal: we took the feeding event of a single pig (each record of entering the feeding station) as the basic unit and removed unreasonable data through dynamic outlier detection, such as weight anomalies (marked as abnormal if the current weight deviated from the 3-day rolling median by more than 20% or 3 standard deviations) and feeding rate anomalies (defined as abnormal if the feeding rate exceeded 10 g/s), and the output was the cleaned single feeding event record. (2) Daily summary aggregation: grouping by animal_number and visit_date, calculating the daily average weight (daily_avg_weight) and the last value of the day (daily_last_weight) for weight, the sum of all feeding events on that day (daily_total_feed) for feed intake, the total duration of all feeding events on that day (daily_total_duration) for feeding time, the average value of all records on that day (daily_avg_feed_rate) for feeding rate, and the first recorded value (gender_first) for static attributes such as gender. (3) Trend verification (validate_animals function): further removing abnormal pigs by checking whether the weight dropped by more than 5% for two consecutive days and whether the feed conversion ratio (total cumulative feed intake/total weight gain) was lower than 2.0 (assuming FCR should not be lower than 2.0). For the distribution characteristics of feed samples, the original data contained 438,552 feeding event records (code input data), which were reduced to daily single-pig summary records after daily aggregation. Each summary record included 6 core indicators (average weight, last weight, total feed intake, total time, average rate, and gender). Finally, the data were aggregated according to the pig weight stage of 5 kg, and the large-interval FCR was calculated.
Step 2—Removing missing values: Rows with any missing values were removed from the dataset. This approach is effective when the proportion of missing data is small (5% in our dataset).
Step 3—Eliminating outliers beyond three standard deviations and normalization: The mean (μ) and standard deviation (σ) of the dataset were calculated. Features were normalized to place them on the same scale of measurement. The formula used was
z = x u s
where x is the current sample value, and μ and σ are the mean and standard deviation of all samples, respectively.
Step 4—Feature standardization: Features were standardized by subtracting the mean and dividing by the standard deviation, resulting in a distribution with a mean of 0 and a standard deviation of 1. This enables optimization algorithms like gradient descent to converge faster [11]. Models such as SVM, Linear Regression, and logistic regression perform better on standardized data. Feature scaling can also reduce model training time by eliminating the influence of dimensionality. The formula used is
y = x x   m i n ( x   m a x x   m i n ) 1
where x represents the sample value, x min and x max represent the minimum and maximum values of all samples, respectively, and y is the scaled value of the feature.
After quality control, the dataset showed a mean of 2.1090, a median of 2.1021, a maximum of 2.4114, a minimum of 1.7187, a standard deviation of 0.1330, and a coefficient of variation of 6.3049%.
These preprocessing steps are particularly important for distance-based algorithms (SVM, K-NN), gradient-based optimization (neural networks), and regularized linear models (Ridge, Lasso) which are sensitive to feature scaling.

2.3. Dataset Splitting

The data were divided into a training set and a validation set in a 7:3 ratio. To assess the impact of input FCR interval size on long-term FCR prediction, the validation set data were partitioned at intervals of 5 kg, 10 kg, 15 kg, 20 kg, 25 kg, 30 kg, 35 kg, and 40 kg.

2.4. Machine Learning Algorithms

In our study, we systematically compare a variety of machine learning models to assess their performance in handling both discrete and continuous data types. This comparative analysis involves nineteen distinct models, each chosen for its unique capabilities and theoretical underpinnings. These models are categorized based on their primary functions and methodologies.
We begin with fundamental linear models such as Linear Regression, Ridge Regression, Lasso, and ElasticNet. These models are chosen for their simplicity and effectiveness in linear approximation tasks [13]. Ridge and Lasso regression address issues of multicollinearity among independent variables, yielding more stable parameter estimates. Meanwhile, Lasso regression and ElasticNet achieve feature selection through penalty terms, which helps reduce model complexity and improve interpretability.
Additionally, we include Bayesian Ridge Regression, which incorporates Bayesian reasoning principles. This model provides a probabilistic approach to Linear Regression by offering uncertainty estimates for parameter estimates through posterior distributions, enabling the calculation of confidence intervals for predictions. By integrating prior distributions, Bayesian Ridge Regression effectively mitigates overfitting, particularly in scenarios with small datasets or high-dimensional feature spaces. Although Bayesian computations are generally more complex, the implementation of Bayesian Ridge Regression remains relatively straightforward and can be efficiently solved using existing computational tools.
In the realm of ensemble methods, our study incorporates random forest, Gradient Boosting, AdaBoost, Extra Trees, and specialized Gradient Boosting machines (LightGBM, XGBoost, and CatBoost). These models are known for their robustness and ability to improve predictive accuracy through techniques such as bagging and boosting.
We also examine traditional support vector machines, which are binary classification models that map input data to a high-dimensional space and find an optimal hyperplane for classification tasks. Furthermore, we explore kernel-based methods, including Support Vector Regression (SVR), NuSVR, and Linear SVR. These are crucial for their capacity to model non-linear relationships using different kernel functions.
For a non-parametric perspective, we utilize the decision tree model, valued for its interpretability and simplicity [35]. We also employ the K-Nearest Neighbors (K-NN) algorithm, which makes predictions based on the proximity to its neighbors.
Additionally, we investigate the Gaussian Process model, which offers insights into the uncertainty of predictions, and the Multi-Layer Perceptron (MLP), a basic form of neural networks. These models are evaluated for their effectiveness in more complex and high-dimensional spaces [16,17].
This comprehensive suite of models enables a thorough comparison and evaluation, allowing us to determine the most effective techniques for various types of data and modeling challenges.
The Gradient Boosting model, which exhibited the best performance in this study, is an ensemble model that combines multiple weak learners (usually decision trees) to create a strong learner. It builds the model in a stage-wise fashion, with each new tree attempting to correct the errors made by the previous trees. The model is trained using a differentiable loss function and gradient descent optimization [22]. These 19 algorithms were selected based on their diverse approaches to data modeling, ranging from simple linear models to complex ensemble methods, to comprehensively evaluate different machine learning techniques for FCR prediction.
Although the above algorithms are designed to directly handle continuous numerical inputs, in this specific study on feed conversion ratio (FCR) prediction, we preprocessed the continuous variable of body weight through binning (at 5 kg intervals), which made this feature appear as an ordered categorical variable (ordinal discrete data) when input into the model.
This binning preprocessing is primarily based on the following two considerations:
(1)
Domain Practice and Interpretability: In livestock management and animal nutrition research, animal body weight stages (e.g., <5 kg, 5–10 kg, 10–15 kg, etc.) are typically closely associated with specific feeding strategies, nutritional requirements, and health management. Converting body weight into explicit stage categories facilitates the interpretability of model predictions (i.e., FCR values for specific weight stages) by domain experts and their application in practical production decisions.
(2)
Feature Engineering and Mitigation of Potential Issues: Body weight data may exhibit skewed distributions or contain potential measurement outliers. Binning at 5 kg intervals can mitigate the impact of these issues and may help simplify the learning boundaries of certain complex models (such as tree-based ensemble methods). The superior performance of Gradient Boosting in this study partially validates the effectiveness of this feature representation in capturing the relationship between weight stages and FCR.
It should be noted that while the underlying algorithms can handle continuous data, our choice to bin weight data into 5 kg intervals was a deliberate preprocessing decision based on domain expertise and practical considerations in livestock management, where weight-based feeding stages are commonly used in commercial operations.
Hyperparameter optimization was performed using a grid search with cross-validation for all models to ensure fair comparison. The optimal hyperparameter configurations are detailed in Appendix A. For specific hyperparameter settings, please refer to Tables S1 and S2.

2.5. Model Evaluation

All analyses were performed using Python 3.8.18 with the scikit-learn 1.3.2 library. Four basic metrics were used to evaluate the models’ performance:
R 2 = 1 i = 1 n ( y i f i ) 2 i = 1 n ( y i y _ ) 2 R M S E = 1 n i = 1 n ( y i f i ) 2 , M A E = 1 n i = 1 n | y i f i | , M A P E = 1 n i = 1 n | y i f i y i | ,
where n is the number of samples in the data, and y_mean is the average of the samples.
Additionally, the Pearson correlation coefficient was used to evaluate the correlation between predicted and true values, indirectly reflecting the predictive performance of the model. It measures the linear relationship between two quantitative variables, requiring the data to be continuous and at least approximately normally distributed. The Pearson coefficient is sensitive to outliers. The formula is:
r = ( ( x i ( x ¯ ) ) ( y i ( y ¯ ) ) ) ( ( x i ( x ) ¯ ) 2 ( y i ( y ) ¯ ) 2

3. Results

Having established our methodology, we now turn to the results of our analysis, which demonstrate the efficacy of our approach in predicting long-term FCR using short-term data.

3.1. Data Preprocessing Results

After handling missing values and performing data validation, individual data from 138 pigs with weights ranging between 30 and 105 kg were retained. The FCR was calculated for each 5 kg interval, as well as the overall FCR for the entire 30–105 kg range.

3.2. Model Training

Among the 19 models tested, Gradient Boosting, LightGBM, and CatBoost consistently outperformed others. Therefore, we focus on these three models for further analysis. The model training employed a grid search strategy for hyperparameter optimization to identify the best-performing configuration for each model. The optimal hyperparameters identified through this process are detailed in Appendix A Figure A1. For specific hyperparameter settings, please refer to Tables S1 and S2.
The model training results are presented in Table 1. Among the models tested, Gradient Boosting, LightGBM, and CatBoost achieved the highest performance, with R2 values around 0.5. The Gradient Boosting model demonstrated the best overall performance (R2 = 0.51, RMSE = 0.09, MAE = 0.07, MAPE = 0.03). The reason some other models performed poorly may be due to the relatively complex linear relationships in the data, which limited the effectiveness of models such as Linear Regression, Ridge Regression, and Lasso regression. Additionally, some models may suffer from overfitting, particularly in cases of high-dimensional data or insufficient feature selection, leading to their poor performance on the test set.

3.3. Model Prediction Performance Under Different Interval Sizes

The trained Gradient Boosting model was used to evaluate the impact of different FCR interval sizes on 30~105 kg FCR prediction, using the validation set (Table 2). When the interval reached 40 kg, the model achieved an R2 of 0.72, and the correlation between predicted and actual values reached 0.85.
The trained LightGBM model was also evaluated using the validation set (Table 3). When the interval reached 40 kg, the model achieved an R2 of 0.71 and a correlation of 0.84.
Similarly, the trained CatBoost model was evaluated (Table 4). When the interval reached 40 kg, the model achieved an R2 of 0.72 and a correlation of 0.85.
Figure 2 shows the R2 curves for the Gradient Boosting, LightGBM, and CatBoost models. As the prediction interval expands from smaller intervals (every 5 kg) to larger ones, the models’ R2 values improve significantly.

3.4. Optimal Prediction Interval with a Fixed Interval Size of 40 kg

Table 5, Table 6 and Table 7 present the prediction performance of the Gradient Boosting, LightGBM, and CatBoost models, respectively, for different starting weights while maintaining a consistent interval size of 40 kg. The analysis reveals that all three models demonstrate excellent predictive performance, particularly in the weight ranges of 50–90 kg and 65–105 kg.
Notably, the results indicate that measuring feed intake for pigs between 50 and 90 kg is most effective for predicting overall FCR in actual production scenarios. This finding has significant implications for optimizing data collection practices in livestock management.

4. Discussion

This study employs machine learning models to predict long-term (30–105 kg) FCR using short-term FCR data and provides a detailed evaluation of these models’ performance. The Gradient Boosting model outperforms all other models, achieving an R2 value of 0.72 when the prediction interval reaches 40 kg. Using data from the 50–90 kg and 65–105 kg weight ranges with a 40 kg prediction interval ensures the model’s accuracy and reliability.

4.1. Impact of Prediction Interval Size on FCR

The results demonstrate that as the prediction interval expands from smaller intervals (every 5 kg) to larger ones, the R2 values of the models improve significantly. For instance, from 5 kg to 40 kg, the R2 values of the Gradient Boosting model increase from 0.06 to 0.72. This indicates that larger prediction intervals provide more data aggregation, helping the model capture more stable and significant trends while reducing noise in the predictions.
Within smaller weight intervals, individual differences and random fluctuations may be more pronounced, potentially interfering with model predictions. In larger weight intervals, these short-term fluctuations are averaged out, allowing the model to better learn and predict long-term trends.
The study also observes that as the interval size increases, the correlation between predicted and actual values improves. This further confirms that larger intervals are more effective at simulating and reflecting actual growth and feeding responses during prediction.
The horizontal clusters observed in Figure 3, particularly evident in the CatBoost model, are characteristic of tree-based ensemble methods. These horizontal clusterings result from the decision tree splitting logic and the discrete nature of our 5 kg weight interval binning, representing how tree-based algorithms partition the feature space rather than indicating model deficiencies.

4.2. Impact of Different Prediction Intervals with the Same Interval Size

The 50–90 kg and 65–105 kg weight ranges exhibit the best model prediction performance. Since the 65–105 kg range is already in the later stages of pig growth in this model, it has limited significance in actual production. In practice, measuring feed intake for pigs between 50 and 90 kg, which focuses on the mid-growth stage, is most important for animal feeding and pathological management. Accurate FCR prediction can significantly impact cost-effectiveness, particularly in pig farming. While our method demonstrates strong predictive power, it is important to note its limitations. The model’s performance may vary in different farming environments and with different pig breeds. Future studies should validate these findings across diverse settings.

4.3. Biological Factors Affecting FCR

The biological processes governing growth and feed efficiency can vary significantly across different weight ranges. In the 50–90 kg range, animals tend to be in a phase of rapid growth, where their metabolism is optimized for feed conversion. During this stage, the relationship between feed intake and weight gain is more linear and predictable, making it easier for machine learning models to learn from historical data. Conversely, outside this range, factors such as reduced growth rates, changes in dietary needs, or the onset of maturity may introduce more variability in FCR, making predictions less reliable.

4.4. Data Availability and Quality

The dataset utilized in this study likely reflects more consistent and comprehensive records for the 50–90 kg range. Data quality plays a crucial role in the performance of machine learning models. As animals grow, variations in feed composition, management practices, and environmental factors may lead to inconsistent data outside the 50–90 kg range. Therefore, models trained on data from this interval can achieve higher accuracy due to the inherent consistency in the growth dynamics and feeding practices observed within this weight category.

4.5. Practical Implications for Production

Understanding that FCR predictions are most reliable in the 50–90 kg range has significant implications for production practices:
Optimized Feeding Strategies: Producers can focus on implementing targeted feeding strategies during this critical growth phase to maximize feed efficiency. By closely monitoring FCR in this range, adjustments can be made to feeding regimens in real time, ensuring that animals receive the appropriate nutrients to maintain optimal growth rates.
Resource Allocation: With accurate FCR predictions, producers can allocate resources more effectively. This includes better planning of feed purchases and inventory management, ultimately leading to cost savings and reduced waste.
Predictive Management: By utilizing machine learning models for ongoing FCR monitoring, producers can establish predictive management practices that anticipate growth patterns. This proactive approach can mitigate issues related to underperformance or health problems that may arise outside the optimal weight range.

5. Conclusions and Future Directions

This study demonstrates the feasibility of using machine learning techniques, particularly the Gradient Boosting model, to predict long-term FCR based on short-term FCR data. The findings highlight the importance of appropriate data preprocessing, model selection, and interval size determination for optimal prediction performance. The insights gained from this research can guide precision livestock farming practices, leading to more efficient resource management and sustainable animal husbandry. Real-time analysis of short-term feed conversion ratio (FCR) data can be conducted through data recorded at feed testing stations. Livestock producers can immediately understand the current feed efficiency, and predict long-term FCR trends based on short-term changes. This allows for proactive management strategies, such as adjusting immunization programs or modifying feed management practices. To ensure the accuracy of FCR predictions, it is still necessary to regularly evaluate model performance and adjust model parameters based on feedback, as well as to remove any anomalous data. These findings have immediate practical applications for commercial pig production, enabling producers to optimize feed efficiency monitoring while reducing equipment occupancy time and costs.
However, this study has certain limitations: the dataset used is from a specific region, potentially lacking complete representation of the diversity within pig farming environments. Subsequent research could incorporate data from multiple sources to improve the generalizability of the models.
While the Gradient Boosting model performed best in this study, other advanced machine learning algorithms, such as deep learning models, could be explored to potentially further enhance prediction accuracy. Incorporating diverse farming environments and breeds into the model training will enhance its robustness when faced with new data. Additionally, algorithms such as Gradient Boosting, with their powerful feature learning capabilities, can identify key factors influencing FCR across different settings, thereby improving performance in varied scenarios. Furthermore, considering changes in farming conditions, such as feed types and husbandry methods, the application of transfer learning can effectively bolster the model’s generalization ability. Therefore, future research should further validate the model’s performance across different farming environments to ensure its broad applicability in real-world production settings. Model Optimization: Investigate hyperparameter tuning and ensemble methods to refine the existing models for better performance. To further advance the research in this domain and contribute to the enhancement of global animal husbandry practices, several promising avenues for future work can be pursued. One valuable direction is to examine the impact of various environmental and management factors on pig farming outcomes more comprehensively. This could involve potentially integrating IoT data for real-time analysis, which would enable a more dynamic understanding of how different elements in the farming environment interact and influence results. Another important aspect is the exploration of different algorithms. Future research can evaluate the effectiveness of a wider range of machine learning techniques, such as neural networks and support vector machines. By doing so, it becomes possible to identify which approaches yield the best results in diverse contexts, considering the unique characteristics and requirements of pig farming scenarios. By building upon the foundation established in this study through these kinds of explorations, future research can make significant contributions to enhancing the efficiency and sustainability of global animal husbandry practices. Ultimately, this would lead to better outcomes for both farmers and livestock, fostering a more productive and environmentally friendly agricultural sector.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ani15121773/s1, Table S1: “best_hyperparameters” shows the model hyperparameter settings; Table S2: “optimized_model_results” presents the pre-training results of the model during grid search.

Author Contributions

Conceptualization, X.Y. and L.Z. (Liangyu Zhu); methodology, Y.Y.; software, W.J.; validation, X.Y.; formal analysis, X.Y., M.G. and L.Z. (Liangyu Zhu); investigation, Y.Y.; resources, L.S. and M.G.; data curation, X.Y.; writing—original draft preparation, X.Y. and L.Z. (Liangyu Zhu); writing—review and editing, L.S. and L.Z. (Li Zhu); visualization, X.Y. and M.G.; supervision, M.G.; project administration, L.Z. (Li Zhu); funding acquisition, L.S. and L.Z. (Li Zhu). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by STI 2030-Major Projects (2023ZD04045); the Sichuan Science and Technology Program (2021ZDZX0008, 2021YFYZ0030); the China Agriculture Research System (CARS-35); The Program for Pig Industry Technology System Innovation Team of Sichuan Province (SCCXTD-2025-8).

Institutional Review Board Statement

All animal experimental procedures were formally approved by the Animal Ethics and Welfare Committee of Sichuan Agricultural University, Chengdu, China (Approval number 2021302137, approved in March 2021).

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article; further inquiries can be directed to the corresponding author. Code availability for access: https://github.com/DeepBLUP/FCR-Predication.git (accessed on 3 June 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Model Training Parameter Settings In this study, the main parameter settings are as follows:
Figure A1. Model Training Parameter Settings.
Figure A1. Model Training Parameter Settings.
Animals 15 01773 g0a1
The selection of these parameters is based on the following theoretical and empirical considerations:
Learning rate: This parameter controls the step size of each iteration, typically ranging between 0.01 and 0.1 [27]. A rate that is too small may cause slow convergence, while one that is too large may prevent convergence to an optimal solution.
Model complexity: The number of iterations (n_estimators) or maximum tree depth (max_depth) limits the model’s complexity to prevent overfitting. Generally, shallower trees and fewer iterations are sufficient [28].
Tree growth control: Parameters such as the minimum number of samples required for leaf nodes (min_samples_split, min_data_in_leaf) and subsample ratio help control the growth of each tree. Appropriate settings can improve generalization.
Regularization: Parameters such as l2_leaf_reg can limit the size of leaf node output values and reduce the risk of overfitting.
To optimize these hyperparameters, we employed grid search and cross-validation techniques. This approach helps balance training errors and generalization errors [36]. The meticulous selection and optimization of these parameters allow the model to thoroughly learn the intrinsic patterns of the data while preventing the memorization of noise. The robust performance of the final model on unseen data indicates its excellent generalization capabilities.
In summary, this study introduces a novel method based on machine learning techniques to predict long-term feed conversion ratio (FCR) utilizing short-term FCR data. Empirical studies have demonstrated that this method can effectively improve the prediction accuracy of feed conversion rate, serving as a robust tool for precision livestock farming. We anticipate that this study will make a significant contribution to the sustainable development of animal husbandry practices.

References

  1. Petrović, B.; Bumbálek, R.; Zoubek, T.; Kuneš, R.; Smutný, L.; Bartoš, P. Application of precision agriculture technologies in Central Europe-review. J. Agric. Food Res. 2024, 15, 101048. [Google Scholar] [CrossRef]
  2. Yamsani, N.; Muthukumaran, K.; Kumar, B.S.; Asha, V.; Singh, N.; Dhanraj, J. IoT-Based Livestock Monitoring and Management System Using Machine Learning Algorithms. In Proceedings of the 2024 International Conference on Science Technology Engineering and Management (ICSTEM), Putrajaya, Malaysia, 19–20 December 2024; pp. 1–6. [Google Scholar]
  3. Prakash, A.; Saxena, V.K.; Singh, M.K. Genetic analysis of residual feed intake, feed conversion ratio and related growth parameters in broiler chicken: A review. World’s Poult. Sci. J. 2020, 76, 304–317. [Google Scholar] [CrossRef]
  4. Davison, C.; Michie, C.; Tachtatzis, C.; Andonovic, I.; Bowen, J.; Duthie, C.A. Feed Conversion Ratio (FCR) and Performance Group Estimation Based on Predicted Feed Intake for the Optimisation of Beef Production. Sensors 2023, 23, 4621. [Google Scholar] [CrossRef] [PubMed]
  5. France, J.; Theodorou, M.; Lowman, R.; Beever, D.J.F.S. Feed evaluation models. In Feed Evaluation for Animal Production; Theodorou, M., France, J., Eds.; CAB International: Oxfordshire, UK, 2000; pp. 1–9. [Google Scholar]
  6. Morota, G.; Ventura, R.V.; Silva, F.F.; Koyama, M.; Fernando, S.C. Big data analytics and precision animal agriculture symposium: Machine learning and data mining advance predictive big data analysis in precision animal agriculture. J. Anim. Sci. 2018, 96, 1540–1550. [Google Scholar] [CrossRef]
  7. Sharma, A.; Jain, A.; Gupta, P.; Chowdary, V.J.I.A. Machine learning applications for precision agriculture: A comprehensive review. IEEE Access 2020, 9, 4843–4873. [Google Scholar] [CrossRef]
  8. Mahmud, M.S.; Zahid, A.; Das, A.K.; Muzammil, M.; Khan, M.U.J.C. A systematic literature review on deep learning applications for precision cattle farming. Comput. Electron. Agric. 2021, 187, 106313. [Google Scholar] [CrossRef]
  9. Fan, Z.; Mengting, Z.; Benhai, X.; Zhengang, Y.; Minze, L.; Wenxiao, F.; Xiangfang, T.J.S.A. Research Advances and Prospect of Intelligent Monitoring Systems for the Physiological Indicators of Beef Cattle. Agriscience 2024, 6, 1. [Google Scholar]
  10. Gaillard, C.; Brossard, L.; Dourmad, J.-Y.J.A.F.S. Improvement of feed and nutrient efficiency in pig production through precision feeding. Technology 2020, 268, 114611. [Google Scholar] [CrossRef]
  11. Zhou, Y.; Hooker, G.J.D.M.; Discovery, K. Decision tree boosted varying coefficient models. Data Min. Knowl. Discov. 2022, 36, 2237–2271. [Google Scholar] [CrossRef]
  12. Li, Y.; Ma, R.; Qi, R.; Li, J.; Liu, W.; Wan, Y.; Li, S.; Sun, Z.; Xu, J.; Zhan, K.J.P.S. Novel insight into the feed conversion ratio in laying hens and construction of its prediction model. Poult. Sci. 2024, 103, 104013. [Google Scholar] [CrossRef]
  13. Montgomery, D.C.; Peck, E.A.; Vining, G.G. Introduction to Linear Regression Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  14. Cortes, C.J.M.L. Support-Vector Networks. In Maching Learning; Springer: Berlin/Heidelberg, Germany, 1995. [Google Scholar]
  15. Fu, Q.; Shen, W.; Wei, X.; Zhang, Y.; Xin, H.; Su, Z.; Zhao, C.J.C. Prediction of the diet energy digestion using kernel extreme learning machine: A case study with Holstein dry cows. Comput. Electron. Agric. 2020, 169, 105231. [Google Scholar] [CrossRef]
  16. LeCun, Y.; Bengio, Y.; Hinton, G.J.N. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  17. Piles, M.; Tusell, L.; Velasco-Galilea, M.; Helies, V.; Drouilhet, L.; Zemb, O.; Sánchez, J.; Garreau, H. Machine Learning Algorithms for the Prediction of Feed Efficiency Based on Caecal Microbiota. In Proceedings of the12th World Rabbit Congress, Nantes, France, 1–3 July 2021. [Google Scholar]
  18. Shirzadifar, A.; Manafiazar, G.; Davoudi, P.; Do, D.; Hu, G.; Miar, Y.J.A. Prediction of growth and feed efficiency in mink using machine learning algorithms. Animal 2024, 19, 101330. [Google Scholar] [CrossRef]
  19. Johansen, S.V. Automated Optimization of Broiler Production. Ph.D. Thesis, Aalborg Universitetsforlag, Aalborg, Denmark, 2019. [Google Scholar]
  20. Le Boucher, R.; Chung, W.; Lin, J.N.K.; Tan, L.S.E.; Poon, Z.W.J.; Lee, C.S.J.A. Optimizing automated feeder algorithms for feed intake prediction in barramundi (Lates calcarifer). Aquaculture 2023, 577, 739909. [Google Scholar] [CrossRef]
  21. Ellis, J.; Jacobs, M.; Dijkstra, J.; Van Laar, H.; Cant, J.; Tulpan, D.; Ferguson, N.J.A. Synergy between mechanistic modelling and data-driven models for modern animal production systems in the era of big data. Animal 2020, 14, S223–S237. [Google Scholar] [CrossRef]
  22. Moniz, N.; Vale, Z.; Cascalho, J.; Silva, C.; Sebastião, R. Progress in Artificial Intelligence. In Proceedings of the 22nd EPIA Conference on Artificial Intelligence, EPIA 2023, Faial Island, Portugal, 5–8 September 2023; Proceedings, Part II. Springer Nature: Berlin/Heidelberg, Germany, 2024; Volume 14116. [Google Scholar]
  23. Fahrurrozi, I.; Wahyono, W.; Sari, A.K.; Usuman, I.; Ariyadi, B. Integrating random forest model and internet of things-based sensor for smart poultry farm monitoring system. Indones. J. Electr. Eng. Comput. Sci. 2024, 33, 1283–1292. [Google Scholar] [CrossRef]
  24. Mishra, S.J.H. Internet of things enabled deep learning methods using unmanned aerial vehicles enabled integrated farm management. Heliyon 2023, 9, e18659. [Google Scholar] [CrossRef]
  25. Cravero, A.; Pardo, S.; Sepúlveda, S.; Muñoz, L.J.A. Challenges to use machine learning in agricultural big data: A systematic literature review. Agronomy 2022, 12, 748. [Google Scholar] [CrossRef]
  26. Koppe, G.; Toutounji, H.; Kirsch, P.; Lis, S.; Durstewitz, D.; Isik, L. Identifying nonlinear dynamical systems via generative recurrent neural networks with applications to, fMRI. PLoS Comput. Biol. 2019, 15, e1007263. [Google Scholar] [CrossRef]
  27. Griesbach, C.; Säfken, B.; Waldmann, E. Gradient boosting for linear mixed models. Int. J. Biostat. 2021, 17, 317–329. [Google Scholar] [CrossRef]
  28. Konstantinov, A.V.; Utkin, L.V. Interpretable machine learning with an ensemble of gradient boosting machines. Knowl.-Based Syst. 2021, 222, 106993. [Google Scholar] [CrossRef]
  29. Cheng, L.; Chen, X.; De Vos, J.; Lai, X.; Witlox, F. Applying a random forest method approach to model travel mode choice behavior. Travel Behav. Soc. 2019, 14, 1–10. [Google Scholar] [CrossRef]
  30. Kriegeskorte, N. Deep neural networks: A new framework for modeling biological vision and brain information processing. Annu. Rev. Vis. Sci. 2015, 1, 417–446. [Google Scholar] [CrossRef]
  31. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y.J.A. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  32. Sutton, R.S.; Barto, A.G.J.C. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998; Volume 22447, p. 10. [Google Scholar]
  33. Gu, S.; Holly, E.; Lillicrap, T.; Levine, S. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, Singapore, 29 May–3 June 2017; pp. 3389–3396. [Google Scholar]
  34. Benos, L.; Tagarakis, A.C.; Dolias, G.; Berruto, R.; Kateris, D.; Bochtis, D.J.S. Machine learning in agriculture: A comprehensive updated review. Sensors 2021, 21, 3758. [Google Scholar] [CrossRef]
  35. Welper, G.J.N.N. Universality of gradient descent neural network training. Neural Netw. 2022, 150, 259–273. [Google Scholar] [CrossRef]
  36. Wen, C.; Yan, W.; Zheng, J.; Ji, C.; Zhang, D.; Sun, C.; Yang, N. Feed efficiency measures and their relationships with production and meat traits traits in slower growing broilers. Poult. Sci. 2018, 97, 2356–2364. [Google Scholar] [CrossRef]
  37. Lamanna, M.; Ovoovo, M.; Cavallini, D. Wearable Collar Technologies for Dairy Cows: A Systematized Review of the Current Applications and Future Innovations in Precision Livestock Farming. Animals 2025, 15, 458. [Google Scholar] [CrossRef]
Figure 1. FCR experimental flowchart.
Figure 1. FCR experimental flowchart.
Animals 15 01773 g001
Figure 2. R2 curves for Gradient Boosting, LightGBM, and CatBoost models.
Figure 2. R2 curves for Gradient Boosting, LightGBM, and CatBoost models.
Animals 15 01773 g002
Figure 3. Predicted scatter plots of Gradient Boosting, LightGBM, and CatBoost models.
Figure 3. Predicted scatter plots of Gradient Boosting, LightGBM, and CatBoost models.
Animals 15 01773 g003
Table 1. Prediction performance of different models for 30–105 kg FCR.
Table 1. Prediction performance of different models for 30–105 kg FCR.
ModelR2RMSEMAEMAPE
Linear Regression0.360.11 0.080.04
Random Forest0.410.10 0.080.04
Gradient Boosting0.510.09 0.070.03
SVR0.220.12 0.100.05
Ridge0.360.11 0.080.04
Lasso0.000.13 0.110.05
ElasticNet0.000.13 0.110.05
Bayesian Ridge0.360.11 0.080.04
Decision Tree0.070.13 0.100.05
AdaBoost0.370.11 0.080.04
Extra Trees0.270.11 0.090.04
K-Nearest Neighbors0.420.10 0.080.04
MLP0.420.10 0.080.04
Linear SVR0.250.12 0.090.04
NuSVR0.240.12 0.090.04
Gaussian Process−3.810.29 0.090.04
LightGBM0.510.09 0.070.03
XGBoost0.460.10 0.070.04
CatBoost0.500.09 0.070.03
Table 2. Prediction performance of the Gradient Boosting model for 30–105 kg FCR under different intervals.
Table 2. Prediction performance of the Gradient Boosting model for 30–105 kg FCR under different intervals.
R2Pearson_Corr.RMSEMAEMAPE
50.060.290.130.110.05
100.220.470.110.090.04
150.390.630.090.080.04
200.470.700.100.080.04
250.540.740.090.070.03
300.610.780.080.070.03
350.650.810.080.060.03
400.720.850.070.050.02
450.770.880.060.050.02
500.790.890.060.050.02
550.800.900.060.040.02
600.860.930.050.040.02
650.860.940.050.040.02
700.930.980.040.030.01
Table 3. Prediction performance of the LightGBM model under different intervals.
Table 3. Prediction performance of the LightGBM model under different intervals.
R2Pearson_Corr.RMSEMAEMAPE
50.09 0.31 0.13 0.10 0.05
100.18 0.44 0.11 0.09 0.04
150.37 0.61 0.10 0.08 0.04
200.49 0.71 0.10 0.07 0.04
250.53 0.74 0.09 0.07 0.03
300.63 0.80 0.08 0.06 0.03
350.67 0.82 0.08 0.06 0.03
400.71 0.84 0.07 0.05 0.02
450.78 0.88 0.06 0.05 0.02
500.79 0.89 0.06 0.05 0.02
550.80 0.90 0.06 0.04 0.02
600.87 0.93 0.05 0.04 0.02
650.88 0.94 0.05 0.03 0.02
700.96 0.98 0.03 0.02 0.01
Table 4. Prediction performance of the CatBoost model under different intervals.
Table 4. Prediction performance of the CatBoost model under different intervals.
R2Pearson_Corr.RMSEMAEMAPE
50.060.280.13 0.11 0.05
100.160.420.12 0.09 0.04
150.380.620.10 0.08 0.04
200.480.700.10 0.08 0.04
250.530.730.09 0.07 0.03
300.620.790.08 0.07 0.03
350.660.810.08 0.06 0.03
400.720.850.07 0.05 0.02
450.780.880.06 0.05 0.02
500.790.890.06 0.05 0.02
550.80 0.900.06 0.04 0.02
600.80.930.05 0.04 0.02
650.900.950.04 0.03 0.02
700.950.970.03 0.02 0.01
Table 5. Prediction performance of the Gradient Boosting model for different starting weights with a 40 kg interval.
Table 5. Prediction performance of the Gradient Boosting model for different starting weights with a 40 kg interval.
R2Pearson_Corr.RMSEMAEMAPE
30–700.53 0.73 0.08 0.06 0.03
35–750.66 0.82 0.09 0.07 0.03
40–800.60 0.78 0.08 0.06 0.03
45–850.71 0.84 0.07 0.05 0.02
50–900.81 0.90 0.06 0.05 0.02
55–950.72 0.87 0.06 0.05 0.02
60–1000.74 0.87 0.06 0.05 0.02
65–1050.85 0.93 0.05 0.04 0.02
Table 6. Prediction performance of the LightGBM model for different starting weights with a 40 kg interval.
Table 6. Prediction performance of the LightGBM model for different starting weights with a 40 kg interval.
R2Pearson_Corr.RMSEMAEMAPE
30–700.49 0.72 0.08 0.06 0.03
35–750.69 0.84 0.08 0.07 0.03
40–800.57 0.76 0.08 0.06 0.03
45–850.68 0.84 0.07 0.05 0.02
50–900.80 0.90 0.06 0.05 0.02
55–950.71 0.87 0.06 0.05 0.02
60–1000.76 0.88 0.06 0.05 0.02
65–1050.86 0.93 0.05 0.04 0.02
Table 7. Prediction performance of the CatBoost model for different starting weights with a 40 kg interval.
Table 7. Prediction performance of the CatBoost model for different starting weights with a 40 kg interval.
R2Pearson_CorrRMSEMAEMAPE
30–700.50 0.73 0.08 0.06 0.03
35–750.68 0.82 0.08 0.07 0.03
40–800.59 0.77 0.08 0.06 0.03
45–850.70 0.85 0.07 0.05 0.02
50–900.81 0.90 0.06 0.05 0.02
55–950.71 0.87 0.06 0.05 0.02
60–1000.78 0.88 0.05 0.04 0.02
65–1050.87 0.93 0.05 0.04 0.02
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Zhu, L.; Jiang, W.; Yang, Y.; Gan, M.; Shen, L.; Zhu, L. Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction. Animals 2025, 15, 1773. https://doi.org/10.3390/ani15121773

AMA Style

Yang X, Zhu L, Jiang W, Yang Y, Gan M, Shen L, Zhu L. Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction. Animals. 2025; 15(12):1773. https://doi.org/10.3390/ani15121773

Chicago/Turabian Style

Yang, Xidi, Liangyu Zhu, Wenyu Jiang, Yiting Yang, Mailin Gan, Linyuan Shen, and Li Zhu. 2025. "Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction" Animals 15, no. 12: 1773. https://doi.org/10.3390/ani15121773

APA Style

Yang, X., Zhu, L., Jiang, W., Yang, Y., Gan, M., Shen, L., & Zhu, L. (2025). Machine Learning-Based Prediction of Feed Conversion Ratio: A Feasibility Study of Using Short-Term FCR Data for Long-Term Feed Conversion Ratio (FCR) Prediction. Animals, 15(12), 1773. https://doi.org/10.3390/ani15121773

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop