Next Article in Journal
Cafeteria and Fast-Food Diets Induce Neuroinflammation, Social Deficits, but a Different Cardiometabolic Phenotype
Previous Article in Journal
Longitudinal Association Between Menopausal Transition and Obstructive Sleep Apnea with Effect Modification by Salt Intake: A Prospective Cohort Study
Previous Article in Special Issue
Prognostic Value of Computed Tomography-Derived Muscle Density for Postoperative Complications in Enhanced Recovery After Surgery (ERAS) and Non-ERAS Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image-Based Dietary Energy and Macronutrients Estimation with ChatGPT-5: Cross-Source Evaluation Across Escalating Context Scenarios

by
Marcela Rodríguez-Jiménez
1,
Gustavo Daniel Martín-del-Campo-Becerra
2,
Sandra Sumalla-Cano
1,3,4,
Jorge Crespo-Álvarez
5,* and
Iñaki Elio
1,6,7,*
1
Research Group on Foods, Nutritional Biochemistry and Health, Universidad Europea del Atlántico, 39011 Santander, Spain
2
German Aerospace Center (DLR), 82234 Weßling, Germany
3
Faculty of Health Sciences, Universidade do Cuanza, Cuito EN250, Bié, Angola
4
Faculty of Health Sciences, Universidad de La Romana, La Romana 22000, Dominican Republic
5
Higher Polytechnic School, Universidad Europea del Atlántico, 39011 Santander, Spain
6
Faculty of Health Sciences, Universidad Internacional Iberoamericana, Arecibo, PR 00613, USA
7
Department of Health, Nutrition and Sport, Universidad Internacional Iberoamericana, Campeche 24560, Mexico
*
Authors to whom correspondence should be addressed.
Nutrients 2025, 17(22), 3613; https://doi.org/10.3390/nu17223613
Submission received: 24 October 2025 / Revised: 14 November 2025 / Accepted: 17 November 2025 / Published: 19 November 2025

Abstract

Background/Objectives: Estimating energy and macronutrients from food images is clinically relevant yet challenging, and rigorous evaluation requires transparent accuracy metrics with uncertainty and clear acknowledgement of reference data limitations across heterogeneous sources. This study assessed ChatGPT-5, a general-purpose vision-language model, across four scenarios differing in the amount and type of contextual information provided, using a composite dataset to quantify accuracy for calories and macronutrients. Methods: A total of 195 dishes were evaluated, sourced from Allrecipes.com, the SNAPMe dataset, and Home-prepared, weighed meals. Each dish was evaluated under Case 1 (image only), Case 2 (image plus standardized non-visual descriptors), Case 3 (image plus ingredient lists with amounts), and Case 4 (replicates Case 3 but excluding the image). The primary endpoint was kcal Mean Absolute Error (MAE); secondary endpoints included Median Absolute Error (MedAE) and Root Mean Square Error (RMSE) for kcal and macronutrients (protein, carbohydrates, and lipids), all reported with 95% Confidence Intervals (CIs) via dish-level bootstrap resampling and accompanied by absolute differences (Δ) between scenarios. Inference settings were standardized to support reproducibility and variance estimation. Source stratified analyses and quartile summaries were conducted to examine heterogeneity by curation level and nutrient ranges, with additional robustness checks for error complexity relationships. Results and Discussion: Accuracy improved from Case 1 to Case 2 and further in Case 3 for energy and all macronutrients when summarized by MAE, MedAE, and RMSE with 95% CIs, with absolute reductions (Δ) indicating material gains as contextual information increased. In contrast to Case 3, estimation accuracy declined in Case 4, underscoring the contribution of visual cues. Gains were largest in the Home-prepared dietitian-weighed subset and smaller yet consistent for Allrecipes.com and SNAPMe, reflecting differences in reference curation and measurement fidelity across sources. Scenario-level trends were concordant across sources, and stratified and quartile analyses showed coherent patterns of decreasing absolute errors with the provision of structured non-visual information and detailed ingredient data. Conclusions: ChatGPT-5 can deliver practically useful calorie and macronutrient estimates from food images, particularly when augmented with standardized nonvisual descriptors and detailed ingredients, as evidenced by reductions in MAE, MedAE, and RMSE with 95% CIs across scenarios. The decline in accuracy observed when the image was omitted, despite providing detailed ingredient information, indicates that visual cues contribute meaningfully to estimation performance and that improvements are not solely attributable to arithmetic from ingredient lists. Finally, to promote generalizability, it is recommended that future studies include repeated evaluations across diverse datasets, ensure public availability of prompts and outputs, and incorporate systematic comparisons with non-artificial-intelligence baselines.

1. Introduction

The increasing prevalence of sedentary lifestyles and the widespread consumption of ultra-processed foods have contributed to the rising incidence of chronic diseases, including type 2 diabetes, cardiovascular disorders, and certain types of cancer [1]. Against this backdrop, accurate dietary intake assessment is key for detecting unhealthy eating patterns, guiding nutritional interventions, and supporting dietitians in both disease prevention and patient management. Therefore, there is a reinforced need for accurate and scalable dietary assessment beyond self-report methods such as 24 h recalls and food diaries, which are prone to recall bias, underreporting, and variable adherence, compromising data quality and downstream decision-making in clinical and public health practice [2,3,4].
Image-based dietary assessment offers a pragmatic alternative by leveraging meal photographs with technologies such as computer vision and, more recently, artificial intelligence (AI) [3,4,5,6,7,8,9,10]. Still, performance in real-world conditions is limited by portion size estimation, occlusion, mixed dishes, lighting variability, camera angle, and cultural diversity of foods, all of which can introduce systematic and random errors in nutrient prediction from images.
Zheng et al. [4] carried out an exploratory review of AI models employing different input modalities (text, food images, jaw motion, and multimodal data), reporting variations in accuracy depending on the input type. Shonkoff et al. [5] conducted a systematic review on the use of AI in dietary assessment through food photographs, concluding that AI-based estimations of food volume and caloric content are comparable to those of dietitians and may even surpass them. Similarly, Sultana et al. [6] performed a systematic review of deep learning (DL) approaches applied to food photographs for classification, portion size estimation, and nutrient analysis, concluding that current models still face challenges in accuracy and generalizability for real-world dietary assessment.
In parallel, several studies have proposed dedicated AI models. Fang et al. [7] introduced an energy-distribution approach; Akpa et al. [8] tested the use of reference objects to improve calorie estimation; Ege et al. [9] compared convolutional neural networks in multitask and monotask configurations; and Wang et al. [10] trained a DL model on large-scale food image datasets. Additional efforts have focused on the development of databases and algorithmic resources to strengthen reliability. Wang et al. [10] employed the Nutrition5k and ChinaMartFood109 datasets to validate novel models, while Larke et al. [11] created the Surveying Nutrient Assessment with Photographs of Meals (SNAPMe) database to support standardized evaluation. Collectively, these studies highlight the critical role of robust, well-structured, and diverse image repositories in advancing AI-driven dietary assessment.
Despite encouraging research findings, current AI models for dietary assessment based on food photographs remain largely impractical for routine use by nutrition professionals. Their implementation often requires advanced expertise in programming and machine learning, as well as access to high-performance computing resources that are typically beyond the reach of most dietitians [10,12]. Thus, the use or adaptation of these systems depends heavily on collaboration with data scientists and engineers, creating a gap between research innovation and everyday clinical practice.
Advances in vision-language models (VLMs) suggest potential to fuse visual and textual cues to improve recognition and portion reasoning, but reported that accuracy varies widely across tasks, datasets, and protocols [13,14,15]. Evaluation quality depends critically on curated references, standardized prompts/inputs, and transparent error metrics with uncertainty to enable reproducibility and fair comparison across studies and time.
Consequently, this study adopts the perspective of a dietitian-nutritionist who, despite lacking knowledge in AI, seeks to use it independently in daily practice without the need for expensive computer equipment or support from programming specialists. In this context, ChatGPT-5 [16] offers an accessible alternative, providing advanced multimodal AI through a conversational interface. Moreover, as one of the most widely recognized and publicly adopted AI systems, it aligns with the familiarity and usability required for non-technical professionals, thereby enabling practical nutritional estimation from food images. Other VLMs are currently available, such as CLIP by OpenAI [17], PaLM-E by Google DeepMind [18], and Kosmos-2 by Microsoft [19]. However, these systems require dedicated setup and Application Programming Interface (API) integration, limiting their accessibility for healthcare professionals.
Building on the experience gained from previous work in this field, this study evaluates the accuracy of ChatGPT-5 in estimating caloric and macronutrient content from food photographs, using nutritional reference data as a benchmark. The evaluation framework is defined according to key considerations from the literature, which encompass three main areas: (i) model performance, including factors that affect accuracy and strategies to improve estimation; (ii) evaluation metrics and methods for comparison with ground truth values; and (iii) database design and nutritional annotation for training and validation purposes. These are summarized in Table 1.
In line with Table 1, the present study adopts an evaluation strategy that integrates controlled scenarios with validated nutritional benchmarks, serving as the ground truth. To support the evaluation of ChatGPT-5 across diverse cultural contexts, the analysis draws on a heterogeneous set of images, including dishes prepared and annotated by dietitians, samples from SNAPMe [11], and recipe images obtained from Allrecipes.com [20]. Four case studies are addressed: Case 1 includes only the meal image, while the next two cases progressively incorporate supplementary contextual information. Case 4 uses only detailed ingredient information, omitting the corresponding image, to assess the contribution of visual input to model performance. In addition, this work analyses the relationship between the number of ingredients, as a measure of dish complexity, and the absolute error in kcal.
The remainder of the paper is organized as follows. Section 2 describes the materials and methods employed, Section 3 presents the results, followed by the discussion in Section 4. Finally, Section 5 concludes this work.

2. Materials and Methods

This section describes the organization of the reference nutritional data, which served as the ground truth for assessing the accuracy of the estimates generated by ChatGPT-5. It also outlines the evaluation metrics, visualization tools, and the correlation and inferential analyses applied. Finally, it presents the study cases considered and concludes by describing how ChatGPT-5 was queried to obtain the nutritional estimations.

2.1. Database

Recommendations from the literature emphasize the importance of using diverse, standardized, and well-annotated data sources to minimize potential biases [4,6,9,11]. Accordingly, a database of 195 dishes was compiled, combining images and reference nutritional information obtained from three complementary sources, described below:
  • Allrecipes.com [20]: Provides caloric content and macronutrient distribution of meal recipes spanning diverse gastronomic traditions, accompanied by user-shared images that provide authentic visual representations of the dishes. From an estimated population of 51,000 dishes, a sample of 96 was drawn, meeting the criteria for a 10% margin of error at a 95% confidence level.
  • SNAPMe [11]: Collection of food photographs annotated with portion sizes and nutritional values, contributed by U.S. participants. Images include a reference object to assist with accurate volume estimation. For a population of 275 dishes, a sample of 74 was selected for a 10% margin of error at a 95% confidence level.
  • Home-prepared: Comprises 25 dishes prepared and carefully annotated by nutritionists, serving as a baseline for comparison. Nutritional data were obtained through direct quantity measurements and energy content analysis using Nutrix 1.0 [21] (a specialized software developed by Universidad Europea del Atlántico) and food equivalence tables, including the Mexican System of Food Equivalents (SMAE) [22] and the U.S. Department of Agriculture (USDA) FoodData Central [23]. A reference object (i.e., a ruler marked in centimeters) was placed in each photograph to support accurate volume and depth estimation. Images were captured with a mobile device (iPhone 16 Pro Max) under varying lighting conditions.
The reference datasets differ in origin and structure, which could introduce uncontrolled variability. To mitigate this, all nutritional entries were verified so that energy and macronutrient values were expressed in consistent units (kcal and grams (g)) and accurately reflected both the ingredient quantities and the portion sizes depicted in the corresponding images. When obvious inconsistencies were identified, macronutrient and energy values were adjusted using the SMAE [22] and USDA [23] food equivalence tables in combination with Nutrix 1.0 [21]. Records with incomplete or implausible nutritional information were excluded prior to analysis. All curated data were then organized into a consolidated database comprising all dishes, each including the corresponding image, detailed ingredient list, and reference nutritional values. Appendix A presents an example of the database structure, providing traceability and standardized access to the information required for model evaluation. Although some heterogeneity among datasets is unavoidable, this diversity enhances the ecological validity of the study by reflecting the variability typically observed in dietary records and real-world recipe data.

2.2. Evaluation Metrics

The performance of ChatGPT-5 was evaluated using a set of complementary quantitative error metrics. For all accuracy measures, 95% Confidence Intervals (CIs) were estimated using a non-parametric bootstrap with 10,000 resamples at the dish level.
The Mean Absolute Error (MAE) [10] and Median Absolute Error (MedAE) [24] quantified the magnitude of the errors between the ChatGPT-5 estimates g ^ j j = 1 J and the reference values g j j = 1 J :
M A E = 1 J j = 1 J g j g ^ j ,
M e d A E = m e d i a n g j g ^ j .
In contrast to MAE, which averages all errors and may be disproportionately influenced by extreme values, MedAE provides a more robust measure by reducing the impact of outliers. It captures the central tendency of the estimation errors across observations, offering a more representative assessment of model performance under typical conditions.
Relative accuracy was evaluated using the Mean Absolute Percentage Error (MAPE) [10] and the Median Absolute Percentage Error (MedAPE) [24]:
M A P E = 1 J j = 1 J g j g ^ j / g j ×   100 ,
M e d A P E = m e d i a n g j g ^ j / g j × 100 .
Both MAPE and MedAPE express errors as a percentage of the reference value, facilitating interpretability. Nonetheless, these metrics are asymmetric, as they penalize overestimations more heavily than underestimations of the same magnitude. Similar to MedAE, MedAPE is less sensitive to outliers.
Error dispersion was quantified using the Interquartile Range (IQR) [25], which represents the spread of the middle 50% of the data and is robust to extreme values. Specifically, the IQR of Absolute Errors (IQRAE) and Absolute Percentage Errors (IQRAPE) were computed as:
I Q R A E = Q 3 g j g ^ j Q 1 g j g ^ j ,
I Q R A P E = Q 3 g j g ^ j / g j × 100 Q 1 g j g ^ j / g j × 100 ,
where Q 1 and Q 3 denote the first and third quartiles of the distributions of absolute and absolute percentage errors, respectively. These metrics help assess the consistency of the model’s performance across the dataset, highlighting whether errors are concentrated around the median or spread out over a wider range.
Lastly, the Root Mean Square Error (RMSE) was computed as [5,24]:
R M S E = 1 J j = 1 J g j g ^ j 2 .
This metric assigns greater weight to large estimation errors due to the squaring operation.

2.3. Visual Analysis

Scatter plots were employed to visually assess estimation performance, depicting ground truth values g j j = 1 J on the x-axis and corresponding ChatGPT-5 estimates g ^ j j = 1 J on the y-axis. Ideally, all points would align along the identity line y = x ; any displacement from this line indicates estimation errors, with points above and below representing over- and underestimations, respectively. This visualization approach eases rapid identification of bias, dispersion, and overall agreement between estimated and reference values [10].
Boxplots of absolute errors for calories and macronutrients were used to compare estimation performance across the three data sources. These visualizations summarize both the central tendency and the dispersion of errors (expressed in kcal or g), allowing straightforward identification of variability and outliers. Each boxplot displays the median, IQR, and Tukey whiskers (1.5 × IQR) [26], while data points lying beyond the whiskers are shown as outliers.

2.4. Correlation Analysis

To evaluate whether dish complexity influenced estimation accuracy, the relationship between the number of ingredients and the absolute error in dietary energy estimation (kcal) was examined using Pearson’s correlation coefficient [9]:
r = m i m ¯ n i n ¯ m i m ¯ 2 · n i n ¯ 2 ,
where r denotes the coefficient, m i and n i are the observed values, and m ¯ and n ¯ represent their respective means. Values of r close to 1 indicate a strong positive linear relationship, values near −1 indicate a strong negative linear relationship, and values around 0 suggest the absence of correlation. To quantify the uncertainty associated with r , 95% CIs were computed using Fisher’s z -transformation [27].
To further characterize this relationship, a linear regression was fitted [9]:
y = a   +   b x ,
where y represents the dependent variable, x the independent variable, a the intercept and b the slope. This method allows the identification of systematic trends, assessment of predictive relationships, and detection of potential under- or overestimation patterns. A 95% CI for the regression line was included, along with a confidence ellipse representing the joint distribution of predicted and observed values. The CI quantifies the uncertainty in the slope b and intercept a estimates, while the ellipse captures the covariance structure of the data. A narrow ellipse reflects strong agreement and low dispersion, whereas a wider ellipse indicates greater variability or systematic bias.

2.5. Inferential Analysis of Estimation Errors

To complement the descriptive statistics (MAE and MedAE), inferential analyses were conducted to formally test whether estimation errors differed significantly among the three data sources (Allrecipes.com, SNAPMe, and Home-prepared). Since the dishes from each source are distinct (i.e., samples are independent and not paired), a non-parametric approach for independent samples was employed. The following hypotheses were evaluated:
  • Null hypothesis (H0): There are no significant differences in estimation errors among the three data sources.
  • Alternative hypothesis (H1): At least one data source differs significantly from the others.
The global hypothesis was tested using the Kruskal–Wallis rank-sum test [28], which evaluates whether the distributions of absolute errors differ significantly among the sources. The test statistic is given by:
H = 12 L L + 1 i = 1 I l i R ¯ i 2 3 L + 1 ,
where L is the total number of observations, I is the number of groups, l i is the sample size of group i , and R ¯ i is the mean rank of group i . Under the null hypothesis (no significant differences between sources), H approximately follows a χ 2 distribution with I 1 degrees of freedom. A result of p < 0.05 indicates that at least one group differs significantly from the others.
When the Kruskal–Wallis was significant, pairwise comparisons were conducted using the Mann–Whitney U test [29] to identify which specific pairs of sources differ. The test statistic is:
U = k x k y + k x k x + 1 2 R x ,
where k x and k y are the sample sizes of groups X and Y , and R x is the sum of ranks in group X .
To account for multiple testing, Holm’s sequential correction [30] was applied in two stages: (i) within each variable, to adjust pairwise comparisons among the three data sources (Allrecipes.com, SNAPMe, and Home-prepared); and (ii) across variables (calories, proteins, lipids, and carbohydrates) within each experimental case, to control the overall family-wise error rate given that the errors are correlated. This two-level strategy provides a more stringent control of false positives and ensures that only statistically robust differences (unlikely to occur by chance) are interpreted as true differences among datasets. A result of p H o l m < 0.05 indicates that the difference between the two groups is statistically significant.
To assess the magnitude of the detected differences, Cliff’s δ [31] was computed. This measure quantifies the degree of overlap between two distributions:
δ = # x i > y j # x i < y j k x k y ,
where x i and y j are observations from groups X and Y . Operator # · denotes the number of pairwise comparisons in which the condition holds. Values of δ range from −1 to +1, with δ 0.147,0.333,0.474 corresponding to small, medium, and large effects, respectively.

2.6. Test Cases

The evaluation of ChatGPT-5 was conducted under four scenarios differing in the type and amount of information available. The assessment was performed using MAE, MedAE, MAPE, MedAPE, IQRAE, IQRAPE and RMSE, complemented with scatter plots and boxplots. These test cases were designed to identify model limitations, analyze the effect of non-visual information, and evaluate performance when complete ingredient data were available. Specifically, the following scenarios were considered:
  • Image only: performance assessment in the absence of contextual information.
  • Image with non-visual data: inclusion of characteristics that are difficult to capture visually, namely type and amount of fat, type and amount of sweetener, fat content of dairy products, and type of meat.
  • Image with detailed ingredient information: provision of complete ingredient descriptions and quantities to enable the most accurate estimation.
  • Detailed ingredient information only: replicates the conditions of Case 3, except that the image is omitted to isolate the contribution of visual input.
For each test case, the relationship between the number of ingredients and the absolute error in dietary energy estimation (kcal) was analyzed using Pearson’s correlation and linear regression. Given that dishes with more ingredients are commonly more complex, this analysis evaluated whether dish complexity affected estimation error and to what extent this effect is consistent across scenarios.

2.7. ChatGPT

This work used the Plus version of ChatGPT-5 (released in August 2025) to avoid the restrictions of the free tier and usage limits. This allowed uninterrupted access to the tool, including the uploading and processing of multiple food images. The data-sharing opt-out setting of the Plus account was activated so that uploaded images and data were processed only within the active session and not used for model training. According to OpenAI’s data policy, session content remains private and is not publicly shared, although it may be temporarily retained for security or compliance purposes. The exact retention period is not publicly specified.
The study focused on evaluating ChatGPT-5 under conditions representative of real-world professional use. Accordingly, all estimations were obtained through the ChatGPT web interface (chat.openai.com) rather than the API. This setup reflects the typical context of a dietitian-nutritionist without programming expertise, using accessible tools rather than specialized computational environments. As the chat user interface does not allow control over generation parameters such as temperature, the model operates in its default probabilistic mode.
Prompts were intentionally designed to remain simple and natural, resembling how a dietitian-nutritionist would typically interact with the system. Each case study involved specific prompts tailored to the task. The prompt for the first scenario was: “Based on the image, calculate the energy content in kilocalories (kcal) and the amount of protein, fat, and carbohydrates in grams (g). Organize the information in a table, indicating amounts for each food item and totals”. In the subsequent scenarios, this same prompt was adapted, for example, by adding contextual information. Appendix B illustrates how ChatGPT-5 was applied in each case study.
For Cases 1–3, each prompt included a single image corresponding to one dish, and the chat session was reset prior to each evaluation so that ChatGPT-5 had no prior information or contextual memory from previous interactions. This procedure made all estimations independent, without influence from preceding evaluations. Case 4 followed the same procedure but excluded the image.
While ChatGPT-5 is more powerful and flexible than earlier versions, it also tends to engage in broader interpretive reasoning, introducing inferred information not explicitly present in the visual or textual input. To mitigate this issue and promote analytical consistency, a set of explicit behavioral instructions was incorporated at the beginning of each evaluation to limit interpretation, enforce structured outputs, and encourage conservative, evidence-based responses. Specifically:
  • “Act as a technical analyst. Do not make assumptions: describe only what the imagery and data shows.”
  • “Provide responses in a literal and quantitative manner.”
  • “Do not draw conclusions unless they are explicitly supported by the data.”
  • “If ambiguity exists, state it explicitly rather than attempting to resolve it.”
  • “Present intermediate calculations before the final result, and clearly indicate any estimated values.”

3. Results

All analyses were conducted in Python 3.12, using NumPy 1.26.4, pandas 2.3.3, SciPy 1.16.2, and Matplotlib 3.10.6. Results are grouped by test case and structured around the three reference data sources: Allrecipes.com [20], SNAPMe [11], and Home-prepared. The summary tables (Table 2, Table 3, Table 4 and Table 5) report MAE, MedAE, IQRAE, and RMSE in absolute terms (kcal or g), as well as MAPE, MedAPE, and IQRAPE in percentage (%). The values for each independent data source and for all combined data sources (global) are presented as overall results and by quartile. Scatter plots in Figure 1 then compare estimated calories and macronutrients against the corresponding reference values, while boxplots in Figure 2 complement these evaluations by summarizing the distribution and variability of absolute errors across sources, enabling visual assessment of consistency, dispersion, and the presence of outliers.
The last plots in Figure 3 analyze the relationship between the number of ingredients and the absolute error in energy estimation (kcal). The regression slope indicates the average change in error per additional ingredient, and the Pearson correlation coefficient ( r ) quantifies the strength of the association. A pronounced positive slope generally corresponds to r values close to 1 ( 0.7), whereas low r values (0–0.3) indicate high variability [9]. The shaded area represents the 95% CI for the regression line, and the ellipse delineates the joint confidence region between predicted and observed values, providing a visual summary of the dispersion and overall fit for each data source.
Finally, nonparametric inferential analyses (Table 6) were performed to assess differences in estimation errors among the three data sources. The Kruskal–Wallis test was applied to each variable across the test cases. When overall differences were significant, pairwise Mann–Whitney tests with Holm correction were conducted. Cliff’s δ was computed to quantify effect sizes. Results are summarized in Table 6, with statistically significant differences ( p H o l m < 0.05) shown in bold.
Table 2. Case 1 (image only). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Table 2. Case 1 (image only). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Calories (kcal)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal127.81100.50102.7535.0025.6527.83166.99
95% CIs107.41, 150.6287.00, 121.5072.250, 136.25029.30, 41.1823.02, 34.7619.76, 37.59135.27, 201.55
Q1 (≤280)96.4285.0097.0055.0850.9064.61120.30
Q2 (280–368.5)85.1372.0089.0026.5322.8626.25109.46
Q3 (368.5–502.5)113.2596.0061.5026.5923.1417.84123.37
Q4 (>502.5)216.00161.50197.5030.6125.6525.07264.13
SNAPMeTotal133.00108.50138.2528.1928.0328.93175.61
95% CIs107.73, 160.4675.00, 135.50108.49, 189.723.26, 33.7821.00, 30.8920.68, 32.37142.51, 208.48
Q1 (≤348.2)82.1073.0063.5037.0329.0027.75105.90
Q2 (348.2–435.5)72.3841.5096.5018.689.8426.1195.63
Q3 (435.5–592.8)136.90125.0091.2526.4624.6715.12155.84
Q4 (>592.8)237.63245.00210.0030.0132.2925.56277.88
Home-preparedTotal75.1456.0062.0020.1213.2513.05104.52
95% CIs49.60, 105.7634.00, 77.0030.00, 123.0012.46, 30.9310.04, 20.157.38, 24.5863.22, 144.96
Q1 (≤381)56.0740.0039.2532.2420.1517.8570.72
Q2 (381–448)38.6639.5028.259.028.966.2443.96
Q3 (448–489)58.8357.5031.7512.3311.946.7668.69
Q4 (>489)150.16131.0099.5024.8720.2914.46181.76
All sourcesTotal123.0398.00123.0030.5124.3427.49163.80
95% CIs108.61, 138.7885.00, 106.00100.00, 141.0026.86, 34.2522.68, 28.4922.76, 31.71142.58, 186.00
Q1 (≤299)81.6968.0089.0045.2132.0752.18105.20
Q2 (299–407)86.2081.0093.0024.4423.0224.86108.5
Q3 (407–533.5)98.8192.5080.5021.2520.7315.17113.94
Q4 (>533.5)224.91190.00208.0030.9430.8926.20266.83
Proteins (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal8.025.7038.0751.7328.4037.8911.08
95% CIs6.52, 9.604.20, 7.605.55, 12.9037.92, 68.7723.16, 36.3625.89, 60.049.04, 13.02
Q1 (≤12)4.512.303.6093.8240.0090.958.06
Q2 (12–19)7.723.558.9248.0224.0457.1710.83
Q3 (19–28)8.667.5212.2537.0530.5623.4814.43
Q4 (>28)11.468.0012.2525.1526.0023.3314.43
SNAPMeTotal8.324.1510.5437.0121.5529.6512.50
95% CIs6.30, 10.532.80, 6.326.10, 14.9226.20, 51.6114.67, 32.2224.33, 36.309.60, 15.24
Q1 (≤13.2)3.661.382.6055.5922.9225.615.88
Q2 (13.2–22.6)6.082.603.9737.2613.6127.5410.02
Q3 (22.6–33.9)7.306.457.3026.1324.2727.138.76
Q4 (>33.9)16.0614.2020.2528.4931.9429.8820.16
Home-preparedTotal6.144.607.1423.9318.0018.838.04
95% CIs4.24, 8.272.10, 9.003.20, 9.5014.76, 37.918.78, 23.798.75, 27.965.70, 10.32
Q1 (≤18.2)3.072.102.5735.5217.8619.414.52
Q2 (18.2–27)5.094.854.5020.7520.1919.046.06
Q3 (27–38.8)4.043.352.8711.6110.518.984.87
Q4 (>38.8)12.8611.456.8025.9024.5013.1113.61
All sourcesTotal7.895.008.5042.5824.8430.9711.32
95% CIs6.81, 9.094.20, 6.306.65, 11.204.17, 52.4320.38, 28.5326.08, 37.079.77, 12.89
Q1 (≤13)3.982.203.3070.1027.2786.676.81
Q2 (13–22)7.534.608.0143.6924.2952.3210.63
Q3 (22–33.3)7.127.205.4026.5924.8422.758.64
Q4 (>33.3)13.169.7015.0025.6923.7926.9216.69
Carbohydrates (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal11.486.8014.1253.8737.0851.0616.34
95% CIs9.27, 13.855.05, 10.459.07, 19.1241.87, 67.8930.00, 45.0036.28, 67.9413.47, 18.99
Q1 (≤12)6.132.806.7590.4348.4766.939.97
Q2 (12–26)10.147.409.0051.9239.4748.2613.80
Q3 (26–41)15.4212.1519.8745.0730.0062.8519.96
Q4 (>41)14.7512.0015.1023.6722.2231.3519.91
SNAPMeTotal12.998.7511.5032.0124.4126.7918.86
95% CIs10.07, 16.335.98, 11.508.10, 18.7626.09, 38.6218.57, 33.7021.45, 37.9813.74, 23.86
Q1 (≤26)7.984.325.2548.0836.6560.3211.05
Q2 (26–40)7.124.808.6022.6814.0826.108.59
Q3 (40–52.3)13.9711.758.7730.1325.5717.5217.43
Q4 (>52.3)22.6012.5018.6525.2919.7822.5630.06
Home-preparedTotal8.917.1010.3125.5917.1321.9511.56
95% CIs6.12, 11.953.76, 12.505.90, 16.2016.22, 37.5411.36, 30.2911.19, 41.178.23, 14.58
Q1 (≤33.5)11.7114.5013.7549.1748.4436.8414.87
Q2 (33.5–41.8)7.207.124.1318.5717.4412.348.07
Q3 (41.8–46.4)6.304.908.7514.2210.8520.908.12
Q4 (>550)9.976.6814.2116.4710.8121.7412.95
All sourcesTotal11.728.1012.1041.9529.5537.1116.84
95% CIs10.11, 13.506.00, 10.209.60, 16.0535.31, 49.6123.08, 32.5930.52, 45.8714.36, 19.43
Q1 (≤17.9)6.013.406.5068.9745.0055.139.24
Q2 (17.9–34.5)11.467.7012.0045.5036.3342.8215.25
Q3 (34.5–46.2)11.339.6511.0528.1425.1025.2414.55
Q4 (>46.2)18.0812.0020.0024.9019.7825.3224.56
Lipids (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal9.176.807.4559.7637.7138.1813.35
95% CIs7.39, 11.245.30, 8.506.00, 11.5243.76, 80.2228.42, 47.0030.90, 61.719.72, 17.26
Q1 (≤10)6.334.506.50128.3089.0091.258.49
Q2 (10–19.5)5.815.406.6540.5641.1132.427.08
Q3(19.5–27.2)6.545.656.2727.1422.0925.758.12
Q4 (>27.2)17.9515.0512.1239.3637.7431.8822.89
SNAPMeTotal8.905.7912.3246.8241.3339.2912.55
95% CIs7.00, 11.014.30, 8.307.60, 14.7538.03, 57.2635.62, 49.0924.02, 55.379.76, 15.60
Q1 (≤11.2)3.522.603.5067.9153.9629.064.79
Q2 (11.2–19.2)6.495.144.7044.7035.9733.948.43
Q3 (19.2–30.8)8.137.3111.6534.1133.0557.7110.72
Q4 (>30.8)17.2916.609.9040.8941.3919.0420.35
Home-preparedTotal6.365.524.8044.2035.2027.757.98
95% CIs4.63, 8.323.10, 7.402.10, 9.9131.81, 60.0327.57, 42.789.10, 47.955.67, 10.34
Q1 (≤10.5)4.322.903.8066.2930.0071.455.75
Q2 (10.5–15)4.054.802.3031.3835.9315.294.47
Q3 (15–21.7)7.566.951.6738.1335.1111.917.97
Q4 (>21.7)9.8410.058.3237.3435.2334.6111.96
All sourcesTotal8.706.009.1052.8839.4838.5312.48
95% CIs7.51, 10.055.24, 7.407.00, 11.5043.89, 63.6434.10, 43.7531.87, 47.8610.27, 14.94
Q1 (≤10.6)4.823.404.7098.2861.1788.546.78
Q2 (10.6–18.7)6.095.245.1043.4636.6732.927.53
Q3(18.7–27.5)6.655.887.3029.3228.2132.838.27
Q4 (>27.5)17.2115.4013.1040.9241.1330.5821.21
Note: Values for each independent data source and for all combined data sources (global) are presented as overall results and by quartile. 95% CIs for all reported metrics were estimated via nonparametric bootstrap (10,000 resamples, dish-level).
Table 3. Case 2 (image with non-visual data). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Table 3. Case 2 (image with non-visual data). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Calories (kcal)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal95.5880.1596.5028.8518.9532.82127.45
95% CIs79.33, 113.2262.00, 91.0071.50, 128.7523.40, 34.7116.11, 24.3019.54, 43.54104.25, 150.58
Q1 (≤280)88.8581.0081.0050.1343.0352.36107.63
Q2 (280–368.5)70.1343.0068.5021.6613.2721.88100.38
Q3 (368.5–502.5)97.7975.5068.7523.2717.6014.29135.15
Q4 (>502.5)124.7999.50112.7519.1516.9517.68158.06
SNAPMeTotal101.1377.00111.2522.1719.3221.53129.39
95% CIs82.99, 119.8854.50, 110.0081.00, 143.7518.63, 25.9412.88, 25.6517.41, 25.78107.67, 149.83
Q1 (≤348.2)78.2164.0078.5033.0528.4926.28102.59
Q2 (348.2–435.5)49.2239.5033.0012.7310.507.8960.87
Q3 (435.5–592.8)110.40125.0097.0021.3125.5818.56130.05
Q4 (>592.8)164.47156.00142.5021.0319.0618.96187.47
Home-preparedTotal51.5039.0021.0013.5510.419.0071.37
95% CIs35.74, 73.7232.00, 49.0011.50, 57.009.26, 19.027.69, 12.784.02, 15.9841.12, 104.46
Q1 (≤381)38.0737.5016.5020.0817.677.6641.84
Q2 (381–448)44.1642.5031.0010.4810.168.2656.99
Q3 (448–489)36.5034.0013.757.697.183.0642.06
Q4 (>489)89.5053.0037.714.849.343.79119.02
All sourcesTotal92.0468.0095.5024.3517.6723.36122.48
95% CIs80.99, 103.6354.00, 83.0079.50, 118.5021.21, 27.6714.36, 20.1519.01, 28.18107.89, 137.76
Q1 (≤299)68.0353.0061.0037.4028.4946.4587.27
Q2 (299–407)83.3253.0080.0023.6913.7426.45119.37
Q3 (407–533.5)73.7964.5078.2515.8114.0615.8990.77
Q4 (>533.5)142.63119.00149.0020.3317.5721.54172.54
Proteins (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal7.475.108.5247.6627.3040.5210.77
95% CIs5.96, 9.113.50, 6.305.57, 11.7234.97, 63.1422.38, 35.7129.99, 55.128.62, 12.80
Q1 (≤12)3.432.302.2081.2323.0063.796.48
Q2 (12–19)7.495.209.6047.0234.7359.7110.67
Q3 (19–28)8.587.607.4536.7930.7029.9810.55
Q4 (>28)10.676.109.7023.1919.6820.9514.27
SNAPMeTotal8.114.6210.0936.1124.5529.5112.06
95% CIs6.16, 10.243.10, 6.776.22, 13.9225.84, 50.2817.23, 32.5723.42, 37.849.17, 14.85
Q1 (≤13.2)3.952.402.9359.6127.5045.656.02
Q2 (13.2–22.6)4.762.654.1229.1016.2825.617.61
Q3 (22.6–33.9)7.617.956.9626.9728.2126.189.01
Q4 (>33.9)15.9315.3020.7027.9331.6128.4719.97
Home-preparedTotal6.556.205.9024.2718.2412.888.29
95% CIs4.70, 8.634.40, 7.902.56, 8.2716.82, 34.0515.59, 26.965.92, 29.565.93, 10.55
Q1 (≤18.2)3.732.205.8235.1915.7134.884.79
Q2 (18.2–27)4.054.854.4216.5820.1918.104.91
Q3 (27–38.8)5.905.752.9217.2316.917.946.48
Q4 (>38.8)13.0011.858.8526.2522.4817.0413.91
All sourcesTotal7.595.108.3540.2825.5332.9911.01
95% CIs6.52, 8.764.30, 6.106.55, 11.0532.54, 49.3521.43, 28.5726.65, 39.839.45, 12.57
Q1 (≤13)3.612.303.0065.3127.2749.835.99
Q2 (13–22)7.075.006.6740.8827.3938.639.94
Q3 (22–33.3)6.766.107.5025.1825.5328.308.49
Q4 (>33.3)13.179.3013.7025.8524.6425.8016.71
Carbohydrates (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal10.766.9013.2050.8632.4147.7015.02
95% CIs8.78, 12.894.95, 10.609.92, 17.3039.51, 64.0522.50, 40.7135.42, 73.7812.56, 17.37
Q1 (≤12)5.982.706.1584.9842.9489.2710.14
Q2 (12–26)9.417.0010.0549.6435.7946.1012.98
Q3 (26–41)14.6412.9514.5742.7733.6845.6018.54
Q4 (>41)13.4611.4016.0021.9422.2224.5317.29
SNAPMeTotal11.298.0512.2727.7423.2624.6415.37
95% CIs9.05, 13.785.37, 11.158.36, 17.2123.05, 33.0018.41, 27.8017.16, 32.9112.13, 18.77
Q1 (≤26)6.684.154.8538.5024.0546.359.61
Q2 (26–40)7.675.709.0024.0117.0724.989.12
Q3 (40–52.3)11.7710.409.5225.2722.3719.2613.91
Q4 (>52.3)18.9014.2018.5022.1024.6416.0523.76
Home-preparedTotal9.508.0012.5026.8017.6727.6412.19
95% CIs6.56, 12.583.00, 13.305.75, 18.0017.39, 38.8711.25, 33.3313.21, 46.388.86, 15.14
Q1 (≤33.5)11.6313.3014.3048.3948.4436.0814.82
Q2 (33.5–41.8)8.707.574.4122.4318.5613.0910.61
Q3 (41.8–46.4)7.135.608.0515.9912.2519.058.91
Q4 (>550)10.176.6815.2616.8110.8123.6013.08
All sourcesTotal10.807.5012.6539.0024.8136.2414.83
95% CIs9.41, 12.266.00, 9.4010.35, 15.8532.90, 46.0321.42, 30.0027.11, 42.1912.99, 16.67
Q1 (≤17.9)5.553.206.1063.1938.3359.768.78
Q2 (17.9–34.5)10.796.8011.8042.6934.7841.5214.50
Q3 (34.5–46.2)11.069.4511.7827.5823.5725.7813.89
Q4 (>46.2)15.7912.2017.7022.3122.2220.8619.96
Lipids (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal6.254.007.1651.1722.0839.889.06
95% CIs4.97, 7.602.70, 5.254.37, 9.8231.32, 79.6015.45, 30.7727.46, 60.827.11, 10.96
Q1 (≤10)5.813.307.10121.4462.5068.339.63
Q2 (10–19.5)5.542.704.3037.6624.5539.957.93
Q3 (19.5–27.2)4.312.603.5517.659.9214.516.16
Q4 (>27.2)9.328.7010.224.4420.8822.6711.55
SNAPMeTotal5.614.195.9035.1522.7339.837.76
95% CIs4.44, 6.883.00, 4.953.40, 9.2027.06, 44.7818.92, 31.8022.05, 51.366.07, 9.45
Q1 (≤11.2)3.081.803.4060.5155.0051.234.31
Q2 (11.2–19.2)5.544.644.3737.4132.0238.527.36
Q3 (19.2–30.8)5.224.203.8522.4618.1021.347.12
Q4 (>30.8)8.579.358.3521.0121.5722.5910.80
Home-preparedTotal5.565.526.7236.1238.2534.026.82
95% CIs4.05, 7.102.50, 7.903.50, 9.1027.44, 45.0720.00, 43.9116.17, 50.455.36, 8.11
Q1 (≤10.5)3.081.684.7141.0636.3653.974.18
Q2 (10.5–15)4.964.804.1239.4439.6943.045.85
Q3 (15–21.7)6.477.212.9331.9436.7912.857.35
Q4 (>21.7)8.169.446.5531.2339.1519.999.22
All sourcesTotal5.924.306.6543.2024.3640.848.32
95% CIs5.13, 6.763.30, 4.905.48, 8.0632.63, 57.5120.25, 30.4831.54, 51.917.12, 9.54
Q1 (≤10.6)4.212.204.4088.2951.6770.527.29
Q2 (10.6–18.7)5.454.505.2038.8729.3843.657.19
Q3 (18.7–27.5)5.114.255.5422.7017.6225.756.80
Q4 (>27.5)8.909.309.5123.4521.5023.9711.19
Note: Values for each independent data source and for all combined data sources (global) are presented as overall results and by quartile. 95% CIs for all reported metrics were estimated via nonparametric bootstrap (10,000 resamples, dish-level).
Table 4. Case 3 (image with detailed ingredient information). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Table 4. Case 3 (image with detailed ingredient information). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Calories (kcal)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal64.9940.0068.5018.649.8420.2097.84
95% CIs51.05, 80.0826.50, 54.0052.50, 103.7514.16, 23.608.11, 12.2911.66, 26.8073.95, 121.58
Q1 (≤280)49.0638.0066.0029.2614.2935.6367.87
Q2 (280–368.5)28.5020.0019.508.876.006.0441.19
Q3 (368.5–502.5)94.6763.5093.5022.2115.7121.32135.19
Q4 (>502.5)86.8960.00103.7513.3810.4510.64116.58
SNAPMeTotal46.4833.5045.7510.168.0711.6465.82
95% CIs36.38, 57.6824.00, 42.0032.20, 65.758.13, 12.435.12, 10.958.58, 14.6950.56, 80.75
Q1 (≤348.2)24.4215.0023.5011.978.1613.5035.27
Q2 (348.2–435.5)36.0535.0056.759.458.7815.5845.92
Q3 (435.5–592.8)44.4538.0036.258.828.156.8357.15
Q 4 (>592.8)80.3653.00109.0010.287.3911.13102.66
Home-preparedTotal28.8222.0038.006.926.028.1737.55
95% CIs19.88, 38.4410.00, 45.0018.00, 51.004.84, 9.192.44, 9.735.09, 11.1127.21, 46.86
Q1 (≤281)19.7816.0021.508.406.029.0327.05
Q2 (281–448)35.1635.0020.758.298.455.7738.15
Q3 (448–489)7.668.504.751.611.790.978.50
Q4 (>489)54.1658.0017.259.139.552.5759.10
All sourcesTotal53.3334.0056.0013.928.4012.7380.86
95% CIs45.08, 62.2727.00, 42.0044.00, 70.0011.54, 16.667.18, 9.9110.09, 15.7665.90, 96.38
Q1 (≤299)34.2024.0039.0019.979.1619.1151.40
Q2 (299–407)46.1329.0044.0012.738.6311.5778.80
Q3 (407–533.5)55.5836.0046.7511.968.1410.1083.77
Q4 (>533.5)77.4556.0093.0010.988.8410.48101.44
Proteins (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal4.592.805.1527.2414.6927.086.91
95% CIs3.61, 5.682.20, 3.853.97, 6.3719.96, 35.9112.76, 18.7116.52, 37.795.32, 8.42
Q1 (≤12)2.081.302.7046.2221.6755.832.96
Q2 (12–19)2.802.303.1517.8212.9823.004.08
Q3 (19–28)6.343.309.9526.7214.9537.029.28
Q4 (>28)7.375.703.7516.9915.3510.589.07
SNAPMeTotal3.382.204.2215.1910.1713.845.02
95% CIs2.59, 4.271.47, 3.252.81, 5.5011.37, 19.937.78, 14.6810.46, 19.333.67, 6.39
Q1 (≤13.2)1.660.961.9824.3311.4924.282.47
Q2 (13.2–22.6)2.261.523.3213.219.4616.913.19
Q3 (22.6–33.9)3.342.753.7011.6110.7811.874.11
Q4 (>33.9)6.194.706.9211.3010.339.868.14
Home-preparedTotal3.772.302.9013.1012.567.975.47
95% CIs2.45, 5.521.60, 4.341.70, 5.109.90, 16.577.86, 15.334.77, 18.062.98, 7.97
Q1 (≤18.2)1.911.701.2415.4213.824.092.23
Q2 (18.2–27)3.233.202.1213.3012.2710.703.68
Q3 (27–38.8)4.503.802.0213.0610.927.615.74
Q4 (38.8)5.763.254.8710.227.9911.428.50
All sourcesTotal4.032.504.3020.8512.8017.036.07
95% CIs3.41, 4.692.20, 3.303.75, 5.3516.91, 25.7110.33, 14.7113.55, 23.125.02, 7.09
Q1 (≤13)1.850.962.2732.8513.8235.002.68
Q2 (13–22)3.202.292.4718.1712.6113.615.32
Q3 (22–33.3)4.683.604.8517.4614.2316.426.38
Q4 (>33.3)6.665.205.8013.5912.7910.678.62
Carbohydrates (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal6.893.807.7033.5215.6534.0111.55
95% CIs5.15, 8.832.20, 4.704.37, 12.8224.60, 44.0910.93, 22.8623.02, 54.358.58, 14.34
Q1 (≤12)4.641.304.6060.2128.5766.047.93
Q2 (12–26)5.683.906.0531.1522.9430.777.85
Q3 (26–41)8.854.157.2525.3411.0719.1115.12
Q4 (>41)8.613.5010.6514.268.3321.7513.63
SNAPMeTotal5.512.206.6213.416.1312.089.42
95% CIs3.89, 7.371.20, 3.003.20, 10.349.63, 17.825.00, 8.936.99, 21.896.58, 12.11
Q1 (≤26)3.131.002.4217.126.1112.676.09
Q2 (26–40)2.541.601.508.245.335.703.93
Q3 (40–52.3)8.053.0511.0017.946.1325.5112.70
Q4 (>52.3)8.296.909.369.848.9211.4611.81
Home-preparedTotal5.564.206.3015.0410.9513.607.67
95% CIs3.62, 7.772.20, 6.803.01, 9.1910.17, 20.606.57, 19.557.06, 25.854.86, 10.24
Q1 (≤33.5)4.053.602.3522.0919.6015.604.71
Q2 (33.5–41.8)1.441.671.813.824.524.771.75
Q3 (41.8–46.4)9.188.556.3220.2619.1113.3710.60
Q4 (>46.4)7.806.503.2212.8010.896.2610.20
All sourcesTotal6.202.907.3623.5210.4022.5810.34
95% CIs5.12, 7.412.20, 4.005.05, 8.7518.61, 29.208.28, 12.9217.84, 28.848.47, 12.20
Q1 (≤17.9)4.162.204.5044.8022.9448.726.79
Q2 (17.9–34.5)5.232.204.3019.836.5717.279.42
Q3 (34.5–46.2)6.762.757.6716.487.1318.2311.46
Q4 (>46.2)8.666.2012.7012.829.1714.5812.73
Lipids (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal4.813.256.6227.2016.3432.926.82
CI95%3.87, 5.802.30, 4.304.50, 7.7221.55, 33.2710.42, 23.3321.46, 42.805.58, 8.03
Q1 (≤10)2.230.702.8035.1517.1457.003.73
Q2 (10–19.5)5.674.507.1336.9928.1245.467.89
Q3 (19.5–27.2)4.693.834.5719.2715.8218.506.02
Q4 (>27.2)6.785.357.1517.4610.1219.398.69
SNAPMeTotal3.541.922.8118.9612.4421.285.85
95% CIs2.57, 4.671.40, 2.601.80, 5.6514.89, 23.478.87, 16.8612.31, 31.033.99, 7.59
Q1 (≤11.2)1.541.101.8530.1123.6141.332.08
Q2 (11.2–19.2)2.671.851.3517.0213.3610.213.66
Q3 (19.2–30.8)2.671.502.4511.376.5911.483.83
Q4 (>30.8)7.193.609.0017.439.6819.4010.12
Home-preparedTotal4.262.404.9227.6421.7426.946.48
95% CIs2.54, 6.381.20, 5.002.60, 9.1018.35, 37.869.44, 33.8017.43, 51.003.63, 9.33
Q1 (≤10.5)1.571.070.6926.5515.0032.282.42
Q2 (10.5–15)3.783.602.9528.9331.5924.085.10
Q3 (15–21.7)4.954.455.3225.1321.0123.775.97
Q4 (>21.7)7.194.555.9930.1419.4121.9010.33
All sourcesTotal4.262.505.0224.1614.9427.876.42
95% CIs3.62, 4.972.00, 3.003.64, 6.8520.67, 27.9011.43, 17.7821.39, 34.875.45, 7.38
Q1 (≤10.6)1.770.702.1031.6916.9049.872.94
Q2 (10.6–18.7)3.992.504.5727.3517.6929.445.67
Q3 (18.7–27.5)4.573.404.6220.3113.8120.226.55
Q4 (>27.5)6.714.307.6917.369.7720.179.02
Note: Values for each independent data source and for all combined data sources (global) are presented as overall results and by quartile. 95% CIs for all reported metrics were estimated via nonparametric bootstrap (10,000 resamples, dish-level).
Table 5. Case 4 (detailed ingredient information only). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Table 5. Case 4 (detailed ingredient information only). Evaluation metrics for caloric and macronutrient estimation by data source and for all combined data sources.
Calories (kcal)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal88.4258.00111.2525.7617.2434.61125.75
95% CIs70.69, 106.5435.00, 77.0076.75, 136.7520.37, 31.418.97, 23.8624.81, 42.84102.82, 147.11
Q1 (≤280)61.6057.0073.0037.1923.8439.2181.64
Q2 (280–368.5)61.0625.0094.5018.547.8228.5889.71
Q3 (368.5–502.5)127.52104.00137.8029.5125.7332.89167.50
Q4 (>502.5)103.4655.00137.5017.029.1222.69143.33
SNAPMeTotal43.0123.5050.0010.135.8011.2466.11
95% CIs32.15, 55.0016.00, 35.5031.00, 62.507.56, 13.064.35, 8.467.29, 14.7448.67, 81.84
Q1 (≤348.2)28.8916.0030.5013.358.4611.7944.82
Q2 (348.2–435.5)33.7819.5030.008.765.047.5757.57
Q3 (435.5–592.8)54.5133.5060.2510.566.5213.2776.65
Q4 (>592.8)55.0036.0060.007.804.336.7779.40
Home-preparedTotal51.9531.0052.0011.919.4711.3978.09
95% CIs31.90, 76.3819.00, 66.0023.00, 73.008.11, 16.254.75, 14.206.69, 20.0742.22, 112.64
Q1 (≤381)31.0721.0038.2513.0211.6512.2539.90
Q2 (381–448)41.0538.6522.179.719.345.1544.02
Q3 (448–489)23.0019.0014.254.854.033.1731.35
Q4 (>489)116.16112.0089.7519.8817.8718.37143.64
All sourcesTotal66.5135.0073.0018.058.9721.03101.12
95% CIs56.06, 77.4525.00, 50.0054.50, 99.0015.01, 21.407.55, 11.6514.74, 27.4185.70, 115.85
Q1 (≤299)44.7423.0056.0025.5515.0929.4867.94
Q2 (299–407)60.6339.0060.0016.8610.1217.1788.28
Q3 (407–533.5)76.3432.1596.5016.527.5922.80118.42
Q4 (>533.5)84.5354.00117.0013.267.0219.40120.60
Proteins (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal6.364.106.6043.1520.6540.769.49
95% CIs4.98, 7.812.85, 5.404.50, 8.7230.08, 59.2317.00, 28.4230.27, 50.927.42, 11.45
Q1 (≤12)3.782.002.9082.1828.5747.926.40
Q2 (12–19)3.413.204.6722.2620.0333.534.44
Q3 (19–28)11.379.4013.4547.6244.8049.9814.49
Q4 (>28)6.994.505.8017.8811.7315.729.51
SNAPMeTotal4.122.305.0221.4410.8214.086.02
95% CIs3.18, 5.171.75, 3.102.70, 7.5914.00, 33.119.16, 14.919.02, 26.714.65, 7.31
Q1 (≤13.2)2.070.801.3634.2010.4213.923.91
Q2 (13.2–22.6)4.142.453.8923.1511.9628.786.05
Q3 (22.6–33.9)5.203.056.3719.0111.8220.257.16
Q4 (>33.9)5.113.705.909.358.6211.526.52
Home-preparedTotal3.132.283.5011.329.299.654.01
95% CIs2.22, 4.201.50, 4.401.80, 4.808.36, 14.775.73, 14.084.77, 12.512.82, 5.22
Q1 (≤18.2)1.401.001.0213.529.299.261.57
Q2 (18.2–27)2.151.500.968.785.643.322.86
Q3 (27–38.8)4.744.801.6514.0415.295.035.11
Q4 (>38.8)4.534.201.078.578.843.905.45
All sourcesTotal5.092.905.4430.8315.0026.777.75
95% CIs4.30, 5.942.30, 3.904.35, 7.0023.28, 39.7111.73, 17.0618.13, 35.366.46, 9.04
Q1 (≤13)3.041.803.3055.4720.0035.005.28
Q2 (13–22)4.702.554.5025.4313.6534.027.27
Q3 (22–33.3)7.404.108.0527.8414.4031.2110.67
Q4 (>33.3)5.684.506.2012.3110.4711.297.44
Carbohydrates (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal7.815.259.9042.5424.9537.2612.00
95% CIs6.07, 9.673.00, 6.806.75, 13.7028.19, 61.6215.78, 29.2930.10, 46.739.5, 14.78
Q1 (≤12)4.981.254.3083.9424.7250.6710.32
Q2 (12–26)7.046.704.1037.5632.3123.338.58
Q3 (26–41)10.248.0511.7028.8523.8531.7913.32
Q4 (>41)9.213.7012.9014.985.8022.5914.90
SNAPMeTotal4.382.903.5713.847.109.657.52
95% CIs3.14, 5.932.32, 3.702.72, 6.108.16, 22.305.35, 8.967.00, 13.114.43, 10.36
Q1 (≤26)4.182.602.8728.0812.0215.238.57
Q2 (26–40)2.632.601.808.338.677.223.33
Q3 (40–52.3)4.782.704.9310.645.779.738.90
Q4 (>52.3)5.784.204.506.836.563.897.59
Home-preparedTotal6.483.707.5917.9812.6819.749.17
95% CIs4.18, 9.281.60, 8.803.81, 11.5011.82, 25.225.56, 23.539.07, 31.435.53, 12.71
Q1 (≤33.5)4.713.704.8525.6725.3018.565.64
Q2 (33.5–41.8)2.641.351.076.623.802.814.01
Q3 (41.8–46.4)11.508.3013.1826.0618.2128.6914.95
Q4 (>550)7.387.204.6212.2912.748.158.58
All sourcesTotal6.343.706.6728.5012.2224.9110.15
95% CIs5.28, 7.512.90, 4.305.70, 8.3020.70, 38.618.64, 14.7920.04, 33.568.24, 12.07
Q1 (≤ 17.9)5.162.205.5063.6424.6245.339.83
Q2 (17.9–34.5)5.273.706.6020.6613.5323.757.10
Q3 (34.5–46.2)7.222.958.5217.997.1122.6311.41
Q4 (>46.2)7.724.416.8011.507.129.0311.66
Lipids (g)
SourceRangeMAEMedAEIQRAEMAPE (%)MedAPE (%)IQRAPE (%)RMSE
Allrecipes.comTotal6.123.007.0734.4116.8542.069.58
95% CIs4.68, 7.641.70, 4.605.00, 12.7026.15, 43.6210.00, 34.6733.70, 55.487.57, 11.42
Q1 (≤10)2.431.302.5036.8026.0044.294.19
Q2 (10–19.5)7.906.009.3554.1538.8963.2011.99
Q3 (19.5–27.2)6.164.607.6525.2517.7830.658.31
Q4 (>27.2)8.234.0513.6722.188.6142.3211.91
SNAPMeTotal4.581.504.2821.6412.2120.218.42
95% CIs3.11, 6.291.10, 2.351.85, 8.2215.88, 28.107.51, 17.3813.73, 32.695.67, 10.96
Q1 (≤11.2)0.950.800.7520.6617.8312.611.16
Q2 (11.2–19.2)3.191.271.7819.329.1212.466.98
Q3 (19.2–30.8)5.672.208.1724.548.6438.259.05
Q4 (>30.8)8.485.4010.6522.0311.8819.9112.29
Home-preparedTotal4.543.334.0030.9524.4427.746.24
95% CIs2.98, 6.261.90, 5.002.10, 10.0819.99, 44.8915.47, 33.9411.22, 43.134.04, 8.16
Q1 (≤10.5)3.581.903.2340.7927.1431.136.46
Q2 (10.5–15)4.973.653.1738.6932.0733.396.41
Q3 (15–21.7)3.653.812.1918.2317.6411.924.00
Q4 (>21.7)6.105.406.5524.4623.0330.207.52
All sourcesTotal5.332.405.7029.1616.4234.248.78
95% CIs4.39, 6.351.70, 3.004.35, 8.2024.06, 34.7311.88, 20.8329.60, 41.937.31, 10.18
Q1 (≤10.6)2.061.001.8032.3520.0033.453.91
Q2 (10.6–18.7)5.092.305.6035.4116.0033.508.98
Q3 (18.7–27.5)5.934.457.7726.2517.7832.958.16
Q4 (>27.5)8.264.4013.2022.6410.1832.8812.07
Note: Values for each independent data source and for all combined data sources (global) are presented as overall results and by quartile. 95% CIs for all reported metrics were estimated via nonparametric bootstrap (10,000 resamples, dish-level).
Figure 1. Scatter plots of estimated vs. reference values for calories and macronutrients by data source.
Figure 1. Scatter plots of estimated vs. reference values for calories and macronutrients by data source.
Nutrients 17 03613 g001aNutrients 17 03613 g001b
Figure 2. Boxplots of absolute errors for calories and macronutrients by data source. Each boxplot displays the median, IQR, and Tukey whiskers (1.5 × IQR). Data points lying beyond the whiskers are presented as outliers.
Figure 2. Boxplots of absolute errors for calories and macronutrients by data source. Each boxplot displays the median, IQR, and Tukey whiskers (1.5 × IQR). Data points lying beyond the whiskers are presented as outliers.
Nutrients 17 03613 g002aNutrients 17 03613 g002b
Figure 3. Linear regression analysis between the number of ingredients and the absolute estimation error (kcal) by data source. The plots display Pearson’s correlation coefficient r with its 95% CIs in brackets, along with the fitted linear regression equation. The shaded band represents the 95% CI, and the ellipse illustrates the joint confidence region between predicted and observed values.
Figure 3. Linear regression analysis between the number of ingredients and the absolute estimation error (kcal) by data source. The plots display Pearson’s correlation coefficient r with its 95% CIs in brackets, along with the fitted linear regression equation. The shaded band represents the 95% CI, and the ellipse illustrates the joint confidence region between predicted and observed values.
Nutrients 17 03613 g003aNutrients 17 03613 g003b
Table 6. Inferential analysis of estimation errors across test cases and data sources.
Table 6. Inferential analysis of estimation errors across test cases and data sources.
CaseVariableKruskal–Wallis
( H ,   p )
Pairwise ComparisonMann–Whitney
( U ,   p )
Holm
( p H o l m )
Cliff s   δ
1Calories (kcal)H = 7.863,  p = 0.020Allrecipes.com vs. SNAPMeU = 3553.5, p = 0.998
Allrecipes.com vs. Home-preparedU = 1644.0,  p = 0.0050.018Large
SNAPMe vs. Home-preparedU = 1219.5,  p = 0.0180.072Large
Proteins (g)H = 0.777, p = 0.678
Carbohydrates (g)H = 2.018, p = 0.365
Lipids (g)H = 1.205, p = 0.547
2Calories (kcal)H = 9.043,  p = 0.011Allrecipes.com vs. SNAPMeU = 3367.5, p = 0.563
Allrecipes.com vs. Home-preparedU = 1616.5,  p = 0.0080.031Large
SNAPMe vs. Home-preparedU = 1286.5,  p = 0.0040.015Large
Proteins (g)H = 0.083, p = 0.960
Carbohydrates (g)H = 0.700, p = 0.705
Lipids (g)H = 0.208, p = 0.901
3Calories (kcal)H = 5.743, p = 0.057
Proteins (g)H = 1.936, p = 0.380
Carbohydrates (g)H = 1.428, p = 0.490
Lipids (g)H = 3.669, p = 0.160
4Calories (kcal)H = 14.130,  p < 0.001Allrecipes.com vs. SNAPMeU = 4716.5,  p < 0.0010.001Large
Allrecipes.com vs. Home-preparedU = 1477.5, p = 0.076
SNAPMe vs. Home-preparedU = 803.0, p = 0.328
Proteins (g)H = 2.005, p = 0.367
Carbohydrates (g)H = 7.753,  p = 0.021Allrecipes.com vs. SNAPMeU = 4373.5,  p = 0.0100.030Small
Allrecipes.com vs. Home-preparedU = 1176.0, p = 0.880
SNAPMe vs. Home-preparedU = 684.0, p = 0.053
Lipids (g)H = 3.229, p = 0.199
Note: p-values were adjusted using Holm’s sequential correction in two stages: first across pairwise comparisons among the three data sources (Allrecipes.com, SNAPMe, and Home-prepared) within each variable, and subsequently across variables (calories, proteins, lipids, and carbohydrates) within each test case to control the family-wise error rate. Only the final adjusted p-values ( p H o l m ) are reported. Values in bold indicate statistically significant differences ( p H o l m < 0.05). Cliff’s δ values range from −1 to +1, where | δ |   0.147, 0.333, and 0.474 correspond to small, medium, and large effect sizes, respectively.

4. Discussion

4.1. Results Interpretation

The results show that ChatGPT-5’s accuracy in estimating caloric and macronutrient content from meal photographs depends strongly on the amount and specificity of contextual information given. As shown in Table 2, Table 3 and Table 4, MAE, MAPE and RMSE decreased systematically across the three first experimental cases. Taking the total energy as an example, MAE fell from 123.03 kcal in Case 1 (image only) to 92.04 kcal in Case 2 (image with non-visual information) and further to 53.33 kcal in Case 3 (image with detailed ingredient list). The corresponding RMSE values decreased from 163.80 kcal to 122.48 kcal and 80.86 kcal, respectively. Likewise, the MAPE showed a substantial reduction, from 30.51% in Case 1 to 24.35% in Case 2 and 13.92% in Case 3. Both absolute and relative errors decreased as more contextual cues were provided.
When compared with Cases 1 and 2, Case 4 (summarized in Table 5) showed markedly lower errors, confirming that detailed ingredient information alone substantially improved estimation accuracy relative to image-only or limited-context inputs. However, when compared with Case 3, where the same textual information was provided but accompanied by the image, accuracy declined once the visual component was removed. For example, the total energy MAE increased from 53.33 kcal in Case 3 to 66.51 kcal in Case 4, while MAPE increased from 13.92% to 18.05% and RMSE from 80.86 kcal to 101.12 kcal. The qualitative analysis of ChatGPT-5’s responses revealed that, without access to the image, the model was sometimes uncertain about whether specific ingredients were raw or cooked, leading to greater variability in its estimations. In contrast, when visual cues were available, the model could infer cooking state and other preparation details, resulting in more accurate and consistent estimations across food items.
Scatter plots in Figure 1 further illustrate the trends. In Case 1, data points were widely dispersed around the identity line (particularly for lipids and proteins), indicating systematic over- and underestimation. As context was added (Cases 2 and 3), the scatter became progressively tighter and more symmetric, with points clustering closely around the y = x line, showing improved agreement between estimated and reference values. In Case 4, dispersion increased in comparison to Case 3. Data points deviated more noticeably from the identity line, reflecting a decline in estimation accuracy. Regarding proteins, Case 4 shows more overestimations than Case 3. This pattern likely arises from ChatGPT-5 interpreting certain protein sources (e.g., meats) as cooked rather than raw. Since such cooked items have a higher protein concentration per unit weight due to water loss, this misinterpretation can lead to inflated protein estimates and systematic overestimations.
Boxplots of absolute errors (Figure 2) corroborate the quantitative findings. MedAE and IQRAE decreased steadily from Case 1 to Case 3 for all variables and data sources. The Home-prepared dataset displayed consistently fewer outliers and, in most cases, narrower distributions, indicating greater consistency and less dispersion in estimation performance. In contrast, data from Allrecipes.com [20] and SNAPMe [11] exhibited broader variability, likely due to inconsistencies in recipe annotation, portion size, and photographic conditions. In all sources, greater errors (outliers) were observed in dishes served in bowls, which may relate to challenges in perceiving depth. Contrasting Case 4 with Case 3, error distributions widened again across most variables. MedAE increased slightly, and the interquartile ranges expanded, reflecting the model’s reduced accuracy when visual cues were removed.
To further investigate the differences in performance among data sources, an inferential analysis was conducted, as described in Section 2.5. As summarized in Table 6, significant differences in estimation errors among data sources were found primarily for caloric values. In Cases 1 and 2, the Allrecipes.com dataset produced significantly higher absolute errors compared to the Home-prepared dataset ( p H o l m   = 0.018 and 0.031, respectively). SNAPMe also differed from Home-prepared in both cases ( p H o l m = 0.072 and 0.015, respectively). These contrasts were associated with large effect sizes (Cliff’s δ > 0.474), indicating substantial disparities in performance across datasets. In contrast, macronutrient errors (proteins, carbohydrates, and lipids) did not show statistically significant differences across data sources in Cases 1–3.
Although total calories are computed from macronutrient content, the estimation errors for proteins, carbohydrates, and lipids do not necessarily vary in the same direction. Overestimation in one macronutrient (e.g., carbohydrates) may be offset by underestimation in another (e.g., lipids), resulting in a net caloric error that is smaller and more consistent across samples. This compensatory behavior can reduce variability in total energy estimates while keeping macronutrient-specific errors high and uneven, which diminishes the likelihood of detecting statistically significant differences for individual nutrients. Consequently, the overall caloric estimation becomes a more stable indicator of performance differences among data sources.
In Case 3, no significant differences were detected between sources, since the inclusion of detailed ingredient lists and quantities homogenized the input information and reduced between-source variability. Under these conditions, ChatGPT-5 achieved uniformly low errors across all datasets, indicating that sufficient contextual detail allows the model to reach a consistent estimation performance regardless of the data source.
Unlike Case 3, Case 4 exhibited statistically significant differences for calories and carbohydrates, with Allrecipes.com showing higher errors than SNAPMe ( p H o l m = 0.001 and 0.030, correspondingly). These results suggest that, when visual information was removed, variability between datasets increased again, particularly affecting energy and carbohydrate estimations. In Case 4, ChatGPT-5 appeared more sensitive to the structure of textual descriptions, as visual information was unavailable. In particular, the sometimes more narrative format of Allrecipes.com seems to make ingredient interpretation more ambiguous, leading to higher estimation errors compared with the more organized entries from SNAPMe.
The relationship between dish complexity (measured as the number of ingredients) and the absolute caloric estimation error was analyzed through Pearson’s correlation coefficient and linear regression across the experimental cases and data sources. Overall, weak positive correlations ( r   0.37 ) were observed in all cases (Figure 3), indicating that the estimation error tended to increase slightly as the number of ingredients grew, but the effect remained modest. This suggests that while greater ingredient diversity may add visual and contextual complexity, it does not substantially compromise ChatGPT-5’s ability to generate nutritional estimates.
Across test cases, the correlation strength varied mildly: the lowest values were found in Case 2, reflecting the stabilizing effect of non-visual contextual information, while Case 3 showed higher coefficients, likely because detailed ingredient lists in combination with food images exposed residual inconsistencies in complex dishes. Despite these slight variations, the regression slopes remained low in all cases, and the confidence ellipses progressively narrowed from Case 1 to Case 3, consistent with the overall reduction in MAE and RMSE. These results confirm that as contextual information becomes richer, ChatGPT-5’s estimates grow more reliable. In Case 4, correlations decreased slightly when compared with Case 3. When analyzed by data source, the Home-prepared dataset exhibited the weakest correlations, denoting consistent performance even for multi-ingredient meals. In contrast, Web-based and SNAPMe datasets displayed broader variability and slightly higher regression slopes.

4.2. Comparison with Related Work

The findings of this study confirm the potential of ChatGPT-5 for dietary assessment from meal photographs, especially when supported by complementary information. This aligns with prior work by Zheng [4] and Ege [9], who concluded that AI models achieve higher accuracy when provided with contextual information such as ingredients or cooking methods. In particular, Ege [9] showed that integrating such variables reduces relative error and strengthens correlations with reference values.
Notably, although ChatGPT-5 is not specifically designed for nutritional assessment, its performance was comparable to that of models explicitly trained for this purpose. Fang [7] reported an average error of 209 kcal, whereas ChatGPT-5 achieved a MAE of 123.03 kcal using images alone. When full contextual information was incorporated, the error decreased to 53.33 kcal, approaching the 37.9 kcal reported by Wang [10] with a more advanced architecture.
In contrast to the findings of Sultana [6] and Wang [10], who reported major difficulties for AI models in assessing multi-ingredient dishes, our study identified only a weak positive correlation between the number of ingredients and absolute error. This indicates that dish complexity, in terms of component count, is not a critical determinant of ChatGPT-5’s estimations. Rather, elements such as image quality, level of detail, and the inclusion of reference objects seem to have a stronger influence on accuracy, consistent with observations by Akpa [8] and Ege [9].
When contrasted with the recent evaluation of ChatGPT-4 by O’Hara et al. [32], our findings provide complementary evidence. Their study assessed 114 standardized meal photographs from the Irish National Adult Nutrition Survey (NANS), focusing on the model’s ability to identify foods and estimate nutrient content. ChatGPT-4 achieved high accuracy in food recognition (accuracy 93%, recall 84.6%, harmonic mean 88.6%). For nutrient estimation, correlation with reference values ranged from weak (0.29) to strong (0.83), indicating that the model’s performance varied depending on the nutrient considered. Nutrient underestimation was frequent, with a MAPE of 26.9% and marked discrepancies for certain micronutrients, such as vitamin D (−100%) and potassium (−49.5%). Our framework, which incorporated contextual information, yielded substantially lower errors for energy and macronutrients, suggesting that enriched prompts and detailed input data can mitigate systematic underestimations observed in baseline evaluations.
While the study by O’Hara et al. [32] shows the feasibility of applying ChatGPT-4 to dietary assessment, it was constrained by reliance on a single, relatively homogeneous dataset focused on Irish meal types. In contrast, the present work incorporates a more heterogeneous database, combining meals prepared and annotated by nutritionists, images from the SNAPMe repository [11], and recipe photographs sourced from the web. In addition, the latest version of ChatGPT was utilized, and the model was systematically evaluated under different levels of contextual information. Overall, the approach of this study not only allows for a more realistic and culturally diverse assessment but also offers a clearer understanding of how non-visual information contributes to the model’s accuracy.
An important observation concerns the influence of data source on estimation accuracy. Photographs of Home-prepared meals, annotated with directly measured ingredients and standardized equivalence tables, consistently showed lower errors, whereas external sources (i.e., Allrecipes.com and SNAPMe) displayed greater variability. This supports Larke’s [11] observation that databases built by individuals without formal training in nutrition may contain inaccuracies in portion sizes, ingredient descriptions, or nutritional values.
Although the employed dataset underwent a curation process to remove inconsistent entries and standardize nutritional information, residual inaccuracies cannot be fully excluded. The higher errors observed in external sources may partly reflect inaccuracies in the ground truth. During database development, several records had to be discarded or corrected due to clearly erroneous annotations, revealing issues in data collection. In the case of SNAPMe, limitations of the data-entry application and user inexperience often resulted in inaccurate quantities or the inclusion of false ingredients. Likewise, nutritional information obtained from Allrecipes.com frequently failed to match the portion sizes or ingredients displayed in the images. In addition, user-uploaded web photographs exhibited greater heterogeneity in angle, lighting, and overall quality, further contributing to variability in estimation accuracy.

4.3. Practical Application of ChatGPT

As a complementary outcome of this work and to exemplify the versatility and ease of use of ChatGPT-5, a customized version of the model was developed, tailored for dietary assessment from dish photographs. The process is straightforward and requires only simple instructions in natural language. Details of its creation and an illustrative example are provided in Appendix C.
Building upon the experience gained throughout this study, the customized model is designed to prompt users for all available contextual information, such as ingredient lists, portion sizes, and preparation methods. Yet, it retains the capability to generate estimates in the absence of any accompanying metadata, i.e., when only an image is provided.

4.4. Open Issues

  • Higher estimation errors were consistently observed for dishes served in bowls, likely reflecting ChatGPT-5’s limited ability to infer depth and volume from monocular images. Future work could evaluate the effectiveness of including a vertically oriented reference object placed beside the dish to improve volumetric estimation.
  • Since ChatGPT-5 is a pre-trained, closed-weight model, conventional validation schemes such as holdout test sets or k-fold cross-validation cannot be directly applied. The model’s parameters are fixed and not updated during use, so its outputs reflect generalization learned during large-scale pretraining rather than task-specific adaptation. Nevertheless, such validation procedures would be highly relevant in the context of fine-tuned or customized derivatives of ChatGPT-5, as described in Appendix C. For these tailored versions, systematic partitioning of annotated data into training, validation, and test sets (or stratified k-fold cross-validation) would enable quantitative assessment of generalizability and overfitting. Furthermore, iterative prompt engineering and supervised fine-tuning on domain-specific image–nutrient pairs could improve accuracy, reduce variance across food types, and enhance reliability for practical dietary assessment applications.
  • A major limitation for future research is the absence of broad, high-quality, and systematically annotated datasets linking food images to verified nutritional composition. The development of such a database, curated by nutrition professionals, would provide a reliable foundation for both model evaluation and fine-tuning. At Universidad Europea del Atlántico, such an initiative is being considered, involving active participation of students in Human Nutrition and Dietetics or related disciplines. Leveraging their technical training and the institution’s multicultural environment would ensure both the accuracy and diversity of the dataset, enhancing its representativeness for real-world dietary applications.
  • Future research should include direct benchmarking of ChatGPT-5’s nutritional estimations against both traditional computer vision models and professional dietitian assessments. Such head-to-head evaluations would provide a clearer understanding of the model’s relative strengths and limitations, quantify its added value beyond established analytical frameworks, and support evidence-based assessment of its potential integration into clinical nutrition and dietary evaluation workflows.

5. Conclusions

The main objective of this study was to evaluate the accuracy of ChatGPT-5 in estimating caloric content and macronutrient distribution from meal photographs. A total of 195 dishes were analyzed. The imagery and nutritional information were obtained from three sources: two external datasets (Allrecipes.com [20] and SNAPMe [11]) and one internally developed by nutritionists. Four scenarios with varying types and levels of contextual information were tested, showing that accuracy improved consistently when visual inputs were complemented with additional non-visual cues and detailed ingredient descriptions. However, the decline in accuracy observed when the image was omitted, despite the inclusion of detailed ingredient information, indicates that visual cues contribute markedly to estimation performance, and that the observed improvements cannot be attributed solely to arithmetic derived from ingredient lists. Dish complexity, measured by the number of ingredients, showed a weak positive correlation, suggesting that complexity alone is not decisive.
The lowest errors were consistently observed in self-prepared and annotated meals, whereas external sources exhibited greater variability. This finding underscores the limitations of relying on datasets created without nutritional expertise and the need for professionally curated image-based resources. Developing robust and multicultural databases under the supervision of nutrition professionals is, therefore, essential to ensure accuracy, reduce variability, and make the evaluation of AI models (such as ChatGPT-5) in dietary assessment fairer and more representative.
In summary, this study provides novel evidence supporting a potential complementary role of ChatGPT-5 within clinical nutrition workflows. However, confirmatory head-to-head evaluations against registered dietitians and validated assessment tools are still needed before clinical adoption. Although limitations persist, particularly with external data sources, the findings emphasize its potential as a complementary tool for dietary evaluation, with promising applications in both clinical practice and nutrition research.

Author Contributions

Conceptualization, M.R.-J., S.S.-C. and I.E.; methodology, G.D.M.-d.-C.-B., M.R.-J. and J.C.-Á.; software, G.D.M.-d.-C.-B. and J.C.-Á.; validation, S.S.-C., I.E. and M.R.-J.; formal analysis, G.D.M.-d.-C.-B. and J.C.-Á.; investigation, M.R.-J.; resources, S.S.-C. and I.E.; data curation, M.R.-J. and G.D.M.-d.-C.-B.; writing—original draft preparation, M.R.-J.; writing—review and editing, G.D.M.-d.-C.-B., S.S.-C., J.C.-Á. and I.E.; visualization, G.D.M.-d.-C.-B. and J.C.-Á.; supervision, S.S.-C. and I.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

No human participant data were collected in this work; therefore, the study is exempt from Institutional Review Board review. All food images used in the analyses were either publicly available or photographs created by the authors, ensuring that no privacy or ethical concerns are involved.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
APIApplication Programming Interface
CIConfidence Interval
DLDeep Learning
GPTGenerative Pre-trained Transformer
IQRInterquartile Range
IQRAEIQR of Absolute Errors
IQRAPEIQR of Absolute Percentage Errors
MAEMean Absolute Error
MedAEMedian Absolute Error
MAPEMean Absolute Percentage Error
MedAPEMedian Absolute Percentage Error
NANSNational Adult Nutrition Survey
RMSERoot Mean Square Error
SMAEMexican System of Food Equivalents
SNAPMeSurveying Nutrient Assessment with Photographs of Meals
USDAU.S. Department of Agriculture

Appendix A

The database for this study comprises photographs, reference nutritional values, and ingredient lists for 195 dishes. Table A1 illustrates its structure, providing an example of the sources from which the data were collected, namely, Allrecipes.com [20], SNAPMe [11], and Home-prepared.
Table A1. Database organization.
Table A1. Database organization.
Image Ingredients Energy (kcal) Proteins (g) Lipids (g) Carbohydrates (g) 
Nutrients 17 03613 i001
Allrecipes.com [20]
¼ cup vegetable oil; 3 corn tortillas (6 inches) cut into strips; ¾ dried pasilla chile (seeded); ½ tomato (seeded and chopped); ⅛ large onion (chopped); ¼ garlic clove; a pinch of dried oregano (≈1/16 teaspoon); ¼ teaspoon vegetable oil; 1 cup of water; 1 teaspoon chicken bouillon granules; ¼ sprig fresh parsley; ¼ avocado (peeled, pitted, and diced); 3 tablespoons cotija cheese (diced); ½ tablespoon crème fraîche (or Mexican crema).944528013
Nutrients 17 03613 i002
SNAPMe [11]
120 g breaded chicken fillet; 15 g honey mustard dip; 180 g potato, French fries (baked); 13.5 g olive oil; 66.6 g French bread; 5 g salted butter stick.960295595
Nutrients 17 03613 i003
Home-prepared
82 g grilled beef; 200 g boiled potatoes; 170 g baked pumpkin; ½ teaspoon coconut oil; ½ teaspoon olive oil.45938.8411.0058.60

Appendix B

This appendix presents an example of the application of ChatGPT-5 to each case study described in Section 2.6. The Home-prepared dish from Appendix A is used as the basis.
Figure A1. Use of ChatGPT-5 for test case 1 (image only).
Figure A1. Use of ChatGPT-5 for test case 1 (image only).
Nutrients 17 03613 g0a1
Figure A2. Use of ChatGPT-5 for test case 2 (image with non-visual data).
Figure A2. Use of ChatGPT-5 for test case 2 (image with non-visual data).
Nutrients 17 03613 g0a2
Figure A3. Use of ChatGPT-5 for test case 3 (image with detailed ingredient information).
Figure A3. Use of ChatGPT-5 for test case 3 (image with detailed ingredient information).
Nutrients 17 03613 g0a3
Figure A4. Use of ChatGPT-5 for test case 4 (detailed ingredient information only).
Figure A4. Use of ChatGPT-5 for test case 4 (detailed ingredient information only).
Nutrients 17 03613 g0a4

Appendix C

The observations derived from the analysis in this study lay the foundation for the development of NutriFoto, a customized Generative Pre-trained Transformer (GPT) designed for dietary assessment based on meal photographs. ChatGPT permits the creation of personalized versions of the GPT model through a user-friendly interface [33]. The steps for generating a customized GPT are described below, followed by the specific instructions used to develop NutriFoto. Finally, an example of its application is presented.

Appendix C.1. Steps to Create a Customized GPT

  • Access to the creation function. Log into chat.openai.com with a Plus account. In the side panel, select the “Explore GPTs” option, then click “Create a GPT” to launch the setup assistant.
  • Creation assistant. The system guides the user through a step-by-step wizard, in which the purpose of the GPT is described. Based on this information, the assistant generates an initial configuration that can be modified.
  • Customization of instructions. The user may edit the system instructions, which define the GPT’s behavior. These instructions can include, for example, the tone of the responses (formal, colloquial, etc.) and the area of specialization (e.g., nutrition).
  • Inclusion of tools and resources (optional). The GPT’s functionality can be expanded through the incorporation of additional resources, such as files or external links.
  • Model testing. Before finalizing the process, the GPT’s behavior can be tested within the platform. This stage allows for adjustments to the instructions or the correction of any issues identified during testing.
  • Publication and use. Finally, the GPT can be private (for exclusive use by the creator), shared via link (accessible only to recipients), or public (available in the GPT gallery for all users). A distinctive name, brief description, and logo may also be added.

Appendix C.2. Instructions Used to Create NutriFoto

The following instructions correspond to Step 3 of Appendix C.1, where the GPT’s behavior is defined and adapted to the intended purpose.
  • This GPT is a nutrition-specialized assistant designed to calculate the calories and macronutrients (protein, lipids, and carbohydrates) of dishes based on photographs uploaded by the user. Its objective is to provide an estimate based on image analysis, with optional support from additional user-provided data.
  • The photograph is the only essential requirement. The assistant should greet the user, request the image, and suggest including a reference object (such as a ruler, spoon, fork, or chopsticks) to improve accuracy. If no reference object is present, it will proceed with the image and inform the user that the estimate may be less accurate.
  • Before providing any estimate, the assistant should gather as much contextual information as possible, in the following order of priority: type and amount of fats, type of meat, type and content of sweeteners, fat content of dairy products, cooking method, individual ingredients (if the user can provide them), and the name of the dish (if known). If the image contains wrapped or stuffed foods, it should suggest (only when feasible) cutting them in half to view the contents.
  • Once the information is collected, the assistant will present the results in a clear table with the fields: ingredients (if known), calories, protein, lipids, and carbohydrates. The table will be accompanied by a brief interpretation written in accessible language for the user. The tone should be warm, educational, and non-judgmental.

Appendix C.3. Illustrative Application of NutriFoto

This section provides an illustrative example of NutriFoto, showing how the tool processes a meal photograph and generates nutritional estimates based on both visual and contextual information. Figure A5 displays the input interface of NutriFoto. Figure A6 then presents the model’s initial response, where contextual information is requested from the user. Once these data are supplied, NutriFoto generates a structured output including estimated caloric content and macronutrient distribution, as shown in Figure A7.
In the illustrative example, the reference values were 557 kcal, 23.5 g of protein, 64.6 g of carbohydrates, and 23 g of lipids. To refine the estimation, additional essential information was provided, including type and amount of fats, type of meat, and cooking method.
NutriFoto has been made publicly available and can be accessed online at: https://chatgpt.com/g/g-680616e7ec848191aebe3495da3d5587-nutrifoto (accessed on 13 November 2025).
Figure A5. Input interface of the custom GPT NutriFoto.
Figure A5. Input interface of the custom GPT NutriFoto.
Nutrients 17 03613 g0a5
Figure A6. NutriFoto requesting additional information.
Figure A6. NutriFoto requesting additional information.
Nutrients 17 03613 g0a6
Figure A7. Result of the nutritional evaluation from NutriFoto.
Figure A7. Result of the nutritional evaluation from NutriFoto.
Nutrients 17 03613 g0a7

References

  1. Dhawan, D.; Sharma, S. Abdominal Obesity, Adipokines and Non-communicable Diseases. J. Steroid Biochem. Mol. Biol. 2020, 203, 105737. [Google Scholar] [CrossRef]
  2. Bailey, R.L. Overview of dietary assessment methods for measuring intakes of foods, beverages, and dietary supplements in research studies. Curr. Opin. Biotechnol. 2021, 70, 91–96. [Google Scholar] [CrossRef]
  3. Boushey, C.J.; Spoden, M.; Zhu, F.M.; Delp, E.J.; Kerr, D.A. New mobile methods for dietary assessment: Review of image-assisted and image-based dietary assessment methods. Proc. Nutr. Soc. 2017, 76, 283–294. [Google Scholar] [CrossRef] [PubMed]
  4. Zheng, J.; Wang, J.; Shen, J.; An, R. Artificial Intelligence Applications to Measure Food and Nutrient Intakes: Scoping Review. J. Med. Internet. Res. 2024, 26, e54557. [Google Scholar] [CrossRef] [PubMed]
  5. Shonkoff, E.; Cara, K.C.; Pei, X.; Chung, M.; Kamath, S.; Panetta, K.; Hennessy, E. AI-based digital image dietary assessment methods compared to humans and ground truth: A systematic review. Ann. Med. 2023, 55, 2273497. [Google Scholar] [CrossRef] [PubMed]
  6. Sultana, J.; Ahmed BMd Masud, M.M.; Huq, A.K.O.; Ali, M.E.; Naznin, M. A Study on Food Value Estimation From Images: Taxonomies, Datasets, and Techniques. IEEE Access 2023, 11, 45910–45935. [Google Scholar] [CrossRef]
  7. Fang, S.; Shao, Z.; Kerr, D.A.; Boushey, C.J.; Zhu, F. An End-to-End Image-Based Automatic Food Energy Estimation Technique Based on Learned Energy Distribution Images: Protocol and Methodology. Nutrient 2019, 11, 877. [Google Scholar] [CrossRef]
  8. Akpa, E.A.H.; Suwa, H.; Arakawa, Y.; Yasumoto, K. Smartphone-Based Food Weight and Calorie Estimation Method for Effective Food Journaling. SICE J. Control Meas. Syst. Integr. 2017, 10, 360–369. [Google Scholar] [CrossRef]
  9. Ege, T.; Yanai, K. Image-Based Food Calorie Estimation Using Knowledge on Food Categories, Ingredients and Cooking Directions. In Proceedings of the on Thematic Workshops of ACM Multimedia 2017, Mountain View, CA, USA, 23–27 October 2017; ACM: New York, NY, USA; pp. 367–375. Available online: https://dl.acm.org/doi/10.1145/3126686.3126742 (accessed on 15 March 2025).
  10. Wang, H.; Tian, H.; Ju, R.; Ma, L.; Yang, L.; Chen, J.; Liu, F. Nutritional composition analysis in food images: An innovative Swin Transformer approach. Front. Nutr. 2024, 11, 1454466. [Google Scholar] [CrossRef]
  11. Larke, J.A.; Chin, E.L.; Bouzid, Y.Y.; Nguyen, T.; Vainberg, Y.; Lee, D.H.; Pirsiavash, H.; Smilowitz, J.T.; Lemay, D.G. Surveying Nutrient Assessment with Photographs of Meals (SNAPMe): A Benchmark Dataset of Food Photos for Dietary Assessment. Nutrients 2023, 15, 4972. [Google Scholar] [CrossRef]
  12. Konstantakopoulos, F.S.; Georga, E.I.; Fotiadis, D.I. An Automated Image-Based Dietary Assessment System for Mediterranean Foods. IEEE Open J. Eng. Med. Biol. 2023, 4, 45–54. [Google Scholar] [CrossRef]
  13. Hattab, S.; Badrasawi, M.; Anabtawi, O.; Zidan, S. Development and validation of a smartphone image-based app for dietary intake assessment among Palestinian undergraduates. Sci. Rep. 2022, 12, 15467. [Google Scholar] [CrossRef] [PubMed]
  14. Shao, W.; Min, W.; Hou, S.; Luo, M.; Li, T.; Zheng, Y.; Jiang, S. Vision-based food nutrition estimation via RGB-D fusion network. Food Chem. 2023, 424, 136309. [Google Scholar] [CrossRef] [PubMed]
  15. Lee, L.; Hall, R.; Krebs, J. Tailored prompting to improve adherence to image-based dietary assessment: Mixed methods study. JMIR Mhealth Uhealth 2024, 12, e52074. [Google Scholar] [CrossRef] [PubMed]
  16. OpenAI. Introducing GPT-5. San Francisco (CA): OpenAI. Available online: https://openai.com/es-ES/index/introducing-gpt-5/ (accessed on 16 November 2025).
  17. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. arXiv 2021, arXiv:2103.00020. [Google Scholar] [CrossRef]
  18. Driess, D.; Xia, F.; Sajjadi, M.S.M.; Lynch, C.; Chowdhery, A.; Ichter, B.; Wahid, A.; Tompson, J.; Vuong, Q.; Yu, T.; et al. PaLM-E: An Embodied Multimodal Language Model. arXiv 2023, arXiv:2303.03378. [Google Scholar]
  19. Peng, Z.; Wang, W.; Dong, L.; Hao, Y.; Huang, S.; Ma, S.; Wei, F. Kosmos-2: Grounding Multimodal Large Language Models to the World. arXiv 2023, arXiv:2306.14824. [Google Scholar] [CrossRef]
  20. Allrecipes. Allrecipes|Recipes, How-Tos, Videos and More. Available online: https://www.allrecipes.com/ (accessed on 30 March 2025).
  21. Mataix Verdú, J. Tabla de Composición de Alimentos, 5th ed.; Universidad de Granada: Granada, Spain, 2009. [Google Scholar]
  22. Pérez Lizaur, A.B. SMAE Sistema Mexicano de Alimentos Equivalentes, 5th ed.; Fomento de Nutrición y Salud: Delegación Cuauhtémoc, Mexico, 2022; 209p. [Google Scholar]
  23. USDA. Available online: https://www.usda.gov/ (accessed on 31 August 2025).
  24. Hewamalage, H.; Ackermann, K.; Bergmeir, C. Forecast evaluation for data scientists: Common pitfalls and best practices. Data Min. Knowl. Discov. 2023, 37, 788–832. [Google Scholar] [CrossRef]
  25. Passarelli, S.; Free, C.M.; Allen, L.H.; Batis, C.; Beal, T.; Biltoft-Jensen, A.P.; Bromage, S.; Cao, L.; Castellanos-Gutiérrez, A.; Christensen, T.; et al. Estimating national and subnational nutrient intake distributions of global diets. Am. J. Clin. Nutr. 2022, 116, 551–560. [Google Scholar] [CrossRef]
  26. Tukey, J.W. Exploratory Data Analysis, 1st ed.; Addison-Wesley: Boston, MA, USA, 1977; 688p. [Google Scholar]
  27. Zar, J.H. Biostatistical Analysis, 5th ed.; Pearson new international edition; Pearson Education Limited: Harlow, UK, 2014; 756p. [Google Scholar]
  28. Dohms, P.O.D.S.; Stangarlin-Fiori, L.; Medeiros, C.O. Preparations offered to workers in different food services: Application of the score for qualitative assessment of preparations. Front. Nutr. 2024, 11, 1354841. [Google Scholar] [CrossRef]
  29. Gray, V.; Gustin, L.; Jensen, L.; Smith, K. Validating the Made-From-Scratch Versus Traditional Style Approach for Improving the Nutritional Quality of Heart Start Menus. ICAN Infant Child Adolesc Nutr. 2015, 7, 355–366. [Google Scholar] [CrossRef]
  30. Rempelos, L.; Wang, J.; Barański, M.; Watson, A.; Volakakis, N.; Hadall, C.; Hasanaliyeva, G.; Chatzidimitriou, E.; Magistrali, A.; Davis, H.; et al. Diet, but not food type, significantly affects micronutrient and toxic metal profiles in urine and/or plasma; a randomized, controlled intervention trial. Am. J. Clin. Nutr. 2022, 116, 1278–1290. [Google Scholar] [CrossRef]
  31. Rowley, C.E.; Lodge, S.; Egan, S.; Itsiopoulos, C.; Christophersen, C.T.; Silva, D.; Kicic-Starcevich, E.; O’Sullivan, T.A.; Wist, J.; Nicholson, J.; et al. Altered dietary behaviour during pregnancy impacts systemic metabolic phenotypes. Front. Nutr. 2023, 10, 1230480. [Google Scholar] [CrossRef]
  32. O’Hara, C.; Kent, G.; Flynn, A.C.; Gibney, E.R.; Timon, C.M. An Evaluation of ChatGPT for Nutrient Content Estimation from Meal Photographs. Nutrients 2025, 17, 607. [Google Scholar] [CrossRef]
  33. Introducing GPTs. 2024. Available online: https://openai.com/index/introducing-gpts/ (accessed on 23 March 2025).
Table 1. Key considerations from the literature.
Table 1. Key considerations from the literature.
TopicReferencesConsiderations
Model PerformanceSultana et al. [6], Akpa et al. [8], Ege et al. [9].Use of reference objects (e.g., cutlery and chopsticks) to improve food volume and depth estimation.
Zheng et al. [4], Fang et al. [7], Wang et al. [10].Difficulty in identifying ingredients in complex dishes. Incorporation of contextual information (e.g., type and content of fat and sweeteners) is recommended.
Akpa et al. [8], Ege et al. [9], Wang et al. [10].Estimates affected by camera angle, lighting, occlusions, tableware color, and image resolution.
Evaluation metrics and methodsShonkoff et al. [5].Absolute and relative error, Root Mean Square Error (RMSE).
Ege et al. [9].Pearson correlation coefficient and regression analysis.
Wang et al. [10].Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and visual analysis using scatter plots.
DatabasesZheng et al. [4], Sultana et al. [6], Ege et al. [9].Use of images from social media and websites to represent diverse cuisines, including real-life scenarios and multi-ingredient dishes.
Larke et al. [11].SNAPMe: publicly available database of food photographs with portion size and nutritional annotations.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodríguez-Jiménez, M.; Martín-del-Campo-Becerra, G.D.; Sumalla-Cano, S.; Crespo-Álvarez, J.; Elio, I. Image-Based Dietary Energy and Macronutrients Estimation with ChatGPT-5: Cross-Source Evaluation Across Escalating Context Scenarios. Nutrients 2025, 17, 3613. https://doi.org/10.3390/nu17223613

AMA Style

Rodríguez-Jiménez M, Martín-del-Campo-Becerra GD, Sumalla-Cano S, Crespo-Álvarez J, Elio I. Image-Based Dietary Energy and Macronutrients Estimation with ChatGPT-5: Cross-Source Evaluation Across Escalating Context Scenarios. Nutrients. 2025; 17(22):3613. https://doi.org/10.3390/nu17223613

Chicago/Turabian Style

Rodríguez-Jiménez, Marcela, Gustavo Daniel Martín-del-Campo-Becerra, Sandra Sumalla-Cano, Jorge Crespo-Álvarez, and Iñaki Elio. 2025. "Image-Based Dietary Energy and Macronutrients Estimation with ChatGPT-5: Cross-Source Evaluation Across Escalating Context Scenarios" Nutrients 17, no. 22: 3613. https://doi.org/10.3390/nu17223613

APA Style

Rodríguez-Jiménez, M., Martín-del-Campo-Becerra, G. D., Sumalla-Cano, S., Crespo-Álvarez, J., & Elio, I. (2025). Image-Based Dietary Energy and Macronutrients Estimation with ChatGPT-5: Cross-Source Evaluation Across Escalating Context Scenarios. Nutrients, 17(22), 3613. https://doi.org/10.3390/nu17223613

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop