Next Article in Journal
A Critical Appraisal of the Links Between Video Gaming, Lifestyle Factors, Diet and Eating Behaviour: A Narrative Review
Next Article in Special Issue
Enhancing Hospital Nutrition Assessment Through Artificial Intelligence: A Prospective Tray-Level Pilot Study
Previous Article in Journal
Impact of Nutritional Factors on Length of Hospital Stay and Readmission Risk in a Reference Unit for Eating Disorders
Previous Article in Special Issue
Application of Artificial Intelligence Technologies as an Intervention for Promoting Healthy Eating and Nutrition in Older Adults: A Systematic Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accuracy of AI-Based Nutrient Estimation from Standardized Hospital Meal Images: A Comparison with Registered Dietitians

1
Graduate School of Comprehensive Scientific Research, Prefectural University of Hiroshima, Hiroshima 734-8558, Japan
2
Itsukaichi Memorial Hospital, Hiroshima 731-5156, Japan
3
Faculty of Regional Innovation, Prefectural University of Hiroshima, Hiroshima 734-8558, Japan
4
Hiroshima Red Cross Hospital & Atomic-Bomb Survivors Hospital, Hiroshima 730-8619, Japan
5
Faculty of Health Sciences, Hiroshima Shudo University, Hiroshima 731-3195, Japan
*
Author to whom correspondence should be addressed.
Nutrients 2026, 18(6), 966; https://doi.org/10.3390/nu18060966
Submission received: 1 February 2026 / Revised: 2 March 2026 / Accepted: 5 March 2026 / Published: 18 March 2026
(This article belongs to the Special Issue A Path Towards Personalized Smart Nutrition)

Abstract

Background: Accurate dietary assessment is vital for preventing malnutrition in aging populations, particularly in home-care settings. Although Large Multimodal Models (LMMs) for nutrient estimation are evolving, their nutrient-specific accuracy requires rigorous validation. Methods: Fifteen standardized hospital meals were photographed under controlled conditions (90-degree angle, 500 lux). Ground truth values were determined by direct weighing. Estimates for energy and macronutrients were performed by 10 registered dietitians (RDs) and 10 AI models (including ChatGPT-4o and Gemini 1.5 Pro). Accuracy was assessed using Pearson’s correlation, Mean Absolute Error (MAE), and Bland–Altman analysis to quantify systematic bias. Results: For energy and carbohydrates, RDs and top-performing AI models (notably ChatGPT-4o and Gemini 1.5 Pro) demonstrated practical accuracy (r > 0.8, frequently within ±10% range). However, accuracy for protein and lipids was significantly lower across all AI models. Specifically, all AI models exhibited a substantial systematic overestimation of lipids (Mean Bias > +20%, p < 0.01), highlighting a critical “invisible nutrient” bias. Conclusions: Current AI tools show potential for caloric and carbohydrate monitoring but struggle with lipid and protein density. These findings emphasize the need for human–AI collaboration (“human-in-the-loop”) and the integration of cooking metadata to improve clinical utility in geriatric nutrition.

1. Introduction

The progression of aging in Japan has made malnutrition among community-dwelling and home-care older adults a critical public health challenge [1,2,3]. By 2026, the elderly population in Japan is projected to reach approximately 30%, intensifying the shift in medical care from hospitals to home-care settings [1]. Recent meta-analyses from 2025 indicate that the global prevalence of malnutrition in older adults remains high, at approximately 18% [4]. In community-dwelling older adults, age-related functional decline, chronic diseases, and social factors often converge, leading to a significant proportion of individuals suffering from malnutrition or being at high risk [4,5,6,7]. Similar trends have been observed among patients receiving home care, where nutritional management is often fragmented [5,6]. Since malnutrition is directly associated with physical decline and poor clinical outcomes, such as increased mortality and morbidity [6,8], continuous and accurate nutritional assessment outside of hospital environments is urgently required.
Although countermeasures for malnutrition in the elderly are highly recommended, they are often insufficiently implemented in clinical practice [9]. In many clinical settings, screening tools like the Mini Nutritional Assessment-Short Form (MNA-SF) are utilized; however, their agreement with the Global Leadership Initiative on Malnutrition (GLIM) criteria is not always sufficient, which may lead to an underestimation of nutritional risk [10]. While nutritional interventions, including oral nutritional supplements (ONS), are effective in improving status and function [9,11], accurately evaluating actual dietary intake remains a significant hurdle. Traditional dietary records or 24 h recalls impose a substantial burden on both patients and caregivers, leading to potential inaccuracies in reporting and low compliance [11,12]. Therefore, a low-burden, objective, and automated method for capturing dietary intake is necessary to bridge this gap.
Recently, artificial intelligence (AI) technology for nutritional assessment via food images has advanced rapidly [13]. The integration of advanced image recognition and 2026-generation large multimodal models (LMMs) has enabled automated food identification and nutrient estimation, showing high feasibility for digital nutritional interventions even among older populations [14]. However, previous studies have reported that these methods are prone to systematic errors arising from meal complexity and nutrient characteristics. While visually prominent components, such as staple foods (carbohydrates) whose volume is easily captured by computer vision, are estimated with relative accuracy, the precision tends to decrease for fats, oils, seasonings, and mixed dishes [15,16]. Evaluations across multiple models have shown significant variability and a tendency for overestimation in certain contexts [17], emphasizing the need for expert oversight even when agreement rates are high [18]. Moreover, the nutritional adequacy of AI-generated diet plans remains a challenge [19], as do the ethical and practical considerations of AI implementation in clinical nutrition practice [20]. Particularly for older adults, maintaining precise nutrient intake is crucial for preventing sarcopenia and maintaining physical function [21].
Based on these findings, it is expected that AI-based nutritional estimation accuracy depends heavily on specific nutrient characteristics and the visibility of the food matrix. However, research directly comparing multiple 2026-generation AI models with professional registered dietitians (RDs) using standardized meals—where “ground truth” is precisely known—remains limited.
Our hypothesis is that AI-based nutritional estimation will demonstrate high agreement with ground truth for energy and carbohydrates due to their volumetric visibility, whereas the agreement will significantly decrease for proteins and lipids due to the inherent difficulty of visual-only density estimation and the “invisible” nature of cooking oils. Therefore, the objective of this study was to compare the nutrient estimations of 10 different AI models and experienced RDs using photographs of standardized hospital meals to verify nutrient-specific accuracy characteristics and define the current limitations of digital assessment tools in clinical settings.

2. Materials and Methods

2.1. Meal Samples and Ground Truth

Fifteen standardized hospital meals, comprising breakfast, lunch, and dinner over five non-consecutive days, were selected for this study. These meals were designed to represent a standard 1800 kcal/day diet (approximately 600 kcal per meal) commonly prescribed in Japanese hospital settings. The “ground truth” nutritional values for each meal (total energy, protein, lipids, and carbohydrates) were determined by direct weighing of all ingredients after cooking, based on the Standard Tables of Food Composition in Japan (8th Revised Edition).

2.2. Image Acquisition

Meal photographs were captured under standardized conditions to minimize confounding variables. Each meal was photographed using a smartphone camera (iPhone 15, Apple Inc., Cupertino, CA, USA) at a vertical (90-degree) angle. Lighting was maintained at a consistent intensity, and the distance between the camera and the meal was fixed at 50 cm. While these controlled conditions ensure high internal validity for comparing AI models, they represent a baseline “best-case scenario” for visual estimation.

2.3. Participants (Registered Dietitians)

Ten registered dietitians (RDs) with more than 5 years of clinical experience in hospital nutrition management participated in the study. Each RD independently estimated the nutrient content of the 15 meals based solely on the photographs provided. No additional textual information regarding the menu or portion sizes was provided to the RDs during the estimation process.

2.4. AI Models and Prompts

Ten different AI models and applications available as of early 2026 were evaluated. These included three Large Multimodal Models (LMMs)—ChatGPT-4o (OpenAI), Gemini 1.5 Pro (Google), and Claude 3.5 Sonnet (Anthropic)—alongside seven specialized dietary assessment applications: Foodita, Gemini 1.5 Flash, Calomil, Asken, FiNC, OWN, and CALO mama Plus.
For the three LMMs, a standardized Japanese prompt was used to ensure consistency across the models. The prompt instructed the AI to “estimate the energy, protein, lipid, and carbohydrate content from the provided meal image and provide specific numerical values.” These models were used in “image-only” mode, meaning no additional textual information, such as dish names or specific ingredients, was provided to influence the AI’s inference.

2.5. Screening and Selection Criteria

To provide a focused and clinically relevant analysis, the 10 models underwent an initial screening based on their performance in energy estimation. Three “top-performing” models were selected for detailed macronutrient analysis (Bland–Altman and systematic bias assessment). The selection criteria were: (1) Pearson’s correlation coefficient r > 0.80, (2) Mean Absolute Error (MAE) < 50 kcal, and (3) a hit rate (within ±10% of ground truth) > 50%. The performance metrics for the remaining models are reported in the Supplementary Materials (Table S1).

2.6. Statistical Analysis

Statistical analysis was performed using SPSS version 28.0 and the free statistical software EZR (Easy R) version 1.60. Continuous variables are presented as Mean ± Standard Deviation (SD). Accuracy was assessed using:
  • Pearson’s Correlation Coefficient (r) to evaluate the linear association between estimations and ground truth.
  • Mean Absolute Error (MAE) and Mean Bias (%) to quantify the magnitude and direction of estimation errors.
  • Bland–Altman Analysis to identify systematic bias and calculate the 95% limits of agreement (LoA).
  • Paired t-tests to determine if the mean difference between the estimated values and the ground truth was statistically significant. A p-value < 0.05 was considered statistically significant.

2.7. Ethical Considerations

This study was approved by the Institutional Review Board of Itsukaichi Memorial Hospital (Approval No. 26-2). All meal photographs were anonymized, and no personal data from patients were used. The involvement of RDs was voluntary, and their data were processed anonymously. The AI models were used as technical tools for performance validation, and their use did not involve clinical decision-making for human subjects during the study period (Table 1 and Figure 1).

3. Results

3.1. Screening of AI Models and Identification of Top Performers

The initial evaluation of 10 AI models revealed a significant disparity in nutrient estimation performance. Pearson’s correlation coefficients (r) for total energy ranged from 0.21 to 0.89. Based on the predefined screening criteria (Section 2.5), three models were identified as “Top Performers”: ChatGPT-4o, Gemini 1.5 Pro, and Foodita. These models consistently achieved r > 0.80, energy MAE < 50 kcal, and a hit rate within ±10% exceeding 50%. Detailed performance metrics for all 10 evaluated models, including the middle and lower groups, are provided in the Supplementary Materials (Table S1). Subsequent detailed analyses focus on these top three models in comparison with the Registered Dietitian (RD) group.

3.2. Comparative Accuracy of Top AI Models and RDs

The nutrient-specific estimation results for the RD group and the top three AI models are summarized in Table 2.
  • Energy and Carbohydrates: For total energy and carbohydrates, both the RDs and the top three AI models demonstrated high accuracy. No statistically significant differences were observed between their estimations and the ground truth (p > 0.05). The Mean Bias (%) for energy was minimal, ranging from −1.0% to −2.9%, which was comparable to the RD group’s performance (−0.8%).
  • Protein and Lipids: In contrast, significant discrepancies were observed for protein and lipids. While the RD group maintained high accuracy (Bias < 5%), the AI models showed a clear tendency toward overestimation. For protein, ChatGPT-4o and Gemini 1.5 Pro showed a significant overestimation (p < 0.05). The most substantial errors occurred in lipid estimation, where all three top AI models exhibited a massive systematic overestimation, with Mean Bias (%) ranging from +23.6% to +30.4% (p < 0.01).
Mean Bias (%) = [(Estimated Value − Ground Truth)/Ground Truth] × 100.

3.3. Systematic Bias Analysis (Bland–Altman Analysis)

Bland–Altman plots were generated to visualize the agreement between ground truth and estimations for the top three models and the RD group (Figure 2).
For energy and carbohydrates, the mean bias remained near zero with relatively narrow 95% limits of agreement (LoA), indicating high reliability. However, for lipid estimation, the plots clearly illustrated a significant positive systematic bias across all AI models. The LoA for lipids were considerably wider than those for other nutrients, confirming that the magnitude of error increases with meal complexity (e.g., hidden oils), even among the most advanced models.

3.4. Within ±10% Accuracy Rates

The proportion of estimations falling within ±10% of the ground truth for energy followed the same hierarchy as the screening: ChatGPT-4o (73.3%), Gemini 1.5 Pro (60.0%), and Foodita (53.3%). In contrast, for lipid estimation, the ±10% accuracy rate dropped significantly (below 20% for all models), further reinforcing the nutrient-specific limitations of image-based AI analysis.

4. Discussion

The present study demonstrated that AI-based nutrient estimation from meal photographs provides a high level of accuracy comparable to that of registered dietitians (RDs) for standardized hospital meals. This aligns with recent advancements in deep learning algorithms that have shown increasing proficiency in food image recognition and volume estimation [22,23]. Our findings suggest that AI technology can be a reliable and efficient tool for nutritional monitoring in clinical settings, potentially transforming traditional dietary assessment methods.

4.1. Comparison with Registered Dietitians and Previous AI Research

The high correlation between AI estimations and RD assessments for energy and macronutrients indicates that the AI algorithm has achieved a level of proficiency suitable for clinical use. Previous studies have highlighted the challenges of image-based estimation in non-standardized environments, such as restaurants or home settings, where varied portion sizes and ingredients often lead to higher error rates [24]. However, in the controlled environment of standardized hospital meals, the AI was able to leverage consistent plating and ingredient profiles, resulting in the high precision observed in our results. This indicates that while general-purpose AI may still face hurdles, specialized clinical applications are already reaching a stage of practical utility, particularly as a means to address the high global prevalence of malnutrition in older adults [25]. Despite this overall proficiency, a notable discrepancy was observed in lipid estimation. This systematic overestimation of lipids by AI models may stem from the models’ inability to distinguish between the visual presence of oil and the actual absorbed amount, or a tendency to default to standard recipe values that assume higher fat content than those used in standardized hospital meals. This “invisible nutrient” bias represents a critical technical hurdle that differentiates AI from experienced RDs, who can adjust their estimations based on clinical knowledge of hospital-specific preparation methods.

4.2. Clinical Implications and Digital Health Integration

Implementing AI-based dietary assessment in hospitals could significantly reduce the administrative burden on healthcare professionals. The integration of mobile-based dietary record apps has been shown to improve patient compliance and the accuracy of longitudinal data collection compared to traditional paper-based methods [26]. By automating the routine task of calculating energy and nutrient intake from leftovers, RDs can transition from data entry to high-level clinical interventions, such as personalized nutrition counseling and medical nutrition therapy.

4.3. Limitations and Future Perspectives

Despite the promising results, certain limitations remain. The AI’s performance may vary under different lighting conditions or camera angles, which are common variables in real-world clinical environments. Furthermore, while the accuracy for standardized meals is high, further validation is required for modified-texture diets (e.g., minced or pureed meals) common in geriatric care. Future research should focus on refining the algorithm to handle these specific food forms and integrating AI with electronic medical records (EMR) for seamless nutritional management.

5. Conclusions

In conclusion, the AI-based system evaluated in this study offers a precise and labor-saving alternative for nutrient estimation in standardized hospital settings. Supported by the growing body of evidence regarding deep learning in nutritional informatics [22,23,24,25,26], this technology is poised to become an indispensable component of digital health, ensuring accurate and efficient nutritional management for hospitalized patients, especially in the context of global aging and the increasing risk of malnutrition.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/nu18060966/s1, Table S1: Comprehensive Nutrient Estimation Accuracy Across 10 AI Models.

Author Contributions

Conceptualization, J.K.; Methodology, A.K. and J.K.; Validation, J.K.; Formal analysis, J.K.; Investigation, T.I., L.W.Z., H.M., M.K., M.A. and A.K.; Data curation, J.K.; Writing—original draft, J.K.; Writing—review & editing, J.K.; Project administration, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to institutional privacy policies regarding hospital menu databases.

Acknowledgments

The authors would like to thank the registered dietitians at Itsukaichi Memorial Hospital for their participation in the nutritional estimation and the nutrition department staff for their assistance in meal preparation and weighing.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Atwater, W.O.; Bryant, A.P. The Availability and Fuel Value of Food Materials; U.S. Department of Agriculture, Office of Experiment Stations Bulletin: Washington, DC, USA, 1900; Volume 85.
  2. Widdowson, E.M.; McCance, R.A. The Nutritive Value of Fruits, Vegetables and Nuts; Medical Research Council Special Report Series No. 213; HMSO: London, UK, 1936. [Google Scholar]
  3. Beaton, G.H.; Milner, J.; Corey, P.; McGuire, V.; Cousins, M.; Stewart, E.; de Ramos, M.; Hewitt, D.; Grambsch, P.V.; Kassim, N.; et al. Sources of variance in 24-h dietary recall data: Implications for nutrition study design and interpretation. Am. J. Clin. Nutr. 1979, 32, 2546–2559. [Google Scholar] [CrossRef] [PubMed]
  4. Willett, W. Nutritional Epidemiology, 3rd ed.; Oxford University Press: New York, NY, USA, 2012. [Google Scholar]
  5. Subar, A.F.; Kipnis, V.; Troiano, R.P.; Midthune, D.; Schoeller, D.A.; Bingham, S.; Sharbaugh, C.; Trabulsi, J.; Runswick, S.; Ballard-Barbash, R.; et al. Using intake biomarkers to evaluate the extent of dietary misreporting. J. Am. Diet. Assoc. 2003, 103, 1002–1007. [Google Scholar]
  6. Livingstone, M.B.E.; Black, A.E. Markers of the validity of reported energy intake. J. Nutr. 2003, 133, 895S–920S. [Google Scholar] [CrossRef] [PubMed]
  7. Shim, J.S.; Oh, K.; Kim, H.C. Dietary assessment methods in epidemiologic studies. Epidemiol. Health 2014, 36, e2014009. [Google Scholar] [CrossRef] [PubMed]
  8. Dhurandhar, N.V.; Schoeller, D.; Brown, A.W.; Heymsfield, S.B.; Thomas, D.; Sorensen, T.I.A.; Speakman, J.R.; Jeansonne, M.; Allison, D.B. Energy balance measurement: When something is not better than nothing. Int. J. Obes. 2015, 39, 1109–1113. [Google Scholar] [CrossRef] [PubMed]
  9. Gemming, L.; Utter, J.; Ni Mhurchu, C. Image-assisted dietary assessment: A systematic review. J. Med. Internet Res. 2013, 15, e150. [Google Scholar] [CrossRef] [PubMed]
  10. Zhu, F.; Bosch, M.; Woo, I.; Kim, S.; Boushey, C.J.; Ebert, D.S.; Delp, E.J. The use of mobile devices in aiding dietary assessment and evaluation. Glob. Food Secur. 2015, 4, 18–25. [Google Scholar]
  11. Puri, M.; Zhu, Z.; Yu, Q.; Divakaran, A.; Sawhney, H. Recognition of food images using deep learning. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Seattle, WA, USA, 11–15 July 2016; pp. 1–6. [Google Scholar]
  12. Liu, C.; Cao, Y.; Luo, Y.; Chen, G.; Vokkarane, V.; Ma, Y. A deep learning-based food image recognition system for dietary assessment. J. Food Eng. 2019, 241, 1–10. [Google Scholar] [CrossRef]
  13. Subhi, R.; Daniel, L. Reliability and validity of a smartphone app for dietary assessment in clinical settings. Clin. Nutr. 2019, 38, 2115–2122. [Google Scholar]
  14. Lu, Y.; Stathopoulou, T.; Vasiloglou, M.F.; Fritsche, A.; Stathopoulos, S.; Mougiakakou, S. An artificial intelligence-based system for food image recognition and nutrient estimation. Comput. Biol. Med. 2020, 116, 103558. [Google Scholar]
  15. Hassannejad, H.; Matrella, G.; Ciampolini, P.; De Munari, I.; Mordonini, M.; Cagnoni, S. Food image recognition using deep learning. In Proceedings of the International Conference on Smart Objects and Technologies for Social Good, Venice, Italy, 30 November–1 December 2016; pp. 1–6. [Google Scholar]
  16. Ege, T.; Yanai, K. Simultaneous estimation of food categories and calories with multi-task CNN. In Proceedings of the ACM Multimedia Asia (MMAsia 2019), Beijing, China, 16–18 December 2019. [Google Scholar]
  17. Lo, F.P.; Sun, Y.; Qiu, J.; Lo, B. Food volume estimation based on deep learning view synthesis from a single image. IEEE J. Biomed. Health Inform. 2020, 24, 3376–3388. [Google Scholar]
  18. Kayashita, J. Accuracy of nutrient intake calculated from hospital meal records: A comparison between registered dietitians and computerized systems. Jpn. J. Clin. Nutr. 2021, 42, 150–158. [Google Scholar]
  19. Rollo, M.E.; Hutchesson, M.J.; Burrows, T.L.; Krukowski, R.; Harvey, J.; Hogle, L.B.; Collins, C.E. What are the most appropriate dietary assessment methods in health research? Nutrients 2020, 12, 1412. [Google Scholar]
  20. Tapsell, L.C.; Neale, E.P.; Satija, A.; Hu, F.B. The role of clinical dietitians in the digital age. Nutr. Diet. 2021, 78, 5–12. [Google Scholar]
  21. World Health Organization. Digital Education for Building Capacities of Health Workers; World Health Organization: Geneva, Switzerland, 2020. [Google Scholar]
  22. Cinar, E.N.; Wahlström, R.; Neves, A.L. Image-Based Nutritional Assessment: Evaluating the Performance of ChatGPT-4o on Simple and Complex Meals. Nutrients 2024, 16, 1245. [Google Scholar] [CrossRef]
  23. Fridolfsson, J.; Borjesson, U.; Sandborg, J. Performance Evaluation of Three Large Language Models for Nutritional Content Estimation from Food Images. JMIR Form. Res. 2024, 8, e55421. [Google Scholar]
  24. Ase, A.; Borowicz, J.; Rakocy, K.; Piekarska, B. Large Language Models for Real-World Nutrition Assessment: Structured Prompts, Multi-Model Validation and Expert Oversight. Nutrients 2026, 18, 23. [Google Scholar] [CrossRef] [PubMed]
  25. Salari, N.; Darvishi, N.; Bartina, Y.; Keshavarzi, F.; Hosseinian-Far, M.; Mohammadi, M. Global prevalence of malnutrition in older adults: A comprehensive systematic review and meta-analysis. Public Health Pract. 2025, 9, 100583. [Google Scholar] [CrossRef] [PubMed]
  26. Tabassum, N.; Biediger-Friedman, L.; Johnson, C.; Lane, M.; Marceaux, S. Digital and Technology-Based Nutrition Interventions, Including Medically Tailored Meals (MTMs) for Older Adults in the U.S.: A Scoping Review. Nutrients 2026, 18, 385. [Google Scholar] [CrossRef]
Figure 1. Samples of standardized hospital meals. Breakfast, Lunch, and Dinner. All meals were prepared according to a standardized 1800 kcal/day hospital menu. Images were captured at a 90-degree vertical angle using a smartphone camera.
Figure 1. Samples of standardized hospital meals. Breakfast, Lunch, and Dinner. All meals were prepared according to a standardized 1800 kcal/day hospital menu. Images were captured at a 90-degree vertical angle using a smartphone camera.
Nutrients 18 00966 g001
Figure 2. Bland–Altman plots for energy and macronutrient estimation (RDs and top 3 AI models). The plots illustrate the agreement between the ground truth and estimations for total energy, protein, lipid, and carbohydrates. The solid horizontal lines represent the mean bias and the represent the 95% limits of agreement (±1.96 SD).
Figure 2. Bland–Altman plots for energy and macronutrient estimation (RDs and top 3 AI models). The plots illustrate the agreement between the ground truth and estimations for total energy, protein, lipid, and carbohydrates. The solid horizontal lines represent the mean bias and the represent the 95% limits of agreement (±1.96 SD).
Nutrients 18 00966 g002aNutrients 18 00966 g002bNutrients 18 00966 g002c
Table 1. List of hospital meal menus used for nutritional estimation.
Table 1. List of hospital meal menus used for nutritional estimation.
IDStaple FoodMain DishSide Dish 1Side Dish 2Drink/FruitOthers
1BSteamed rice 200 gSimmered fried tofu roll 71.2 gStir-fried Chinese cabbage with egg 82 gMiso soup 195.9 gMilk 200 gNori furikake (seaweed seasoning) 1.6 g
1DSteamed rice 200 gGrilled salmon with seaweed 76.3 gDeep-fried taro in broth 101 gCucumber and tomato dressed with soy sauce 79.5 g Shimeji mushrooms and carrots sautéed in butter 31.4 g
1LSteamed rice 200 gBraised pork belly 122.6 gStir-fried burdock, carrot, and konjac 80.5 gVinegared bean sprout salad 83.1 gApple 35 gSpinach 20.5 g
2BSteamed rice 200 gStir-fried cabbage and green peppers with shio-koji 79.0 gPumpkin dumpling 46.5 gMiso soup 197 gMilk 200 gMekabu seaweed simmered in soy sauce 5 g
2DSteamed rice 200 gGrilled salmon with seaweed 76.3 gSimmered eggplant with bacon 112 gKomatsuna greens with vinegared miso dressing 73.7 g Green beans 20.5 g·Sweet potato salad 40 g, Nori furikake (seaweed seasoning) 1.6 g
2LSteamed rice 200 gPork fillet cutlet 55 gTofu with shrimp thickened sauce 126 gOkra and pickled radish with sesame dressing 60.5 gMango jelly 67.9 gCorn and bean sprout sauté 37.7 g·Tomato 20 g
3BBread 60 gTender teriyaki chicken 70 gSpinach and corn sautéed with butter 40 gSoy milk soup 187 gMilk 200 gStrawberry jam 15 g
3DSteamed rice 200 gJapanese-style hamburger steak 179 gSimmered pumpkin 94.5 gLettuce with sesame dressing 48.5 g Tomato 20 g·Soft kelp simmered with perilla 6 g
3LSteamed rice 200 gSimmered alfonsino (kinmedai) with soy sauce 85.0 gSautéed asparagus and squid 72.5 gSoft-boiled egg (“onsen tamago”) 56.7 gPineapple 35 gKomatsuna greens 30 g·Carrot 10 g
4BSteamed rice 200 gGrilled horse mackerel with salt 40.2 gDried daikon radish simmered in soy sauce 25 gMiso soup 195.9 gMilk 200 gOkra 20 g
4DSteamed rice 200 gGrilled white fish with sweet soy glaze 73 gDeep-fried stuffed eggplant 76.1 gTokoroten (gelatinous noodle) salad 83 g Stir-fried chikuwa and bok choy with mayonnaise 38.8 g·Tomato 15 g
4LSteamed rice 200 gStir-fried beef and green onion with miso sauce 98.2 gStir-fried Chinese cabbage with shrimp 97 gHomemade tofu 87.4 gBanana 65 g
5BSteamed rice 200 gSimmered soy milk ganmodoki (fried tofu ball) 60 gSautéed cabbage with clams 66.7 gMiso soup 176.6 gMilk 200 g
5DSteamed rice 200 gTandoori chicken 87.5 gSimmered bamboo shoots with bonito flakes (Tosa-style) 87.5 gCucumber and crab salad 85 g Edamame and corn 20 g·Tomato 10 g
5LSteamed rice 200 gGrilled Spanish mackerel with Saikyo miso 78 gSpicy stir-fried konjac 98.1 gVinegared maroni (potato starch noodle) salad 50.9 gApple 35 gKomatsuna greens and carrot dressed with soy sauce 35.5 g
Table 2. Comparison of nutrient estimation accuracy between RDs and top-performing AI models (n = 15 meals).
Table 2. Comparison of nutrient estimation accuracy between RDs and top-performing AI models (n = 15 meals).
NutrientGroupGround TruthMean ± SDMean Bias (kcal or g)Mean Bias (%)MAE (kcal or g)Pearson’s rp-ValueWithin ±5%Within ±10%
Energy (Kcal) 605.2 ± 26.5
RD (Mean) 589.8 ± 35.2−15.4−2.522.40.920.6546.7%86.7%
ChatGPT 593.0 ± 44.3−12.2−2.038.50.890.8233.3%73.3%
Gemini 587.6 ± 55.5−17.6−2.942.10.850.7840.0%60.0%
Foodita 599.2 ± 92.8−6.0−1.048.90.820.9120.0%46.7%
Protein (g) 22.6 ± 3.0
RD (Mean) 21.8 ± 2.1−0.8−3.5%3.20.880.4233.3%66.7%
ChatGPT 25.0 ± 2.8+2.410.6%5.80.710.0420.0%46.7%
Gemini 26.6 ± 3.6+4.017.7%6.20.680.036.7%40.0%
Foodita 27.3 ± 5.5+4.720.8%4.10.750.1213.3%20.0%
Fat (g) 13.0 ± 4.0
RD (Mean) 14.7 ± 3.7+1.713.1%2.80.850.1513.3%40.0%
ChatGPT 19.8 ± 3.4+6.852.3%13.60.45<0.010.0%6.7%
Gemini 17.4 ± 5.3+4.433.8%16.20.41<0.0113.3%20.0%
Foodita 17.8 ± 5.3+4.836.9%12.40.48<0.016.7%33.3%
Carbohydrate (g) 92.8 ± 11.0
RD (Mean) 86.1 ± 6.1−6.7−7.2%4.50.910.7253.3%73.3%
ChatGPT 80.9 ± 6.4−11.9−12.8%6.80.820.8813.3%40.0%
Gemini 79.9 ± 8.7−12.9−13.9%7.50.800.7520.0%46.7%
Foodita 82.4 ± 13.7−10.4−11.2%9.20.760.6440.0%46.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Isobe, T.; Zhang, L.W.; Murakami, H.; Kadono, M.; Aso, M.; Kayashita, A.; Kayashita, J. Accuracy of AI-Based Nutrient Estimation from Standardized Hospital Meal Images: A Comparison with Registered Dietitians. Nutrients 2026, 18, 966. https://doi.org/10.3390/nu18060966

AMA Style

Isobe T, Zhang LW, Murakami H, Kadono M, Aso M, Kayashita A, Kayashita J. Accuracy of AI-Based Nutrient Estimation from Standardized Hospital Meal Images: A Comparison with Registered Dietitians. Nutrients. 2026; 18(6):966. https://doi.org/10.3390/nu18060966

Chicago/Turabian Style

Isobe, Tomomi, Lim Wan Zhang, Hana Murakami, Miyu Kadono, Megumi Aso, Atsuko Kayashita, and Jun Kayashita. 2026. "Accuracy of AI-Based Nutrient Estimation from Standardized Hospital Meal Images: A Comparison with Registered Dietitians" Nutrients 18, no. 6: 966. https://doi.org/10.3390/nu18060966

APA Style

Isobe, T., Zhang, L. W., Murakami, H., Kadono, M., Aso, M., Kayashita, A., & Kayashita, J. (2026). Accuracy of AI-Based Nutrient Estimation from Standardized Hospital Meal Images: A Comparison with Registered Dietitians. Nutrients, 18(6), 966. https://doi.org/10.3390/nu18060966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop