Next Article in Journal
Bi-Level Collaborative Optimization for Medical Consumable Order Splitting and Reorganization Considering Multi-Dimensional and Multi-Scale Characteristics
Previous Article in Journal
Seasonal Biochemical Variations in Mediterranean Halophytes and Salt-Tolerant Plants: Targeting Sustainable Innovations in Ruminant Health
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Deep Learning in Food Image Recognition: A Comprehensive Review

1
School of Intelligence Science and Technology, Xinjiang University, Urumqi 830017, China
2
Xinjiang Key Laboratory of Signal Detection and Processing, Urumqi 830017, China
3
Department of Electronic Engineering, Beijing National Research Center for Information Science and Technology, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(14), 7626; https://doi.org/10.3390/app15147626
Submission received: 19 May 2025 / Revised: 3 July 2025 / Accepted: 4 July 2025 / Published: 8 July 2025

Abstract

Food not only fulfills basic human survival needs but also significantly impacts health and culture. Research on food-related topics holds substantial theoretical and practical significance, with food image recognition being a core task in fine-grained image recognition. This field has broad applications and promising prospects in smart dining, intelligent healthcare, and smart retail. With the rapid advancement of artificial intelligence, deep learning has emerged as a key technology that enhances recognition efficiency and accuracy, enabling more practical applications. This paper comprehensively reviews the techniques and challenges of deep learning in food image recognition. First, we outline the historical development of food image recognition technologies, categorizing the primary methods into manual feature extraction-based and deep learning-based approaches. Next, we systematically organize existing food image datasets and summarize the characteristics of several representative datasets. Additionally, we analyze typical deep learning models and their performance on different datasets. Finally, we discuss the practical applications of food image recognition in calorie estimation and food safety, identify current research challenges, and propose future research directions.

1. Introduction

Food plays a vital role in daily life, serving not only as a necessity for survival but also as a cultural and health cornerstone. With the advancement of society, heightened health consciousness has driven an increasing demand for personalized and balanced diets. Consequently, food-related research has emerged as a critical interdisciplinary field, bridging computer vision, nutrition science, and industrial automation. Among these efforts, food image recognition—a core task in fine-grained image analysis—has garnered significant attention for its potential to revolutionize dietary assessment, healthcare management, and intelligent retail systems. Applications range from automated checkout in unmanned restaurants to real-time calorie tracking for patients with chronic diseases and fitness enthusiasts.
In computer vision, food images are categorized as fine-grained data due to their inherent complexity [1]. These images typically exhibit small inter-class variations (e.g., subtle differences between regional cuisines) and large intra-class diversity (e.g., varying presentations of the same dish), compounded by the absence of consistent spatial structures. Current research primarily focuses on dish recognition using subsidiary tasks such as ingredient identification, food scene analysis, and cooking method classification are also actively explored.
The evolution of food image recognition methodologies can be broadly divided into two eras: manual feature extraction and deep learning-based approaches. Early efforts, such as Parrish et al.’s fruit recognition system (1977) [2] and Bolle et al.’s Veggie Vision (1996) [3], relied on handcrafted features like color histograms and texture descriptors. Subsequent advancements introduced hybrid frameworks, including Kitamura et al.’s multimedia food logging system (2008) [4] and Yang et al.’s pixel-level segmentation techniques (2010) [5]. However, these methods struggled with scalability and robustness due to their dependence on explicit feature engineering.
A paradigm shift occurred post-2014 with the advent of deep learning. Kagaya et al. pioneered the application of convolutional neural networks (CNNs) to food recognition (2014) [6], demonstrating superior accuracy over traditional methods. This breakthrough catalyzed rapid innovation. FoodCam (2014) [7] integrated mobile-based calorie estimation, while Im2Calories (2015) [8] leveraged CNNs for single-image nutrient prediction. Subsequent years witnessed architectural refinements, such as attention mechanisms (2020, 2024) [9,10], multi-task learning frameworks (2021) [11], and lightweight models (2022, 2024) [12,13], addressing challenges like fine-grained discrimination and computational efficiency. Notably, recent advancements like GSNet (2024) [14] and self-supervised pre-training (2024) [10] highlight the field’s ongoing evolution toward robustness and adaptability (Table S1).
This paper systematically reviews the technological trajectory of food image recognition, emphasizing deep learning innovations, dataset development, and real-world applications. Key contributions include a taxonomy of methodologies, an analysis of dataset limitations, and a discussion of emerging trends in edge computing and multi-modal integration.
The main contributions of this survey are summarized in the following four aspects:
  • A comprehensive review of food image recognition methods, emphasizing deep learning models.
  • A systematic organization of existing datasets to aid researchers in selecting appropriate resources.
  • A discussion of practical applications in nutrition assessment, food services, and safety.
  • An analysis of current challenges and future directions for research.
This survey selects a number of papers (2014-2024) from major databases and adopts a keyword strategy targeting food recognition and deep learning. Priority was given to studies with reproducible results, excluding work lacking empirical validation. The reviewed methods are classified according to their architectures and application scenarios, and performance indicators are extracted for cross-comparison (Table 1 and Table S2).
The remainder of this paper is structured as follows: Section 2 introduces food image datasets, Section 3 reviews recognition methods, Section 4 explores applications, and Section 5 concludes with open issues and future trends (Figure 1).

2. Datasets

Nowadays, the world’s mainstream datasets are mainly based on various dishes, covering different dishes around the world. Among these, the types of dishes are mainly Western dishes, and the number of Chinese dishes is relatively small. There are also some fruit and vegetable image datasets and other types of food. These datasets can be divided into small-scale, medium-scale, and large-scale dish datasets. With the data amount of 10,000 as the limit, the dataset with the total data amount less than 10,000 is called the small-scale dish dataset, the data amount between 10,000 and 100,000 is called the medium-scale dish dataset, and the data amount more than 100,000 is called the large-scale dish dataset. Many large-scale datasets have significantly promoted the development of food image-related technology.
In 2009, Chen et al. published the first fast food dataset, PFID, for dish recognition, which contains 4545 still images, 606 stereo pairs, 303 360-degree motion structure videos, and 27 related videos [16]. Subsequently, Joutou et al. published Food50 [17] and Food80 [18]. In 2014, Bossard et al. [19] published the first large-scale Western dish image dataset, ETHZ Food-101, which contains 101,000 dish images composed of images from 101 Western dishes. Due to its large data scale, this dataset is currently widely used to test the performance of dish image recognition model algorithms and as one of the benchmarks of recognition methods based on deep learning. Since then, food image datasets have been continuously expanded, increasing in size and scope (Table S3, Figure 2).
For the existing datasets, there are still some limitations restricting the research progress:
  • Low cultural coverage: The proportion of non-Western dishes in the mainstream dataset is relatively low, and the annotation standard does not consider the difference of national and regional names. For example, in the ETHZ Food-101 dataset, Asian dishes make up only 30% of the total, and African dishes make up even less, around 1%. The same is true for the ISIA Food-500 dataset, where Asian dishes make up about 40% of the total and African dishes make up only about 4%. In the single non-Western food dataset, there is still a lack of refinement of regional variants.
  • Coarse annotation granularity: Most datasets only provide food category labels and lack fine-grained annotations such as ingredients and cooking methods, which limits the application of multi-task learning.
  • Lack of authenticity: There are significant domain differences between the images taken in the laboratory environment and the blurred and occluded pictures actually taken by users, resulting in the performance degradation of the model in the real scene and the lack of dynamically updated datasets.
To mitigate the impact of dataset limitations (e.g., cultural bias and annotation granularity) while improving model performance, current research employs complementary strategies at both data and algorithm levels:
  • Data Augmentation Techniques:
    • Geometric transformation: By rotating, flipping, scaling, and performing other operations on the image, it can simulate the food form under different shooting angles and perspectives and effectively alleviate the problem of model recognition bias caused by different shooting angles.
    • Color perturbation: Adjusting the color parameters of the image such as brightness and contrast can reduce the influence of light changes on the visual characteristics of the food image and enhance the robustness of the model under different lighting conditions.
    • Generative data augmentation: Ito et al. [20] utilized the Conditional Generative Adversarial Network (CGAN) for food image generation. In the experiment, the CGAN was trained using the “ramen” image dataset and the recipe image dataset. Through generative data augmentation, the diversity of the data was enriched. A plate discriminator was added to the “ramen” GAN to make the shape of the dishes in the generated image more rounded. The “recipe” GAN was used to generate dish images from cooking ingredients and use them for image search in the recipe database.
  • Domain adaptation method
    • Transfer learning: This approach pretrains models on large-scale general-purpose datasets such as ImageNet and then fine-tunes them to the task of food image recognition. This method reduces the dependence of the model on large-scale labeled food data and makes full use of the common visual features learned by the pre-trained model.
    • Self-supervised learning: In the case of masked image modeling, large amounts of unlabeled data can be effectively utilized to improve the generalization ability of the model by occlusion part of the image region and letting the model predict the occluded content.
  • Multi-source data fusion
    • Combining auxiliary information such as recipe texts and geolocation with visual data of food images can make up for the cultural bias problem existing in pure visual data. Herranz et al. [21] simplified the classification problem by using context information such as geolocation and proposed a food recognition framework including semantic feature learning and location-adaptive classification. Experiments showed that using geographical location could improve the recognition performance by approximately 30%.
In summary, the existing food image datasets have made significant progress in scale and diversity, but there are still problems such as insufficient cultural coverage, rough annotation granularity, and a lack of authenticity. Mainstream datasets such as Food-101 and ISIA Food-500 provide benchmarks for research. However, the proportion of non-Western foods is low, and there is a lack of fine-grained annotations such as ingredients and cooking methods. To cope with these limitations, researchers have employed techniques such as data augmentation, domain adaptation, and multi-source data fusion. In the future, large-scale datasets with more cultural representation and fine-grained annotation should be constructed to improve the generalization ability of the model.

3. Food Image Recognition Methods

With the rapid advancement of deep learning technologies, researchers have integrated deep neural networks into food image analysis to address the limitations of traditional manual feature extraction methods in terms of representational capacity and adaptability. Currently, deep learning-based food image recognition approaches can be categorized into two main technical paradigms: convolutional neural network (CNN)-driven recognition frameworks and dedicated deep network architectures specifically designed for food characteristics.

3.1. Convolutional Neural Network (CNN)-Based Methods

The paradigm shift in food image recognition began with Kagaya et al. [6], who pioneered the application of convolutional neural networks (CNNs) to this domain. While early implementations directly employed off-the-shelf CNN architectures for feature extraction, these approaches overlooked domain-specific characteristics of food imagery, resulting in suboptimal recognition accuracy. Subsequent research has focused on architectural adaptations of baseline CNNs to enhance model performance through task-specific optimizations.
Recent advancements demonstrate three distinct optimization strategies:
  • Attention-Augmented Architectures: Abiyev et al. [22] introduced the FRCNNSAM framework, integrating self-attention mechanisms with deep CNNs through weight-sharing and data compression. Ensemble predictions from multiple FRCNNSAM variants via averaging achieved state-of-the-art accuracies of 96.40% on Food-101 and 95.11% on MA Food-121, outperforming transfer learning baselines by 8.12% (Figure 3).
  • Transfer Learning Ensembles: Bu et al. [23] developed an ensemble model leveraging ImageNet-pretrained networks (VGG19, ResNet50, MobileNetV2, AlexNet) for generic feature extraction. Fine-tuning and strategic combination of base learners through feature-space fusion yielded 96.88% accuracy on FOOD-11, demonstrating superior generalization over individual models (Figure 4).
  • Spatiotemporal Fusion Networks: Phiphitphatphaisit et al. [24] proposed ASTFF-Net, a novel architecture combining the following (Figure 5):
    • Spatial encoder: ResNet50 with parameter reduction for robust feature extraction.
    • Temporal processor: LSTM-enhanced Conv1D layers for sequential pattern learning.
    • Adaptive fusion module: Softmax-optimized spatiotemporal feature integration. This framework addresses key challenges in food recognition-including occlusions, image blur, and inter-class similarity while maintaining computational efficiency. Benchmark evaluations across Food11, UEC FOOD-100/256, and ETH Food-101 confirm its competitive performance.

3.2. Dedicated Deep Networks

Although the local convolution operation of convolutional neural network has a strong ability to capture the local features of food (e.g., texture and shape, etc.), there are still problems such as the risk of overfitting, insufficient interpretability, and high computational cost. In view of the fine-grained image characteristics of food images, researchers have proposed some special networks for food image recognition in recent years. Inspired by the great achievements of Transformers in the field of natural language processing, in recent years, Vision Transformer (ViT) structures based on self-attention mechanisms have been widely used in the field of computer vision and achieved certain results. Although the self-attention mechanism used by the ViT structure has some flexibility in characterizing visual content, it fails to achieve the best results when encoding the fine-grained features of images due to the limitation of image factors such as size. On this basis, Jiang et al. [10] proposed a food image recognition algorithm based on ViT. Aiming at the demand for fine-grained feature extraction of food images, a local-attention mechanism based on the attention mechanism was proposed to enable the model to extract richer fine-grained features. At the same time, Jiang et al. proposed an image self-supervised pre-training algorithm to solve the problem of a small number of samples in the food image dataset, which used mask images to train network parameters to effectively alleviate the problem of insufficient sample training. Experimental results show that the Top-1 accuracy and Top-5 accuracy values of the proposed model on the ISIA Food-500 dataset are 65.58% and 90.03%, which are the best accuracies at present (Figure 6).
Food images are unstructured images with complex and unfixed visual patterns. Modeling the long-range semantic relationships between ingredients, as well as semantic interactions between categories and ingredients, is beneficial for ingredient identification and food analysis. Based on the above factors, Liu et al. [25] proposed a multi-task learning framework for food category and ingredient recognition. The framework consists of a food-oriented Transformer named Convolution-Enhanced Bi-Branch Adaptive Transformer (CBiAFormer) and a multi-task category-ingredient recognition network called Structural Learning and Cross-Task Interaction (SLCI). In CBiAFormer, a query-aware data adaptive attention mechanism consisting of a local fine-grained branch and a global coarse-grained branch is established, and an adaptive candidate key/value set is assigned to each query to mine the local and global semantic-aware regions of different input images. At the same time, the framework introduces a convolutional patch embedding module for extracting fine-grained features that the transformer ignores. The SLCI is composed of Cross-Layer Attention (CLA) and two cross-task interaction modules, namely Category-guided Ingredient Relationship Learning (CIRL) and Ingredient-guide Category Feature Enhancement (ICFE). The SLCI is used to model the semantic relationship between the components and mine the semantic interaction between the categories and components so as to make full use of the information contained in images. This method has obtained good experimental results on the mainstream datasets (ETH food-101, Vireo food-172, and ISIA food-200) (Figure 7).
In summary, the current mainstream deep learning methods in the field of food image recognition can be summarized into four paradigms: CNN-based, Vision Transformer architecture, multi-modal fusion, and lightweight neural network models. In order to more clearly compare the characteristics of these methods, Table 2 systematically summarizes them from the dimensions of computational efficiency, recognition accuracy, and data requirements. Table S4 lists the comparison and summary of some lightweight neural network models. Table S5 summarizes the main models mentioned above, including the dataset evaluated, performance metrics, and key features.
The performance of food recognition models proposed in current research often depends on the scale and cultural coverage of the training set, and its performance may degrade when the model is deployed to the real scene or changed to the test scene. For example, in cross-cultural tests, models trained on Western food data show significant performance degradation on Asian food datasets. Although the CBiAFormer model proposed by Liu et al. [25] achieved an excellent performance of 92.40% Top-1 accuracy on ETHZ Food-101 (Western dish dataset), the Top-1 accuracy on ISIA Food-200 (mixed dish dataset) was only 72.92%. Compared with its performance on Food-101, it is decreased by about 21%. This significant performance difference reveals the generalization limitations of current models in cross-cultural scenarios.
Due to the particularity of food images, there is a semantic gap between the general features of natural images (e.g., edges and textures, etc.) and the high-order features specific to food (e.g., food ingredient combination patterns, etc.), which makes the recognition effect of traditional transfer learning limited, and there is a problem of “feature mismatch”. For example, in the experiments of Bu et al. [23], the average accuracy of a single model is only 88.35% when directly using ImageNet pre-trained models such as VGG19, ResNet50, and MobileNetV2 for transfer learning on the FOOD-11 dataset. It is much lower than the accuracy of 96.88% after introducing ensemble learning on the basis of transfer learning. It indicates that more refined feature extraction and domain adaptation methods are needed for food image recognition.
Therefore, in order to improve the practical value of the model, it is necessary to construct a more representative multicultural and large-scale food image dataset and develop algorithms that can effectively utilize limited samples. Recently, some studies [10] have begun to explore methods such as self-supervised learning to deal with the problem of insufficient data, which will be an important research direction in the future.

4. Application

4.1. Nutrient Intake Assessment

Food computing is one of the main interdisciplinary studies of food science and computer science at present. It aims to use artificial intelligence, data processing, and data analysis technology and integrate information such as nutritional characteristics of food itself, changes in nutritional attributes of raw materials, and manufacturing processes. By using these data, the nutritional content of corresponding food can be predicted and estimated, providing a scientific basis for people to eat healthily. In addition, diet quality assessment is also one of the important measures to prevent and treat various chronic diseases. Through the application of food image recognition, classification, and retrieval technology, it can help people to complete the purpose of food calculation and has important practical value for the study of people’s eating habits and health management. At present, the main related applications include intelligent health management software, food calorie prediction platforms, and so on. For example, Nadeem et al. [26] developed “Smart Diet Diary,” an application for diet tracking, using a pre-trained R-CNN model that uses deep learning to identify foods and calculate their nutritional value based on calories.

4.2. Food Services

Research related to food image has been widely used in the food industry, such as automatic commodity settlement, food retrieval recommendations, and smart kitchen utensils. Food image recognition technology can play an important role in monitoring food sales, automatic settlement, and so on. For example, Xiao et al. [27] embedded a partial-and-imbalanced domain adaptation technique (tree adaptation network) in the deep learning model and proposed a solution to realize automatic procurement of food raw materials using electronic scales. In addition, food image recognition technology can play an important role in the freshness and type recognition of ingredients, which is mainly used in smart kitchenware. For example, Mohammad et al. [28] developed a smart refrigerator using convolutional neural network architecture and transfer learning techniques as well as Inception V3 pre-trained models to identify and detect items in a refrigerator. In the era of big data, food image retrieval technology is applied to all kinds of social media, especially diet or recipe-related search recommendation websites, to provide convenience for people’s daily diets. Food image recognition technology is also widely used in intelligent catering robots, food recommendations, and social entertainment scenes. For example, Herranz et al. [29] proposed a probabilistic framework combining multiple pieces of evidence (visual, geolocation, and external knowledge) for food recognition in restaurant scenes.

4.3. Food Safety

As the saying goes, “Food is the priority of the people, and food is the priority of safety.” Food safety concerns the national economy and people’s livelihood. Compared with traditional chemical detection methods, food safety detection methods based on computer vision have the characteristics of a short cycle, a relatively simple process, and easy to be put into large-scale use, and it has gradually become the focus of market research in recent years. Efficient and accurate food safety assessment research and early warnings can provide data support for the regulatory authorities to conduct more targeted sampling checks; provide safety protection for the production, sales, and purchase of food by enterprises and merchants; and reduce the probability of social negative events caused by food safety problems. In addition, food safety testing, as a downstream step of the industrial chain, provides an important means of quality assurance for the food industry. Through testing, it can ensure that the products produced by the food industry meet the safety, hygiene, and quality standards and protect the health rights and interests of consumers. For example, Wang et al. [30] proposed a two-stage learning mode of YOLO-SIMM by combining Tiny-YOLO and a twin network. In addition, a food image recognition and food safety detection method based on deep learning by threshold segmentation technology is proposed that can effectively separate foreign bodies from food and help to improve the level of food safety.

5. Open Issues and Future Direction

At present, more progress has been made in the research of algorithms and datasets related to food image recognition, and they have been applied to real life. The following sections mainly discuss the existing problems in food image-related research and make prospects for future development.
  • Dataset Limitations
Although the current mainstream food image datasets (e.g., Vireo Food-172, ISIA Food-500, etc.) have established a benchmark system, there are still prominent problems of limited categories and insufficient scale compared with general large-scale datasets, such as ImageNet. This limitation is extremely significant in cross-cultural scenarios. First, food images are strongly culturally specific. Dishes with the same name in different regions have significant differences in ingredient ratios, cooking techniques, and presentation styles, while the cultural coverage of existing datasets is seriously unbalanced. For example, the proportion of Western cuisines in the mainstream dataset far exceeds that of Southeast Asian and African cuisines, which limits the cross-cultural generalization ability of the model. In addition, the fine-grained nature of food itself exacerbates the difficulty of data construction. Dishes of the same category may present completely different visual features due to regional differences, which makes it difficult to unify the annotation standards. Although ISIA Food-500 tries to introduce cultural labels [31], it increases the labeling cost in terms of microscopic details. The above problems make it difficult to build a large-scale food image dataset. In addition, when acquiring multi-channel images, how to effectively reduce the cost of dataset construction and improve the quality of the dataset is also a key issue to be solved.
  • Research on Food Computing Integrating Multiple Attributes of Food
Different from conventional fine-grained images, food images do not have a fixed semantic pattern, lack a unified spatial layout, and have more noise. It is difficult to extract common semantic information between different food images for recognition. In addition, it is difficult to accurately identify food images with high confusability due to the similarities between and differences within food images. How to accurately extract and study similar food local discriminant feature images becomes the key to research.
At the same time, food images often contain rich additional information, such as ingredient information, ingredient state geographical location, etc. At present, the existing food image recognition systems mainly focus on the types of food ingredients, and the recognition of the specific state of food ingredients is less important. In addition, deep learning models still face challenges in extracting higher-order semantic features of food, such as food ingredient combination patterns. The performance of traditional transfer learning methods is limited by the semantic differences between natural images and food images, such as the mismatch between texture and food ingredient combination. Multimodal fusion methods such as vision-text joint modeling [32] and spatio-temporal feature fusion [24] have been proposed to effectively improve the accuracy of food image recognition. Making full use of the additional information contained in food images can help to improve the recognition performance of food images. How to make full use of image features, design a dedicated network architecture for food semantics, and break through key technical bottlenecks such as modal alignment and computational efficiency will have important research value.
  • New Applications and Ethical Privacy Challenges
At present, food image recognition technology has been widely used in many fields such as diet quality assessment, food safety monitoring, intelligent catering sales, and personalized diet recommendation and plays an important role in industrial and agricultural production. Examples include intelligent retail and unmanned settlement systems, automated food production processing, and agricultural picking and sorting robots. Majil et al. developed a cooking assistance system to identify ingredient types and provide suggested recipes to the user [33]. In addition, food image recognition technologies, such as kitchen cooking robots and intelligent refrigerator food ingredient recognition and management, have gradually become more and more widely used in families. With the popularity of portable smart devices, the market demand for lightweight food recognition technology for mobile terminals is increasing [34]. At present, there are also some open source tools. For example, Kaur et al. [35] proposed an open source dataset for fine-grained food classification, containing 251 kinds of fine-grained food and 158k images, which has been used in international competitions. Sahoo et al. [36] proposed FoodAI, an open source smart food recording system that provides pre-trained models and API interfaces for mobile applications. At the level of technical implementation, the lightweight and real-time performance of deep learning models have become the key. Although previous studies have proposed mobile solutions based on MobileNetV3 [37] and EfficientNet [38], model compression and edge deployment in cross-cultural scenarios still need to be further optimized.
Food is a special carrier of culture and health. On the one hand, through the analysis of food imagery, the transmission path of traditional food culture is traced. On the other hand, it can provide objective dietary evaluation means for chronic disease management [39], which further highlights the application value of this technology. However, the deep application of technology also brings new ethical challenges. Diet data are sensitive health information, and privacy protection problems are becoming increasingly prominent. A user’s diet may reveal health conditions, and there is a risk of data misuse. Dietary preference analysis in cross-cultural scenarios should follow local data protection regulations. In some countries, some generative AI service management policies have been introduced to regulate applications. In commercial applications, personalized recommendation algorithms may exacerbate dietary bias. With the increasing maturity of recognition technology, it is urgent to establish a development framework that takes into account both technological innovation and ethical constraints while expanding emerging markets, so as to realize the balance between technical benefits and social responsibility.
Current research faces challenges such as dataset limitations, insufficient model generalization, and ethical privacy issues. In the future, we need to focus on the following: developing large-scale datasets with multicultural, fine-grained annotations. Multi-modal fusion is used to improve the recognition performance by combining multi-source data such as vision, text, and geographic information. The model was optimized to adapt to mobile terminal and edge computing scenarios and improve the lightweight and real-time performance of the model in practical applications. In addition, it is also an important research direction to establish a data protection framework, balance technological innovation and social responsibility, and ensure that the technology will not infringe on the privacy of others in practical applications.

6. Conclusions

This survey reviews advancements in food image recognition, emphasizing deep learning techniques, datasets, and applications. While significant progress has been made, challenges such as dataset scarcity and model interpretability remain. Future research should focus on multi-modal learning, real-time deployment, and cross-cultural adaptability to unlock the full potential of food image recognition.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app15147626/s1, Table S1. Summary of main methods; Table S2. Literature selection criteria; Table S3. Summary of datasets; Table S4. Comparison of lightweight neural network models; Table S5. Comparison of deep learning models in food image recognition [40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118].

Author Contributions

D.L.: Writing—original draft, Visualization, Validation, Resources. D.W.: Validation, Software, Methodology, Investigation. E.Z.: Formal analysis, Data curation. D.W. and L.H.: Supervision, Resources, Project administration, Funding acquisition. L.D.: Resources, Methodology, Investigation. X.L.: Writing—review and editing, Software, Project. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key Science and Technology Projects in Key Areas of Xinjiang Production and Construction Corps (2023AB062), and the National Undergraduate Innovation Training Program of China (202410755138).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Min, W.; Liu, L.; Liu, Y.; Luo, M.; Jiang, S. A Survey on Food Image Recognition. Chin. J. Comput. 2022, 45, 542–566. [Google Scholar]
  2. Parrish, E.A.; Goksel, A.K. Pictorial Pattern Recognition Applied to Fruit Harvesting. Trans. ASABE 1977, 20, 822–0827. [Google Scholar] [CrossRef]
  3. Bolle, R.M.; Connell, J.H.; Haas, N.; Mohan, R.; Taubin, G. VeggieVision: A produce recognition system. In Proceedings of the Third IEEE Workshop on Applications of Computer Vision, WACV’96, Sarasota, FL, USA, 2–4 December 1996; pp. 244–251. [Google Scholar]
  4. Kitamura, K.; Yamasaki, T.; Aizawa, K. Food log by analyzing food images. In Proceedings of the ACM Multimedia, Vancouver, BC, Canada, 27–31 October 2008. [Google Scholar]
  5. Yang, S.; Chen, M.; Pomerleau, D.; Sukthankar, R. Food recognition using statistics of pairwise local features. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2249–2256. [Google Scholar]
  6. Kagaya, H.; Aizawa, K.; Ogawa, M. Food Detection and Recognition Using Convolutional Neural Network. In Proceedings of the 22nd ACM international conference on Multimedia 2014, Orlando, FL, USA, 3–7 November 2014. [Google Scholar]
  7. Kawano, Y.; Yanai, K. FoodCam: A real-time food recognition system on a smartphone. Multimed. Tools Appl. 2015, 74, 5263–5287. [Google Scholar] [CrossRef]
  8. Myers, A.; Johnston, N.; Rathod, V.; Korattikara, A.; Gorban, A.; Silberman, N.; Guadarrama, S.; Papandreou, G.; Huang, J.; Murphy, K. Im2Calories: Towards an Automated Mobile Vision Food Diary. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1233–1241. [Google Scholar]
  9. Song, G.; Tao, Z.; Huang, X.; Cao, G.; Liu, W.; Yang, L. Hybrid Attention-Based Prototypical Network for Unfamiliar Restaurant Food Image Few-Shot Recognition. IEEE Access 2020, 8, 14893–14900. [Google Scholar] [CrossRef]
  10. Feng, J.; Lili, Z. Method for Recognition of Food Images Based on Improved Attention Model. J. Comput. Eng. Appl. 2024, 60, 153. [Google Scholar]
  11. Liang, H.; Wen, G.; Hu, Y.; Luo, M.; Yang, P.; Xu, Y. MVANet: Multi-Task Guided Multi-View Attention Network for Chinese Food Recognition. IEEE Trans. Multimed. 2021, 23, 3551–3561. [Google Scholar] [CrossRef]
  12. Sheng, G.; Sun, S.; Liu, C.; Yang, Y. Food recognition via an efficient neural network with transformer grouping. Int. J. Intell. Syst. 2022, 37, 11465–11481. [Google Scholar] [CrossRef]
  13. Sheng, G.; Min, W.; Zhu, X.; Xu, L.; Sun, Q.; Yang, Y.; Wang, L.; Jiang, S. A Lightweight Hybrid Model with Location-Preserving ViT for Efficient Food Recognition. Nutrients 2024, 16, 200. [Google Scholar] [CrossRef]
  14. Sheng, G.; Min, W.; Yao, T.; Song, J.; Yang, Y.; Wang, L.; Jiang, S. Lightweight Food Image Recognition with Global Shuffle Convolution. IEEE Trans. AgriFood Electron. 2024, 2, 392–402. [Google Scholar] [CrossRef]
  15. Min, W.; Wang, Z.; Liu, Y.; Luo, M.; Kang, L.; Wei, X.; Wei, X.; Jiang, S. Large Scale Visual Food Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9932–9949. [Google Scholar] [CrossRef]
  16. Chen, M.; Dhingra, K.; Wu, W.; Yang, L.; Sukthankar, R.; Yang, J. PFID: Pittsburgh fast-food image dataset. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 289–292. [Google Scholar]
  17. Taichi, J.; Keiji, Y. A food image recognition system with Multiple Kernel Learning. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 285–288. [Google Scholar]
  18. Hoashi, H.; Joutou, T.; Yanai, K. Image Recognition of 85 Food Categories by Feature Fusion. In Proceedings of the 2010 IEEE International Symposium on Multimedia, Taichung, Taiwan, 13–15 December 2010; pp. 296–301. [Google Scholar]
  19. Bossard, L.; Guillaumin, M.; Van Gool, L. Food-101—Mining Discriminative Components with Random Forests. In Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 446–461. [Google Scholar]
  20. Ito, Y.; Shimoda, W.; Yanai, K. Food image generation using a large amount of food images with conditional GAN: ramenGAN and recipeGAN. In Proceedings of the Joint Workshop on Multimedia for Cooking and Eating Activities and Multimedia Assisted Dietary Management, Stockholm, Sweden, 15 July 2018. [Google Scholar] [CrossRef]
  21. Wang, W.; Duan, P.; Zhang, W.; Gong, F.; Zhang, P.; Rao, Y. Towards a Pervasive Cloud Computing Based Food Image Recognition. In Proceedings of the 2013 IEEE International Conference on Green Computing and Communications and IEEE Internet of Things and IEEE Cyber, Physical and Social Computing, Beijing, China, 20–23 August 2013; pp. 2243–2244. [Google Scholar]
  22. Abiyev, R.; Adepoju, J. Automatic Food Recognition Using Deep Convolutional Neural Networks with Self-attention Mechanism. Hum. Centric Intell. Syst. 2024, 4, 171–186. [Google Scholar] [CrossRef]
  23. Bu, L.; Hu, C.; Zhang, X. Recognition of food images based on transfer learning and ensemble learning. PLoS ONE 2024, 19, e0296789. [Google Scholar] [CrossRef] [PubMed]
  24. Phiphitphatphaisit, S.; Surinta, O. Multi-layer adaptive spatial-temporal feature fusion network for efficient food image recognition. Expert Syst. Appl. 2024, 255, 124834. [Google Scholar] [CrossRef]
  25. Liu, Y.; Min, W.; Jiang, S.; Rui, Y. Convolution-Enhanced Bi-Branch Adaptive Transformer with Cross-Task Interaction for Food Category and Ingredient Recognition. IEEE Trans. Image Process. 2024, 33, 2572–2586. [Google Scholar] [CrossRef]
  26. Nadeem, M.; Shen, H.; Choy, L.; Barakat, J. Smart Diet Diary: Real-Time Mobile Application for Food Recognition. Appl. Syst. Innov. 2023, 6, 53. [Google Scholar] [CrossRef]
  27. Xiao, G.; Wu, Q.; Chen, H.; Cao, D.; Guo, J.; Gong, Z. A Deep Transfer Learning Solution for Automating Food Material Procurement using Electronic Scales. IEEE Trans. Ind. Inform. 2019, 16, 2290–2300. [Google Scholar] [CrossRef]
  28. Mohammad, I.; Mazumder, M.S.I.; Saha, E.K.; Razzaque, S.T.; Chowdhury, S. A Deep Learning Approach to Smart Refrigerator System with the assistance of IOT. In Proceedings of the International Conference on Computing Advancements, Dhaka, Bangladesh, 10–12 January 2020; p. 57. [Google Scholar]
  29. Herranz, L.; Ruihan, X.; Shuqiang, J. A probabilistic model for food image recognition in restaurants. In Proceedings of the 2015 IEEE International Conference on Multimedia and Expo (ICME), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
  30. Wang, Y.; Wu, J.; Deng, H.; Zeng, X. Food Image Recognition and Food Safety Detection Method Based on Deep Learning. In Computational Intelligence and Neuroscience; John Wiley & Sons Ltd.: Hoboken, NJ, USA, 2021; pp. 1–13. [Google Scholar]
  31. Min, W.; Liu, L.; Wang, Z.; Luo, Z.; Wei, X.; Wei, X.; Jiang, S. ISIA Food-500: A Dataset for Large-Scale Food Recognition via Stacked Global-Local Attention Network. In Proceedings of the 28th ACM International Conference on Multimedia 2020, Seattle, WA, USA, 12–16 October 2020. [Google Scholar]
  32. Chen, J.; Ngo, C.-W. Deep-based Ingredient Recognition for Cooking Recipe Retrieval. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 32–41. [Google Scholar]
  33. Majil, I.; Yang, M.T.; Yang, S. Augmented Reality Based Interactive Cooking Guide. Sensors 2022, 22, 8290. [Google Scholar] [CrossRef]
  34. Haque, R.U.; Khan, R.; Shihavuddin, A.; Syeed, M.; Uddin, F. Lightweight and Parameter-Optimized Real-Time Food Calorie Estimation from Images Using CNN-Based Approach. Appl. Sci. 2022, 12, 9733. [Google Scholar] [CrossRef]
  35. Kaur, P.; Sikka, K.; Wang, W.; Belongie, S.J.; Divakaran, A. FoodX-251: A Dataset for Fine-grained Food Classification. arXiv 2019, arXiv:1907.06167. [Google Scholar]
  36. Sahoo, D.; Wang, H.; Ke, S.; Wu, X.; Le, H.; Achananuparp, P.; Lim, E.-P.; Hoi, S.C.H. FoodAI: Food Image Recognition via Deep Learning for Smart Food Logging. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining 2019, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  37. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, South Korea, 27 October–2 November 2019. [Google Scholar]
  38. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
  39. Mezgec, S.; Koroušić Seljak, B. NutriNet: A Deep Learning Food and Drink Image Recognition System for Dietary Assessment. Nutrients 2017, 9, 657. [Google Scholar] [CrossRef] [PubMed]
  40. Pornpanomchai, C.; Srikeaw, K.; Harnprasert, V.; Promnurakkit, K. Thai fruit recognition system (TFRS). In Proceedings of the International Conference on Internet Multimedia Computing and Service, Kunming, China, 23–25 November 2009. [Google Scholar]
  41. Wen, W.; Jie, Y. Fast food recognition from videos of eating for calorie estimation. In Proceedings of the 2009 IEEE International Conference on Multimedia and Expo, New York, NY, USA, 28 June–3 July 2009; pp. 1210–1213. [Google Scholar]
  42. Kong, F.; Tan, J. DietCam: Regular Shape Food Recognition with a Camera Phone. In Proceedings of the 2011 International Conference on Body Sensor Networks, Dallas, TX, USA, 23–25 May 2011; pp. 127–132. [Google Scholar]
  43. Matsuda, Y.; Hoashi, H.; Yanai, K. Recognition of Multiple-Food Images by Detecting Candidate Regions. In Proceedings of the 2012 IEEE International Conference on Multimedia and Expo, Melbourne, VIC, Australia, 9–13 July 2012; pp. 25–30. [Google Scholar]
  44. Chen, M.-Y.; Yang, Y.-H.; Ho, C.-J.; Wang, S.-H.; Liu, S.-M.; Chang, E.; Yeh, C.-H.; Ouhyoung, M. Automatic chinese food identification and quantity estimation. In SIGGRAPH Asia 2012 Technical Briefs; Machinery: New York, NY, USA, 2012; pp. 1–4. [Google Scholar]
  45. Anthimopoulos, M.M.; Gianola, L.; Scarnato, L.; Diem, P.; Mougiakakou, S.G. A Food Recognition System for Diabetic Patients Based on an Optimized Bag-of-Features Model. IEEE J. Biomed. Health Inform. 2014, 18, 1261–1271. [Google Scholar] [CrossRef] [PubMed]
  46. Song, X.; Jiang, S.; Xu, R.; Herranz, L. Semantic Features for Food Image Recognition with Geo-Constraints. In Proceedings of the 2014 IEEE International Conference on Data Mining Workshop, Shenzhen, China, 14 December 2014; pp. 1020–1025. [Google Scholar]
  47. Tammachat, N.; Pantuwong, N. Calories analysis of food intake using image recognition. In Proceedings of the 2014 6th International Conference on Information Technology and Electrical Engineering (ICITEE), Yogyakarta, Indonesia, 7–8 October 2014; pp. 1–4. [Google Scholar]
  48. Yanai, K.; Kawano, Y. Food image recognition using deep convolutional network with pre-training and fine-tuning. In Proceedings of the 2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
  49. Ahmed, A.; Ozeki, T. Food Image Recognition by Using Bag-of-SURF Features and HOG Features. In Proceedings of the 3rd International Conference on Human-Agent Interaction 2015, Daegu, Republic of Korea, 21–24 October 2015. [Google Scholar]
  50. Bi, Y.; Lv, M.; Song, C.; Xu, W.; Guan, N.; Yi, W. AutoDietary: A Wearable Acoustic Sensor System for Food Intake Recognition in Daily Life. IEEE Sens. J. 2016, 16, 806–816. [Google Scholar] [CrossRef]
  51. Singla, A.; Yuan, L.; Ebrahimi, T. Food/Non-food Image Classification and Food Categorization using Pre-Trained GoogLeNet Model. In Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management, Amsterdam, The Netherlands, 16 October 2016. [Google Scholar]
  52. Liu, C.; Cao, Y.; Luo, Y.; Chen, G.; Vokkarane, V.; Ma, Y. DeepFood: Deep Learning-Based Food Image Recognition for Computer-Aided Dietary Assessment. In Proceedings of the Inclusive Smart Cities and Digital Health, Wuhan, China, 25–27 May 2016; pp. 37–48. [Google Scholar]
  53. Zheng, J.; Wang, Z.J.; Ji, X. Superpixel-based image recognition for food images. In Proceedings of the 2016 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE), Vancouver, BC, Canada, 15–18 May 2016; pp. 1–4. [Google Scholar]
  54. Tatsuma, A.; Aono, M. Food Image Recognition Using Covariance of Convolutional Layer Feature Maps. IEICE Trans. Inf. Syst. 2016, 99, 1711–1715. [Google Scholar] [CrossRef]
  55. Pouladzadeh, P.; Shirmohammadi, S. Mobile multi-food recognition using deep learning. ACM Trans. Multimed. Comput. Commun. Appl. 2017, 13, 1–21. [Google Scholar] [CrossRef]
  56. Giovany, S.; Putra, A.; Hariawan, A.S.; Wulandhari, L.A. Machine Learning and SIFT Approach for Indonesian Food Image Recognition. Procedia Comput. Sci. 2017, 116, 612–620. [Google Scholar] [CrossRef]
  57. Termritthikun, C.; Muneesawang, P.; Kanprachar, S. NU-InNet: Thai Food Image Recognition Using Convolutional Neural Networks on Smartphone. J. Telecommun. Electron. Comput. Eng. 2017, 9, 2289–8131. [Google Scholar]
  58. Zheng, J.; Wang, Z.J.; Zhu, C. Food Image Recognition via Superpixel Based Low-Level and Mid-Level Distance Coding for Smart Home Applications. Sustainability 2017, 9, 856. [Google Scholar] [CrossRef]
  59. Termritthikun, C.; Kanprachar, S. NU-ResNet: Deep Residual Networks for Thai Food Image Recognition. J. Telecommun. Electron. Comput. Eng. 2018, 10, 29–33. [Google Scholar]
  60. Horiguchi, S.; Amano, S.; Ogawa, M.; Aizawa, K. Personalized Classifier for Food Image Recognition. IEEE Trans. Multimed. 2018, 20, 2836–2848. [Google Scholar] [CrossRef]
  61. Ming, Z.-Y.; Chen, J.; Cao, Y.; Forde, C.; Ngo, C.-W.; Chua, T.S. Food Photo Recognition for Dietary Tracking: System and Experiment. In Proceedings of the MultiMedia Modeling: 24th International Conference, MMM 2018, Bangkok, Thailand, 5–7 February 2018; pp. 129–141. [Google Scholar]
  62. Reddy, V.H.; Kumari, S.; Muralidharan, V.; Gigoo, K.; Thakare, B.S. Food Recognition and Calorie Measurement using Image Processing and Convolutional Neural Network. In Proceedings of the 2019 4th International Conference on Recent Trends on Electronics, Information, Communication & Technology (RTEICT), Bangalore, India, 17–18 May 2019; pp. 109–115. [Google Scholar]
  63. Jiang, S.; Min, W.; Liu, L.; Luo, Z. Multi-Scale Multi-View Deep Feature Aggregation for Food Recognition. IEEE Trans. Image Process. 2020, 29, 265–276. [Google Scholar] [CrossRef] [PubMed]
  64. Merchant, K.; Pande, Y. ConvFood: A CNN-Based Food Recognition Mobile Application for Obese and Diabetic Patients. In Proceedings of the Emerging Research in Computing, Information, Communication and Applications, Singapore, 3 May 2019; pp. 493–502. [Google Scholar]
  65. Jeny, A.A.; Junayed, M.S.; Ahmed, I.; Habib, M.T.; Rahman, M.R. FoNet—Local Food Recognition Using Deep Residual Neural Networks. In Proceedings of the 2019 International Conference on Information Technology (ICIT), Bhubaneswar, India, 19–21 December 2019; pp. 184–189. [Google Scholar]
  66. Tasci, E. Voting combinations-based ensemble of fine-tuned convolutional neural networks for food image recognition. Multimed. Tools Appl. 2020, 79, 30397–30418. [Google Scholar] [CrossRef]
  67. Metwalli, A.S.; Shen, W.; Wu, C.Q. Food Image Recognition Based on Densely Connected Convolutional Neural Networks. In Proceedings of the 2020 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan, 19–21 February 2020; pp. 027–032. [Google Scholar]
  68. He, J.; Shao, Z.; Wright, J.; Kerr, D.; Boushey, C.; Zhu, F. Multi-task Image-Based Dietary Assessment for Food Recognition and Portion Size Estimation. In Proceedings of the 2020 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Shenzhen, China, 6–8 August 2020; pp. 49–54. [Google Scholar]
  69. Jiang, S.; Min, W.; Lyu, Y.; Liu, L. Few-shot food recognition via multi-view representation learning. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 16, 1–20. [Google Scholar] [CrossRef]
  70. Zhao, H.; Yap, K.H.; Kot, A.C.; Duan, L. JDNet: A Joint-Learning Distilled Network for Mobile Visual Food Recognition. IEEE J. Sel. Top. Signal Process. 2020, 14, 665–675. [Google Scholar] [CrossRef]
  71. Minija, S.J.; Emmanuel, W.R.S. Food recognition using neural network classifier and multiple hypotheses image segmentation. Imaging Sci. J. 2020, 68, 100–113. [Google Scholar] [CrossRef]
  72. He, X.; Wang, F.; Qu, Z. An Improved YOLOv3 Model for Asian Food Image Recognition and Detection. Open J. Appl. Sci. 2021, 11, 1287–1306. [Google Scholar] [CrossRef]
  73. Liu, C.; Liang, Y.; Xue, Y.; Qian, X.; Fu, J. Food and Ingredient Joint Learning for Fine-Grained Recognition. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 2480–2493. [Google Scholar] [CrossRef]
  74. Xiao, L.; Lan, T.; Xu, D.; Gao, W.; Li, C. A Simplified CNNs Visual Perception Learning Network Algorithm for Foods Recognition. Comput. Electr. Eng. 2021, 92, 107152. [Google Scholar] [CrossRef]
  75. Liu, Y.X.; Min, W.Q.; Jiang, S.Q.; Rui, Y. Food Image Recognition via Multi-scale Jigsaw and Reconstruction Network. J. Softw. 2022, 33, 4379–4395. [Google Scholar]
  76. Zhang, Q.; He, C.; Qin, W.; Liu, D.; Yin, J.; Long, Z.; He, H.; Sun, H.C.; Xu, H. Eliminate the hardware: Mobile terminals-oriented food recognition and weight estimation system. Front. Nutr. 2022, 9, 965801. [Google Scholar] [CrossRef]
  77. Liu, Y.-C.; Onthoni, D.; Mohapatra, S.; Irianti, D.; Sahoo, P. Deep-Learning-Assisted Multi-Dish Food Recognition Application for Dietary Intake Reporting. Electronics 2022, 11, 1626. [Google Scholar] [CrossRef]
  78. Cherpanath, E.D.; Nasreen, P.R.F.; Pradeep, K.; Menon, M.; Jayanthi, V.S. Food Image Recognition and Calorie Prediction Using Faster R-CNN and Mask R-CNN. In Proceedings of the 2023 9th International Conference on Smart Computing and Communications (ICSCC), Kochi, Kerala, 17–19 August 2023; pp. 83–89. [Google Scholar]
  79. Luo, L. Research on Food Image Recognition of Deep Learning Algorithms. In Proceedings of the 2023 International Conference on Computers, Information Processing and Advanced Education (CIPAE), Ottowa, ON, Canada, 26––28 August 2023; pp. 733–737. [Google Scholar]
  80. Zuo, W.; Zhang, W.; Ren, Z. Food Recognition and Classification Based on Image Recognition: A Study Utilizing PyTorch and PReNet. In Proceedings of the 2023 IEEE International Conference on Image Processing and Computer Applications (ICIPCA), Changchun, China, 11–13 August 2023; pp. 1325–1330. [Google Scholar]
  81. Liu, Y. Automatic food recognition based on efficientnet and ResNet. J. Phys. Conf. Ser. 2023, 2646, 012037. [Google Scholar] [CrossRef]
  82. Ramkumar, G.; Venkatramanan, C.B.; Manonmani, V.; Palanivel, S.; Sakthisaravanan, N.; Kumar, A.K. A Real-time Food Image Recognition System to Predict the Calories by Using Intelligent Deep Learning Strategy. In Proceedings of the 2023 International Conference on Research Methodologies in Knowledge Management, Artificial Intelligence and Telecommunication Engineering (RMKMATE), Chennai, India, 1–2 November 2023; pp. 1–7. [Google Scholar]
  83. Zhu, S.; Ling, X.; Zhang, K.; Niu, J. Food Image Recognition Method Based on Generative Self-supervised Learning. In Proceedings of the 9th International Conference on Computing and Artificial Intelligence, Tianjin, China, 17–20 March 2023; pp. 203–207. [Google Scholar]
  84. Rokhva, S.; Teimourpour, B.; Soltani, A.H. Computer vision in the food industry: Accurate, real-time, and automatic food recognition with pretrained MobileNetV2. Food Humanit. 2024, 3, 100378. [Google Scholar] [CrossRef]
  85. Xiao, Z.; Diao, G.; Deng, Z. Fine grained food image recognition based on swin transformer. J. Food Eng. 2024, 380, 112134. [Google Scholar] [CrossRef]
  86. Boyd, L.; Nnamoko, N.; Lopes, R. Fine-Grained Food Image Recognition: A Study on Optimising Convolutional Neural Networks for Improved Performance. J. Imaging 2024, 10, 126. [Google Scholar] [CrossRef]
  87. Wang, H.; Tian, H.; Ju, R.; Ma, L.; Yang, L.; Chen, J.; Liu, F. Nutritional composition analysis in food images: An innovative Swin Transformer approach. Front. Nutr. 2024, 11, 1454466. [Google Scholar] [CrossRef]
  88. Matsuda, Y.; Yanai, K. Multiple-food recognition considering co-occurrence employing manifold ranking. In Proceedings of the Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; pp. 2017–2020. [Google Scholar]
  89. Kawano, Y.; Yanai, K. Automatic Expansion of a Food Image Dataset Leveraging Existing Categories with Domain Adaptation. In Proceedings of the ECCV Workshops, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  90. Xin, W.; Kumar, D.; Thome, N.; Cord, M.; Precioso, F. Recipe recognition with large multimodal food dataset. In Proceedings of the 2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Turin, Italy, 29 June–3 July 2015; pp. 1–6. [Google Scholar]
  91. Xu, R.; Herranz, L.; Jiang, S.; Wang, S.; Song, X.; Jain, R. Geolocalized Modeling for Dish Recognition. IEEE Trans. Multimed. 2015, 17, 1187–1199. [Google Scholar] [CrossRef]
  92. Ciocca, G.; Napoletano, P.; Schettini, R. Food Recognition and Leftover Estimation for Daily Diet Monitoring. In Proceedings of the New Trends in Image Analysis and Processing—ICIAP 2015 Workshops, Genoa, Italy, 8 September 2015; pp. 334–341. [Google Scholar]
  93. Farinella, G.M.; Allegra, D.; Stanco, F. A Benchmark Dataset to Study the Representation of Food Images. In Proceedings of the Computer Vision—ECCV 2014 Workshops, Zurich, Switzerland, 6–12 September 2014; pp. 584–599. [Google Scholar]
  94. Ciocca, G.; Napoletano, P.; Schettini, R. Food Recognition: A New Dataset, Experiments, and Results. IEEE J. Biomed. Health Inform. 2017, 21, 588–598. [Google Scholar] [CrossRef]
  95. Klasson, M.; Zhang, C.; Kjellström, H. A Hierarchical Grocery Store Image Dataset with Visual and Semantic Labels. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 491–500. [Google Scholar]
  96. Zhou, F.; Lin, Y. Fine-Grained Image Classification by Exploring Bipartite-Graph Labels. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1124–1133. [Google Scholar]
  97. Merler, M.; Wu, H.; Uceda-Sosa, R.; Nguyen, Q.-B.; Smith, J. Snap, Eat, RepEat: A Food Recognition Engine for Dietary Logging. In Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management, Amsterdam, The Netherlands, 16 October 2016; pp. 31–40. [Google Scholar]
  98. Farinella, G.M.; Allegra, D.; Moltisanti, M.; Stanco, F.; Battiato, S. Retrieval and classification of food images. Comput. Biol. Med. 2016, 77, 23–39. [Google Scholar] [CrossRef]
  99. Ciocca, G.; Napoletano, P.; Schettini, R. Learning CNN-based Features for Retrieval of Food Images. In Proceedings of the New Trends in Image Analysis and Processing—ICIAP 2017, Catania, Italy, 11–15 September 2017; pp. 426–434. [Google Scholar]
  100. Chen, X.; Zhu, Y.; Zhou, H.; Diao, L.; Wang, D. ChineseFoodNet: A large-scale Image Dataset for Chinese Food Recognition. arXiv 2017, arXiv:1705.02743. [Google Scholar]
  101. Hou, S.; Feng, Y.; Wang, Z. VegFru: A Domain-Specific Dataset for Fine-Grained Visual Categorization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 541–549. [Google Scholar]
  102. Waltner, G.; Schwarz, M.; Ladstätter, S.; Weber, A.; Luley, P.; Lindschinger, M.; Schmid, I.; Scheitz, W.; Bischof, H.; Paletta, L. Personalized Dietary Self-Management Using Mobile Vision-Based Assistance. In Proceedings of the New Trends in Image Analysis and Processing—ICIAP 2017, Catania, Italy, 11–15 September 2017; pp. 385–393. [Google Scholar]
  103. Kaur, P.; Sikka, K.; Divakaran, A. Combining Weakly and Webly Supervised Learning for Classifying Food Images. arXiv 2017, arXiv:1712.08730. [Google Scholar]
  104. Lee, K.H.; He, X.; Zhang, L.; Yang, L. CleanNet: Transfer Learning for Scalable Image Classifier Training with Label Noise. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5447–5456. [Google Scholar]
  105. Min, W.; Jiang, S.; Sang, J.; Wang, H.; Liu, X.; Herranz, L. Being a Supercook: Joint Food Attributes and Multimodal Content Modeling for Recipe Retrieval and Exploration. IEEE Trans. Multimed. 2017, 19, 1100–1113. [Google Scholar] [CrossRef]
  106. Min, W.; Bao, B.K.; Mei, S.; Zhu, Y.; Rui, Y.; Jiang, S. You Are What You Eat: Exploring Rich Recipe Information for Cross-Region Food Analysis. IEEE Trans. Multimed. 2018, 20, 950–964. [Google Scholar] [CrossRef]
  107. Min, W.; Liu, L.; Luo, Z.; Jiang, S. Ingredient-Guided Cascaded Multi-Attention Network for Food Recognition. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1331–1339. [Google Scholar]
  108. Muresan, H.B.; Oltean, M. Fruit recognition from images using deep learning. Acta Univ. Sapientiae Inform. 2017, 10, 26–42. [Google Scholar] [CrossRef]
  109. Qiu, J.; Lo, F.P.-W.; Sun, Y.; Wang, S.; Lo, B.P.L. Mining Discriminative Food Regions for Accurate Food Recognition. In Proceedings of the British Machine Vision Conference, Cardiff, UK, 9–12 September 2019. [Google Scholar]
  110. Zhu, Y.; Zhao, X.; Zhao, C.; Wang, J.; Lu, H. Food det: Detecting foods in refrigerator with supervised transformer network. Neurocomputing 2020, 379, 162–171. [Google Scholar] [CrossRef]
  111. Tan, Y.; Lu, K.; Rahman, M.M.; Xue, J. Rgbd-Fg: A Large-Scale Rgb-D Dataset for Fine-Grained Categorization. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), Virtual, 6–10 July 2020; pp. 1–6. [Google Scholar]
  112. Chen, J.; Zhu, B.; Ngo, C.W.; Chua, T.S.; Jiang, Y.G. A Study of Multi-Task and Region-Wise Deep Learning for Food Ingredient Recognition. IEEE Trans. Image Process. 2021, 30, 1514–1526. [Google Scholar] [CrossRef]
  113. Wu, X.; Fu, X.; Liu, Y.; Lim, E.-P.; Hoi, S.C.H.; Sun, Q. A Large-Scale Benchmark for Food Image Segmentation. In Proceedings of the 29th ACM International Conference on Multimedia 2021, Virtual, 20–24 October 2021. [Google Scholar]
  114. Hou, Q.; Min, W.; Wang, J.; Hou, S.; Zheng, Y.; Jiang, S. FoodLogoDet-1500: A Dataset for Large-Scale Food Logo Detection via Multi-Scale Feature Decoupling Network. arXiv 2021, arXiv:2108.04644. [Google Scholar]
  115. Fan, B.; Li, W.; Dong, L.; Li, J.; Nie, Z. Automatic Chinese Food recognition based on a stacking fusion model. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney Australia, 24–27 July 2023; pp. 1–4. [Google Scholar]
  116. Chen, C.-S.; Chen, G.-Y.; Zhou, D.; Jiang, D.; Chen, D. Res-VMamba: Fine-Grained Food Category Visual Classification Using Selective State Space Models with Deep Residual Learning. arXiv 2024, arXiv:2402.15761. [Google Scholar]
  117. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  118. Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Computer Vision—ECCV 2018, 15th European Conference, Munich, Germany, 8–14 September 2018, Proceedings, Part XIV; Springer: Munich, Germany, 2018. [Google Scholar] [CrossRef]
Figure 1. Overview of the survey.
Figure 1. Overview of the survey.
Applsci 15 07626 g001
Figure 2. Sample images from a few food datasets.
Figure 2. Sample images from a few food datasets.
Applsci 15 07626 g002
Figure 3. The FRCNNSAM high-level model overview.
Figure 3. The FRCNNSAM high-level model overview.
Applsci 15 07626 g003
Figure 4. The whole process of food image recognition based on transfer learning and ensemble learning.
Figure 4. The whole process of food image recognition based on transfer learning and ensemble learning.
Applsci 15 07626 g004
Figure 5. The ASTFF-Net architecture.
Figure 5. The ASTFF-Net architecture.
Applsci 15 07626 g005
Figure 6. Framework of method based on attention mechanism.
Figure 6. Framework of method based on attention mechanism.
Applsci 15 07626 g006
Figure 7. The illustration for the whole framework. (a) The Convolution-Enhanced Bi-Branch Adaptive Transformer (CBiAFormer). (bd) The Structural Learning and Cross-Task Interaction (SLCI).
Figure 7. The illustration for the whole framework. (a) The Convolution-Enhanced Bi-Branch Adaptive Transformer (CBiAFormer). (bd) The Structural Learning and Cross-Task Interaction (SLCI).
Applsci 15 07626 g007
Table 1. Comparison of our review and previous surveys.
Table 1. Comparison of our review and previous surveys.
AspectOur ReviewPrevious Surveys [1,15]
ScopeCovers both historical and cutting-edge methods, including CNN-based, ViT, hybrid models, and lightweight architecturesPrimarily focused on CNNs or specific paradigms
Dataset AnalysisThis paper quantitatively analyzes the cultural bias of the dataset (such as the proportion of Asian dishes) and the annotation granularity problem and introduces the relevant methods to deal with the shortcomings of the datasetLimited discussion of dataset cultural coverage or real-world applicability
Method TaxonomyFour paradigms: CNN-based, ViT, multimodal fusion, lightweight models; includes performance comparisons (Tables S3 and S4)Often focused on single paradigms (e.g., CNNs) without cross-paradigm comparisons
ApplicationsExpands to nutrient assessment, food safety, and ethical challengesFocus on traditional scenarios, such as practical applications such as calorie estimation or dish identification
Challenges and FutureHighlights cross-cultural generalization, feature mismatch, and ethical risksEthical issues and multicultural dataset requirements are rarely addressed, and challenges are mostly technical
Table 2. Comparison of deep learning models in food image recognition.
Table 2. Comparison of deep learning models in food image recognition.
MethodsAdvantagesDisadvantages
CNN basedStrong ability to extract local features such as texture and shape
Simple structure, easier to train and deploy
Limited ability to model global semantic relationships
Overfitting can occur on small datasets
Vision TransformerSelf-attention mechanisms capture global dependencies
Strong robustness to data distribution changes
Suitable for large-scale pre-training
Performance declines on small datasets
High computational complexity, performance is more obvious on the high resolution image
Multimodal fusionCombine visual and non-visual information
Strong generalization ability in cross-cultural scenes
Support complex downstream tasks such as nutritional assessment
High labeling cost
Multi-source data alignment is complex
High complexity of model design
Lightweight neural network modelsSmall computational resource requirements
It has good flexibility and real-time performance in actual deployment
Accuracy may drop for larger models
Need a targeted optimization design
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, D.; Zuo, E.; Wang, D.; He, L.; Dong, L.; Lu, X. Deep Learning in Food Image Recognition: A Comprehensive Review. Appl. Sci. 2025, 15, 7626. https://doi.org/10.3390/app15147626

AMA Style

Liu D, Zuo E, Wang D, He L, Dong L, Lu X. Deep Learning in Food Image Recognition: A Comprehensive Review. Applied Sciences. 2025; 15(14):7626. https://doi.org/10.3390/app15147626

Chicago/Turabian Style

Liu, Detianjun, Enguang Zuo, Dingding Wang, Liang He, Liujing Dong, and Xinyao Lu. 2025. "Deep Learning in Food Image Recognition: A Comprehensive Review" Applied Sciences 15, no. 14: 7626. https://doi.org/10.3390/app15147626

APA Style

Liu, D., Zuo, E., Wang, D., He, L., Dong, L., & Lu, X. (2025). Deep Learning in Food Image Recognition: A Comprehensive Review. Applied Sciences, 15(14), 7626. https://doi.org/10.3390/app15147626

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop