Next Article in Journal
The Influence of the Vacuum Impregnation, Beetroot Juice, and Various Drying Methods on Selected Properties of Courgette and Broccoli Snacks
Previous Article in Journal
Comparison of Biological Activities and Protective Effects on PAH-Induced Oxidative Damage of Different Coffee Cherry Pulp Extracts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DPF-Nutrition: Food Nutrition Estimation via Depth Prediction and Fusion

1
School of Electronic Information and Communication, Huazhong University of Science and Technology, Wuhan 430074, China
2
Institute of Agricultural Products Processing and Nuclear Agricultural Technology, Hubei Academy of Agricultural Science, Wuhan 430064, China
*
Author to whom correspondence should be addressed.
Foods 2023, 12(23), 4293; https://doi.org/10.3390/foods12234293
Submission received: 26 October 2023 / Revised: 20 November 2023 / Accepted: 22 November 2023 / Published: 28 November 2023

Abstract

:
A reasonable and balanced diet is essential for maintaining good health. With advancements in deep learning, an automated nutrition estimation method based on food images offers a promising solution for monitoring daily nutritional intake and promoting dietary health. While monocular image-based nutrition estimation is convenient, efficient and economical, the challenge of limited accuracy remains a significant concern. To tackle this issue, we proposed DPF-Nutrition, an end-to-end nutrition estimation method using monocular images. In DPF-Nutrition, we introduced a depth prediction module to generate depth maps, thereby improving the accuracy of food portion estimation. Additionally, we designed an RGB-D fusion module that combined monocular images with the predicted depth information, resulting in better performance for nutrition estimation. To the best of our knowledge, this was the pioneering effort that integrated depth prediction and RGB-D fusion techniques in food nutrition estimation. Comprehensive experiments performed on Nutrition5k evaluated the effectiveness and efficiency of DPF-Nutrition.

1. Introduction

Dietary health has become the predominant focus in modern life. Excessive or imbalanced intake may lead to different kinds of diet-related diseases, especially obesity, which will dramatically increase the risk of hypertension, cardiovascular disease and diabetes [1]. Misestimating of nutrition content is a significant factor in excessive and imbalanced intake. The International Food Information Council (IFIC) Foundation reported that most people tend to overestimate their vegetable intake while underestimating their fat intake [2]. Therefore, there is an urgent demand for effective nutrition estimation methods to help individuals monitor their daily dietary intake and guide them to a healthier diet. Previous dietary assessment methods heavily relied on human involvement. Specifically, the commonly used 24 h Dietary Recalls [3] require participants to report their food types and portion sizes over a 24 h period, thereby understanding their eating behavior. Many popular applications, i.e., MyFitnessPal, MyDietCoach, Yazio, FatSecret, MyFoodDiary and Foodnotes, are all developed based on this method. Despite the advantage of being easy to implement, it is burdensome and unreliable due to its high dependence on the subjective judgements of the participants.
Fortunately, recent development in Artificial Intelligence (AI), especially in deep learning techniques, have made the automated and reliable dietary assessment a reality [4,5]. Vision-based nutrition estimation methods allow users to monitor their food intake by capturing images using their mobile devices, which heavily reduces user burden. According to the type of input data, existing methods can be broadly divided into three categories [6]: monocular image-based methods, multi-view image-based methods and RGB-D methods. Earlier works [7,8] have predominantly relied on multi-view images to reconstruct the 3D structure of food objects and estimate their volume. Subsequently, they calculated the sum of nutrition by combining the volume with the nutritional information of the food. Nevertheless, multi-view image methods are troublesome and inefficient since they require users to capture images from specific angles. In contrast, monocular image-based methods only relied on a single food image and demonstrated good performance. Specifically, Shao et al. [9] proposed a method for estimating calories based on an energy density map, which maps RGB images to the energy density of food on a pixel-to-pixel basis. They compared the proposed method with the manual 24 h Dietary Recalls and the results showed an obvious advantage. Similarly, Thames et al. [10] demonstrated that the performance of the vison-based method outperformed the professional human nutritionists where they used a multi-task convolutional neural network to estimate multiple nutrients. Shao et al. [11] employed a combination of non-destructive detection technology and deep learning to analyze the nutritional content of food. They improved the detection of small target foods to improve the nutrient estimation accuracy. However, estimating food nutrition from a monocular image is an ill-posed problem [12]. This is because the process of mapping food objects to monocular images often leads to a loss of crucial 3D information, which is vital for food portion estimation. In order to resolve the problem, depth information is utilized to complement the 3D information lost in monocular images. Lu et al. [13] utilized RGB-D pairs captured from real eating scenarios as input, and integrated techniques of food segmentation, recognition and 3D surface reconstruction to estimate nutrient intake for hospitalized patients. Thames et al. [10] incorporated the monocular images with depth maps as four-channel data, which were subsequently sampled into a three-channel tensor as the input of the model. However, these methods only treated food images and depth maps as image pairs, neglecting the inherent differences between RGB and depth images, which limits the performance of nutrition estimation. Recently, cross-modal fusion has been explored to enhance the performance of nutrition estimation. Vinod et al. [14] employed normalization techniques to address disparities in the feature space between the energy density map [9] and the depth map. Shao et al. [15] employed the balanced feature pyramid [16] and the convolutional block attention module [17] to enhance the fused features, but the fusion of RGB and depth features still employed a straightforward concatenation approach. These methods have facilitated the research on RGB-D nutrition estimation, primarily through their focus on modal disparity. However, the enhancements failed to explore the interactive effect of cross-modal features, thereby restricting the extraction of complementary information that could further enhance the accuracy of nutrition estimation. In addition, the acquisition of a depth map heavily relies on professional depth sensors, which increases the cost and restricts the application scenarios of RGB-D nutrition estimation methods.
In this paper, we proposed a novel food nutrition estimation method based on Depth Prediction and Fusion, referred to as DPF-Nutrition. We employed a depth prediction model that generated predicted depth maps instead of actual depth maps captured by depth sensors. These predicted depth maps recovered the missing 3D information in monocular images, thereby enhancing the accuracy of food portion estimation without additional cost. Unlike existing RGB-D fusion approaches in nutrition estimation [14,15] that enhanced features on a single modality, we proposed a Cross-modal Attention Block (CAB) that focused on the interaction effect of the cross-modal features. The CAB utilized cross-modal attention features to enhance single-modal features, leading to increased complementarity between modalities and the generation of more discriminative fused features. This improvement allows the model to accurately focus on the correct nutrient regions. In addition, we proposed a multi-scale fusion network that enhanced the semantic features through combining the fused features at different resolutions, thereby enabling the model to capture the co-occurring food features in the feature maps. The proposed CAB and multi-scale fusion network consist of the RGB-D fusion module in DFP-Nutrition. This module effectively exploits the features of both RGB and depth images to enhance the performance of food nutrition estimation. During the inference stage, DPF-Nutrition relies solely on monocular food images as input, making it essentially a monocular image-based method. We evaluated the effectiveness of DPF-Nutrition on the public Nutrition5k dataset and obtained encouraging results. Specifically, the percentage mean absolute errors (PMAEs) for calories, mass, protein, fat and carbohydrate estimation reached 14.7%, 10.6%, 20.2%, 22.6% and 20.7%, respectively. Compared with the previous monocular image-based method proposed by Thames et al. [10], the mean PMAE of our DPF-Nutrition reached 17.8%, improved by 11.3%. Furthermore, when compared with the state-of-the-art RGB-D method proposed by Shao et al. [15], our DPF-Nutrition demonstrated competitive performance with a 0.7% improvement.
The contribution of this paper can be summarized as three aspects:
  • We proposed a novel monocular image-based nutrition estimation method based on Depth Prediction and Fusion, referred as DPF-Nutrition. It was the first attempt to incorporate depth prediction and RGB-D fusion techniques in vision-based nutrition estimation.
  • We designed an RGB-D fusion module that incorporated the proposed cross-modal attention block (CAB) and multi-scale fusion network to fully exploit the informative image features for nutrition estimation.
  • Our proposed DPF-Nutrition demonstrated effectiveness in accurately estimating multiple nutrients, which has been evaluated on the public dataset Nutrition5k.

2. Materials and Methods

2.1. Dataset

We summarized diversity, size, annotations and inclusion of depth information in nutrition datasets. The result is shown in Table 1. Overall, the Nutrition5k dataset [10] contained the highest number of unique dishes and images. Furthermore, it contained extensive nutritional annotations, whereas other datasets only included single calorie or portion annotations. In this paper, we evaluated our DPF-Nutrition method on the Nutrition5k dataset. This dataset comprises 20k short videos and 3.5k RGB-D images captured by an Intel RealSense camera, involving approximately 5k distinct food dishes. Each dish within the dataset includes detailed information such as ingredient names, quantities and associated macronutrient data calculated using the reliable USDA Food and Nutrient Database [18]. Figure 1 demonstrates examples of images in the Nutrition5k dataset. In the depth map, objects that are closer to the camera are represented in blue, while objects that are farther away are shown in red. The color bar serves as a visual indicator, reflecting the varying distances from each point to the camera, with the unit of measurement in centimeters. As an image-based nutrition estimation method, our DPF-Nutrition was evaluated on the 3.5k food images from the Nutrition5k dataset. The Nutrition5k dataset provides a predefined split, dividing the data into training and testing subsets with a ratio of 5:1. There exists a small number of erroneous images, as shown Figure 2, which have the potential to mislead the training of the model, resulting in suboptimal performance. However, in order to maintain consistency in the comparison of methods in this research, we have chosen to preserve the integrity of the dataset and refrain from performing any additional data cleaning.

2.2. Methods

The overall architecture of our DPF-Nutrition is illustrated in Figure 3. The model comprises two main modules:
  • Module1: Depth prediction module aims to reconstruct the 3D depth information based on 2D monocular images. The depth prediction module employs a vision transformer as the encoder which can reduce the loss of granularity and feature resolution for more accurate 3D information recovery.
  • Module2: RGB-D fusion module is specifically designed to fully leverage the features of RGB and predicted depth images for nutrition estimation. This module integrates the proposed multi-scale fusion network and cross-modal attention block (CAB). The multi-scale fusion network effectively enriches the intricate semantic features of fine-grained food images, while the CAB further enhances the complementarity of RGB and depth features.
To achieve accurate nutrition estimation, both food recognition and portion estimation are essential. The depth prediction module plays a crucial role in generating depth maps, which serve as the basis for precise portion estimation. The multi-scale fusion network combines multi-level feature maps to generate a high-resolution, strongly semantic feature map, which enhance the extensively recognition of the co-occurring food items in the images. DPF-Nutrition is designed to estimate the nutritional value of food in an end-to-end manner; thus, CAB integrates the complementary information from RGB images and predicted depth maps, enabling the cross-modal information to collaborate in nutrition estimation.
In general, our DPF-Nutrition consists of depth prediction module and RGB-D fusion module, as shown in Figure 3. Specifically, we utilized the ResNet101 network [23] as the backbone of our DPF-Net. In depth prediction module, we employed the Dense Prediction Transformer (DPT) [24] to generate a predicted depth map. The generated depth map was then combined with the RGB image as the input of RGB-D fusion module. In RGB-D fusion module, we extracted RGB features and depth features using separate ResNet networks. The extracted features at the same ResNet layer were then fed into the Cross-modal Attention Block (CAB). The CAB suppressed redundant features within each modality and combined the complementary features. Then, the fused RGB-D features of different resolutions were fed into the third ResNet network and fused from shallow layers to deep layers. This multi-scale fusion process ensured that the final feature map contained comprehensive and detailed information. The feature map was finally processed through global average pooling and a separate multi-task head to estimate the nutritional composition.

2.2.1. The Depth Prediction Module

Depth prediction plays a crucial role in computer vision since it enhances the understanding and perception of real 3D scenes. The depth prediction model typically follows a pattern comprising an encoder and a decoder [25,26,27]. The encoder extracts the features from the input images, while the decoder combines the features from the encoder and converts them into the ultimate depth prediction. The choice of the backbone network for the encoder is essential since the feature information that is lost in encoder cannot be recovered in following decoder. Compared with the depth information of other outdoor scenes or large objects, the depth information of food is more intricate due to its varied geometry and abundant texture. The loss of resolution and granularity during feature extraction in the encoder is prone to causing distortion in the final depth prediction. However, the convolutional neural network [23,28] inevitably loses granularity in deeper stages, since the increase in receptive field and abstraction of features are reliant on the downsampling operation. In comparison, Vision Transformer (ViT) [29] abandons the downsampling operation and maintain the global receptive field of all stages. This makes it more suitable to be used as an encoder for fine-grained food images.
The depth prediction module comprised a transformer encoder and a convolutional decoder. The transformer encoder was responsible for extracting the bag-of-words representation, while the decoder reconstructed the bag-of-words representation into image-like features of different scales. Finally, image-like features were combined into the depth estimation. The structure of DPT is showed in Figure 4a.
Specifically, the input images were abstracted as two-dimensional feature vectors by ResNet-50 network to satisfy the input format of transformer. The feature vectors were then combined with a trainable position embedding to retain positional information. The resulting sequence of embedding vectors, referred to as tokens, were fed into transformer encoder to extract image feature information. Transformer block comprised multi-head self-attention layer, multi-layer perceptron (MLP), two layers of LayerNorm (LN) and residual connections. The structure of transformer encoder is shown in Figure 4b. Multi-head attention enabled transformer encoders to learn representations of both global and local image features. This capability empowered Vision Transformers (ViTs) to capture features at multiple scales, eliminating the requirement for traditional convolutional networks’ gradual downsampling operations. This advancement significantly contributed to maintaining feature resolution and granularity, which are crucial for accurate depth prediction. Given that the resolution of the input image is H × W, after transformer encoders, we obtained a set of tokens t = { t 0 t N p }, t n R D , where N p = H × W p 2 , D refers to the dimension of tokens and p is the sampling rate of ResNet50.
Then, the N p tokens were spatially concatenated into an image-like feature map through placing each token based on the information of the position embedding:
C o n c a t n a t e : R N p × D R H p × W p × D
We resampled the image-like representations into specific size H s × W s with D dimensions:
R e s a m p l e : R H p × W p × D R H s × W s × D
Finally, we employed a RefineNet-based decoder [30] to progressively combine the feature maps at different resolutions and generate the depth prediction. More detailed implementation of DPT can be obtained in the paper [24].

2.2.2. RGB-D Fusion Module

RGB-D fusion is widely applied in various visual tasks including image classification [31], food intake detection [32] and food nutrition assessment [15]. The essence of RGB-D fusion is to combine the complementary information from RGB and depth images to produce more informative and enhanced features. RGB images and depth images pertain to distinct modalities, featuring fundamentally distinct information. Thus, straightforward concatenating or summing cannot fully exploit the cross-modal features. To address this issue, some methods explored improving the efficiency of RGB-D fusion through feature enhancement.
We summarized existing enhancement methods for RGB-D fusion in nutrition estimation into two paradigms: enhancement–fusion [14] and fusion–enhancement [15], as is shown in Figure 5. The specific enhancement techniques can be various. For example, Vinod et al. [14] employed normalization techniques to address disparities in the feature space while Shao et al. [15] employed the balanced feature pyramid and the convolutional block attention module to enhance the fused features. Although these two paradigms could enhance the efficiency of cross-modal fusion, they have certain limitations. The fusion–enhancement method focused on independently enhancing each single-modal feature, without considering the modal complementarity. As a result, this approach might lead to the loss of useful information. In contrast, the enhancement–fusion method enhanced the features after fusion, which might result in residual redundant information from each modal. Differently from the existing methods, we proposed a “fusion–enhancement–fusion” paradigm. Firstly, we performed a straightforward fusion of RGB and depth features to acquire cross-modal interaction information which is then used to enhance both of the single-modal features. Finally, the enhanced single-modal features were combined to generate the fused RGB-D features. This method can effectively remove redundant information from single-modal features while retaining complementary information from cross-modal features. Following this fusion paradigm, we propose a novel cross-modal attention block (CAB) for RGB-D fusion, which is demonstrated in Figure 6. By utilizing the CAB, we enhanced informative features and filter out redundant ones to obtain the most informative fused features. This enables our model to accurately prioritize regions with high nutritional value, rather than solely focusing on specific food items or larger portions.
Specifically, after the depth prediction module, the RGB images together with estimated depth images were input into the feature extraction networks to generate hierarchical features, denoted as RGB features {R i } i = 0 4 and depth features {D i } i = 0 4 . The RGB and depth feature maps, with the same resolution, were fed into CAB to generate the complementary fused representation. The structure of CAB is shown in Figure 6. The CAB explicitly built the feature correspondences among different modalities based on the channel and spatial attention vectors of additive features, which emphasized crucial features and suppressed redundant ones. The re-calibrated features were concatenated to obtain complementary cross-modality features. Specifically, given two input features R i and D i , we first added the two features pixel by pixel along the channel dimension. Then, we processed the additive features in two branches to obtain channel attention vectors and spatial attention vectors. In channel attention branch, we employed global average pool to obtain global channel descriptor. Then, the global channel descriptor was fed into a 1 × 1 convolution with BN and ReLU to increase non-linearity. Finally, it went through a sigmoid activation function to produce the channel attention vector; the procedure can be defined as:
C A = S i g ( C o n v 1 × 1 ( G A P ( R i D i ) ) )
where GAP(·) denotes the global average pooling, ⨁ denotes the element-wise addition, Sig(·) represents the sigmoid function, and C o n v 1 × 1 indicates a convolutional layer with 1 × 1 kernel size, followed by BN and ReLU.
In spatial attention branch, we calculated the average value for all pixels of the additive feature map along the channel dimension to obtain spatial descriptor. Then, we applied a 3 × 3 convolution with BN and ReLU to smooth the spatial descriptor. Finally, the spatial attention vector was obtained by passing it through a sigmoid activation function; the procedure can be defined as:
S A = S i g ( C o n v 3 × 3 ( M e a n ( R i D i ) ) )
where Mean(·) denotes mean function along the channel dimension. The attention weights of channel and spatial dimensions enhanced the correlation and complementarity of the RGB and depth features. Based on cross-modal attention, the enhanced RGB feature map and depth feature map were obtained. The enhanced features were then further concatenated and fed into a 1 × 1 convolution to obtain the complementary RGB-D features C i ; the procedure can be defined as:
C i = C o n v 1 × 1 ( C o n c a t ( R i C A S A , D i C A S A ) )
where Concat(·) denotes cross-channel concatenation and ⨂ indicates pixel-wise multiplication.
After the proposed CAB, we obtained the cross-modal features {C i } i = 0 4 . To further enhance semantic features, we adopted a multi-scale fusion network to combine the local detailed information of low-level features and the global context information of the high-level features, progressively. This enabled our model to extensively capture the co-occurring food items, especially the small objects in the images, resulting in accurate nutrition estimation. We used the first cross-modal feature map F 0 as the input of the feature fusion network; the feature was then refined by a ResNet convolutional block and combined with the next cross-modal feature map C 1 to generate the fused feature map F 1 . We repeated this operation at the next stage. In this way, we combined the cross-modal features at different scales. The procedure can be defined as:
F i = C i R e s i ( C i 1 )
where Res i indicates the i-th ResNet convolutional block.
Through the multi-scale fusion network and CAB, we fully utilized the complementary RGB and depth information to generate the final feature representation F 4 . Then, it was fed into a full connected (FC) layer and five multi-task FC heads (with dimensions 2048 and 1, respectively) to generate the estimated nutrition values.

2.2.3. Loss Function

As a multi-task learning model, our DPF-Nutrition predicts the contents of calories, mass and three essential macronutrients of food (fat, carbohydrates and protein). For each subtask, L1 loss was used to measure the bias between the estimated nutritional values and ground-truth ones, which can be defined as:
L c a l = 1 N i = 1 N y c a l y c a l
where L c a l denotes the subtask loss of calories, y c a l indicates the estimated value of calorie and y c a l represents the ground-truth calorie value. The loss of other subtasks follow this equation.
The scale of the subtask losses are various, which can cause some tasks to dominate other tasks during the training phase. The geometric loss combination [33] is invariant to the scale of individual losses, thereby maintaining a balanced approach towards subtask losses of varying scales. Thus, we used the geometric loss strategy as our loss function, which can be defined as:
L t o t a l = L c a l L m a s s L f a t L c a r b L p r o t e i n 5
where L t a t a l denotes the overall loss function.

2.3. Evaluation Metrics

In this paper, we adopted the two evaluation metrics of mean absolute error (MAE) and percentage of mean absolute error (PMAE), which are defined as:
M A E = 1 N i = 1 N y i y i
P M A E = MAE 1 N i = 1 N y i
where y is the estimated nutrient value while y is the ground-truth nutrient value. Caloric values are measured in standard kilocalorie units, while the other three nutrient values are measured in grams. A higher level of accuracy in nutrient estimation is achieved when the MAE and PMAE values for evaluation are lower.
In addition, to evaluate the accuracy of depth estimation, we adopt the three evaluation metrics of absolute relative error (AbsRel), root mean squared error (RMSE) and accuracy with threshold, which are defined as:
A b s R e l = 1 N i = 1 N d i d i d i
R M S E = 1 N i = 1 N d i d i 2
A c c u r a c y = % o f d i s . t . m a x ( d i d i , d i d i ) = δ < t h r
where d represents the ground-truth depth value while d i represents the predicted depth value. A higher level of accuracy in depth prediction is achieved when the AbsRel and RMSE values are lower and accuracy is nearer to 1.

3. Results

3.1. Experimental Detail

All the experiments were conducted on a 24G NVIDIA GTX 3090 GPU. To maintain the experimental fairness, the same setup was utilized for all experiments. For the training of the depth prediction module, we resized the input images to have a long side of 384 pixels and train on random square crops of size 384. The process not only meets the input size requirement of the vision transformer but also serves as an augmentation technique to enhance the model’s generalization capabilities. The encoder network was initialized with ImageNet1K pre-trained weight while the decoder network was initialized randomly. We utilized the Adam optimizer [34] with an initial learning rate of 1 × 10 5 and implemented a cosine annealing strategy for learning rate decay. The learning rate declined to 1 × 10 6 after cosine annealing. We trained the model for 60 epochs with a batch size of 8. For the training of the RGB-D fusion module, the input images were resized to 336 × 448 pixels to reduce memory consumption while maintaining the original aspect ratio. Image augmentation methods including center cropped and random horizontal flip were applied to training images to enhance the model’s generalization capabilities. The backbone network was initialized with Food2K [35] pre-trained weight. Food2K is a large dataset for fine-grained food recognition that offers features specifically suitable for transfer learning in food-related vision tasks. We chose the Adam optimizer with an initial learning rate of 5 × 10 5 and implemented an exponential decay strategy for updating the learning rate, with a decay rate set to 0.98. We trained the model for 150 epochs with a batch size of 8. The optimization techniques, including optimizer selection, learning rate strategies and batch size are all carefully chosen based on experimental experiences, in order to achieve the best possible results.

3.2. Backbone Comparison

It is essential to select a suitable backbone networks for the model. In this section, we made a comparison among several widely used convolutional neural networks (CNNs) and the recently popular vision transformer [29]. According to Table 2, ResNet101 achieved better performance, with the best mean PMAE of 20.9%. Therefore, we used ResNet101 [23] as the backbone network for our DPF-Nutrition, unless otherwise specified.

3.3. Depth Prediction Analysis

To evaluate the performance of the Dense Prediction Transformer (DPT) in food images, we compare it with three depth prediction methods based on fully convolutional networks. To ensure a fair comparison, we consistently maintained the experimental settings and dataset splits across all methods. We compare the methods including: DPT [24], FCRN [37], UNet [26] and HRNet [38]. The experimental results are shown in Table 3. Due to the vision transformer’s capability of preserving image resolution and granularity throughout the encoding process, the Dense Prediction Transformer (DPT) achieved the best performance in depth estimation.
Furthermore, in order to assess the influence of depth map quality on nutrition estimation, we evaluated the performance of nutrition estimation models that integrated RGB images and depth maps. These depth maps included the ground-truth ones and those generated by DPT, FRCN, UNet and HRNet. All the experiments employed the same RGB-D fusion module and experimental settings. The result is shown in Table 4. The results of the actual depth maps were best, with a mean PMAE of 17.2%. And a better performance of nutrition estimation was achieved when the quality of the predicted depth maps was better.

3.4. Cross-Modal Fusion Analysis

To evaluate the effectiveness of our proposed CAB, we compared it with two RGB-D fusion methods. The result is shown in Table 5. These two methods followed the cross-modal fusion paradigms demonstrated in Figure 5, referred to as “fusion–enhancement” and “enhancement–fusion”. To ensure fairness in the comparison, we maintained consistency in the specific feature enhancement method, which included spatial attention and channel attention mechanisms. Additionally, the experimental settings and dataset splits remained the same. CAB utilized cross-modal interaction information to enhance single-modal features so as to generate complementary fused features. Among the three methods compared, CAB demonstrated the best performance by following a “fusion–enhancement–fusion” paradigm, resulting in a mean PMAE of 17.8%.

3.5. Method Comparison

To evaluate the performance of DFP-Nutrition, we compared it with three representative nutrition estimation methods. To ensure a fair comparison, we consistently maintained the experimental settings and dataset splits across all methods. The experimental results are shown in Table 6. We referred to the methods proposed by Thames et al. [10] and Shao et al. [15] as Google-Nutrition and RGB-D Nutrition, respectively. We compared the methods including: Google-Nutrition, RGB-D Nutrition, Swin-Nutrition [11] and our proposed DPF-Nutrition.
Google-Nutrition was introduced in conjunction with the publication of the Nutrition5k dataset, which was widely regarded as the baseline method for nutrition estimation. Google-Nutrition offered two variations: Google-Nutrition-monocular, which utilized monocular images, and Google-Nutrition-depth, which incorporated food images and depth data. Swin-Nutrition achieved the state-of-the-art performance in nutrition estimation based on monocular images. Swin-Nutrition employed Swin-Transformer [39] as the backbone and used a feature fusion module to obtain discriminative feature representation to improve the accuracy of nutrition estimation. It should be noted that the results of Swin-Nutrition in this paper differed from the original ones published in [11]. This is because we replicated their method to avoid the data inconsistency caused by their experimental data cleaning. RGB-D Nutrition achieved the state-of-the-art results in nutrition estimation based on RGB-D images. RGB-D Nutrition employed balanced feature pyramid [16] and convolutional block attention module [17] to perform effective RGB-D fusion to improve the accuracy of nutrition estimation.
As demonstrated by the experimental results, DPF-Nutrition showed advantages in monocular image-based nutrition estimation. In comparison to Google-Nutrition-monocular and Swin-Nutrition, the mean PMAE improved by 11.3% and 2.6%, respectively. This improvement highlights the effectiveness of the deep prediction module in enhancing the accuracy of nutrition prediction by incorporating 3D information. Furthermore, when compared to the state-of-the-art RGB-D method, DPF-Nutrition demonstrated a competitive performance with a 0.7% improvement in the mean PMAE. Considering the potential impact caused by the bias between predicted depth maps and actual depth maps, the improvement underscored the effectiveness of our exploration of RGB-D fusion for nutrition estimation. Overall, our DPF-Nutrition exhibited a highly competitive performance when compared to existing methods.

3.6. Ablation Study

To validate the effectiveness of the proposed modules, comprehensive ablation studies were conducted on the various components comprising our DPF-Nutrition. The experimental results are shown in Table 7. The baseline was the RGB stream that used only a ResNet101 network to estimate nutrients from RGB images. Similarly, depth stream used a single network to estimate nutrients from depth images generated by the depth prediction module. Model (c) and Model (d) employed simple feature vector concatenation and the proposed multi-scale fusion network, respectively, to integrate the RGB stream and depth stream. Model (e) represented the complete DPF-Nutrition which incorporates the CAB on the basis of Model (d).
According to Table 7, Model (a) achieved a mean PMAE of 20.9%, whereas model (b) achieves a mean PMAE of 38.6%. The results demonstrated that relying solely on depth information alone cannot accurately estimate nutrients. Model (c) concatenated the RGB and depth features at the last layer of the feature extraction network, and achieves a mean PMAE of 19.6%, which is a 1.5% improvement compared to Model (a). The result demonstrated the effectiveness of complementing RGB images with estimated images for nutrition estimation and validated the efficacy of our depth prediction module. The improvements achieved by Model (d) and Model (e) evaluated the effectiveness of the proposed multi-scale fusion network and Cross-modal Attention Block (CAB), respectively. Model (e) achieved an improvement over Model (c), with a 1.8% decrease in mean PMAE, as well as reductions of 4.8 kCal in calories MAE and 3.8 g in mass MAE. The results of the ablation experiment demonstrated that all of the depth prediction module, multi-scale fusion network and CAB can effectively improve the accuracy of nutrient prediction.

3.7. Visualization Analysis

To visually showcase the efficacy of our method, we began by visualizing the performance of the depth prediction module. As depicted in Figure 7, our depth prediction module successfully recovered the depth information and smoothed out the noise in the original depth.
Next, we employed Grad-Cam [40] to visualize the RGB-D fusion module, which generated a heat map highlighting the regions of interest (ROIs) identified by the model. This visualization allowed for a more intuitive understanding of how our method leveraged food images to make accurate predictions. The visualization results of certain dishes and the corresponding nutrition facts are shown in Figure 8. The regions that the model focuses on are marked in red, while the opposite regions are marked in blue. As illustrated in the visualization, our model demonstrated a focused attention on specific regions for each nutrient task. For instance, when estimating the fat and protein content in dish_156278816, the model exhibited a strong focus on the pork and fish components. Conversely, when estimating the carbohydrate content, the model primarily emphasized the rice and corn elements. This indicated that our model effectively captured relevant visual cues and assigned appropriate importance to different regions based on the specific nutrient being estimated.

4. Discussion

Daily nutrient intake is essential for people’s health. According to previous studies, it has been found that a moderate caloric restriction (ranging from 18% to 30%) can lead to various positive effects on the health of obese patients, including a reduction in visceral fat, an improvement in insulin sensitivity and a decrease in the risk of developing metabolic diseases [41,42]. Furthermore, the results of a long-term experiment revealed that, after therapy, the weight of participants of obesity class II decreased by 22.4% while their calorie and carbohydrate intake decreased by 18.3% and 15.6%, respectively [43]. Although our DPF-Nutrition demonstrated an improvement in nutrition estimation, it is important to note that there were still errors in estimating the contents of calories, mass, protein, fat and carbohydrates, which reached 14.7%, 10.6%, 20.2%, 22.6% and 20.7% respectively. These errors may be disadvantageous, especially in the context of dietary monitoring for medical purposes.
In order to investigate the limitations of our model, we examined the test samples that yielded poor results. We discovered that the most significant errors primarily originated from three categories of images. The first category includes images with food stacking and covering. For instance, in the case of dish_1560367980, where the low-calorie spinach significantly obscured the high-calorie pizza, the estimated calorie value was 31.6% lower than the actual value. The second category comprises images containing minuscule components that are imperceptible, such as oil and sugar. For example, in the case of dish_1562617939, where a large amount of olive oil was added to the dish, the calorie PMAE was up to 52.5%, while the fat PMAE was dramatically 68.3%. The third category comprises images with food items that are rarely encountered in the training data. For example, in the case of dish_1562617703, which involved the uncommon ice-cream in the Nutrition5k dataset, the calorie PMAE reached up to 39.5%, while the carbohydrate PMAE reached up to 57.4%. It is evident that food stacking, imperceptible components and insufficient data are the primary factors that restrict the performance of DPF-Nutrition. When we removed these samples in the experiments, the performance of DPF-Nutrition in the estimation of calories, mass, protein, fat and carbohydrates reached 13.4%, 10.2%, 21.2%, 19.2% and 18.9%, respectively. The mean PMAE improved by 1.2%. Although removing these challenging samples can result in improved results, the dietary scenarios containing food stacks and imperceptible ingredients cannot be ignored, as they are real-life situations. In addition, the impact of scarcity of datasets on food nutrition estimation is significant. Deep learning techniques rely heavily on the dataset; however, the diversity and scale of the existing nutrition datasets are far from sufficient for practical application. Furthermore, due to the high local and cultural characteristics of food, the transferability between datasets of different cuisines are relatively weak. For example, the model trained on the Western food dataset cannot be directly used for nutrition estimation of Chinese food. Generally, Chinese dishes offer a wide range of variety, with complex ingredients and richer cooking methods, which can pose more challenges in vision-based nutrition estimation.
In the future, we will focus on overcoming the problems of food stacking and imperceptible components for more accurate nutrition estimation. One potential solution for food stacking involves detecting the stacked portion in the food image, conducting structural analysis on this segment and ultimately calculating nutrition adjustments. Furthermore, we are committed to creating an extensive public food dataset to facilitate comprehensive dietary assessment. Given the limited transferability across food datasets, our goal is to construct a specialized dataset, specifically focusing on a Chinese breakfast dataset. This dataset will be compact yet of high quality, ensuring its practical usability in daily applications. In addition to further improving our algorithms, we are also ambitious to integrate DPF-Nutrition into practical applications. Many existing dietary tracking apps such as MyFitnessPal, MyDietCoach and FatSecret require some manual input, such as recalling what you ate or weighing the food. In contrast, our proposed DPF-Nutrition achieves a fully automated process from capturing an image to estimating its nutritional content. The practical application of DPF-Nutrition holds immense potential in alleviating user burden, reducing costs and enhancing accuracy in dietary monitoring. However, the practical application of DPF-Nutrition should consider additional issues such as computational complexity and image quality. Our current design primarily focuses on improving accuracy, without taking into account the computational complexity of the model. Yet, in practical scenarios, the computational complexity directly affects resource utilization and running speed, which are closely associated with user experience. Therefore, in the next stage, we need to address the problem of reducing the model’s complexity while maintaining accuracy. Moreover, factors like angles, distances and lighting conditions during the image capture process can all impact image quality. It is crucial to find solutions to mitigate the impact of low image quality on nutrition estimation, as this is another practical issue that needs consideration in the application of DPF-Nutrition.
We strongly believe that the study of automated nutritional assessment holds great promise. By automating dietary assessment, individuals can receive real-time feedback on the nutritional content of their meals, thereby increasing their awareness and understanding of their dietary choices. This empowers individuals to make informed decisions about their food intake and ultimately adopt healthier eating habits. The convenience, affordability and efficiency of automated dietary assessment make it suitable for meeting the dietary monitoring needs of diverse populations. It is our hope that our study will contribute to the advancement of automated dietary assessment, ultimately enhancing dietary education and improving public health outcomes.

5. Conclusions

In this paper, we proposed our DFP-Nutrition for dietary assessment, aiming to develop an automated, cost-effective and precise method for nutrition estimation. Our proposed method offered a novel approach by predicting the depth map from a monocular image and incorporating the recovered 3D information with RGB food images to improve nutrition estimation. To assess the effectiveness of our method, we performed experiments on the Nutrition5k dataset and compared its performance with that of state-of-the-art image-based nutrition estimation methods. The results clearly demonstrated the effectiveness of our proposed method, showcasing its competitive performance when compared to other existing methods. In the future, we envision the widespread utilization of automated vision-based nutrition estimation methods in our daily lives, making a significant contribution to improving dietary education and public health.

Author Contributions

Conceptualization, Q.C. and Y.H.; methodology, Y.H.; software, Y.H.; validation, Y.H.; data curation, Y.H. and Z.H.; writing—original draft preparation, Y.H.; writing—review and editing, Q.C. and Y.H.; visualization, Y.H.; supervision, Q.C. and Y.H.; project administration, Q.C. and W.W.; funding acquisition, Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China under Grant (No. 42271352).

Data Availability Statement

The data used to support the findings of this study can be made available by the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Greenhalgh, S. Soda industry influence on obesity science and policy in China. J. Public Health Policy 2019, 40, 5–16. [Google Scholar] [CrossRef] [PubMed]
  2. Matthews, J. 2011 Food & Health Survey Consumer Attitudes toward Food Safety, Nutrition & Health; International Food Information Council Foundation: Washington, DC, USA, 2011; Volume 31. [Google Scholar]
  3. Subar, A.F.; Kirkpatrick, S.I.; Mittl, B.; Zimmerman, T.P.; Thompson, F.E.; Bingley, C.; Willis, G.; Islam, N.G.; Baranowski, T.; McNutt, S.; et al. The automated self-administered 24-hour dietary recall (ASA24): A resource for researchers, clinicians and educators from the National Cancer Institute. J. Acad. Nutr. Diet. 2012, 112, 1134. [Google Scholar] [CrossRef] [PubMed]
  4. Meyers, A.; Johnston, N.; Rathod, V.; Korattikara, A.; Gorban, A.; Silberman, N.; Guadarrama, S.; Papandreou, G.; Huang, J.; Murphy, K.P. Im2Calories: Towards an automated mobile vision food diary. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1233–1241. [Google Scholar] [CrossRef]
  5. Ege, T.; Yanai, K. Image-based food calorie estimation using knowledge on food categories, ingredients and cooking directions. In Proceedings of the on Thematic Workshops of ACM Multimedia 2017, Mountain View, CA, USA, 23–27 October 2017; pp. 367–375. [Google Scholar] [CrossRef]
  6. Lo, F.P.W.; Sun, Y.; Qiu, J.; Lo, B. Image-based food classification and volume estimation for dietary assessment: A review. IEEE J. Biomed. Health Inform. 2020, 24, 1926–1939. [Google Scholar] [CrossRef] [PubMed]
  7. Puri, M.; Zhu, Z.; Yu, Q.; Divakaran, A.; Sawhney, H. Recognition and volume estimation of food intake using a mobile device. In Proceedings of the 2009 Workshop on Applications of Computer Vision (WACV), Snowbird, UT, USA, 7–8 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1–8. [Google Scholar] [CrossRef]
  8. Dehais, J.; Anthimopoulos, M.; Shevchik, S.; Mougiakakou, S. Two-view 3D reconstruction for food volume estimation. IEEE Trans. Multimed. 2016, 19, 1090–1099. [Google Scholar] [CrossRef]
  9. Shao, Z.; Fang, S.; Mao, R.; He, J.; Wright, J.L.; Kerr, D.A.; Boushey, C.J.; Zhu, F. Towards learning food portion from monocular images with cross-domain feature adaptation. In Proceedings of the 2021 IEEE 23rd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, 6–8 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
  10. Thames, Q.; Karpur, A.; Norris, W.; Xia, F.; Panait, L.; Weyand, T.; Sim, J. Nutrition5k: Towards automatic nutritional understanding of generic food. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8903–8911. [Google Scholar] [CrossRef]
  11. Shao, W.; Hou, S.; Jia, W.; Zheng, Y. Rapid Non-Destructive Analysis of Food Nutrient Content Using Swin-Nutrition. Foods 2022, 11, 3429. [Google Scholar] [CrossRef] [PubMed]
  12. Fang, S.; Shao, Z.; Mao, R.; Fu, C.; Delp, E.J.; Zhu, F.; Kerr, D.A.; Boushey, C.J. Single-view food portion estimation: Learning image-to-energy mappings using generative adversarial networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 251–255. [Google Scholar] [CrossRef]
  13. Lu, Y.; Stathopoulou, T.; Vasiloglou, M.F.; Christodoulidis, S.; Stanga, Z.; Mougiakakou, S. An artificial intelligence-based system to assess nutrient intake for hospitalised patients. IEEE Trans. Multimed. 2020, 23, 1136–1147. [Google Scholar] [CrossRef]
  14. Vinod, G.; Shao, Z.; Zhu, F. Image Based Food Energy Estimation With Depth Domain Adaptation. In Proceedings of the 2022 IEEE 5th International Conference on Multimedia Information Processing and Retrieval (MIPR), Virtual, 2–4 August 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 262–267. [Google Scholar] [CrossRef]
  15. Shao, W.; Min, W.; Hou, S.; Luo, M.; Li, T.; Zheng, Y.; Jiang, S. Vision-based food nutrition estimation via RGB-D fusion network. Food Chem. 2023, 424, 136309. [Google Scholar] [CrossRef] [PubMed]
  16. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 821–830. [Google Scholar] [CrossRef]
  17. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef]
  18. Montville, J.B.; Ahuja, J.K.; Martin, C.L.; Heendeniya, K.Y.; Omolewa-Tomobi, G.; Steinfeldt, L.C.; Anand, J.; Adler, M.E.; LaComb, R.P.; Moshfegh, A. USDA food and nutrient database for dietary studies (FNDDS), 5.0. Procedia Food Sci. 2013, 2, 99–112. [Google Scholar] [CrossRef]
  19. Beijbom, O.; Joshi, N.; Morris, D.; Saponas, S.; Khullar, S. Menu-match: Restaurant-specific food logging from images. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 5–9 January 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 844–851. [Google Scholar] [CrossRef]
  20. Liang, Y.; Li, J. Computer vision-based food calorie estimation: Dataset, method, and experiment. arXiv 2017, arXiv:1705.07632. [Google Scholar]
  21. Fang, S.; Liu, C.; Zhu, F.; Delp, E.J.; Boushey, C.J. Single-view food portion estimation based on geometric models. In Proceedings of the 2015 IEEE International Symposium on Multimedia (ISM), Miami, FL, USA, 14–16 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 385–390. [Google Scholar] [CrossRef]
  22. Ando, Y.; Ege, T.; Cho, J.; Yanai, K. Depthcaloriecam: A mobile application for volume-based foodcalorie estimation using depth cameras. In Proceedings of the 5th International Workshop on Multimedia Assisted Dietary Management, Nice, France, 21 October 2019; pp. 76–81. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  24. Ranftl, R.; Bochkovskiy, A.; Koltun, V. Vision transformers for dense prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 12179–12188. [Google Scholar] [CrossRef]
  25. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  27. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  30. Lin, G.; Milan, A.; Shen, C.; Reid, I. Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1925–1934. [Google Scholar] [CrossRef]
  31. Kim, W.; Son, B.; Kim, I. Vilt: Vision-and-language transformer without convolution or region supervision. In Proceedings of the International Conference on Machine Learning. PMLR, Virtual, 18–24 July 2021; pp. 5583–5594. [Google Scholar]
  32. Bahador, N.; Ferreira, D.; Tamminen, S.; Kortelainen, J. Deep learning–based multimodal data fusion: Case study in food intake episodes detection using wearable sensors. JMIR mHealth uHealth 2021, 9, e21926. [Google Scholar] [CrossRef] [PubMed]
  33. Chennupati, S.; Sistu, G.; Yogamani, S.; A Rawashdeh, S. Multinet++: Multi-stream feature aggregation and geometric loss strategy for multi-task learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Min, W.; Wang, Z.; Liu, Y.; Luo, M.; Kang, L.; Wei, X.; Wei, X.; Jiang, S. Large scale visual food recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9932–9949. [Google Scholar] [CrossRef] [PubMed]
  36. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
  37. Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; Navab, N. Deeper depth prediction with fully convolutional residual networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 239–248. [Google Scholar] [CrossRef]
  38. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef] [PubMed]
  39. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  40. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  41. Aucott, L.; Poobalan, A.; Smith, W.C.S.; Avenell, A.; Jung, R.; Broom, J. Effects of weight loss in overweight/obese individuals and long-term hypertension outcomes: A systematic review. Hypertension 2005, 45, 1035–1041. [Google Scholar] [CrossRef] [PubMed]
  42. Zhu, M.; Miura, J.; Lu, L.X.; Bernier, M.; DeCabo, R.; Lane, M.A.; Roth, G.S.; Ingram, D.K. Circulating adiponectin levels increase in rats on caloric restriction: The potential for insulin sensitization. Exp. Gerontol. 2004, 39, 1049–1059. [Google Scholar] [CrossRef] [PubMed]
  43. Poli, V.F.S.; Sanches, R.B.; dos Santos Moraes, A.; Fidalgo, J.P.N.; Nascimento, M.A.; Bresciani, P.; Andrade-Silva, S.G.; Cipullo, M.A.T.; Clemente, J.C.; Caranti, D.A. The excessive caloric intake and micronutrient deficiencies related to obesity after a long-term interdisciplinary therapy. Nutrition 2017, 38, 113–119. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The example images from Nutrition5k dataset. (a) RGB images. (b) Depth maps. (c) Nutritional annotations.
Figure 1. The example images from Nutrition5k dataset. (a) RGB images. (b) Depth maps. (c) Nutritional annotations.
Foods 12 04293 g001
Figure 2. Incorrect image samples. (a) Food is not fully incorporated in the image. (b) Dishes are overlapping. (c) Non-food image.
Figure 2. Incorrect image samples. (a) Food is not fully incorporated in the image. (b) Dishes are overlapping. (c) Non-food image.
Foods 12 04293 g002
Figure 3. The overall framework of our DPF-Nutrition, which consists of depth prediction module and RGB-D fusion module. We adopt depth prediction transformer (DPT) to generate the predicted depth map. We design a cross-modal attention block (CAB) to extract and integrate the complementary features of RGB and depth images. ⨁ indicates element-wise addition, Ⓖ denotes global average pool.
Figure 3. The overall framework of our DPF-Nutrition, which consists of depth prediction module and RGB-D fusion module. We adopt depth prediction transformer (DPT) to generate the predicted depth map. We design a cross-modal attention block (CAB) to extract and integrate the complementary features of RGB and depth images. ⨁ indicates element-wise addition, Ⓖ denotes global average pool.
Foods 12 04293 g003
Figure 4. (a) The structure of depth prediction module. The input image is transformed into feature vectors by ResNet-50 feature extractor and consequently embedded into two-dimensional tokens. The tokens are then fed into transformer encoder. The tokens from different transformer stages are reassembled into image-like feature maps at various resolutions. Finally, the image-like feature maps are fused progressively to generate the depth prediction. (b) The structure of transformer encoder. ⨁ indicates element-wise addition.
Figure 4. (a) The structure of depth prediction module. The input image is transformed into feature vectors by ResNet-50 feature extractor and consequently embedded into two-dimensional tokens. The tokens are then fed into transformer encoder. The tokens from different transformer stages are reassembled into image-like feature maps at various resolutions. Finally, the image-like feature maps are fused progressively to generate the depth prediction. (b) The structure of transformer encoder. ⨁ indicates element-wise addition.
Foods 12 04293 g004
Figure 5. The structures of the RGB-D fusion paradigms. (a) Fusion–enhancement. (b) Enhancement–fusion. (c) Our proposed. ⨁ denotes element-wise addition, ⨂ indicates pixel-wise multiplication, Ⓒ represents cross-channel concatenation.
Figure 5. The structures of the RGB-D fusion paradigms. (a) Fusion–enhancement. (b) Enhancement–fusion. (c) Our proposed. ⨁ denotes element-wise addition, ⨂ indicates pixel-wise multiplication, Ⓒ represents cross-channel concatenation.
Foods 12 04293 g005
Figure 6. The structure of CAB. GAP indicates global average pooling, ⨁ denotes element-wise addition, ⨂ indicates pixel-wise multiplication, Ⓒ represents cross-channel concatenation, Mean represents mean function along the channel dimension.
Figure 6. The structure of CAB. GAP indicates global average pooling, ⨁ denotes element-wise addition, ⨂ indicates pixel-wise multiplication, Ⓒ represents cross-channel concatenation, Mean represents mean function along the channel dimension.
Foods 12 04293 g006
Figure 7. The sample results of the depth estimation. (a) RGB images. (b) Estimated depth maps. (c) Actual depth maps.
Figure 7. The sample results of the depth estimation. (a) RGB images. (b) Estimated depth maps. (c) Actual depth maps.
Foods 12 04293 g007
Figure 8. The visualization results. (a) The ROI heat-maps of different nutrients. (b) The nutrition facts.
Figure 8. The visualization results. (a) The ROI heat-maps of different nutrients. (b) The nutrition facts.
Foods 12 04293 g008
Table 1. Summary of nutrition datasets. Diversity represents the number of unique dishes and size represents the number of the images.
Table 1. Summary of nutrition datasets. Diversity represents the number of unique dishes and size represents the number of the images.
DatasetDiversitySizeAnnotationDepth
MenuMatch [19]41646CaloriesN
ECUSTFD [20]1602978Volume and MassN
Fang et al. [21]345CaloriesN
Ando et al. [22]3270CaloriesY
Nutrition5k [10]50663490Calories, Mass and MacronutrientsY
Table 2. Comparison of the performance of different backbones. The best results were highlighted in bold.
Table 2. Comparison of the performance of different backbones. The best results were highlighted in bold.
MethodsCalories
MAE (kCal)/PMAE (%)
Mass
MAE (g)/PMAE (%)
Fat
MAE (g)/PMAE (%)
Carb
MAE (g)/PMAE (%)
Protein
MAE (g)/PMAE (%)
Mean
PMAE (%)
ViT [29]52.4/20.432.5/16.33.78/29.45.72/28.95.72/28.924.6
VGG16 [28]47.9/18.629.2/14.63.61/28.05.18/26.24.71/26.822.8
InceptionV3 [36]46.2/18.028.4/14.23.24/25.15.1/25.84.37/24.821.6
ResNet50 [23]46.9/18.227.7/13.83.61/28.04.51/22.84.56/25.921.7
ResNet101 [23]46.0/17.927.2/13.63.42/26.54.56/23.04.31/24.521.1
Table 3. Comparison of the performance of different depth prediction methods.The best results were highlighted in bold.
Table 3. Comparison of the performance of different depth prediction methods.The best results were highlighted in bold.
Method δ > 1.25 δ > 1 . 25 2 δ > 1 . 25 3 AbsRelRMSE
UNet [26]0.5730.7710.8600.6610.211
FCRN [37]0.6340.7990.8770.5290.189
HRNet [38]0.6920.8640.9060.3870.162
DPT [24]0.7430.8930.9420.3220.128
Table 4. Comparison of the performance of nutrition estimation using depth maps from different sources.The best results were highlighted in bold.
Table 4. Comparison of the performance of nutrition estimation using depth maps from different sources.The best results were highlighted in bold.
Depth SourceCalories
MAE (kCal)/PMAE (%)
Mass
MAE (g)/PMAE (%)
Fat
MAE (g)/PMAE (%)
Carb
MAE (g)/PMAE (%)
Protein
MAE (g)/PMAE (%)
Mean
PMAE (%)
UNet [26]41.6/16.222.6/11.33.29/25.54.43/22.44.01/22.819.6
FRCN [37]40.7/15.822.7/11.33.17/24.64.28/21.63.87/22.019.1
HRNet [38]39.3/15.321.6/10.83.01/23.34.13/20.93.72/21.218.3
DPT [24]37.9/14.721.2/10.62.92/22.64.09/20.73.56/20.217.8
Depth sensor36.5/14.220.4/10.22.76/21.44.08/20.63.48/19.817.2
Table 5. Comparison of the performance of different cross-modal fusion methods.The best results were highlighted in bold.
Table 5. Comparison of the performance of different cross-modal fusion methods.The best results were highlighted in bold.
MethodCalories
MAE (kCal)/PMAE (%)
Mass
MAE (g)/PMAE (%)
Fat
MAE (g)/PMAE (%)
Carb
MAE (g)/PMAE (%)
Protein
MAE (g)/PMAE (%)
Mean
PMAE (%)
Enhancement–Fusion39.1/15.222.6/11.33.03/23.54.14/20.93.80/21.618.5
Fusion–Enhancement39.3/15.321.6/10.83.01/23.34.13/20.93.72/21.218.3
CAB37.9/14.721.2/10.62.92/22.64.09/20.73.56/20.217.8
Table 6. Comparison of the performance of different methods.The best results were highlighted in bold.
Table 6. Comparison of the performance of different methods.The best results were highlighted in bold.
InputMethodsCalories
MAE (kCal)/
PMAE (%)
Mass
MAE (g)/
PMAE (%)
Fat
MAE (g)/
PMAE (%)
Carb
MAE (g)/
PMAE (%)
Protein
MAE (g)/
PMAE (%)
Mean
PMAE (%)
RGB-D imagesGoogle-Nutrition-depth [10]47.6/18.840.7/18.92.27/18.14.6/23.83.7/20.920.1
RGB-D Nutrition [15]38.5/15.021.6/10.83.0/23.54.43/22.43.69/21.018.5
Monocular imagesGoogle-Nutrition-monocular [10]70.6/26.140.4/18.85.0/34.26.1/31.95.5/29.529.1
Swin-Nutrition [11]41.5/16.227.5/13.73.21/24.94.32/21.84.47/25.420.4
DPF-Nutrition (ours)37.9/14.721.2/10.62.92/22.64.09/20.73.56/20.217.8
Table 7. Comparison with different ablation settings.The best results were highlighted in bold.
Table 7. Comparison with different ablation settings.The best results were highlighted in bold.
IndexModelCalories
MAE (kCal)/
PMAE (%)
Mass
MAE (g)/
PMAE (%)
Fat
MAE (g)/
PMAE (%)
Carb
MAE (g)/
PMAE (%)
Protein
MAE (g)/
PMAE (%)
Mean
PMAE (%)
(a)RGB Stream46.0/17.927.2/13.63.42/26.54.56/23.04.31/24.521.1
(b)Depth Stream83.5/32.544.7/22.36.29/48.88.08/40.88.53/ 8.538.6
(c)(a) + (b) + direct fusion42.7/16.725.0/12.53.17/24.64.31/21.83.97/22.619.6
(d)(a) + (b) + multi-scale fusion40.7/15.824.8/12.43.01/23.34.13/20.93.88/22.118.9
(e)(d) + CAB37.9/14.721.2/10.62.92/22.64.09/20.73.56/20.217.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, Y.; Cheng, Q.; Wu, W.; Huang, Z. DPF-Nutrition: Food Nutrition Estimation via Depth Prediction and Fusion. Foods 2023, 12, 4293. https://doi.org/10.3390/foods12234293

AMA Style

Han Y, Cheng Q, Wu W, Huang Z. DPF-Nutrition: Food Nutrition Estimation via Depth Prediction and Fusion. Foods. 2023; 12(23):4293. https://doi.org/10.3390/foods12234293

Chicago/Turabian Style

Han, Yuzhe, Qimin Cheng, Wenjin Wu, and Ziyang Huang. 2023. "DPF-Nutrition: Food Nutrition Estimation via Depth Prediction and Fusion" Foods 12, no. 23: 4293. https://doi.org/10.3390/foods12234293

APA Style

Han, Y., Cheng, Q., Wu, W., & Huang, Z. (2023). DPF-Nutrition: Food Nutrition Estimation via Depth Prediction and Fusion. Foods, 12(23), 4293. https://doi.org/10.3390/foods12234293

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop