Next Article in Journal
A Pathway to Hallux Valgus Correction: Intra- and Interexaminer Reliability of Hallux Alignment
Next Article in Special Issue
Machine Vision-Based Chinese Walnut Shell–Kernel Recognition and Separation
Previous Article in Journal
Knowledge-Graph- and GCN-Based Domain Chinese Long Text Classification Method
Previous Article in Special Issue
ACapMed: Automatic Captioning for Medical Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bi-LS-AttM: A Bidirectional LSTM and Attention Mechanism Model for Improving Image Captioning

School of Information Science and Technology, Nantong University, Nantong 226019, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7916; https://doi.org/10.3390/app13137916
Submission received: 14 June 2023 / Revised: 28 June 2023 / Accepted: 3 July 2023 / Published: 6 July 2023
(This article belongs to the Special Issue Recent Trends in Automatic Image Captioning Systems)

Abstract

:
The discipline of automatic image captioning represents an integration of two pivotal branches of artificial intelligence, namely computer vision (CV) and natural language processing (NLP). The principal functionality of this technology lies in transmuting the extracted visual features into semantic information of a higher order. The bidirectional long short-term memory (Bi-LSTM) has garnered wide acceptance in executing image captioning tasks. Of late, scholarly attention has been focused on modifying suitable models for innovative and precise subtitle captions, although tuning the parameters of the model does not invariably yield optimal outcomes. Given this, the current research proposes a model that effectively employs the bidirectional LSTM and attention mechanism (Bi-LS-AttM) for image captioning endeavors. This model exploits the contextual comprehension from both anterior and posterior aspects of the input data, synergistically with the attention mechanism, thereby augmenting the precision of visual language interpretation. The distinctiveness of this research is embodied in its incorporation of Bi-LSTM and the attention mechanism to engender sentences that are both structurally innovative and accurately reflective of the image content. To enhance temporal efficiency and accuracy, this study substitutes convolutional neural networks (CNNs) with fast region-based convolutional networks (Fast RCNNs). Additionally, it refines the process of generation and evaluation of common space, thus fostering improved efficiency. Our model was tested for its performance on Flickr30k and MSCOCO datasets (80 object categories). Comparative analyses of performance metrics reveal that our model, leveraging the Bi-LS-AttM, surpasses unidirectional and Bi-LSTM models. When applied to caption generation and image-sentence retrieval tasks, our model manifests time economies of approximately 36.5% and 26.3% vis-a-vis the Bi-LSTM model and the deep Bi-LSTM model, respectively.

1. Introduction

Image captioning is a hot topic involving several fields such as computer vision (CV) and natural language processing (NLP), known as image semantic description or “talking about pictures” [1,2,3,4,5,6]. Image captioning technology not only needs to recognize the entity object information in the image and the relationship between objects but also needs to learn how to integrate them into the ability to describe reasonable sentence descriptions. Traditional methods use models based on visual space search, sentence template usage, and the most matching sentence in the dataset to accomplish the tasks of image captioning. The disadvantage of these methods is the low efficiency of generating real and accurate sentences and the poor ability to generate structurally novel sentences. In recent research [7,8,9,10,11], visual and language information has been embedded into a common space via recurrent neural networks (RNNs) initially. Convolutional neural networks (CNNs) were then embedded within the visual space and combined with long short-term memory (LSTM) to produce more effective results.
Most models extract image features by embedding the CNN into visual space. While this method can achieve good results, the extracted image features are not highly accurate and efficient, wasting a lot of time. Many models embed LSTM and Bi-LSTM into language space to generate sentences, but the results are not accurate enough. Therefore, it is challenging for subtitling models to perform novel subtitling tasks with accurate and efficient image-sentence retrieval.
To address these issues, we propose a model leveraging a bidirectional LSTM coupled with an attention mechanism (Bi-LS-AttM). This innovative model substitutes the region convolutional neural network (RCNN)—commonly used for feature extraction—with a more efficient fast region convolutional neural network (Fast RCNN). This adjustment enhances the identification and extraction of features within the image’s regions of interest (RoIs). The optimized model is then applied to refine the LSTM network’s performance. By juxtaposing forward and backward outcomes and incorporating the attention boost, the Bi-LS-AttM is able to predict word vectors with greater precision and generate more fitting image captions.
Why do we use the model? We employed the model to break through the boundaries of the traditional Bi-LSTM model, which is not focused enough on the comparison of historical and future word results. In the traditional LSTM cells, the prediction of the next word  x t  using the visible context  V  and historical context  x 1 : t - 1  is performed by estimating  log P ( x t | V , x 1 : t - 1 ) . However, in the Bi-LS-AttM, the prediction of the word  x t  depends on the forward and backward results of separately maximizing  log P ( x t | V , x 1 : t - 1 )  and  log P ( x t | V , x t + 1 : t )  at time  t . By combining the Bi-LSTM with the attention model, the model focuses increasingly on comparing historical and future word results and using their dependencies to predict and generate appropriate image captions. Figure 1 shows the example image of the Bi-LS-AttM model generating a sentence that supports our hypothesis that the Bi-LS-AttM model can generate more complementary and focused captions.
We tested the efficiency of our model on the datasets Flickr30K and MSCOCO and performed a qualitative analysis. The analysis showed that the method performs efficiently, and the proposed Bi-LS-AttM model outperforms other published models. The principal contributions of this paper are threefold:
  • We proposed a trainable model incorporating a bidirectional LSTM and attention mechanism. This model embeds image captions and scores into a region by capitalizing on the long-term forward and backward context.
  • We upgraded the feature extraction mechanism, replacing the conventional CNN and RCNN with a Fast RCNN. This improvement enhances the model’s ability to rapidly detect and extract features from items within an image’s regions of interest.
  • We verified the efficiency of the framework on two datasets Flickr30K and MSCOCO. The evaluation demonstrated that the Bi-LSTM and attention mechanism model achieved highly competitive performance results relative to current techniques in the tasks of generating captions and image sentence retrieval.

2. Related Works

Initially, researchers utilized computers to analyze identified content in image captioning, which was the original task for image recognition [12,13,14]. Later, they introduced additional requirements such as processing and determining object properties, identifying object relationships, and describing image content in natural language. Since then, numerous image captioning techniques have been introduced, broadly categorized into three groups: template-based, retrieval-based, and deep-learning-based methods.
Template-based methods, which utilize fixed templates for sentence generation, identify image elements such as objects, actions, and scenes based on visual dependency grammar. For instance, Farhadi [15] used a support vector machine (SVM) [16,17] to detect image items and pre-established templates for sentence descriptions. However, the limitations of datasets and template algorithms impeded their performance. Similarly, Li [18] employed Web-scale N-grams for phrase extraction linked to objects, actions, and relationships in 2011. Later, Kulkarni [19] used a conditional random field (CRF) [20,21] for data extraction from a large pool of visual descriptive text, thereby improving computer vision recognition and sentence generation. Despite these efforts, the performance of these methods was suboptimal due to the inherent restrictions of template-based approaches.
The retrieval-based method stores all image descriptions in a collection. The image to be described is then compared to the training set and filtered to find similar images. Using a similar image description to the one found, the candidate description is modified accordingly. Kuznetsova [22] proposed to search for images with attached titles on the Internet and obtain expressive phrases as tree fragments from the test images. Then, new descriptions are composed by filtering and merging the fragments of the extraction tree. Mason [23] proposed a nonparametric density estimation (NDE) technique that estimates the frequency of visual content words of the image to be detected and transforms caption generation into an extractive summarization problem. Sun [24] proposed a concept automatic recognition method that uses parallel text and visual corpora. It can filter out text terms by matching the visual characteristics of similar images in the image library and the image to be described. Retrieval-based methods can be more natural language-like, although relying heavily on the capacity of the database makes it difficult to generate sentences for specific images.
In recent years, with the continued advancement of deep learning, neural networks have been extensively used in image caption tasks. Kiros [25] first used deep neural networks and LSTM to construct two different multimodal neural network models in 2014, continuously integrating semantic information to generate words. For the encoding part, they applied an RNN to convert vocabulary into  D -dimensional word vectors. The sentence described can be written as matrix  V   x   D , where  V  is the number of words in a sentence, and  D  is the size of the word vector. Finally, they used a decoder consisting of LSTM cells to generate the final picture subtitle result word by word with the combination of image features and the language model sentence by sentence. In subsequent research, Xu et al. [26] incorporated attention mechanisms into the encoder and decoder structural models to describe images. By establishing an attention matrix, they can automatically focus on different areas when predicting different words at different times to enhance the description effectiveness of the model. Bo [27] used generative adversarial networks to generate diversified descriptions by controlling random noise vectors.
In contemporary research, Ayoub [28] deployed the Bahdanau attention mechanism and transfer learning techniques for image caption generation. They incorporated a pre-trained image feature extractor alongside the attention mechanism, thus improving captioning quality and precision. Muhammad [29] proposed a model blending the attention mechanism and object features for image captioning, enhancing the model’s ability to leverage extracted object features from images. Chun [30] demonstrated an advanced deep learning approach for image captioning that combined CNN for image feature extraction and RNN for caption generation, enhanced by the attention mechanism. This innovative method facilitated the automated creation of comprehensive bridge damage descriptions. Lastly, Wu [31] addressed the challenge of describing novel objects through a switchable attention mechanism and multimodal fusion approach, resulting in the generation of accurate and meaningful descriptions.
Wang [8] used a Bi-LSTM model to perform image caption. Wang has developed a deep Bi-LSTM model based on this and has achieved good results. Fazin [9] simplified Wang’s model by reducing many parameters and improving the efficiency of the model. Unlike the above models, the mapping relationship between vision and language in our Bi-LS-AttM model is reverse-crossed, and the forward and backward attention of visual language are dynamically learned. As shown in Section 4, this has been demonstrated to be extremely beneficial for picture caption and image sentence retrieval.

3. Methodology

This section outlines our proposed model, an enhanced version of the deep Bi-LSTM model for image captioning as proposed by Wang [8] and Fazin [9]. In our design, we replace the RCNN used in Wang and Vahid’s model with Fast RCNN to expedite the feature extraction process. Furthermore, we substitute the Bi-LSTM with the Bi-LS-AttM, representing our unique contribution to this study.
Our model framework comprises three components: a Fast RCNN for detecting objects within images; a Bi-LSTM paired with an attention mechanism to provide attentional representation for each word; and a common space to compile all sentences and their respective final scores. The specifics of each module will be elaborated in the subsequent sections.

3.1. Detect Object by Fast RCNN

In this section, we adopt the method proposed by Girshick [32] for feature extraction and recognition. The selective search algorithm is utilized to extract candidate regions from the input image. These regions are then mapped to the final convolutional feature layer based on their spatial positional relationship.
For each candidate region on the convolutional feature layer, RoI pooling is performed to secure fixed-dimensional features. These extracted features are then fed into the fully connected layer for subsequent classification and regression tasks.
Fast RCNN outputs the probability of each category for each candidate region, as well as the calculated position of each candidate box through regression. For each candidate region, the following loss function is calculated:
L p , u , t u , v = L c l s p , u 1 + λ u 1
where  p  is the probability of each category belonging to the candidate region and  u  is the ground truth category.  t  is the predicted position for each category, and  v  is the ground truth position for the candidate field. Compared with the previous version of RCNN, Fast RCNN improves the calculation speed, saving the time and cost of object detection.
Fast RCNN combines classification and regression into a common network, enabling consistent training. In particular, its main enhancement on RCNN is that it eliminates the practice of using separate SVM classifiers and bounding regressors, which greatly improves speed.

3.2. Long Short-Term Memory

The LSTM cells form the basis of this work. They are a unique form of RNN able to memorize long-term associations. Figure 2 shows that an LSTM cell is made up of four important components: a memory cell  g  and three gate circuits ( i  is the gate of the input,  f  is the gate of forget, and  o  is the gate of the output) [9].
In the formula below,  f t i t , and  o t  are the values of forget, input, and output at time  t , respectively.  a t  is the intermediate feature extract result of  h t - 1  and  x t  at time  t :
a t = tanh W a h t - 1 + U a x t + b a
i t = σ W i h t - 1 + U i x t + b i
f t = σ W f h t - 1 + U f x t + b f
o t = σ W o h t - 1 + U o x t + b o
where  x t  is the entrance, and  h t - 1  is the hidden state value at time  t - 1 . The results calculated by forgetting and inputting operate on the cell state, expressed as the formula below:
c t = c t - 1 f t + i t a t
where   represents the Hadamard product. Finally, the hidden state is at  t h t  is obtained by multiplying the gate output  o t  and the current cell state  c t  using the Hadamard product:
h t = o t t a n h c t
The following equation uses the parameters  W s  and  b S  to predict the next word:
F p t i ; W s , b S = exp W s h t i + b S j = 1 K exp W s h t j + b S
where  p t i  is the tipping probability of the forecast value.

3.3. Bi-LSTM

Both RNN and LSTM units leverage past temporal information to predict forthcoming outputs. However, in some instances, the desired output is associated with not just the previous state but also the future state. For instance, predicting a missing word within a textual context requires comprehension of both the preceding and succeeding context. This dual-directional context analysis provides a more comprehensive and accurate interpretation, thereby achieving a genuine contextual understanding and decision-making process.
In the traditional LSTM, the forecast of the word  x t  using the optical context  V  and historical context  x 1 : t - 1  is performed by estimating  log P ( x t | V , x 1 : t - 1 ) . However, in the Bi-LSTM with attention, the prediction of the word  x t  depends on the forward and backward results of separately maximizing  log P ( x t | V , x 1 : t - 1 )  and  log P ( x t | V , x t + 1 : t )  at time  t .
In the Bi-LSTM cell structure, the input sequence is processed in both forward and backward directions by two distinct LSTM cells to extract features. As illustrated in Figure 3, the output vectors generated are amalgamated to form the final word representation. The core concept behind the Bi-LSTM cell is to facilitate the capture of features at any given time point, encompassing information from both preceding and succeeding time steps. It is worth noting that the two LSTM units within the Bi-LSTM cell operate with independent parameters while sharing a common word-embedding vector space.

3.4. Architecture Model

The general layout of the model is illustrated in Figure 4. It is mainly composed of three modules: Fast RCNN for encoding image input, Bi-LS-AttM for encoding sentence input, and embedding picture and caption into common space and decoding it into image captions and evaluation scores.
The Bi-LS-AttM generates word vectors by comparing similarity using the context information from the frontend and the backend. More accurate words are selected after passing by attention. In our work, the model calculates the front hidden vector  h  and the back hidden vector  h . The front cell starts at  t = 1 , while the back cell starts at  t = T . Our framework works such that for an initial input frame I, the encoding is performed as follows:
I t = F I ; θ m
h t 1 = B M ; θ n
h t 1 = B M ; θ n
where  F  and  B  represent Fast RCNN and Bi-LSTM, respectively.  θ m  and  θ n  are their corresponding weight coefficients.  M  and  M  are forward and backward vectors learned from the neural network, respectively. Afterward, the obtained vectors  h  and  h  are input into attention. The bilinear scoring procedure is applied to calculate the correlation between the query  q  and  h  and  h . Next, a SoftMax is applied to these scores to normalize them and obtain the attention distribution  a = [ a 1 , a 2 , , a t ] . The bilinear scoring function and SoftMax are defined as follows:
s h t 1 , h t 1 , q = h t 1 T h t 1 T W q
a t = exp s h t 1 , h t 1 , q j = 1 n exp s h j 1 , h j 1 , q
where  W  is a trainable parameter matrix.  s  is a bilinear function.
After training the model, it can predict the word  x t  with a given image context  V  and forward word context  x 1 : t - 1 , predicted either in a forward direction using  P x t x 1 : t - 1 , V  or in a backward direction using  P x t x t + 1 : t , V . For both forward and backward directions,  x 1 = x T = 0  is set at the starting point. Finally, for sentences generated from both directions, the last sentence of the given image  P x 1 : T V  is determined by the sum of all words’ probabilities in the caption:
p x 1 : T V = m a x t = 1 T p x t V , t = 1 T p x t V
p x t V = t = 1 T p x t x 1 , x 2 , , x t - 1 , V
p x t V = t = 1 T p x t x t + 1 , x t + 2 , , x T , V
The Bi-LSTM module and its training parameters are similar to those presented in Wang [8]. The difference is that an attention mechanism is added to it. It can focus more on comparing the forward and backward context information to obtain the attention distribution. When extracting features, the Fast RCNN is more efficient and saves time.

4. Experiments

4.1. Dataset

Experiments were performed to validate the effectiveness, generality, and robustness of the model compared to other methods on two datasets, Flickr30K [33] and MSCOCO [34]:
Flickr30K. This is an extended version of Flickr8K. The dataset can be accessed via the following link: http://shannon.cs.illinois.edu/DenotationGraph (accessed 2 May 2023). It contains 31,783 images, each with 5 captions. The dataset does not explicitly categorize the images into different types or categories. We followed the dataset partitioning proposed by Karpathy [4]. In this split of the dataset, 29,000/1000/1000 pictures were utilized for training, validation, and testing, respectively.
MSCOCO. The dataset can be accessed via the following link: https://cocodataset.org (accessed 2 May 2023). This dataset, published several years ago, includes 82,783 training, 40,504 validation, and 40,775 test images. The dataset contains 80 different object categories. Five sentences are annotated for each frame. The focus is on describing all important parts of the scene rather than unimportant details. In the absence of a standard classification, we follow the classification of Wang et al. [8], which uses 80,000 images to train and 5000 to validate and test.

4.2. Evaluation Metrics

The evaluation methods of machine translation can be referred to as the evaluation criteria, which match the generated sentences with human descriptions to obtain a similarity score to measure the accuracy of the task.
For caption generation, the previous work is continued, and the BLEU-N score [35] is used:
U ( d , s ) = b ( d , s ) exp z = 1 Z k log p ( d , s )
where  d  represents the candidate description, which is the reference description,  b  is the penalty function,  k  represents the probability of selecting a specific caption, and  p  represents the accuracy measurement function. Comparing the results of METEOR [36] and CIDEr [37], METEOR can overcome the inherent deficiencies of the BLUE standard, while CIDEr computes the closeness of reference and modeled descriptions as the evaluation standard.
In the retrieval of an image-sentence, we use R@K and Medr as assessment scores. R@K is the recall rate of the top captions. Medr is the average score of the first basic fact image and caption retrieved.

4.3. Implementation Details

During the image encoding process, we utilize the VggNet model [38] for pre-training and employ Fast RCNN to obtain the features from the final fully connected layer. This allows Fast RCNN to share features and parameters in the feature extraction and RoI pooling stages, thereby enhancing processing efficiency. The Bi-LS-AttM is deployed for training the language module. In addition, we selected the widely used and enhanced VggNet [38] and GoogleNet [39] models for our experiments. We tested our model on two datasets specifically designed for image captioning: Flickr30k and MSCOCO.
The server hardware configuration was as shown below: Intel(R) Core (TM) i5-6200U 2.30 GHz CPU, NVIDIA RTX3080Ti GPU, and Win10 OS. The respective version levels needed for the software are Python 3.9, Torch 1.13.1, Scipy 1.2.1, H5py, and Tqdm. The parameters set for models are shown in Table 1, Table 2 and Table 3.
All words in the caption are taken from the vector used to generate the caption. Words appearing less than five times in the training set are marked and removed. A vocabulary of 7200 and 8600 words is provided for the Flickr30K and MSCOCO datasets, respectively. Additionally, 048 Bi-LSTM hidden units are used, and the initialization range of the weight coefficients is set to [−0.06, 0.06].

4.4. Experimental Results on the Generated Image Caption

Our image captioning model’s efficacy was evaluated through comparative experiments utilizing the BLUE-N metric, with the resultant data exhibited in Table 4. The additional attention layer implemented within our model contributed significantly to its strong performance on both evaluated datasets. Substituting AlexNet with VggNet [40] resulted in substantial performance improvements across all BLUE metrics. Our model ranks predominantly within the top two positions across these metrics. While our model lags marginally behind the top-rated Hard-attention [24] model in the B-1 metric, it surpasses the performance of both the Bi-LSTM and deep Bi-LSTM models in all other assessed metrics.
Figure 5 illustrates the comparison of our model with others on the METEOR and CIDEr metrics. We compared our model with the Bi-LSTM [9] and deep Bi-LSTM [8] models. As shown in the figure, the Bi-LS-AttM outperformed the leading-edge methods on the metrics. In the Flickr30K dataset, we improved the METEOR and CIDEr scores by about 8.0% and 12.5%, respectively. In the MSCOCO dataset, our model improved the METEOR and CIDEr scores by about 6.8% and 15.6%, respectively. We can also speculate that it can give better results on larger datasets.
We performed a comparative evaluation of our model against other advanced models on METEOR and CIDEr metrics, as depicted in Table 5. This analysis reveals that our Bi-LS-AttM model demonstrates robust competitiveness within the MSCOCO Karpathy split dataset. Particularly, on the METEOR score, our model trails slightly behind Wu’s model, potentially due to their superior parameter optimization techniques and the dataset’s compatibility with their switchable novel object captioner. This calls for further investigations into novel datasets. However, our model excels in the CIDEr score. A considerable score variation is evident when comparing the performance of our model with and without the attention mechanism.
In a similar vein, we conducted a thorough comparison between the performance scores of the baseline model, which lacks an attention mechanism, and our proposed model, employing diverse evaluation metrics. The results presented in Table 6 clearly depict the relative performance of the baseline model in contrast to our model. Remarkably, our model consistently surpasses the baseline model in terms of performance metrics on both the MSCOCO and Flickr30k test sets. Furthermore, the line graph depicted in Figure 6 visually demonstrates the competitive advantage of our model across various evaluation metrics. It is noteworthy that our model exhibits a more pronounced competitive edge, particularly on the MSCOCO test set. We attribute this observation to the larger scale of the MSCOCO test set, enabling a more comprehensive evaluation and, subsequently, yielding higher performance scores.

4.5. Experimental Results on the Retrieval of Image-Sentence

In assessing image-sentence retrieval, we primarily focused on retrieval scores. Table 7 presents the R@K and Medr scores from our model’s image-sentence retrieval across various datasets. Generally, our model surpasses previous methodologies on most metrics, with a particularly strong performance on the MSCOCO dataset. Notably, the Bi-LS-AttM outstrips advanced models in both image-to-sentence and sentence-to-image retrieval tasks. However, some metrics reveal suboptimal performance; for instance, the Mind’s Eye model [42], which effectively integrates image and text features, outperforms our model on the Flickr30K dataset. We posit that incorporating an adaptive attention mechanism could enhance efficiency in image-sentence retrieval tasks, a hypothesis we aim to investigate in future research.
We conducted additional experiments to authenticate our model’s performance in image-sentence retrieval tasks. Figure 7 presents examples from several retrieval experiments on the MSCOCO validation set. In each caption query, the model retrieves visually congruent images and captions, illustrating its proficiency in discerning the visual–textual association in image caption rankings. The upper dashed line represents image retrieval predicated on keywords, while the lower one symbolizes sentence retrieval based on images.
In the sentence-image retrieval task, we generated three images that mirror the keywords and sentences and then selected the matching image based on its similarity. For the image-sentence retrieval task, we produced three appropriate captions grounded on the images and then chose the generated sentences based on their high scores in the shared space. The given examples underscore the efficiency of our model in executing image-sentence retrieval tasks.

4.6. Discussion

Effect of Bi-LS-AttM: To gauge the influence of the Bi-LS-AttM, we compared the computational time of the Bi-LS-AttM model with the Bi-LSTM and deep Bi-LSTM models for caption creation and image-to-sentence retrieval tasks. Table 8 delineates the computational durations of these models for the respective tasks. We randomly selected 20 images from the Flickr30K validation set and assessed each model ten times for caption creation and image-to-sentence retrieval. The table provides the average time duration across the ten trials, excluding model initialization and training time.
The caption generation time expenses encompass the extraction of image features, bidirectional caption content sampling, computation of the final caption result, and caption accuracy evaluation. Conversely, the retrieval time accounts for the computation of the image-to-sentence retrieval score, image and sentence query, and sorting operations in descending order. By employing the Fast R-CNN framework and fine-tuning the relevant parameters, our model demonstrates significant time savings in accomplishing the given task. From Table 8, we can see that our model saves about 36.5% and 26.3% of the time compared to the Bi-LSTM and deep Bi-LSTM models, respectively, in solving image captioning and image-sentence retrieval tasks. We have verified the efficiency of the Bi-LS-AttM.
Effect of image caption: We used the Bi-LS-AttM model to generate real, accurate, and novel image descriptions. Figure 8 shows the comparison of the baseline model and our model in generating captions on the datasets. We evaluated generated captions from various perspectives. In some descriptions, the relationships between objects are well expressed (e.g., “A hot dog and a red bottle of drink are on the table”). In the example above, the objects ‘hot dog’ and ‘table’ are accurately identified, and the relationship between them is established. Finally, the image is described accurately and in a novel way (e.g., “A boy dressed in black surfs the sea with a red surfboard.”). However, the baseline model solely provides descriptive accounts of the images, lacking the generation of novel and expressive sentences to depict them. From the perspective of object detection, the object recognized in the baseline model is “bread” rather than the more accurate “hot dog”. The results of the studies show our model has good efficiency. Our model can achieve a balance between performance and efficiency.
Examples of failed experiments: Figure 9 depicts a notable number of anomalies arising from our experimental approach. It is pertinent to note that these inaccuracies primarily originate from the Flickr30K validation set, which we hypothesize may be due to the limited range and diversity present in the training dataset of this source. For instance, in the preceding images, our model exhibits imprecision in object identification (i.e., identifying “white clothes” when the man is not clothed). Another example demonstrates an illogical caption suggesting a man cycling on water. We surmise that these limitations could be ameliorated through improvements in our visual feature extraction aspect.
Despite these anomalies, we see them as potential avenues for further research rather than setbacks. Nonetheless, it is crucial to emphasize that a substantial number of remaining cases were accurately represented, as exemplified in Figure 8.

5. Conclusions

In this study, we have introduced a model that leverages the capabilities of the Bi-LS-AttM approach to generate captions that are precise, inventive, and context-sensitive. This was accomplished by incorporating bidirectional information and an attention mechanism. For the dual purposes of feature extraction and time optimization, we utilized the Fast RCNN. Additionally, to provide a comprehensive understanding of the proposed model’s structure, we generated a detailed visualization outlining the word generation process at consecutive timesteps. The model’s robustness and stability were thoroughly assessed across various datasets pertinent to image captioning and image-sentence retrieval tasks.
In terms of future work, we intend to delve into more intricate domains of image captioning, including those related to remote sensing and medical imaging. We anticipate broadening the application scope of our model to encapsulate other forms of captioning tasks such as video captioning. Furthermore, we plan to explore the integration of multi-task learning methodologies with an aim to enhance the model’s general applicability.

Author Contributions

Conceptualization, T.X. and J.W.; methodology, J.W. and T.X.; software, T.X.; validation, T.X., W.D. and J.W.; formal analysis, J.W.; resources, J.W.; data curation, T.X. and X.W.; writing—original draft preparation, T.X.; writing—review and editing, T.X.; visualization, W.D. and J.W.; supervision, J.Z. and J.W.; project administration, T.X; funding acquisition, W.D. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61976120, and the Basic Science Research Project of Nantong, grant number JC2020143.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code and required datasets of the experiments can be obtained upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The abbreviations used in this manuscript are listed in the following tables:
CVComputer Vision
NPLNatural Language Processing
Bi-LSTMBidirectional Long Short-Term Memory
LSTMLong Short-Term Memory
Bi-LS-AttMBidirectional LSTM and Attention Mechanism
Fast RCNNFast Region-based Convolutional Network
CNNConvolutional Neural Network
RNNRecurrent Neural Network
RoIRegion of Interest
SVMSupport Vector Machine
CRFConditional Random Field
NDENonparametric Density Estimation
VggNetVisual geometry group Net
Deep VSDeep Visual Semantic
m-RNNMultimodal Recurrent Neural Network
BULEBilingual Evaluation Understudy
MSCOCOMicrosoft Common Objects in Context
METEORMetric for Evaluation of Translation with Explicit Ordering
CIDErConsensus-based Image Description Evaluation
R@KRecall@K

References

  1. Lu, J.; Xiong, C.; Parikh, D.; Socher, R. Knowing when to look: Adaptive attention via a visual sentinel for image captioning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3242–3250. [Google Scholar]
  2. Lu, J.; Yang, J.; Batra, D.; Parikh, D. Neural Baby Talk. In Proceedings of the Name of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6566–7296. [Google Scholar]
  3. Ren, Z.; Wang, X.; Zhang, N.; Lv, X.; Li, L.-J. Deep reinforcement learning-based captioning with embedding reward. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1151–1159. [Google Scholar]
  4. Hossain, M.Z.; Sohel, F.; Shiratuddin, M.F.; Laga, H.A. Comprehensive Survey of Deep Learning for Image Captioning. ACM Comput. Surv. 2019, 51, 118. [Google Scholar] [CrossRef] [Green Version]
  5. Yu, N.; Hu, X.; Song, B.; Yang, J.; Zhang, J. Topic-Oriented Image Captioning Based on Order-Embedding. IEEE Trans. Image Process. 2019, 28, 2743–2754. [Google Scholar] [CrossRef]
  6. Jiang, W.; Zhu, M.; Fang, Y.; Shi, G.; Zhao, X.; Liu, Y. Visual Cluster Grounding for Image Captioning. IEEE Trans. Image Process. 2022, 31, 3920–3934. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, C.; Yang, H.; Bartz, C.; Meinel, C. Image captioning with deep bidirectional LSTMs. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 988–997. [Google Scholar]
  9. Vahid, C.; Fadaeieslam, J.; Yaghmaee, F. Improvement of image description using bidirectional LSTM. Int. J. Multimed. Inf. Retr. 2018, 7, 147–155. [Google Scholar]
  10. Ahmed, S.; Saif, A.; Hanif, M.; Shakil, M.; Jaman, M.; Haque, M.; Shawkat, S.; Hasan, J.; Sonok, B.; Rahman, F. Att-BiL-SL: Attention-Based Bi-LSTM and Sequential LSTM for Describing Video in the Textual Formation. Appl. Sci. 2022, 12, 317. [Google Scholar] [CrossRef]
  11. Cho, S.; Oh, H. Generalized Image Captioning for Multilingual Support. Appl. Sci. 2023, 13, 2446. [Google Scholar] [CrossRef]
  12. Guo, L.; Liu, J.; Lu, S.; Lu, H. Show, Tell, and Polish: Ruminant Decoding for Image Captioning. IEEE Trans. Multimed. 2020, 22, 2149–2162. [Google Scholar] [CrossRef]
  13. Zhang, L.; Zhang, Y.; Zhao, X.; Zou, Z. Image captioning via proximal policy optimization. Image Vis. Comput. 2021, 108, 104126. [Google Scholar] [CrossRef]
  14. Wang, J.; Xu, W.; Wang, Q.; Chan, A. On Distinctive Image Captioning via Comparing and Reweighting. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 2088–2103. [Google Scholar] [CrossRef]
  15. Farhadi, A.; Hejrati, S.; Sadeghi, M.; Young, P.; Forsyth, D. Every picture tells a story: Generating sentences from images. In Proceedings of the 11th European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 15–29. [Google Scholar]
  16. Chen, C.; Lin, C.; Schlkopf, B. A tutorial on ν-support vector machines. Appl. Stoch. Model. Bus. Ind. 2005, 21, 111–136. [Google Scholar] [CrossRef]
  17. Chang, C.; Lin, C. Libsvm: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  18. Li, S.; Kulkarni, G.; Berg, T.; Berg, A.; Choi, Y. Composing simple image descriptions using web-scale n-grams. In Proceedings of the Fifteenth Conference on Computational Natural Language Learning, Portland Oregon, OR, USA, 23–24 June 2011; pp. 220–228. [Google Scholar]
  19. Kulkarni, G.; Premraj, V.; Dhar, S.; Li, S.; Berg, T. Baby talk: Understanding and generating simple image descriptions. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2891–2903. [Google Scholar] [CrossRef] [Green Version]
  20. Yuan, Q.; Szummer, M.; Minka, T. Bayesian conditional random fields. In Proceedings of the 10th International Conference on Artificial Intelligence and Statistics, Bridgetown, Barbados, 6–8 January 2005. [Google Scholar]
  21. Sutton, C.; Rohanimanesh, K.; McCallum, A. Dynamic conditional random fields: Factorized probabilistic models for labeling and segmenting sequence data. J. Mach. Learn. Res. 2007, 8, 693–723. [Google Scholar]
  22. Kuznetsova, P.; Ordonez, V.; Berg, T.; Choi, Y. Treetalk: Composition and compression of trees for image descriptions. Trans. Assoc. Comput. Linguist. 2014, 2, 351–362. [Google Scholar] [CrossRef]
  23. Mason, R.; Charniak, E. Nonparametric Method for Data-driven Image Captioning. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MA, USA, 23–24 June 2014; pp. 592–598. [Google Scholar]
  24. Sun, C.; Gan, C.; Nevatia, R. Automatic concept discovery from parallel text and visual corpora. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2596–2604. [Google Scholar]
  25. Kiros, R.; Salakhutdinov, R.; Zemel, R. Multimodal Neural Language Models. In Proceedings of the 31st International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 595–603. [Google Scholar]
  26. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhutdinov, R.; Zemel, R.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 2048–2057. [Google Scholar]
  27. Bo, D.; Fidler, S.; Urtasun, R.; Lin, D. Towards diverse and natural image descriptions via a conditional gan. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2989–2998. [Google Scholar]
  28. Ayoub, S.; Gulzar, Y.; Reegu, F.A.; Turaev, S. Generating Image Captions Using Bahdanau Attention Mechanism and Transfer Learning. Symmetry 2022, 14, 2681. [Google Scholar] [CrossRef]
  29. Muhammad, A.; Jafar, A.; Ghneim, N. Image captioning model using attention and object features to mimic human image understanding. J. Big Data 2022, 9, 1–16. [Google Scholar]
  30. Chun, J.; Yamane, T.; Maemura, Y. A deep learning-based image captioning method to automatically generate comprehensive explanations of bridge damage. Comput.-Aided Civ. Infrastruct. Eng. 2022, 37, 1387–1401. [Google Scholar] [CrossRef]
  31. Wu, Y.; Jiang, L.; Yang, Y. Switchable Novel Object Captioner. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1162–1173. [Google Scholar] [CrossRef]
  32. Girshick, R. Fast r-cnn. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  33. Plummer, B.; Wang, L.; Cervantes, C.; Caicedo, J.; Lazebnik, S. Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  34. Vinyals, O.; Toshev, A.; Bengio, S.; Erhan, D. Show and Tell: Lessons Learned from the 2015 MSCOCO Image Captioning Challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 652–663. [Google Scholar] [CrossRef] [Green Version]
  35. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, Philadelphia, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar]
  36. Denkowski, M.; Lavie, A. Meteor universal: Language specific translation evaluation for any target language. In Proceedings of the 9th Workshop on statistical machine translation (WMT 2014), Baltimore, MD, USA, 26–27 June 2014; pp. 376–380. [Google Scholar]
  37. Vedantam, R.; Zitnick, C.; Parikh, D. CIDEr: Consensus-based Image Description Evaluation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 4566–4575. [Google Scholar]
  38. Ke, H.; Chen, D.; Li, X.; Tang, Y.; Shah, T.; Ranjan, R. Towards brain big data classification: Epileptic eeg identification with a lightweight vggnet on global mic. IEEE Access 2018, 6, 14722–14733. [Google Scholar] [CrossRef]
  39. Muthiah, M.; Logashamugam, E.; Nandhitha, N. Performance evaluation of googlenet, squeezenet, and resnet50 in the classification of herbal images. Int. J. Eng. Trends Technol. 2021, 69, 229–232. [Google Scholar]
  40. Mao, J.; Xu, W.; Yang, Y.; Wang, J.; Huang, Z.; Yuille, A. Deep captioning with multimodal recurrent neural networks (m-rnn). arXiv 2014, arXiv:1412.6632. [Google Scholar]
  41. Karpathy, A.; Fei-Fei, L. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3128–3137. [Google Scholar]
  42. Chen, X.; Zitnick, C. Mind’s eye: A recurrent visual representation for image caption generation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2422–2431. [Google Scholar]
Figure 1. Example captions generated by the model. (a) Caption generation (by the unidirectional model (upper) and by our model (lower)) on Flickr30K. (b) Caption generation (by the unidirectional model (upper) and by our model (lower)) on MSCOCO.
Figure 1. Example captions generated by the model. (a) Caption generation (by the unidirectional model (upper) and by our model (lower)) on Flickr30K. (b) Caption generation (by the unidirectional model (upper) and by our model (lower)) on MSCOCO.
Applsci 13 07916 g001
Figure 2. LSTM cell structure.
Figure 2. LSTM cell structure.
Applsci 13 07916 g002
Figure 3. Bi-LSTM cell structure.
Figure 3. Bi-LSTM cell structure.
Applsci 13 07916 g003
Figure 4. Proposed model architecture.
Figure 4. Proposed model architecture.
Applsci 13 07916 g004
Figure 5. (a) Comparison of METEOR scores of three models on two benchmark datasets; (b) comparison of CIDEr scores of three models on two benchmark datasets.
Figure 5. (a) Comparison of METEOR scores of three models on two benchmark datasets; (b) comparison of CIDEr scores of three models on two benchmark datasets.
Applsci 13 07916 g005
Figure 6. Figure 6. (a) Comparison of Captioning Models on Flickr30K; (b) Comparison of Captioning Models on MSCOCO.
Figure 6. Figure 6. (a) Comparison of Captioning Models on Flickr30K; (b) Comparison of Captioning Models on MSCOCO.
Applsci 13 07916 g006
Figure 7. Example of using our model for image retrieval and caption retrieval on the MSCOCO validation set. (a) To search for three images using captions. (b) To search for three captions using images.
Figure 7. Example of using our model for image retrieval and caption retrieval on the MSCOCO validation set. (a) To search for three images using captions. (b) To search for three captions using images.
Applsci 13 07916 g007
Figure 8. Examples of image captioning for the baseline and our model on the datasets. The captions generated by the baseline model are above, while the captions generated by our model are below.
Figure 8. Examples of image captioning for the baseline and our model on the datasets. The captions generated by the baseline model are above, while the captions generated by our model are below.
Applsci 13 07916 g008
Figure 9. Examples of failed experiments: (a) feature extraction error, (b) Image representation error, (c) caption logic error. We mark the extracted features on the image with red boxes and use blue fonts to distinguish errors.
Figure 9. Examples of failed experiments: (a) feature extraction error, (b) Image representation error, (c) caption logic error. We mark the extracted features on the image with red boxes and use blue fonts to distinguish errors.
Applsci 13 07916 g009
Table 1. Model parameters.
Table 1. Model parameters.
ParametersDescriptionsValue
Emb-dimDimension of word embeddings.256
Attention-dimDimension of Bi-LSTM attention linear layers.256
Frcnn-dimDimension of fast RCNN.128
DropoutThe phenomenon of learning to adapt can be considerably reduced by training batches.0.5
Cudnn-benchmarkSet to true only if inputs to model are fixed size; otherwise, lots of computational overhead.True
Table 2. Bi-LSTM attention parameters.
Table 2. Bi-LSTM attention parameters.
ParametersDescriptionsValue
Lay-NumNumber of layers3
Time-foreTime of data used to make the forecast.5
Hidden-sizeThe size of the featured area in the hidden status.8
EpochDisplaying the number of forward and backward calculations that could be performed.120
Batch-sizeThe amount of data transferred to the trainer.0.5
Learning-rateAdjustment of the network weighting rate by the loss gradient.0.002
OptimizerTraining and parameter tuning to reduce the loss of function.Adam
Table 3. Training parameters.
Table 3. Training parameters.
ParametersDescriptionsValue
Lay-NumNumber of layers.3
Hidden-dimSize of the featured area in the hidden status.8
EpochDisplaying the number of forward and backward calculations that could be performed.120
Batch-sizeThe amount of data transferred to the trainer.64
Learning-rateAdjustment of the network weighting rate by the loss gradient.0.004
OptimizerTraining and parameter tuning to reduce the loss function.Adam
Grad-clipClipping to threshold, and gradient to update weights.5
Alpha-cRegularization parameter for attention mechanisms.1
Table 4. Compare the BLEU score of each model on Flickr30K and MSCOCO.
Table 4. Compare the BLEU score of each model on Flickr30K and MSCOCO.
ModelFlickr30KMSCOCO
B-1B-2B-3B-4B-1B-2B-3B-4
Deep VS [41]57.336.924.015.762.545.032.123.0
m-RNN(AlexNet) [40]54.036.023.015.0----
m-RNN(VGGNet) [40]60.041.028.019.067.049.035.025.0
Hard-attention [26]66.943.929.619.971.850.435.725.0
Bi-LSTM [9]62.142.628.119.367.249.235.224.4
Deep Bi-LSTM [8]63.144.229.720.067.549.035.524.9
Our model (Bi-LSTM
and attention mechanism)
64.544.629.820.268.851.035.925.2
The results shown in bold type are the best values.
Table 5. Comparison of the METEOR and CIDEr scores between our and the state-of-the-art models.
Table 5. Comparison of the METEOR and CIDEr scores between our and the state-of-the-art models.
ModelMETEORCIDEr
Muhammad’s model [29]16.339.0
Wu’s model (SNOC) [31]21.939.5
Our model (Bi-LSTM)17.834.5
Our model (Bi-LS-AttM)21.541.2
The results shown in bold type are the best values.
Table 6. Performance scores of the baseline model and our model across various metrics.
Table 6. Performance scores of the baseline model and our model across various metrics.
ModelFlickr30K MSCOCO
B-1B-2B-3B-4METEORB-1B-2B-3B-4METEOR
Baseline models (without attention)56.238.923.014.814.759.645.828.419.715.9
Our model (Bi-LS-AttM)64.544.629.820.218.668.851.035.925.221.5
The results shown in bold type are the best values.
Table 7. R@K (a high score is good) and Medr (a low score is good) comparison of each model on Flickr30K and MSCOCO.
Table 7. R@K (a high score is good) and Medr (a low score is good) comparison of each model on Flickr30K and MSCOCO.
DatasetModelImage to SentenceSentence to Image
R@1R@2R@3MedrR@1R@2R@3Medr
Flickr30kDeep VS [41]22.248.261.44.815.237.750.59.2
m-RNN(AlexNet) [40]18.440.250.910.012.631.241.516.0
Mind’s Eye [42]18.545.758.17.016.642.558.98.0
Bi-LSTM [9]28.153.164.24.019.643.855.87.0
Deep Bi-LSTM [8]29.254.064.93.820.844.556.76.7
Our model (Bi-LSTM
and attention mechanism)
29.553.865.03.621.044.757.36.5
MSCOCODeep VS [41]16.539.252.09.010.729.642.214.0
m-RNN(AlexNet) [40]12.429.348.615.09.525.438.218.0
Mind’s Eye [42]12.835.650.111.011.633.748.510.0
Bi-LSTM [9]13.433.144.713.09.426.537.719.0
Deep Bi-LSTM [8]16.639.452.49.011.630.943.413.0
Our model (Bi-LSTM
and attention mechanism)
17.240.053.68.012.235.649.811.0
The results shown in bold type are the best values.
Table 8. The cost of checking 10 images on Flickr30K.
Table 8. The cost of checking 10 images on Flickr30K.
TaskBi-LSTMDeep Bi-LSTMOur Models
Caption creation1.04s1.07 s0.67 s
Image-Sentence Retrieval5.78 s5.81 s4.28 s
The results shown in bold type are the best values.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, T.; Ding, W.; Zhang, J.; Wan, X.; Wang, J. Bi-LS-AttM: A Bidirectional LSTM and Attention Mechanism Model for Improving Image Captioning. Appl. Sci. 2023, 13, 7916. https://doi.org/10.3390/app13137916

AMA Style

Xie T, Ding W, Zhang J, Wan X, Wang J. Bi-LS-AttM: A Bidirectional LSTM and Attention Mechanism Model for Improving Image Captioning. Applied Sciences. 2023; 13(13):7916. https://doi.org/10.3390/app13137916

Chicago/Turabian Style

Xie, Tian, Weiping Ding, Jinbao Zhang, Xusen Wan, and Jiehua Wang. 2023. "Bi-LS-AttM: A Bidirectional LSTM and Attention Mechanism Model for Improving Image Captioning" Applied Sciences 13, no. 13: 7916. https://doi.org/10.3390/app13137916

APA Style

Xie, T., Ding, W., Zhang, J., Wan, X., & Wang, J. (2023). Bi-LS-AttM: A Bidirectional LSTM and Attention Mechanism Model for Improving Image Captioning. Applied Sciences, 13(13), 7916. https://doi.org/10.3390/app13137916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop