Next Article in Journal
Multifunctional Cement Mortars Enhanced with Graphene Nanoplatelets and Carbon Nanotubes
Next Article in Special Issue
DailyCog: A Real-World Functional Cognitive Mobile Application for Evaluating Mild Cognitive Impairment (MCI) in Parkinson’s Disease
Previous Article in Journal
Research on the Prediction of Green Plum Acidity Based on Improved XGBoost
Previous Article in Special Issue
Using a Stochastic Agent Model to Optimize Performance in Divergent Interest Tacit Coordination Games
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NMN-VD: A Neural Module Network for Visual Dialog

Department of Computer Science, Kyonggi University, Suwon 16227, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(3), 931; https://doi.org/10.3390/s21030931
Submission received: 26 December 2020 / Revised: 25 January 2021 / Accepted: 26 January 2021 / Published: 30 January 2021
(This article belongs to the Special Issue Human-Computer Interaction in Smart Environments)

Abstract

:
Visual dialog demonstrates several important aspects of multimodal artificial intelligence; however, it is hindered by visual grounding and visual coreference resolution problems. To overcome these problems, we propose the novel neural module network for visual dialog (NMN-VD). NMN-VD is an efficient question-customized modular network model that combines only the modules required for deciding answers after analyzing input questions. In particular, the model includes a Refer module that effectively finds the visual area indicated by a pronoun using a reference pool to solve a visual coreference resolution problem, which is an important challenge in visual dialog. In addition, the proposed NMN-VD model includes a method for distinguishing and handling impersonal pronouns that do not require visual coreference resolution from general pronouns. Furthermore, a new Compare module that effectively handles comparison questions found in visual dialogs is included in the model, as well as a Find module that applies a triple-attention mechanism to solve visual grounding problems between the question and the image. The results of various experiments conducted using a set of large-scale benchmark data verify the efficacy and high performance of our proposed NMN-VD model.

1. Introduction

Recent developments in computer vision and natural language processing technologies have contributed to increasing interests in multimodal artificial intelligence, which involves the simultaneous understanding of images and language. Representative multimodal artificial intelligence tasks based on images and language include image/video captioning, visual question answering (VQA) [1], and visual dialog [2]. VQA is the task of providing an accurate natural language answer to a natural language question about a given image. By contrast, visual dialog is the task of generating a natural language answer A t to a question Q t , given an image and a caption about the image, dialog history about the image < ( Q 1 , A 1 ) ,   ( Q t 1 , A t 1 )>, and new natural language question Q t , as illustrated in Figure 1. As visual dialog extends VQA to a multi-round dialog, it encounters the visual grounding problem, which is a major challenge in VQA.
The visual grounding problem involves identifying the region of the image that the natural language question seeks to understand. For example, in question Q3 in Figure 1, the region of the image indicated by “the surfboard” should be found. Furthermore, as visual dialog exchanges questions and answers consecutively about one image, the question–answer pairs comprising the dialog are interdependent. Owing to this characteristic, visual dialog encounters the visual coreference resolution problem, which involves determining the pronouns and noun phrases in a multi-round natural language dialog that co-refer to the same object instance in an image. For example, to produce the correct answer to question Q4 in Figure 1, it must be known that the pronoun “him” in question Q4, “he” in question Q2, and “man” in caption C refer to the same male region in the image.
The existing models for visual dialog have been mostly implemented with a large monolithic neural network [3,4,5,6,7,8,9,10,11]. However, VQA and visual dialog are composable in nature in that the process of generating an answer to one natural language question can be completed by composing multiple basic neural network modules. Moreover, various types of questions can appear in a visual dialog, each requiring a slightly different detailed process. For example, in the example of Figure 1, question Q3 must find a region corresponding to “the surfboard” in the image and determine the color of “the surfboard” in that region. By contrast, question Q4 must find regions indicated by “the surfboard” and “him” in the image and compare their heights. Thus, it is difficult to learn one monolithic neural network structure that can generate answers to various types of questions. It is likewise difficult for the designer to understand the reasoning process of the neural network via which a large-scale monolithic neural network structure generates answers.
To overcome these problems of monolithic neural network structures, a neural module network (NMN) model for visual dialog [5] was adopted. As natural language questions are inherently compositional, they require multiple levels of reasoning for each question. The NMN model analyzes a question with linguistic substructures and dynamically composes modules on the basis of the structures. Visual grounding is another difficulty encountered; this refers to the understanding of the linguistic elements (e.g., nouns and verbs) that appear in a natural language question in connection with the regions and visual elements in the input image. In addition, visual grounding is highly useful for making a visual dialog model more interpretable because it can explain the rationale for the answer using visual elements. Furthermore, in visual dialog, various comparison questions exist to compare the length, size, color, and other characteristics of two objects in an image. For example, question Q4 in Figure 1 is a comparison question that can be answered only by comparing the heights of “the surfboard” and “him” in the image. Lastly, the visual coreference resolution of pronouns is a major challenge in visual dialog; the impersonal pronouns that appear in dialogs do not need visual coreference resolution, unlike general pronouns. Impersonal pronouns are only used formally, and there are no particular nouns in the dialog history or visual elements in the image that they indicate.
To address these problems, we propose the novel neural module network for visual dialog (NMN-VD), which can lower the complexity of neural network structures and reduce the number of parameters to be learned by selectively composing neural network modules required for the processing of each question. The proposed model has the advantage of more clearly understanding the answer generation process of each question than the monolithic neural network model, and it provides the basic neural network modules for visual dialog: Find, Refer, And, Relocate, Describe, and Compare.
In the proposed model, answers to natural language questions are generated through program generation and execution steps. To solve the visual grounding problem, the proposed model adopts the Find module that applies a triple-attention mechanism to improve the find performance, unlike the Find modules of existing models that use single- or double-attention mechanisms. Furthermore, to process these comparison questions effectively, the proposed model contains the Compare module that determines a minimum bounding area that includes two object regions in the image and extracts a context for comparison operation from this area.
Lastly, the proposed model composes modules to process impersonal pronouns differently from general pronouns. Various experiments were conducted using two large-scale benchmark datasets, VisDial v0.9 and v1.0 [2], to analyze the performance of the proposed NMN-VD.
The remainder of this paper is organized as follows: Section 2 reviews existing work related to this study. Section 3 presents the detailed design of the proposed NMN-VD model. Section 4 describes the implementation of the proposed model and presents experimental results. The conclusions of the study are then presented in Section 5.

2. Related Work

2.1. Visual Question and Answering (VQA)

VQA [1] is a task in which an appropriate answer is automatically generated to a question when an image and a natural language question about the image are given. To solve the visual grounding problem, which is a major challenge in VQA, early research focused on the attention mechanism. In particular, the visual attention mechanism is a technology that generates a heatmap on the image to indicate candidate image regions from which the answer can be derived. Early studies mainly investigated unidirectional visual attention to find regions related to entities or relationships mentioned in natural language questions [12,13,14,15,16].
Later studies often researched bidirectional attention mechanisms that focus on the relationships between images and natural language questions [17,18,19,20,21]. Lu et al. [17] proposed a model that applied the question attention and visual attention for the first time. Their model was designed to learn the region of the input image and part of the question that should be focused on simultaneously by applying co-attention between the image and natural language questions. Lao et al. [18], Nam et al. [19], and Yu et al. [20] proposed multilevel attention models between image and natural language questions. Furthermore, Yang et al. [21] proposed a model that uses not only the co-attention mechanism between the question and input image but also the question type information.

2.2. Visual Dialog

Visual dialog researchers have developed attention mechanisms for images, questions, and previous dialogs to solve the visual grounding problem. Das et al. [2] and Lu et al. [7] adopted a dialog history attention mechanism to find and focus on past dialog history related to the present question. Wu et al. [10], Guo et al. [4], and Yang et al. [11] proposed a model for applying the co-attention mechanism among the three elements of current question, image, and past dialog history to determine the answer to the current question. Gan et al. [3] proposed a model that repeats co-attention among the three elements several times. Idan et al. [22] developed a factor graph-based attention framework, where nodes correspond to utilities and factors model their interactions. Park et al. [23] proposed the multiview attention network (MVAN), which leverages multiple views about heterogeneous inputs on the basis of attention mechanisms.
Unlike the VQA, which is performed once, visual dialog has interdependence between questions and answers because questions and answers are exchanged several times consecutively for one image. Owing to this characteristic, visual dialog has a visual coreference resolution problem. For visual coreference resolution, the visual dialog model of [9] stored pairs of dialog context and its corresponding visual attention map together in an attention memory network for each round of dialog. A dialog context consists of a dialog history and a new question. Furthermore, this model searched for and used a visual attention map of a past dialog context that has the highest relevance in the attention memory network when finding a reference image region of a new question. By contrast, the model proposed by Kottur et al. [5] used a memory network that stores pairs of each noun word appearing in a natural language question or answer during the dialog and the corresponding visual attention region together. In other words, for the visual coreference resolution problem, the models of [5,9] operate a separate attention memory network at the sentence level and word level, respectively. However, the models of [8,6] adopted visual coreference resolution methods that first find the object indicated by a pronoun of the new natural language question in the past history of natural language dialogs before finding the corresponding visual attention map. However, existing studies on visual dialog including [5] have not attempted to process impersonal pronouns separately from general pronouns, unlike the model proposed in this study.

2.3. Neural Module Network

In recent times, deep neural networks have shown excellent performance in various computer vision tasks. However, a more complex task leads to a larger size and a more complex structure of the deep neural network. Hence, the network requires several calculation resources and considerable training time to train a large, robust monolithic neural network model. Meanwhile, different computer vision tasks often require several common subtasks. Therefore, when the complexity of tasks is high, the modular approach can be more efficient as it divides tasks into subtasks and composes neural modules appropriate for each subtask [24,25]. Some previous studies applied such NMNs to VQA and visual dialog tasks [5,26,27,28].
Andreas et al. [26] proposed the first NMN model for VQA. This model decomposes a natural language question into linguistic substructures using a semantic parser and generates a customized deep neural network that can provide an answer to the given question by composing the required neural network modules on the basis of these substructures. However, this model has the limitation that a separately developed semantic parser must be used to generate a neural module network that is suitable for the natural language question. Follow-up studies [27,28] proposed an NMN model that can suggest a composition of modules appropriate for the question by itself without using a semantic parser. Kottur et al. [5] proposed the first NMN model for visual dialog. However, the model of [5] cannot distinguish impersonal pronouns from general pronouns. Moreover, it does not contain any independent module to process comparison questions. Thus, the NMN-VD model proposed in this present study can be considered as an improvement to that of [5], as ours separately processes impersonal pronouns, includes a new compare module, and adopts a revised find module.

3. Neural Module Network Model

3.1. Model Overview

The NMN-VD model proposed in this study selects the most correct answer to the given question by considering the dialog context in the visual dialog. It decomposes a natural language question into language substructures and generates a question-customized program by composing modules appropriate for each substructure. The program is then run to determine the most appropriate answer to the question.
The structure of the proposed model is shown in Figure 2. The model receives an image I , a dialog history including a short caption C about the image I , H t = < ( C a p t i o n   C ) , ( Q 1 , A 1 ) , ( Q 2 ,   A 2 ) , ,   ( Q t 1 , A t 1 ) , and a question of the current round t , Q t , as inputs. Then, using these inputs, the proposed NMN-VD model selects the answer A t that is most appropriate for question Q t among the candidate natural language answers of the current round t . The entire process of the proposed model consists of two steps: (a) program generation and (b) program execution. In (a), the program generation step, a question-customized program P t is generated, which includes the modules required for deciding the appropriate answer A t to the question Q t and the execution sequence of the modules. In (b), the program execution step, the context vector C t is generated by dynamically connecting and executing neural network modules according to the generated program P t . Then, the most appropriate answer A t is determined in the candidate answer list L t using the generated context vector C t . For example, in Figure 2, when the program generator of the proposed model analyzes question Q t using language substructures, the result of [NP (the surfboard), VP (looks to be taller than, NP (PRP(him)))] is obtained, where NP denotes the noun phrase, VP denotes the verb phrase, and PRP denotes the personal pronoun. At this point, the program generator matches the Find module to NP, the Refer module to the PRP, and the Compare module to VP, which contains comparatives such as “taller than”. Consequently, the program generator generates the question-customized program P t = [ Refer ,   Find ,   Compare ] to obtain the answer to the question Q t . Next, the program executor of the proposed NMN-VD model determines the most appropriate answer A t in the candidate answer list for the question Q t by connecting and executing the neural network modules according to the generated program P t .
The detailed structure of the program generator is shown in Figure 3. As shown in the figure, the program generator encodes the natural language question Q t as a question feature vector q t through a multilayer long short-term memory (LSTM). Additionally, the program generator generates the dialog history feature vector h t by encoding C and ( Q i , A i ) constituting the dialog history H t through another multilayer LSTM. Then, the program generator performs an inner product operation to calculate the correlation between each question–answer pair ( Q i , A i ) in the dialog history H t and the current question Q t . Furthermore, the program generator generates the final feature vector h ˜ t of the dialog history H t by calculating the weighted sum ⊕ of the question–answer pairs constituting the dialog history H t on the basis of this correlation. The question feature vector q t and the dialog history feature vector h ˜ t generated in this way are concatenated, and the question feature vector q ˜ t considering the dialog history context is generated through the fully connected layer (FC). Lastly, the program generator generates pairs of module labels and textual attention maps ( m t , i , a t , i ) required to determine the answer to the question Q t one per time through multilayer LSTM as a function of the question feature vector q ˜ t . Consequently, the program generator generates the final output program of the form P t = [ ( m t , 1 , a t , 1 ) ,   ( m t , 2 , a t , 2 ) ,   ,   ( m t , i , a t , i ) ]. In the case of Figure 3, the final program for question Q 4 is P 4 = [ ( R e f e r , a 4 , 1 ) ,   ( F i n d , a 4 , 2 ) ,   ( C o m p a r e , a 4 , 3 ) ], where m 4 , 1 of program P 4 denotes the label “ R e f e r ”, which is the neural network module that must be executed first, and a 4 , 1 denotes the attention map for “him”, which is a region that must be processed by the R e f e r module in the natural language question Q 4 . The learning process of the program generator aims to minimize the cross-entropy loss l p g between the program P t predicted by the program generator and the ground-truth program P t * .
The detailed structure of the program executor is shown in Figure 4. The program executor extracts the visual features x v i s from the input image I using the convolutional neural network VGG16. Then, the program executor determines the textual feature x t x t of the question region that must be processed by each neural network module m t , i by applying the textual attention map a t , i to the question feature vector q t . For most neural network modules, the visual features x v i s extracted from the image and the textual features x t x t extracted from the natural language question are given as default inputs. Depending on the case, a visual attention map or reference pool is given as an additional input. Once the program executor executes all the neural network modules according to the sequence specified in the program P t , the context vector C t to be used for deciding the answer is generated. Subsequently, the program executor calculates the correlation between each candidate answer A t , i (which is encoded through the LSTM) and the context vector C t by conducting an inner product operation of both. The probability P r t of each candidate answer A t , i is calculated on the basis of this correlation. Among the candidate answers, the candidate answer with the highest probability A t , i is selected as the most appropriate predicted answer A t . The learning process of the program executor aims to minimize the negative log-likelihood loss l p e between the predicted answer A t and ground-truth answer A t * .

3.2. Neural Network Module

The list of modules used in the NMN program is provided in Table 1. The modules can be largely divided into two types based on output: the Find, Relocate, And, and Refer modules generate an attention map for the input image as an output, whereas the Describe and Compare modules generate the context vector for deciding the answer as an output. The Find, Refer, and Compare modules are explained in more detail in Section 3.3, Section 3.4 and Section 3.5. The inherent function of each module is described here.
As presented in Table 1, the And module receives two visual attention maps, a 1   and   a 2 , and outputs the image region that is commonly indicated by them. In the example of Figure 5, the visual attention maps corresponding to “him” and “the elephant” are given as inputs of the And module; the output of the And module here is the common region where objects in the image overlap.
The relocate module performs the function of moving the visual attention map a for the image x v i s to the direction indicated by x v i s . Here, x t x t corresponds to a prepositional phrase that appears in the natural language question, such as “behind”, “in front of”, “up”, and “down.” In the example of Figure 5, x t x t , which indicates the moving direction, corresponds to “behind” that appears in a natural language question. Accordingly, the Relocate module moves the common regions of “him” and “the elephant” in this image further back. The operation equation of the Relocate module is as follows:
y = c o n v 2 ( c o n v 1 ( x v i s ) W 1 ( s u m ( a x v i s ) )   W 2 x t x t ) ,
where x v i s denotes the visual feature of the image, a denotes the visual attention map of the image, and x t x t denotes the natural language question part that indicates a direction.
The Describe module plays the role of generating the context vector C t , which is required to determine the answer on the basis of the visual attention map a and the natural language question x t x t for the image x v i s . The operation equation for the Describe module is as follows:
y = W 3 T ( W 1 s u m ( a x v i s ) W 2 x t x t ) .

3.3. Refer Module

The Refer module is a neural network module for solving the visual coreference resolution problem. To find the object indicated by a noun phrase or a pronoun in the question Q t , the Refer module searches the reference pool. Using a dictionary format composed of key–value pairs, the reference pool stores nouns and noun phrases that appeared in previous questions and answers of dialogs and the image regions corresponding to them. Thus, the key k i in a reference pool is a noun or noun phrase that appears in a question or answer, and the corresponding value is a visual attention map a i indicating the region of the corresponding object in the image. The proposed NMN-VD model finds the visual attention map, which is the region of the corresponding object in the image, using the Find module whenever a noun or noun phrase appears in the natural language question of the dialog. At this point, the pair ( x t x t , a ) of the noun text x t x t (the input of the Find module) and the visual attention map a (the output of the Find module) is stored in the reference pool.
Figure 6 illustrates the work of the reference pool and Refer module. In this figure, x t x t indicates the pronoun “him” of the new question Q t , which requires coreference resolution. The Refer module of the proposed NMN-VD model searches for the reference key k i that has the highest correlation with the pronoun x t x t in the reference pool. The correlation between x t x t , which is the pronoun part of the question Q t , and each key k i of the reference pool is calculated by the scoring network as follows:
y i ˜ = S o f t m a x ( W 1 ( [ x t x t ,   k i ,   i R D ] ) ) ,
where i R D denotes the round distance between the current dialog round t and the past dialog round in which k i appeared. Considering the sequential nature of dialog, the pronoun of the current question Q t is highly likely to indicate an object that appeared in the most recent dialog round. The scoring network concatenates the pronoun x t x t , key k i of the reference pool, and dialog round difference i R D between these two into one and evaluates the relevance of each key k i through the softmax layer. As a function of the relevance score y i ˜ of each reference key, the equation for obtaining the final attention map a , which is the image region of the target object indicated by the pronoun x t x t , is as follows:
a =   i = 1 | P r e f | y i ˜ a i ,
where | P r e f | denotes the number of key–value pairs in the reference pool. The relevance score y i ˜ of each key k i for the pronoun x t x t is used as the weight of the visual attention map a i , which is the image region indicated by the key. In other words, the Refer module determines the visual attention map a , which is the object region indicated by the pronoun x t x t , by calculating the weighted sum of the visual attention maps a i stored in the reference pool according to the relevance score y i ˜ of each reference key k i for the pronoun x t x t .

3.4. Processing Impersonal Pronouns

We further propose a method of distinguishing between general pronouns that require visual coreference resolution and impersonal pronouns that do not. As mentioned in Section 3.3, in the case of general pronouns, the proposed NMN-VD model solves the visual coreference resolution problem using the reference pool and Refer module. However, impersonal pronouns do not require visual coreference resolution because they generally do not refer to specific objects that appear in the previous dialog history. When impersonal pronouns are processed using the reference pool and Refer module as in general pronouns, a wrong answer can be derived by connecting the target object regions of the past dialog history unreasonably with impersonal pronouns that are unrelated. Therefore, the proposed NMN-VD model generates a program for impersonal pronouns that is different from the one for general pronouns in the program generation step.
Figure 7 illustrates an example of the program generated by the proposed NMN-VD model to process impersonal pronouns. In question Q1, “Can you tell if it is daytime?”, “it” is an impersonal pronoun and is used to ask about time. When “it” in Q1 is processed as a general pronoun, the generated program will be [Refer module, Describe module]. However, the proposed NMN-VD model was designed to generate a correct program such as [Find module, Describe module] by distinguishing “it” as an impersonal pronoun.

3.5. Compare Module

This study proposes a new Compare module to answer effectively various comparison questions that require comparison of properties (e.g., height and size) between two objects in an image. The inputs of the Compare module are the visual attention maps a 1 and   a 2 of two objects, the visual features of the entire image x v i s , and the expression of comparison relation x t x t . Furthermore, the output is the context vector C that indicates the comparison result of the two objects. The Compare module requires visual information of the image regions in which the two compared objects are included. Therefore, the Compare module in Equation (5) of the proposed NMN-VD model determines the minimum bounding area y ˜ , which includes the two objects in the image, according to the visual attention maps a 1   a n d   a 2 of two objects, as follows:
y ˜ = min ( T h r e s h o l d ( a 1 , a 2 ) ) .
Figure 8 illustrates an example of calculating the minimum bounding areas of two objects for comparison question Q4, “Does the surfboard look to be taller than him?”, which includes the comparison expression “taller than”. In Figure 8, a 1 and a 2 denote the visual attention maps of “him” and “the surfboard”, respectively, which are objects of comparison, and y ˜ denotes the minimum bounding area of the two objects determined by the Compare module.
C =   W 3 T ( W 1 s u m ( y ˇ x v i s ) W 2 x t x t ) .
Furthermore, the Compare module generates the context vector C , which is the final result of comparison, as shown in Equation (6), according to the minimum bounding area y ˜ of the two objects, visual features x v i s of the entire image, and expression of the comparison relationship x t x t .

3.6. Find Module

The proposed NMN-VD model uses the Find module to solve the visual grounding problem in a visual dialog. To improve the detection performance, the Find module of the NMN-VD model applies a triple-attention mechanism.
Figure 9 illustrates the effects of a more accurate detection of the “the surfboard” region in the image with an increasing number of attentions. However, excessive attention above a certain number narrows the detection area, and the correct object region cannot be found. Therefore, the NMN-VD model applies triple attention that detects the object region most accurately.
y s = c o n v 1 ( x v i s x t x t ) .
y d = c o n v 2 ( ( ( s u m ( y s x v i s ) x t x t ) x t x t ) x v i s ) .
y t =   c o n v 3 ( ( ( s u m ( y d x v i s ) x t x t ) x t x t ) x v i s ) .
Equations (7)–(9) are the equations for obtaining the single-, double-, and triple-attention maps.

4. Implementation and Experiments

4.1. Datasets and Model Training

We conducted experiments to analyze the performance of the proposed NMN-VD model using VisDial v0.9 and v1.0, which are two benchmark datasets for visual dialog. The VisDial v0.9 dataset is a collection of dialogs that people exchanged about each image through Amazon Mechanical Turk (AMT) on the basis of the MS-COCO image dataset. Each dialog consists of 10 question–answer pairs on average, and 100 candidate answers are given to each question of a dialog. The VisDial v0.9 dataset is composed of training and validation datasets that consist of 82,783 and 40,504 images and dialog data, respectively. VisDial v1.0 is an extension of VisDial v0.9. It has 123,000 images for the training split that combine the training and validation splits of VisDial v0.9. An additional 10,000 images from the Flickr dataset are utilized to construct the validation and test splits in VisDial v1.0, which contain 2000 and 8000 images, respectively. Additionally, ground-truth programs for each question of the dialogs were used for training the program generator of the NMN-VD model in addition to the original VisDial v0.9 and v1.0 datasets.
The proposed NMN-VD model was implemented using the Python deep learning library TensorFlow. Furthermore, the training and performance evaluation experiments of the NMN-VD model were performed in a computer environment with a GeForce GTX 1080 Ti graphics processing unit (GPU) and Ubuntu 16.04 LTS operating system. The model was trained end-to-end, and the total loss for the model training l t o t was calculated as follows:
l t o t =   l p g +   l p e .
In other words, the total loss l t o t is the sum of the losses of the program generation step l p g and program execution step l p e . The optimization algorithm used for model training was Adam, and the batch size and epoch were set to 5 and 20, respectively.

4.2. Performance Evaluation of Each Module (Ablation Study)

Experiments were conducted to analyze the effect of each module of the proposed NMN-VD model using four different performance metrics. The first performance metric was the mean reciprocal rank (MRR) of the human response in the returned ranked list. MRR is expressed as follows:
M R R =   1 | Q | i = 1 | Q | 1 r a n k i ,
where | Q | denotes the total set of questions, and rank i is the rank of the predicted human answer to the i -th question in the returned ranked list. The second performance metric is Recall@k, which represents the probability of the existence of the human response in the top-k responses to each question. The third performance metric is mean, which is the mean rank of human responses. For example, when the rank of the predicted human responses to five questions is [1, 3, 6, 8, 10], the mean of the ranks is 5.6. The fourth performance metric is normalized discounted cumulative gain (NDCG), which takes into account all relevant answers from the 100-answer list by using the densely annotated relevance scores. It penalizes the lower-ranked answers with high relevance scores. NDCG is computed as follows:
D C G @ k =   i = 1 k r e l e v a n c e i l o g 2 ( i + 1 ) ,
N D C G @ k =   D C G @ k   f o r   s u b m i t t e d   r a k i n g D C G @ k   f o r   i d e l   r a k i n g ,
where k is the number of answer options whose relevance is greater than zero. Of these metrics, a higher score is better for NDCG, MRR, and R@1, 5, 10, but a lower score is better for mean rank.
The first experiment analyzed the effect of attention applied to the Find module of the proposed NMN-VD model. The experimental results are summarized in Table 2, where “single attention,” “dual attention,” “triple attention,” and “quad attention” denote the cases of the Find module applying one, two, three, and four attentions to the image, respectively.
When the performances of the Find module applying single attention, dual attention, and triple attention were compared as shown in Table 2, it was observed that higher numbers of attention to the image improved the performance in almost all metrics. However, when triple attention and quad attention were compared, the case of quad attention, which applied too much attention, showed a lower performance than that of triple attention. Therefore, in the experiment of this study using the VisDial v0.9 dataset, the proposed NMN-VD model demonstrated the best performance when the Find module applying triple attention was used. Thus, the Find module applying triple attention was used in the remaining experiments.
The second experiment analyzed the Refer module (which is used for the coreference resolution of general pronouns in the proposed NMN-VD model) and the effect of the separate processing of impersonal pronouns. To analyze the effect of the separate processing of impersonal pronouns, the VisDialR v0.9 dataset was used in addition to the existing VisDial v0.9 dataset. The VisDialR v0.9 dataset is a partial dataset of dialogs that includes at least one pronoun in the existing VisDial v0.9 dataset. In Table 2, “Refer” indicates the Refer module of the CorefNMN model [5], which processes impersonal pronouns in the same way as general pronouns; “Refer + Impersonality” indicates the Refer module of the proposed NMN-VD model, which distinguishes general and impersonal pronouns and processes them separately.
As observed in the experimental results in Table 3, for both the VisDialR v0.9 and the VisDial v0.9 datasets, the “Refer + Impersonality” Refer module showed a higher performance than the “Refer” Refer module. This experimental result confirms that the separate processing of general pronouns (which requires coreference resolution) and impersonal pronouns (which does not) as in the proposed NMN-VD model helped to improve the performance. Furthermore, when the VisDialR v0.9 and VisDial v0.9 datasets were compared, both the “Refer” and “Refer + Impersonality” Refer modules showed a higher performance with the VisDialR v0.9 dataset. This result suggests that both Refer modules effectively solved the coreference resolution problem of general pronouns.
The third experiment analyzed the performance of the Compare module of the proposed NMN-VD model. Table 4 outlines the results of this experiment. In the table, “No Compare” indicates the case of not using a separate Compare module, whereas “Compare (Inner Product)” indicates the Compare module that performs the comparison operation as shown in Equation (11) [17].
y =   W 4 T ( W 1 s u m ( a 1 x v i s ) W 2 s u m ( a 1 x v i s ) W 3 x t x t ) .
The Compare (Inner Product) Compare module expressed by Equation (11) applies the attention maps a 1 and a 2 for two objects separately to the image without combining them. Furthermore, “Compare (Or)” in Table 4 indicates the Compare module that performs a union of attention maps a 1 and a 2 of the two objects. Lastly, “Compare (Ours)” indicates the Compare module of the proposed NMN-VD model using the minimum bounding area of a 1 and a 2 .
In the experimental results presented in Table 4, the Compare (Ours) Compare module proposed in this study showed the best performance. When the other three modules (excluding Compare (Ours)) were compared, the cases of Compare (Inner Product) and Compare (Or), which added the Compare module, showed a slightly lower performance than the case of “No Compare”, which did not include the Compare module. The comparison method of Compare (Inner Product) makes it difficult to find the relationship between two objects in an image because they are processed separately rather than in a combined manner. Furthermore, the comparison method of Compare (Or) obtains a distributed visual attention map, although it contains both objects because it uses the union of two visual attention maps. These factors are presumed to be the cause of the lower performance compared with the cases without the Compare module.
Contrastingly, the Compare (Ours) Compare module proposed in this study uses the minimum bounding area of two objects; thus, it can obtain the information of two objects and their surroundings when answering a comparison question. Therefore, this characteristic of the Compare (Ours) Compare module was likely a factor in obtaining a better performance than the other three cases.

4.3. Performance Comparison with Existing Models

To prove the efficacy of the NMN-VD model proposed in this study, some experiments were performed to compare the performance with existing visual dialog models. Table 5 outlines the results of the performance comparison experiment with VisDial v0.9 and v1.0. In these experiments, the proposed NMN-VD (Ours) model was compared with the existing models, including late fusion (LF) [2], hierarchical recurrent encoder (HRE) [2], memory network (MN) [2], history-conditioned image attention encoder (HCIAE) [7], attention memory (AMEM) [9], co-attention network (CoAtt) [10], and coreference neural module network (CorefNMN) [5].
As summarized in Table 5, AMEM, CorefNMN, and the proposed NMN-VD model (Ours) demonstrated much higher performances in all metrics in solving the visual coreference resolution problem, compared with the basic models proposed in [2]. This experimental result suggests that solving the visual coreference resolution problem in visual dialog has a significant effect on performance improvement. Furthermore, compared to the monolithic neural network models of [2,7,9,10], CorefNMN and NMN-VD (Ours), which use a modular neural network, showed higher performances. This result means that the question-customized NMN was more effective in processing various question types of visual dialog than a large monolithic neural network.
Lastly, in comparing CorefNMN and NMN-VD (Ours) that use similar reference pools and Refer modules for the visual coreference resolution of pronouns, NMN-VD (Ours) showed higher performance improvements in both VisDial v0.9 and v1.0 datasets. Specifically, NMN-VD (Ours) achieved significant improvements in NDCG from 54.79 to 56.90, in MRR from 61.50 to 63.00, in R@1 from 47.55 to 49.23, and in mean from 4.40 to 4.31 for VisDial v1.0. This is because, unlike the conventional CorefNMN model, the proposed model NMN-VD (Ours) includes not only a function to process general and impersonal pronouns separately but also a newly developed Compare module for comparison questions and the Find module with improved performance. Therefore, the efficacy of the NMN-VD (Ours) proposed in this study compared with existing models was verified through the experimental results presented in Table 5.

4.4. Qualitative Evaluation

Figure 10, Figure 11 and Figure 12 illustrate examples of visual dialog to which the proposed NMN-VD model was applied. Each figure shows not only the program and predicted answers generated for each question of consecutive questions about one input image but also the changes in the visual attention map, linguistic attention map, and reference pool for each module.
The dialog in Figure 10 illustrates an example of using the Refer module of the NMN-VD model. To examine the linguistic and visual attention maps generated first in this dialog, the proposed model detected tennis players and tennis courts mentioned in caption C using the Find module and saved them in the reference pool. Then, the model used the Refer module for visual reference resolution for the pronoun “they” that appears in questions Q2 and Q3. The pronoun “they” in question Q2 was processed to refer to the same visual attention map as that of the “tennis players” mentioned in caption C. Furthermore, the pronoun “they” in question Q3 was processed to refer to the same visual attention map as that of the previous answer A2.
The dialog in Figure 11 shows an example of using the Find module of the NMN-VD model in caption C, question Q1, and question Q2. An examination of the linguistic and visual attention maps reveals that the Find module of NMN-VD correctly detected the image region corresponding to the zebra in caption C. Furthermore, it can be seen that the Find module of NMN-VD correctly detected the image regions of the corresponding objects required in the answers to the questions about the existence of certain objects in the input image, such as questions Q1 and Q2. Meanwhile, in determining the answer to question Q3 in Figure 11, the Compare module of the NMN-VD model was used. To compare the heights of “zebra” and “the rocks”, mentioned in question Q3, the NMN-VD model found the image regions of the corresponding objects in the input image using the Refer and Find modules. The minimum bounding area of the two objects was then determined using the Compare module, and the correct answer to the “taller” comparison relationship was determined as a function of the minimum bounding area.
Question Q3 in Figure 12 illustrates an example of applying the impersonal pronoun processing method of the NMN-VD model. Question Q3 asks about the weather of the input image and contains the impersonal pronoun “it”. The NMN-VD model identified “it” in question Q3 as an impersonal pronoun rather than a general pronoun that requires visual coreference resolution. Therefore, the proposed model detected the sky region required for weather judgment using the Find module instead of the Refer module and subsequently determined the correct answer. The case analyses of Figure 10, Figure 11 and Figure 12 to which the NMN-VD model was applied confirmed that the Find, Refer, and Compare modules, as well as the impersonal pronoun processing method of the NMN-VD model, were effectively used for various question-answering cases of visual dialog.

5. Conclusions

This study proposed a novel neural module network for visual dialog (NMN-VD). It is an efficient question-customized neural module network model that composes only the modules required to determine answers by analyzing the input question. In particular, the model includes a reference module that effectively finds the visual region indicated by pronouns using a reference pool to solve the visual coreference resolution problem, which is an important problem in visual dialog. Furthermore, the proposed NMN-VD model includes a method for processing impersonal pronouns, which do not require visual coreference resolution, separately from general pronouns. Additionally, the model includes a new Compare module that effectively processes comparison questions that appear often in visual dialog. Moreover, the proposed model includes the Find module, which applies the triple-attention mechanism to solve the visual grounding problem between the question and image. The efficacy of the proposed NMN-VD model was verified through various experiments using two large-scale benchmark datasets, VisDial v0.9 and v1.0.
As a limitation, the current NMN-VD model sometimes composes the same module twice unnecessarily during the program generation process. Therefore, follow-up research is needed to improve the accuracy of program generators. Additionally, the visual dialog dataset VisDial v0.9 contains several unusual questions with low frequencies. Therefore, the NMN-VD model will be extended in a future study to process effectively unusual questions with low frequencies.

Author Contributions

Conceptualization, Y.C. and I.K.; methodology, I.K.; software, Y.C.; validation, I.K.; formal analysis, Y.C.; investigation, I.K.; resources, Y.C. and I.K.; writing—original draft preparation, Y.C.; writing—review and editing, I.K. All authors read and agreed to the published version of the manuscript.

Funding

This work was supported by the Technology Innovation Program (or Industrial Strategic Technology Development Program) (10077538, Development of manipulation technologies in social contexts for human-care service robots) funded by the Ministry of Trade, Industry, and Energy (MOTIE, Korea).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Agrawal, A.; Lu, J.; Antol, S.; Mitchell, M.; Zitnick, C.L.; Parikh, D.; Batra, D. VQA: Visual question answering. Int. J. Comput. Vis. 2017, 123, 4–31. [Google Scholar] [CrossRef]
  2. Das, A.; Kottur, S.; Gupta, K.; Singh, A.; Yadav, D.; Moura, J.M.F.; Parikh, D.; Batra, D. Visual dialog. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Honolulu, HI, USA, 22–25 July 2017; pp. 326–335. [Google Scholar] [CrossRef]
  3. Gan, Z.; Cheng, Y.; Kholy, A.E.; Li, L.; Liu, J.; Gao, J. Multi-step reasoning via recurrent dual attention for visual dialog. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; Association for Computational Linguistics: Florence, Italy, 2019; pp. 6463–6474. [Google Scholar] [CrossRef]
  4. Guo, D.; Xu, C.; Tao, D. Image-question-answer synergistic network for visual dialog. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Long Beach, CA, USA, 16–20 June 2019; pp. 10434–10443. [Google Scholar] [CrossRef] [Green Version]
  5. Kottur, S.; Moura, J.M.; Parikh, D.; Batra, D.; Rohrbach, M. Visual coreference resolution in visual dialog using neural module networks. In Proceedings of the 15th European Conference on Computer Vision, ECCV, Munich, Germany, 8–14 September 2018; pp. 153–169. [Google Scholar] [CrossRef] [Green Version]
  6. Kang, G.; Lim, J.; Zhang, B.T. Dual attention networks for visual reference resolution in visual dialog. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, EMNLP, Hong Kong, China, 3–7 November 2019; pp. 2024–2033. [Google Scholar] [CrossRef]
  7. Lu, J.; Kannan, A.; Yang, J.; Parikh, D.; Batra, D. Best of both worlds: Transferring knowledge from discriminative learning to a generative visual dialog model. In Proceedings of the 31st Conference on Neural Information Processing Systems, NIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 314–324. [Google Scholar]
  8. Niu, Y.; Zhang, H.; Zhang, M.; Zhang, J.; Lu, Z.; Wen, J.-R. Recursive visual attention in visual dialog. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Long Beach, CA, USA, 16–20 June 2019; pp. 6679–6688. [Google Scholar] [CrossRef] [Green Version]
  9. Seo, P.H.; Lehrmann, A.; Han, B.; Sigal, L. Visual reference resolution using attention memory for visual dialog. In Proceedings of the 31st Conference on Neural Information Processing Systems, NIPS, Long Beach, CA, USA, 4–9 December 2017; pp. 3719–3729. [Google Scholar]
  10. Wu, Q.; Wang, P.; Shen, C.; Reid, I.; van den Hegel, A. Are you talking to me? Reasoned visual dialog generation through adversarial learning. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6106–6115. [Google Scholar] [CrossRef] [Green Version]
  11. Yang, T.; Zha, Z.J.; Zhang, H. Making history matter: History-advantage sequence training for visual dialog. In Proceedings of the International Conference on Computer Vision, ICCV, Seoul, Korea, 27 October–2 November 2019; pp. 2561–2569. [Google Scholar] [CrossRef] [Green Version]
  12. Anderson, P.; He, X.; Buehler, C.; Teney, D.; Johnson, M.; Gould, S.; Zhang, L. Bottom-up and top-down attention for image captioning and VQA. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6077–6086. [Google Scholar]
  13. Chen, K.; Wang, J.; Chen, L.-C.; Gao, H.; Xu, W.; Nevatia, R. ABC-CNN: An attention based convolutional neural network for visual question answering. arXiv 2015, arXiv:1511.05960. [Google Scholar]
  14. Ilievski, I.; Yan, S.; Feng, J. A focused dynamic attention model for visual question answering. arXiv 2016, arXiv:1604.01485. [Google Scholar]
  15. Xu, H.; Saenko, K. Ask, attend and answer: Exploring question-guided spatial attention for visual question answering. In Proceedings of the 14th European Conference on Computer Vision, ECCV, Amsterdam, The Netherlands, 8–16 October 2016; pp. 451–466. [Google Scholar] [CrossRef] [Green Version]
  16. Yang, Z.; He, X.; Gao, J.; Smola, A. Stacked attention networks for image question answering. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Las Vegas, NE, USA, 27–30 June 2016; pp. 21–29. [Google Scholar] [CrossRef] [Green Version]
  17. Lu, J.; Yang, J.; Batra, D.; Parikh, D. Hierarchical question-image co-attention for visual question answering. In Proceedings of the 30th Conference on Neural Information Processing Systems, NIPS, Barcelona, Spain, 5–10 December 2016; pp. 289–297. [Google Scholar]
  18. Lao, M.; Guo, Y.; Wang, H.; Zhang, X. Cross-modal multistep fusion network with co-attention for visual question answering. IEEE Access 2018, 6, 31516–31524. [Google Scholar] [CrossRef]
  19. Nam, H.; Ha, J.W.; Kim, J. Dual attention networks for multimodal reasoning and matching. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Honolulu, HI, USA, 22–25 July 2017; pp. 299–307. [Google Scholar] [CrossRef] [Green Version]
  20. Yu, D.; Fu, J.; Mei, T.; Rui, Y. Multi-level attention networks for visual question answering. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Honolulu, HI, USA, 22–25 July 2017; pp. 4709–4717. [Google Scholar] [CrossRef]
  21. Yang, C.; Jiang, M.; Jiang, B.; Zhou, W.; Li, K. Co-attention network with question type for visual question answering. IEEE Access 2019, 7, 40771–40781. [Google Scholar] [CrossRef]
  22. Idan, S.; Yu, S.; Hazan, T.; Schwing, A. Factor graph attention. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Long Beach, CA, USA, 16–20 June 2019; pp. 2039–2048. [Google Scholar] [CrossRef]
  23. Park, S.; Whang, T.; Yoon, Y.; Lim, H. Multi-view attention network for visual dialog. arXiv 2020, arXiv:2004.14025. [Google Scholar]
  24. Kirsch, L.; Kunze, J.; Barber, I. Modular networks: Learning to decompose neural computation. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS, Montreal, QC, Canada, 3–8 December 2018; pp. 2408–2418. [Google Scholar]
  25. Neelakantan, A.; Le, Q.V.; Sutskever, I. Neural programmer: Inducing latent programs with gradient descent. In Proceedings of the 4th International Conference on Learning Representations, ICLR, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  26. Andreas, J.; Rohrbach, M.; Darrell, T.; Klein, D. Neural module networks. In Proceedings of the IEEE Computer Vision and Pattern Recognition, CVPR, IEEE Xplore, Las Vegas, NE, USA, 27–30 June 2016; pp. 39–48. [Google Scholar] [CrossRef]
  27. Hu, R.; Andreas, J.; Rohrbach, M.; Darell, T.; Saenko, K. Learning to reason: End-to-end module networks for visual question answering. In Proceedings of the IEEE International Conference on Computer Vision ICCV, IEEE Xplore, Venice, Italy, 22–29 October 2017; pp. 804–813. [Google Scholar] [CrossRef] [Green Version]
  28. Hu, R.; Andreas, J.; Darrell, T.; Saenko, K. Explainable neural computation via stack neural module networks. In Proceedings of the 15th European Conference on Computer Vision, ECCV, Munich, Germany, 8–14 September 2018; pp. 53–69. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Example of a visual dialog.
Figure 1. Example of a visual dialog.
Sensors 21 00931 g001
Figure 2. Overall structure of the proposed model.
Figure 2. Overall structure of the proposed model.
Sensors 21 00931 g002
Figure 3. Program generation.
Figure 3. Program generation.
Sensors 21 00931 g003
Figure 4. Program execution.
Figure 4. Program execution.
Sensors 21 00931 g004
Figure 5. Example of a generated neural program.
Figure 5. Example of a generated neural program.
Sensors 21 00931 g005
Figure 6. Reference pool and Refer module.
Figure 6. Reference pool and Refer module.
Sensors 21 00931 g006
Figure 7. Creating a program for processing impersonal pronouns.
Figure 7. Creating a program for processing impersonal pronouns.
Sensors 21 00931 g007
Figure 8. Example of minimum bounding area.
Figure 8. Example of minimum bounding area.
Sensors 21 00931 g008
Figure 9. Attention effect.
Figure 9. Attention effect.
Sensors 21 00931 g009
Figure 10. Dialog example 1 derived by the proposed model.
Figure 10. Dialog example 1 derived by the proposed model.
Sensors 21 00931 g010
Figure 11. Dialog example 2 derived by the proposed model.
Figure 11. Dialog example 2 derived by the proposed model.
Sensors 21 00931 g011
Figure 12. Dialog example 3 derived by the proposed model.
Figure 12. Dialog example 3 derived by the proposed model.
Sensors 21 00931 g012
Table 1. Neural network modules for visual dialog.
Table 1. Neural network modules for visual dialog.
ModuleInputsOutputFunction
Find x v i s ,   x t x t Attention y s = c o n v 1 ( x v i s x t x t )
y d = c o n v 2 ( ( ( s u m ( y s x v i s ) x t x t ) x t x t ) x v i s )
y t = c o n v 3 ( ( ( s u m ( y d x v i s ) x t x t ) x t x t ) x v i s )
Relocate x v i s ,   x t x t ,   a   Attention y = c o n v 2 ( c o n v 1 ( x v i s ) W 1 ( s u m ( a x v i s ) ) W 2 x t x t )
And a 1 ,   a 2 Attention y = m i n { a 1 , a 2 }
Refer x t x t ,   P r e f Attention y ˜ i = s o f t m a x ( W 1 ( [ x t x t ,   k i , i R D ] ) ) ,       y = i = 1 | P r e f | y ˜ i a i
Describe x v i s ,   x t x t ,   a Context y = W 3 T ( W 1 s u m ( a x v i s ) W 2 x t x t )
Compare x v i s ,   x t x t ,   a 1 ,   a 2 Context y ˜ = m i n ( T h r e s h o l d ( a 1 , a 2 ) )
y = W 3 T ( W 1 s u m ( y ˜ x v i s ) W 2 x t x t )
Table 2. Analysis of the attention effects of the Find module. MRR, mean reciprocal rank.
Table 2. Analysis of the attention effects of the Find module. MRR, mean reciprocal rank.
AttentionMRRR@1R@5R@10Mean
Single attention64.1050.9280.1888.814.53
Dual attention64.5051.3180.5289.044.35
Triple attention64.7051.5680.5789.074.32
Quad attention64.5051.4180.5088.994.35
Table 3. Performance comparison between two Refer modules.
Table 3. Performance comparison between two Refer modules.
DatasetRefer ModulesMRRR@1R@5R@10Mean
VisDialR v0.9 Refer 64.8051.6480.8789.284.29
Refer + Impersonality64.9051.7080.9489.354.27
VisDial v0.9 Refer 64.7051.5680.5789.074.32
Refer + Impersonality64.9051.7480.8889.294.30
Table 4. Performance comparison among Compare modules.
Table 4. Performance comparison among Compare modules.
Compare ModulesMRRR@1R@5R@10Mean
No compare64.9051.7480.8889.294.30
Compare (Inner Product)64.5051.2880.5088.984.36
Compare (Or)64.8051.6780.7989.244.30
Compare (Ours)64.9051.8280.8889.344.27
Table 5. Performance comparison with existing models.
Table 5. Performance comparison with existing models.
VisDial v1.0VisDial v0.9
ModelNDCGMRRR@1R@5R@10MeanMRRR@1R@5R@10Mean
LF [2]45.3155.4240.9572.4582.835.9558.0743.8274.6884.075.78
HRE [2]45.4654.1639.9370.4581.506.4158.6844.8274.8184.365.66
MN [2]47.5055.4940.9872.3083.305.9259.6545.5576.2285.375.46
HCIAE [7]------62.2248.4878.7587.594.81
AMEM [9] ------62.2748.5378.6687.434.86
CoAtt [10]------63.9850.2980.1888.814.47
CorefNMN [5]54.7061.5047.5578.1088.804.4064.1050.9280.1888.814.53
NMN-VD (Ours)56.9063.0049.2379.8188.894.3164.9051.8280.8889.344.27
LF, late fusion; HRE, hierarchical recurrent encoder; MN, memory network; HCIAE, history-conditioned image attention encoder; AMEM, at-tention memory; CoAtt, co-attention network; CorefNMN, coreference neural module network; NMN-VD, neural module network for visual dialog.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cho, Y.; Kim, I. NMN-VD: A Neural Module Network for Visual Dialog. Sensors 2021, 21, 931. https://doi.org/10.3390/s21030931

AMA Style

Cho Y, Kim I. NMN-VD: A Neural Module Network for Visual Dialog. Sensors. 2021; 21(3):931. https://doi.org/10.3390/s21030931

Chicago/Turabian Style

Cho, Yeongsu, and Incheol Kim. 2021. "NMN-VD: A Neural Module Network for Visual Dialog" Sensors 21, no. 3: 931. https://doi.org/10.3390/s21030931

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop