Next Article in Journal
Virtualized Microphone Array for Moving Sources Mapping
Previous Article in Journal
An Investigation of the Domain Gap in CLIP-Based Person Re-Identification
Previous Article in Special Issue
Depth Video-Based Secondary Action Recognition in Vehicles via Convolutional Neural Network and Bidirectional Long Short-Term Memory with Spatial Enhanced Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios

1
Department of Electronics and Electrical Engineering, Faculty of Science and Technology, Keio University, 3-14-1, Hiyoshi, Kohoku-ku, Yokohama 223-8522, Japan
2
National Institute of Advanced Industrial Science and Technology (AIST), 1-1-1 Umezono, Tsukuba 305-8560, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(2), 364; https://doi.org/10.3390/s25020364
Submission received: 21 November 2024 / Revised: 7 January 2025 / Accepted: 8 January 2025 / Published: 9 January 2025
(This article belongs to the Special Issue Multi-Sensor Systems for Object Tracking—2nd Edition)

Abstract

:
Traditional Vision-and-Language Navigation (VLN) tasks require an agent to navigate static environments using natural language instructions. However, real-world road conditions such as vehicle movements, traffic signal fluctuations, pedestrian activity, and weather variations are dynamic and continually changing. These factors significantly impact an agent’s decision-making ability, underscoring the limitations of current VLN models, which do not accurately reflect the complexities of real-world navigation. To bridge this gap, we propose a novel task called Dynamic Vision-and-Language Navigation (DynamicVLN), incorporating various dynamic scenarios to enhance the agent’s decision-making abilities and adaptability. By redefining the VLN task, we emphasize that a robust and generalizable agent should not rely solely on predefined instructions but must also demonstrate reasoning skills and adaptability to unforeseen events. Specifically, we have designed ten scenarios that simulate the challenges of dynamic navigation and developed a dedicated dataset of 11,261 instances using the CARLA simulator (ver.0.9.13) and large language model to provide realistic training conditions. Additionally, we introduce a baseline model that integrates advanced perception and decision-making modules, enabling effective navigation and interpretation of the complexities of dynamic road conditions. This model showcases the ability to follow natural language instructions while dynamically adapting to environmental cues. Our approach establishes a benchmark for developing agents capable of functioning in real-world, dynamic environments and extending beyond the limitations of static VLN tasks to more practical and versatile applications.

1. Introduction

The Vision-and-Language Navigation (VLN) [1] task requires an agent to be able to navigate in environments based on visual inputs and natural language instructions. This capability is crucial for various real-world applications, including household robots, caregiver assistance, navigation aids for visually impaired individuals, disaster area assessment, and delivery services. There is a strong societal demand for advancements in this technology, as it has the potential to transform everyday life and support critical operations. To support these applications, researchers have proposed a variety of VLN datasets that capture various navigation challenges, including Room-to-Room (R2R) [1] for indoor settings and Touchdown [2] for exploring outdoor urban navigation, as well as specialized datasets such as ALFRED [3] that introduce tasks that combine navigation with object interaction, further enriching the scope of VLN research.
However, these datasets were designed based on static environments, where the objects and layout remain unchanged, failing to encapsulate the unpredictable nature of the real world. The absence of dynamic elements such as moving cars, pedestrians, fluctuating traffic lights, and variable weather conditions limits the applicability of these datasets for preparing agents to navigate in environments that closely mimic daily scenarios.
To address the limitations of traditional Vision-and-Language Navigation (VLN) in handling real-world dynamic scenarios, we introduce the Dynamic Vision-and-Language Navigation (DynamicVLN) task. DynamicVLN builds upon the traditional VLN framework by introducing a temporal stop action to handle the unpredictability of dynamic obstacles encountered during navigation. In traditional VLN, navigation is inherently a discrete, instruction-driven task, where the agent follows a predefined path strictly dictated by navigation instructions. The agent navigates step-by-step on a graph-based map, selecting the next adjacent node based on the given instruction. DynamicVLN extends this framework by incorporating temporal stop points, allowing agents to pause momentarily when encountering dynamic elements before continuing along the predefined path. Specifically, DynamicVLN is structured around four key scenarios: vehicles, pedestrians, traffic signals, and weather conditions. Each of these elements introduces variability that requires agents to dynamically adjust their behavior, closely mirroring humans’ challenges in everyday navigation. In response to these elements, agents may encounter numerous situations where they must decide whether to perform a temporal stop or to continue according to instructions. As shown in Figure 1, at timestep T, while the instruction indicates “turn right”, the agent must instead briefly stop to avoid a collision with an oncoming vehicle, exemplifying the need for adaptive decision-making in dynamic environments. This scenario highlights the challenges addressed by DynamicVLN, which requires agents to respond to their surroundings while adhering to navigation instructions dynamically.
Scenarios like this are among the 10 types of dynamic variations incorporated into the DynamicVLN dataset, designed to evaluate agents’ adaptability across diverse real-world challenges. Specific scenarios requiring a temporal stop are systematically detailed in Section 3. For example, in vehicle-related scenarios, an agent must navigate sudden stops by cars or adjust its path in response to vehicles merging into its lane. Pedestrian scenarios test the agent’s ability to safely navigate around individuals crossing the street unexpectedly or moving in unpredictable patterns. In scenarios involving traffic signals, agents are required to interpret changes in traffic lights, making split-second decisions that ensure compliance with traffic laws while progressing toward their goal. Weather scenarios introduce visual and physical challenges, such as reduced visibility due to fog or altered road conditions caused by rain or snow, requiring agents to modify their navigation strategies to maintain safety and efficiency. We constructed DynamicVLN using the CARLA simulator [4] and automatically generated instructions with GPT-4 [5], resulting in a dataset of 11,261 unique navigation instances. Each of the 10 dynamic scenarios—vehicles, pedestrians, traffic signals, and weather conditions—is divided between cases requiring a temporal stop and those that do not. This balanced design ensures comprehensive coverage of dynamic change detection, providing a robust foundation for training and evaluating navigation systems in real-world-like environments.
Along with the DynamicVLN task, we introduced DynaNav, a baseline model designed to address its distinct challenges effectively. At the core of DynaNav is a dynamic detection module, which recognizes dynamic elements within the environment. This module enables DynaNav to discern when to execute a ‘temporal stop’ or proceed, ensuring effective and adaptive navigation in dynamic scenarios. In summary, the contribution of our work is four-fold:
  • We introduce Dynamic Vision-and-Language Navigation (DynamicVLN), a novel task that incorporates dynamic real-world scenarios such as moving vehicles, pedestrians, fluctuating traffic signals, and varying weather conditions, addressing the limitations of traditional static VLN tasks.
  • We constructed the DynamicVLN dataset, comprising 11,261 navigation instances across ten dynamic scenarios. Data collection was automated using the CARLA simulator, and captions were generated automatically by GPT-4, ensuring both realism and diversity.
  • We propose DynaNav, a baseline model equipped with a Dynamic Detection Module, enabling agents to recognize dynamic elements and make context-aware decisions, such as when to execute a ‘temporal stop’.

2. Related Works

2.1. Vision-and-Language Navigation Dataset

Vision-and-Language Navigation (VLN) tasks require agents to navigate environments using natural language instructions. Existing VLN datasets encompass a variety of scenarios. For indoor navigation, the Room-to-Room (R2R) dataset [1], based on Matterport3D [6], serves as a foundational benchmark. It has been extended by Room-Across-Room (RxR) [7] and XL-R2R [8], which include multilingual instructions. These datasets are designed for simple and structured navigation tasks. For outdoor navigation, the Touchdown dataset [2], utilizing Google Street View (https://developers.google.com/maps/documentation/streetview, accessed on 15 September 2024), provides a benchmark for navigating complex urban environments. Similarly, StreetLearn [9], Retouchdown [10], StreetNav [11], Talk2Nav [12], map2seq [13], and VLN-VIDEO [14] focus on urban navigation tasks, with VLN-VIDEO augmenting navigation performance using driving videos. In terms of object interaction, the ALFRED dataset [3], built upon AI2-THOR 2.0 [15], emphasizes complex tasks requiring agents to interact with objects while navigating indoor environments. For aerial navigation, the AerialVLN dataset [16], based on the AirSim (version 1.2.0) [17], introduces challenges that require agents to interpret instructions and navigate elevated perspectives. Some datasets focus on task-specific navigation; for instance, CARLA-NAV [18] explores the grounding of navigable regions corresponding to textual descriptions, while DOROTHIE [19] highlights dialogue-based navigation in dynamic environments. However, these datasets primarily address static or predictable dynamic cues and lack realistic, sudden changes commonly encountered in real-world traffic scenarios.
In contrast, DynamicVLN addresses this limitation by introducing sudden, unpredictable events—such as abrupt vehicle stops, unexpected pedestrian crossings, and changing traffic signals—into the navigation process. It preserves the discrete, instruction-driven structure of traditional VLN tasks, ensuring agents’ decision-making remains interpretable and consistent with existing VLN frameworks. Additionally, the integration of temporal stop actions allows agents to effectively respond to real-world dynamic challenges without deviating from predefined paths. (The key differences and contributions of DynamicVLN compared to other datasets are systematically summarized in Table 1).

2.2. Approach for Vision-and-Language Navigation

With the introduction of numerous Vision-and-Language Navigation (VLN) benchmarks, researchers have developed various methods to enhance navigation performance. Initially, many approaches employed sequence-to-sequence (seq2seq) models, integrating images and instructions into pre-trained models [21,22,23], which have great understanding for images and instructions, then utilizing cross-modal attention mechanisms for action prediction [1,24]. Transformer-based models have further advanced VLN performance. ORIST [25] introduced object-level and word-level inputs to learn fine-grained relationships across textual and visual modalities, enabling more precise decision-making. VLN-BERT [26] extended this approach by developing a recurrent vision-and-language BERT, which incorporates historical states to enhance sequential decision-making in navigation tasks. SOTA [27] proposed a scene- and object-aware transformer, emphasizing context-specific understanding by focusing on relevant objects and environmental details. Given the challenges of environmental understanding in VLN tasks, researchers have also incorporated fine-grained attention mechanisms to improve agents’ comprehension of their surroundings. For example, previous work [28,29] leveraged landmarks along the route to divide the navigation path into smaller segments, using these landmarks as references to enhance navigation performance by providing clear intermediate goals. Recently, LLM-based approaches have emerged as a promising direction in VLN. For instance, VELMA [30] employs large language models to generate natural language explanations, helping agents reason about their navigation steps. VirtuWander [31] introduced a method for augmenting VLN tasks with detailed descriptions generated by LLMs, improving interpretability and task success. MapGPT [32] integrates GPT-based language generation to facilitate map-based navigation, showcasing the potential of LLMs to handle complex, multimodal navigation tasks.

2.3. Large Language Model for Dataset Generation

Recent advancements in large language models (LLMs) have significantly enhanced data annotation processes. Traditional methods often involve manual labeling, which is time-consuming and costly. Applications of synthetic data generation have garnered significant attention due to their remarkable capability to understand and generate human-like text based on input prompts and a few examples [33]. Early works utilized LLMs primarily for generating textual data, particularly in the natural language processing (NLP) field, enabling tasks such as text classification [34], summarization [35], and translation with minimal human intervention [36]. Building on this foundation, LLMs have also been applied in multimodal contexts. For instance, LLaVA [37] demonstrated how language-only models like GPT-4 could generate multimodal instruction-following data by combining vision encoders with LLMs to create datasets for visual understanding tasks. Moreover, LLMs have been leveraged to augment existing datasets to boost task performance. AttrPrompt [38] utilized LLMs to enhance attribute-based reasoning tasks, while other studies employed LLMs to expand datasets with diverse, high-quality examples [39]. Beyond general-purpose datasets, there has been a growing focus on using LLMs to generate domain-specific datasets [40,41].
In this work, we design a multistage workflow to generate high-quality navigation instructions for DynamicVLN automatically. To address the hallucination issues, we utilize two LLMs in our workflow. One LLM generates the instructions, while the other supervises and corrects the outputs, ensuring the realism, diversity, and accuracy of the dataset.

3. DynamicVLN Dataset

3.1. Task Definition

DynamicVLN is a task that challenges an agent to navigate through an environment based on natural language instructions X = { x 1 , x 2 , , x l } while dynamically adapting to changes within that environment to reach a specified target location. The instructions consist of a sequence of l word tokens, each represented by x i . The environment is structured as an undirected graph, with nodes v V representing specific locations connected by labeled edges ( v , u ) E , where u denotes an adjacent location to v. Each node is linked to an RGB image, providing visual context, and edges denote possible navigation paths with heading angles α ( v , u ) between images.
At any given time t, the agent’s state is s t S , s t = ( v t , α ( v t 1 , v t ) ) , incorporating the current panoramic view v t and the heading angle α ( v t 1 , v t ) from the previous state to the current one. To navigate, the agent performs actions from the set a t { forward , left , right , temporalstop , stop } , where the temporal stop action is introduced to allow brief pauses in response to dynamic events such as moving pedestrians or vehicles.
The objective in DynamicVLN is for the agent to generate a sequence of state-action pairs s 1 , a 1 , s 2 , a 2 , , s n , a n , culminating in a a n = STOP action that indicates the goal location has been reached, as defined by the instructions. Including the temporal stop action enhances the agent’s ability to navigate more effectively by adapting to real-time changes in the environment, thus making DynamicVLN a more realistic and challenging task that mirrors the complexities of navigating dynamic, real-world scenarios.

3.2. Scenario Design

We designed four dynamic element types to realistically simulate complex navigation environments: vehicles, pedestrians, traffic signals, and weather. Each element introduces specific trigger conditions that require agents to make adaptive decisions while maintaining alignment with the discrete, instruction-driven structure of VLN tasks.
For instance, these scenarios often necessitate a temporal stop when agents encounter immediate risks or obstructions—such as yielding to cross-traffic, halting for a pedestrian crossing, or stopping in response to a traffic signal change. These trigger conditions and their corresponding decision logic are systematically summarized in Table 2, which highlights how each dynamic element aligns with temporal stop or continue actions based on specific environmental cues. Figure 2 shows examples in DynamicVLN when the vehicle needs to temporarily stop at timestep T during driving for each type of dynamic element.

3.3. Dataset Collection

We designed driving scenarios using the open source CARLA simulator [4] to simulate realistic driving conditions and collect images and driving data on vehicle navigation. Specifically, we utilized CARLA’s pre-defined town maps (Town 1, 2, 3, 4, 5, 6, and 10 HD) as our environment. In each map, we defined waypoints spaced 5 m apart to serve as navigable points along the routes. We generated random traffic flows, including vehicles and pedestrians, to simulate real-world traffic conditions and create dynamic scenarios. Initially, we select each defined waypoint as a start waypoint to create static routes. Each path spans 30–50 waypoints and includes at least one turning intersection, mimicking the design of traditional VLN datasets. These static paths serve as the baseline for our dataset. We introduced dynamic elements to construct dynamic routes based on these static routes, simulating real-world driving conditions with unpredictable events.
Specifically, we defined a vehicle automatically driving on the routes and set up a camera sensor on the car to capture images during driving behavior. We collected the actions (forward, right, left, temporal stop, stop) and landmarks for each waypoint on the routes. We collected the actions according to the route yaw changes and vehicle speed. The landmarks are mainly about traffic signs.

3.4. Instruction Generation

Based on the data collected from CARLA, we proposed a pipeline to generate high-quality navigation instructions using GPT-4o automatically. In this work, we aimed to train the agent to adapt to sudden events, so we did not add the temporal stop action in the instructions. This pipeline employs two LLM agents: an Instruction Generator and an Instruction Supervisor. As shown in Figure 3, the pipeline begins by feeding the route overview, the sequence of actions at each waypoint, and the landmarks encountered along the route into the Instruction Generator. The generator produces an initial navigation instruction formatted similarly to traditional VLN instructions. Next, the generated instruction and a simplified action list are passed to the Instruction Supervisor. The supervisor evaluates the alignment between the simplified action list and the generated instruction, checking for missing actions, incorrect sequencing, or unnecessary additions. Based on its analysis, the supervisor refines the instruction to ensure accuracy and consistency with the action list. This iterative process ensures that the final instruction is coherent, is aligned with the route’s actions, and accurately reflects the dynamic elements of the environment.
Specifically, in Figure 4, we provide the prompts used by the Instruction Generator and the Instruction Supervisor during the instruction creation process. The Instruction Generator receives structured input detailing the route overview, actions, and landmarks, generating an initial instruction. The Instruction Supervisor takes the simplified action list and the initial instruction as input and performs a consistency check, identifying necessary corrections before refining the final instruction.

3.5. Data Statistics

Following the pipeline introduced above, we collected a total of 11,261 routes. Among these, 2786 routes are associated with dynamic elements of vehicles, 1680 with pedestrian activities, 3370 with traffic conditions, and 601 with weather-related scenarios. These represent various dynamic factors encountered during navigation. However, the occurrence of a dynamic factor does not necessarily require the vehicle to stop temporally. In many cases, vehicles may need to stop at multiple points along the route before they can proceed, rather than at a single temporal stop. To better understand these dynamics, we analyzed the distribution of temporal stops within these routes. Each dynamic route involves between two and seven temporal stop actions, reflecting the complexity of real-world scenarios where vehicles must repeatedly adjust to changing conditions. Furthermore, each dynamic instance has 2–7 temporal stops. Figure 5 illustrates the distribution of routes based on the number of temporal stops they contain, providing insight into the frequency and variation of such actions across all collected routes.

4. Method

In this section, we introduce the DynaVLN model, designed to detect dynamic events and improve navigation performance in dynamic environments, building upon traditional VLN models. As illustrated in Figure 6, DynaVLN consists of four key components: the Image Encoder, Instruction Encoder, Dynamic Event Detector, and Action Predictor. At each decoding timestep, DynaVLN computes a visual representation of the agent’s current and previous states in the environment, integrating previously predicted actions, instruction features, and dynamic event detection results to predict the next action.

4.1. Model Details

Image Encoder. At each timestep t, the agent captures an image view of its surroundings. The visual representation of the current agent position is computed by extracting features from the panorama using a pre-trained CLIP Vision Encoder [42]. This step provides a robust and compact visual embedding v t for navigating and detecting changes during the driving process.
Instruction Encoder. The instruction encoder processes the natural language navigation instructions x = { x 1 , x 2 , , x L } , where L denotes the number of tokens in the instruction sequence. Each token x i is embedded and encoded using a bidirectional LSTM [43]:
x ^ i = embedding ( x i )
( w 1 , w 2 , , w L ) , z L w = Bi-LSTM ( x ^ 1 , x ^ 2 , , x ^ L ) ,
where w i represents the hidden representation of token x i , and z L w is the final cell state of the LSTM. These outputs capture both local (token-level) and global (sequence-level) contextual information from the navigation instructions.
Dynamic Event Detector. To enable the agent to handle unforeseen dynamic events without explicit instructions intelligently, we incorporate a Dynamic Event Detector (DED) and propose an Attention Modulation Mechanism that dynamically adjusts the model’s focus between visual and language features based on the detected event. This mechanism allows the model to autonomously decide actions, including temporal stop, to address sudden changes in the environment.
At each timestep t, the DED processes the current visual feature v t and the previous visual feature v t 1 to compute the event signal e t , representing the likelihood of a dynamic event. Formally, the event signal is defined as follows:
e t = σ ( MLP ( [ v t v t 1 ] ) )
where σ ( · ) is the sigmoid activation, and MLP denotes a lightweight multi-layer perceptron. e t [ 0 , 1 ] indicates the intensity of the detected dynamic event.
The event signal e t is then used to modulate the attention mechanism, dynamically altering the interaction between the visual feature v t and the instruction feature sequence { w 1 , , w L } . The modulation affects both the Query ( q t ) and the Key/Value representations ( k t , v k ) as follows:
q t = Linear q ( [ v t ; e t ] )
k t , v k = Linear k ( [ w 1 , , w L ] ) · ( 1 e t ) , Linear v ( [ w 1 , , w L ] ) · ( 1 e t )
where [ v t ; e t ] denotes the concatenation of the visual feature and the event signal. The Key and Value representations are scaled by ( 1 e t ) , reducing the influence of language instructions when a significant dynamic event is detected ( e t 1 ).
The attention output f t is computed via the scaled dot-product attention mechanism:
Attention ( q t , k t , v k ) = softmax q t k t d k v k
where d k is the dimensionality of the Key vectors. The resulting feature f t incorporates the adjusted contribution of visual and language features, dynamically weighted by the presence or absence of a detected event.
Action Predictor. Finally, the attention output f t is concatenated with the event signal e t and passed to the Action Predictor to generate the next action:
a t = ActionPredictor ( [ f t ; e t ] , a t 1 )
where a t is the predicted action (e.g., move forward, turn left, or stop) and a t 1 is the previous action.

4.2. Loss Function

To train the proposed model effectively, we design a loss function that jointly optimizes the performance of the Dynamic Event Detector (DED) and the Action Predictor.
Dynamic Event Detector Loss. The DED outputs an event signal e t [ 0 , 1 ] , representing the probability of a dynamic event occurring at timestep t. The ground truth event label y t { 0 , 1 } indicates whether a dynamic event is present. The detection loss is formulated as a binary cross-entropy loss:
L DED = 1 N t = 1 N y t log ( e t ) + ( 1 y t ) log ( 1 e t ) ,
where N is the total number of samples. This loss ensures that the DED module learns to predict event probabilities that align with the ground truth labels.
Action Prediction Loss. The Action Predictor generates the probability distribution over possible actions { a 1 , a 2 , , a C } , where C is the number of action classes (e.g., move forward, turn left, stop). The ground truth action label is denoted as a t true { 1 , , C } . We use a categorical cross-entropy loss to optimize the predicted action distribution:
L Action = 1 N t = 1 N i = 1 C 1 a t true = i log P a t = i f t , e t , a t 1 ,
where P ( a t = i | f t , e t , a t 1 ) is the predicted probability of action i, and 1 [ · ] is the indicator function.
Joint Loss. To jointly optimize both components of the model, we combine the above losses into a single objective:
L = λ 1 L DED + λ 2 L Action ,
where λ 1 and λ 2 are balancing weights that control the relative importance of the two loss terms.

5. Experiments

This section presents extensive experiments on our DynamicVLN dataset to evaluate our DynaVLN models’ performance on the outdoor VLN task and emergency accident adaptation.

5.1. Implementation Details

Data Processing. We adopted a clustering-based approach to group spatially proximate waypoints into unified nodes to construct a node-based navigation graph for each town map for Vision-and-Language Navigation (VLN) tasks. Specifically, all waypoints extracted from the routes, including those used during navigation, were collected and clustered based on their spatial proximity. For clustering, we utilized the DBSCAN algorithm [44], which is effective for identifying arbitrarily shaped clusters without requiring a predefined number of clusters. Each waypoint was treated as a three-dimensional point ( x , y , z ) in Cartesian coordinates, and waypoints within a distance threshold ( ϵ = 0.5 m) were grouped into the same cluster. After clustering, each cluster was assigned a unique node ID. The navigation graph was then constructed by treating each cluster as a node and connecting nodes adjacent to the same route.
Model Details. The proposed framework and baseline models were implemented using PyTorch (version 2.0.1) (https://pytorch.org/, accessed on 1 October 2024). We employed a pre-trained CLIP vision encoder with a ViT backbone to extract visual features with a size of 512 from images at each waypoint. Navigation instructions were tokenized into byte pair encodings (BPE) using a vocabulary size of 2000 tokens. The instruction tokens were lower-cased and embedded into vectors of size 256. For the loss function, we balanced the objectives of navigation performance and dynamic event detection using weights λ 1 = 0.5 and λ 2 = 1.0 . This prioritization ensures the agent achieves reliable navigation while accurately detecting dynamic events. In scenarios where ground truth event labels y t are unavailable, pseudo-labels were generated heuristically. These heuristics included abrupt changes in the visual scene (e.g., the appearance of obstacles) or significant deviations in the agent’s planned actions (e.g., an unexpected stop or turn). This allows the model to handle dynamic events in complex environments adaptively. Since each town map is unique, it was necessary to construct a separate navigation graph for each town and conduct training individually for each map. However, in this section, we only present experimental results using the Town05 map. Town05 was selected because it contains the largest number of routes, with a total of 2611 routes. These routes were randomly split into training, development, and test sets in a 7:2:1 ratio, resulting in 1828 routes for training, 522 for development, and 261 for testing.

5.2. Baseline Models

ORAR [45] is a VLN model proposed for outdoor VLN tasks. This model uses an LSTM to encode the instruction text and an LSTM to decode the multimodal features and predict the following action. We selected ORAR because it is a model without collision detection and only uses whole images and instruction features for decision-making. By comparing our results with ORAR, we demonstrated that collision detection benefits adaptability to unforeseen events.

5.3. Metrics

The following metrics are used to evaluate VLN performance: (1) Task Completion (TC): the accuracy of navigating to the correct location. The correct location is defined as the exact goal panorama or one of its neighboring panoramas. (2) Shortest-Path Distance (SPD) [46]: the mean distance between the agent’s final position and the goal position in the environment graph. (3) Success Weighted by Edit Distance (SED): the normalized Levenshtein edit distance [47] between the predicted path and the ground truth path, with points only awarded for successful paths. (4) Coverage Weighted by Length Score (CLS) [48]: a measurement of the fidelity of the agent’s path to the ground truth path. (5) Normalized Dynamic Time Warping (nDTW) [49]: the minimized cumulative distance between the predicted path and the ground truth path. (6) Success Weighted Dynamic Time Warping (SDTW): the nDTW value where the summation is only over successful navigation.

5.4. Quantitative Results

We provide a comprehensive evaluation of the experimental results to assess DynaVLN’s performance in comparison to the baseline ORAR across various metrics. These results highlight the proposed model’s strengths and limitations, particularly in handling dynamic navigation environments. The experimental results are summarized in Table 3.
DynaVLN demonstrates higher TC, indicating its ability to navigate more effectively and reach target destinations more reliably. In terms of SPD, DynaVLN maintains a closer adherence to the target trajectory, reflecting its capability to navigate with greater precision. For SED, DynaVLN achieves better alignment with the ground truth action sequences, highlighting its accuracy in predicting correct actions during navigation. Conversely, ORAR achieves a higher CLS, suggesting stronger coverage of the navigation path but potentially less responsiveness to dynamic changes. Regarding temporal alignment, ORAR performs better in static environments, as indicated by its nDTW score. However, DynaVLN excels in dynamic scenarios, with a significantly higher sDTW score, showcasing its ability to adapt and succeed in environments with dynamic changes. Overall, these results suggest that while ORAR is well suited for structured and static environments, DynaVLN offers superior adaptability and robustness in dynamic and complex real-world navigation tasks.

6. Conclusions

In this paper, we addressed the challenge of navigating dynamic environments by introducing the Dynamic Vision-and-Language Navigation (DynamicVLN) task. Our work recognizes the limitations of traditional VLN tasks, which largely focus on static environments, and proposes a more realistic dataset that incorporates the unpredictability of real-world conditions.
Looking ahead, we plan to refine and perfect the DynamicVLN task. We will focus on developing the DynaNav model to meet the unique challenges presented by dynamic elements. DynaNav will incorporate advanced real-time decision-making techniques, leveraging visual and linguistic inputs to navigate effectively and safely.

Author Contributions

Y.S. proposed and implemented the approaches and conducted the experiments. She also wrote the paper together with Y.Q. and Y.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JST SPRING, Grant Number JPMJSP2123.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting the findings of this study are available from the corresponding author upon reasonable request. Additionally, a subset of our dataset is publicly available on GitHub at https://github.com/AuderySun/DynamicVLN, accessed on 10 November 2024. For further inquiries or access to the full dataset, please contact the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
VLNVision-and-Language Navigation
LLMLarge Language Model
GPTGenerative Pre-Trained Transformers
LDLinear Dichroism

References

  1. Anderson, P.; Wu, Q.; Teney, D.; Bruce, J.; Johnson, M.; Sünderhauf, N.; Reid, I.; Gould, S.; van den Hengel, A. Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  2. Chen, H.; Suhr, A.; Misra, D.; Snavely, N.; Artzi, Y. TOUCHDOWN: Natural Language Navigation and Spatial Reasoning in Visual Street Environments. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar] [CrossRef]
  3. Shridhar, M.; Thomason, J.; Gordon, D.; Bisk, Y.; Han, W.; Mottaghi, R.; Zettlemoyer, L.; Fox, D. ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  4. Dosovitskiy, A.; Ros, G.; Codevilla, F.; Lopez, A.; Koltun, V. CARLA: An Open Urban Driving Simulator. In Proceedings of the 1st Annual Conference on Robot Learning, Mountain View, CA, USA, 13–15 November 2017; pp. 1–16. [Google Scholar]
  5. OpenAI; Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; et al. GPT-4 Technical Report. arXiv 2024, arXiv:2303.08774. [Google Scholar]
  6. Chang, A.; Dai, A.; Funkhouser, T.; Halber, M.; Niessner, M.; Savva, M.; Song, S.; Zeng, A.; Zhang, Y. Matterport3D: Learning from RGB-D Data in Indoor Environments. arXiv 2017, arXiv:1709.06158. [Google Scholar]
  7. Ku, A.; Anderson, P.; Patel, R.; Ie, E.; Baldridge, J. Room-Across-Room: Multilingual Vision-and-Language Navigation with Dense Spatiotemporal Grounding. In Proceedings of the EMNLP, Online, 16–20 November 2020. [Google Scholar]
  8. Yan, A.; Wang, X.E.; Feng, J.; Li, L.; Wang, W.Y. Cross-Lingual Vision-Language Navigation. arXiv 2019, arXiv:1910.11301. [Google Scholar]
  9. Mirowski, P.; Grimes, M.K.; Malinowski, M.; Hermann, K.M.; Anderson, K.; Teplyashin, D.; Simonyan, K.; Kavukcuoglu, K.; Zisserman, A.; Hadsell, R. Learning to Navigate in Cities Without a Map. In Proceedings of the NeurIPS, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  10. Mehta, H.; Artzi, Y.; Baldridge, J.; Ie, E.; Mirowski, P. Retouchdown: Releasing Touchdown on StreetLearn as a Public Resource for Language Grounding Tasks in Street View. In Proceedings of the EMNLP-SpLU, Online, 16–20 November 2020. [Google Scholar] [CrossRef]
  11. Hermann, K.; Malinowski, M.; Mirowski, P.; Banki-Horvath, A.; Anderson, K.; Hadsell, R. Learning to Follow Directions in Street View. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34. [Google Scholar] [CrossRef]
  12. Vasudevan, A.B.; Dai, D.; Van Gool, L. Talk2Nav: Long-Range Vision-and-Language Navigation with Dual Attention and Spatial Memory. Int. J. Comput. Vis. 2021, 129, 246–266. [Google Scholar] [CrossRef]
  13. Schumann, R.; Riezler, S. Generating Landmark Navigation Instructions from Maps as a Graph-to-Text Problem. In Proceedings of the ACL, Online, 5–10 July 2021. [Google Scholar]
  14. Li, J.; Padmakumar, A.; Sukhatme, G.; Bansal, M. VLN-Video: Utilizing Driving Videos for Outdoor Vision-and-Language Navigation. arXiv 2024, arXiv:2402.03561. [Google Scholar]
  15. Kolve, E.; Mottaghi, R.; Han, W.; VanderBilt, E.; Weihs, L.; Herrasti, A.; Gordon, D.; Zhu, Y.; Gupta, A.; Farhadi, A. AI2-THOR: An Interactive 3D Environment for Visual AI. arXiv 2017, arXiv:1712.05474. [Google Scholar]
  16. Liu, S.; Zhang, H.; Qi, Y.; Wang, P.; Zhang, Y.; Wu, Q. AerialVLN: Vision-and-language Navigation for UAVs. In Proceedings of the International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023. [Google Scholar]
  17. Shah, S.; Dey, D.; Lovett, C.; Kapoor, A. AirSim: High-Fidelity Visual and Physical Simulation for Autonomous Vehicles. In Proceedings of the Field and Service Robotics. arXiv 2017, arXiv:1705.05065. [Google Scholar]
  18. Jain, K.; Chhangani, V.; Tiwari, A.; Krishna, K.M.; Gandhi, V. Ground then Navigate: Language-guided Navigation in Dynamic Scenes. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation, London, UK, 29 May–2 June 2023. [Google Scholar]
  19. Ma, Z.; VanDerPloeg, B.; Bara, C.P.; Huang, Y.; Kim, E.I.; Gervits, F.; Marge, M.; Chai, J. DOROTHIE: Spoken Dialogue for Handling Unexpected Situations in Interactive Autonomous Driving Agents. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 4800–4822. [Google Scholar]
  20. Krantz, J.; Wijmans, E.; Majundar, A.; Batra, D.; Lee, S. Beyond the Nav-Graph: Vision and Language Navigation in Continuous Environments. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  21. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  24. Fried, D.; Hu, R.; Cirik, V.; Rohrbach, A.; Andreas, J.; Morency, L.P.; Berg-Kirkpatrick, T.; Saenko, K.; Klein, D.; Darrell, T. Speaker-Follower Models for Vision-and-Language Navigation. In Proceedings of the Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  25. Qi, Y.; Pan, Z.; Hong, Y.; Yang, M.; van den Hengel, A.; Wu, Q. The Road to Know-Where: An Object-and-Room Informed Sequential BERT for Indoor Vision-Language Navigation. In Proceedings of the ICCV, Montreal, QC, Canada, 10–17 October 2021; pp. 1655–1664. [Google Scholar]
  26. Hong, Y.; Wu, Q.; Qi, Y.; Rodriguez-Opazo, C.; Gould, S. A Recurrent Vision-and-Language BERT for Navigation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 11–15 June 2021; pp. 1643–1653. [Google Scholar]
  27. Moudgil, A.; Majumdar, A.; Agrawal, H.; Lee, S.; Batra, D. SOAT: A Scene- and Object-Aware Transformer for Vision-and-Language Navigation. arXiv 2021, arXiv:2110.14143. [Google Scholar]
  28. Sun, Y.; Qiu, Y.; Aoki, Y.; Kataoka, H. Outdoor Vision-and-Language Navigation Needs Object-Level Alignment. Sensors 2023, 23, 6028. [Google Scholar] [CrossRef] [PubMed]
  29. Sun, Y.; Qiu, Y.; Aoki, Y.; Kataoka, H. Guided by the Way: The Role of On-the-route Objects and Scene Text in Enhancing Outdoor Navigation. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 5198–5204. [Google Scholar] [CrossRef]
  30. Schumann, R.; Zhu, W.; Feng, W.; Fu, T.J.; Riezler, S.; Wang, W.Y. VELMA: Verbalization Embodiment of LLM Agents for Vision and Language Navigation in Street View. arXiv 2023, arXiv:2307.06082. [Google Scholar]
  31. Wang, Z.; Yuan, L.P.; Wang, L.; Jiang, B.; Zeng, W. VirtuWander: Enhancing Multi-modal Interaction for Virtual Tour Guidance through Large Language Models. In Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24), New York, NY, USA, 11–16 May 2024. [Google Scholar] [CrossRef]
  32. Chen, J.; Lin, B.; Xu, R.; Chai, Z.; Liang, X.; Wong, K.Y. MapGPT: Map-Guided Prompting with Adaptive Path Planning for Vision-and-Language Navigation. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; pp. 9796–9810. [Google Scholar] [CrossRef]
  33. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2020; Volume 33, pp. 1877–1901. [Google Scholar]
  34. Sun, X.; Li, X.; Li, J.; Wu, F.; Guo, S.; Zhang, T.; Wang, G. Text Classification via Large Language Models. arXiv 2023, arXiv:2305.08377. [Google Scholar]
  35. Wang, Y.; Zhang, Z.; Wang, R. Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought Method. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 8640–8665. [Google Scholar] [CrossRef]
  36. Zhu, W.; Liu, H.; Dong, Q.; Xu, J.; Huang, S.; Kong, L.; Chen, J.; Li, L. Multilingual Machine Translation with Large Language Models: Empirical Results and Analysis. arXiv 2023, arXiv:2304.04675. [Google Scholar]
  37. Liu, H.; Li, C.; Wu, Q.; Lee, Y.J. Visual Instruction Tuning. In Proceedings of the Advances in Neural Information Processing Systems, Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  38. Yu, Y.; Zhuang, Y.; Zhang, J.; Meng, Y.; Ratner, A.; Krishna, R.; Shen, J.; Zhang, C. Large language model as attributed training data generator: A tale of diversity and bias. In Proceedings of the 37th International Conference on Neural Information Processing Systems (NIPS ’23), Red Hook, NY, USA, 10–16 December 2023. [Google Scholar]
  39. Chung, J.; Kamar, E.; Amershi, S. Increasing Diversity While Maintaining Accuracy: Text Data Generation with Large Language Models and Human Interventions. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 575–593. [Google Scholar] [CrossRef]
  40. Seo, M.; Baek, J.; Thorne, J.; Hwang, S.J. Retrieval-Augmented Data Augmentation for Low-Resource Domain Tasks. arXiv 2024, arXiv:2402.13482. [Google Scholar]
  41. Lee, D.H.; Pujara, J.; Sewak, M.; White, R.W.; Jauhar, S.K. Making Large Language Models Better Data Creators. In Proceedings of the The 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), Singapore, 6–10 December 2023. [Google Scholar]
  42. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. arXiv 2021, arXiv:2103.00020. [Google Scholar]
  43. Graves, A.; Fernández, S.; Schmidhuber, J. Bidirectional LSTM Networks for Improved Phoneme Classification and Recognition. In Proceedings of the Artificial Neural Networks: Formal Models and Their Applications—ICANN 2005, Warsaw, Poland, 11–15 September 2005. [Google Scholar]
  44. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD’96), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  45. Schumann, R.; Riezler, S. Analyzing Generalization of Vision and Language Navigation to Unseen Outdoor Areas. In Proceedings of the ACL, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
  46. Rizi, F.S.; Schloetterer, J.; Granitzer, M. Shortest Path Distance Approximation Using Deep Learning Techniques. In Proceedings of the 2018 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), Barcelona, Spain, 28–31 August 2018; pp. 1007–1014. [Google Scholar] [CrossRef]
  47. Miller, F.P.; Vandome, A.F.; McBrewster, J. Levenshtein Distance: Information Theory, Computer Science, String (Computer Science), String Metric, Damerau? Levenshtein Distance, Spell Checker, Hamming Distance; Alpha Press: Orlando, FL, USA, 2009. [Google Scholar]
  48. Jain, V.; Magalhaes, G.; Ku, A.; Vaswani, A.; Ie, E.; Baldridge, J. Stay on the Path: Instruction Fidelity in Vision-and-Language Navigation. arXiv 2019, arXiv:1905.12255. [Google Scholar]
  49. Ilharco, G.; Jain, V.; Ku, A.; Ie, E.; Baldridge, J. General Evaluation for Instruction Conditioned Navigation using Dynamic Time Warping. arXiv 2019, arXiv:1907.05446. [Google Scholar]
Figure 1. In traditional VLN tasks, agents predict actions based only on instructions, without accounting for real-time environmental changes. In Dynamic VLN tasks, however, agents must consider both instructions and dynamic elements, such as moving vehicles. For example, although the instruction here directs the agent to ‘turn right’, the agent must temporarily stop to yield to an oncoming car, adapting its actions to avoid a potential accident.
Figure 1. In traditional VLN tasks, agents predict actions based only on instructions, without accounting for real-time environmental changes. In Dynamic VLN tasks, however, agents must consider both instructions and dynamic elements, such as moving vehicles. For example, although the instruction here directs the agent to ‘turn right’, the agent must temporarily stop to yield to an oncoming car, adapting its actions to avoid a potential accident.
Sensors 25 00364 g001
Figure 2. Example of temporal stop under each dynamic element type setting.
Figure 2. Example of temporal stop under each dynamic element type setting.
Sensors 25 00364 g002
Figure 3. Pipeline of instruction generation for DynamicVLN, and this example shows an emergency scenario with oncoming cars.
Figure 3. Pipeline of instruction generation for DynamicVLN, and this example shows an emergency scenario with oncoming cars.
Sensors 25 00364 g003
Figure 4. Prompts used in the instruction generation pipeline. The (a) shows the generation process, while the (b) illustrates the supervision and refinement process. (a) The Instruction Generator processes the route overview, action list, and landmarks to generate an initial navigation instruction. (b) The Instruction Supervisor refines the initial instruction by ensuring alignment with the simplified action list and correcting any discrepancies.
Figure 4. Prompts used in the instruction generation pipeline. The (a) shows the generation process, while the (b) illustrates the supervision and refinement process. (a) The Instruction Generator processes the route overview, action list, and landmarks to generate an initial navigation instruction. (b) The Instruction Supervisor refines the initial instruction by ensuring alignment with the simplified action list and correcting any discrepancies.
Sensors 25 00364 g004
Figure 5. Distribution of routes based on the number of temporal stops. This figure shows the frequency of routes containing different numbers of temporal stops, reflecting the complexity and variability of dynamic scenarios in the collected dataset.
Figure 5. Distribution of routes based on the number of temporal stops. This figure shows the frequency of routes containing different numbers of temporal stops, reflecting the complexity and variability of dynamic scenarios in the collected dataset.
Sensors 25 00364 g005
Figure 6. Overview for proposed model DynaVLN.
Figure 6. Overview for proposed model DynaVLN.
Sensors 25 00364 g006
Table 1. Comparison of various Vision-and-Language Navigation datasets highlighting environment type, data source, presence of dynamic elements, use of automatic annotation, and primary task focus (the highlighted color in the table shows our DynamicVLN).
Table 1. Comparison of various Vision-and-Language Navigation datasets highlighting environment type, data source, presence of dynamic elements, use of automatic annotation, and primary task focus (the highlighted color in the table shows our DynamicVLN).
DatasetEnvironmentData SourceDynamic
Elements
Automatic
Annotation
Emergent
Adaptation
Complex Navigation
Conditions
Room-to-Room [1]indoorMatterport3DStructured, static
Room-Across-Room [7]indoorMatterport3DStructured, static
VLN-CE [20]indoorMatterport3DContinuous navigation
ALFRED [3]indoorAI2-THOR 2.0Object interactions
Touchdown [2]outdoorGoogle Street ViewUrban navigation
map2seq [13]outdoorGoogle Street ViewUrban navigation
AerialVLN [16]outdoorAirSimAerial navigation
CARLA-NAV [18]outdoorCARLAGrounded navigation
DOROTHIE [19]outdoorCARLADialogue-based navigation
VLN-VIDEO [14]outdoorGoogle Street ViewUrban navigation
DynamicVLN (our)outdoorCARLAEmergent adaptation
Table 2. Dynamic elements in DynamicVLN are abstracted into higher-level trigger conditions and corresponding decision logic, maintaining the discrete, instruction-driven nature of VLN tasks. Specific scenarios illustrate how agents respond to each condition through temporal stop or continuation decisions.
Table 2. Dynamic elements in DynamicVLN are abstracted into higher-level trigger conditions and corresponding decision logic, maintaining the discrete, instruction-driven nature of VLN tasks. Specific scenarios illustrate how agents respond to each condition through temporal stop or continuation decisions.
Dynamic ElementTrigger ConditionDecision LogicIncluding Scenario
VehicleObstacle appears in agent’s pathTemporal stopVehicle stops abruptly ahead
Potential collision risk detectedTemporal stop or notVehicle approaches from side lane
Priority vehicle detectedTemporal stopAmbulance or fire engine approaches
Lane change detectedTemporal stop or notVehicle changes lanes abruptly
PedestrianPedestrian enters agent’s intended pathTemporal stopPedestrian crosses at crosswalk
Sudden pedestrian movement detectedTemporal stopChild runs across road
Traffic ConditionChange in traffic signal stateTemporal stop or notTraffic light turns red
Regulatory signs encounteredTemporal stop or notStop sign detected
WeatherVisibility reduced due to weather conditionsTemporal stopHeavy fog or rain obscures view
Road condition compromisedTemporal stopSlippery road detected
Table 3. Quantitative results comparing ORAR and DynaVLN on navigation performance metrics. Higher TC and CLS scores indicate better trajectory completion and coverage length, respectively. Lower SPD, SED, nDTW, and sDTW scores indicate better alignment with ground truth trajectories. The quantity of each metric is defined as follows: TC is measured as a percentage (%), SPD represents graph distance, SED corresponds to edit distance, CLS measures coverage length in graph distance, and both nDTW and sDTW represent dynamic time warping scores. The up-arrow means higher is better, and the down-arrow means lower is better.
Table 3. Quantitative results comparing ORAR and DynaVLN on navigation performance metrics. Higher TC and CLS scores indicate better trajectory completion and coverage length, respectively. Lower SPD, SED, nDTW, and sDTW scores indicate better alignment with ground truth trajectories. The quantity of each metric is defined as follows: TC is measured as a percentage (%), SPD represents graph distance, SED corresponds to edit distance, CLS measures coverage length in graph distance, and both nDTW and sDTW represent dynamic time warping scores. The up-arrow means higher is better, and the down-arrow means lower is better.
MethodTC ↑SPD ↓SED ↑CLS ↑nDTW ↑sDTW ↑
ORAR1.6524.021.0816.354.151.05
DynaVLN2.7423.651.4115.073.262.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, Y.; Qiu, Y.; Aoki, Y. DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios. Sensors 2025, 25, 364. https://doi.org/10.3390/s25020364

AMA Style

Sun Y, Qiu Y, Aoki Y. DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios. Sensors. 2025; 25(2):364. https://doi.org/10.3390/s25020364

Chicago/Turabian Style

Sun, Yanjun, Yue Qiu, and Yoshimitsu Aoki. 2025. "DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios" Sensors 25, no. 2: 364. https://doi.org/10.3390/s25020364

APA Style

Sun, Y., Qiu, Y., & Aoki, Y. (2025). DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios. Sensors, 25(2), 364. https://doi.org/10.3390/s25020364

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop