Next Article in Journal
Performance-Enhancing Market Risk Calculation Through Gaussian Process Regression and Multi-Fidelity Modeling
Previous Article in Journal
Revolutionizing Sperm Analysis with AI: A Review of Computer-Aided Sperm Analysis Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing

by
Huthaifa I. Ashqar
1,2,*,
Ahmed Jaber
3,*,
Taqwa I. Alhadidi
4 and
Mohammed Elhenawy
5,6
1
AI and Data Science Department, Arab American University, 13 Zababdeh, Jenin P.O. Box 240, Palestine
2
Artificial Intelligence Program, Fu Foundation School of Engineering and Applied Science, Columbia University, 500 W 120th Street, New York, NY 10027, USA
3
Department of Transport Technology and Economics, Budapest University of Technology and Economics, Műegyetem rkp. 3., H-1111 Budapest, Hungary
4
Civil Engineering Department, Al-Ahliyya Amman University, Al-Saro Al-Salt, Amman 19111, Jordan
5
CARRS-Q, Queensland University of Technology, 130 Victoria Park Rd, Kelvin Grove, Brisbane, QLD 4059, Australia
6
Hourani Center for Applied Scientific Research, Al-Ahliyya Amman University, Al-Saro Al-Salt, Amman 19111, Jordan
*
Authors to whom correspondence should be addressed.
Computation 2025, 13(6), 133; https://doi.org/10.3390/computation13060133
Submission received: 14 April 2025 / Revised: 10 May 2025 / Accepted: 20 May 2025 / Published: 3 June 2025
(This article belongs to the Special Issue Object Detection Models for Transportation Systems)

Abstract

:
This study aims to comprehensively review and empirically evaluate the application of multimodal large language models (MLLMs) and Large Vision Models (VLMs) in object detection for transportation systems. In the first fold, we provide a background about the potential benefits of MLLMs in transportation applications and conduct a comprehensive review of current MLLM technologies in previous studies. We highlight their effectiveness and limitations in object detection within various transportation scenarios. The second fold involves providing an overview of the taxonomy of end-to-end object detection in transportation applications and future directions. Building on this, we proposed empirical analysis for testing MLLMs on three real-world transportation problems that include object detection tasks, namely, road safety attribute extraction, safety-critical event detection, and visual reasoning of thermal images. Our findings provide a detailed assessment of MLLM performance, uncovering both strengths and areas for improvement. Finally, we discuss practical limitations and challenges of MLLMs in enhancing object detection in transportation, thereby offering a roadmap for future research and development in this critical area.

1. Introduction

Object detection is becoming more significant in the field of transportation because it is of critical importance in terms of transportation security, efficiency, and safety. Object detection is the methodology that is used to identify the items and categorize them, including vehicles, traffic signs, pedestrians, etc., which are needed for several applications, including autonomous driving, traveler behavior analysis, and urban infrastructure planning [1,2,3]. In autonomous driving, reliable object detection is the most important method to navigate safely and avoid crashes and traffic signal violations. In traffic monitoring, it provides on-the-spot traffic flow analysis, deals with congestion, and provides real-time incident detection. Accurate object detection also allows urban planners and traffic controllers to build better infrastructures by understanding the different ways people and traffic move in an area. Transportation is a sector with increasing activity as urbanization and population growth continue, and the accurate detection and interpretation of objects within these systems becomes significantly essential [4]. The results of enhanced object detection capabilities can lead to significant reductions in crash rates, optimized traffic flow, and better urban planning, thereby substantially contributing to the overall improvement of transportation systems.
Multimodal large language models (MLLMs) represent a breakthrough advancement as an artificial intelligence (AI) model that has the ability to integrate multimodal data processing capabilities [5]. Specifically, the models that use their vision capability are called in the literature Large Vision Models (VLMs), which use texts and images as input with relatively less reasoning capabilities. MLLMs can handle and analyze data from several sources, including text, images, videos, and sensor data, to provide a comprehensive understanding of complex environments in different settings [6,7,8,9]. MLLMs are built with advanced architectures including transformers, which concurrently process multiple data input streams along with their interpretation. It might lead them to the point of achieving tasks that require a combination of linguistic and perceptual knowledge. The combination permits MLLMs to understand and deal with challenging situations as effectively as real-life single-model scenarios [10]. As shown in Figure 1, a general architecture of an MLLM includes three components. Modality encoder, which condenses raw data into a more streamlined representation that can be used by the AI model. This process utilizes a pre-trained encoder (e.g., CLIP) that has been calibrated to other modalities. The large language model (LLM) backbone is the second component, which is required to output responses in text. The third component is the modality interface, which can be used as a link between the encoder and the LLM components as LLMs can only interpret text.
In transportation, MLLMs have the potential to combine visual data from cameras with textual data from traffic reports and sensor data from vehicles to create a more accurate and comprehensive understanding of the environment as well as provide real-time recommendations as a user-friendly output [11]. This ability to integrate and interpret multiple data types and sources can uniquely position MLLMs to handle the complexities of object detection for advanced transportation systems, where diverse data streams are constantly generated and need to be analyzed in real-time.

1.1. Would MLLM Be Leading the Future of Object Detection?

MLLMs can potentially offer several advantages over traditional object detection models. One of the primary advantages of using MLLMs is their flexibility in understanding various complex settings and environments without the need for bounding boxes, which can be a hurdle for traditional object detection models [5]. MLLMs also effectively handle diverse data types, resources, and tasks, which makes them adaptable to various transportation scenarios and tasks [12]. Moreover, MLLMs have wider capabilities as they can search and locate data simply and broaden their understanding from different types of information, especially in visual form [13,14,15]. Utilizing multimodal data, MLLMs are able to have a higher accuracy rate for object detection tasks versus the models that are based on a particular data type. This potential of improved accuracy is so crucially important for applications such as autonomous driving, where the detection of objects is necessary for safe driving [13,14,15]. MLLMs also offer enhanced contextual understanding as they can interpret complex scenes by combining information from multiple sources, providing a more profound contextual understanding. This capability enables them to make more informed decisions and recommendations, such as recognizing a pedestrian in a crosswalk even if partially obscured by other objects [16,17,18].
Scalability and transferability can be regarded as another advantageous point. MLLMs can be brought to the scale of large data sets processing and applications working in real-time, and that is crucial in dynamic transportation facilities. These data can be transformed into AI-based system solutions. The support of scalability is, as a result, they can endure the enormous quantity of data in modern transportation systems [19,20]. Another positive aspect of language models is that they decrease the need for huge amounts of training data, which is one of the main problems with machine learning systems. MLLMs, thanks to abilities such as zero-shot (zero-shot in-context learning refers to a model’s ability to perform a new task without any prior task-specific training, simply by providing task instructions and examples in the prompt. In our context, MLLMs interpret visual scenes based on textual prompts without being fine-tuned on specific datasets (e.g., iRAP or hazard detection), making them flexible tools for novel applications) and few-shot learning can work outstandingly even with insufficient training data, which in turn diminishes the dependence on big and annotated datasets. This can make MLLMs a potential cost- and time-saving alternative in the context of object detection in transportation [21,22,23].
MLLMs may provide more accurate and reliable object detection capabilities for transportation applications. It becomes possible to use them in location estimation, high-resolution image matching, and object recognition algorithms that increase the transportation system performance by faster and more effective data processing and collaboration [24]. In autonomous driving, MLLMs could enhance the perception system of the vehicle for real-time and correct detection and reaction to road signs, vehicles, and pedestrians [10]. Better detection capability is paramount for instant decision-making that ensures safety and efficiency in vehicle operation. MLLMs will be able to integrate data from cameras, sensors, and reports in monitoring traffic to obtain detail about the current state of the traffic, hence better management and incident responses in the future [25,26]. Contextual understanding is superior; therefore, it recognizes patterns and anomalies in traffic flow, hence improving safety and efficiency. Moreover, MLLMs can guide city planners with elaborate insights into traffic patterns and pedestrian behaviors, thereby giving planners better infrastructure and optimized strategies related to traffic management [27,28]. By leveraging the strengths of MLLMs, transportation systems can achieve higher levels of performance and reliability, addressing the growing demands and complexities of modern urban environments. Besides, connectivity between MLLMs and transportation programs offers a promising future to the technologies for ensuring safer and more convenient transportation.

1.2. Study Contribution

In this paper, we present a comprehensive review study that discusses the use of MLLMs for object detection applications in transportation systems. As opposed to the traditional approaches directed to single-modality data, our study is about the integration of multimodal data streams to improve object detection capabilities. Our work aims to highlight the transformative potential of MLLMs in transportation systems with the help of an extensive review of current technologies but also by presenting empirical tests. The intensity, effectiveness, and safety of transportation systems can be improved by MLLMs more effectively when the accuracy is high; hence, we analyze the theoretical and practical aspects of this topic. We stimulate new research and innovation in the area and promote deeper interaction of the MLLMs with the end-to-end object detection paradigm through this study. The main contribution of this study includes:
  • Provides a detailed review of MLLMs and VLMs in transportation object detection, highlighting benefits and limitations.
  • Introduces a structured taxonomy for end-to-end object detection methods in transportation using MLLMs.
  • Proposes future directions and applications for MLLMs in transportation object detection.
  • Conducts real-world empirical tests on MLLMs for three transportation problems: road safety attribute extraction, safety-critical event detection, and visual reasoning of thermal images.
  • Highlights the potential of MLLMs to enhance intelligent transportation systems, contributing to safer, more efficient AVs and traffic management.
  • Identifies key challenges and limitations in MLLMs, such as order compositional understanding, fine-detail recognition, object hallucination, and computational limitations.

2. Existing Object Detection Technologies

Traditional methods for object detection in the transportation sector have historically relied on manual bounding box annotations in input images [29], followed by the extraction of features such as gradient histogram features, scale invariant features, and Haar-like features [30]. These features are then classified to determine the class of the object being detected. Approaches such as the YOLO series algorithms and SSDs have introduced regression-based object detection algorithms that are more efficient and accurate, such as the YOLO-MFE method that utilizes multiscale feature extraction for improved accuracy [30]. These newer algorithms leverage deep learning techniques and neural networks to enhance object detection performance. For example, object detection using SSD and MobileNets has been shown to be efficient in quickly detecting objects with fewer resources while maintaining high performance [31]. The development of advanced object detection algorithms has led to the creation of specialized methods that cater to specific applications in transportation, such as the RRPN algorithm designed for object detection in autonomous driving vehicles using radar-based real-time region proposal networks [32]. These specialized algorithms showcase the progression of object detection techniques to fulfill unique requirements of transportation systems, particularly in safety-critical applications, such as autonomous driving.
Several traditional techniques were used to detect objects, including CNN, YOLO, and vision transformers. CNN was used extensively in detecting different objects using images and videos. Researchers enhanced the CNN performance by 44.6% when they used the maximal clique algorithm. The improved CNN shows its capability of detecting small objects [33]. In other works, researchers proposed to use signals with different lengths with the lighter convolutional neural networks without using the heavier Recurrent Neural Networks (RNNs), which achieved a comparable performance. Results indicated that the modified CNN reduced the number of parameters and improved the processing methods [34]. Another enhancement on CNN was done by incorporating the classic CNN, and hand-crafted feature extraction was done to assist drivers, including HOG, ICF, and ACF. The results of the work improved object detection for ADAS [35].
YOLO was used extensively in detecting objects, including traffic signs [36], pavement cracks [37], traffic scenes [38], and unmanned vehicles [39]. YOLOv4 was enhanced by using SwinT to detect cars and persons in different traffic scenes. Model performance showed a high prediction accuracy where the car detection precision was 89% and the person detection precision was 94%. In using the real-time application, the YOLOv5 was modified and used to detect ship in real-time. The modified YOLOv5 showed an improvement over the traditional YOLOv5 by 1.3%. Model average precision was 96.6% under different detection schemes [39]. Researchers proposed a YOLO-MXANet for small object detection in traffic scenes, offering improved accuracy, reduced complexity, and faster detection speed compared to traditional methods. Ding et al. [40] proposed an anti-disturbance and variable-scale spatial context features (AVD) detector for road detection, where the training of the multi-layer features of the detector is always taken under the imposition of fake-feature-disturbance from an independent generator, which is trained to exacerbate the detector errors and the mistakes of the feature discriminator. The results indicated that the proposed method improves road detection accuracy by 3% on the Munich remote sensing dataset and 0.4% on the urban road dataset [40].
A novel deep learning anchor-free approach based on CenterNet using the Atrous Spatial Pyramid Pooling (ASPP) to improve the detection performance while not increasing the computational cost and the number of parameters. A large-scale naturalistic driving dataset (BDD100K) was used to examine the effectiveness of the proposed approach. The experimental results show that the proposed approach can effectively improve the detection performance on small objects in various traffic situations [41].

Limitations of Existing Object Detection Technologies

However, these traditional methods have limitations in terms of efficiency and accuracy. For instance, in complex railway scenes, traditional object detection approaches may be inefficient or lack the necessary accuracy, especially when dealing with small objects [42]. Additionally, traditional object detection algorithms often use horizontal bounding boxes to label objects in images, which can lead to accuracy issues and include excessive background information [43]. Moreover, some traditional object detection systems rely on static traffic cameras or expensive mobile units for deployment, which can be costly to establish and maintain or lack diversity in deployment options [44]. These systems may not be as adaptable or cost-effective as desired for widespread transportation applications. Furthermore, traditional methods may struggle with detecting objects with arbitrary orientations, as they typically use horizontal boxes for object labeling, potentially leading to accuracy loss and increased background interference [43].

3. Object Detection Using MLLMs

Recently, the integration of MLLMs and VLMs into object detection systems has seen growing interest, especially in the transportation engineering field. The ability of MLLMs to process diverse types of data, such as images, videos, and text, opens new avenues for improving the accuracy and context-awareness of object detection in critical domains such as autonomous driving and traffic safety. We provide the following comprehensive literature review of the research that used MLLMs, VLMs, and LLMs for object detection in transportation engineering. The methodology for this comprehensive review is illustrated in Figure 2.
In the identification phase, we began by conducting a comprehensive search using Google Scholar. We used specific search terms such as “MLLM or LLM”, “VLM”, “object detection”, and “transportation engineering” to ensure we captured a wide range of relevant studies. This initial search yielded around 782 articles, encompassing various approaches and methodologies in object detection within transportation systems or other fields.
During the screening phase, we aimed to refine the list of articles to focus on the most relevant and high-quality studies. First, we excluded preprints to concentrate on peer-reviewed and published work. Following this, we conducted a title screening to eliminate articles that did not directly relate to our topic, which reduced the pool to about 116 articles. Next, we performed an abstract screening, which involved a more detailed evaluation of each article’s abstract to ensure they aligned with our specific interest in MLLM and object detection. This step further narrowed the list to about 25 articles that were highly relevant to our research.
Finally, in the inclusion phase, we carefully evaluated the full text of the eligible articles. This step involved a thorough review of each paper to determine whether it specifically employed MLLM in the context of object detection within the transportation engineering field. After this detailed evaluation, we identified about eight key articles that directly applied MLLM for object detection, making them highly relevant to our research objectives. These final articles will form the basis of our analysis and review.

3.1. Focused State-of-the-Art

3.1.1. Relevance to Transportation

In the following, we describe the common aspects and differences to thoroughly discuss the use of MLLMs in object detection for transportation engineering. Table 1 shows these eight articles, their taxonomy, and their relevance to transportation engineering.
As shown in Table 1, Knapik et al. [45] and Tami et al. [6] emphasize transportation safety, with Knapik focusing on driver fatigue detection using thermal imaging and CNNs, while Tami et al. address automated detection of traffic safety-critical events. Both are highly relevant to improving road and driver safety. Liao et al. [46] explores object detection for autonomous driving using GPT-4 and Vision Transformer models, enhancing the safety and efficiency of autonomous vehicles. However, de Curtò et al. [47] investigates scene understanding using MLLMs on UAVs, which could have potential applications in aerial traffic monitoring, though its transportation relevance is indirect. Similarly, Zang et al. [48] focuses on general object detection in human-AI interaction scenarios, with indirect transportation relevance in broader object detection contexts. Additionally, Chi et al. [49] investigates the use of GPT-4 for action planning and reasoning in self-driving vehicles. Authors focused on explainable decision-making using multimodal data from monocular cameras and demonstrated how LLMs can provide human-understandable driving rationales. Continually, Zhou et al. [50] presents OpenAnnotate2, a multimodal tool for auto-annotating data in autonomous driving systems, while Mu et al. [51] proposes Pix2Planning, an end-to-end vision-language model for autonomous driving, which translates visual inputs into language-based trajectory planning. The system uses bird’s-eye view (BEV) features to enhance planning precision on the CARLA simulator.

3.1.2. Models and Architectures

As Table 2 shows, de Curtò et al. [47] used YOLOv7 with CLIP prefix captioning for object detection from UAVs, while Knapik et al. [45] incorporated far infrared and visual inputs using various sensors and multimodal fusion. Tami et al. [6] deployed a framework that integrates textual, visual, and auditory modalities to detect safety-critical events using models such as Gemini Pro Vision 1.5. However, Zang et al. [48] highlighted zero-shot detection models for road detection using a pipeline of CLIP and object detectors. Finally, Liao et al. [46] focused on autonomous driving by leveraging GPT-4’s capabilities for grounding vision and language, as well as Chi et al. [49] and Zhou et al. [50].
It was found that the most common architectures are GPT-based models and CNNs such as YOLOv8, as well as vision transformers, which provide high accuracy in object detection. It was noted that fine-tuning was essential for models that required real-time performance, particularly in the Multimodal Driver Condition Monitoring System and the traffic safety-critical event detection.

3.1.3. Dataset Description

As Table 2 shows, the datasets used in the analyzed studies vary widely in size and data type. Public datasets such as KITTI and COCO are frequently used, particularly for autonomous vehicle perception tasks. Custom datasets are often used for specific applications such as driver monitoring and UAV-based scene understanding.
In most of the reviewed studies, the integration of models follows a modular architecture, where vision encoders (e.g., CLIP, YOLOv8) are first used to extract visual features from images or video streams. These encoded visual features are then passed to an LLM (e.g., GPT-3 or GPT-4) either via a joint embedding layer or a prompt interface. For example, CLIP embeddings can be translated into text tokens, allowing the LLM to interpret visual scenes through language. YOLO-based detectors often perform the primary object recognition, with outputs fed as structured inputs into the LLM for scene-level reasoning.

3.1.4. Performance Metrics

The studies reported various performance metrics, with precision being the most commonly reported. As Table 2 shows, using MLLMs for traffic safety-critical events achieved a high accuracy of about 79% in detecting safety-critical events using few-shot learning, outperforming older models [6]. Semantic scene understanding with UAVs demonstrated success in zero-shot scene descriptions but faces limitations in handling complex and crowded scenes. Contextual object detection highlighted how combining multimodal inputs can improve detection accuracy, though model performance depends on data quality. The GPT-4 enhanced multimodal grounding study achieved its highest mAP (mean average precision) at 75%. Furthermore, Chi et al. [49] exhibited a strong performance in generating accurate, interpretable actions with high decision-making transparency, while Zhou et al. [50] showed notable improvements in data annotation efficiency and accuracy for autonomous driving datasets, but detailed performance metrics such as precision or recall were not reported. Finally, Pix2Planning achieved a strong performance on CARLA benchmarks, with a Driving Score (DS) of 92.39 and Route Completion (RC) of 99.60, outperforming other visual-only methods by leveraging the autoregressive transformer to enhance planning precision.

3.1.5. Model Complexity

The model complexity across the eight reviewed articles varies depending on the tasks and modalities integrated into the detection systems. In Knapik et al., the complexity stems from combining thermal imaging and visual data using CNNs and MLLMs, which adds to the computational load due to real-time processing and multimodal fusion of far-infrared and visual inputs. Tami et al. presented a framework that integrates textual, visual, and audio data for detecting traffic safety-critical events [6]. The use of object-level question-answer prompts in the MLLM increases complexity by requiring the model to synthesize data from multiple sources for accurate hazard detection [52,53,54]. Liao et al. incorporated GPT-4 and Vision Transformer models for object detection in autonomous driving, resulting in high model complexity due to the cross-modal attention mechanisms and large-scale pre-training required to understand complex driving environments. Moreover, although de Curtò et al. focused on UAVs, their study involved significant complexity in scene understanding using MLLMs, which require zero-shot learning and real-time data processing. However, Zang et al. explored contextual object detection in varied scenarios, further increasing complexity by integrating multimodal data for human-AI interaction, requiring the models to process diverse data inputs and provide contextualized responses. Chi et al. increased model complexity through its use of the Graph-of-Thought (GoT) structure, adding substantial computational overhead due to its interpretability and explainability mechanisms. In contrast, Zhou et al. simplified the annotation process using a deep neural network for multimodal auto-annotating, though it introduces some complexity due to handling large-scale multi-sensor data. Mu et al. added significant computational complexity due to the large receptive field and the transformer’s ability to maintain spatial information over long sequences. Nonetheless, the integration of multiple data modalities and the need for real-time, high accuracy detection contribute to the high computational and architectural complexity across these studies.

3.1.6. Prompt Engineering

Prompt engineering is pivotal across the eight studies for enhancing MLLM performance. Knapik et al. used tailored prompts to interpret far-infrared images and detect driver fatigue, while Tami et al. employed object-level question-answer prompts to focus on identifying traffic safety-critical events [6]. Liao et al., on the other hand, leveraged advanced prompts with GPT-4, guiding the model to detect key elements in autonomous driving scenes. De Curtò et al. used prompts for zero-shot scene descriptions in UAVs, and Zang et al. applied prompts to help MLLMs detect objects in varied human-AI interaction contexts. Finally, Chi et al. used Graph-of-Thought (GoT) prompts, while Mu et al. employed a vision-language prompt to translate waypoints into language sequences. These engineered prompts ensured more accurate and relevant object detection across the studies.

3.1.7. Limitations of MLLMs from Previous Studies

The following is a list of the limitations reported by the eight studies.
  • Knapik et al.: the main limitation lies in the reliance on far-infrared imaging, which may not perform well in diverse weather or lighting conditions. Additionally, the model’s ability to generalize to different types of drivers or vehicles is not fully explored, as it is limited to specific driver conditions (e.g., yawning).
  • Tami et al.: while the framework effectively detects traffic safety-critical events, its reliance on multimodal data inputs may result in computational inefficiencies, making real-time deployment in complex urban settings a challenge [6].
  • Liao et al.: the complexity of using GPT-4 and vision transformers for object detection creates significant computational demand, limiting its scalability for real-world autonomous driving applications. Furthermore, fine-tuning the model for diverse driving conditions and weather scenarios remains an unresolved challenge.
  • de Curtò et al.: the study’s focus on UAV scene understanding presents limitations when applied to crowded or fast-changing environments. The zero-shot learning approach may also struggle with novel or highly complex scenes that it has not been pre-trained on.
  • Zang et al.: the general focus on human-AI interaction object detection is broad, leading to limitations in transportation-specific contexts. The model’s ability to provide contextually relevant outputs in real-time may falter when faced with highly dynamic or ambiguous scenes.
  • Chi et al.: the limitations are noted in real-world applicability due to safety concerns in dynamic environments.
  • Zhou et al.: the challenges in scaling the tool for larger datasets, especially when dealing with high-dimensional sensor data.
  • Mu et al.: the high computational demand of the Pix2Planning model, which may limit real-time deployment.
All papers discussed the challenge of generalizing models across different environments, such as urban vs. rural, weather conditions, or lighting. Some models struggled with small object detection or suffered from hallucinations in text outputs. Additionally, computational complexity and delays in processing multimodal data in real-time are common limitations.

3.2. Not Focused State-of-the-Art

While this review primarily focuses on the intersection of MLLMs/LLMs, object detection, and transportation engineering, there are several relevant studies that do not directly incorporate MLLMs or LLMs but still contribute significantly to advancements in transportation systems and object detection. Below are summaries of key studies that, while not the primary focus, offer valuable insights into the broader landscape of object detection in transportation and related fields.
The paper Xi et al. [55] focused on using GSENet (Global Semantic Enhancement Network) for roadway feature interpretation in autonomous driving systems. The primary contribution is enhancing semantic segmentation of road features using integrated domain adaptation methods. The model works by improving the interpretation of unstructured road environments, which is crucial for autonomous vehicles operating in complex urban settings. While the paper uses deep learning for road feature detection, it does not involve MLLMs or LLMs and is focused on semantic segmentation rather than object detection. Similarly, Ou et al. [56] introduced Drone-TOOD, a lightweight object detection model designed specifically for UAVs to detect vehicles from aerial imagery. The model is task-aligned, meaning it is optimized for the unique challenges of object detection in UAV images, such as varying altitudes, angles, and scales. Finally, Alaba et al. [57] examined the use of multimodal fusion techniques to enhance 3D object detection in autonomous vehicles. It explored the integration of data from multiple sensors, such as LiDAR, cameras, and radar, to improve the accuracy and robustness of object detection in autonomous driving systems.

3.3. Future Work from Previous Studies

As previous studies have argued, future research should focus on improving the scalability and robustness of MLLMs for real-world transportation applications. For example, it is needed to handle diverse environmental conditions, reduce computational costs, and optimize real-time performance. Also, there are potential improvements by incorporating new sensor types (e.g., radar, LiDAR) or refining multimodal fusion techniques to enhance object detection. Further research on streamlining data annotation tools could also help manage the growing demand for large, high-quality datasets. Finally, reducing the computational complexity of transformer-based models used in end-to-end planning will be crucial for making these models more practical for real-time deployment in large-scale autonomous driving systems.

4. Future Directions and Potential Applications

This section provides a comprehensive taxonomy on the use of MLLMs (or VLMs) within the domain of transportation for object detection and classification. Figure 3 organizes the taxonomy for object detection into around three major, basic, and downstream tasks: perception and understanding, navigation and planning, and decision-making and control. Each of these tasks represents a core area of object detection for transportation applications, including autonomous driving and intelligent transportation systems (ITSs). Perception and understanding include tasks such as Vulnerable Road Users (VRUs) detection and visual scene reasoning for the surrounding environment, which are considered essential for the vehicle to accurately and safely perceive and understand the environment. Navigation and planning include the tasks that VLMs assist in guiding the vehicle and providing recommendations about its speed and acceleration, such as localization and motion planning. In the area of decision-making and control, MLLMs play an important role in providing decisions on both the open-loop and closed-loop levels. This includes tasks that ensure the vehicle can respond appropriately to real-time scenarios by understanding the environment. The taxonomy in Figure 3 depicts a structured view on how MLLMs can be potentially integrated into different aspects of end-to-end object detection. It shows how these models can contribute to the overall functionality and safety of different transportation tasks.
To investigate the potential of MLLMs, we present in the next section three different case studies that can be linked to the taxonomy of object detection tasks and applications in autonomous driving and ITS. The first case study, namely road safety attributes extraction, falls under the perception and understanding category and aligns with tasks such as VRU detection and what is called Open-Vocabulary Object Detection (OVD). In OVD, MLLM combines the power of vision and language to provide accurate and actionable descriptions of objects in scenes, including their attributes, behaviors, relations, and interactions. This case study focuses on extracting and understanding various safety-related events from the images, which can be important for perceiving and interpreting the surrounding environment in ITS applications. The second case study, namely safety-critical event detection, closely represents the decision-making and control category. It is related to the areas of open-loop and closed-loop decision making. The detection of safety-critical events requires the MLLM to identify and react to potential hazardous events and provide a user-friendly decision in real-time that ensures the safety and control of the autonomous vehicle and the surrounding environment and traffic. In the third case study, namely, visual reasoning of thermal images, we investigate the potential of MLLMs within the navigation and planning category. This case study emphasizes the ability of automated vehicles to navigate during adverse environmental conditions using thermal images. The three cases show MLLMs can contribute to the broader goals of perception, decision-making, and navigation in this field. Our goal is to demonstrate the potential of integrating MLLMs into object detection applications, which highlights some of their advantages in improving accuracy, robustness, and versatility in various scenarios.
The fusion methodology in the presented case studies involves sequential processing: visual data is fed into vision encoders (e.g., CLIP or YOLOv8), and relevant textual data (e.g., scene descriptions or road attributes) is combined using prompt engineering for input into the MLLM. In the absence of real-time vehicle sensor data, contextual prompts were used to simulate sensor metadata (e.g., speed, object proximity). This creates a multimodal environment that mimics real-world sensor fusion in intelligent transportation systems.
While this study primarily utilized visual and textual data, future work should consider integrating LiDAR, radar, and inertial sensor modalities. The fusion of these sensor streams with MLLMs could enhance scene understanding, depth perception, and robustness under adverse conditions. Multimodal sensor fusion may also mitigate key limitations observed in MLLMs, such as hallucination and poor fine-detail recognition, by providing richer context and redundancy across modalities.

4.1. Case Study 1: Road Safety Attributes Extraction

Road safety is a critical concern worldwide, with the International Road Assessment Program (iRAP) playing a pivotal role in evaluating and improving road conditions to reduce accidents and fatalities [58,59]. The iRAP is an umbrella program for Road Assessment Programs (RAPs) worldwide that are working to reduce crashes and save lives using a robust, evidence-based approach to prevent unnecessary deaths and suffering. Nonetheless, traditional assessment methods are labor-intensive and time-consuming [60]. The introduction of Google Street View imagery and advances in MLLMs offer promising alternatives [61]. This case study explores the integration of zero-shot in-context learning and MLLMs to automate the extraction of road safety attributes from Google Street View images, providing a scalable and efficient solution for road safety assessments.
In this case study, we collected 168 street view images from various driving environments in Brisbane, Queensland, Australia, encompassing multiple road conditions and features. Utilizing zero-shot in-context learning, we prompted GPT-4-vision-preview to recognize and interpret these images in the context of iRAP safety attributes. This approach allowed the model to generate relevant safety features without explicit prior training on the iRAP dataset, showcasing the potential of MLLMs in understanding complex driving environments and replacing humans in some of the routine inspections. Table 3 shows an example of a utilized prompt for interpreting Google Street View images for eleven selected iRAP safety attributes and the resulting response.
The model was employed in a zero-shot learning approach, meaning it was not specifically trained on these criteria beforehand. Instead, it was presented with the prompts and required to generate accurate responses based on its base knowledge and understanding. Each response was then assessed for accuracy, with results scaled from 0 to 100. Figure 4 shows the accuracies of a GPT-4-vision-preview model in responding to prompts related to eleven selected iRAP criteria using a zero-shot learning approach.
Results showed that in eight attributes out of eleven, the model was able to achieve an accuracy of more than 80%. However, here are some observations on its performance. The model frequently misidentifies barriers on highways as pedestrian fencing. Emergency bays and shoulders are often mistaken for additional traffic lanes. The model struggles with accurately counting lanes, especially when there are more than three. Moving vehicles on urban streets are sometimes incorrectly classified as parked cars. The model misidentifies the materials of medians. It also confuses low medians for mere road lines.
This case study introduced a novel approach to extracting critical attributes for the iRAP using street view images combined with zero-shot in-context learning techniques and MLLMs. By analyzing about 168 diverse street view images collected from different driving environments, we demonstrate the promising capability of MLLMs to accurately identify and categorize key road safety features without prior direct training on specific iRAP attributes. This methodology reduces the reliance on extensive labeled datasets. The application of zero-shot learning allows for the flexible adaptation of the model to new attributes or changes in assessment criteria with no additional training. However, the common misclassifications identified signal a need for model enhancements, either through few-shot learning or fine-tuning with a specialized iRAP dataset. Future research will explore few-shot in-context learning, employing a small set of example images to teach the model traffic and road safety concepts more effectively. It also needs to evaluate other multimodal models capable of local execution, refining them through supervised learning to tap into the MLLM’s inherent foundational knowledge. By incorporating images that delineate the differences between various iRAP attribute values, the goal is to refine the model’s accuracy in extracting attributes, thereby setting the stage for more advanced and detailed road safety evaluations.

4.2. Case Study 2: Safety-Critical Event Detection

Recent advancements in MLLMs are enhancing autonomous driving, focusing on textual and multimodal analyses to interpret complex driving environments [62,63]. These technologies are promising for understanding dynamic road conditions and offering early hazard warnings [64]. Despite their potential, a gap exists in applying these models to safety-critical event analysis. This case study addresses this gap by exploring the use of MLLMs for analyzing safety-critical events in driving, combining text and image analyses to provide illustrative and actionable insights.
We used a multi-stage question-answer (Q/A) approach and applied it for two MLLMs including Gemini Pro Vision 1.0 and Llava-7B 1.5. We started with frame extraction, where the system automatically collects video frames from the ego vehicle’s camera at uniform intervals. The frames were then tested for hazard detection, where the MLLMs assessed the scene for potential danger. Upon detecting a hazard, the framework used a multilateral Q/A strategy to identify the characteristics of the critical event further. For this stage, we used what is called what, which, and where queries to investigate the object-level details. In the “What” phase, the MLLMs classified the objects detected by the camera, while the “Which” phase involved the MLLMs identifying specific features and attributes of the agents involved in the event, and the final “Where” phase tasks the MLLMs to determine the spatial location and distance of the hazardous agents. To evaluate MLLMs, human experts first assessed ground truth data annotating the events, which was performed by one transportation safety researcher expert who independently reviewed each video. The annotation was done by a five researchers’ team who labeled six elements per event: (1) risk presence, (2) hazard type, (3) agent object, (4) agent attribute (e.g., speed, signal violation), (5) spatial position, and (6) appropriate action. Disagreements were resolved through consensus. These annotations served as the benchmark for evaluating model accuracy in the Q/A stages. We collected a sample of 200 videos from a YouTube channel, namely Dash Cam Owners Australia, offering diverse scenarios for testing. Comparative analysis measured the model’s performance against this data to determine its accuracy in hazard detection and categorization.
Figure 5 shows four separate examples from the Gemini Pro Vision 1.0 predictions, demonstrating the model’s reasoning process under different visual inputs. To evaluate the effectiveness of MLLMs in detecting safety-critical events, a series of experiments were carried out using Gemini Pro Vision 1.0 and Llava-7B 1.5. For Gemini, we evaluated the model in two folds. In the first experiment, we used two frames as input to the model, and for the second experiment, we used the video directly with no processing. The examples shown in Figure 5 are as follows:
Figure 5a. Risk: yes, scene: intersection, what: vehicle, which: truck, where: middle of the intersection, action: stop.
Figure 5b. Risk: yes, scene: highway, what: vehicle, which: the car ahead has lost a tier, where: middle of the road, action: decelerate.
Figure 5c. Risk: no, scene: rural road.
Figure 5d. Risk: yes, scene: intersection, what: vehicle, which: white ahead, where: middle of the intersection, action: stop.
Table 4 presents a comparative analysis of the performance of three different configurations of vision models—Gemini Pro Vision 1.0 with two-frame input, Llava-7B 1.5 with two-frame input, and Gemini Pro Vision 1.0 with video input—across various stages of the Q/A process. The stages include risk, scene, what, which, where, and recommended action, with corresponding accuracy percentages for each configuration. Gemini Pro Vision 1.0 with video input consistently outperforms the other configurations, achieving the highest accuracy in most categories, particularly in the scene (100%) and recommended action (75%) stages. The overall performance metrics indicate that Gemini Pro Vision 1.0 with video input (74.67%) is superior to both its two-frame input counterpart (71.8%) and the Llava-7B 1.5 model (58.6%). This suggests that utilizing video input can significantly enhance the model’s ability to accurately process and respond to queries in dynamic and complex visual scenarios.
This case study aims to enhance the efficiency of processing naturalistic driving dash cam data, ultimately improving the detection of safety-critical events using an MLLM-empowered model. It demonstrates the potential of leveraging the capabilities of MLLMs in analyzing safety-critical event scenarios using multimodal data integration and dynamic contextual data reduction for guiding the model’s output. Future research in this area will focus on expanding the dataset and scenario range to further validate our innovative approach to driving safety.

4.3. Case Study 3: Visual Reasoning of Thermal Images

Thermal imaging is vital for advancing automatic driving systems due to its ability to detect heat signatures, enhancing the overall perception capabilities of autonomous vehicles [65,66]. One primary benefit of thermal imaging is its enhanced visibility in adverse conditions in both urban and highway environments [67]. Thermal cameras can detect heat emitted by objects and living beings, providing clear images even in complete darkness, which is essential for detecting pedestrians, animals, and other vehicles at night [68,69].
This case study explores the capabilities of MLLMs, specifically GPT-4 Vision Preview and Gemini 1.0 Pro Vision, in visual reasoning of thermal images. We evaluate the MLLMs’ ability to detect and enumerate objects within thermal images using zero-shot in-context learning. This case study not only shows the potential of MLLMs to enhance object detection accuracy and environmental understanding but also opens venues for safer and more reliable autonomous driving systems. We used the Teledyne FLIR Free ADAS Thermal Dataset V2 [70], which includes fully annotated frames covering more than fifteen object classes, captured using a thermal and visible camera pair mounted on a vehicle. The dataset ensures diversity by sampling frames from a wide range of footage, excluding redundant footage, to provide robust training and validation data.
Using zero-shot in-context learning, the models were tested on a sample of 200 images and demonstrated an ability to process and analyze these modalities. Table 5 shows the performance metrics for GPT-4 Vision Preview and Gemini 1.0 Pro Vision evaluated on four object categories: person, bike, car, and motorcycle. The metrics include True Positive Rate (TPR) and False Positive Rate (FPR). For person detection, GPT-4 demonstrates a higher TPR (0.57) compared to Gemini (0.39), with both models maintaining an FPR of 0, indicating better accuracy in identifying persons for GPT-4. In bike detection, Gemini shows a slightly higher TPR of about 0.41 than GPT-4 (0.31). For car detection, both models perform well, but Gemini has a slightly higher TPR of 0.90 and a slightly lower FPR of 0.08. In motorcycle detection, Gemini significantly outperforms GPT-4, with a TPR of 0.24 versus 0.08. Overall, the insights suggest that while both models are effective, Gemini 1.0 Pro Vision shows superior performance in bike, car, and motorcycle detection, whereas GPT-4 Vision Preview surpasses in person detection. Detection accuracy also varied among different images and different objects. Although thermal images posed a unique challenge due to their reliance on heat signatures rather than visible light and their reduced visual information, MLLMs achieved moderate success in recognizing objects such as cars and people, establishing a strong foundation for future model development.
This case study shows the potential of MLLMs, specifically GPT-4 Vision Preview and Gemini 1.0 Pro Vision, in processing and analyzing thermal images for applications in autonomous driving and intelligent transportation systems. The results suggest the need for ongoing enhancements to enhance the accuracy of MLLMs, especially in diverse environmental conditions that autonomous vehicles may encounter. The findings validate the ongoing use and enhancement of MLLMs for image-based processing in autonomous driving.
Overall, Gemini 1.0 Pro Vision outperformed GPT-4 in safety-critical event detection when using video inputs, particularly in recommending actions and scene interpretation. Conversely, GPT-4 demonstrated higher accuracy in recognizing pedestrian-related events in thermal imagery, suggesting a strength in detecting humans over other classes. These complementary strengths highlight the need for hybrid MLLM systems or task-specific fine-tuning.

5. Challenges and Limitations

In this section, we offer our understanding of the limitations and challenges of using MLLMs for object detection and classification in transportation applications. We derived these limitations and challenges from implementing the three case studies, studying the state-of-the-art of using MLLMs in object detection, as well as from other studies that we published previously or are in progress. Figure 6 depicts a summary of the challenges and limitations.

5.1. Order Compositional Understanding

We found that MLLMs struggle with correctly ordering and composing the relationships between objects and their attributes in complex scenes [71,72]. Results showed that these models usually perform well on tasks such as image-text retrieval, but they still frequently show shortages in capturing the precise relationships and the correct order within the visual context. MLLMs seemed to misinterpret and sometimes overlook the importance of the sequence of the scene and in which objects and their corresponding attributes are presented. This frequently led to errors in predicting, such as confusing the relationships between objects (e.g., mistaking “the cat on the mat” with “the mat on the cat”). This limitation can be relaxed by more robust training techniques and benchmarking, which ensure that models will accurately interpret and logically reason about the compositional structure inherent in multimodal data.

5.2. Fine-Detail Recognition: Perceptual Limitation

We found that the perceptual limitations of MLLMs were primarily revolved around the struggle of accurately identifying and interpreting small visual objects within images. Although MLLMs showed promising capabilities in visual question answering (Q/A), their performance (sometimes significantly) drops when dealing with relatively smaller objects. This has been majorly influenced by several factors including image and object quality, relative size, its location in the image, and the presence of other surrounding distractors [73]. We also found that even with relatively high-quality images, MLLMs still struggled to recognize smaller objects if they were positioned peripherally or between visual distractors. Additionally, the models’ performance varied based on the object’s location within the image, which suggested a positional bias in its training.

5.3. Cross-Modal Alignment

Another limitation that we found in the vision encoders of MLLMs was primarily centered upon achieving effective cross-modal alignment and visual correspondence. These two factors were found to be important for the model’s performance, as they significantly influence how well the vision representations integrate with the language model [74]. Vision encoders might struggle to align visual features with corresponding textual elements and other related modes of data, which made them fail to maintain high fidelity in visual correspondence. This was found particularly when fine-grained details were involved.

5.4. Object Hallucination

In the exploration of MLLM, we found that one notable limitation was object hallucination, in which the model generates objects that do not exist in the target image [75]. This issue presented a significant challenge as it undermined the reliability of MLLMs in some of the tasks, and it might become an issue in other applications as well. For example, we found that in scenarios that require high precision and reliability, such as navigation and planning in AVs, MLLM tends to hallucinate with regard to objects, which might lead to unsafe decisions and increased risks. The problem of object hallucination not only highlights a fundamental flaw in the current capabilities of MLLMs but also restricts their practical deployment across sensitive and downstream tasks.

5.5. Object Counting Accuracy

One crucial application for ITS is the counting of objects in the scenes to be used for different applications, such as a simple traffic management system. Despite MLLMs’ advanced capabilities, we found that MLLMs showed a significant challenge in object enumeration, as shown directly in Case Study 3. The study highlighted how these models can be proficient in identifying objects across diverse modalities such as RGB and thermal imaging but still struggle with precise object counting. This was particularly shown when interpreting complex or overlapping object scenes.

5.6. Computational Limitations

We found that the computational demands for training, fine-tuning, and inference of MLLMs can be substantially driven by the complexity and the scale of these models [76,77]. We found that zero-shot learning, which does not require training and depends on the knowledge base of the MLLM, showed relatively low performance. We also found that few-shot learning, training, and fine-tuning showed a higher performance. Nonetheless, training an MLLM often involves managing vast amounts of data across multiple modalities (e.g., text, images, and video). This requires extensive computational power involving GPUs or TPUs over long periods. For instance, we found that in some cases the pre-training phase alone can consume hundreds of thousands of GPU hours, depending on the model size and dataset. Fine-tuning these models also demands significant resources, though it is generally more efficient than training from scratch. The fine-tuning process can be optimized through techniques such as quantization and distributed training, which help to reduce memory usage and computational load [78]. Moreover, we found that inference also presents challenges, especially when deploying these models in real-time applications. The memory requirements for managing key-value caches during inference are a major bottleneck [79]. Techniques such as PagedAttention, which optimize memory allocation by using non-contiguous memory blocks, can help alleviate some of these constraints.

6. Conclusions

This study provides a comprehensive evaluation of MLLMs and VLMs in the context of object detection for transportation systems. Through an extensive review of current MLLM technologies, we highlight their significant potential in enhancing transportation applications, particularly in tasks such as road safety monitoring and traffic management. While MLLMs and VLMs hold considerable potential for advancing autonomous driving and ITS, there remains a gap between their current capabilities and the practical demands of real-time, safety-critical applications.
The empirical analysis across three real-world transportation scenarios—road safety attributes extraction, safety-critical event detection, and visual reasoning of thermal images—offers valuable insights into the strengths and limitations of MLLMs. While these models show promise in interpreting multimodal data and addressing complex transportation problems, their limitations indicate the need for further refinement in cross-modal alignment, order compositional understanding, fine-detail recognition, object counting accuracy, object hallucination, and computational demands.
This study serves as a roadmap for future research, suggesting taxonomy for the use of MLLMs in end-to-end object detection. It also identifies areas for improvement in model training, computational efficiency, and the broader integration of multimodal data for safer and more efficient transportation systems. By addressing these challenges, MLLMs can play a transformative role in shaping the future of autonomous vehicles and intelligent transportation infrastructure.

Author Contributions

Conceptualization, H.I.A. and M.E.; methodology, H.I.A., A.J., T.I.A. and M.E.; software, M.E.; validation, H.I.A., A.J., T.I.A. and M.E.; investigation, H.I.A., A.J., T.I.A. and M.E.; data curation, H.I.A., A.J., T.I.A. and M.E.; writing—original draft preparation, H.I.A., A.J., T.I.A. and M.E.; writing—review and editing, H.I.A.; visualization, H.I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The prompts used, evaluation scripts, and the annotated labels are available upon request. We intend to upload a subset of the scripts and sample prompts to these repositories [https://huggingface.co/Tami3/HazardNet (accessed on 1 February 2025) and https://universe.roboflow.com/detection-flj91/irap-zcrw6-uif6x (accessed on 1 February 2025)].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Campbell, A.; Both, A.; Sun, C. Detecting and mapping traffic signs from Google Street View images using deep learning and GIS. Comput. Environ. Urban. Syst. 2019, 77, 101350. [Google Scholar] [CrossRef]
  2. Jain, N.K.; Saini, R.K.; Mittal, P. A Review on Traffic Monitoring System Techniques. In Soft Computing: Theories and Applications; Ray, K., Sharma, T.K., Rawat, S., Saini, R.K., Bandyopadhyay, A., Eds.; Springer: Singapore, 2019; Volume 742, pp. 569–577. [Google Scholar] [CrossRef]
  3. Zitar, R.A.; Alhadhrami, E.; Abualigah, L.; Barbaresco, F.; Seghrouchni, A.E. Optimum sensors allocation for drones multi-target tracking under complex environment using improved prairie dog optimization. Neural Comput. Appl. 2024, 36, 10501–10525. [Google Scholar] [CrossRef]
  4. Sivaraman, S.; Trivedi, M.M. Looking at Vehicles on the Road: A Survey of Vision-Based Vehicle Detection, Tracking, and Behavior Analysis. IEEE Trans. Intell. Transp. Syst. 2013, 14, 1773–1795. [Google Scholar] [CrossRef]
  5. Huang, D.; Yan, C.; Li, Q.; Peng, X. From Large Language Models to Large Multimodal Models: A Literature Review. Appl. Sci. 2024, 14, 5068. [Google Scholar] [CrossRef]
  6. Tami, M.A.; Ashqar, H.I.; Elhenawy, M.; Glaser, S.; Rakotonirainy, A. Using Multimodal Large Language Models (MLLMs) for Automated Detection of Traffic Safety-Critical Events. Vehicles 2024, 6, 1571–1590. [Google Scholar] [CrossRef]
  7. Jaradat, S.; Alhadidi, T.I.; Ashqar, H.I.; Hossain, A.; Elhenawy, M. Exploring Traffic Crash Narratives in Jordan Using Text Mining Analytics. arXiv 2024, arXiv:2406.09438. [Google Scholar]
  8. Elhenawy, M.; Abutahoun, A.; Alhadidi, T.I.; Jaber, A.; Ashqar, H.I.; Jaradat, S.; Abdelhay, A.; Glaser, S.; Rakotonirainy, A. Visual Reasoning and Multi-Agent Approach in Multimodal Large Language Models (MLLMs): Solving TSP and mTSP Combinatorial Challenges. Mach. Learn. Knowl. Extr. 2024, 6, 1894–1920. [Google Scholar] [CrossRef]
  9. Belhaouari, S.B.; Kraidia, I. Efficient self-attention with smart pruning for sustainable large language models. Sci. Rep. 2025, 15, 10171. [Google Scholar] [CrossRef]
  10. Luo, S.; Chen, W.; Tian, W.; Liu, R.; Hou, L.; Zhang, X.; Shen, H.; Wu, R.; Geng, S.; Zhou, Y.; et al. Delving Into Multi-Modal Multi-Task Foundation Models for Road Scene Understanding: From Learning Paradigm Perspectives. IEEE Trans. Intell. Veh. 2024, 1–25. [Google Scholar] [CrossRef]
  11. Zhang, D.; Zheng, H.; Yue, W.; Wang, X. Advancing ITS Applications with LLMs: A Survey on Traffic Management, Transportation Safety, and Autonomous Driving. In Proceedings of the Rough Sets: International Joint Conference, IJCRS 2024, Halifax, NS, Canada, 17–20 May 2024; pp. 295–309. [Google Scholar] [CrossRef]
  12. Veres, M.; Moussa, M. Deep Learning for Intelligent Transportation Systems: A Survey of Emerging Trends. IEEE Trans. Intell. Transp. Syst. 2020, 21, 3152–3168. [Google Scholar] [CrossRef]
  13. Naveed, Q.N.; Alqahtani, H.; Khan, R.U.; Almakdi, S.; Alshehri, M.; Rasheed, M.A.A. An Intelligent Traffic Surveillance System Using Integrated Wireless Sensor Network and Improved Phase Timing Optimization. Sensors 2022, 22, 3333. [Google Scholar] [CrossRef] [PubMed]
  14. Yoon, S.; Cho, J. Deep Multimodal Detection in Reduced Visibility Using Thermal Depth Estimation for Autonomous Driving. Sensors 2022, 22, 5084. [Google Scholar] [CrossRef]
  15. Wang, Y.; Mao, Q.; Zhu, H.; Deng, J.; Zhang, Y.; Ji, J.; Li, H.; Zhang, Y. Multi-Modal 3D Object Detection in Autonomous Driving: A Survey. Int. J. Comput. Vis. 2023, 131, 2122–2152. [Google Scholar] [CrossRef]
  16. Guo, J.; Song, B.; Chen, S.; Yu, F.R.; Du, X.; Guizani, M. Context-Aware Object Detection for Vehicular Networks Based on Edge-Cloud Cooperation. IEEE Internet Things J. 2020, 7, 5783–5791. [Google Scholar] [CrossRef]
  17. Lee, W.-Y.; Jovanov, L.; Philips, W. Cross-Modality Attention and Multimodal Fusion Transformer for Pedestrian Detection. In Proceedings of the Computer Vision—ECCV 2022 Workshops, Tel Aviv, Israel, 23–27 October 2022; pp. 608–623. [Google Scholar] [CrossRef]
  18. Chen, Y.; Ye, J.; Wan, X. TF-YOLO: A Transformer–Fusion-Based YOLO Detector for Multimodal Pedestrian Detection in Autonomous Driving Scenes. World Electr. Veh. J. 2023, 14, 352. [Google Scholar] [CrossRef]
  19. Yang, Y.; Zhang, W.; Lin, H.; Liu, Y.; Qu, X. Applying masked language model for transport mode choice behavior prediction. Transp. Res. Part A Policy Pract. 2024, 184, 104074. [Google Scholar] [CrossRef]
  20. Wang, X.; Chen, G.; Qian, G.; Gao, P.; Wei, X.-Y.; Wang, Y.; Tian, Y.; Gao, W. Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey. Mach. Intell. Res. 2023, 20, 447–482. [Google Scholar] [CrossRef]
  21. Zhou, S.; Deng, C.; Piao, Z.; Zhao, B. Few-shot traffic sign recognition with clustering inductive bias and random neural network. Pattern Recognit. 2020, 100, 107160. [Google Scholar] [CrossRef]
  22. Bansal, A.; Sikka, K.; Sharma, G.; Chellappa, R.; Divakaran, A. Zero-Shot Object Detection. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 397–414. [Google Scholar] [CrossRef]
  23. Liu, Y.; Dong, L.; He, T. A Closer Look at Few-Shot Object Detection. In Pattern Recognition and Computer Vision; Springer: Singapore, 2024; pp. 430–447. [Google Scholar] [CrossRef]
  24. Liu, Y.; Meng, S.; Wang, H.; Liu, J. Deep learning based object detection from multi-modal sensors: An overview. Multimed. Tools Appl. 2023, 83, 19841–19870. [Google Scholar] [CrossRef]
  25. Hu, Z.; Lam, W.H.K.; Wong, S.C.; Chow, A.H.F.; Ma, W. Turning traffic surveillance cameras into intelligent sensors for traffic density estimation. Complex. Intell. Syst. 2023, 9, 7171–7195. [Google Scholar] [CrossRef]
  26. Radwan, A.; Amarneh, M.; Alawneh, H.; Ashqar, H.I.; AlSobeh, A.; Magableh, A.A.A.R. Predictive Analytics in Mental Health Leveraging LLM Embeddings and Machine Learning Models for Social Media Analysis. Int. J. Web Serv. Res. (IJWSR) 2024, 21, 1–22. [Google Scholar] [CrossRef]
  27. Fan, Z.; Loo, B.P.Y. Street life and pedestrian activities in smart cities: Opportunities and challenges for computational urban science. Comput. Urban. Sci. 2021, 1, 26. [Google Scholar] [CrossRef]
  28. Masri, S.; Ashqar, H.I.; Elhenawy, M. Leveraging Large Language Models (LLMs) for Traffic Management at Urban Intersections: The Case of Mixed Traffic Scenarios. arXiv 2024, arXiv:2408.00948. [Google Scholar]
  29. Elhenawy, M.; Ashqar, H.I.; Masoud, M.; Almannaa, M.H.; Rakotonirainy, A.; Rakha, H.A. Deep transfer learning for vulnerable road users detection using smartphone sensors data. Remote Sens. 2020, 12, 3508. [Google Scholar] [CrossRef]
  30. Li, H.; Yin, Z.; Fan, C.; Wang, X. YOLO-MFE: Towards More Accurate Object Detection Using Multiscale Feature Extraction. In Proceedings of the Sixth International Conference on Intelligent Computing, Communication, and Devices (ICCD 2023), Hong Kong, China, 3–5 March 2023. [Google Scholar] [CrossRef]
  31. Channakeshnava Gowda, S.V.; Lokesh, D.S.; Peddinni Sai, K.; Praveen Kumar, S.; Pushpanathan, G. Object Detection Using Open CV and Deep Learning. Int. J. Sci. Res. Eng. Manag. 2022, 6, 1–4. [Google Scholar] [CrossRef]
  32. Nabati, R.; Qi, H. RRPN: Radar Region Proposal Network for Object Detection in Autonomous Vehicles. arXiv 2019, arXiv:1905.00526. [Google Scholar] [CrossRef]
  33. García-Aguilar, I.; García-González, J.; Luque-Baena, R.M.; López-Rubio, E. Object detection in traffic videos: An optimized approach using super-resolution and maximal clique algorithm. Neural Comput. Appl. 2023, 35, 18999–19013. [Google Scholar] [CrossRef]
  34. Moreau, H.; Vassilev, A.; Chen, L. The devil is in the details: An efficient convolutional neural network for transport mode detection. IEEE Trans. Intell. Transp. Syst. 2021, 23, 12202–12212. [Google Scholar] [CrossRef]
  35. Wei, J.; He, J.; Zhou, Y.; Chen, K.; Tang, Z.; Xiong, Z. Enhanced object detection with deep convolutional neural networks for advanced driving assistance. IEEE Trans. Intell. Transp. Syst. 2019, 21, 1572–1583. [Google Scholar] [CrossRef]
  36. Nafaa, S.; Ashour, K.; Mohamed, R.; Essam, H.; Emad, D.; Elhenawy, M.; Ashqar, H.I.; Hassan, A.A.; Alhadidi, T.I. Advancing Roadway Sign Detection with YOLO Models and Transfer Learning. In Proceedings of the 2024 IEEE 3rd International Conference on Computing and Machine Intelligence (ICMI), Mount Pleasant, MI, USA, 13–14 April 2024; IEEE: New York, NY, USA, 2024; pp. 1–4. [Google Scholar]
  37. Nafaa, S.; Ashour, K.; Mohamed, R.; Essam, H.; Emad, D.; Elhenawy, M.; Ashqar, H.I.; Hassan, A.A.; Alhadidi, T.I. Automated Pavement Cracks Detection and Classification Using Deep Learning. In Proceedings of the 2024 IEEE 3rd International Conference on Computing and Machine Intelligence (ICMI), Mount Pleasant, MI, USA, 13–14 April 2024; IEEE: New York, NY, USA, 2024; pp. 1–5. [Google Scholar]
  38. Lu, K.; Zhao, F.; Xu, X.; Zhang, Y. An object detection algorithm combining self-attention and YOLOv4 in traffic scene. PLoS ONE 2023, 18, e0285654. [Google Scholar] [CrossRef]
  39. Zhang, J.; Jin, J.; Ma, Y.; Ren, P. Lightweight object detection algorithm based on YOLOv5 for unmanned surface vehicles. Front. Mar. Sci. 2023, 9, 1058401. [Google Scholar] [CrossRef]
  40. Ding, Q.; Liu, H.; Luo, H.; Chen, X. Road detection network based on anti-disturbance and variable-scale spatial context detector. IEEE Access 2021, 9, 114640–114648. [Google Scholar] [CrossRef]
  41. Li, G.; Xie, H.; Yan, W.; Chang, Y.; Qu, X. Detection of road objects with small appearance in images for autonomous driving in various traffic situations using a deep learning based approach. IEEE Access 2020, 8, 211164–211172. [Google Scholar] [CrossRef]
  42. Tao, Y.; Zhang, X.; Zhou, F. Autonomous Railway Traffic Object Detection Using Feature-Enhanced Single-Shot Detector. IEEE Access 2020, 8, 145182–145193. [Google Scholar] [CrossRef]
  43. Chen, Y.; Duan, Z.; Yan, H. R-YOLOv5: A Lightweight Rotation Detector for Remote Sensing. In Proceedings of the International Conference on Computer, Artificial Intelligence, and Control Engineering (CAICE 2022), Zhuhai, China, 25–27 February 2022. [Google Scholar] [CrossRef]
  44. Gaidash, V.; Grakovski, A. ‘Mass Centre’ Vectorization Algorithm for Vehicle’s Counting Portable Video System. Transp. Telecommun. J. 2016, 17, 289–297. [Google Scholar] [CrossRef]
  45. Knapik, M.; Cyganek, B.; Balon, T. Multimodal Driver Condition Monitoring System Operating in the Far-Infrared Spectrum. Electronics 2024, 13, 3502. [Google Scholar] [CrossRef]
  46. Liao, H.; Shen, H.; Li, Z.; Wang, C.; Li, G.; Bie, Y.; Xu, C. GPT-4 enhanced multimodal grounding for autonomous driving: Leveraging cross-modal attention with large language models. Commun. Transp. Res. 2024, 4, 100116. [Google Scholar] [CrossRef]
  47. de Curtò, J.; de Zarzà, I.; Calafate, C.T. Semantic Scene Understanding with Large Language Models on Unmanned Aerial Vehicles. Drones 2023, 7, 114. [Google Scholar] [CrossRef]
  48. Zang, Y.; Li, W.; Han, J.; Zhou, K.; Loy, C.C. Contextual Object Detection with Multimodal Large Language Models. Int. J. Comput. Vis. 2024, 133, 825–843. [Google Scholar] [CrossRef]
  49. Chi, F.; Wang, Y.; Nasiopoulos, P.; Leung, V.C.M. Multi-Modal GPT-4 Aided Action Planning and Reasoning for Self-driving Vehicles. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; IEEE: New York, NY, USA, 2024; pp. 7325–7329. [Google Scholar] [CrossRef]
  50. Zhou, Y.; Cai, L.; Cheng, X.; Zhang, Q.; Xue, X.; Ding, W.; Pu, J. OpenAnnotate2: Multi-Modal Auto-Annotating for Autonomous Driving. IEEE Trans. Intell. Veh. 2024, 1–13. [Google Scholar] [CrossRef]
  51. Mu, X.; Qin, T.; Zhang, S.; Xu, C.; Yang, M. Pix2Planning: End-to-End Planning by Vision-language Model for Autonomous Driving on Carla Simulator. In Proceedings of the 2024 IEEE Intelligent Vehicles Symposium (IV), Jeju Island, Republic of Korea, 2–5 June 2024; IEEE: New York, NY, USA, 2024; pp. 2383–2390. [Google Scholar] [CrossRef]
  52. Sammoudi, M.; Habaybeh, A.; Ashqar, H.I.; Elhenawy, M. Question-Answering (QA) Model for a Personalized Learning Assistant for Arabic Language. arXiv 2024, arXiv:2406.08519. [Google Scholar]
  53. Tami, M.; Ashqar, H.I.; Elhenawy, M. Automated Question Generation for Science Tests in Arabic Language Using NLP Techniques. arXiv 2024, arXiv:2406.08520. [Google Scholar]
  54. Rouzegar, H.; Makrehchi, M. Generative AI for Enhancing Active Learning in Education: A Comparative Study of GPT-3.5 and GPT-4 in Crafting Customized Test Questions. arXiv 2024, arXiv:2406.13903. [Google Scholar]
  55. Xi, S.; Liu, Z.; Wang, Z.; Zhang, Q.; Ding, H.; Kang, C.C.; Chen, Z. Autonomous Driving Roadway Feature Interpretation Using Integrated Semantic Analysis and Domain Adaptation. IEEE Access 2024, 12, 98254–98269. [Google Scholar] [CrossRef]
  56. Ou, K.; Dong, C.; Liu, X.; Zhai, Y.; Li, Y.; Huang, W.; Qiu, W.; Wang, Y.; Wang, C. Drone-TOOD: A Lightweight Task-Aligned Object Detection Algorithm for Vehicle Detection in UAV Images. IEEE Access 2024, 12, 41999–42016. [Google Scholar] [CrossRef]
  57. Alaba, S.Y.; Gurbuz, A.C.; Ball, J.E. Emerging Trends in Autonomous Vehicle Perception: Multimodal Fusion for 3D Object Detection. World Electr. Veh. J. 2024, 15, 20. [Google Scholar] [CrossRef]
  58. Murozi, A.-F.M.; Ishak, S.Z.; Nusa, F.N.M.; Hoong, A.P.W.; Sulistyono, S. The application of international road assessment programme (irap) as a road infrastructure risk assessment tool. In Proceedings of the 2022 IEEE 13th Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 23 July 2022; IEEE: New York, NY, USA, 2022; pp. 237–242. [Google Scholar]
  59. World Health Organization. Global Status Report on Road Safety 2015; World Health Organization: Geneva, Switzerland, 2015. [Google Scholar]
  60. Angelo, A.A.; Sasai, K.; Kaito, K. Assessing critical road sections: A decision matrix approach considering safety and pavement condition. Sustainability 2023, 15, 7244. [Google Scholar] [CrossRef]
  61. Alhadidi, T.I.; Jaber, A.; Jaradat, S.; Ashqar, H.I.; Elhenawy, M. Object Detection Using Oriented Window Learning Vision Transformer: Roadway Assets Recognition. arXiv 2024, arXiv:2406.10712. Available online: https://api.semanticscholar.org/CorpusID:270560797 (accessed on 1 January 2025).
  62. Cui, C.; Ma, Y.; Cao, X.; Ye, W.; Wang, Z. Receive, reason, and react: Drive as you say, with large language models in autonomous vehicles. IEEE Intell. Transp. Syst. Mag. 2024, 16, 81–94. [Google Scholar] [CrossRef]
  63. Sha, H.; Mu, Y.; Jiang, Y.; Chen, L.; Xu, C.; Luo, P.; Li, S.E.; Tomizuka, M.; Zhan, W.; Ding, M. Languagempc: Large language models as decision makers for autonomous driving. arXiv 2023, arXiv:2310.03026. [Google Scholar]
  64. Xu, Z.; Zhang, Y.; Xie, E.; Zhao, Z.; Guo, Y.; Wong, K.-Y.K.; Li, Z.; Zhao, H. Drivegpt4: Interpretable end-to-end autonomous driving via large language model. arXiv 2023, arXiv:2310.01412. [Google Scholar] [CrossRef]
  65. Yu, X.; Marinov, M. A study on recent developments and issues with obstacle detection systems for automated vehicles. Sustainability 2020, 12, 3281. [Google Scholar] [CrossRef]
  66. He, Y.; Deng, B.; Wang, H.; Cheng, L.; Zhou, K.; Cai, S.; Ciampa, F. Infrared machine vision and infrared thermography with deep learning: A review. Infrared Phys. Technol. 2021, 116, 103754. [Google Scholar] [CrossRef]
  67. Nidamanuri, J.; Nibhanupudi, C.; Assfalg, R.; Venkataraman, H. A progressive review: Emerging technologies for ADAS driven solutions. IEEE Trans. Intell. Veh. 2021, 7, 326–341. [Google Scholar] [CrossRef]
  68. Krišto, M.; Ivasic-Kos, M.; Pobar, M. Thermal object detection in difficult weather conditions using YOLO. IEEE Access 2020, 8, 125459–125476. [Google Scholar] [CrossRef]
  69. Leira, F.S.; Helgesen, H.H.; Johansen, T.A.; Fossen, T.I. Object detection, recognition, and tracking from UAVs using a thermal camera. J. Field Robot. 2021, 38, 242–267. [Google Scholar] [CrossRef]
  70. Krišto, M.; Ivašić-Kos, M. Thermal imaging dataset for person detection. In Proceedings of the 2019 42nd International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 20–24 May 2019; IEEE: New York, NY, USA, 2019; pp. 1126–1131. [Google Scholar]
  71. Yuksekgonul, M.; Bianchi, F.; Kalluri, P.; Jurafsky, D.; Zou, J. When and why vision-language models behave like bags-of-words, and what to do about it? In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  72. Liu, H.; Xue, W.; Chen, Y.; Chen, D.; Zhao, X.; Wang, K.; Hou, L.; Li, R.; Peng, W. A survey on hallucination in large vision-language models. arXiv 2024, arXiv:2402.00253. [Google Scholar]
  73. Zhang, J.; Hu, J.; Khayatkhoei, M.; Ilievski, F.; Sun, M. Exploring perceptual limitation of multimodal large language models. arXiv 2024, arXiv:2402.07384. [Google Scholar]
  74. Yang, S.; Zhai, B.; You, Q.; Yuan, J.; Yang, H.; Xu, C. Law of Vision Representation in MLLMs. arXiv 2024, arXiv:2408.16357. [Google Scholar]
  75. Li, Y.; Du, Y.; Zhou, K.; Wang, J.; Zhao, W.X.; Wen, J.-R. Evaluating object hallucination in large vision-language models. arXiv 2023, arXiv:2305.10355. [Google Scholar]
  76. Jaradat, S.; Nayak, R.; Paz, A.; Ashqar, H.I.; Elhenawy, M. Multitask Learning for Crash Analysis: A Fine-Tuned LLM Framework Using Twitter Data. Smart Cities 2024, 7, 2422–2465. [Google Scholar] [CrossRef]
  77. Masri, S.; Raddad, Y.; Khandaqji, F.; Ashqar, H.I.; Elhenawy, M. Transformer Models in Education: Summarizing Science Textbooks with AraBART, MT5, AraT5, and mBART. arXiv 2024, arXiv:2406.07692. [Google Scholar]
  78. Bai, G.; Chai, Z.; Ling, C.; Wang, S.; Lu, J.; Zhang, N.; Shi, T.; Yu, Z.; Zhu, M.; Zhang, Y.; et al. Beyond efficiency: A systematic survey of resource-efficient large language models. arXiv 2024, arXiv:2401.00625. [Google Scholar]
  79. Kwon, W.; Li, Z.; Zhuang, S.; Sheng, Y.; Zheng, L.; Yu, C.H.; Gonzalez, J.; Zhang, H.; Stoica, I. Efficient memory management for large language model serving with pagedattention. In Proceedings of the 29th Symposium on Operating Systems Principles, Koblenz, Germany, 23–26 October 2023; pp. 611–626. [Google Scholar]
Figure 1. A general architecture of an MLLM.
Figure 1. A general architecture of an MLLM.
Computation 13 00133 g001
Figure 2. Methodology of literature review in this study.
Figure 2. Methodology of literature review in this study.
Computation 13 00133 g002
Figure 3. An overview of the taxonomy of end-to-end object detection in transportation applications.
Figure 3. An overview of the taxonomy of end-to-end object detection in transportation applications.
Computation 13 00133 g003
Figure 4. Analysis of ChatGPT-4’s zero-shot learning in road feature recognition.
Figure 4. Analysis of ChatGPT-4’s zero-shot learning in road feature recognition.
Computation 13 00133 g004
Figure 5. Four example outputs from Gemini Pro Vision analysis with two frames.
Figure 5. Four example outputs from Gemini Pro Vision analysis with two frames.
Computation 13 00133 g005
Figure 6. Challenges and limitations of MLLMs in object detection for transportation applications.
Figure 6. Challenges and limitations of MLLMs in object detection for transportation applications.
Computation 13 00133 g006
Table 1. The relevance of the focused state-of-the-art to transportation applications.
Table 1. The relevance of the focused state-of-the-art to transportation applications.
StudyTitleTaxonomyTransportation Relevance
[45]Multimodal Driver Condition Monitoring System Operating in the Far-Infrared Spectrum Driver fatigue detection via thermal imagesHigh relevance for transportation safety
[6]Using Multimodal LLMs for Automated Detection of Traffic Safety-Critical EventsSafety-critical traffic events detectionHigh relevance for road safety and potential hazards
[46]GPT-4 Enhanced Multimodal Grounding for Autonomous Driving: Leveraging Cross-Modal Attention with Large Language ModelsObject detection for autonomous drivingHigh relevance for autonomous driving systems
[47]Semantic Scene Understanding with Large Language Models on Unmanned Aerial Vehicles Scene understanding for Unmanned Aerial VehiclesIndirect relevance but a potential application in aerial traffic monitoring
[48]Contextual Object Detection with Multimodal Large Language ModelsGeneral object detection in human-AI interaction scenariosIndirect relevance but a potential for object detection in varied contexts
[49]Multi-Modal GPT-4 Aided Action Planning and Reasoning for Self-driving VehiclesExplainable decision-making from monocular camerasHigh relevance for planning and reasoning in self-driving vehicles
[50]OpenAnnotate2: Multi-Modal Auto-Annotating for Autonomous DrivingIntegrates various modalities to improve annotation accuracy for large datasetsHigh relevance for auto-annotating data in autonomous driving systems
[51]Pix2Planning: End-to-End Planning by Vision-language Model for Autonomous Driving on Carla SimulatorEnd-to-end vision-language modeling for autonomous drivingHigh relevance for translation visual inputs into language-based trajectory planning
Table 2. Overview of the models and datasets used in previous studies.
Table 2. Overview of the models and datasets used in previous studies.
StudyDatasetData TypeMLLM/LLM UsedSizePerformance
[45]Thermal image datasetImagesChatGPT3.5-turbo + YOLOv8LargemAP = 0.94
[6]DRAMA datasetImages, textGemini Pro-Vision 1.5 MLLMMediumAccuracy = 79%
[46]Talk2CarImages, textGPT-4, Vision TransformerLargemAP = 0.75
[47]Custom UAV datasetImages, videoGPT-3, CLIP, YOLOv7Medium-
[48]COCO, CODE BenchmarkImages, textGPT-3, BLIP-2LargeAP = 43.4
[49]CARLAImages, text, and sensor dataGPT-4, Graph-of-Thought (GoT)Large-
[50]SemanticKITTIImages and multimodal sensor dataGPT-4, LangchainLarge-
[51]CARLAImages, BEV representations, and textAutoregressive transformerLargeDS = 92.39
RC = 99.60
  • mAP (mean average precision): indicates the average detection precision across various intersection-over-union (IoU) thresholds.
  • AP (average precision): used for specific object categories, reflecting detection accuracy.
  • TPR (True Positive Rate) and FPR (False Positive Rate): measure the ratio of correctly identified and wrongly identified objects, respectively.
  • Driving Score (DS): represents the percentage of successful driving decisions based on MLLM outputs.
  • Route Completion (RC): measures the model’s ability to follow a planned trajectory to completion.
Table 3. An example of a prompt utilized for interpreting Google Street View images to extract iRAP safety attributes.
Table 3. An example of a prompt utilized for interpreting Google Street View images to extract iRAP safety attributes.
Your role is systematically inspecting an image for various road features.
---
Road Inspection Checklist
1. Number of Lanes—Count the number of lanes in the direction closest to the camera. Provide the number without additional comment.
2. School Zone Warning—Inspect for school zone warning signs and select the appropriate option:
    1. School zone flashing beacons
    2. School zone static signs or road markings
    3. No school zone warning
    4. Not applicable (no school at the location)
3. School Zone Crossing Supervisor—Observe for the presence of a school zone crossing supervisor and select one:
    1. Supervisor present at school start and finish times
    2. Supervisor not present
    3. Not applicable (no school at the location)
4. Roads that Cars Can Read
    - Inspect the following and provide ‘yes’, ‘no’, or an applicable comment:
      - Clear and consistent road signage visible and unobstructed.
      - Well-defined road markings, maintained with high contrast for visibility.
5. Roadworks—Examine the area for roadworks and choose one:
    1. No roadworks
    2. Minor roadworks in progress
    3. Major roadworks in progress
6. Pedestrian Fencing—Check for pedestrian fencing (guard rails) and select one:
    1. Not present
    2. Present
7. Vehicle Parking—Inspect the area for vehicle parking and select one:
    1. None
    2. Each roadside has at least one parked vehicle
    3. At least one vehicle is parked on one side of the road
8. Speed Management/Traffic Calming —Identify any present speed management features such as curb buildouts, speed humps, raised tables, etc.
9. Median Type
    - Examine the image and identify any present median type:
    - Options include centreline, wide centreline, central hatching, continuous central turning lane, various physical median widths, safety barriers, etc.
10. Carriageway—Inspect the image and select one:
    1. Divided carriageway
    2. Undivided carriageway
11. Quality of curve—Inspect the image to identify any present road curvature then select one:
  1. Poor
  2. Not applicable
  3. Adequate
Computation 13 00133 i001
Map data: © 2024 Google.
Number of Lanes: there are four lanes in the direction closest to the camera.
School Zone Warning: no school zone warning
School Zone Crossing Supervisor: not applicable (no school at the location)
Roads that Cars Can Read: yes, clear and consistent road signage is visible and unobstructed.
Yes, well-defined road markings, maintained with high contrast for visibility.
Roadworks: no roadworks
Pedestrian Fencing: present, as there is a barrier between the pedestrian walkway and the road.
Vehicle Parking: none
Speed Management/Traffic Calming: No visible speed management features such as curb buildouts, speed humps, raised tables, etc.
Median Type: safety barriers are present between the opposite lanes of traffic.
Carriageway: divided carriageway
Quality of Curve: not applicable, as no significant road curvature is visible in the image provided.
Table 4. Comparative performance analysis of safety-critical event detection.
Table 4. Comparative performance analysis of safety-critical event detection.
Q/A StageGemini Pro Vision 1.0
with Two-Frame Input
Llava-7B 1.5
with Two-Frame Input
Gemini Pro Vision 1.0
with Video as Input
Risk55%46%75%
Scene90%76%100%
What83%71%66%
Which87%59%66%
Where61%49%66%
Recommended Action55%51%75%
Overall71.8%58.6%74.67%
Table 5. Performance breakdown of GPT4 and Gemini analyzing thermal images.
Table 5. Performance breakdown of GPT4 and Gemini analyzing thermal images.
MLLM ModelGPT4 Vision Preview Gemini 1.0 Pro Vision
Evaluation Metrics per CategoryTPRFPRTPRFPR
Category ID 1: person0.5700.390
Category ID 2: bike0.310.010.410.02
Category ID 3: car0.860.090.900.08
Category ID 4: motorcycle0.0800.240.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ashqar, H.I.; Jaber, A.; Alhadidi, T.I.; Elhenawy, M. Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing. Computation 2025, 13, 133. https://doi.org/10.3390/computation13060133

AMA Style

Ashqar HI, Jaber A, Alhadidi TI, Elhenawy M. Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing. Computation. 2025; 13(6):133. https://doi.org/10.3390/computation13060133

Chicago/Turabian Style

Ashqar, Huthaifa I., Ahmed Jaber, Taqwa I. Alhadidi, and Mohammed Elhenawy. 2025. "Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing" Computation 13, no. 6: 133. https://doi.org/10.3390/computation13060133

APA Style

Ashqar, H. I., Jaber, A., Alhadidi, T. I., & Elhenawy, M. (2025). Advancing Object Detection in Transportation with Multimodal Large Language Models (MLLMs): A Comprehensive Review and Empirical Testing. Computation, 13(6), 133. https://doi.org/10.3390/computation13060133

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop