Next Article in Journal
Advanced In Situ TEM Microchip with Excellent Temperature Uniformity and High Spatial Resolution
Previous Article in Journal
E2E-RDS: Efficient End-to-End Ransomware Detection System Based on Static-Based ML and Vision-Based DL Approaches
Previous Article in Special Issue
Smart Task Assistance in Mixed Reality for Astronauts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Review of IoT Sensor Systems Used for Monitoring the Road Infrastructure

Department of Cybernetics and Artificial Intelligence, Faculty of Electrical Engineering and Informatics, Technical University of Kosice, 042 00 Kosice, Slovakia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(9), 4469; https://doi.org/10.3390/s23094469
Submission received: 6 April 2023 / Revised: 26 April 2023 / Accepted: 2 May 2023 / Published: 4 May 2023
(This article belongs to the Special Issue Vision Sensors for Object Detection and Recognition)

Abstract

:
An intelligent transportation system is one of the fundamental goals of the smart city concept. The Internet of Things (IoT) concept is a basic instrument to digitalize and automatize the process in the intelligent transportation system. Digitalization via the IoT concept enables the automatic collection of data usable for management in the transportation system. The IoT concept includes a system of sensors, actuators, control units and computational distribution among the edge, fog and cloud layers. The study proposes a taxonomy of sensors used for monitoring tasks based on motion detection and object tracking in intelligent transportation system tasks. The sensor’s taxonomy helps to categorize the sensors based on working principles, installation or maintenance methods and other categories. The sensor’s categorization enables us to compare the effectiveness of each sensor’s system. Monitoring tasks are analyzed, categorized, and solved in intelligent transportation systems based on a literature review and focusing on motion detection and object tracking methods. A literature survey of sensor systems used for monitoring tasks in the intelligent transportation system was performed according to sensor and monitoring task categorization. In this review, we analyzed the achieved results to measure, sense, or classify events in intelligent transportation system monitoring tasks. The review conclusions were used to propose an architecture of the universal sensor system for common monitoring tasks based on motion detection and object tracking methods in intelligent transportation tasks. The proposed architecture was built and tested for the first experimental results in the case study scenario. Finally, we propose methods that could significantly improve the results in the following research.

1. Introduction

Transportation systems are the backbone of every economy in the world. This study will focus on the Internet of Things (IoT) concepts used in road transportation infrastructure. IoT concepts applied to road transportation infrastructure can create the part of an intelligent transportation system. An intelligent transportation system could increase traffic efficiency and safety because the road’s monitoring system would be reliable. The automatization of monitoring systems can be performed in different ways, and this article compares the advantages and disadvantages of the most commonly used ways.
Organizing fluent and safe traffic is a priority in every country. By defining monitoring tasks in intelligent transportation systems, it is necessary to achieve this goal and characterize the data to be collected. In compliance with [1], there are measured data about safety, the diagnostics of a vehicle or infrastructure, traffic management, driver’s assistance, environment’s impact and user’s ability to drive. We focused on a comparison of the sensor systems collecting traffic management and safety data. The sensor systems used for traffic management and safety data use the techniques for vehicle motion detection and object tracking. The traffic and safety management system requires monitoring tasks such as vehicle type classification and counting, automatic license plate recognition, incident detection, parking management and speed measurement. Regarding the opinion of the authors in the scientific article [2], many sensor systems have been developed to solve these tasks, but their effectiveness has not significantly improved over the past decade. We propose a research direction where using the IoT concept via computational distribution on edge computing in sensor systems can improve the sensor system’s effectiveness.
This paper is organized as follows. The IoT concept is introduced in Section 2. The IoT concept includes several main parts, such as sensors, control units, actuators and computational distribution. We focus on computational distribution and propose our taxonomy according to its purpose in transportation tasks as well as a taxonomy from the view of heterogeneity sensor data. Section 3 briefly explains the main working principles of sensors used for intelligent transportation tasks. Section 4 describes transportation tasks based on motion detection and object tracking. Section 5 compares sensor systems’ effectiveness, as mentioned in Section 3, in collecting data for the tasks mentioned in Section 4 based on a state-of-the-art literature survey. We rely on the comparison results in Section 5 to select the sensors for our architecture of the universal sensor system in Section 6. Section 7 presents a case study scenario where the proposed architecture of a universal sensor system is described, as well as its implementation and the first experimental results. We discuss research challenges and gaps in Section 8 and conclude them in Section 9.

2. IoT Concepts

The IoT is a concept of connecting devices to the internet. This connection helps to create a digital copy of the physical world. The main attributes of the IoT concept are sensors, actuators, control units, and the internet connecting these entities. The Internet is an instrument that connects each part of the IoT concept, and actuators are devices used to execute actions in the IoT concept. In the opinion of the authors [3,4], the IoT concept used for intelligent transportation systems is an important step in improving the quality of the transportation industry. In this section, we introduce the taxonomy of sensors and the possibility of computation distribution on edge, fog, and cloud computing.

2.1. Taxonomy of Sensors

Further, a sensor is defined as an entity that retrieves the state of a measured or sensed object and sends the data or information to a control entity [5]. We can use living organisms [6] or some devices [7] as sensors.
We propose the categorization sensors according to the taxonomy mentioned in Figure 1. Sensors could be divided into physical or virtual ones. Physical sensors are devices directly measuring the state of sensed things, and the particular state is being send to a control or storage unit [8]. Virtual sensors are entities that indirectly measure an object’s state via computation, aggregation, or other processing methods from gathered data [9,10].
Virtual sensors can be divided into three categories [11,12]:
  • Software-based;
  • Hardware-based;
  • A combination of software- and hardware-based.
Software-based virtual sensors are devices with a deployed program that senses the state of an object in a virtual world. For example, it describes the model of a neural network that is trained for sentiment analysis used to measure aggressiveness behaviour in forum conversation. This model changes the computer into a social behaviour sensor [13,14,15].
The hardware-based virtual sensor shows the object’s measured value from the computed value of two or more physical sensors connected to a computer or a microcontroller unit (MCU). For instance, temperature and humidity sensors connected to MCU, fuse data using the computation of atmospheric pressure in the observed environment. The MCU computes the pressure value and behaves like a virtual barometer [12].
Virtual sensors, made with fusion software- and hardware-based approach, are physical sensors connected to a control unit collecting data from the other virtual sensor to compute the measured objects [11]. For example, two cameras of the same type could be used for stereovision purposes. The original purpose of these cameras was to take pictures or videos. The algorithms for chess calibration and matching pixels’ positions in pictures taken by both cameras, based on the fact that the positions of each camera are known, are used for distance calculation of the objects of interest from the taken pictures. The software using the algorithm to calculate the distance of each pixel from taken pictures is a virtual sensor for distance measurement. A virtual sensor transforms pixels to voxels. The cameras are physical sensors that provide the data for calculation.
The physical sensors mentioned in this paper are divided into three main categories [16,17]:
  • mechanical;
  • electronic;
  • biological.
The mechanical sensor we describe is a sensor that measures the value of an object’s state in the mechanical meter; for example, the volume of the mercury in the glass cylinder measures temperature [18].
An electronic sensor senses the object’s state by converting physical quantities into electrical signals; for example, piezoelectric sensors make electric impulses when they are impacted by pressure deformation [19].
There is correlation to the fact that a biological sensor is a living organism that shows the status of the measured environment’s state by changing its behaviour, such as Syngonium Podophyllum becoming fluorescent in soil that contains Arsenicum or other heavy metals [6].

2.2. Edge and Fog Computing

There are many definitions of edge and fog computing. Some authors consider these words synonymous, but we distinguish between the edge and fog computing paradigms (Figure 2). Both paradigms are based on computing in the nearest node to the data source, but they can be used for other purposes [20,21,22].
Edge computing runs on edge devices such as smartphones, single-board computers (SBC), or MCUs directly connected to the sensors or actuators. Within them there are implemented programs that process the data received from sensors that are sent to the central computational node [23] or directly to the actuator. It is possible to convert the edge device to a virtual sensor. The edge device with implemented software can sense other measured phenomena via computation from acquired data from the connected physical sensor [12].
We propose to use terminology in fog computing, defined as computational devices controlling and making complex analysis above the other edge devices connected in the same local network. As fog computing, we can use all devices mentioned in the edge computing paradigm as well as personal computers (PC) or server computers. The condition is to design enough computational capacities for complex algorithms, methods and analysis [24].
Figure 2. The architecture of IoT computing distribution [25].
Figure 2. The architecture of IoT computing distribution [25].
Sensors 23 04469 g002

2.3. Cloud Computing

The cloud computing paradigm has gained popularity in the last decade. In the literature, we can find many definitions of this concept. These can be defined as a computing paradigm serving the central complex computations and storage in specialized data centres. Cloud has a centralized approach, and the fog has decentralized control over the edge devices and sensors. Hereby, the centralized approach in cloud computing from the side of software abstraction. Physically, the applications implemented in the cloud could run across many connected computational devices. In cloud computing, all connected devices have to behave as one computational device. The user should not recognize the distribution of computational capacity across nodes between servers in cloud applications [23,26].

3. Transportation Sensoric Systems

Most countries and cities try to optimize the traffic and transportation control via IoT concepts to develop smart cities and smart transportation concepts. A lot of different routes have implemented IoT concepts. Still, the development of new sensors, computational devices, and protocol concepts create new possibilities to improve the effectiveness of these systems including maintenance, financial and computational capacities [27,28]. The mentioned tasks in transportation have their specifications creating a constraint to install some kinds of sensors. One of the vital requirements of the sensor is the ability to produce actual data where the control unit considers these data. In the intelligent transportation system, we use physical sensors categorized into electrical sensors in our taxonomy mentioned in Section 2.1. There could be created either a virtual sensor based on these physical sensors. In this section, there is describe a survey of the most common sensors used in transportation and traffic systems and their taxonomy shown in Table 1 from the view of deployment.
In real-time data analysis, intelligent transportation systems use sensors that measure humidity, temperature, wind speed, precipitation, and visibility. They play a critical role in forecasting weather conditions that impact traffic safety. However, this paper does not include the sensors used for the weather forecast in the taxonomy of sensors mentioned in Table 1 because they do not solve tasks based on motion detection and object tracking.

3.1. Intrusive Sensors

Intrusive sensors are sensors installed on pavement surfaces. Intrusive sensors are also categorized as in-roadway-based sensors. They are accurate for vehicle classification performance for transportation monitoring tasks and processing other data types such as speed or weight measurement [29].

3.1.1. Piezoelectric Sensors

Piezoelectric sensors are based on piezoelectric materials such as barium titanate (BaTiO3) or lead titanate (PbTiO3). These materials generate an electrical signal when subjected to mechanical impact or vibration. The measured signal is proportional to the force or vehicle weight [19].

3.1.2. Fiber Bragg Gratting Sensors

The fiber Bragg grating (FBG) sensor is used in civil engineering applications to measure temperature, strains, and loads [30,31]. FBG sensors are sensitive to damage from careless manipulation throughout the installation process because they are built from silica material. There are used many kinds of materials for packaging. The packaging material, made from glass-fiber-reinforced polymer, provides reliable protection to the FBG sensor. The reliable packaging material enables the use of FBG sensors for monitoring the traffic situation in the intelligent transportation system [32].

3.1.3. Pneumatic Road Tube

A pneumatic road tube placed on the road’s pavement being connected to the sensor mechanism. This sensor’s mechanism is proposed for producing an electrical signal from a burst of air pressure created when the vehicle’s tires pass over the tube. The received signals from this sensor can classify and count vehicles passing through the road section [33].

3.1.4. Vibration Sensor

It is possible to build vibration sensors using various technologies. Choosing the optimal materials depends on the environment and the degree of sensitivity that we have to use to implement the vibration sensing system. The study [34] proposed using the road pavement as a source of vibration transduced to the electrical signal. Microelectromechanical system (MEMS) accelerometers transduce generated vibrations to electrical signals. A control unit processes these electrical signals to classify and count the vehicles on the road section.

3.1.5. Loop Detector

An inductive loop detector is a coil of metal wire. The wire captures changes in inductance. Changes in inductance generate a time-variable signal of vehicles passed through this detector. According to the generated signal, the computation unit can classify the vehicles’ presence and categories via amplitude, phase, and frequency spectrum analysis [35].
Depending on the installation method, there are two types of loop detectors: saw-cut and performed loops. In the saw-cut method, the installation process needs to saw-cut the pavement to lay the loop wires under the pavement. The loop wires require protection by filling the pavement after installation. The performed loop detectors are not installed under the pavement. In this method, instead of saw-cut pavement, the loop wires are installed inside the PVC pipe. The PVC pipe is attached to the pavement [36].

3.1.6. Magnetic Sensors

Every vehicle’s construction consists of ferrous metals inducing disturbance to the Earth’s magnetic field. The function of magnetic sensors is to measure the Earth’s magnetic field distortion caused by passing vehicles through the sensor. Magnetic sensors are more energy-efficient with smaller size, weight and cost than inductive loop detectors [37].

3.2. Non-Intrusive Sensors

Non-intrusive sensors are not installed in the pavement surface to compare intrusive sensors. They could be installed on the sides of or over the roadway. Non-intrusive sensors are more expensive to produce than intrusive sensors. However, the installation process of non-intrusive sensors is cheaper than that of intrusive ones because saw-cutting or digging under the road surface is considered technically difficult and expensive. The maintenance of road surfaces with installed intrusive sensors as well as standalone intrusive sensors requires an expensive technological approach. In conclusion, the development cost of reliable non-intrusive sensors for intelligent transportation tasks, including the installation, maintenance and service costs, might be cheaper than intrusive sensors. Still, non-intrusive sensors could classify with less accuracy because weather conditions and other environmental events not connected to the transportation could significantly impact classification accuracy processing of other traffic monitoring information such as speed measurement or free parking slot management [29].

3.2.1. Radio Frequency Identification

Radio frequency identification (RFID) sensors can be categorized into two main categories:
  • RFID with passive tags;
  • RFID with active tags.
RFID with passive tags consists of two components. The first component is the RFID reader, which reads (senses) data and sends an electromagnetic signal. The second component is the RFID tag, which transmits the data stored in a memory chip via the specific radio frequency signal. The passive RFID tags do not use an internal power source. They consider obtaining electrical energy from the RFID reader’s electromagnetic signal conducted to electricity. Passive RFID tags can only be applied for short distances due to the fact that they miss power source limits by the reader’s electromagnetic signal strength.
The RFID with active tags works similarly to the RFID with passive tags. The difference is the working principle of the RFID tag. Active RFID tag has an internal power source and automatically transmits the data to the reader after connecting to the electricity. We can use active RFID tags for longer distances than passive RFID tags [38].
RFID sensors are used for free parking slot management or vehicle identification. Active RFID tags enable the transmission of a vehicle’s diagnostics about the expected service date or daily fuel consumption. Vehicles can then be located inside tunnels via active RFID tags. This solution is a good alternative to the GPS navigation system because GPS navigation does not work inside tunnels or buildings [39].

3.2.2. Acoustic Sensor

Acoustic sensors are based on an array of microphones. They capture audio signals induced by passing vehicles. For vehicle classification, we have to extract features from the captured sound. This task is challenging because a sound can contain much noise, and every vehicle has its specific sound. The production of electromobility cars responds to the fact that these cars might be more silent than traditional ones using petrol or diesel vehicles. Acoustic sensors can be used in addition to other sensors [40].

3.2.3. Laser Scanner

A laser scanner is a sensor measuring the distances between its position and the object of interest via a laser beam. Laser scanners are divided into several categories depending on the working principle of distance calculation. One type of laser scanner is a light detection and ranging (Lidar) scanner. The Lidar working principle and its categorization are described as a standalone sensor in our taxonomy (Section 3) because Lidars use similar techniques to radars for distance measuring [41]. However, comparing the radars, Lidars use lasers instead the radio waves. As mentioned above, there exist other types of laser scanners except Lidars. Those types of laser scanners are based on triangulation, time-of-flight (ToF), phase-shift and structured light scanners [42].
Laser scanners based on triangulation use a determined camera position, laser position and angle of laser beam emission to calculate the point’s depth.
ToF laser scanners measure the time of laser beam flight to the object of interest. Regarding the measured time, the ToF laser scanner computes the distance between the laser beam emitter and the object of interest. ToF scanners use a camera or similar photosensitive device to simultaneously measure the depth of many points in the camera’s specific narrow range of visibility.
Phase-shift laser scanners use a continuous laser beam. The sensor inside the scanner calculates the distance via phase-shift of emitted and reflected waveforms of the laser beam.
Structured light laser scanners use a pattern emitted via a laser beam onto an object. The deformation of pattern shape onto an object’s surface is processed through the camera system.
The laser scanner is used to take the shapes of vehicles via a calculated depth map [43]. Laser scanners based on ToF or the phase-shift principle are more expensive than cameras, and laser scanners produce data of lower quality than cameras [41].

3.2.4. Radar

Radio detection and ranging (Radar) is a sensor that measures the presence and distance of various objects. The main principle consists of transmitting and reflecting radio waves from scanned environments [44]. Radar is less accurate than Lidar. Radar provides only a 2D map of object localization, and Lidar provides a 3D representation of the scanned environment. On the other hand, Radar is less weather-sensitive than Lidar [41].

3.2.5. Wi-Fi

A wireless field (Wi-Fi) has a purpose as a network protocol for sharing internet access by radio waves. This protocol was used for collecting data along highways. The principle of using radio waves for data exchange between devices enables the localisation of the connected device through the strength of the signal from the Wi-Fi transceiver [45]. Comparing the signal’s strength between the Wi-Fi receiver and transmitter also enables detection of vehicle motion [46,47].

3.2.6. Lidar

Lidar works on similar principles to radar, but it has some specifications. Some light sources are used, commonly lasers, to transmit the signal into the environment, and photosensitive elements receive the reflected signal. The calculation of the object’s range, presence, and shape is the same as in the Radar sensor mentioned above [48].

3.2.7. Passive Infrared Sensor

A passive infrared sensor (PIR) detects the object’s presence by generated heat energy from vehicles or other objects. Every object that generates heat shines a ray in the infrared spectrum. A Fresnel lens focuses the ray onto the sensing element, which transforms the infrared signal into an electric signal. The PIR sensor is widely used for motion detection in a small range [49].

3.2.8. Ultrasonic Sensor

Ultrasonic sensors use sound waves transmitted at high frequencies not audible to human ears to compute distances between objects. Sensors detect reflected sound wave energy and convert it into an electric signal. The control unit has to process the electric signal. It measures the time difference between transmitted and received sound waves to calculate the distance of the obstacle. Ultrasonic sensors are widely used for measuring vehicle flow or free parking slot management [50,51]. The main disadvantage of ultrasonic sensors are their high sensitivity to environmental noises [52].

3.2.9. Camera Systems

Cameras are image sensors (Figure 3). The sensor captures images in 2D that could reconstruct a 3D object with special techniques, such as stereovision. All image sensors consist of a photodetector array called pixels. These pixels transform electromagnetic waves in the visible light spectrum into electric signals [53]. The methods of transforming visible light into electric signals divide the image sensors into two types: charged coupled device (CCD) and complementary metal oxide semiconductor (CMOS) [54].
CCD sensor pixels consist of individual metal oxide semiconductor (MOS) diodes. These diodes are arranged in rows and columns. The rows and columns create a matrix where the energy of visible light charges every diode proportional to the light’s intensity. The control unit in the CCD camera uses a mechanism to transform the diode’s charge from each pixel in the matrix to the electric signal. The CCD image sensors take pictures with high quality. The disadvantage of CCD cameras is the processing method of pixels. The camera’s control unit has to process every pixel sequentially to reconstruct the image, which is time-consuming and energy-consuming [53,54,56].
CMOS sensor pixels are composed of photodiodes and CMOS transistors. Every CMOS transistor works as a switch generating a digital electric signal from transformed visible light. The main advantage is the possibility of parallel processing of every pixel by the camera’s control unit. It increases the speed of taking pictures with low power consumption. The CMOS image sensors are cheaper than CCD image sensors [57].
The increased quality of camera sensors and suitable computer vision (CV) methods and algorithms enable us to process and classify many tasks in intelligent transportation. The possibility of camera sensors is deeply described in Section 5 and Section 6.

3.3. Vehicle Sensors

The reliability of autonomous or semi-autonomous vehicles depends on collecting the actual data from the traffic situation. It is necessary to install sensors in the vehicle to obtain actual data from the road. Table 2 shows the taxonomy of in-vehicle sensors and non-vehicle sensors. This paper defines in-vehicle sensors as sensors that could be used inside a vehicle’s construction to monitor the vehicle status or environment around the vehicle. Non-vehicle sensors are defined as sensors that cannot be installed into a vehicle’s construction, and are only used as part of road infrastructure.
In-vehicle sensors can also be used to transform cars into mobile sensor systems in the IoT concept of communication vehicle to infrastructure (V2I). The possibility of using a car as a sensor is described in (Section 5).

4. Transportation Tasks Based on Motion Detection and Object Tracking

This section describes the main transportation tasks used for monitoring traffic or infrastructure. The monitoring tasks are focused on the object’s motion detection and object tracking. Automatized monitoring tasks via IoT concepts in transportation could improve the conditions of traffic situations.
In-vehicle sensors could detect or warn the driver before a collision with other objects [58]. Development and research in artificial intelligence (AI) applications enable the building of driverless vehicles that consider data collected from sensors in real-time and process these to control vehicles while maintaining security standards [59].

4.1. High-Occupancy Vehicle Lane Management

Several countries try to reduce traffic congestion by motivating citizens to share cars or use public transportation. For this purpose, high-occupancy vehicle (HOV) lanes have been built, which serve only vehicles with two or more passengers. These HOV lanes enable drivers to reach their target destination quickly, avoiding potential traffic congestion [60]. HOV lane management systems are proposed to check the count of passengers in every vehicle. The system sorts the vehicles that have permission to use the HOV lanes. There are two options for checking the passenger count in the vehicles. The first one is using sensors for person detection from the checking station [61]. The second one is using seat occupancy sensors installed on the car’s interior [62]. The information extracted from the sensors could be sent to the checking station via the dedicated communication protocol.

4.2. Incident Detection

We detect incidents via various methods and different systems. Transportation incidents could happen in various scenarios, such as a vehicle fire inside a tunnel [63], traffic jams resulting from a damaged car [64], or chain car accidents [65]. For this purpose, many different types of sensors and communication devices are used to contact emergency services. We can divide incident identification into two aspects. The identification of incidents via infrastructure—for example, the camera surveillance system in a tunnel identifies a fire in a tunnel, which alarms the operator to begin the rescue of people inside the tunnel [66]. The second possibility is using in-vehicle sensors. For example, the ultrasonic sensors detect the collision, and the car’s board computer contacts the emergency service by wireless internet connection [67].

4.3. Vehicle Counting

Many paid highways use a counting vehicle system to monitor usage [68]. The counting vehicle system produces information about the frequency of visiting the routes. This information is used in future planning for maintenance or building new highways or roads [69,70]. The advanced counting vehicle system categorizes vehicles into many categories, such as buses, trucks, or cars (Section 4.5). These collected data could help manage traffic.

4.4. License Plate Recognition

License plate recognition is an essential task in road surveillance. This task can identify cars with detected anomalous behavior. We can usually apply the methods of automatic license plate reading after finding the car’s position via some object detector. According to [71], the automatic license plate recognition (ALPR), process is divided into three steps:
  • Localize the position of the license plate in car.
  • Segment characters from the background.
  • Apply optical character recognition (OCR) methods.
This task is essential for checking the priced parking places [72] or identification of stolen cars [73].

4.5. Vehicle Type Classification

Many roads have some restrictions on the types of vehicles they allow because of safety, flow, and maintenance costs. Trucks, due to their heavy weight, are prohibited from using some routes [74]. In many countries, truck drivers have to pay extra charges to use highways due to higher maintenance costs [75]. Sometimes, the large size of trucks or buses limits passing through some sections without a traffic incident [76]. The data collected from the categorized count of vehicles by particular sections of the roads can improve the management of traffic lights to avoid traffic jams [66,77]. This information can be used by the dispatchers of public transport to react to special events.

4.6. Speed Measurement

Safety is considered a target priority in transportation. High-speed driving vehicles increase the risk of an accident. The increased kinetic energy of high-speed driving impacts the crash consequences after the collision with another vehicle or object. The risk of accidents in high-speed driving is related to limited human reaction time. States worldwide use speed limit restrictions in traffic management to reduce accident risks [78]. One of the aims of an intelligent transportation system is to check adherence to speed limit restrictions. The police use sensor systems that detect vehicle motion and track vector movements for speed measurement [79].

4.7. Parking Management

The increased number of people buying the cars influence the fact that they are widely used as a kind of individual transport method. Many people ride to work, school, or shop by car, occupying parking slots across a city or town. Free parking slots are a limited resource in many cities; therefore, many parking houses or places are priced [80]. Priced parking places require some automatic system of checking all the car owners to pay for parking slots [72]. A lot of automatic parking place checking is based on a sensor system similar to the vehicle counting system.

4.8. Weight Measurement

Worldwide, there are many specialized sections of roads that have specific vehicle weight constraints. These constraints exist because of road maintenance or security reasons [74]. To avoid the entry of overweight vehicles, sensors are installed that measure the vehicle’s weight, and the system can close the entry gate. For example, bridges are places where a vehicle’s axle weight restriction is necessary to keep traffic safe [81].

5. Applications the Sensor System in Transportation Tasks

In this section, we introduce the transportation monitoring tasks categories mentioned in (Section 4), which could be solvable sensor systems, as mentioned in (Section 3). Table 3 represents state of the art in sensor systems used for intelligent monitoring transportation tasks in the last decade. The literature review in Table 3 refers to identification of research challenge for each sensor system in tasks mentioned in the table. Table 3 uses a literature survey and review to compare the transportation tasks mentioned in (Section 4) and the sensor system mentioned in (Section 3).
Table 3 describes the potential of mentioned sensors to solve transportation monitoring tasks based on motion detection and object tracking. From the sensors mentioned (Section 3), we can conclude that we may solve many monitoring transportation tasks via the camera system in almost all applications except for weight measurement. CV methods and algorithms enable the creation of a virtual sensor (Figure 1) from the camera to sense many objects. The research’s progress in traditional CV methods and artificial deep neural networks makes motion-detection and object-tracking methods more effective [157,158].
Intrusive sensors such as magnetic sensors, loop detectors or pneumatic road tubes use a specific installation configuration for each purpose (Table 3). This installation configuration impacts the detector sensitivity level. Each monitoring task requires a different sensitivity level. For example, single- and dual-loop detector configurations have different sensitivity levels. The parking slot management system does not need to categorize the vehicle; therefore, the single-loop detector is efficient for parking slot management. In monitoring highways, the information about counted trucks and other vehicle types has a significant role. Dual-loop detectors increase the vehicle categorization accuracy [41].
FBG sensors are the most prospective intrusive sensors (Table 3). Suitable installation configurations and machine learning models can classify many events. The disadvantage is a calibration system. The advantage of the FBG sensor is the low-cost manufacturing investments compared with metallic sensors, long service life, and low maintenance cost. The FBG sensor is categorised as a virtual sensor source for traffic monitoring tasks in intelligent transportation systems because it measures a reflected light spectrum due to the fibre strain changes. It requires a complex, trained machine learning model to filter and extract valuable data for vehicle presence, speed estimation or weight measurement [32,159].
This article proposes categorising intrusive sensors, except FBG sensors, into physical sensors because the installation configuration classifies the vehicle presence directly via simple signal processing. These sensors consider common vehicle features that every vehicle consists of metal parts. The metal parts change the magnetic field’s behaviour (magnetometer, loop detector) or impact the vehicle’s weight characteristic (pneumatic road tube) [41].
In conclusion, intrusive sensors except FBG sensors have limited usage for collecting complex data about the actual situation on the road.
Article [2] categorises a camera system, Lidars, Radars, and passive infrared sensors as vision-based sensors. These sensors produce an output in map or image form. We propose categorising these sensors as virtual sensors for intelligent transportation system monitoring tasks because CV methods can extract valuable information from vision-based sensor data (Section 6). Sensors such as cameras or Lidars require other types of CV methods than radars or PIR sensors. PIR sensors detect a level of infrared radiation. Thresholding or neural networks are the only CV methods that can effectively use a control unit from data obtained from one PIR sensor. PIR sensors use thresholding for motion detection, but the sensors do not recognize what makes a motion [160]. Radar produces a 2D map of an object’s shapes and positions. On this map, we can use pattern recognition and advanced feature extraction approaches from CV methods, such as corner points or edge detectors [161,162]. Lidar can make a 3D map of the sensed environment. The measured points of the Lidar’s 3D map with the suitable transformations enable the reconstruction image with shades of selected colours palette where the shades represent the measured points’ distance from the selected perspective. The reconstructed image can use similar CV methods to images made using camera systems [163]. However, the Lidar scanning depth and missing objects’ colour data do not enable using all CV methods.
In Table 2 (Section 3.3), we mention the division of sensors by non-vehicle and in-vehicle sensors. and we used in-vehicle sensors in vehicle-to-vehicle (V2V) interaction or V2I interaction. Non-vehicle sensors are used for infrastructure monitoring systems [1]. In-vehicle sensors are served for the driver’s safety management to avoid collision with another object [164]. The Wi-Fi system can serve not only as a sensor system but also as a communication way to interact with emergency services in case of an accident [1]. Recently, many companies have tried to implement autonomous driverless vehicles that have to gain complex information about the actual situation from the road environment. Sensors such as ultrasonic or radar sensors are inefficient in obtaining the necessary data for autonomous vehicles [2]. Ultrasonic sensors have a small range of detection objects, and Radar has less accuracy than Lidar. For Lidar, it is possible to reconstruct the 3D scene for long distances, which can help to avoid collisions with other vehicles on the road. The reconstruction of the 3D scene is also possible with camera systems, achieving similar results to Lidars [42]. Furthermore, Lidars cannot read a traffic sign, where a camera system can.
Referring to Table 3, intrusive sensors such as loop detectors or magnetometers are limited to use in the HOV lanes management system. However, almost all vision-based sensors can check a vehicle’s occupancy status. Despite the fact that vision-based sensors are suitable for this purpose, there are still some problems. Lidars and laser scanners are too expensive for installing to the car’s interior only for counting the passengers. An automatic passenger-counting system based on a camera system needs to be installed at distances from which the person’s face could be recognizable. This approach is intrusive. The optimal solution is using sensors detecting passengers via seat occupation and sending the passenger count from the vehicle with V2I protocol to the HOV lane station.
The parking assistance system is one of the standard systems being implemented in cars. This system helps the driver search for a free parking slot and successfully park without collision. Parking houses can monitor free parking slots with a huge number of intrusive sensors that are costly to implement and maintain or with few cameras which can cover the same area with cheaper investment. Cameras can navigate cars across a parking lot and help calculate the trajectory necessary to reach a free parking slot. Parking management systems will become a vital part of the intelligent transportation system, with which V2I will cooperate [80,165,166,167].
Intrusive sensors have limitations in gaining complex data about the actual situation in monitoring road tasks. All intrusive sensors have a smaller distance detection range than non-intrusive sensors. These sensors can reliably categorize and count vehicles, and some of them, such as a magnetometer or FBG sensor, can measure speed. On the other hand, the expensive installation process limits their multiple usages in the same route. The most significant advantage of intrusive sensors is that they can only measure a vehicle’s weight. However, not all intrusive sensors can measure a vehicle’s weight, for example, loop detectors or magnetometers.
FBG sensors can monitor almost all tasks mentioned in (Table 3) except the ALPR task. This sensor can be used as a universal intrusive virtual hardware-based sensor for intelligent transportation systems tasks because it can solve most tasks from intrusive sensors (Table 3).
Camera systems visually observe roads by making video records. CV methods can extract a lot more information from these video records. High-resolution images or videos increase details and data quality for the following analysis process. Optimized convolutional neural network (CNN) models detect many objects of interest via classification rectangle (object detection methods) [168]. Object-detection methods are used for almost all monitoring tasks (Table 3). The CV methods that run on the camera system videos form a universal non-intrusive virtual sensor in the intelligent transportation system (Table 3).

6. Universal Sensor

This literature survey in Section 5 concludes and summarises the possibility and flexibility of the sensors mentioned in Section 3. Many camera systems connected to computational resources in edge, fog or cloud servers provide the specific purpose of monitoring tasks in the intelligent transportation system. The camera system produces image data containing useful information for selected monitoring tasks in Section 4. On the same image or video stream, other CV methods and algorithms can be simultaneously applied that extract or sense different events or objects independently. Table 3 summarises the camera system’s usability as well as the usability of other sensors for selected monitoring tasks (Section 4). Table 3 concludes that the camera system is a non-intrusive sensor that can solve most tasks by comparing other sensors. Camera system can be a declared as universal sensor for monitoring tasks because mentioned reasons above (Section 4).
The construction sensor system consists of a suitable surveillance camera and embedded SBC to the camera with Linux distribution. The embedded SBC in the camera allows the creation of a universal software-based virtual edge-enabled intelligent sensor system (EEISS) for most monitoring tasks (Section 4) in the intelligent transportation system. The purpose of the EEISS depends on the implemented and uploaded software solution based on plug-in (service-oriented) architecture. Each monitoring task (Section 4) has its own software plug-in (service) for processing and classifying the same camera’s data. The plug-ins (services) provide the flexibility of purpose EEISS based on the requirements for the system. In our case, the plug-ins (services) represent the software-based virtual part of the sensing system in the EEISS. For example, independent plug-ins (services) can be the speed estimation, ALPR or the vehicle’s motion detection and tracking software embedded into EEISS. The EEISS also allows additional data from other sensors, such as FBG sensors or Lidar, to these software-based plug-ins (services). In Section 7, we introduce the case study scenario of the development and usage of the EEISS.

7. Case Study Scenario

In this section, we present a case study that shows the effectiveness of the EEISS in providing complex information from selected monitoring tasks in the intelligent transportation system. The EEISS allows significantly improved flexibility of the sensor system for mentioned monitoring tasks. This flexibility helps to fill the market gap where a similar solution to the EEISS in the market is not available at this moment. This case study aims to prove the achievability of developing and building the EEISS for intelligent transportation system purposes. The case study consists of the proposed architecture (Figure 4) based on a state-of-the-art literature review, implementing the architecture’s prototype and demonstrating the first experimental results in real-life conditions.
We implemented and deployed the testing of a mast-mounted surveillance station based on this architecture in real-world conditions on a selected highway. Our research group can upgrade and optimize hardware and software solutions for this testing station, including changing various vision-based sensors according to their effectiveness. The testing station architecture is described below in 5 points.
(1) The intrusive pressure sensor (for example FBG sensor) is installed on the road surface to measure the vehicle’s weight. The sensor sends data to the fog computing layer (4) for following analysis.
(2) The mast-mounted surveillance system consists of a camera system. This system sends a video stream to the edge computing layer (3).
(3) The edge computing layer consists of SBCs embedded in the camera system construction (2). SBCs run firmware with computer vision methods and algorithms directly into the Camera System. These methods and algorithms of vehicles detect motion and classify the situation on the road. If the SBC system detects a vehicle presence, it starts the ALPR system or other services to check the traffic situation mentioned in (Section 4). The edge computing layer sends the classified, collected and aggregated data to the fog computing layer (4) to analyse the complex situation.
(4) The fog computing layer represents the microserver connected near the mast-mounted surveillance system. The microserver directly connects the intrusive pressure sensor and the system of SBCs with its cameras (EEISS). This microserver collects sensor data and controls the data flow to store or send to the cloud computing layer.
(5) The cloud computing layer uses a rent server to provide many applications for traffic monitoring and management system. It stores data for the following data mining projects or contacts emergency services in case of accidents.
The proposed and built architecture demonstrated the interesting first partial results described below in Figure 5. The picture on the left in Figure 5 shows the yellow rectangle (1) representing the road surface’s free slot from vehicle presence. The picture in the middle of Figure 5 describes the same rectangle (1) with green colour representing a positive motion detection of vehicle presence. After positive motion detection, the SBC’s program takes data from the intrusive piezoelectric sensor (2) in Figure 5 about the vehicle’s weight or category and starts the ALPR program from another camera. The ALPR camera’s picture is shown on the right of Figure 5.
The piezoelectric sensor has been installed into the road surface in this route for over two years, so the firmware is calibrated with high reliability to measure the weight and classify the vehicle type. The firmware accuracy for both tasks is over 97%. There is tendency to use the accuracy of piezoelectric sensors for the automatic annotation of image-based data from the cameras for machine learning purposes. Collectin more annotated data increases the machine-learning model accuracy. On the other hand, the piezoelectric sensor cannot be used for reliable speed measurement and other tasks mentioned in (Section 4). In the following research, we plan to replace the piezoelectric sensor with the FBG sensor to solve more monitoring tasks in intelligent transportation systems.
The installed camera systems showed us as a reliable monitoring system for many tasks mentioned in Section 4. In future work, we plan to buy a high-speed camera system to enable us to collect data to estimate vehicles’ speed. We also plan to try different angles of installation as well as to buy cameras with night vision mode. Cameras with night vision mode can reduce the impact of extreme weather conditions at night to obtain image data usable for CV methods.

8. Discussion

The research community has made significant progress in developing various sensor systems for intelligent transportation. A huge number of mentioned physical sensors directly measure the observed values and classify the observed events through these values. Sensor systems, except the vision-based sensor systems, mostly focus on a small range of transportation tasks.
Camera systems can scale sensitivity via various lenses, sampling rates, and resolutions. These input conditions enable us to create an universal virtual sensor for many tasks with a wide range of data quality input. In last decade, scientists developed many open-source as well as paid frameworks for pre-trained neural network models for image data. We can also use non-machine learning approaches in CV methods to solve tasks requiring motion detection and object tracking.
The semiconductor industry enables us to manufacture strong SBC and MCU edge devices where we can deploy the neural network models as well as other CV approaches. The edge devices are also small in size, with small energy consumption. These edge devices can be connected directly to the same places where the Camera Systems are installed. The connection of edge devices (with deployed CV methods or algorithm software) that are used as a virtual sensor, and a Camera System, which is used as a physical sensor, create a new physical sensor to sense or measure new phenomena such as speed estimation, fire detection, or distance measuring. However, edge devices have limited computational power to compute complex machine-learning models in real-time applications.
Pre-trained neural network models are unreliable for images taken in extreme weather conditions. To improve theses models’ accuracy in extreme weather conditions, the models need to learn from the additional dataset. The dataset can be collected using a vehicle’s cameras. The car’s camera can collect a dataset from traffic situations in extreme weather conditions from various view angles that can be used to improve existing machine learning models.
To obtain this dataset, engineers have to implement communication interfaces and protocols in the infrastructure’s monitoring system. The communication protocols are to exchange valuable data between V2I and V2V that need to be standardized, and they are in development. A reliable, standardized protocol can be used in every vehicle as the sensor data source for monitoring traffic in intelligent transportation monitoring tasks.
Creating a universal intelligent sensor system for monitoring tasks in the intelligent transportation system is one of the research challenges. In the following research, we focus on a camera-based system. To this purpose, it is necessary to find a suitable camera system with parameters such as resolution, sampling rate, and other features that will help to achieve this aim. Then, we should propose, develop and test a communication protocol between SBCs or MCUs. Based on these conditions, the camera system classifies the monitored tasks mentioned in Section 4. If there is a need to use complex machine learning models in real-time applications installed in SBCs or MCUs, then the computational limitations of SBCs or MCUs have to be solved with hardware-based or software-based techniques to simplify the model computation. Hardware-based inference acceleration techniques increase the power consumption and make the system less compact. Software-based techniques focus on the simplicity of machine-learning models or model explainability techniques to extract features programmatically. The appropriate SBC or MCU selection can reduce the machine-learning model or other algorithm optimization via hardware-based or software-based techniques. Improvement of the existing pre-trained machine-learning models for extreme weather conditions requires installing the camera-based monitoring system in many places around the world to obtain data or obtaining data from vehicles’ camera system records. Standardizing the V2I communication protocol is necessary to get the vehicle’s sensor data and camera records regarding the driver’s security and privacy. Integrating all necessary data including the extreme weather conditions to make an efficient machine-learning model is a technically complicated process, as well as the training model process of the suitable machine-learning architecture model. SBC or MCU have to use the optimized variant of the trained model. We propose using the traditional CV methods, not using the machine-learning approach and not requiring a complex dataset to make the CV’s software solution [169,170].

9. Conclusions

Sensors are the essential element of IoT in collecting data from the physical world to the digital form. Implementing IoT concepts into transportation can improve monitoring systems for the maintenance of roads. The quality of the road’s pavement impacts traffic safety and fluency. The improved Lidar and camera manufacturing technology partially solved the research challenges in monitoring roads. There are many real-time algorithms and methods to recognise vehicles and pedestrians from Lidars and camera data. These algorithms and methods are suitable for use in edge devices. The edge device’s optimised firmware based on CV methods can be used as a universal physical sensor based on the fusion software-based virtual sensor and camera or Lidar as a physical sensor. Cameras have the disadvantage of taking photos in bad weather conditions. Lidars could solve this disadvantage and use similar CV algorithms and methods to classify the motion detection of vehicles or pedestrians and track their movement.
In this study, we performed a literature survey of the available sensor systems used for intelligent transportation system purposes. These sensor systems were categorized according to our proposed taxonomy from many angles of view. We also surveyed the monitoring tasks in the intelligent transportation system based on motion detection and object tracking methods. The taxonomy and literature survey of the sensor systems, as well as monitoring tasks, were used for making Table 3, which reviews the sensor systems’ effectiveness for each monitoring task.
The result of this review was the case study where we proposed the architecture of EEISS. The EEISS uses vision-based sensors to create a flexible universal sensor system for monitoring tasks in the intelligent transportation system. The flexibility of the EEISS provides software-based plug-ins (service-oriented) architecture. These software plug-ins (services) use CV methods and algorithms to classify and process image data. These plug-ins (services) are the kind of software-based virtual sensors for monitoring tasks such as ALPR, vehicle counting and classification or speed measurement.
The following research steps focus on optimizing CV methods and algorithms for real-time applications capable of the SBCs’ performance. Our aim was to discover minimal SBC hardware requirements for some selected software-based plug-ins (services). It will be compared the performance of traditional CV and deep neural network approaches in the vehicle’s motion detection and object tracking tasks suitable for SBC from inference time and reliability. To gain reliable image data from the mast-mounted surveillance system, various camera systems will be tested with different frame rates, lenses, angles of the view and resolution matrix. The analytics have used intrusive sensors to count and classify types of vehicles for the last two decades because they are reliable. Furthermore, installing and maintaining intrusive sensors on the road surface is expensive. Therefore, we aim to develop and implement plug-ins (services) based on CV methods in which the classification and data process accuracy will be similar to or better than intrusive sensors. The EEISS based on these plug-ins (services) could replace some intrusive because maintenance and installation costs could be cheaper of intrusive sensors.

Author Contributions

Conceptualization and methodology, K.M. and P.P.; formal analysis and supervision, P.P. and I.Z.; writing—original draft preparation, K.M.; writing—review and editing, K.M., P.P. and I.Z.; funding acquisition and project administration, I.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research work is the result and funded of the APVV grant ENISaC - Edge-eNabled Intelligent Sensing and Computing (APVV-20-0247).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The mast-mounted surveillance system of the Betamont company.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guerrero-Ibáñez, J.; Zeadally, S.; Contreras-Castillo, J. Sensor technologies for intelligent transportation systems. Sensors 2018, 18, 1212. [Google Scholar] [CrossRef] [PubMed]
  2. Shokravi, H.; Shokravi, H.; Bakhary, N.; Heidarrezaei, M.; Rahimian Koloor, S.S.; Petrů, M. A review on vehicle classification and potential use of smart vehicle-assisted techniques. Sensors 2020, 20, 3274. [Google Scholar] [CrossRef] [PubMed]
  3. Vaidya, R.B.; Kulkarni, S.; Didore, V. Intelligent transportation system using IOT: A Review. Int. J. Res. Trends Innov. 2021, 6, 80–87. [Google Scholar]
  4. Chhillar, S.R.S. A Review of Intelligent Transportation Systems in Existing Framework using IoT. Int. J. Eng. Trends Technol. 2022, 70, 137–143. [Google Scholar] [CrossRef]
  5. Robin, A.; Botts, M. OGC® SensorML: Model and XML Encoding Standard; Version 2.0.0 (OGC 12-000); Open Geospatial Consortium: Wayland, MA, USA, 2014; 196p. [Google Scholar] [CrossRef]
  6. Liang, J.; Zulkifli, M.Y.; Choy, S.; Li, Y.; Gao, M.; Kong, B.; Yun, J.; Liang, K. Metal–organic framework–plant nanobiohybrids as living sensors for on-site environmental pollutant detection. Environ. Sci. Technol. 2020, 54, 11356–11364. [Google Scholar] [CrossRef]
  7. Kostal, P.; Roman, R.; Holubek, R. Novel Trends in Manufacturing Systems with View on Implementation Possibilities of Intelligent Automation. Int. J. Ind. Eng. 2013, 7, 858–862. [Google Scholar]
  8. Paepae, T.; Bokoro, P.N.; Kyamakya, K. From fully physical to virtual sensing for water quality assessment: A comprehensive review of the relevant state-of-the-art. Sensors 2021, 21, 6971. [Google Scholar] [CrossRef]
  9. Gupta, A.; Mukherjee, N. Poster: Virtual sensor: The purpose and applications. In Proceedings of the 14th Annual International Conference on Mobile Systems, Applications, and Services Companion, Singapore, 26–30 June 2016; p. 28. [Google Scholar]
  10. Abdelwahab, S.; Hamdaoui, B.; Guizani, M.; Znati, T. Cloud of things for sensing-as-a-service: Architecture, algorithms, and use case. IEEE Internet Things J. 2016, 3, 1099–1112. [Google Scholar] [CrossRef]
  11. Martin, D.; Kühl, N.; Satzger, G. Virtual sensors. Bus. Inf. Syst. Eng. 2021, 63, 315–323. [Google Scholar] [CrossRef]
  12. Armando, N.; Rodrigues, A.; Pereira, V.; Sá Silva, J.; Boavida, F. An outlook on physical and virtual sensors for a socially interactive internet. Sensors 2018, 18, 2578. [Google Scholar] [CrossRef]
  13. Bachiller, R.; Matthys, N.; del Cid, J.; Joosen, W.; Hughes, D.; Van Laerhoven, K. @migo: A comprehensive middleware solution for participatory sensing applications. In Proceedings of the 2015 IEEE 14th International Symposium on Network Computing and Applications, Cambridge, MA, USA, 28–30 September 2015; pp. 1–8. [Google Scholar]
  14. Daniel, M.; Neves, R.F.; Horta, N. Company event popularity for financial markets using Twitter and sentiment analysis. Expert Syst. Appl. 2017, 71, 111–124. [Google Scholar] [CrossRef]
  15. Nakashima, K.; Yokoyama, M.; Taniyama, Y.; Yohshihisa, T.; Hara, T. s3 system: A system for sharing social sensor data and analytical programs. In Proceedings of the Adjunct Proceedings of the 13th International Conference on Mobile and Ubiquitous Systems: Computing Networking and Services, Hiroshima, Japan, 28 November–1 December 2016; pp. 147–152. [Google Scholar]
  16. Inci, E.; Topcu, G.; Guner, T.; Demirkurt, M.; Demir, M.M. Recent developments of colorimetric mechanical sensors based on polymer composites. J. Mater. Chem. C 2020, 8, 12036–12053. [Google Scholar] [CrossRef]
  17. Moldovan, O.; Iñiguez, B.; Deen, M.J.; Marsal, L.F. Graphene electronic sensors–review of recent developments and future challenges. IET Circ. Dev. Syst. 2015, 9, 446–453. [Google Scholar] [CrossRef]
  18. Gerensea, H.; Murugan, R. Is there significant difference between digital and glass mercury thermometer? Adv. Nurs. 2016, 2016, 3474503. [Google Scholar] [CrossRef]
  19. Tewolde, G.S. Sensor and network technology for intelligent transportation systems. In Proceedings of the 2012 IEEE International Conference on Electro/Information Technology, Indianapolis, IN, USA, 6–8 May 2012; pp. 1–7. [Google Scholar]
  20. Chiang, M.; Ha, S.; Risso, F.; Zhang, T.; Chih-Lin, I. Clarifying fog computing and networking: 10 questions and answers. IEEE Commun. Mag. 2017, 55, 18–20. [Google Scholar] [CrossRef]
  21. Pérez, J.; Díaz, J.; Berrocal, J.; López-Viana, R.; González-Prieto, Á. Edge computing. Computing 2022, 104, 2711–2747. [Google Scholar] [CrossRef]
  22. Mohan, N.; Kangasharju, J. Edge-Fog cloud: A distributed cloud for Internet of Things computations. In Proceedings of the 2016 Cloudification of the Internet of Things (CIoT), Paris, France, 23–25 November 2016; pp. 1–6. [Google Scholar]
  23. Shi, W.; Cao, J.; Zhang, Q.; Li, Y.; Xu, L. Edge computing: Vision and challenges. IEEE Internet Things J. 2016, 3, 637–646. [Google Scholar] [CrossRef]
  24. Solutions, C.F.C. Unleash the Power of the Internet of Things; Cisco Systems Inc.: San Jose, CA, USA, 2015. [Google Scholar]
  25. Atlam, H.F.; Walters, R.J.; Wills, G.B. Fog Computing and the Internet of Things: A Review. Big Data Cogn. Comput. 2018, 2, 10. [Google Scholar] [CrossRef]
  26. Dillon, T.; Wu, C.; Chang, E. Cloud computing: Issues and challenges. In Proceedings of the 2010 24th IEEE International Conference on Advanced Information Networking and Applications, Perth, Australia, 20–13 April 2010; pp. 27–33. [Google Scholar]
  27. Lee, H.; Coifman, B. Using LIDAR to validate the performance of vehicle classification stations. J. Intell. Transp. Syst. 2015, 19, 355–369. [Google Scholar] [CrossRef]
  28. Muthuramalingam, S.; Bharathi, A.; Rakesh Kumar, S.; Gayathri, N.; Sathiyaraj, R.; Balamurugan, B. IoT based intelligent transportation system (IoT-ITS) for global perspective: A case study. In Internet of Things and Big Data Analytics for Smart Generation; Springer: Cham, Switzerland, 2019; pp. 279–300. [Google Scholar]
  29. Barbagli, B.; Manes, G.; Facchini, R.; Manes, A. Acoustic sensor network for vehicle traffic monitoring. In Proceedings of the 1st International Conference on Advances in Vehicular Systems, Technologies and Applications, Venice, Italy, 24–29 June 2012; pp. 24–29. [Google Scholar]
  30. Hashemi Vaziri, S. Investigation of environmental impacts on piezoelectric weigh-in-motion sensing system. In UWSpace; University of Waterloo: Waterloo, ON, Canada, 2011; Available online: http://hdl.handle.net/10012/6344 (accessed on 1 May 2023).
  31. Zhang, L.; Haas, C.; Tighe, S.L. Evaluating weigh-in-motion sensing technology for traffic data collection. In Proceedings of the Annual Conference of the Transportation Association of Canada, Saskatoon, SK, Canada, 14–17 October 2007; pp. 1–17. [Google Scholar]
  32. Al-Tarawneh, M.; Huang, Y.; Lu, P.; Bridgelall, R. Weigh-In-Motion System in Flexible Pavements Using Fiber Bragg Grating Sensors Part A: Concept. IEEE Trans. Intell. Transp. Syst. 2020, 21, 5136–5147. [Google Scholar] [CrossRef]
  33. Trafford, R.; Linden, R.; Donovan, J.; Neville, E.; Marroquin, W.; Tan, S. Retrofitting Rural Infrastructure for Smart Parking and Traffic Monitoring. In Proceedings of the IEEE Sensors Applications Symposium 2017, Glassboro, NJ, USA, 13–15 March 2017. [Google Scholar]
  34. Bajwa, R.; Rajagopal, R.; Coleri, E.; Varaiya, P.; Flores, C. In-pavement wireless weigh-in-motion. In Proceedings of the 12th International Conference on Information Processing in Sensor Networks, Philadelphia, PA, USA, 8–11 April 2013; pp. 103–114. [Google Scholar]
  35. Meta, S.; Cinsdikici, M.G. Vehicle-classification algorithm based on component analysis for single-loop inductive detector. IEEE Trans. Veh. Technol. 2010, 59, 2795–2805. [Google Scholar] [CrossRef]
  36. Martin, P.T.; Feng, Y.; Wang, X. Detector Technology Evaluation; Technical Report; Mountain-Plains Consortium: Fargo, ND, USA, 2003. [Google Scholar]
  37. Cheung, S.Y.; Coleri, S.; Dundar, B.; Ganesh, S.; Tan, C.W.; Varaiya, P. Traffic measurement and vehicle classification with single magnetic sensor. Transp. Res. Rec. 2005, 1917, 173–181. [Google Scholar] [CrossRef]
  38. Khan, M.A.; Sharma, M.; Prabhu, B.R. A survey of RFID tags. Int. J. Recent Trends Eng. 2009, 1, 68. [Google Scholar]
  39. Wang, J.; Dhanapal, R.K.; Ramakrishnan, P.; Balasingam, B.; Souza, T.; Maev, R. Active RFID based indoor localization. In Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019; pp. 1–7. [Google Scholar]
  40. Jiang, D.; Huang, D.; Song, Y.; Wu, K.; Lu, H.; Liu, Q.; Zhou, T. An audio data representation for traffic acoustic scene recognition. IEEE Access 2020, 8, 177863–177873. [Google Scholar] [CrossRef]
  41. Won, M. Intelligent traffic monitoring systems for vehicle classification: A survey. IEEE Access 2020, 8, 73340–73358. [Google Scholar] [CrossRef]
  42. Verykokou, S.; Ioannidis, C. An Overview on Image-Based and Scanner-Based 3D Modeling Technologies. Sensors 2023, 23, 596. [Google Scholar] [CrossRef]
  43. Chidlovskii, B.; Csurka, G.; Rodriguez-Serrano, J. Vehicle type classification from laser scans with global alignment kernels. In Proceedings of the 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), Qingdao, China, 8–11 October 2014; pp. 2840–2845. [Google Scholar]
  44. Raja Abdullah, R.S.A.; Abdul Aziz, N.H.; Abdul Rashid, N.E.; Ahmad Salah, A.; Hashim, F. Analysis on target detection and classification in LTE based passive forward scattering radar. Sensors 2016, 16, 1607. [Google Scholar] [CrossRef]
  45. Chang, N.; Rashidzadeh, R.; Ahmadi, M. Robust indoor positioning using differential Wi-Fi access points. IEEE Trans. Consum. Electron. 2010, 56, 1860–1867. [Google Scholar] [CrossRef]
  46. Alsharif, M.H.; Kim, S.; Kuruoğlu, N. Energy harvesting techniques for wireless sensor networks/radio-frequency identification: A review. Symmetry 2019, 11, 865. [Google Scholar] [CrossRef]
  47. Kanschat, R.; Gupta, S.; Degbelo, A. Wireless-Signal-Based Vehicle Counting and Classification in Different Road Environments. IEEE Open J. Intell. Transp. Syst. 2022, 3, 236–250. [Google Scholar] [CrossRef]
  48. Sakib, S.M.N. LiDAR Technology—An Overview. IUP J. Electr. Electron. Eng. 2022, 15, 36–57. [Google Scholar]
  49. Mukhopadhyay, B.; Srirangarajan, S.; Kar, S. Modeling the analog response of passive infrared sensor. Sens. Actuators A Phys. 2018, 279, 65–74. [Google Scholar] [CrossRef]
  50. Sitanayah, L.; Angdresey, A.; Utama, J.W. Poster: An Internet of Things-based Vehicle Counting System. In Proceedings of the EWSN, Delft, The Netherlands, 17–19 February 2021; pp. 175–176. [Google Scholar]
  51. Kianpisheh, A.; Mustaffa, N.; Limtrairut, P.; Keikhosrokiani, P. Smart parking system (SPS) architecture using ultrasonic detector. Int. J. Softw. Eng. Appl. 2012, 6, 55–58. [Google Scholar]
  52. Hodges, L. Ultrasonic and Passive Infrared Sensor Integration for Dual Technology User Detection Sensors; Application Note; Michigan State University: East Lansing, MI, USA, 2009; pp. 5–6. [Google Scholar]
  53. Pandey, M.; Mishra, G. Types of sensor and their applications, advantages, and disadvantages. In Proceedings of the Emerging Technologies in Data Mining and Information Security: Proceedings of IEMIS 2018; Springer: New York, NY, USA, 2019; Volume 3, pp. 791–804. [Google Scholar]
  54. RadhaKrishna, M.; Govindh, M.V.; Veni, P.K. A review on image processing sensor. J. Phys. Conf. Ser. 2021, 1714, 012055. [Google Scholar] [CrossRef]
  55. Deformable Mirrors and “Fake” Stars in Space Photography. Available online: https://www.findlight.net/blog/deformable-mirrors-in-astronomy/ (accessed on 18 April 2023).
  56. Theuwissen, A.J. CCD or CMOS image sensors for consumer digital still photography? In Proceedings of the 2001 International Symposium on VLSI Technology, Systems, and Applications, Proceedings of Technical Papers (Cat. No. 01TH8517), Hsinchu, Taiwan, 18–20 April 2001; pp. 168–171. [Google Scholar]
  57. Choi, B.S.; Kim, S.H.; Lee, J.; Chang, S.; Park, J.; Lee, S.J.; Shin, J.K. CMOS image sensor using pixel aperture technique for single-chip 2D and 3D imaging. In Proceedings of the 2017 IEEE SENSORS, Glasgow, UK, 29 October–1 November 2017; pp. 1–3. [Google Scholar]
  58. Kilicarslan, M.; Zheng, J.Y. Predict vehicle collision by TTC from motion using a single video camera. IEEE Trans. Intell. Transp. Syst. 2018, 20, 522–533. [Google Scholar] [CrossRef]
  59. Sagar, V.D.; Nanjundeswaraswamy, T. Artificial intelligence in autonomous vehicles-a literature review. i-Manag. J. Future Eng. Technol. 2019, 14, 56. [Google Scholar]
  60. Boysen, N.; Briskorn, D.; Schwerdfeger, S.; Stephan, K. Optimizing carpool formation along high-occupancy vehicle lanes. Eur. J. Oper. Res. 2021, 293, 1097–1112. [Google Scholar] [CrossRef]
  61. Lee, J.; Byun, J.; Lim, J.; Lee, J. A framework for detecting vehicle occupancy based on the occupant labeling method. J. Adv. Transp. 2020, 2020, 8870211. [Google Scholar] [CrossRef]
  62. Yashwanth, S.D.; Nimish, S.K.; Likhitha, S. Design of Piezoelectric Transducer for Automobile Seat Belt Locking System; Research Square: Durham, NC, USA, 2023. [Google Scholar] [CrossRef]
  63. Aralt, T.; Nilsen, A. Automatic fire detection in road traffic tunnels. Tunn. Undergr. Space Technol. 2009, 24, 75–83. [Google Scholar] [CrossRef]
  64. White, J.; Thompson, C.; Turner, H.; Dougherty, B.; Schmidt, D.C. Wreckwatch: Automatic traffic accident detection and notification with smartphones. Mob. Netw. Appl. 2011, 16, 285–303. [Google Scholar] [CrossRef]
  65. Brooks, R.M.; Cetin, M. Pavement Friction V. Initial Velocity of Vehicle in Chain Accidents on Highways. Int. J. Eng. Invent. 2013, 3, 27–31. [Google Scholar]
  66. Maha Vishnu, V.; Rajalakshmi, M.; Nedunchezhian, R. Intelligent traffic video surveillance and accident detection system with dynamic traffic signal control. Clust. Comput. 2018, 21, 135–147. [Google Scholar] [CrossRef]
  67. Popescu, O.; Sha-Mohammad, S.; Abdel-Wahab, H.; Popescu, D.C.; El-Tawab, S. Automatic incident detection in intelligent transportation systems using aggregation of traffic parameters collected through V2I communications. IEEE Intell. Transp. Syst. Mag. 2017, 9, 64–75. [Google Scholar] [CrossRef]
  68. Chattopadhyay, D.; Rasheed, S.; Yan, L.; Lopez, A.A.; Farmer, J.; Brown, D.E. Machine learning for real-time vehicle detection in all-electronic tolling system. In Proceedings of the 2020 Systems and Information Engineering Design Symposium (SIEDS), Charlottesville, VA, USA, 24 April 2020; pp. 1–6. [Google Scholar]
  69. Karami, Z.; Kashef, R. Smart transportation planning: Data, models, and algorithms. Transp. Eng. 2020, 2, 100013. [Google Scholar] [CrossRef]
  70. Litman, T. Introduction to Multi-Modal Transportation Planning; Victoria Transport Policy Institute Canada: Victoria, BC, Canada, 2017. [Google Scholar]
  71. Tang, J.; Wan, L.; Schooling, J.; Zhao, P.; Chen, J.; Wei, S. Automatic number plate recognition (ANPR) in smart cities: A systematic review on technological advancements and application cases. Cities 2022, 129, 103833. [Google Scholar] [CrossRef]
  72. Rashid, M.M.; Musa, A.; Rahman, M.A.; Farahana, N.; Farhana, A. Automatic parking management system and parking fee collection based on number plate recognition. Int. J. Mach. Learn. Comput. 2012, 2, 94. [Google Scholar] [CrossRef]
  73. Hwang, C.M.; Shu, S.Y.; Chen, W.Y.; Chen, Y.W.; Wen, K.P. PC-based car license plate reader. In Proceedings of the Machine Vision Applications, Architectures, and Systems Integration, Boston, MA, USA, 1 November 1992; Volume 1823, pp. 272–283. [Google Scholar]
  74. Wanshui, H.; Xiaodong, L.; Guangzhong, G.; Qing, X.; Yangguang, Y. Site-Specific Extra-Heavy Truck Load Characteristics and Bridge Safety Assessment; American Society of Civil Engineers: Reston, VA, USA, 2018. [Google Scholar]
  75. Schindler, N. Making the Slovak Tolling Network the Largest in the EU with the Use of GNSS Technology. In Proceedings of the 10th ITS Europeen Congress, Helsinki, Finland, 16–19 June 2014. [Google Scholar]
  76. Cao, R.; Agrawal, A.K.; El-Tawil, S. Overheight impact on bridges: A computational case study of the Skagit River bridge collapse. Eng. Struct. 2021, 237, 112215. [Google Scholar] [CrossRef]
  77. Moridpour, S.; Mazloumi, E.; Mesbah, M. Impact of heavy vehicles on surrounding traffic characteristics. J. Adv. Transp. 2015, 49, 535–552. [Google Scholar] [CrossRef]
  78. Aarts, L.; Van Schagen, I. Driving speed and the risk of road crashes: A review. Accid. Anal. Prev. 2006, 38, 215–224. [Google Scholar] [CrossRef]
  79. Kopper, J.M. The Scientific Reliability of Radar Speedmeters. NCL Rev. 1954, 33, 343. [Google Scholar]
  80. Micko, K.; Babic, F.; Papcun, P.; Zolotova, I. Temporary Parking via Computer Vision and Deep Learning. In Proceedings of the 2022 IEEE 20th Jubilee World Symposium on Applied Machine Intelligence and Informatics (SAMI), Poprad, Slovakia, 19–22 January 2022; pp. 455–460. [Google Scholar]
  81. Richardson, J.; Jones, S.; Brown, A.; O’Brien, E.; Hajializadeh, D. On the use of bridge weigh-in-motion for overweight truck enforcement. Int. J. Heavy Veh. Syst. 2014, 21, 83–104. [Google Scholar] [CrossRef]
  82. Yuan, F.; Cheu, R.L. Incident detection using support vector machines. Transp. Res. Part C: Emerg. Technol. 2003, 11, 309–328. [Google Scholar] [CrossRef]
  83. Rizvi, S.M.A.; Ahmed, A.; Shen, Y. Real-Time Incident Detection and Capacity Estimation Using Loop Detector Data. J. Adv. Transp. 2020, 2020, 8857502. [Google Scholar] [CrossRef]
  84. Amin, M.S.; Nasir, S.S.; Reaz, M.B.I.; Ali, M.A.M.; Chang, T.G. Preference and placement of vehicle crash sensors. Tech. Gaz. 2014, 21, 889–896. [Google Scholar]
  85. Rajab, S.A.; Mayeli, A.; Refai, H.H. Vehicle classification and accurate speed calculation using multi-element piezoelectric sensor. In Proceedings of the 2014 IEEE Intelligent Vehicles Symposium Proceedings, Dearborn, MI, USA, 8–11 June 2014; pp. 894–899. [Google Scholar]
  86. Sochor, J.; Juránek, R.; Španhel, J.; Maršık, L.; Širokỳ, A.; Herout, A.; Zemcık, P. BrnoCompSpeed: Review of traffic camera calibration and comprehensive dataset for monocular speed measurement. arXiv 2017, arXiv:1702.06441. [Google Scholar]
  87. Jain, A.; Shambavi, K. Energy Efficient Smart Parking System Using Peizoelectric Material And Sensors. Int. Res. J. Mod. Eng. Technol. Sci. 2021, 3, 3148–3150. [Google Scholar]
  88. Gajda, J.; Sroka, R.; Zeglen, T.; Burnos, P. The influence of temperature on errors of WIM systems employing piezoelectric sensor. Metrol. Meas. Syst. 2013, 20, 171–182. [Google Scholar] [CrossRef]
  89. Krehel, M.; Rossi, R.M.; Bona, G.L.; Scherer, L.J. Characterization of flexible copolymer optical fibers for force sensing applications. Sensors 2013, 13, 11956–11968. [Google Scholar] [CrossRef]
  90. Tavares, C.; Silva, J.O.E.; Mendes, A.; Rebolo, L.; Domingues, M.D.F.; Alberto, N.; Lima, M.; Silva, H.P.; Antunes, P.F.D.C. Instrumented Office Chair with Low-Cost Plastic Optical Fiber Sensors for Posture Control and Work Conditions Optimization. IEEE Access 2022, 10, 69063–69071. [Google Scholar] [CrossRef]
  91. Haroglu, D.; Powell, N.; Seyam, A.F.M. The response of polymer optical fiber (POF) to bending and axial tension for the application of a POF sensor for automotive seat occupancy sensing. J. Text. Inst. 2017, 108, 132–139. [Google Scholar] [CrossRef]
  92. Hu, Z.; Hu, Y. Fundamental Study on Fiber Bragg Grating Network for High Speed Collisions Detection. In Proceedings of the 2011 Second International Conference on Digital Manufacturing & Automation, Zhangjiajie, China, 5–7 August 2011; pp. 556–559. [Google Scholar]
  93. Yang, X.; Ahmad, S.; Huang, Y.; Lu, P. Automatic Vehicle Counting by Using In-Pavement Fiber Bragg Grating Sensor. In Proceedings of the International Conference on Transportation and Development 2022, Seattle, WA, USA, 31 May–3 June 2022; pp. 225–234. [Google Scholar]
  94. Huang, Y.; Lu, P.; Tolliver, D.; Al-Tarawneh, M. Vehicle classification system using in-pavement fiber Bragg grating sensors. IEEE Sens. J. 2018, 18, 2807–2815. [Google Scholar]
  95. Yuksel, K.; Kinet, D.; Chah, K.; Caucheteur, C. Implementation of a mobile platform based on fiber bragg grating sensors for automotive traffic monitoring. Sensors 2020, 20, 1567. [Google Scholar] [CrossRef] [PubMed]
  96. Prasad, A.G.; Sharath, U.; Amith, B.; Supritha, B.; Asokan, S.; Hegde, G. Fiber Bragg Grating sensor instrumentation for parking space occupancy management. In Proceedings of the 2012 International Conference on Optical Engineering (ICOE), Belgaum, India, 26–28 July 2012; pp. 1–4. [Google Scholar]
  97. Al-Tarawneh, M.A.; Huang, Y. Glass fiber-reinforced polymer packaged fiber Bragg grating sensors for low-speed weigh-in-motion measurements. Opt. Eng. 2016, 55, 086107. [Google Scholar] [CrossRef]
  98. Puan, O.C.; Nor, N.S.M.; Mashros, N.; Hainin, M.R. Applicability of an automatic pneumatic-tube-based traffic counting device for collecting data under mixed traffic. IOP Conf. Ser. Earth Environ. Sci. 2019, 365, 012032. [Google Scholar] [CrossRef]
  99. Sparks, C. The influence of tension, pressure and weight on pipe and riser deformations and stresses. J. Energy Resour. Technol. Trans. ASME 1984, 106, 46–54. [Google Scholar] [CrossRef]
  100. Orlewski, P. Occupant Detection System for an Automotive Vehicle. U.S. Patent 12/669,550, 7 July 2008. [Google Scholar]
  101. Chaudhari, A.; Agrawal, H.; Poddar, S.; Talele, K.; Bansode, M. Smart Accident Detection And Alert System. In Proceedings of the 2021 IEEE India Council International Subsections Conference (INDISCON), Nagpur, India, 27–29 August 2021; pp. 1–4. [Google Scholar] [CrossRef]
  102. Yoshida, M.; Akiyama, S.; Moriyama, Y.; Takeshima, Y.; Kondo, Y.; Suwa, H.; Yasumoto, K. Traffic Census Sensor Using Vibration Caused by Passing Vehicles. Sens. Mater. 2021, 33, 1–16. [Google Scholar] [CrossRef]
  103. Smith, A.; Goley, S.; Vongsy, K.; Shaw, A.; Dierking, M. Effects of fundamental frequency normalization on vibration-based vehicle classification. In Proceedings of the Conference on Signal Processing, Sensor/Information Fusion, and Target Recognition XXIV; Kadar, I., Ed.; SPIE: Baltimore, MD, USA, 2015; Volume 9474, pp. 20–22. [Google Scholar] [CrossRef]
  104. Wang, X.; Shi, X.; Wang, J.; Yu, X.; Han, B. Train Speed Estimation from Track Structure Vibration Measurements. Appl. Sci. 2020, 10, 4742. [Google Scholar] [CrossRef]
  105. Wang, H.; He, W. A Reservation-based Smart Parking System. In Proceedings of the 2011 IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Shanghai, China, 10–15 April 2011; pp. 690–695. [Google Scholar] [CrossRef]
  106. Li, B.; Zhan, S.; Hu, J.; Sun, W.; Zhang, G. New strategy for weigh-in-motion measurement based on vibration sensors. In Proceedings of the 2018 IEEE 8th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER), Tianjin, China, 19–23 July 2018; pp. 1595–1599. [Google Scholar]
  107. Gajda, J.; Sroka, R.; Stencel, M.; Wajda, A.; Zeglen, T. A vehicle classification based on inductive loop detectors. In Proceedings of the IMTC 2001, 18th IEEE Instrumentation and Measurement Technology Conference, Rediscovering Measurement in the Age of Informatics (Cat. No.01CH 37188), Budapest, Hungary, 21–23 May 2001; Volume 1, pp. 460–464. [Google Scholar] [CrossRef]
  108. Oluwatobi, A.N.; Tayo, A.O.; Oladele, A.T.; Adesina, G.R. The design of a vehicle detector and counter system using inductive loop technology. Procedia Comput. Sci. 2021, 183, 493–503. [Google Scholar] [CrossRef]
  109. Sheik Mohammed, A.S.; Joshi, N.; George, B.; Vanajakshi, L. Application of random forest algorithm to classify vehicles detected by a multiple inductive loop system. In Proceedings of the 2012 15th International IEEE Conference on Intelligent Transportation Systems, Anchorage, AK, USA, 16–19 September 2012; pp. 491–495. [Google Scholar]
  110. He, Y.; Du, Y.; Sun, L. Vehicle Classification Method Based on Single-Point Magnetic Sensor. Procedia—Soc. Behav. Sci. 2012, 43, 618–627. [Google Scholar] [CrossRef]
  111. Bao, X.; Li, H.; Xu, D.; Jia, L.; Ran, B.; Rong, J. Traffic Vehicle Counting in Jam Flow Conditions Using Low-Cost and Energy-Efficient Wireless Magnetic Sensors. Sensors 2016, 16, 1868. [Google Scholar] [CrossRef]
  112. Trigona, C.; Andò, B.; Sinatra, V.; Vacirca, C.; Rossino, E.; Palermo, L.; Kurukunda, S.; Baglio, S. Implementation and characterization of a smart parking system based on 3-axis magnetic sensors. In Proceedings of the 2016 IEEE International Instrumentation and Measurement Technology Conference Proceedings, Taipei, Taiwan, 23–26 May 2016; pp. 1–6. [Google Scholar]
  113. Oberli, C.; Torres-Torriti, M.; Landau, D. Performance evaluation of UHF RFID technologies for real-time passenger recognition in intelligent public transportation systems. IEEE Trans. Intell. Transp. Syst. 2010, 11, 748–753. [Google Scholar] [CrossRef]
  114. Diyasa, I.G.S.M.; Purbasari, I.Y.; Setiawan, A.; Winardi, S. Smart Passenger Information System Based On IoT. In Proceedings of the 2019 TRON Symposium (TRONSHOW), Tokyo, Japan, 11–13 December 2019; pp. 1–5. [Google Scholar]
  115. Krausz, N.; Lovas, T.; Barsi, Á. Radio frequency identification in supporting traffic safety. Period. Polytech. Civ. Eng. 2017, 61, 727–731. [Google Scholar] [CrossRef]
  116. Li, C. Automatic vehicle identification (AVI) system based on RFID. In Proceedings of the 2010 International Conference on Anti-Counterfeiting, Security and Identification, Chengdu, China, 18–20 July 2010; pp. 281–284. [Google Scholar] [CrossRef]
  117. Mainetti, L.; Palano, L.; Patrono, L.; Stefanizzi, M.L.; Vergallo, R. Integration of RFID and WSN technologies in a Smart Parking System. In Proceedings of the 2014 22nd International Conference on Software, Telecommunications and Computer Networks (SoftCOM), Split, Croatia, 17–19 September 2014; pp. 104–110. [Google Scholar]
  118. Chiang, C.Y.; Jaber, M.; Hayward, P. A Distributed Acoustic Sensor System for Intelligent Transportation using Deep Learning. arXiv 2022, arXiv:2209.05978. [Google Scholar]
  119. Sim, K.Y.H.; Chen, Y.; Wan, Y.; Chau, K. Advanced automobile crash detection by acoustic methods. In Proceedings of the Meetings on Acoustics 178ASA; Acoustical Society of America: New York, NY, USA, 2019; Volume 39, p. 055001. [Google Scholar]
  120. Ishida, S.; Liu, S.; Mimura, K.; Tagashira, S.; Fukuda, A. Design of acoustic vehicle count system using DTW. In Proceedings of the ITS World Congress, Melbourne, Australia, 10–14 October 2016; pp. 1–10. [Google Scholar]
  121. Kandpal, M.; Kakar, V.K.; Verma, G. Classification of ground vehicles using acoustic signal processing and neural network classifier. In Proceedings of the 2013 International Conference on Signal Processing and Communication (ICSC), Noida, India, 12–14 December 2013; pp. 512–518. [Google Scholar]
  122. Göksu, H. Vehicle speed measurement by on-board acoustic signal processing. Meas. Control 2018, 51, 138–149. [Google Scholar] [CrossRef]
  123. Na, K.; Kim, Y.; Cha, H. Acoustic sensor network-based parking lot surveillance system. In Proceedings of the European Conference on Wireless Sensor Networks, Cork, Ireland, 11–13 February 2009; pp. 247–262. [Google Scholar]
  124. Dieng, L.; Girardeau, C.; Gaillet, L.; Falaise, Y.; Žnidarič, A.; Ralbovsky, M. Bridge assessment using Weigh-In-Motion and acoustic emission methods. In Dynamics of Civil Structures; Springer: New York, NY, USA, 2016; Volume 2, pp. 205–215. [Google Scholar]
  125. Kuchár, P.; Pirník, R.; Tichỳ, T.; Rástočnỳ, K.; Skuba, M.; Tettamanti, T. Noninvasive passenger detection comparison using thermal imager and IP cameras. Sustainability 2021, 13, 12928. [Google Scholar] [CrossRef]
  126. Jouper, J.A.; Peabody, M. Apparatus and Method to Monitor the Occupancy of Seating. U.S. Patent 9,302,781, 25 February 2014. [Google Scholar]
  127. Imaki, M.; Kameyama, S.; Ishimura, E.; Nakaji, M.; Yoshinaga, H.; Hirano, Y. Line scanning time-of-flight laser sensor for intelligent transport systems, combining wide field-of-view optics of 30 deg, high scanning speed of 0.9 ms/line, and simple sensor configuration. Opt. Eng. 2017, 56, 031205. [Google Scholar] [CrossRef]
  128. Sandhawalia, H.; Rodriguez-Serrano, J.A.; Poirier, H.; Csurka, G. Vehicle type classification from laser scanner profiles: A benchmark of feature descriptors. In Proceedings of the 16th International IEEE Conference on Intelligent Transportation Systems (ITSC 2013), The Hague, The Netherlands, 6–9 October 2013; pp. 517–522. [Google Scholar]
  129. Shinohara, S.; Suzuki, T.; Yoshida, H.; Ikeda, H.; Sumi, M. Acquisition of 3-D image of still or moving objects utilizing laser diode range-finding speedometer. In Proceedings of the IECON’93-19th Annual Conference of IEEE Industrial Electronics, Maui, HI, USA, 15–19 November 1993; pp. 1730–1735. [Google Scholar]
  130. Groh, B.H.; Friedl, M.; Linarth, A.G.; Angelopoulou, E. Advanced real-time indoor parking localization based on semi-static objects. In Proceedings of the 17th International Conference on Information Fusion (FUSION), Salamanca, Spain, 7–10 July 2014; pp. 1–7. [Google Scholar]
  131. Munte, N.; Lazaro, A.; Villarino, R.; Girbau, D. Vehicle Occupancy Detector Based on FMCW mm-Wave Radar at 77 GHz. IEEE Sens. J. 2022, 22, 24504–24515. [Google Scholar] [CrossRef]
  132. Yang, Z.; Shi, F.; Liang, H. A Portable Traffic Counting, Speed Estimation, and Classification Terminal Using IR-UWB Radar. IEEE Sens. J. 2022, 22, 13365–13374. [Google Scholar] [CrossRef]
  133. Ibrahim, N.; Abdullah, R.; Saripan, M. Artificial neural network approach in radar target classification. J. Comput. Sci. 2009, 5, 23. [Google Scholar] [CrossRef]
  134. Cai, Z.; Zhou, Y.; Qi, Y.; Zhuang, W.; Deng, L. A millimeter wave dual-lens antenna for IoT-based smart parking radar system. IEEE Internet Things J. 2020, 8, 418–427. [Google Scholar] [CrossRef]
  135. Jiang, H.; Chen, S.; Xiao, Z.; Hu, J.; Liu, J.; Dustdar, S. Pa-Count: Passenger Counting in Vehicles Using Wi-Fi Signals. IEEE Trans. Mob. Comput. 2023, 1–14, early access. [Google Scholar] [CrossRef]
  136. Yuan, C.; Fei, L.; Jianxin, C.; Wei, J. A smart parking system using WiFi and wireless sensor network. In Proceedings of the 2016 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), Nantou County, Taiwan, 27–29 May 2016; pp. 1–2. [Google Scholar]
  137. Liu, J.; Mu, H.; Vakil, A.; Ewing, R.; Shen, X.; Blasch, E.; Li, J. Human occupancy detection via passive cognitive radio. Sensors 2020, 20, 4248. [Google Scholar] [CrossRef] [PubMed]
  138. Wang, Z.; Yu, B.; Chen, J.; Liu, C.; Zhan, K.; Sui, X.; Xue, Y.; Li, J. Research on Lidar Point Cloud Segmentation and Collision Detection Algorithm. In Proceedings of the 2019 6th International Conference on Information Science and Control Engineering (ICISCE), Shanghai, China, 20–22 December 2019. [Google Scholar] [CrossRef]
  139. Zhang, J.; Xiao, W.; Coifman, B.; Mills, J.P. Vehicle tracking and speed estimation from roadside lidar. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5597–5608. [Google Scholar] [CrossRef]
  140. Toth, C.K.; Barsi, A.; Lovas, T. Vehicle recognition from LiDAR data. Int. Arch. Photogramm. Remote Sens. 2003, 34, W13. [Google Scholar]
  141. Vasile, I.; Tudor, E.; Sburlan, I.C.; Gheți, M.A.; Popa, G. Experimental Validation of LiDAR Sensors Used in Vehicular Applications by Using a Mobile Platform for Distance and Speed Measurements. Sensors 2021, 21, 8147. [Google Scholar] [CrossRef]
  142. Jiménez, F.; Clavijo, M.; Cerrato, A. Perception, positioning and decision-making algorithms adaptation for an autonomous valet parking system based on infrastructure reference points using one single LiDAR. Sensors 2022, 22, 979. [Google Scholar] [CrossRef]
  143. Gerland, H.E.; Sutter, K. Automatic passenger counting (apc): Infra-red motion analyzer for accurate counts in stations and rail, light-rail and bus operations. In Proceedings of the 1999 APTA Bus Conference, Proceedings American Public Transportation Association, Cleveland, OH, USA, 2–6 May 1999. [Google Scholar]
  144. Xavier, K.L.B.L.; Nanayakkara, V.K. Development of an Early Fire Detection Technique Using a Passive Infrared Sensor and Deep Neural Networks. Fire Technol. 2022, 58, 3529–3552. [Google Scholar] [CrossRef]
  145. Janahan, S.K.; Veeramanickam, M.; Arun, S.; Narayanan, K.; Anandan, R.; Parvez, S.J. IoT based smart traffic signal monitoring system using vehicles counts. Int. J. Eng. Technol. 2018, 7, 309–312. [Google Scholar] [CrossRef]
  146. Khatri, H.S.; Somani, S.B. Infrared-based system for vehicle counting and classification. In Proceedings of the 2015 International Conference on Pervasive Computing (ICPC), Pune, India, 8–10 January 2015. [Google Scholar]
  147. Bachhav, J.D.; Mechkul, M. Smart car parking system. Int. Res. J. Eng. Technol. (IRJET) 2017, 4, 3036–3038. [Google Scholar]
  148. Chimba, D.; Camp, J. High Occupancy Vehicle (HOV) Detection System Testing; Technical Report; Bureau of Transportation Statistics: Washington, DC, USA, 2018.
  149. Kaluthantri, S.T.; Shanmugavel, M.; Menon, R. Low-Cost, Real Time Vehicle Classifcation and Speed Estimation using Raspberry pi 3 Vision Based and Ultra-Sonic Sensor Based Systems. In Proceedings of the 4th IEEE International Symposium in Robotics and Manufacturing Automation (ROMA), Perambalur, India, 10–12 December 2018. [Google Scholar]
  150. Chakraborty, P.; Sharma, A.; Hegde, C. Freeway Traffic Incident Detection from Cameras: A Semi-Supervised Learning Approach. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 1840–1845. [Google Scholar] [CrossRef]
  151. Li, P.; Zhao, W. Image fire detection algorithms based on convolutional neural networks. Case Stud. Therm. Eng. 2020, 19, 100625. [Google Scholar] [CrossRef]
  152. Espinoza, F.T.; Gabriel, B.G.; Barros, M.J. Computer vision classifier and platform for automatic counting: More than cars. In Proceedings of the 2017 IEEE Second Ecuador Technical Chapters Meeting (ETCM), Salinas, Ecuador, 16–20 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
  153. Laroca, R.; Severo, E.; Zanlorensi, L.A.; Oliveira, L.S.; Gonçalves, G.R.; Schwartz, W.R.; Menotti, D. A robust real-time automatic license plate recognition based on the YOLO detector. In Proceedings of the 2018 International Joint Conference on Neural Networks (Ijcnn), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–10. [Google Scholar]
  154. Derrouz, H.; Elbouziady, A.; Ait Abdelali, H.; Haj Thami, R.O.; El Fkihi, S.; Bourzeix, F. Moroccan Video Intelligent Transport System: Vehicle Type Classification Based on Three-Dimensional and Two-Dimensional Features. IEEE Access 2019, 7, 72528–72537. [Google Scholar] [CrossRef]
  155. Wu, J.D.; Chen, B.Y.; Shyr, W.J.; Shih, F.Y. Vehicle Classification and Counting System Using YOLO Object Detection Technology. Trait. Signal 2021, 38, 1087–1093. [Google Scholar] [CrossRef]
  156. Afifah, F.; Nasrin, S.; Mukit, A. Vehicle speed estimation using image processing. J. Adv. Res. Appl. Mech 2019, 48, 9–16. [Google Scholar]
  157. Valenzuela, W.; Saavedra, A.; Zarkesh-Ha, P.; Figueroa, M. Motion-Based Object Location on a Smart Image Sensor Using On-Pixel Memory. Sensors 2022, 22, 6538. [Google Scholar] [CrossRef]
  158. Chou, J.Y.; Chang, C.M. Image Motion Extraction of Structures Using Computer Vision Techniques: A Comparative Study. Sensors 2021, 21, 6248. [Google Scholar] [CrossRef]
  159. Campanella, C.E.; Cuccovillo, A.; Campanella, C.; Yurt, A.; Passaro, V.M. Fibre Bragg grating based strain sensors: Review of technology and applications. Sensors 2018, 18, 3115. [Google Scholar] [CrossRef]
  160. Hong, S.G.; Kim, N.S.; Kim, W.W. Reduction of false alarm signals for PIR sensor in realistic outdoor surveillance. ETRI J. 2013, 35, 80–88. [Google Scholar] [CrossRef]
  161. Savelonas, M.A.; Veinidis, C.N.; Bartsokas, T.K. Computer Vision and Pattern Recognition for the Analysis of 2D/3D Remote Sensing Data in Geoscience: A Survey. Remote Sens. 2022, 14, 6017. [Google Scholar] [CrossRef]
  162. Wei, Q.; Feng, D.; Jia, W. UDR: An Approximate Unbiased Difference-Ratio Edge Detector for SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6688–6705. [Google Scholar] [CrossRef]
  163. Heško, F.; Fiľko, M.; Novotňák, J.; Kašper, P. Perimeter protection of the areas of interest. Acta Avion. 2021, 23, 32–44. [Google Scholar] [CrossRef]
  164. Abdelhamid, S.; Hassanein, H.S.; Takahara, G. Vehicle as a mobile sensor. Procedia Comput. Sci. 2014, 34, 286–295. [Google Scholar] [CrossRef]
  165. Qin, T.; Chen, T.; Chen, Y.; Su, Q. AVP-SLAM: Semantic Visual Mapping and Localization for Autonomous Vehicles in the Parking Lot. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2020; pp. 5939–5945. [Google Scholar] [CrossRef]
  166. Oladimeji, D.; Gupta, K.; Kose, N.A.; Gundogan, K.; Ge, L.; Liang, F. Smart Transportation: An Overview of Technologies and Applications. Sensors 2023, 23, 3880. [Google Scholar] [CrossRef] [PubMed]
  167. Ravi, S.; Mamdikar, M.R. A Review on ITS (Intelligent Transportation Systems) Technology. In Proceedings of the 2022 International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 9–11 May 2022; pp. 155–159. [Google Scholar] [CrossRef]
  168. Ouaknine, A. Review of Deep Learning Algorithms for Object Detection. Available online: https://medium.com/zylapp/review-of-deep-learning-algorithms-for-object-detection-c1f3d437b852 (accessed on 5 February 2023).
  169. HCL Technologies. Decoding the dichotomy: Traditional Image Processing vs. Deep Learning. Available online: https://www.imveurope.com/sites/default/files/content/white-paper/pdfs/HCL_IMVE_WP-ImageProcessing_vs_DL.pdf (accessed on 10 February 2023).
  170. Domanska, O. Machine Learning Vs Traditional Programming-Avenga. Available online: https://www.avenga.com/magazine/machine-learning-programming/ (accessed on 17 March 2023).
Figure 1. The taxonomy of sensors.
Figure 1. The taxonomy of sensors.
Sensors 23 04469 g001
Figure 3. The comparison of CCD and CMOS sensors [55].
Figure 3. The comparison of CCD and CMOS sensors [55].
Sensors 23 04469 g003
Figure 4. The architecture of proposed intelligent road’s infrastructure monitoring system based on the mast-mounted surveillance station and intrusive sensors.
Figure 4. The architecture of proposed intelligent road’s infrastructure monitoring system based on the mast-mounted surveillance station and intrusive sensors.
Sensors 23 04469 g004
Figure 5. 1—The area of motion detection interest, 2—Saw-cut area of intrusive sensor presence.
Figure 5. 1—The area of motion detection interest, 2—Saw-cut area of intrusive sensor presence.
Sensors 23 04469 g005
Table 1. Taxonomy of the sensors used in transportation.
Table 1. Taxonomy of the sensors used in transportation.
Intrusive Sensors In Roadway Based SensorsPiezoelectric Sensor
Fiber Bragg Gratting Sensor
Pneumatic Tube Detector
Vibration Sensor
Loop Detector
Magnetic Sensor
Non-intrusive Sensors Sensors Side Roadway Based SensorsRFID
Acoustic Sensor
Laser Scanner
Radar
Wi-Fi
Over Roadway Based SensorsLidar
Passive Infrared Sensor
Ultrasonic Sensor
Camera Systems
Table 2. Taxonomy of the sensors that can be used in vehicles.
Table 2. Taxonomy of the sensors that can be used in vehicles.
Non-Vehicle SensorsPiezoelectric Sensor
FBG Sensor
Pneumatic Tube Detector
Loop Detector
Magnetic Sensor
Acoustic Sensor
Laser Scanner
In-Vehicle SensorsVibration Sensor
RFID
Radar
Wi-Fi
Lidar
Passive Infrared Sensor
Ultrasonic Sensor
Camera Systems
Table 3. Table of the transportation tasks solved via the sensors.
Table 3. Table of the transportation tasks solved via the sensors.
HOV Lane ManagementIncident DetectionVehicle CountingLicense Plate RecognitionVehicle Type ClassificationSpeed MeasurementParking ManagementWeight Measuring
Piezoelectric SensorIn-Vehicle monitoring
the seat occupancy approach.
[62]
Indirectly with the combination
of vehicle counting
and machine learning approach
as well as the sensor installing
as a part of the vehicles detect
crashes.
[82,83,84]
This approach uses simple statistical models
for analysis electrical signal generated
from the mechanical pressure. [85]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [85]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [87]
It uses the feature of increasing
weight impacting on strength
of generated electrical signals. [88]
FBG sensorIn-Vehicle monitoring
the seat occupancy approach.
[89,90,91]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83,92]
This approach uses simple statistical models
for analysis electrical signal generated
from the mechanical pressure. [93]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [94]
Signal wave analysis [95].Parking management system
solved via vehicle counting
entry in park houses. [96]
It uses the feature of increasing
weight impacting on strength
of generated electrical signals. [97]
Pneumatic Road
Tube
Indirectly via categorization
of vehicles proposed
for public transport.
[98]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach uses simple statistical models
for analysis electrical signal generated
from the mechanical pressure. [98]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [98]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [98]
It uses the feature of increasing
weight impacting on strength
of generated electrical signals. [99]
Vibration sensorIn-Vehicle monitoring
the seat occupancy approach.
[100]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83,101]
This approach uses simple statistical models
for analysis electrical signal generated
from the vibration. [102]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [103]
For railway transportation
or heavy vehicles,
signal processing and analysis are used. [104]
Parking management system
solved via vehicle counting
entry in park houses. [105]
It uses the feature of increasing
weight impacting on strength
of generated electrical signals. [106]
Loop DetectorIndirectly via categorization
of vehicles proposed
for public transport.
[107]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach uses simple statistical models
for analysis electrical signal generated
from the magnetic field change. [108]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [107]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [109]
Not found publication article
on using this sensor as a solution
to this problem at this moment.
Magnetic SensorIndirectly via categorization
of vehicles proposed
for public transport.
[110]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach uses simple statistical models
for analysis electrical signal generated
from the magnetic field change. [111]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses an obtained weight
features to categorize
the vehicles. [110]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [112]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Radio frequency
identification
Identification of every
passenger approach.
[113,114]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach reads the data from device
[115]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
All vehicles identified
via this system
have to carry information
about the type
of vehicle implicitly. [116]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [117]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Acoustic SensorVehicle’s engine sound analysis
correlated to vehicle occupancy.
[118]
The typical sound of a vehicle
collision detects the incident.
[119]
This approach uses simple statistical models
for analysis electrical signal generated
from the sound. [120]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses the engine’s sound
to categorize the vehicles. [121]
Signal processing and analysis
through neural network. [122]
Parking management system
solved via vehicle counting
entry in park houses. [123]
It uses the feature of increasing
weight impacting on strength
of generated electrical signals. [124]
Laser ScannerIn-Vehicle monitoring
the seat occupancy approach.
[125,126]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach scans the environment
to find vehicles. [127]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [128]
It uses the distance change
in the time interval to compute
the average speed. [129]
The possibility of exact
occupancy-free parking slots
checking. [130]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
RadarIn-Vehicle monitoring
the seat occupancy approach.
[131]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach scans the environment
to find vehicles. [132]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [133]
It uses the distance change
in the time interval to compute
the average speed. [79]
The possibility of exact
occupancy-free parking slots
checking. [134]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Wi-FiIn-Vehicle monitoring
the seat occupancy approach.
[135]
Indirectly with the combination
of vehicle counting
and machine learning approach.
[82,83]
This approach
uses the signal interuption
for detect a vehicle presence. [47]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [47]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
The possibility of exact
occupancy-free parking slots
checking. [136]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
LidarIn-Vehicle monitoring
the seat occupancy approach.
[137]
The installed sensor
could detect the distance
that leads to a collision
with other object in the traffic.
[138]
This approach scans the environment
to find vehicles. [139]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [140]
It uses the distance change
in the time interval to compute
the average speed. [141]
The possibility of exact
occupancy-free parking slots
checking. [142]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Passive
Infrared Sensor
In-Vehicle monitoring
the seat occupancy approach.
[143]
Passive Infrared Sensor
could detect a fire.
[144]
This approach scans the environment
to find vehicles. [145]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [146]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
Parking management system
solved via vehicle counting
entry in park houses. [147]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Ultrasonic SensorIn-Vehicle monitoring
the seat occupancy approach.
[148]
The installed sensor
could detect the distance
that leads to a collision
with other objects in the traffic.
[101]
This approach scans the environment
to find vehicles. [50]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
It uses
spatial parameters information. [149]
Indirectly, installing this system
as a vehicle counter in many places
in the same road section. [86]
The possibility of exact
occupancy-free parking slots
checking. [51]
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Camera SystemsPersons counting
via computer vision approach
on dedicated stations.
[61]
Object detection,
Object tracking,
and image classification
approach [150,151].
Object detection and
image classification approach [152].
Object detection
and image classification aproaches.
[153]
It uses
spatial parameters information.
Image classification
and Object Detection [154,155].
Object detection
and Object tracking [156].
Object detection
and other CV
methods [80].
Not found publication articles
on using this sensor as a solution
to this problem at this moment.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Micko, K.; Papcun, P.; Zolotova, I. Review of IoT Sensor Systems Used for Monitoring the Road Infrastructure. Sensors 2023, 23, 4469. https://doi.org/10.3390/s23094469

AMA Style

Micko K, Papcun P, Zolotova I. Review of IoT Sensor Systems Used for Monitoring the Road Infrastructure. Sensors. 2023; 23(9):4469. https://doi.org/10.3390/s23094469

Chicago/Turabian Style

Micko, Kristian, Peter Papcun, and Iveta Zolotova. 2023. "Review of IoT Sensor Systems Used for Monitoring the Road Infrastructure" Sensors 23, no. 9: 4469. https://doi.org/10.3390/s23094469

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop