Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (31)

Search Parameters:
Keywords = full scene solution

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 6117 KiB  
Article
Enhancing Dense-Scene Millet Appearance Quality Inspection Based on YOLO11s with Overlap-Partitioning Strategy for Procurement
by Leilei He, Ruiyang Wei, Yusong Ding, Juncai Huang, Xin Wei, Rui Li, Shaojin Wang and Longsheng Fu
Agronomy 2025, 15(6), 1284; https://doi.org/10.3390/agronomy15061284 - 23 May 2025
Viewed by 490
Abstract
Accurate millet appearance quality assessment is critical for fair procurement pricing. Traditional manual inspection is time-consuming and subjective, necessitating an automated solution. This study proposes a machine-vision-based approach using deep learning for dense-scene millet detection and quality evaluation. High-resolution images of standardized millet [...] Read more.
Accurate millet appearance quality assessment is critical for fair procurement pricing. Traditional manual inspection is time-consuming and subjective, necessitating an automated solution. This study proposes a machine-vision-based approach using deep learning for dense-scene millet detection and quality evaluation. High-resolution images of standardized millet samples were collected via smartphone and annotated into seven categories covering impurities, high-quality grains, and various defects. To address the challenges with small object detection and feature loss, the YOLO11s model with an overlap-partitioning strategy were introduced, dividing the high-resolution images into smaller patches for improved object representation. The experimental results show that the optimized model achieved a mean average precision (mAP) of 94.8%, significantly outperforming traditional whole-image detection with a mAP of 15.9%. The optimized model was deployed in a custom-developed mobile application, enabling low-cost, real-time millet inspection directly on smartphones. It can process full-resolution images (4608 × 3456 pixels) containing over 5000 kernels within 6.8 s. This work provides a practical solution for on-site quality evaluation in procurement and contributes to real-time agricultural inspection systems. Full article
Show Figures

Figure 1

11 pages, 3390 KiB  
Article
Material Sensing with Spatial and Spectral Resolution Based on an Integrated Near-Infrared Spectral Sensor and a CMOS Camera
by Ben Delaney, Sjors Buntinx, Don M. J. van Elst, Anne van Klinken, René P. J. van Veldhoven and Andrea Fiore
Sensors 2025, 25(11), 3295; https://doi.org/10.3390/s25113295 - 23 May 2025
Viewed by 447
Abstract
Measuring the composition of materials at a distance is a key requirement in industrial process monitoring, recycling, precision agriculture, and environmental monitoring. Spectral imaging in the visible or near-infrared (NIR) spectral bands provides a potential solution by combining spatial and spectral information, and [...] Read more.
Measuring the composition of materials at a distance is a key requirement in industrial process monitoring, recycling, precision agriculture, and environmental monitoring. Spectral imaging in the visible or near-infrared (NIR) spectral bands provides a potential solution by combining spatial and spectral information, and its application has seen significant growth over recent decades. Low-cost solutions for visible multispectral imaging (MSI) have been developed due to the widespread availability of silicon detectors, which are sensitive in this spectral region. In contrast, development in the NIR has been slower, primarily due to the high cost of indium gallium arsenide (InGaAs) detector arrays required for imaging. This work aims to bridge this gap by introducing a standoff material sensing concept which combines spatial and spectral resolution without the hardware requirements of traditional spectral imaging systems. It combines spatial imaging in the visible range with a CMOS camera and NIR spectral measurement at selected points of the scene using an NIR spectral sensor. This allows the chemical characterization of different objects of interest in a scene without acquiring a full spectral image. We showcase its application in plastic classification, a key functionality in sorting and recycling systems. The system demonstrated the capability to classify visually identical plastics of different types in a standoff measurement configuration and to produce spectral measurements at up to 100 points in a scene. Full article
Show Figures

Figure 1

14 pages, 4647 KiB  
Article
Rotary Panoramic and Full-Depth-of-Field Imaging System for Pipeline Inspection
by Qiang Xing, Xueqin Zhao, Kun Song, Jiawen Jiang, Xinhao Wang, Yuanyuan Huang and Haodong Wei
Sensors 2025, 25(9), 2860; https://doi.org/10.3390/s25092860 - 30 Apr 2025
Viewed by 455
Abstract
To address the adaptability and insufficient imaging quality of conventional in-pipe imaging techniques for irregular pipelines or unstructured scenes, this study proposes a novel radial rotating full-depth-of-field focusing imaging system designed to adapt to the structural complexities of irregular pipelines, which can effectively [...] Read more.
To address the adaptability and insufficient imaging quality of conventional in-pipe imaging techniques for irregular pipelines or unstructured scenes, this study proposes a novel radial rotating full-depth-of-field focusing imaging system designed to adapt to the structural complexities of irregular pipelines, which can effectively acquire tiny details with a depth of 300–960 mm inside the pipeline. Firstly, a fast full-depth-of-field imaging method driven by depth features is proposed. Secondly, a full-depth rotating imaging apparatus is developed, incorporating a zoom camera, a miniature servo rotation mechanism, and a control system, enabling 360° multi-view angles and full-depth-of-field focusing imaging. Finally, full-depth-of-field focusing imaging experiments are carried out for pipelines with depth-varying characteristics. The results demonstrate that the imaging device can acquire depth data of the pipeline interior and rapidly obtain high-definition characterization sequence images of the inner pipeline wall. In the depth-of-field segmentation with multiple view angles, the clarity of the fused image is improved by 75.3% relative to a single frame, and the SNR and PSNR reach 6.9 dB and 26.3 dB, respectively. Compared to existing pipeline closed-circuit television (CCTV) and other in-pipeline imaging techniques, the developed rotating imaging system exhibits high integration, faster imaging capabilities, and adaptive capacity. This system provides an adaptive imaging solution for detecting defects on the inner surfaces of irregular pipelines, offering significant potential for practical applications in pipeline inspection and maintenance. Full article
(This article belongs to the Special Issue Sensors in 2025)
Show Figures

Figure 1

41 pages, 7054 KiB  
Review
Seeking Solutions for Inclusively Economic, Rapid, and Safe Molecular Detection of Respiratory Infectious Diseases: Comprehensive Review from Polymerase Chain Reaction Techniques to Amplification-Free Biosensing
by Yaping Xie, Zisheng Zong, Qin Jiang, Xingxing Ke and Zhigang Wu
Micromachines 2025, 16(4), 472; https://doi.org/10.3390/mi16040472 - 15 Apr 2025
Viewed by 700
Abstract
Frequent outbreaks of respiratory infectious diseases, driven by diverse pathogens, have long posed significant threats to public health, economic productivity, and societal stability. Respiratory infectious diseases are highly contagious, characterized by short incubation periods, diverse symptoms, multiple transmission routes, susceptibility to mutations, and [...] Read more.
Frequent outbreaks of respiratory infectious diseases, driven by diverse pathogens, have long posed significant threats to public health, economic productivity, and societal stability. Respiratory infectious diseases are highly contagious, characterized by short incubation periods, diverse symptoms, multiple transmission routes, susceptibility to mutations, and distinct seasonality, contributing to their propensity for outbreaks. The absence of effective antiviral treatments and the heightened vulnerability of individuals with weakened immune systems make them more susceptible to infection, with severe cases potentially leading to complications or death. This situation becomes particularly concerning during peak seasons, such as influenza outbreaks. Therefore, early detection, diagnosis, and treatment are critical, alongside the prevention of cross-infection, ensuring patient safety, and controlling healthcare costs. To address these challenges, this review aims to identify a comprehensive, rapid, safe, and cost-effective diagnostic approach for respiratory infectious diseases. This approach is framed within the existing hierarchical healthcare system, focusing on establishing diagnostic capabilities at hospitals, community, and home levels to effectively tackle the above issues. In addition to PCR and isothermal amplification, the review also explores emerging molecular diagnostic strategies that may better address the evolving needs of respiratory disease diagnostics. A key focus is the transition from amplification technologies to amplification-free biosensing approaches, with particular attention given to their potential for home-based testing. This shift seeks to overcome the limitations of conventional amplification methods, particularly in decentralized and home diagnostics, offering a promising solution to enhance diagnostic speed and safety during outbreaks. In the future, with the integration of AI technologies into molecular amplification technologies, biosensors, and various application levels, the inclusively economic, rapid, and safe respiratory disease diagnosis solutions will be further optimized, and their accessibility will become more widespread. Full article
(This article belongs to the Special Issue Recent Progress of Lab-on-a-Chip Assays)
Show Figures

Figure 1

14 pages, 7668 KiB  
Article
A Machine Learning Method for the Fast Simulation of the Scattering Characteristics of a Target Under a Planar Layered Medium
by Zhaoyu Wang, Qinghe Zhang, Zhaoyang Shen, Lei Zhang and Han Liu
Sensors 2025, 25(8), 2481; https://doi.org/10.3390/s25082481 - 15 Apr 2025
Viewed by 406
Abstract
Numerical simulation of ground-penetrating radar (GPR) has been widely used to enhance the interpretation of GPR data and serves as a key component in Full Waveform Inversion (FWI). In response to the time-consuming numerical computation of layered medium and buried targets, which leads [...] Read more.
Numerical simulation of ground-penetrating radar (GPR) has been widely used to enhance the interpretation of GPR data and serves as a key component in Full Waveform Inversion (FWI). In response to the time-consuming numerical computation of layered medium and buried targets, which leads to inefficiency in full-wave inversion, this paper proposes a machine learning-based forward scattering rapid solution method. Using the detection of rebar buried in concrete under sand as the GPR application scenario, with scene parameters such as concrete moisture content, rebar radius, and burial depth, scattering echo signals are obtained via Finite Difference Time Domain (FDTD) simulation. Principal component analysis (PCA) is applied to reduce the dimensionality of the echo data, and the first 40 principal component weight coefficients are selected as the output of the deep learning network. An innovative cyclic nested deep learning network architecture is designed, which not only fully explores the intrinsic causal relationship between the scene parameters and the principal component weight coefficients, but also refines and corrects each predicted principal component. The numerical results demonstrate that, compared with traditional machine learning methods, the cyclic nested machine learning network architecture offers higher prediction accuracy and learning efficiency, validating the effectiveness of the proposed method. Full article
(This article belongs to the Special Issue Radar Target Detection, Imaging and Recognition)
Show Figures

Figure 1

25 pages, 16357 KiB  
Article
Enhancing Low-Light High-Dynamic-Range Image from Industrial Cameras Using Dynamic Weighting and Pyramid Fusion
by Meihan Dong, Mengyang Chai, Yinnian Liu, Chengzhong Liu and Shibing Chu
Sensors 2025, 25(8), 2452; https://doi.org/10.3390/s25082452 - 13 Apr 2025
Viewed by 672
Abstract
In order to solve the problem of imaging quality of industrial cameras for low-light and large dynamic scenes in many fields, such as smart city and target recognition, this study focuses on overcoming two core challenges: first, the loss of image details due [...] Read more.
In order to solve the problem of imaging quality of industrial cameras for low-light and large dynamic scenes in many fields, such as smart city and target recognition, this study focuses on overcoming two core challenges: first, the loss of image details due to the significant difference in light distribution in complex scenes, and second, the coexistence of dark and light areas under the constraints of the limited dynamic range of a camera. To this end, we propose a low-light high-dynamic-range image enhancement method based on dynamic weights and pyramid fusion. In order to verify the effectiveness of the method, experimental data covering full-time scenes are acquired based on an image acquisition platform built in the laboratory, and a comprehensive evaluation system combining subjective visual assessment and objective indicators is constructed. The experimental results show that, in a multi-temporal fusion task, this study’s method performs well in multiple key indicators such as information entropy (EN), average gradient (AG), edge intensity (EI), and spatial frequency (SF), making it especially suitable for imaging in low-light and high-dynamic-range environments. Specifically in localized low-light high-dynamic-range regions, compared with the best-performing comparison method, the information entropy indexes of this study’s method are improved by 4.88% and 6.09%, which fully verifies its advantages in detail restoration. The research results provide a technical solution with all-day adaptive capability for low-cost and lightweight surveillance equipment, such as intelligent transportation systems and remote sensing security systems, which has broad application prospects. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

18 pages, 1002 KiB  
Article
Optimizing Parameter Extraction in Grid Information Models Based on Improved Convolutional Neural Networks
by Xintong Li and Xiangjun Liu
Electronics 2024, 13(14), 2717; https://doi.org/10.3390/electronics13142717 - 11 Jul 2024
Cited by 2 | Viewed by 1403
Abstract
With the rapid advancement of digital technology, three-dimensional designs of Grid Information Models (GIMs) are increasingly applied in the power industry. Addressing the challenges of extracting key parameters during the GIM’s process for power grid equipment, this paper explores an innovative approach that [...] Read more.
With the rapid advancement of digital technology, three-dimensional designs of Grid Information Models (GIMs) are increasingly applied in the power industry. Addressing the challenges of extracting key parameters during the GIM’s process for power grid equipment, this paper explores an innovative approach that integrates artificial intelligence with image recognition technologies into power design engineering. The traditional methods of “template matching, feature extraction and classification, and symbol recognition” have enabled the automated processing of electrical grid equipment engineering drawings, allowing for the extraction of key information related to grid equipment. However, these methods still rely on manually designed and selected feature regions, which limits their potential for achieving full automation. This study introduces an optimized algorithm that combines enhanced Convolutional Neural Networks (CNNs) with Depth-First Search (DFS) strategies, and is specifically designed for the automated extraction of crucial GIM parameters from power grid equipment. Implemented on the design schematics of power engineering projects, this algorithm utilizes an improved CNN to precisely identify component symbols on schematics, and continues to extract text data associated with these symbols. Utilizing a scene text detector, the text data are matched with corresponding component symbols. Finally, the DFS strategy is applied to identify connections between these component symbols in the diagram, thus facilitating the automatic extraction of key GIM parameters. Experimental results demonstrate that this optimized algorithm can accurately identify basic GIM parameters, providing technical support for the automated extraction of parameters using the GIM. This study’s recognition accuracy is 91.31%, while a traditional CNN achieves 71.23% and a Faster R-CNN achieves 89.59%. Compared to existing research, the main innovation of this paper lies in the application of the combined enhanced CNN and DFS strategies for the extraction of GIM parameters in the power industry. This method not only improves the accuracy of parameter extraction but also significantly enhances processing speed, enabling the rapid and effective identification and extraction of critical information in complex power design environments. Moreover, the automated process reduces manual intervention, offering a novel solution in the field of power design. These features make this research broadly applicable and of significant practical value in the construction and maintenance of smart grids. Full article
Show Figures

Figure 1

17 pages, 15436 KiB  
Article
Developing a Flying Explorer for Autonomous Digital Modelling in Wild Unknowns
by Naizhong Zhang, Yaoqiang Pan, Yangwen Jin, Peiqi Jin, Kewei Hu, Xiao Huang and Hanwen Kang
Sensors 2024, 24(3), 1021; https://doi.org/10.3390/s24031021 - 5 Feb 2024
Viewed by 1665
Abstract
Digital modelling stands as a pivotal step in the realm of Digital Twinning. The future trend of Digital Twinning involves automated exploration and environmental modelling in complex scenes. In our study, we propose an innovative solution for robot odometry, path planning, and exploration [...] Read more.
Digital modelling stands as a pivotal step in the realm of Digital Twinning. The future trend of Digital Twinning involves automated exploration and environmental modelling in complex scenes. In our study, we propose an innovative solution for robot odometry, path planning, and exploration in unknown outdoor environments, with a focus on Digital modelling. The approach uses a minimum cost formulation with pseudo-randomly generated objectives, integrating multi-path planning and evaluation, with emphasis on full coverage of unknown maps based on feasible boundaries of interest. The approach allows for dynamic changes to expected targets and behaviours. The evaluation is conducted on a robotic platform with a lightweight 3D LiDAR sensor model. The robustness of different types of odometry is compared, and the impact of parameters on motion planning is explored. The consistency and efficiency of exploring completely unknown areas are assessed in both indoor and outdoor scenarios. The experiment shows that the method proposed in this article can complete autonomous exploration and environmental modelling tasks in complex indoor and outdoor scenes. Finally, the study concludes by summarizing the reasons for exploration failures and outlining future focuses in this domain. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

18 pages, 9809 KiB  
Article
Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
by Filip Taneski, Istvan Gyongy, Tarek Al Abbas and Robert K. Henderson
Sensors 2023, 23(21), 8943; https://doi.org/10.3390/s23218943 - 3 Nov 2023
Cited by 3 | Viewed by 2875
Abstract
Self-driving vehicles demand efficient and reliable depth-sensing technologies. Lidar, with its capability for long-distance, high-precision measurement, is a crucial component in this pursuit. However, conventional mechanical scanning implementations suffer from reliability, cost, and frame rate limitations. Solid-state lidar solutions have emerged as a [...] Read more.
Self-driving vehicles demand efficient and reliable depth-sensing technologies. Lidar, with its capability for long-distance, high-precision measurement, is a crucial component in this pursuit. However, conventional mechanical scanning implementations suffer from reliability, cost, and frame rate limitations. Solid-state lidar solutions have emerged as a promising alternative, but the vast amount of photon data processed and stored using conventional direct time-of-flight (dToF) prevents long-distance sensing unless power-intensive partial histogram approaches are used. In this paper, we introduce a groundbreaking ‘guided’ dToF approach, harnessing external guidance from other onboard sensors to narrow down the depth search space for a power and data-efficient solution. This approach centers around a dToF sensor in which the exposed time window of independent pixels can be dynamically adjusted. We utilize a 64-by-32 macropixel dToF sensor and a pair of vision cameras to provide the guiding depth estimates. Our demonstrator captures a dynamic outdoor scene at 3 fps with distances up to 75 m. Compared to a conventional full histogram approach, on-chip data is reduced by over twenty times, while the total laser cycles in each frame are reduced by at least six times compared to any partial histogram approach. The capability of guided dToF to mitigate multipath reflections is also demonstrated. For self-driving vehicles where a wealth of sensor data is already available, guided dToF opens new possibilities for efficient solid-state lidar. Full article
Show Figures

Figure 1

38 pages, 112154 KiB  
Article
Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems
by Ana Stojkovic, Jan Aelterman, David Van Hamme, Ivana Shopovska and Wilfried Philips
Sensors 2023, 23(20), 8507; https://doi.org/10.3390/s23208507 - 17 Oct 2023
Cited by 3 | Viewed by 3488
Abstract
High dynamic range (HDR) imaging technology is increasingly being used in automated driving systems (ADS) for improving the safety of traffic participants in scenes with strong differences in illumination. Therefore, a combination of HDR video, that is video with details in all illumination [...] Read more.
High dynamic range (HDR) imaging technology is increasingly being used in automated driving systems (ADS) for improving the safety of traffic participants in scenes with strong differences in illumination. Therefore, a combination of HDR video, that is video with details in all illumination regimes, and (HDR) object perception techniques that can deal with this variety in illumination is highly desirable. Although progress has been made in both HDR imaging solutions and object detection algorithms in the recent years, they have progressed independently of each other. This has led to a situation in which object detection algorithms are typically designed and constantly improved to operate on 8 bit per channel content. This makes these algorithms not ideally suited for use in HDR data processing, which natively encodes to a higher bit-depth (12 bits/16 bits per channel). In this paper, we present and evaluate two novel convolutional neural network (CNN) architectures that intelligently convert high bit depth HDR images into 8-bit images. We attempt to optimize reconstruction quality by focusing on ADS object detection quality. The first research novelty is to jointly perform tone-mapping with demosaicing by additionally successfully suppressing noise and demosaicing artifacts. The first CNN performs tone-mapping with noise suppression on a full-color HDR input, while the second performs joint demosaicing and tone-mapping with noise suppression on a raw HDR input. The focus is to increase the detectability of traffic-related objects in the reconstructed 8-bit content, while ensuring that the realism of the standard dynamic range (SDR) content in diverse conditions is preserved. The second research novelty is that for the first time, to the best of our knowledge, a thorough comparative analysis against the state-of-the-art tone-mapping and demosaicing methods is performed with respect to ADS object detection accuracy on traffic-related content that abounds with diverse challenging (i.e., boundary cases) scenes. The evaluation results show that the two proposed networks have better performance in object detection accuracy and image quality, than both SDR content and content obtained with the state-of-the-art tone-mapping and demosaicing algorithms. Full article
(This article belongs to the Special Issue Vision Sensors: Image Processing Technologies and Applications)
Show Figures

Figure 1

11 pages, 550 KiB  
Review
Enhancing the Evidence with Algorithms: How Artificial Intelligence Is Transforming Forensic Medicine
by Alin-Ionut Piraianu, Ana Fulga, Carmina Liana Musat, Oana-Roxana Ciobotaru, Diana Gina Poalelungi, Elena Stamate, Octavian Ciobotaru and Iuliu Fulga
Diagnostics 2023, 13(18), 2992; https://doi.org/10.3390/diagnostics13182992 - 19 Sep 2023
Cited by 40 | Viewed by 8721
Abstract
Background: The integration of artificial intelligence (AI) into various fields has ushered in a new era of multidisciplinary progress. Defined as the ability of a system to interpret external data, learn from it, and adapt to specific tasks, AI is poised to revolutionize [...] Read more.
Background: The integration of artificial intelligence (AI) into various fields has ushered in a new era of multidisciplinary progress. Defined as the ability of a system to interpret external data, learn from it, and adapt to specific tasks, AI is poised to revolutionize the world. In forensic medicine and pathology, algorithms play a crucial role in data analysis, pattern recognition, anomaly identification, and decision making. This review explores the diverse applications of AI in forensic medicine, encompassing fields such as forensic identification, ballistics, traumatic injuries, postmortem interval estimation, forensic toxicology, and more. Results: A thorough review of 113 articles revealed a subset of 32 papers directly relevant to the research, covering a wide range of applications. These included forensic identification, ballistics and additional factors of shooting, traumatic injuries, post-mortem interval estimation, forensic toxicology, sexual assaults/rape, crime scene reconstruction, virtual autopsy, and medical act quality evaluation. The studies demonstrated the feasibility and advantages of employing AI technology in various facets of forensic medicine and pathology. Conclusions: The integration of AI in forensic medicine and pathology offers promising prospects for improving accuracy and efficiency in medico-legal practices. From forensic identification to post-mortem interval estimation, AI algorithms have shown the potential to reduce human subjectivity, mitigate errors, and provide cost-effective solutions. While challenges surrounding ethical considerations, data security, and algorithmic correctness persist, continued research and technological advancements hold the key to realizing the full potential of AI in forensic applications. As the field of AI continues to evolve, it is poised to play an increasingly pivotal role in the future of forensic medicine and pathology. Full article
(This article belongs to the Special Issue AI and Big Data in Healthcare)
Show Figures

Figure 1

25 pages, 12590 KiB  
Article
High-Accuracy Filtering of Forest Scenes Based on Full-Waveform LiDAR Data and Hyperspectral Images
by Wenjun Luo, Hongchao Ma, Jialin Yuan, Liang Zhang, Haichi Ma, Zhan Cai and Weiwei Zhou
Remote Sens. 2023, 15(14), 3499; https://doi.org/10.3390/rs15143499 - 12 Jul 2023
Cited by 12 | Viewed by 2111
Abstract
Airborne light detection and ranging (LiDAR) technology has been widely utilized for collecting three-dimensional (3D) point cloud data on forest scenes, enabling the generation of high-accuracy digital elevation models (DEMs) for the efficient investigation and management of forest resources. Point cloud filtering serves [...] Read more.
Airborne light detection and ranging (LiDAR) technology has been widely utilized for collecting three-dimensional (3D) point cloud data on forest scenes, enabling the generation of high-accuracy digital elevation models (DEMs) for the efficient investigation and management of forest resources. Point cloud filtering serves as the crucial initial step in DEM generation, directly influencing the accuracy of the resulting DEM. However, forest filtering presents challenges in dealing with sparse point clouds and selecting appropriate initial ground points. The introduction of full-waveform LiDAR data offers a potential solution to the problem of sparse point clouds. Additionally, advancements in multi-source data integration and machine learning algorithms have created new avenues that can address the issue of initial ground point selection. To tackle these challenges, this paper proposes a novel filtering method for forest scenes utilizing full-waveform LiDAR data and hyperspectral image data. The proposed method consists of two main steps. Firstly, we employ the improved dynamic graph convolutional neural network (IDGCNN) to extract initial ground points. In this step, we utilize three types of low-correlation features: LiDAR features, waveform features, and spectral features. To enhance its accuracy and adaptability, a self-attention module was incorporated into the DGCNN algorithm. Comparative experiments were conducted to evaluate the effectiveness of the algorithm, demonstrating that the IDGCNN algorithm achieves the highest classification accuracy with an overall accuracy (OA) value of 99.38% and a kappa coefficient of 95.95%. The second-best performer was the RandLA-net algorithm, achieving an OA value of 98.73% and a kappa coefficient of 91.68%. The second step involves refining the initial ground points using the cloth simulation filter (CSF) algorithm. By employing the CSF algorithm, non-ground points present in the initial ground points are effectively filtered out. To validate the efficacy of the proposed filtering method, we generated a DEM with a resolution of 0.5 using the ground points extracted in the first step, the refined ground points obtained with the combination of the first and second steps, and the ground points obtained directly using the CSF algorithm. A comparative analysis with 23 reference control points revealed the effectiveness of our proposed method, as evidenced by the median error of 0.41 m, maximum error of 0.75 m, and average error of 0.33 m. Full article
Show Figures

Graphical abstract

18 pages, 4295 KiB  
Article
Deep Learning-Based Cost-Effective and Responsive Robot for Autism Treatment
by Aditya Singh, Kislay Raj, Teerath Kumar, Swapnil Verma and Arunabha M. Roy
Drones 2023, 7(2), 81; https://doi.org/10.3390/drones7020081 - 23 Jan 2023
Cited by 105 | Viewed by 7920
Abstract
Recent studies state that, for a person with autism spectrum disorder, learning and improvement is often seen in environments where technological tools are involved. A robot is an excellent tool to be used in therapy and teaching. It can transform teaching methods, not [...] Read more.
Recent studies state that, for a person with autism spectrum disorder, learning and improvement is often seen in environments where technological tools are involved. A robot is an excellent tool to be used in therapy and teaching. It can transform teaching methods, not just in the classrooms but also in the in-house clinical practices. With the rapid advancement in deep learning techniques, robots became more capable of handling human behaviour. In this paper, we present a cost-efficient, socially designed robot called ‘Tinku’, developed to assist in teaching special needs children. ‘Tinku’ is low cost but is full of features and has the ability to produce human-like expressions. Its design is inspired by the widely accepted animated character ‘WALL-E’. Its capabilities include offline speech processing and computer vision—we used light object detection models, such as Yolo v3-tiny and single shot detector (SSD)—for obstacle avoidance, non-verbal communication, expressing emotions in an anthropomorphic way, etc. It uses an onboard deep learning technique to localize the objects in the scene and uses the information for semantic perception. We have developed several lessons for training using these features. A sample lesson about brushing is discussed to show the robot’s capabilities. Tinku is cute, and loaded with lots of features, and the management of all the processes is mind-blowing. It is developed in the supervision of clinical experts and its condition for application is taken care of. A small survey on the appearance is also discussed. More importantly, it is tested on small children for the acceptance of the technology and compatibility in terms of voice interaction. It helps autistic kids using state-of-the-art deep learning models. Autism Spectral disorders are being increasingly identified today’s world. The studies show that children are prone to interact with technology more comfortably than a with human instructor. To fulfil this demand, we presented a cost-effective solution in the form of a robot with some common lessons for the training of an autism-affected child. Full article
(This article belongs to the Topic Artificial Intelligence in Sensors)
Show Figures

Figure 1

18 pages, 872 KiB  
Article
SAUSA: Securing Access, Usage, and Storage of 3D Point CloudData by a Blockchain-Based Authentication Network
by Ronghua Xu, Yu Chen, Genshe Chen and Erik Blasch
Future Internet 2022, 14(12), 354; https://doi.org/10.3390/fi14120354 - 28 Nov 2022
Cited by 6 | Viewed by 2719
Abstract
The rapid development of three-dimensional (3D) acquisition technology based on 3D sensors provides a large volume of data, which are often represented in the form of point clouds. Point cloud representation can preserve the original geometric information along with associated attributes in a [...] Read more.
The rapid development of three-dimensional (3D) acquisition technology based on 3D sensors provides a large volume of data, which are often represented in the form of point clouds. Point cloud representation can preserve the original geometric information along with associated attributes in a 3D space. Therefore, it has been widely adopted in many scene-understanding-related applications such as virtual reality (VR) and autonomous driving. However, the massive amount of point cloud data aggregated from distributed 3D sensors also poses challenges for secure data collection, management, storage, and sharing. Thanks to the characteristics of decentralization and security, Blockchain has great potential to improve point cloud services and enhance security and privacy preservation. Inspired by the rationales behind the software-defined network (SDN) technology, this paper envisions SAUSA, a Blockchain-based authentication network that is capable of recording, tracking, and auditing the access, usage, and storage of 3D point cloud datasets in their life-cycle in a decentralized manner. SAUSA adopts an SDN-inspired point cloud service architecture, which allows for efficient data processing and delivery to satisfy diverse quality-of-service (QoS) requirements. A Blockchain-based authentication framework is proposed to ensure security and privacy preservation in point cloud data acquisition, storage, and analytics. Leveraging smart contracts for digitizing access control policies and point cloud data on the Blockchain, data owners have full control of their 3D sensors and point clouds. In addition, anyone can verify the authenticity and integrity of point clouds in use without relying on a third party. Moreover, SAUSA integrates a decentralized storage platform to store encrypted point clouds while recording references of raw data on the distributed ledger. Such a hybrid on-chain and off-chain storage strategy not only improves robustness and availability, but also ensures privacy preservation for sensitive information in point cloud applications. A proof-of-concept prototype is implemented and tested on a physical network. The experimental evaluation validates the feasibility and effectiveness of the proposed SAUSA solution. Full article
(This article belongs to the Special Issue Security and Privacy in Blockchains and the IoT II)
Show Figures

Figure 1

21 pages, 2333 KiB  
Article
Occlusion and Deformation Handling Visual Tracking for UAV via Attention-Based Mask Generative Network
by Yashuo Bai, Yong Song, Yufei Zhao, Ya Zhou, Xiyan Wu, Yuxin He, Zishuo Zhang, Xin Yang and Qun Hao
Remote Sens. 2022, 14(19), 4756; https://doi.org/10.3390/rs14194756 - 23 Sep 2022
Cited by 8 | Viewed by 2551
Abstract
Although the performance of unmanned aerial vehicle (UAV) tracking has benefited from the successful application of discriminative correlation filters (DCF) and convolutional neural networks (CNNs), UAV tracking under occlusion and deformation remains a challenge. The main dilemma is that challenging scenes, such as [...] Read more.
Although the performance of unmanned aerial vehicle (UAV) tracking has benefited from the successful application of discriminative correlation filters (DCF) and convolutional neural networks (CNNs), UAV tracking under occlusion and deformation remains a challenge. The main dilemma is that challenging scenes, such as occlusion or deformation, are very complex and changeable, making it difficult to obtain training data covering all situations, resulting in trained networks that may be confused by new contexts that differ from historical information. Data-driven strategies are the main direction of current solutions, but gathering large-scale datasets with object instances under various occlusion and deformation conditions is difficult and lacks diversity. This paper proposes an attention-based mask generation network (AMGN) for UAV-specific tracking, which combines the attention mechanism and adversarial learning to improve the tracker’s ability to handle occlusion and deformation. After the base CNN extracts the deep features of the candidate region, a series of masks are determined by the spatial attention module and sent to the generator, and the generator discards some features according to these masks to simulate the occlusion and deformation of the object, producing more hard positive samples. The discriminator seeks to distinguish these hard positive samples while guiding mask generation. Such adversarial learning can effectively complement occluded and deformable positive samples in the feature space, allowing to capture more robust features to distinguish objects from backgrounds. Comparative experiments show that our AMGN-based tracker achieves the highest area under curve (AUC) of 0.490 and 0.349, and the highest precision scores of 0.742 and 0.662, on the UAV123 tracking benchmark with partial and full occlusion attributes, respectively. It also achieves the highest AUC of 0.555 and the highest precision score of 0.797 on the DTB70 tracking benchmark with the deformation attribute. On the UAVDT tracking benchmark with the large occlusion attribute, it achieves the highest AUC of 0.407 and the highest precision score of 0.582. Full article
(This article belongs to the Special Issue Recent Progress in UAV-AI Remote Sensing)
Show Figures

Figure 1

Back to TopTop