Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (18)

Search Parameters:
Keywords = DIoU-NMS

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 6391 KiB  
Article
Lightweight SCD-YOLOv5s: The Detection of Small Defects on Passion Fruit with Improved YOLOv5s
by Yu Zhou, Zhenye Li, Sheng Xue, Min Wu, Tingting Zhu and Chao Ni
Agriculture 2025, 15(10), 1111; https://doi.org/10.3390/agriculture15101111 - 21 May 2025
Viewed by 569
Abstract
Accurate detection of surface defects on passion fruits is crucial for maintaining market competitiveness. Numerous small defects present significant challenges for manual inspection. Recently, deep learning (DL) has been widely applied to object detection. In this study, a lightweight neural network, StarC3SE-CBAM-DIoU-YOLOv5s (SCD-YOLOv5s), [...] Read more.
Accurate detection of surface defects on passion fruits is crucial for maintaining market competitiveness. Numerous small defects present significant challenges for manual inspection. Recently, deep learning (DL) has been widely applied to object detection. In this study, a lightweight neural network, StarC3SE-CBAM-DIoU-YOLOv5s (SCD-YOLOv5s), is proposed based on YOLOv5s for real-time detection of tiny surface defects on passion fruits. Key improvements are introduced as follows: the original C3 module in the backbone is replaced by the enhanced StarC3SE module to achieve a more efficient network structure; the CBAM module is integrated into the neck to improve the extraction of small defect features; and the CIoU loss function is substituted with DIoU-NMS to accelerate convergence and enhance detection accuracy. Experimental results show that SCD-YOLOv5s performs better than YOLOv5s, with precision increased by 13.2%, recall by 1.6%, and F1-score by 17.0%. Additionally, improvements of 6.7% in mAP@0.5 and 5.5% in mAP@0.95 are observed. Compared with manual detection, the proposed model enhances detection efficiency by reducing errors caused by subjective judgment. It also achieves faster inference speed (26.66 FPS), and reductions of 9.6% in parameters and 8.6% in weight size, while maintaining high detection performance. These results indicate that SCD-YOLOv5s is effective for defect detection in agricultural applications. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

26 pages, 10260 KiB  
Article
Only Detect Broilers Once (ODBO): A Method for Monitoring and Tracking Individual Behavior of Cage-Free Broilers
by Chengcheng Yin, Xinjie Tan, Xiaoxin Li, Mingrui Cai and Weihao Chen
Agriculture 2025, 15(7), 669; https://doi.org/10.3390/agriculture15070669 - 21 Mar 2025
Cited by 2 | Viewed by 1548
Abstract
In commercial poultry farming, automated behavioral monitoring systems hold significant potential for optimizing production efficiency and improving welfare outcomes at scale. The behavioral detection of free-range broilers matters for precision farming and animal welfare. Current research often focuses on either behavior detection or [...] Read more.
In commercial poultry farming, automated behavioral monitoring systems hold significant potential for optimizing production efficiency and improving welfare outcomes at scale. The behavioral detection of free-range broilers matters for precision farming and animal welfare. Current research often focuses on either behavior detection or individual tracking, with few studies exploring their connection. To continuously track broiler behaviors, the Only Detect Broilers Once (ODBO) method is proposed by linking behaviors with identity information. This method has a behavior detector, an individual Tracker, and a Connector. First, by integrating SimAM, WIOU, and DIOU-NMS into YOLOv8m, the high-performance YOLOv8-BeCS detector is created. It boosts P by 6.3% and AP by 3.4% compared to the original detector. Second, the designed Connector, based on the tracking-by-detection structure, transforms the tracking task, combining broiler tracking and behavior recognition. Tests on sort-series trackers show HOTA, MOTA, and IDF1 increase by 27.66%, 28%, and 27.96%, respectively, after adding the Connector. Fine-tuning experiments verify the model’s generalization. The results show this method outperforms others in accuracy, generalization, and convergence speed, providing an effective method for monitoring individual broiler behaviors. In addition, the system’s ability to simultaneously monitor individual bird welfare indicators and group dynamics could enable data-driven decisions in commercial poultry farming management. Full article
(This article belongs to the Special Issue Modeling of Livestock Breeding Environment and Animal Behavior)
Show Figures

Figure 1

15 pages, 8890 KiB  
Article
Research on Lightweight Method of Insulator Target Detection Based on Improved SSD
by Bing Zeng, Yu Zhou, Dilin He, Zhihao Zhou, Shitao Hao, Kexin Yi, Zhilong Li, Wenhua Zhang and Yunmin Xie
Sensors 2024, 24(18), 5910; https://doi.org/10.3390/s24185910 - 12 Sep 2024
Cited by 3 | Viewed by 1134
Abstract
Aiming at the problems of a large volume, slow processing speed, and difficult deployment in the edge terminal, this paper proposes a lightweight insulator detection algorithm based on an improved SSD. Firstly, the original feature extraction network VGG-16 is replaced by a lightweight [...] Read more.
Aiming at the problems of a large volume, slow processing speed, and difficult deployment in the edge terminal, this paper proposes a lightweight insulator detection algorithm based on an improved SSD. Firstly, the original feature extraction network VGG-16 is replaced by a lightweight Ghost Module network to initially achieve the lightweight model. A Feature Pyramid structure and Feature Pyramid Network (FPN+PAN) are integrated into the Neck part and a Simplified Spatial Pyramid Pooling Fast (SimSPPF) module is introduced to realize the integration of local features and global features. Secondly, multiple Spatial and Channel Squeeze-and-Excitation (scSE) attention mechanisms are introduced in the Neck part to make the model pay more attention to the channels containing important feature information. The original six detection heads are reduced to four to improve the inference speed of the network. In order to improve the recognition performance of occluded and overlapping targets, DIoU-NMS was used to replace the original non-maximum suppression (NMS). Furthermore, the channel pruning strategy is used to reduce the unimportant weight matrix of the model, and the knowledge distillation strategy is used to fine-adjust the network model after pruning, so as to ensure the detection accuracy. The experimental results show that the parameter number of the proposed model is reduced from 26.15 M to 0.61 M, the computational load is reduced from 118.95 G to 1.49 G, and the mAP is increased from 96.8% to 98%. Compared with other models, the proposed model not only guarantees the detection accuracy of the algorithm, but also greatly reduces the model volume, which provides support for the realization of visible light insulator target detection based on edge intelligence. Full article
(This article belongs to the Special Issue Advanced Fault Monitoring for Smart Power Systems)
Show Figures

Figure 1

19 pages, 9912 KiB  
Article
A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms
by Qianyong Chen, Mengshan Li, Zhenghui Lai, Jihong Zhu and Lixin Guan
J. Imaging 2024, 10(8), 197; https://doi.org/10.3390/jimaging10080197 - 13 Aug 2024
Cited by 3 | Viewed by 2682
Abstract
Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm’s capability in detecting multi-scale targets. This [...] Read more.
Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm’s capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition. Full article
Show Figures

Figure 1

29 pages, 13503 KiB  
Article
YOSMR: A Ship Detection Method for Marine Radar Based on Customized Lightweight Convolutional Networks
by Zhe Kang, Feng Ma, Chen Chen and Jie Sun
J. Mar. Sci. Eng. 2024, 12(8), 1316; https://doi.org/10.3390/jmse12081316 - 3 Aug 2024
Cited by 6 | Viewed by 1863
Abstract
In scenarios such as nearshore and inland waterways, the ship spots in a marine radar are easily confused with reefs and shorelines, leading to difficulties in ship identification. In such settings, the conventional ARPA method based on fractal detection and filter tracking performs [...] Read more.
In scenarios such as nearshore and inland waterways, the ship spots in a marine radar are easily confused with reefs and shorelines, leading to difficulties in ship identification. In such settings, the conventional ARPA method based on fractal detection and filter tracking performs relatively poorly. To accurately identify radar targets in such scenarios, a novel algorithm, namely YOSMR, based on the deep convolutional network, is proposed. The YOSMR uses the MobileNetV3(Large) network to extract ship imaging data of diverse depths and acquire feature data of various ships. Meanwhile, taking into account the issue of feature suppression for small-scale targets in algorithms composed of deep convolutional networks, the feature fusion module known as PANet has been subject to a lightweight reconstruction leveraging depthwise separable convolutions to enhance the extraction of salient features for small-scale ships while reducing model parameters and computational complexity to mitigate overfitting problems. To enhance the scale invariance of convolutional features, the feature extraction backbone is followed by an SPP module, which employs a design of four max-pooling constructs to preserve the prominent ship features within the feature representations. In the prediction head, the Cluster-NMS method and α-DIoU function are used to optimize non-maximum suppression (NMS) and positioning loss of prediction boxes, improving the accuracy and convergence speed of the algorithm. The experiments showed that the recall, accuracy, and precision of YOSMR reached 0.9308, 0.9204, and 0.9215, respectively. The identification efficacy of this algorithm exceeds that of various YOLO algorithms and other lightweight algorithms. In addition, the parameter size and calculational consumption were controlled to only 12.4 M and 8.63 G, respectively, exhibiting an 80.18% and 86.9% decrease compared to the standard YOLO model. As a result, the YOSMR displays a substantial advantage in terms of convolutional computation. Hence, the algorithm achieves an accurate identification of ships with different trail features and various scenes in marine radar images, especially in different interference and extreme scenarios, showing good robustness and applicability. Full article
Show Figures

Figure 1

18 pages, 16213 KiB  
Article
A Lightweight CER-YOLOv5s Algorithm for Detection of Construction Vehicles at Power Transmission Lines
by Pingping Yu, Yuting Yan, Xinliang Tang, Yan Shang and He Su
Appl. Sci. 2024, 14(15), 6662; https://doi.org/10.3390/app14156662 - 30 Jul 2024
Cited by 1 | Viewed by 1358
Abstract
In the context of power-line scenarios characterized by complex backgrounds and diverse scales and shapes of targets, and addressing issues such as large model parameter sizes, insufficient feature extraction, and the susceptibility to missing small targets in engineering-vehicle detection tasks, a lightweight detection [...] Read more.
In the context of power-line scenarios characterized by complex backgrounds and diverse scales and shapes of targets, and addressing issues such as large model parameter sizes, insufficient feature extraction, and the susceptibility to missing small targets in engineering-vehicle detection tasks, a lightweight detection algorithm termed CER-YOLOv5s is firstly proposed. The C3 module was restructured by embedding a lightweight Ghost bottleneck structure and convolutional attention module, enhancing the model’s ability to extract key features while reducing computational costs. Secondly, an E-BiFPN feature pyramid network is proposed, utilizing channel attention mechanisms to effectively suppress background noise and enhance the model’s focus on important regions. Bidirectional connections were introduced to optimize the feature fusion paths, improving the efficiency of multi-scale feature fusion. At the same time, in the feature fusion part, an ERM (enhanced receptive module) was added to expand the receptive field of shallow feature maps through multiple convolution repetitions, enhancing the global information perception capability in relation to small targets. Lastly, a Soft-DIoU-NMS suppression algorithm is proposed to improve the candidate box selection mechanism, addressing the issue of suboptimal detection of occluded targets. The experimental results indicated that compared with the baseline YOLOv5s algorithm, the improved algorithm reduced parameters and computations by 27.8% and 31.9%, respectively. The mean average precision (mAP) increased by 2.9%, reaching 98.3%. This improvement surpasses recent mainstream algorithms and suggests stronger robustness across various scenarios. The algorithm meets the lightweight requirements for embedded devices in power-line scenarios. Full article
Show Figures

Figure 1

26 pages, 12230 KiB  
Article
Research on Real-Time Detection of Maize Seedling Navigation Line Based on Improved YOLOv5s Lightweighting Technology
by Hailiang Gong, Xi Wang and Weidong Zhuang
Agriculture 2024, 14(1), 124; https://doi.org/10.3390/agriculture14010124 - 14 Jan 2024
Cited by 9 | Viewed by 2137
Abstract
This study focuses on real-time detection of maize crop rows using deep learning technology to meet the needs of autonomous navigation for weed removal during the maize seedling stage. Crop row recognition is affected by natural factors such as soil exposure, soil straw [...] Read more.
This study focuses on real-time detection of maize crop rows using deep learning technology to meet the needs of autonomous navigation for weed removal during the maize seedling stage. Crop row recognition is affected by natural factors such as soil exposure, soil straw residue, mutual shading of plant leaves, and light conditions. To address this issue, the YOLOv5s network model is improved by replacing the backbone network with the improved MobileNetv3, establishing a combination network model YOLOv5-M3 and using the convolutional block attention module (CBAM) to enhance detection accuracy. Distance-IoU Non-Maximum Suppression (DIoU-NMS) is used to improve the identification degree of the occluded targets, and knowledge distillation is used to increase the recall rate and accuracy of the model. The improved YOLOv5s target detection model is applied to the recognition and positioning of maize seedlings, and the optimal target position for weeding is obtained by max-min optimization. Experimental results show that the YOLOv5-M3 network model achieves 92.2% mean average precision (mAP) for crop targets and the recognition speed is 39 frames per second (FPS). This method has the advantages of high detection accuracy, fast speed, and is light weight and has strong adaptability and anti-interference ability. It determines the relative position of maize seedlings and the weeding machine in real time, avoiding squeezing or damaging the seedlings. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

16 pages, 4517 KiB  
Article
An Improved YOLOv5 Algorithm for Vulnerable Road User Detection
by Wei Yang, Xiaolin Tang, Kongming Jiang, Yang Fu and Xinling Zhang
Sensors 2023, 23(18), 7761; https://doi.org/10.3390/s23187761 - 8 Sep 2023
Cited by 8 | Viewed by 2894
Abstract
The vulnerable road users (VRUs), being small and exhibiting random movements, increase the difficulty of object detection of the autonomous emergency braking system for vulnerable road users AEBS-VRUs, with their behaviors highly random. To overcome existing problems of AEBS-VRU object detection, an enhanced [...] Read more.
The vulnerable road users (VRUs), being small and exhibiting random movements, increase the difficulty of object detection of the autonomous emergency braking system for vulnerable road users AEBS-VRUs, with their behaviors highly random. To overcome existing problems of AEBS-VRU object detection, an enhanced YOLOv5 algorithm is proposed. While the Complete Intersection over Union-Loss (CIoU-Loss) and Distance Intersection over Union-Non-Maximum Suppression (DIoU-NMS) are fused to improve the model’s convergent speed, the algorithm also incorporates a minor object detection layer to increase the performance of VRU detection. A dataset for complex AEBS-VRUS scenarios is established based on existing datasets such as Caltech, nuScenes, and Penn-Fudan, and the model is trained using migration learning based on the PyTorch framework. A number of comparative experiments using models such as YOLOv6, YOLOv7, YOLOv8 and YOLOx are carried out. The results of the comparative evaluation show that the proposed improved YOLO5 algorithm has the best overall performance in terms of efficiency, accuracy and timeliness of target detection. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

17 pages, 10566 KiB  
Article
LSD-YOLOv5: A Steel Strip Surface Defect Detection Algorithm Based on Lightweight Network and Enhanced Feature Fusion Mode
by Huan Zhao, Fang Wan, Guangbo Lei, Ying Xiong, Li Xu, Chengzhi Xu and Wen Zhou
Sensors 2023, 23(14), 6558; https://doi.org/10.3390/s23146558 - 20 Jul 2023
Cited by 23 | Viewed by 4025
Abstract
In the field of metallurgy, the timely and accurate detection of surface defects on metallic materials is a crucial quality control task. However, current defect detection approaches face challenges with large model parameters and low detection rates. To address these issues, this paper [...] Read more.
In the field of metallurgy, the timely and accurate detection of surface defects on metallic materials is a crucial quality control task. However, current defect detection approaches face challenges with large model parameters and low detection rates. To address these issues, this paper proposes a lightweight recognition model for surface damage on steel strips, named LSD-YOLOv5. First, we design a shallow feature enhancement module to replace the first Conv structure in the backbone network. Second, the Coordinate Attention mechanism is introduced into the MobileNetV2 bottleneck structure to maintain the lightweight nature of the model. Then, we propose a smaller bidirectional feature pyramid network (BiFPN-S) and combine it with Concat operation for efficient bidirectional cross-scale connectivity and weighted feature fusion. Finally, the Soft-DIoU-NMS algorithm is employed to enhance the recognition efficiency in scenarios where targets overlap. Compared with the original YOLOv5s, the LSD-YOLOv5 model achieves a reduction of 61.5% in model parameters and a 28.7% improvement in detection speed, while improving recognition accuracy by 2.4%. This demonstrates that the model achieves an optimal balance between detection accuracy and speed, while maintaining a lightweight structure. Full article
Show Figures

Figure 1

15 pages, 3037 KiB  
Article
3D-DIoU: 3D Distance Intersection over Union for Multi-Object Tracking in Point Cloud
by Sazan Ali Kamal Mohammed, Mohd Zulhakimi Ab Razak and Abdul Hadi Abd Rahman
Sensors 2023, 23(7), 3390; https://doi.org/10.3390/s23073390 - 23 Mar 2023
Cited by 11 | Viewed by 3705
Abstract
Multi-object tracking (MOT) is a prominent and important study in point cloud processing and computer vision. The main objective of MOT is to predict full tracklets of several objects in point cloud. Occlusion and similar objects are two common problems that reduce the [...] Read more.
Multi-object tracking (MOT) is a prominent and important study in point cloud processing and computer vision. The main objective of MOT is to predict full tracklets of several objects in point cloud. Occlusion and similar objects are two common problems that reduce the algorithm’s performance throughout the tracking phase. The tracking performance of current MOT techniques, which adopt the ‘tracking-by-detection’ paradigm, is degrading, as evidenced by increasing numbers of identification (ID) switch and tracking drifts because it is difficult to perfectly predict the location of objects in complex scenes that are unable to track. Since the occluded object may have been visible in former frames, we manipulated the speed and location position of the object in the previous frames in order to guess where the occluded object might have been. In this paper, we employed a unique intersection over union (IoU) method in three-dimension (3D) planes, namely a distance IoU non-maximum suppression (DIoU-NMS) to accurately detect objects, and consequently we use 3D-DIoU for an object association process in order to increase tracking robustness and speed. By using a hybrid 3D DIoU-NMS and 3D-DIoU method, the tracking speed improved significantly. Experimental findings on the Waymo Open Dataset and nuScenes dataset, demonstrate that our multistage data association and tracking technique has clear benefits over previously developed algorithms in terms of tracking accuracy. In comparison with other 3D MOT tracking methods, our proposed approach demonstrates significant enhancement in tracking performances. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

18 pages, 26908 KiB  
Article
Adaptive Slicing-Aided Hyper Inference for Small Object Detection in High-Resolution Remote Sensing Images
by Hao Zhang, Chuanyan Hao, Wanru Song, Bo Jiang and Baozhu Li
Remote Sens. 2023, 15(5), 1249; https://doi.org/10.3390/rs15051249 - 24 Feb 2023
Cited by 19 | Viewed by 6484
Abstract
In the field of object detection, deep learning models have achieved great success in recent years. Despite these advances, detecting small objects remains difficult. Most objects in aerial images have features that are a challenge for traditional object detection techniques, including small size, [...] Read more.
In the field of object detection, deep learning models have achieved great success in recent years. Despite these advances, detecting small objects remains difficult. Most objects in aerial images have features that are a challenge for traditional object detection techniques, including small size, high density, high variability, and varying orientation. Previous approaches have used slicing methods on high-resolution images or feature maps to improve performance. However, existing slicing methods inevitably lead to redundant computation. Therefore, in this article we present a novel adaptive slicing method named ASAHI (Adaptive Slicing Aided Hyper Inference), which can dramatically reduce redundant computation using an adaptive slicing size. Specifically, ASAHI focuses on the number of slices rather than the slicing size, that is, it adaptively adjusts the slicing size to control the number of slices according to the image resolution. Additionally, we replace the standard non-maximum suppression technique with Cluster-DIoU-NMS due to its improved accuracy and inference speed in the post-processing stage. In extensive experiments, ASAHI achieves competitive performance on the VisDrone and xView datasets. The results show that the mAP50 is increased by 0.9% and the computation time is reduced by 20–25% compared with state-of-the-art slicing methods on the TPH-YOLOV5 pretrained model. On the VisDrone2019-DET-val dataset, our mAP50 result is 56.4% higher, demonstrating the superiority of our approach. Full article
Show Figures

Graphical abstract

17 pages, 5179 KiB  
Article
Research on Multi-Scale Pest Detection and Identification Method in Granary Based on Improved YOLOv5
by Jinyu Chu, Yane Li, Hailin Feng, Xiang Weng and Yaoping Ruan
Agriculture 2023, 13(2), 364; https://doi.org/10.3390/agriculture13020364 - 2 Feb 2023
Cited by 24 | Viewed by 3046
Abstract
Accurately detecting and identifying granary pests is important in effectively controlling damage to a granary, ensuring food security scientifically and efficiently. In this paper, multi-scale images of seven common granary pests were collected. The dataset had 5231 images acquired with DSLR-shot, microscope, cell [...] Read more.
Accurately detecting and identifying granary pests is important in effectively controlling damage to a granary, ensuring food security scientifically and efficiently. In this paper, multi-scale images of seven common granary pests were collected. The dataset had 5231 images acquired with DSLR-shot, microscope, cell phone and online crawler. Each image contains different species of granary pests in a different background. In this paper, we designed a multi-scale granary pest recognition model, using the YOLOv5 (You Look Only Once version 5) object detection algorithm incorporating bidirectional feature pyramid network (BiFPN) with distance intersection over union, non-maximum suppression (DIOU_NMS) and efficient channel attention (ECA) modules. In addition, we compared the performance of the different models established with Efficientdet, Faster rcnn, Retinanet, SSD, YOLOx, YOLOv3, YOLOv4 and YOLOv5s, and we designed improved YOLOv5s on this dataset. The results show that the average accuracy of the model we designed for seven common pests reached 98.2%, which is the most accurate model among those identified in this paper. For further detecting the robustness of the proposed model, ablation analysis was conducted. Furthermore, the results show that the average accuracy of models established using the YOLOv5s network model combined with the attention mechanism was 96.9%. When replacing the model of PANet with BiFPN, the average accuracy reached 97.2%. At the same time, feature visualization was analyzed. The results show that the proposed model is good for capturing features of pests. The results of the model have good practical significance for the recognition of multi-scale granary pests. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

13 pages, 2076 KiB  
Article
A Tiny Model for Fast and Precise Ship Detection via Feature Channel Pruning
by Yana Yang, Shuai Xiao, Jiachen Yang and Chen Cheng
Sensors 2022, 22(23), 9331; https://doi.org/10.3390/s22239331 - 30 Nov 2022
Cited by 4 | Viewed by 1912
Abstract
It is of great significance to accurately detect ships on the ocean. To obtain higher detection performance, many researchers use deep learning to identify ships from images instead of traditional detection methods. Nevertheless, the marine environment is relatively complex, making it quite difficult [...] Read more.
It is of great significance to accurately detect ships on the ocean. To obtain higher detection performance, many researchers use deep learning to identify ships from images instead of traditional detection methods. Nevertheless, the marine environment is relatively complex, making it quite difficult to determine features of ship targets. In addition, many detection models contain a large amount of parameters, which is not suitable to deploy in devices with limited computing resources. The two problems restrict the application of ship detection. In this paper, firstly, an SAR ship detection dataset is built based on several databases, solving the problem of a small number of ship samples. Then, we integrate the SPP, ASFF, and DIOU-NMS module into original YOLOv3 to improve the ship detection performance. SPP and ASFF help enrich semantic information of ship targets. DIOU-NMS can lower the false alarm. The improved YOLOv3 has 93.37% mAP, 4.11% higher than YOLOv3 on the self-built dataset. Then, we use the MCP method to compress the improved YOLOv3. Under the pruning ratio of 80%, the obtained compressed model has only 6.7 M parameters. Experiments show that MCP outperforms NS and ThiNet. With the size of 26.8 MB, the compact model can run at 15 FPS on an NVIDIA TX2 embedded development board, 4.3 times faster than the baseline model. Our work will contribute to the development and application of ship detection on the sea. Full article
(This article belongs to the Special Issue Deep Reinforcement Learning and IoT in Intelligent System)
Show Figures

Figure 1

22 pages, 9346 KiB  
Article
Research on Surface Defect Detection of Camera Module Lens Based on YOLOv5s-Small-Target
by Gang He, Jianyun Zhou, Hu Yang, Yuan Ning and Huatao Zou
Electronics 2022, 11(19), 3189; https://doi.org/10.3390/electronics11193189 - 5 Oct 2022
Cited by 8 | Viewed by 3186
Abstract
For the problem of low resolution of camera module lens surface defect image, small target and blurred defect details leading to low detection accuracy, a camera module lens surface defect detection algorithm YOLOv5s-Defect based on improved YOLOv5s is proposed. Firstly, to solve the [...] Read more.
For the problem of low resolution of camera module lens surface defect image, small target and blurred defect details leading to low detection accuracy, a camera module lens surface defect detection algorithm YOLOv5s-Defect based on improved YOLOv5s is proposed. Firstly, to solve the problems arising from the anchor frame generated by the network through K-means clustering, the dynamic anchor frame structure DAFS is introduced in the input stage. Secondly, the SPP-D (Spatial Pyramid Pooling-Defect) improved from the SPP module is proposed. The SPP-D module is used to enhance the reuse rate of feature information in order to reduce the loss of feature information due to the maximum pooling of SPP modules. Then, the convolutional attention module is introduced to the network model of YOLOv5s, which is used to enhance the defective region features and suppress the background region features, thus improving the detection accuracy of small targets. Finally, the post-processing method of non-extreme value suppression is improved, and the improved method DIoU-NMS improves the detection accuracy of small targets in complex backgrounds. The experimental results show that the mean average precision mAP@0.5 of the YOLOv5s-Small-Target algorithm is 99.6%, 8.1% higher than that of the original YOLOv5s algorithm, the detection speed FPS is 80 f/s, and the model size is 18.7M. Compared with the traditional camera module lens surface defect detection methods, YOLOv5s-Small-Target can detect the type and location of lens surface defects more accurately and quickly, and has a smaller model volume, which is convenient for deployment in mobile terminals, meeting the demand for real-time and accuracy of camera module lens surface defect detection. Full article
(This article belongs to the Special Issue Autonomous Robots: Theory, Methods and Applications)
Show Figures

Figure 1

25 pages, 7724 KiB  
Article
YOLOv5-AC: Attention Mechanism-Based Lightweight YOLOv5 for Track Pedestrian Detection
by Haohui Lv, Hanbing Yan, Keyang Liu, Zhenwu Zhou and Junjie Jing
Sensors 2022, 22(15), 5903; https://doi.org/10.3390/s22155903 - 7 Aug 2022
Cited by 44 | Viewed by 7028
Abstract
In response to the dangerous behavior of pedestrians roaming freely on unsupervised train tracks, the real-time detection of pedestrians is urgently required to ensure the safety of trains and people. Aiming to improve the low accuracy of railway pedestrian detection, the high missed-detection [...] Read more.
In response to the dangerous behavior of pedestrians roaming freely on unsupervised train tracks, the real-time detection of pedestrians is urgently required to ensure the safety of trains and people. Aiming to improve the low accuracy of railway pedestrian detection, the high missed-detection rate of target pedestrians, and the poor retention of non-redundant boxes, YOLOv5 is adopted as the baseline to improve the effectiveness of pedestrian detection. First of all, L1 regularization is deployed before the BN layer, and the layers with smaller influence factors are removed through sparse training to achieve the effect of model pruning. In the next moment, the context extraction module is applied to the feature extraction network, and the input features are fully extracted using receptive fields of different sizes. In addition, both the context attention module CxAM and the content attention module CnAM are added to the FPN part to correct the target position deviation in the process of feature extraction so that the accuracy of detection can be improved. What is more, DIoU_NMS is employed to replace NMS as the prediction frame screening algorithm to improve the problem of detection target loss in the case of high target coincidence. Experimental results show that compared with YOLOv5, the AP of our YOLOv5-AC model for pedestrians is 95.14%, the recall is 94.22%, and the counting frame rate is 63.1 FPS. Among them, AP and recall increased by 3.78% and 3.92%, respectively, while the detection speed increased by 57.8%. The experimental results verify that our YOLOv5-AC is an effective and accurate method for pedestrian detection in railways. Full article
(This article belongs to the Special Issue Car Crash: Sensing, Monitoring and Detection)
Show Figures

Figure 1

Back to TopTop