Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (40)

Search Parameters:
Keywords = BRISQUE

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 7391 KiB  
Article
Reliable QoE Prediction in IMVCAs Using an LMM-Based Agent
by Michael Sidorov, Tamir Berger, Jonathan Sterenson, Raz Birman and Ofer Hadar
Sensors 2025, 25(14), 4450; https://doi.org/10.3390/s25144450 - 17 Jul 2025
Viewed by 282
Abstract
Face-to-face interaction is one of the most natural forms of human communication. Unsurprisingly, Video Conferencing (VC) Applications have experienced a significant rise in demand over the past decade. With the widespread availability of cellular devices equipped with high-resolution cameras, Instant Messaging Video Call [...] Read more.
Face-to-face interaction is one of the most natural forms of human communication. Unsurprisingly, Video Conferencing (VC) Applications have experienced a significant rise in demand over the past decade. With the widespread availability of cellular devices equipped with high-resolution cameras, Instant Messaging Video Call Applications (IMVCAs) now constitute a substantial portion of VC communications. Given the multitude of IMVCA options, maintaining a high Quality of Experience (QoE) is critical. While content providers can measure QoE directly through end-to-end connections, Internet Service Providers (ISPs) must infer QoE indirectly from network traffic—a non-trivial task, especially when most traffic is encrypted. In this paper, we analyze a large dataset collected from WhatsApp IMVCA, comprising over 25,000 s of VC sessions. We apply four Machine Learning (ML) algorithms and a Large Multimodal Model (LMM)-based agent, achieving mean errors of 4.61%, 5.36%, and 13.24% for three popular QoE metrics: BRISQUE, PIQE, and FPS, respectively. Full article
Show Figures

Figure 1

36 pages, 26652 KiB  
Article
Low-Light Image Enhancement for Driving Condition Recognition Through Multi-Band Images Fusion and Translation
by Dong-Min Son and Sung-Hak Lee
Mathematics 2025, 13(9), 1418; https://doi.org/10.3390/math13091418 - 25 Apr 2025
Viewed by 538
Abstract
When objects are obscured by shadows or dim surroundings, image quality is improved by fusing near-infrared and visible-light images. At night, when visible and NIR lights are insufficient, long-wave infrared (LWIR) imaging can be utilized, necessitating the attachment of a visible-light sensor to [...] Read more.
When objects are obscured by shadows or dim surroundings, image quality is improved by fusing near-infrared and visible-light images. At night, when visible and NIR lights are insufficient, long-wave infrared (LWIR) imaging can be utilized, necessitating the attachment of a visible-light sensor to an LWIR camera to simultaneously capture both LWIR and visible-light images. This camera configuration enables the acquisition of infrared images at various wavelengths depending on the time of day. To effectively fuse clear visible regions from the visible-light spectrum with those from the LWIR spectrum, a multi-band fusion method is proposed. The proposed fusion process subsequently combines detailed information from infrared and visible-light images, enhancing object visibility. Additionally, this process compensates for color differences in visible-light images, resulting in a natural and visually consistent output. The fused images are further enhanced using a night-to-day image translation module, which improves overall brightness and reduces noise. This night-to-day translation module is a trained CycleGAN-based module that adjusts object brightness in nighttime images to levels comparable to daytime images. The effectiveness and superiority of the proposed method are validated using image quality metrics. The proposed method significantly contributes to image enhancement, achieving the best average scores compared to other methods, with a BRISQUE of 30.426 and a PIQE of 22.186. This study improves the accuracy of human and object recognition in CCTV systems and provides a potential image-processing tool for autonomous vehicles. Full article
Show Figures

Figure 1

17 pages, 8550 KiB  
Article
Enhancing Historical Aerial Photographs: A New Approach Based on Non-Reference Metric and Photo Interpretation Elements
by Abdullah Harun Incekara and Dursun Zafer Seker
Sensors 2025, 25(7), 2126; https://doi.org/10.3390/s25072126 - 27 Mar 2025
Cited by 1 | Viewed by 456
Abstract
Deep learning-based super-resolution (SR) is an effective state-of-the-art technique for enhancing low-resolution images. This study explains a hierarchical dataset structure within the scope of enhancing grayscale historical aerial photographs with a basic SR model and relates it to non-reference image quality metric. The [...] Read more.
Deep learning-based super-resolution (SR) is an effective state-of-the-art technique for enhancing low-resolution images. This study explains a hierarchical dataset structure within the scope of enhancing grayscale historical aerial photographs with a basic SR model and relates it to non-reference image quality metric. The dataset was structured based on the hierarchy of photo interpretation elements. Images of bare land and forestry areas were evaluated as the primary category containing tone and color elements, images of residential areas as the secondary category containing shape and size elements, and images of farmland areas as the tertiary category containing pattern elements. Instead of training all images in all categories at once, which is the issue that any SR model with low number of parameters has difficulty handling, each category was trained separately. Test images containing the features of each category were enhanced separately, which means three enhanced images for one test image. The obtained images were divided into equal parts of 5 × 5 pixel size, and the final image was created by concatenating those that were determined to be of higher quality based on the Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE) metric values. Subsequently, comparative analyses based on visual interpretation and reference-based image quality metrics proved that the approach to the dataset structure positively impacted the results. Full article
(This article belongs to the Special Issue Computational Optical Sensing and Imaging)
Show Figures

Figure 1

18 pages, 22866 KiB  
Article
Real-Time Compensation for Unknown Image Displacement and Rotation in Infrared Multispectral Camera Push-Broom Imaging
by Tongxu Zhang, Guoliang Tang, Shouzheng Zhu, Fang Ding, Wenli Wu, Jindong Bai, Chunlai Li and Jianyu Wang
Remote Sens. 2025, 17(7), 1113; https://doi.org/10.3390/rs17071113 - 21 Mar 2025
Viewed by 654
Abstract
Digital time-delay integration (TDI) enhances the signal-to-noise ratio (SNR) in infrared (IR) imaging, but its effectiveness in push-broom scanning is contingent upon maintaining a stable image shift velocity. Unpredictable image shifts and rotations, caused by carrier or scene movement, can affect the imaging [...] Read more.
Digital time-delay integration (TDI) enhances the signal-to-noise ratio (SNR) in infrared (IR) imaging, but its effectiveness in push-broom scanning is contingent upon maintaining a stable image shift velocity. Unpredictable image shifts and rotations, caused by carrier or scene movement, can affect the imaging process. This paper proposes an advanced technical approach for infrared multispectral TDI imaging. This methodology concurrently estimates the image shift and rotation between frames by utilizing a high-resolution visible camera aligned parallel to the optical axis of the IR camera. Subsequently, parameter prediction is conducted using the Kalman model, and real-time compensation is achieved by dynamically adjusting the infrared TDI integration unit based on the predicted parameters. Simulation and experimental results demonstrate that the proposed algorithm enhances the BRISQUE score of the TDI images by 21.37%, thereby validating its efficacy in push-scan imaging systems characterized by velocity-height ratios instability and varying camera attitudes. This research constitutes a significant contribution to the advancement of high-precision real-time compensation for image shift and rotation in infrared remote sensing and industrial inspection applications. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

23 pages, 3354 KiB  
Article
Simultaneous Learning Knowledge Distillation for Image Restoration: Efficient Model Compression for Drones
by Yongheng Zhang
Drones 2025, 9(3), 209; https://doi.org/10.3390/drones9030209 - 14 Mar 2025
Viewed by 1140
Abstract
Deploying high-performance image restoration models on drones is critical for applications like autonomous navigation, surveillance, and environmental monitoring. However, the computational and memory limitations of drones pose significant challenges to utilizing complex image restoration models in real-world scenarios. To address this issue, we [...] Read more.
Deploying high-performance image restoration models on drones is critical for applications like autonomous navigation, surveillance, and environmental monitoring. However, the computational and memory limitations of drones pose significant challenges to utilizing complex image restoration models in real-world scenarios. To address this issue, we propose the Simultaneous Learning Knowledge Distillation (SLKD) framework, specifically designed to compress image restoration models for resource-constrained drones. SLKD introduces a dual-teacher, single-student architecture that integrates two complementary learning strategies: Degradation Removal Learning (DRL) and Image Reconstruction Learning (IRL). In DRL, the student encoder learns to eliminate degradation factors by mimicking Teacher A, which processes degraded images utilizing a BRISQUE-based extractor to capture degradation-sensitive natural scene statistics. Concurrently, in IRL, the student decoder reconstructs clean images by learning from Teacher B, which processes clean images, guided by a PIQE-based extractor that emphasizes the preservation of edge and texture features essential for high-quality reconstruction. This dual-teacher approach enables the student model to learn from both degraded and clean images simultaneously, achieving robust image restoration while significantly reducing computational complexity. Experimental evaluations across five benchmark datasets and three restoration tasks—deraining, deblurring, and dehazing—demonstrate that, compared to the teacher models, the SLKD student models achieve an average reduction of 85.4% in FLOPs and 85.8% in model parameters, with only a slight average decrease of 2.6% in PSNR and 0.9% in SSIM. These results highlight the practicality of integrating SLKD-compressed models into autonomous systems, offering efficient and real-time image restoration for aerial platforms operating in challenging environments. Full article
(This article belongs to the Special Issue Intelligent Image Processing and Sensing for Drones, 2nd Edition)
Show Figures

Figure 1

24 pages, 6411 KiB  
Article
CNN-Based Kidney Segmentation Using a Modified CLAHE Algorithm
by Abror Shavkatovich Buriboev, Ahmadjon Khashimov, Akmal Abduvaitov and Heung Seok Jeon
Sensors 2024, 24(23), 7703; https://doi.org/10.3390/s24237703 - 2 Dec 2024
Cited by 4 | Viewed by 1840
Abstract
This paper presents an enhanced approach to kidney segmentation using a modified CLAHE preprocessing method, aimed at improving image clarity and CNN performance on the KiTS19 dataset. To assess the impact of the modified CLAHE method, we conducted quality evaluations using the BRISQUE [...] Read more.
This paper presents an enhanced approach to kidney segmentation using a modified CLAHE preprocessing method, aimed at improving image clarity and CNN performance on the KiTS19 dataset. To assess the impact of the modified CLAHE method, we conducted quality evaluations using the BRISQUE metric, comparing the original, standard CLAHE and modified CLAHE versions of the dataset. The BRISQUE score decreased from 28.8 in the original dataset to 21.1 with the modified CLAHE method, indicating a significant improvement in image quality. Furthermore, CNN segmentation accuracy rose from 0.951 with the original dataset to 0.996 with the modified CLAHE method, outperforming the accuracy achieved with standard CLAHE preprocessing (0.969). These results highlight the benefits of the modified CLAHE method in refining image quality and enhancing segmentation performance. This study highlights the value of adaptive preprocessing in medical imaging workflows and shows that CNN-based kidney segmentation accuracy may be greatly increased by altering conventional CLAHE. Our method provides insightful information on optimizing preprocessing for medical imaging applications, leading to more accurate and dependable segmentation results for better clinical diagnosis. Full article
(This article belongs to the Special Issue Machine and Deep Learning in Sensing and Imaging)
Show Figures

Figure 1

13 pages, 4566 KiB  
Article
Assessment of Feldkamp-Davis-Kress Reconstruction Parameters in Overall Image Quality in Cone Beam Computed Tomography
by Hajin Kim, Jun-Seon Choi and Youngjin Lee
Appl. Sci. 2024, 14(16), 7058; https://doi.org/10.3390/app14167058 - 12 Aug 2024
Cited by 1 | Viewed by 2667
Abstract
In low-dose cone beam computed tomography (CT), the insufficient number of photons inevitably results in noise, which reduces the accuracy of disease diagnosis. One approach to improving the image quality of CT images acquired using a low-dose protocol involves the utilization of a [...] Read more.
In low-dose cone beam computed tomography (CT), the insufficient number of photons inevitably results in noise, which reduces the accuracy of disease diagnosis. One approach to improving the image quality of CT images acquired using a low-dose protocol involves the utilization of a reconstruction algorithm that efficiently reduces noise. In this study, we modeled the Feldkamp–Davis–Kress (FDK) algorithm using various filters and projection angles and applied it to the reconstruction process using CT simulation. To quantitatively evaluate the quality of the reconstruction images, we measured the coefficient of variation (COV), and signal-to-noise ratio (SNR) in the air, brain, and bone regions to evaluate the noise level. Furthermore, we calculated root mean square error (RMSE), universal image quality index (UQI), and blind/referenceless image spatial quality evaluator (BRISQUE) as similarity and no-reference evaluation. The Hann filter of the FDK algorithm showed superior performance in terms of COV, SNR, RMSE, and UQI compared to the other filters. In addition, when analyzing the COV and SNR results, we observed that image quality increased significantly at projection angles smaller than approximately 2.8°. Moreover, based on BRISQUE results, we confirm that the Shepp–Logan filter exhibited the most superior performance. In conclusion, we believe that the application of the Hann filter in the FDK reconstruction process offers significant advantages in improving the image quality acquired under a low-dose protocol, and we expect that our study will be a preliminary study of no-reference evaluation of CT reconstruction images. Full article
(This article belongs to the Special Issue Diagnosis of Medical Imaging)
Show Figures

Figure 1

11 pages, 26704 KiB  
Article
No-Reference-Based and Noise Level Evaluations of Cinematic Rendering in Bone Computed Tomography
by Jina Shim and Youngjin Lee
Bioengineering 2024, 11(6), 563; https://doi.org/10.3390/bioengineering11060563 - 2 Jun 2024
Cited by 1 | Viewed by 1222
Abstract
Cinematic rendering (CR) is a new 3D post-processing technology widely used to produce bone computed tomography (CT) images. This study aimed to evaluate the performance quality of CR in bone CT images using blind quality and noise level evaluations. Bone CT images of [...] Read more.
Cinematic rendering (CR) is a new 3D post-processing technology widely used to produce bone computed tomography (CT) images. This study aimed to evaluate the performance quality of CR in bone CT images using blind quality and noise level evaluations. Bone CT images of the face, shoulder, lumbar spine, and wrist were acquired. Volume rendering (VR), which is widely used in the field of diagnostic medical imaging, was additionally set along with CR. A no-reference-based blind/referenceless image spatial quality evaluator (BRISQUE) and coefficient of variation (COV) were used to evaluate the overall quality of the acquired images. The average BRISQUE values derived from the four areas were 39.87 and 46.44 in CR and VR, respectively. The difference between the two values was approximately 1.16, and the difference between the resulting values increased, particularly in the bone CT image, where metal artifacts were observed. In addition, we confirmed that the COV value improved by 2.20 times on average when using CR compared to VR. This study proved that CR is useful in reconstructing bone CT 3D images and that various applications in the diagnostic medical field will be possible. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

19 pages, 2793 KiB  
Article
Key-Point-Descriptor-Based Image Quality Evaluation in Photogrammetry Workflows
by Dalius Matuzevičius, Vytautas Urbanavičius, Darius Miniotas, Šarūnas Mikučionis, Raimond Laptik and Andrius Ušinskas
Electronics 2024, 13(11), 2112; https://doi.org/10.3390/electronics13112112 - 29 May 2024
Cited by 1 | Viewed by 2136
Abstract
Photogrammetry depends critically on the quality of the images used to reconstruct accurate and detailed 3D models. Selection of high-quality images not only improves the accuracy and resolution of the resulting 3D models, but also contributes to the efficiency of the photogrammetric process [...] Read more.
Photogrammetry depends critically on the quality of the images used to reconstruct accurate and detailed 3D models. Selection of high-quality images not only improves the accuracy and resolution of the resulting 3D models, but also contributes to the efficiency of the photogrammetric process by reducing data redundancy and computational demands. This study presents a novel approach to image quality evaluation tailored for photogrammetric applications that uses the key point descriptors typically encountered in image matching. Using a LightGBM ranker model, this research evaluates the effectiveness of key point descriptors such as SIFT, SURF, BRISK, ORB, KAZE, FREAK, and SuperPoint in predicting image quality. These descriptors are evaluated for their ability to indicate image quality based on the image patterns they capture. Experiments conducted on various publicly available image datasets show that descriptor-based methods outperform traditional no-reference image quality metrics such as BRISQUE, NIQE, PIQE, and BIQAA and a simple sharpness-based image quality evaluation method. The experimental results highlight the potential of using key-point-descriptor-based image quality evaluation methods to improve the photogrammetric workflow by selecting high-quality images for 3D modeling. Full article
(This article belongs to the Special Issue IoT-Enabled Smart Devices and Systems in Smart Environments)
Show Figures

Figure 1

13 pages, 2998 KiB  
Technical Note
Image Quality Assessment Tool for Conventional and Dynamic Magnetic Resonance Imaging Acquisitions
by Katerina Nikiforaki, Ioannis Karatzanis, Aikaterini Dovrou, Maciej Bobowicz, Katarzyna Gwozdziewicz, Oliver Díaz, Manolis Tsiknakis, Dimitrios I. Fotiadis, Karim Lekadir and Kostas Marias
J. Imaging 2024, 10(5), 115; https://doi.org/10.3390/jimaging10050115 - 9 May 2024
Cited by 4 | Viewed by 3955
Abstract
Image quality assessment of magnetic resonance imaging (MRI) data is an important factor not only for conventional diagnosis and protocol optimization but also for fairness, trustworthiness, and robustness of artificial intelligence (AI) applications, especially on large heterogeneous datasets. Information on image quality in [...] Read more.
Image quality assessment of magnetic resonance imaging (MRI) data is an important factor not only for conventional diagnosis and protocol optimization but also for fairness, trustworthiness, and robustness of artificial intelligence (AI) applications, especially on large heterogeneous datasets. Information on image quality in multi-centric studies is important to complement the contribution profile from each data node along with quantity information, especially when large variability is expected, and certain acceptance criteria apply. The main goal of this work was to present a tool enabling users to assess image quality based on both subjective criteria as well as objective image quality metrics used to support the decision on image quality based on evidence. The evaluation can be performed on both conventional and dynamic MRI acquisition protocols, while the latter is also checked longitudinally across dynamic series. The assessment provides an overall image quality score and information on the types of artifacts and degrading factors as well as a number of objective metrics for automated evaluation across series (BRISQUE score, Total Variation, PSNR, SSIM, FSIM, MS-SSIM). Moreover, the user can define specific regions of interest (ROIs) to calculate the regional signal-to-noise ratio (SNR) and contrast-to-noise ratio (CNR), thus individualizing the quality output to specific use cases, such as tissue-specific contrast or regional noise quantification. Full article
Show Figures

Figure 1

20 pages, 4604 KiB  
Article
Full-Process Adaptive Encoding and Decoding Framework for Remote Sensing Images Based on Compression Sensing
by Huiling Hu, Chunyu Liu, Shuai Liu, Shipeng Ying, Chen Wang and Yi Ding
Remote Sens. 2024, 16(9), 1529; https://doi.org/10.3390/rs16091529 - 26 Apr 2024
Cited by 2 | Viewed by 1431
Abstract
Faced with the problem of incompatibility between traditional information acquisition mode and spaceborne earth observation tasks, starting from the general mathematical model of compressed sensing, a theoretical model of block compressed sensing was established, and a full-process adaptive coding and decoding compressed sensing [...] Read more.
Faced with the problem of incompatibility between traditional information acquisition mode and spaceborne earth observation tasks, starting from the general mathematical model of compressed sensing, a theoretical model of block compressed sensing was established, and a full-process adaptive coding and decoding compressed sensing framework for remote sensing images was proposed, which includes five parts: mode selection, feature factor extraction, adaptive shape segmentation, adaptive sampling rate allocation and image reconstruction. Unlike previous semi-adaptive or local adaptive methods, the advantages of the adaptive encoding and decoding method proposed in this paper are mainly reflected in four aspects: (1) Ability to select encoding modes based on image content, and maximizing the use of the richness of the image to select appropriate sampling methods; (2) Capable of utilizing image texture details for adaptive segmentation, effectively separating complex and smooth regions; (3) Being able to detect the sparsity of encoding blocks and adaptively allocate sampling rates to fully explore the compressibility of images; (4) The reconstruction matrix can be adaptively selected based on the size of the encoding block to alleviate block artifacts caused by non-stationary characteristics of the image. Experimental results show that the method proposed in this article has good stability for remote sensing images with complex edge textures, with the peak signal-to-noise ratio and structural similarity remaining above 35 dB and 0.8. Moreover, especially for ocean images with relatively simple image content, when the sampling rate is 0.26, the peak signal-to-noise ratio reaches 50.8 dB, and the structural similarity is 0.99. In addition, the recovered images have the smallest BRISQUE value, with better clarity and less distortion. In the subjective aspect, the reconstructed image has clear edge details and good reconstruction effect, while the block effect is effectively suppressed. The framework designed in this paper is superior to similar algorithms in both subjective visual and objective evaluation indexes, which is of great significance for alleviating the incompatibility between traditional information acquisition methods and satellite-borne earth observation missions. Full article
Show Figures

Figure 1

29 pages, 21511 KiB  
Article
Enhancing Surveillance Vision with Multi-Layer Deep Learning Representation
by Dong-Min Son and Sung-Hak Lee
Mathematics 2024, 12(9), 1313; https://doi.org/10.3390/math12091313 - 25 Apr 2024
Cited by 1 | Viewed by 1121
Abstract
This paper aimed to develop a method for generating sand–dust removal and dehazed images utilizing CycleGAN, facilitating object identification on roads under adverse weather conditions such as heavy dust or haze, which severely impair visibility. Initially, the study addressed the scarcity of paired [...] Read more.
This paper aimed to develop a method for generating sand–dust removal and dehazed images utilizing CycleGAN, facilitating object identification on roads under adverse weather conditions such as heavy dust or haze, which severely impair visibility. Initially, the study addressed the scarcity of paired image sets for training by employing unpaired CycleGAN training. The CycleGAN training module incorporates hierarchical single-scale Retinex (SSR) images with varying sigma sizes, facilitating multiple-scaled trainings. Refining the training data into detailed hierarchical layers for virtual paired training enhances the performance of CycleGAN training. Conventional sand–dust removal or dehazing algorithms, alongside deep learning methods, encounter challenges in simultaneously addressing sand–dust removal and dehazing with a singular algorithm. Such algorithms often necessitate resetting hyperparameters to process images from both scenarios. To overcome this limitation, we proposed a unified approach for removing sand–dust and haze phenomena using a single model, leveraging images processed hierarchically with SSR. The image quality and image sharpness metrics of the proposed method were BRIQUE, PIQE, CEIQ, MCMA, LPC-SI, and S3. In sand–dust environments, the proposed method achieved the highest scores, with an average of 21.52 in BRISQUE, 0.724 in MCMA, and 0.968 in LPC-SI compared to conventional methods. For haze images, the proposed method outperformed conventional methods with an average of 3.458 in CEIQ, 0.967 in LPC-SI, and 0.243 in S3. The images generated via this proposed method demonstrated superior performance in image quality and sharpness evaluation compared to conventional algorithms. The outcomes of this study hold particular relevance for camera images utilized in automobiles, especially in the context of self-driving cars or CCTV surveillance systems. Full article
(This article belongs to the Special Issue New Advances and Applications in Image Processing and Computer Vision)
Show Figures

Figure 1

10 pages, 2657 KiB  
Article
Usefulness Evaluation for Nonlocal Means Algorithm in Low-Dose Computed Tomography with Various Iterative Reconstruction Intensities and Kernels: A Pilot Study
by Chaehyeon Song, Yubin Jin, Jina Shim, Seong-Hyeon Kang and Youngjin Lee
Appl. Sci. 2024, 14(8), 3392; https://doi.org/10.3390/app14083392 - 17 Apr 2024
Viewed by 1039
Abstract
The aim of this study was to evaluate the application feasibility of the nonlocal means (NLM) noise reduction algorithm in low-dose computed tomography (CT) images using an advanced modeled iterative reconstruction (ADMIRE) iterative reconstruction technique-based tin filter with various applied parameters. Low-dose CT [...] Read more.
The aim of this study was to evaluate the application feasibility of the nonlocal means (NLM) noise reduction algorithm in low-dose computed tomography (CT) images using an advanced modeled iterative reconstruction (ADMIRE) iterative reconstruction technique-based tin filter with various applied parameters. Low-dose CT images were based on high pitch and tin filters and acquired using slices of the aortic arch, the four chambers of the heart, and the end of the heart. Intensities A2 and A3 as well as kernels B40 and B59 were used as the parameters for the ADMIRE technique. The NLM denoising algorithm was modeled based on the principle of weighting between pixels; the contrast-to-noise ratio (CNR), edge rise distance (ERD), and blind/referenceless image spatial quality evaluator (BRISQUE) were used as image quality evaluation parameters. The CNR result was the highest, with an average of 43.51 in three slices when the proposed NLM denoising algorithm was applied to CT images acquired using the ADMIRE intensity 2 and B59 kernel. The ERD results were similar to those obtained using the ADMIRE intensity 2 and B59 kernel in the CT image acquired using the proposed method. In addition, BRISQUE, which can evaluate the overall image quality, showed a similar trend to the ERD results. In conclusion, the NLM noise reduction algorithm is expected to maximize image quality by preserving efficient edge information while improving noise characteristics in low-dose CT examinations. Full article
(This article belongs to the Special Issue Machine Vision and Machine Learning in Interdisciplinary Research)
Show Figures

Figure 1

17 pages, 6037 KiB  
Article
Evaluation of Mosaic Image Quality and Analysis of Influencing Factors Based on UAVs
by Xiaoyue Du, Liyuan Zheng, Jiangpeng Zhu, Haiyan Cen and Yong He
Drones 2024, 8(4), 143; https://doi.org/10.3390/drones8040143 - 4 Apr 2024
Cited by 1 | Viewed by 2557
Abstract
With the growing prominence of UAV-based low-altitude remote sensing in agriculture, the acquisition and processing of high-quality UAV remote sensing images is paramount. The purpose of this study is to investigate the impact of various parameter settings on image quality and optimize these [...] Read more.
With the growing prominence of UAV-based low-altitude remote sensing in agriculture, the acquisition and processing of high-quality UAV remote sensing images is paramount. The purpose of this study is to investigate the impact of various parameter settings on image quality and optimize these parameters for UAV operations to enhance efficiency and image quality. The study examined the effects of three parameter settings (exposure time, flight altitudes and forward overlap (OF)) on image quality and assessed images obtained under various conditions using signal-to-noise ratio (SNR) and BRISQUE algorithms. The results indicate that the setting of exposure time during UAV image acquisition directly affects image quality, with shorter exposure times resulting in lower SNR. The optimal exposure times for the RGB and MS cameras have been determined as 0.8 ms to 1.1 ms and 4 ms to 16 ms, respectively. Additionally, the best image quality is observed at flight altitudes between 15 and 35 m. The setting of UAV OF complements exposure time and flight altitude; to ensure the completeness of image acquisition, it is suggested that the flight OF is set to approximately 75% at a flight altitude of 25 m. Finally, the proposed image redundancy removal method has been demonstrated as a feasible approach for reducing image mosaicking time (by 84%) and enhancing the quality of stitched images (by 14%). This research has the potential to reduce flight costs, improve image quality, and significantly enhance agricultural production efficiency. Full article
Show Figures

Figure 1

20 pages, 4955 KiB  
Article
Dark-Channel Soft-Constrained and Object-Perception-Enhanced Deep Dehazing Networks Used for Road Inspection Images
by Honglin Wu, Tong Gao, Zhenming Ji, Mou Song, Lianzhen Zhang and Dezhi Kong
Sensors 2023, 23(21), 8932; https://doi.org/10.3390/s23218932 - 2 Nov 2023
Cited by 3 | Viewed by 1302
Abstract
Haze seriously affects the visual quality of road inspection images and contaminates the discrimination of key road objects, which thus hinders the execution of road inspection work. The basic assumptions of the classical dark-channel prior are not suitable for road images containing light-colored [...] Read more.
Haze seriously affects the visual quality of road inspection images and contaminates the discrimination of key road objects, which thus hinders the execution of road inspection work. The basic assumptions of the classical dark-channel prior are not suitable for road images containing light-colored lane lines and vehicles, while typical deep dehazing networks lack physical model interpretability, and they focus on global dehazing effects, neglecting the preservation of object features. For this reason, this paper proposes a Dark-Channel Soft-Constrained and Object-Perception-Enhanced Deep Dehazing Network (DCSC-OPE-Net) for the information recovery of road inspection images. The network is divided into two modules: a dark-channel soft-constrained dehazing module and a near-view object-perception-enhanced module. Unlike the traditional dark-channel algorithms that impose strong constraints on dark pixels, a dark-channel soft-constrained loss function is constructed to ensure that the features of light-colored vehicles and lane lines are effectively maintained. To avoid resolution loss due to patch-based dark-channel processing for image dehazing, a resolution enhancement module is used to strengthen the contrast of the dehazed image. To autonomously perceive and enhance key road features to support road inspection, edge enhancement loss combined with a transmission map is embedded into the network to autonomously discover near-view objects and enhance their key features. The experiments utilize public datasets and real road inspection datasets to validate the performance of the proposed DCSC-OPE-Net compared with typical networks using dehazing evaluation metrics and road object recognition metrics. The experimental results demonstrate that the proposed DCSC-OPE-Net can obtain the best dehazing performance, with an NIQE score of 4.5 and a BRISQUE score of 18.67, and obtain the best road object recognition results (i.e., 83.67%) among the comparison methods. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

Back to TopTop