Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (5)

Search Parameters:
Keywords = salience-aware weighting

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 9691 KiB  
Article
UAV Tracking via Saliency-Aware and Spatial–Temporal Regularization Correlation Filter Learning
by Liqiang Liu, Tiantian Feng, Yanfang Fu, Lingling Yang, Dongmei Cai and Zijian Cao
Symmetry 2024, 16(8), 1076; https://doi.org/10.3390/sym16081076 - 20 Aug 2024
Viewed by 1402
Abstract
Due to their great balance between excellent performance and high efficiency, discriminative correlation filter (DCF) tracking methods for unmanned aerial vehicles (UAVs) have gained much attention. Due to these correlations being capable of being efficiently computed in a Fourier domain by discrete Fourier [...] Read more.
Due to their great balance between excellent performance and high efficiency, discriminative correlation filter (DCF) tracking methods for unmanned aerial vehicles (UAVs) have gained much attention. Due to these correlations being capable of being efficiently computed in a Fourier domain by discrete Fourier transform (DFT), the DFT of an image has symmetry in the Fourier domain. However, DCF tracking methods easily generate unwanted boundary effects where the tracking object suffers from challenging situations, such as deformation, fast motion and occlusion. To tackle the above issue, this work proposes a novel saliency-aware and spatial–temporal regularized correlation filter (SSTCF) model for visual object tracking. First, the introduced spatial–temporal regularization helps build a more robust correlation filter (CF) and improve the temporal continuity and consistency of the model to effectively lower boundary effects and enhance tracking performance. In addition, the relevant objective function can be optimized into three closed-form subproblems which can be addressed by using the alternating direction method of multipliers (ADMM) competently. Furthermore, utilizing a saliency detection method to acquire a saliency-aware weight enables the tracker to adjust to variations in appearance and mitigate disturbances from the surroundings environment. Finally, we conducted numerous experiments based on three different benchmarks, and the results showed that our proposed model had better performance and higher efficiency compared to the most advanced trackers. For example, the distance precision (DP) score was 0.883, and the area under the curve (AUC) score was 0.676 on the OTB2015 dataset. Full article
(This article belongs to the Special Issue Symmetry Applied in Computer Vision, Automation, and Robotics)
Show Figures

Figure 1

13 pages, 839 KiB  
Article
Contextual Patch-NetVLAD: Context-Aware Patch Feature Descriptor and Patch Matching Mechanism for Visual Place Recognition
by Wenyuan Sun, Wentang Chen, Runxiang Huang and Jing Tian
Sensors 2024, 24(3), 855; https://doi.org/10.3390/s24030855 - 28 Jan 2024
Cited by 3 | Viewed by 2480
Abstract
The goal of visual place recognition (VPR) is to determine the location of a query image by identifying its place in a collection of image databases. Visual sensor technologies are crucial for visual place recognition as they allow for precise identification and location [...] Read more.
The goal of visual place recognition (VPR) is to determine the location of a query image by identifying its place in a collection of image databases. Visual sensor technologies are crucial for visual place recognition as they allow for precise identification and location of query images within a database. Global descriptor-based VPR methods face the challenge of accurately capturing the local specific regions within a scene; consequently, it leads to an increasing probability of confusion during localization in such scenarios. To tackle feature extraction and feature matching challenges in VPR, we propose a modified patch-NetVLAD strategy that includes two new modules: a context-aware patch descriptor and a context-aware patch matching mechanism. Firstly, we propose a context-driven patch feature descriptor to overcome the limitations of global and local descriptors in visual place recognition. This descriptor aggregates features from each patch’s surrounding neighborhood. Secondly, we introduce a context-driven feature matching mechanism that utilizes cluster and saliency context-driven weighting rules to assign higher weights to patches that are less similar to densely populated or locally similar regions for improved localization performance. We further incorporate both of these modules into the patch-NetVLAD framework, resulting in a new approach called contextual patch-NetVLAD. Experimental results are provided to show that our proposed approach outperforms other state-of-the-art methods to achieve a Recall@10 score of 99.82 on Pittsburgh30k, 99.82 on FMDataset, and 97.68 on our benchmark dataset. Full article
(This article belongs to the Special Issue Vision Sensors: Image Processing Technologies and Applications)
Show Figures

Figure 1

18 pages, 3915 KiB  
Article
Context-Aware Saliency Guided Radiomics: Application to Prediction of Outcome and HPV-Status from Multi-Center PET/CT Images of Head and Neck Cancer
by Wenbing Lv, Hui Xu, Xu Han, Hao Zhang, Jianhua Ma, Arman Rahmim and Lijun Lu
Cancers 2022, 14(7), 1674; https://doi.org/10.3390/cancers14071674 - 25 Mar 2022
Cited by 14 | Viewed by 2944
Abstract
Purpose: This multi-center study aims to investigate the prognostic value of context-aware saliency-guided radiomics in 18F-FDG PET/CT images of head and neck cancer (HNC). Methods: 806 HNC patients (training vs. validation vs. external testing: 500 vs. 97 vs. 209) from 9 centers [...] Read more.
Purpose: This multi-center study aims to investigate the prognostic value of context-aware saliency-guided radiomics in 18F-FDG PET/CT images of head and neck cancer (HNC). Methods: 806 HNC patients (training vs. validation vs. external testing: 500 vs. 97 vs. 209) from 9 centers were collected from The Cancer Imaging Archive (TCIA). There were 100/384 and 60/123 oropharyngeal carcinoma (OPC) patients with human papillomavirus (HPV) status in training and testing cohorts, respectively. Six types of images were used for radiomics feature extraction and further model construction, namely (i) the original image (Origin), (ii) a context-aware saliency map (SalMap), (iii, iv) high- or low-saliency regions in the original image (highSal or lowSal), (v) a saliency-weighted image (SalxImg), and finally, (vi) a fused PET-CT image (FusedImg). Four outcomes were evaluated, i.e., recurrence-free survival (RFS), metastasis-free survival (MFS), overall survival (OS), and disease-free survival (DFS), respectively. Multivariate Cox analysis and logistic regression were adopted to construct radiomics scores for the prediction of outcome (Rad_Ocm) and HPV-status (Rad_HPV), respectively. Besides, the prognostic value of their integration (Rad_Ocm_HPV) was also investigated. Results: In the external testing cohort, compared with the Origin model, SalMap and SalxImg achieved the highest C-indices for RFS (0.621 vs. 0.559) and MFS (0.785 vs. 0.739) predictions, respectively, while FusedImg performed the best for both OS (0.685 vs. 0.659) and DFS (0.641 vs. 0.582) predictions. In the OPC HPV testing cohort, FusedImg showed higher AUC for HPV-status prediction compared with the Origin model (0.653 vs. 0.484). In the OPC testing cohort, compared with Rad_Ocm or Rad_HPV alone, Rad_Ocm_HPV performed the best for OS and DFS predictions with C-indices of 0.702 (p = 0.002) and 0.684 (p = 0.006), respectively. Conclusion: Saliency-guided radiomics showed enhanced performance for both outcome and HPV-status predictions relative to conventional radiomics. The radiomics-predicted HPV status also showed complementary prognostic value. Full article
Show Figures

Figure 1

20 pages, 5114 KiB  
Article
Tone Mapping of High Dynamic Range Images Combining Co-Occurrence Histogram and Visual Salience Detection
by Ho-Hyoung Choi, Hyun-Soo Kang and Byoung-Ju Yun
Appl. Sci. 2019, 9(21), 4658; https://doi.org/10.3390/app9214658 - 1 Nov 2019
Cited by 5 | Viewed by 3289
Abstract
One of the significant qualities of the human vision, which differentiates it from computer vision, is so called attentional control, which is the innate ability of our human eyes to select what visual stimuli to pay attention to at any moment in time. [...] Read more.
One of the significant qualities of the human vision, which differentiates it from computer vision, is so called attentional control, which is the innate ability of our human eyes to select what visual stimuli to pay attention to at any moment in time. In this sense, the visual salience detection model, which is designed to simulate how the human visual system (HVS) perceives objects and scenes, is widely used for performing multiple vision tasks. This model is also in high demand in the tone mapping technology of high dynamic range images (HDRIs). Another distinct quality of the HVS is that our eyes blink and adjust brightness when objects are in their sight. Likewise, HDR imaging is a technology applied to a camera that takes pictures of an object several times by repeatedly opening and closing a camera iris, which is referred to as multiple exposures. In this way, the computer vision is able to control brightness and depict a range of light intensities. HDRIs are the product of HDR imaging. This article proposes a novel tone mapping method using CCH-based saliency-aware weighting and edge-aware weighting methods to efficiently detect image salience information in the given HDRIs. The two weighting methods combine with a guided filter to generate a modified guided image filter (MGIF). The function of the MGIF is to split an image into the base layer and the detail layer which are the two elements of an image: illumination and reflection, respectively. The base layer is used to obtain global tone mapping and compress the dynamic range of HDRI while preserving the sharp edges of an object in the HDRI. This has a remarkable effect of reducing halos in the resulting HDRIs. The proposed approach in this article also has several distinct advantages of discriminative operation, tolerance to image size variation, and minimized parameter tuning. According to the experimental results, the proposed method has made progress compared to its existing counterparts when it comes to subjective and quantitative qualities, and color reproduction. Full article
(This article belongs to the Section Applied Industrial Technologies)
Show Figures

Graphical abstract

23 pages, 2037 KiB  
Article
Correlation Filter-Based Visual Tracking for UAV with Online Multi-Feature Learning
by Changhong Fu, Fuling Lin, Yiming Li and Guang Chen
Remote Sens. 2019, 11(5), 549; https://doi.org/10.3390/rs11050549 - 6 Mar 2019
Cited by 45 | Viewed by 6325
Abstract
In this paper, a novel online learning-based tracker is presented for the unmanned aerial vehicle (UAV) in different types of tracking applications, such as pedestrian following, automotive chasing, and building inspection. The presented tracker uses novel features, i.e., intensity, color names, and saliency, [...] Read more.
In this paper, a novel online learning-based tracker is presented for the unmanned aerial vehicle (UAV) in different types of tracking applications, such as pedestrian following, automotive chasing, and building inspection. The presented tracker uses novel features, i.e., intensity, color names, and saliency, to respectively represent both the tracking object and its background information in a background-aware correlation filter (BACF) framework instead of only using the histogram of oriented gradient (HOG) feature. In other words, four different voters, which combine the aforementioned four features with the BACF framework, are used to locate the object independently. After obtaining the response maps generated by aforementioned voters, a new strategy is proposed to fuse these response maps effectively. In the proposed response map fusion strategy, the peak-to-sidelobe ratio, which measures the peak strength of the response, is utilized to weight each response, thereby filtering the noise for each response and improving final fusion map. Eventually, the fused response map is used to accurately locate the object. Qualitative and quantitative experiments on 123 challenging UAV image sequences, i.e., UAV123, show that the novel tracking approach, i.e., OMFL tracker, performs favorably against 13 state-of-the-art trackers in terms of accuracy, robustness, and efficiency. In addition, the multi-feature learning approach is able to improve the object tracking performance compared to the tracking method with single-feature learning applied in literature. Full article
(This article belongs to the Special Issue Trends in UAV Remote Sensing Applications)
Show Figures

Graphical abstract

Back to TopTop