You are currently viewing a new version of our website. To view the old version click .

Journal of Imaging

Journal of Imaging is an international, multi/interdisciplinary, peer-reviewed, open access journal of imaging techniques, published online monthly by MDPI.

Indexed in PubMed | Quartile Ranking JCR - Q2 (Imaging Science and Photographic Technology)

All Articles (2,189)

  • Feature Paper
  • Article
  • Open Access

Accurate Segmentation of Vegetation in UAV Desert Imagery Using HSV-GLCM Features and SVM Classification

  • Thani Jintasuttisak,
  • Patompong Chabplan and
  • Sasitorn Issaro
  • + 2 authors

Segmentation of vegetation from images is an important task in precision agriculture applications, particularly in challenging desert environments where sparse vegetation, varying soil colors, and strong shadows pose significant difficulties. In this paper, we present a machine learning approach to robust green-vegetation segmentation in drone imagery captured over desert farmlands. The proposed method combines HSV color-space representation with Gray-Level Co-occurrence Matrix (GLCM) texture features and employs Support Vector Machine (SVM) as the learning algorithm. To enhance robustness, we incorporate comprehensive preprocessing, including Gaussian filtering, illumination normalization, and bilateral filtering, followed by morphological post-processing to improve segmentation quality. The method is evaluated against both traditional spectral index methods (ExG and CIVE) and a modern deep learning baseline using comprehensive metrics including accuracy, precision, recall, F1-score, and Intersection over Union (IoU). Experimental results on 120 high-resolution drone images from UAE desert farmlands demonstrate that the proposed method achieves superior performance with an accuracy of 0.91, F1-score of 0.88, and IoU of 0.82, showing significant improvement over baseline methods in handling challenging desert conditions, including shadows, varying soil colors, and sparse vegetation patterns. The method provides practical computational performance with a processing time of 25 s per image and a training time of 28 min, making it suitable for agricultural applications where accuracy is prioritized over processing speed.

25 December 2025

The workflow of the proposed green vegetation segmentation method.

Accurate six-degree-of-freedom (6-DoF) camera pose estimation is essential for augmented reality, robotics navigation, and indoor mapping. Existing pipelines often depend on detailed floorplans, strict Manhattan-world priors, and dense structural annotations, which lead to failures in ambiguous room layouts where multiple rooms appear in a query image and their boundaries may overlap or be partially occluded. We present Render-Rank-Refine, a two-stage framework operating on coarse semantic meshes without requiring textured models or per-scene fine-tuning. First, panoramas rendered from the mesh enable global retrieval of coarse pose hypotheses. Then, perspective views from the top-k candidates are compared to the query via rotation-invariant circular descriptors, which re-ranks the matches before final translation and rotation refinement. Our method increases camera localization accuracy compared to the state-of-the-art SPVLoc baseline by reducing the translation error by 40.4% and the rotation error by 29.7% in ambiguous layouts, as evaluated on the Zillow Indoor Dataset. In terms of inference throughput, our method achieves 25.8–26.4 QPS, (Queries Per Second) which is significantly faster than other recent comparable methods, while maintaining accuracy comparable to or better than the SPVLoc baseline. These results demonstrate robust, near-real-time indoor localization that overcomes structural ambiguities and heavy geometric assumptions.

25 December 2025

Examples from ZInD [12]. (a–c) represent visually clear layouts; (d–f) represent ambiguous or challenging cases.

The dynamic characteristics of flotation foam, such as velocity and breakage rate, are critical factors that influence mineral separation efficiency. However, challenges inherent in foam images, including weak textures, severe deformations, and motion blur, present significant technical hurdles for dynamic monitoring. These issues lead to a fundamental conflict between the efficiency and accuracy of traditional feature matching algorithms. This paper introduces a novel progressive framework for dynamic feature matching in flotation foam images, termed “stable extraction, efficient coarse screening, and precise matching.” This framework first employs the Accelerated-KAZE (AKAZE) algorithm to extract robust, scale- and rotation-invariant feature points from a non-linear scale-space, effectively addressing the challenge of weak textures. Subsequently, it innovatively incorporates the Grid-based Motion Statistics (GMS) algorithm to perform efficient coarse screening based on motion consistency, rapidly filtering out a large number of obvious mismatches. Finally, the Progressive Sample and Consensus (PROSAC) algorithm is used for precise matching, eliminating the remaining subtle mismatches through progressive sampling and geometric constraints. This framework enables the precise analysis of dynamic foam characteristics, including displacement, velocity, and breakage rate (enhanced by a robust “foam lifetime” mechanism). Comparative experimental results demonstrate that, compared to ORB-GMS-RANSAC (with a Mean Absolute Error, MAE of 1.20 pixels and a Mean Relative Error, MRE of 9.10%) and ORB-RANSAC (MAE: 3.53 pixels, MRE: 27.36%), the proposed framework achieves significantly lower error rates (MAE: 0.23 pixels, MRE: 2.13%). It exhibits exceptional stability and accuracy, particularly in complex scenarios involving low texture and minor displacements. This research provides a high-precision, high-robustness technical solution for the dynamic monitoring and intelligent control of the flotation process.

25 December 2025

The algorithm principle block diagram.

To address the decline in self-consistency and limited spatial adaptability of traditional interpolation methods in complex terrain, this study proposes a terrain-constrained Triangulated Irregular Network (TIN) interpolation method based on UAV point clouds. The method was tested in the southern margin of the Lufeng Dinosaur National Geopark, Yunnan Province, using ground points at different sampling densities (90%, 70%, 50%, 30%, and 10%), and compared with Spline, Kriging, ANUDEM, and IDW methods. Results show that the proposed method maintains the lowest RMSE and MAE across all densities, demonstrating higher stability and self-consistency and better preserving terrain undulations. This provides technical support for high-precision DEM reconstruction from UAV point clouds in complex terrain.

25 December 2025

DEM and Hillshade of the Study Area: (a) Sample 1; (b) Sample 2; (c) Sample 3.

News & Conferences

Issues

Open for Submission

Editor's Choice

Reprints of Collections

Computational Intelligence in Remote Sensing
Reprint

Computational Intelligence in Remote Sensing

2nd Edition
Editors: Yue Wu, Kai Qin, Maoguo Gong, Qiguang Miao

Get Alerted

Add your email address to receive forthcoming issues of this journal.

XFacebookLinkedIn
J. Imaging - ISSN 2313-433X