Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (520)

Search Parameters:
Keywords = light-field camera

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 2235 KB  
Article
Nighttime Encounter Situation Recognition for Unmanned Surface Vessels Based on Images of Vessel Navigation Lights
by Ruoyun Huang, Xiang Zheng, Jianhua Wang, Gongxing Wu, Yu Tian and Yining Tian
J. Mar. Sci. Eng. 2026, 14(8), 761; https://doi.org/10.3390/jmse14080761 (registering DOI) - 21 Apr 2026
Abstract
To address the limitations of existing perception methods for nighttime encounter situation recognition of unmanned surface vessels (USVs), this study proposes an image-based method for navigation-light recognition and encounter situation recognition. In accordance with the International Regulations for Preventing Collisions at Sea (COLREGs), [...] Read more.
To address the limitations of existing perception methods for nighttime encounter situation recognition of unmanned surface vessels (USVs), this study proposes an image-based method for navigation-light recognition and encounter situation recognition. In accordance with the International Regulations for Preventing Collisions at Sea (COLREGs), a parameterized 3D geometric model of vessel navigation lights and encounter scenario models is established. Based on the camera imaging principle, a dataset of navigation-light images under various encounter situations is generated through simulation experiments. By analyzing the variation patterns of navigation-light images in different encounter situations, a feature vector composed of area-domain and azimuth-domain features is constructed, and an encounter situation recognition method is developed accordingly. To mitigate the effects of water reflections and interfering light sources in real images, a navigation-light image-processing method is designed for the stable extraction of feature parameters. Simulation results show that the classification accuracy ranges from 96.6% to 98.3% at different distance conditions. In field experiments conducted with a small USV under a three-light configuration, the proposed method achieves a navigation-light recognition accuracy of 96.2% and an encounter situation recognition accuracy of 94.94%. The proposed method provides an interpretable and lightweight complementary visual solution for nighttime encounter situation recognition, complementing existing nighttime perception technologies. Full article
(This article belongs to the Section Ocean Engineering)
34 pages, 10998 KB  
Article
Computer Vision over 4G/5G Private Network for Real-Time Draft Assessment to Estimate Vessel Productivity
by Tiago Novaes Mathias, Hugo da Silva Bernardes Gonçalves, Vinícius Veiga Paschoal, Danillo Henrique Araujo de Lima, Júlio César de Oliveira Duque, Arthur Lotenberg, Willian Elias Araújo, Rui Carlos Botter and Daniel de Oliveira Mota
Sensors 2026, 26(8), 2443; https://doi.org/10.3390/s26082443 - 16 Apr 2026
Viewed by 154
Abstract
Background: Draft surveys are widely used to estimate cargo mass during bulk vessel loading and unloading; however, conventional procedures depend on manual draft readings that are episodic, labor-intensive, and sensitive to environmental conditions. Existing camera-based automated approaches rely on draft mark recognition or [...] Read more.
Background: Draft surveys are widely used to estimate cargo mass during bulk vessel loading and unloading; however, conventional procedures depend on manual draft readings that are episodic, labor-intensive, and sensitive to environmental conditions. Existing camera-based automated approaches rely on draft mark recognition or explicit waterline detection, which remain vulnerable to illumination variability, hull fouling, and wave-induced disturbances. Methods: This paper proposes a computer vision framework deployed at the Port of Santos, Brazil, using fixed quay-side cameras and a private 4G network infrastructure for continuous image transmission. Unlike prior methods, the framework estimates emergent hull height by segmenting vessel hull contours from bow and stern viewpoints using customized YOLOv8 instance-segmentation models, without relying on draft marks or waterline detection. Pixel measurements are converted to metric units using a nearby bollard of known height as a local physical reference. Results: Field experiments monitor a Panamax bulk carrier over approximately 6.5 days, processing more than 34,000 images from each camera at an average rate of 3.7 images per minute. Both bow and stern segmentation models achieve mAP50-95 mask scores of 0.980 and 0.965, respectively, confirming precise and stable hull boundary delineation. Hull height decreases from 8.27 m to 4.64 m at the bow and from 7.98 m to 3.98 m at the stern over the loading period, with coherent and temporally stable trends across independent viewpoints. Conclusions: The proposed approach delivers repeatable and continuous hull-height estimates under real operational conditions, including variable lighting, background clutter, and partial occlusions, offering a practical and non-intrusive complement to traditional draft surveys for continuous vessel loading monitoring in modern ports. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

13 pages, 4062 KB  
Article
Robotic Harvesting of Apples Using ROS2
by Connor Ruybalid, Christian Salisbury and Duke M. Bulanon
Machines 2026, 14(4), 433; https://doi.org/10.3390/machines14040433 - 14 Apr 2026
Viewed by 291
Abstract
Rising global food demand, increasing labor costs, and farm labor shortages have created significant challenges for specialty crop production, particularly in labor-intensive tasks such as fruit harvesting. Robotic harvesting offers a promising long-term solution, yet its adoption in orchard environments remains limited due [...] Read more.
Rising global food demand, increasing labor costs, and farm labor shortages have created significant challenges for specialty crop production, particularly in labor-intensive tasks such as fruit harvesting. Robotic harvesting offers a promising long-term solution, yet its adoption in orchard environments remains limited due to unstructured conditions, variable lighting, and difficulties in fruit recognition and manipulation. This study presents an improved robotic fruit harvesting system, Orchard roBot (OrBot), developed by the Robotics Vision Lab at Northwest Nazarene University, with the goal of advancing autonomous apple harvesting applications. The updated OrBot platform integrates a dual-camera vision system consisting of an eye-to-hand stereo camera with a wide field of view for fruit detection and an eye-in-hand RGB-D camera for precise manipulation. The control architecture was redesigned using Robot Operating System 2 (ROS2) and Python, enabling modular subsystem development and coordination. Fruit detection was performed using a YOLOv5 deep learning model, and visual servoing was employed to guide the robotic manipulator toward the target fruit. System performance was evaluated through laboratory experiments using artificial trees and field tests conducted in a commercial apple orchard in Idaho. OrBot achieved a 100% harvesting success rate in indoor tests and a 75–80% success rate in outdoor orchard conditions. Experimental results demonstrate that the dual-camera approach significantly enhances fruit search efficiency and harvesting efficiency. Identified limitations include sensitivity to lighting conditions, end effector performance with varying fruit sizes, and depth estimation errors. Overall, the results indicate a positive potential toward effective robotic fruit harvesting and highlight key areas for future improvement in vision, manipulation, and system robustness. Full article
Show Figures

Figure 1

27 pages, 24387 KB  
Article
Green Pepper Harvesting Robot System Based on Multi-Target Tracking with Filtering and Intelligent Scheduling
by Tianyu Liu, Zelong Liu, Jianmin Wang, Dongxin Guo, Yuxuan Tan and Ping Jiang
Horticulturae 2026, 12(4), 464; https://doi.org/10.3390/horticulturae12040464 - 8 Apr 2026
Viewed by 342
Abstract
To address the challenges of unstable target localization and poor multi-module coordination in automated green pepper harvesting—caused by occlusions from branches and leaves, as well as varying lighting conditions—this paper presents the design and implementation of a modular robotic picking system. At the [...] Read more.
To address the challenges of unstable target localization and poor multi-module coordination in automated green pepper harvesting—caused by occlusions from branches and leaves, as well as varying lighting conditions—this paper presents the design and implementation of a modular robotic picking system. At the perception level, the system integrates a YOLOv8 detector with a RealSense D435i camera to identify and locate the calyx–ectocarp junctions of green peppers. An integrated multi-target tracking and filtering framework is proposed, which fuses multi-feature association, trajectory smoothing and coordinate denoising strategies to suppress depth noise and trajectory jitter, thereby enhancing the stability and accuracy of 3D localization. At the control and execution level, a depth-first picking sequence strategy with ID freeze-state management is implemented within a multithreaded software–hardware co-design architecture. This approach avoids task conflicts and duplicate operations while supporting continuous multi-fruit harvesting. Field experiments under natural outdoor lighting and varying occlusion levels demonstrate that the proposed system achieves recognition rates of 91.57% and 80.29% and harvesting success rates of 82.85% and 77.68% for non-occluded and lightly occluded fruits, respectively. The average picking cycle per pepper fruit is 9.8 s. This system provides an effective technical solution for addressing stability control challenges in the automated harvesting process of green peppers. Full article
(This article belongs to the Section Vegetable Production Systems)
Show Figures

Figure 1

29 pages, 111197 KB  
Article
Deep Learning-Driven Sparse Light Field Enhancement: A CNN-LSTM Framework for Novel View Synthesis and 3D Scene Reconstruction
by Vivek Dwivedi, Gregor Rozinaj, Javlon Tursunov, Ivan Minárik, Marek Vanco and Radoslav Vargic
Mach. Learn. Knowl. Extr. 2026, 8(4), 94; https://doi.org/10.3390/make8040094 - 8 Apr 2026
Viewed by 240
Abstract
Sparse light field imaging often limits the quality of 3D scene reconstruction due to insufficient viewpoint coverage, resulting in incomplete or inaccurate reconstructions. This work introduces a hybrid CNN–LSTM-based framework to address this issue by generating novel camera poses and the corresponding synthesized [...] Read more.
Sparse light field imaging often limits the quality of 3D scene reconstruction due to insufficient viewpoint coverage, resulting in incomplete or inaccurate reconstructions. This work introduces a hybrid CNN–LSTM-based framework to address this issue by generating novel camera poses and the corresponding synthesized novel views, effectively densifying the light field representation. The CNN extracts spatial features from the sparse input views, while the LSTM predicts temporal and positional dependencies, enabling smooth interpolation of novel poses and views. The proposed method integrates these synthesized views with the original sparse dataset to produce a comprehensive set of images. Our approach was evaluated on several datasets, including challenging datasets. The inference capability of our method was tested extensively, and it showed good generalization across diverse datasets. The effectiveness of the framework was evaluated not only with local light field fusion (LLFF) but also with NeRF and 3D Gaussian Splatting, which are considered state-of-the-art reconstruction methods. Overall, the enriched dataset generated by our method led to consistent improvements in 3D reconstruction quality, including higher depth estimation accuracy, reduced artifacts, and enhanced structural consistency. Most importantly, LSTM-based approaches have so far attracted limited attention in the context of generating novel views. While LSTMs have been widely applied in sequential data domains such as natural language processing, their use for image generation conditioned on camera poses remains largely unexplored, which underscores the novelty and significance of the proposed work. This approach provides a scalable and generalizable solution to the sparsity problem in light fields, advancing the capabilities of computational imaging, photorealistic rendering, and immersive 3D scene reconstruction. The results firmly establish the proposed method as a robust and versatile tool for improving reconstruction quality in sparse-view settings. Full article
Show Figures

Figure 1

6 pages, 685 KB  
Proceeding Paper
Contactless Footprint Acquisition and Automated Identification Using Convolutional Neural Network
by Angelica A. Claros, Elmo Joaquin D. Estacion and Jocelyn F. Villaverde
Eng. Proc. 2026, 134(1), 30; https://doi.org/10.3390/engproc2026134030 - 3 Apr 2026
Viewed by 193
Abstract
Biometric systems are widely used in security and forensic applications. Conventionally, contact-based footprint scanners require physical contact, which presents significant limitations. These devices raise hygiene concerns and are impractical in field identification conditions, such as forensic investigations or disaster victim identification, where quick [...] Read more.
Biometric systems are widely used in security and forensic applications. Conventionally, contact-based footprint scanners require physical contact, which presents significant limitations. These devices raise hygiene concerns and are impractical in field identification conditions, such as forensic investigations or disaster victim identification, where quick and non-invasive methods are essential. To address these challenges, a contactless footprint acquisition and identification system was developed using image processing techniques and a Convolutional Neural Network (CNN) based on the Visual Geometry Group–16 layer architecture. The system employs a Raspberry Pi 4, a Logitech C922 camera, and a ring light to capture footprints without direct surface contact. Captured images are processed with Contrast Limited Adaptive Histogram Equalization (CLAHE) to improve contrast and mean thresholding to generate binary images for clearer feature extraction. System performance was evaluated using a multiclass confusion matrix. The CNN correctly classified 158 of 160 test images, achieving an accuracy of 98.75%. This result demonstrates higher accuracy than earlier studies that used older CNN models, such as Alex Krizhevsky’s Network and LeCun’s Network-5, which performed with fewer subjects and lower accuracy rates. The developed system shows potential for biometric security, forensic investigations, and disaster response, where contactless and reliable identification is required. Future research can expand the dataset with more diverse footprints, test performance under varied conditions, and extend the approach to other contactless biometrics such as palmprints or ears. Full article
Show Figures

Figure 1

18 pages, 2314 KB  
Article
Efficient Two-Stage Autofocus for Micro-Assembly Based on Joint Spatial-Frequency Image Quality Assessment
by Jianpeng Zhang, Tianbo Kang, Xin Zhao, Mingzhu Sun and Yi Yang
J. Imaging 2026, 12(3), 137; https://doi.org/10.3390/jimaging12030137 - 19 Mar 2026
Viewed by 345
Abstract
Reliable autofocus is a fundamental prerequisite for precise positioning in micro-assembly systems, where complex reflections, scale variations, and narrow depth-of-field often degrade the robustness of traditional sharpness metrics. To address these challenges, we propose an efficient two-stage autofocus method for a dual-camera micro-vision [...] Read more.
Reliable autofocus is a fundamental prerequisite for precise positioning in micro-assembly systems, where complex reflections, scale variations, and narrow depth-of-field often degrade the robustness of traditional sharpness metrics. To address these challenges, we propose an efficient two-stage autofocus method for a dual-camera micro-vision system based on a spatial-frequency image quality assessment (IQA) model. First, we design WaveMamba-IQA for image sharpness estimation, synergistically combining the Discrete Wavelet Transform with Vision Transformers to capture high-frequency details and semantic features, further enhanced by Multi-Linear Transposed Attention and Vision Mamba for global context modeling. Moreover, we implement a coarse-to-fine autofocus workflow, employing the Covariance Matrix Adaptation Evolution Strategy for global optimization on the horizontal camera, followed by geometric prior-based precise adjustment for the oblique camera. Experimental results on a custom microsphere dataset demonstrate that WaveMamba-IQA achieves a Spearman correlation coefficient of 0.9786. Furthermore, the integrated system achieves a 98.33% autofocus success rate across varying lighting conditions. This method significantly improves the robustness and automation level of micro-assembly systems, effectively overcoming the limitations of manual and traditional focusing techniques. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

21 pages, 1469 KB  
Article
Development of Surveillance Robots Based on Face Recognition Using High-Order Statistical Features and Evidence Theory
by Slim Ben Chaabane, Rafika Harrabi, Anas Bushnag and Hassene Seddik
J. Imaging 2026, 12(3), 107; https://doi.org/10.3390/jimaging12030107 - 28 Feb 2026
Viewed by 592
Abstract
The recent advancements in technologies such as artificial intelligence (AI), computer vision (CV), and Internet of Things (IoT) have significantly extended various fields, particularly in surveillance systems. These innovations enable real-time facial recognition processing, enhancing security and ensuring safety. However, mobile robots are [...] Read more.
The recent advancements in technologies such as artificial intelligence (AI), computer vision (CV), and Internet of Things (IoT) have significantly extended various fields, particularly in surveillance systems. These innovations enable real-time facial recognition processing, enhancing security and ensuring safety. However, mobile robots are commonly employed in surveillance systems to handle risky tasks that are beyond human capability. In this paper, we present a prototype of a cost-effective mobile surveillance robot built on the Raspberry PI 4, designed for integration into various industrial environments. This smart robot detects intruders using IoT and face recognition technology. The proposed system is equipped with a passive infrared (PIR) sensor and a camera for capturing live-streaming video and photos, which are sent to the control room through IoT technology. Additionally, the system uses face recognition algorithms to differentiate between company staff and potential intruders. The face recognition method combines high-order statistical features and evidence theory to improve facial recognition accuracy and robustness. High-order statistical features are used to capture complex patterns in facial images, enhancing discrimination between individuals. Evidence theory is employed to integrate multiple information sources, allowing for better decision-making under uncertainty. This approach effectively addresses challenges such as variations in lighting, facial expressions, and occlusions, resulting in a more reliable and accurate face recognition system. When the system detects an unfamiliar individual, it sends out alert notifications and emails to the control room with the captured picture using IoT. A web interface has also been set up to control the robot from a distance through Wi-Fi connection. The proposed face recognition method is evaluated, and a comparative analysis with existing techniques is conducted. Experimental results with 400 test images of 40 individuals demonstrate the effectiveness of combining various attribute images in improving human face recognition performance. Experimental results indicate that the algorithm can identify human faces with an accuracy of 98.63%. Full article
Show Figures

Figure 1

20 pages, 2259 KB  
Article
A Portable Image-Based Detection Device with Improved Algorithms for Real-Time Droplet Deposition Analysis in Plant Protection UAV Spraying
by Ruizhi Chang, Yu Yan, Guobin Wang, Shengde Chen, Yanhua Meng, Cong Ma and Yubin Lan
Agriculture 2026, 16(5), 499; https://doi.org/10.3390/agriculture16050499 - 25 Feb 2026
Viewed by 448
Abstract
Unmanned aerial vehicles (UAVs) have revolutionized plant protection spraying due to their high efficiency and adaptability. However, the lack of rapid, portable tools for assessing droplet deposition remains a bottleneck for optimizing spray quality and improving pesticide utilization. The main purpose of this [...] Read more.
Unmanned aerial vehicles (UAVs) have revolutionized plant protection spraying due to their high efficiency and adaptability. However, the lack of rapid, portable tools for assessing droplet deposition remains a bottleneck for optimizing spray quality and improving pesticide utilization. The main purpose of this study is to develop a portable, image-based detection device with improved algorithms for real-time analysis (<3 s per card) of droplet deposition on spray cards during UAV plant protection spraying, addressing the limitations of existing methods in portability, real-time capability, and field robustness. This study presents a portable detection device integrated with advanced image processing algorithms for real-time analysis of droplet deposition on copperplate paper cards during UAV operations. The device employs a Raspberry Pi 5 as the core processor, coupled with a high-resolution camera and a standard chessboard calibration board for field-portable image acquisition. Key innovations include an adaptive background subtraction and local contrast enhancement method to address variable field lighting conditions, and an improved adhesion droplet segmentation algorithm combining iterative morphological opening operations with refined distance transform-based concave point matching. Validation on 21 field-collected cards using ImageJ as reference demonstrated a droplet extraction accuracy of 89.4%, with coverage rate improvements of 25.4% and 15.2% compared to OTSU and block thresholding methods, respectively. The adhesion segmentation relative error averaged 6.3%. This low-cost, lightweight device provides farmers and researchers with an effective tool for on-site spray quality evaluation, contributing to precision agriculture and reduced pesticide waste. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

22 pages, 4410 KB  
Article
Accelerating Point Cloud Computation via Memory in Embedded Structured Light Cameras
by Yanan Zhang, Shikang Meng, Shijie Wang and Yaheng Ren
J. Imaging 2026, 12(2), 91; https://doi.org/10.3390/jimaging12020091 - 21 Feb 2026
Viewed by 1434
Abstract
Embedded structured light cameras have been widely applied in various fields. However, due to constraints such as insufficient computing resources, it remains difficult to achieve high-speed structured light point cloud computation. To address this issue, this study proposes a memory-driven computational framework for [...] Read more.
Embedded structured light cameras have been widely applied in various fields. However, due to constraints such as insufficient computing resources, it remains difficult to achieve high-speed structured light point cloud computation. To address this issue, this study proposes a memory-driven computational framework for accelerating point cloud computation. Specifically, the point cloud computation process is precomputed as much as possible and stored in memory in the form of parameters, thereby significantly reducing the computational load during actual point cloud computation. The framework is instantiated in two forms: a low-memory method that minimizes memory footprint at the expense of point cloud stability, and a high-memory method that preserves the nonlinear phase–distance relation via an extensive lookup table. Experimental evaluations demonstrate that the proposed methods achieve comparable accuracy to the conventional method while delivering substantial speedups, and data-format optimizations further reduce required bandwidth. This framework offers a generalizable paradigm for optimizing structured light pipelines, paving the way for enhanced real-time 3D sensing in embedded applications. Full article
Show Figures

Figure 1

23 pages, 4564 KB  
Article
Two-Stage Wildlife Event Classification for Edge Deployment
by Aditya S. Viswanathan, Adis Bock, Zoe Bent, Mark A. Peyton, Daniel M. Tartakovsky and Javier E. Santos
Sensors 2026, 26(4), 1366; https://doi.org/10.3390/s26041366 - 21 Feb 2026
Viewed by 670
Abstract
Camera-based wildlife monitoring is often overwhelmed by non-target triggers and slowed by manual review or cloud-dependent inference, which can prevent timely intervention for high stakes human–wildlife conflicts. Our key contribution is a deployable, fully offline edge vision sensor that achieves near-real-time, highly accurate [...] Read more.
Camera-based wildlife monitoring is often overwhelmed by non-target triggers and slowed by manual review or cloud-dependent inference, which can prevent timely intervention for high stakes human–wildlife conflicts. Our key contribution is a deployable, fully offline edge vision sensor that achieves near-real-time, highly accurate wildlife event classification by combining detector-based empty-image suppression with a lightweight classifier trained with a staged transfer-learning curriculum. Specifically, Stage 1 uses a pretrained You Only Look Once (YOLO)-family detector for permissive animal localization and empty-trigger suppression, and Stage 2 uses a lightweight EfficientNet-based binary classifier to confirm puma on detector crops and gate downstream actions. Our design is robust to low-quality nighttime monochrome imagery (motion blur, low contrast, illumination artifacts, and partial-body captures) and operates using commercially available components in connectivity-limited settings. In field deployments running since May 2025, end-to-end latency from camera trigger to action command is approximately 4 s. Ablation studies using a dataset of labeled wildlife images (pumas, not pumas) show that the two-stage approach substantially reduces false alarms in identifying pumas relative to a full-image classifier while maintaining high recall. On the held-out test set (N=1434 events), the proposed two-stage cascade achieves precision 0.983, recall 0.975, F1 0.979, accuracy 0.986, and balanced accuracy 0.983, with only 8 false positives and 12 false negatives. The system can be easily adapted for other species, as demonstrated by rapid retraining of the second stage to classify ringtails. Downstream responses (e.g., notifications and optional audio/light outputs) provide flexible actuation capabilities that can be configured to support intervention. Full article
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems—2nd Edition)
Show Figures

Figure 1

17 pages, 3436 KB  
Article
Design and Experiment of a Reflective Baffle Based on High-Modulus Carbon Fiber Composite Materials
by Heng Zhang, Xuchao Sun, Junsheng Yang, Yibin Liu, Yue Wang and Weimin Tong
Coatings 2026, 16(2), 234; https://doi.org/10.3390/coatings16020234 - 12 Feb 2026
Viewed by 471
Abstract
A reflective baffle for the optical system of a satellite camera based on the carbon fiber composite materials is designed and validated. Firstly, two typical reflective baffles including elliptical type and Stavroudis type are studied. High modulus carbon fiber composite materials are selected [...] Read more.
A reflective baffle for the optical system of a satellite camera based on the carbon fiber composite materials is designed and validated. Firstly, two typical reflective baffles including elliptical type and Stavroudis type are studied. High modulus carbon fiber composite materials are selected to achieve lightweight and high rigidity. The aluminum film is coated on the surface of vanes to enhance the surface spectral reflectivity. Then, temperature field under typical external heat flow is calculated and stray light suppression characteristics are analyzed. Finally, the finite element simulation and mechanical vibration experiment are performed to verify the reliability of the baffle structure. The results show that the reflective baffle meets the requirements of mechanical environment during the launch phase of satellite camera. It provides a reference for the design of the satellite camera baffles structure. Full article
(This article belongs to the Section Environmental Aspects in Colloid and Interface Science)
Show Figures

Figure 1

21 pages, 4711 KB  
Article
An Integrated Framework for Pavement Crack Segmentation and Severity Estimation
by Osama Alsharayah, Dmitry Manasreh and Munir D. Nazzal
Buildings 2026, 16(3), 677; https://doi.org/10.3390/buildings16030677 - 6 Feb 2026
Viewed by 409
Abstract
Pavement maintenance programs rely on timely and accurate crack assessment to preserve roadway quality and reduce long-term rehabilitation costs. Manual inspection remains the prevailing practice, yet it is slow, subjective, and exposes crews to safety risks. Automating crack detection under real-world roadway conditions [...] Read more.
Pavement maintenance programs rely on timely and accurate crack assessment to preserve roadway quality and reduce long-term rehabilitation costs. Manual inspection remains the prevailing practice, yet it is slow, subjective, and exposes crews to safety risks. Automating crack detection under real-world roadway conditions remains challenging due to inconsistent lighting, shadows, stains, and surface textures that obscure distress features. This study examines the applicability of an integrated, vehicle-mounted framework for automated pavement crack segmentation and width-based severity estimation under practical roadway operating conditions. Data were collected from a moving vehicle using a custom camera–GPS system operating under diverse conditions, capturing the variability encountered in practical surveys. The proposed approach employs a state-of-the-art segmentation model and a calibrated width estimation tool that converts pixel-level crack measurements into physical units using a position-dependent regression model. The key contribution of this work is a unified segmentation and severity evaluation pipeline supported by a novel pixel-to-inch calibration surface and validated using images acquired during normal driving operations and manual field crack measurements. By combining advanced computer vision techniques with practical field-oriented data collection, the proposed system provides a deployable solution for roadway crack assessment, enabling safer, faster, and more scalable network-level pavement monitoring. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

25 pages, 2789 KB  
Article
Hybrid Zero-Shot Node-Count Estimation and Growth-Information Sharing for Lisianthus (Eustoma grandiflorum) Cultivation in Fukushima’s Floricultural Revitalization
by Hiroki Naito, Kota Kobayashi, Osamu Inaba, Fumiki Hosoi, Norihiro Hoshi and Yoshimichi Yamashita
Agriculture 2026, 16(3), 296; https://doi.org/10.3390/agriculture16030296 - 23 Jan 2026
Viewed by 545
Abstract
This paper presents a hybrid pipeline based on zero-shot vision models for automatic node count estimation in Lisianthus (Eustoma grandiflorum) cultivation and a system for real-time growth information sharing. The multistage image analysis pipeline integrates Grounding DINO for zero-shot leaf-region detection, [...] Read more.
This paper presents a hybrid pipeline based on zero-shot vision models for automatic node count estimation in Lisianthus (Eustoma grandiflorum) cultivation and a system for real-time growth information sharing. The multistage image analysis pipeline integrates Grounding DINO for zero-shot leaf-region detection, MiDaS for monocular depth estimation, and a YOLO-based classifier, using daily time-lapse images from low-cost fixed cameras in commercial greenhouses. The model parameters are derived from field measurements of 2024 seasonal crops (Trial 1) and then applied to different cropping seasons, growers, and cultivars (Trials 2 and 3) without any additional retraining. Trial 1 indicates high accuracy (R2 = 0.930, mean absolute error (MAE) = 0.73). Generalization performance is confirmed in Trials 2 (MAE = 0.45) and 3 (MAE = 1.14); reproducibility across multiple growers and four cultivars yields MAEs of approximately ±1 node. The model effectively captures the growth progression despite variations in lighting, plant architecture, and grower practices, although errors increase during early growth stages and under unstable leaf detection. Furthermore, an automated Discord-based notification system enables real-time sharing of node trends and analytical images, facilitating communication. The feasibility of combining zero-shot vision models with cloud-based communication tools for sustainable and collaborative floricultural production is thus demonstrated. Full article
Show Figures

Figure 1

11 pages, 4363 KB  
Article
Testing and Characterization of Detection Plane Elements of the XGIS Instrument on Board the THESEUS Mission
by Smiriti Srivastava, Evgeny Demenev, Claudio Labanti, Lorenzo Amati, Riccardo Campana, Giuseppe Baldazzi, Edoardo Borciani, Paolo Calabretto, Francesco Ficorella, Ezequiel J. Marchesini, Giulia Mattioli, Ajay Sharma, David Novel, Giancarlo Pepponi and Enrico Virgilli
Particles 2026, 9(1), 7; https://doi.org/10.3390/particles9010007 - 18 Jan 2026
Viewed by 493
Abstract
This paper presents the procedures employed for experimental functional and performance characterization of a 2 × 2 pixel prototype detection system tailored specifically for the X and Gamma-ray Imaging Spectrometer (XGIS) instrument onboard the THESEUS mission. The XGIS system comprises of two coded [...] Read more.
This paper presents the procedures employed for experimental functional and performance characterization of a 2 × 2 pixel prototype detection system tailored specifically for the X and Gamma-ray Imaging Spectrometer (XGIS) instrument onboard the THESEUS mission. The XGIS system comprises of two coded masked wide field cameras integrated with monolithic SDDs (Silicon Drift Detectors) and CsI:Tl (Thallium doped-Cesium Iodide) scintillators, contributing to its broad X and γ-ray detection range. Given the space instrumentation complexity, thorough requirement qualification and testing procedures are essential. This work focuses on working principle, the testing setup utilized, and observed performance for the small scale four-pixel XGIS prototype. Furthermore, the alignment of light output performance of the four-pixel SDD and scintillator prototype detection system with the XGIS instrument requirements is emphasized. Full article
Show Figures

Figure 1

Back to TopTop