Due to scheduled maintenance work on our servers, there may be short service disruptions on this website between 11:00 and 12:00 CEST on March 28th.
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (560)

Search Parameters:
Keywords = camera configurations

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 4289 KB  
Article
Floor Plan Generation of Existing Buildings Based on Deep Learning and Stereo Vision
by Dejiang Wang and Taoyu Peng
Buildings 2026, 16(7), 1310; https://doi.org/10.3390/buildings16071310 - 26 Mar 2026
Abstract
The reinforcement and renovation of existing buildings constitute an important component of the future development of the civil engineering industry. Such projects typically require the original construction drawings of the building. However, for older structures, the original paper-based drawings may be damaged or [...] Read more.
The reinforcement and renovation of existing buildings constitute an important component of the future development of the civil engineering industry. Such projects typically require the original construction drawings of the building. However, for older structures, the original paper-based drawings may be damaged or lost. Moreover, traditional manual surveying and mapping methods are time-consuming, labor-intensive, and limited in accuracy. To address these issues, this paper proposes a floor plan generation method for existing buildings that integrates deep learning and stereo vision based on a fusion of synthetic and real data. First, collaborative modeling and automated rendering between a large language model and Blender are implemented based on the Model Context Protocol (MCP), enabling indoor scene modeling and image acquisition to construct a synthetic dataset containing structural components such as doors, windows, and walls. Meanwhile, manually annotated real indoor images are incorporated. Synthetic and real data are mixed in different proportions to form multiple dataset configurations for model training and validation. Subsequently, the SegFormer model is employed to perform semantic segmentation of indoor components. Combined with stereo camera calibration results, disparity computation is conducted to extract the three-dimensional spatial coordinates of component corner points. On this basis, the architectural floor plan is generated according to the spatial geometric relationships among structural components. Experimental results demonstrate that the proposed method effectively reduces the need for manual annotation and on-site measurement, providing an efficient technical solution for indoor floor plan generation of existing buildings. Full article
(This article belongs to the Topic Application of Smart Technologies in Buildings)
Show Figures

Figure 1

26 pages, 16104 KB  
Article
Multi-Slot Attention with State Guidance for Egocentric Robotic Manipulation
by Sofanit Wubeshet Beyene and Ji-Hyeong Han
Electronics 2026, 15(7), 1365; https://doi.org/10.3390/electronics15071365 - 25 Mar 2026
Abstract
Visual perception is fundamental to robotic manipulation for recognizing objects, goals, and contextual details. Third-person cameras provide global views but can miss contact-rich interactions and require calibration. Wrist-mounted egocentric cameras reduce these limitations but introduce occlusion, motion blur, and partial observability, which complicate [...] Read more.
Visual perception is fundamental to robotic manipulation for recognizing objects, goals, and contextual details. Third-person cameras provide global views but can miss contact-rich interactions and require calibration. Wrist-mounted egocentric cameras reduce these limitations but introduce occlusion, motion blur, and partial observability, which complicate visuomotor learning. Furthermore, existing perception modules that rely solely on pixels or fuse imagery with proprioception as flat vectors do not explicitly model structured scene representations in dynamic egocentric views. To address these challenges, a multi-slot attention fusion encoder for egocentric manipulation is introduced. Learnable slot queries extract localized visual features from image tokens, and Feature-wise Linear Modulation (FiLM) conditions each slot on the robot’s joint states, producing a structured slot-based latent representation that adapts to viewpoint and configuration changes without requiring object labels or external camera priors. The resulting structured slot-based latent representation is used as input to a Soft Actor–Critic (SAC) agent, which achieves a higher mean cumulative return than pixel-only CNN/DrQ and state-only baselines on a ManiSkill3 egocentric manipulation task. Probing experiments and real-camera evaluation further show that the learned representation remains stable under egocentric viewpoint shifts and partial occlusions, indicating robustness in practical manipulation settings. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

27 pages, 8346 KB  
Article
VNIR/SWIR Multispectral Polarimetric Imager for Polymer Discrimination and Identification
by Ramon Prats Consola and Adriano Camps
Sensors 2026, 26(7), 2040; https://doi.org/10.3390/s26072040 (registering DOI) - 25 Mar 2026
Abstract
This work presents a portable polarimetric multispectral imaging (PMSI) system operating in the visible to shortwave infrared range (VNIR–SWIR: 400–1700 nm) and its application to target detection, discrimination from aquatic backgrounds, and polymer identification. The instrument integrates two synchronized cameras with motorized bandpass [...] Read more.
This work presents a portable polarimetric multispectral imaging (PMSI) system operating in the visible to shortwave infrared range (VNIR–SWIR: 400–1700 nm) and its application to target detection, discrimination from aquatic backgrounds, and polymer identification. The instrument integrates two synchronized cameras with motorized bandpass filters and piezoelectric polarization control, enabling the acquisition of 48 wavelength–polarization measurements per capture. This configuration allows the extraction of both intensity-based and polarimetric features, including the degree of linear polarization (DoLP). A complete radiometric and polarimetric calibration framework is implemented, encompassing system response characterization, polarization-dependent gain correction, and reflectance normalization under variable illumination. Experiments conducted on a representative set of 16 polymer materials show that polarimetric information consistently improves class separability compared to intensity-only features, with a mean gain of 6.9 (95% CI: 6.35–8.47). Although the correlation between intensity- and DoLP-based separability is moderate (r = 0.44), the results indicate complementary identification capability. Material recoverability was further evaluated using spectral unmixing techniques (VCA, N-FINDR, and PPI), with VCA offering the best accuracy–complexity trade-off on the calibrated Stokes reflectance dataset. Despite these gains, identification among chemically similar polyethylene variants remains challenging due to limited spectral and polarimetric contrast. An underwater detectability study under natural illumination reveals strong wavelength-dependent constraints: SWIR penetration is limited to 4 cm, whereas VNIR bands (430–550 nm) preserve detectability up to 20 cm, with DoLP enhancing edge visibility. These results motivate future validation in more complex aquatic conditions and with increased spectral dimensionality. Full article
(This article belongs to the Special Issue Hyperspectral Imaging for Environmental Monitoring)
Show Figures

Figure 1

25 pages, 233246 KB  
Article
Seamlessly Natural: Image Stitching with Natural Appearance Preservation
by Gaetane Lorna N. Tchana, Damaris Belle M. Fotso, Antonio Hendricks and Christophe Bobda
Technologies 2026, 14(3), 186; https://doi.org/10.3390/technologies14030186 - 19 Mar 2026
Viewed by 140
Abstract
Conventional image stitching pipelines predominantly rely on homographic alignment, whose planar assumption often breaks down in dual-camera configurations capturing non-planar scenes, producing geometric warping, bulging, and structural distortion. To address these limitations, this paper presents SENA (Seamlessly Natural), a geometry-driven image stitching approach [...] Read more.
Conventional image stitching pipelines predominantly rely on homographic alignment, whose planar assumption often breaks down in dual-camera configurations capturing non-planar scenes, producing geometric warping, bulging, and structural distortion. To address these limitations, this paper presents SENA (Seamlessly Natural), a geometry-driven image stitching approach with three complementary contributions. First, we propose a hierarchical affine-based warping strategy that combines global affine initialization, local affine refinement, and a smooth free-form deformation field regulated by seamguard adaptive smoothing. This multi-scale design preserves local shape, parallelism, and aspect ratios, thereby reducing the hallucinated distortions commonly associated with homography-based models. Second, SENA incorporates a geometry-driven adequate zone detection mechanism that identifies regions with reduced parallax directly from the disparity consistency of correspondences filtered by RANSAC, without relying on semantic segmentation or depth estimation. Third, within this zone, anchor-based seamline cutting and segmentation enforce one-to-one geometric correspondence between image pairs, reducing ghosting and smearing artifacts. Extensive experiments demonstrate that SENA achieves 26.2 dB PSNR and 0.84 SSIM, obtains the lowest BRISQUE score (33.4) among compared methods, and reduces runtime by 79% on average across resolutions. These results confirm improved structural fidelity and computational efficiency while maintaining competitive alignment accuracy. Full article
(This article belongs to the Special Issue Image Analysis and Processing)
Show Figures

Figure 1

19 pages, 37608 KB  
Article
ZoomPatch: An Adaptive PTZ Scheduling Framework for Small Object Video Analytics
by Shutong Chen, Binhua Liang and Yan Chen
Appl. Sci. 2026, 16(6), 2934; https://doi.org/10.3390/app16062934 - 18 Mar 2026
Viewed by 108
Abstract
Accurate detection of small objects in video analytics is limited by low pixel resolution and insufficient visual cues. While software-based enhancements often fail to recover missing details, Pan–Tilt–Zoom (PTZ) cameras can physically increase spatial resolution through optical zoom. However, mechanical latency and configuration [...] Read more.
Accurate detection of small objects in video analytics is limited by low pixel resolution and insufficient visual cues. While software-based enhancements often fail to recover missing details, Pan–Tilt–Zoom (PTZ) cameras can physically increase spatial resolution through optical zoom. However, mechanical latency and configuration complexity hinder their real-time applicability. We propose ZoomPatch, a real-time video analytics framework tailored for small object detection. ZoomPatch actively schedules PTZ adjustments to capture optically enhanced subframes of regions of interest (ROIs) and fuses inference results back to the global reference frame. Specifically, it introduces a dynamic Cycle Length Proposer to adapt analysis cycles based on scene motion, and a Mixed Integer Linear Programming (MILP)-based Configuration Decider to determine the optimal sequence of pan, tilt, and zoom adjustments under time budget constraints. Simulation-based experimental evaluations across diverse workloads demonstrate that ZoomPatch significantly outperforms fixed-perspective, super-resolution (SR), and greedy baselines. Notably, in the detection task using YOLOv10, ZoomPatch improves the F1-score from 0.33 to 0.47 (a 42% increase) compared to the fixed-perspective baseline. Furthermore, ZoomPatch yields performance gains of 30% and 7% over the SR baseline (0.36) and the greedy baseline (0.44). Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

23 pages, 2962 KB  
Article
Feasibility of Infrared-Based Pedestrian Detectability in Unlit Urban and Rural Road Sections Using Consumer Thermal Cameras
by Yordan Stoyanov, Atanasi Tashev and Penko Mitev
Vehicles 2026, 8(3), 61; https://doi.org/10.3390/vehicles8030061 - 16 Mar 2026
Viewed by 202
Abstract
This study evaluates the feasibility of using two affordable thermal cameras (UNI-T UTi260M and UTi260T), which are not designed as automotive sensors, for observing pedestrians and warm objects during night-time driving under low-illumination conditions. The experimental setup includes mounting the camera on the [...] Read more.
This study evaluates the feasibility of using two affordable thermal cameras (UNI-T UTi260M and UTi260T), which are not designed as automotive sensors, for observing pedestrians and warm objects during night-time driving under low-illumination conditions. The experimental setup includes mounting the camera on the vehicle body (e.g., side mirror area/roof), recording road scenes in urban and rural environments, and selecting representative frames for qualitative and quantitative analysis. The study assesses: (i) observable pedestrian detectability in unlit road sections and under oncoming headlight glare, where visible cameras often lose contrast; (ii) the influence of low ambient temperature and strong cold wind on image appearance (including “whitening”/contrast shifts); and (iii) workflow differences, where UTi260M relies on a smartphone application for streaming/recording, while UTi260T supports PC-based image analysis and temperature-profile visualization. In addition, a calibration-based geometric method is proposed for approximate pedestrian distance estimation from single frames using silhouette pixel height and a regression model based on 1/hpx, valid for a specific mounting configuration and a known subject height. Results indicate that both cameras can highlight warm objects relative to the background and support visual pedestrian identification at low illumination, including in the presence of oncoming headlights, with UTi260M showing more stable behavior in parts of the tests. This work is a feasibility study and does not claim Advanced Driver Assist Systems (ADAS) functionality; it outlines limitations, repeatability considerations, and a minimal set of metrics and procedures for future extension. All quantitative indicators derived from exported frames are explicitly treated as image-level proxy metrics, not as physical sensor characteristics. Full article
(This article belongs to the Special Issue Novel Solutions for Transportation Safety, 2nd Edition)
Show Figures

Figure 1

20 pages, 13437 KB  
Article
Motion Prediction of Moored Platform Using CNN–LSTM for Eco-Friendly Operation
by Omar Jebari, Chungkuk Jin, Byungho Kang, Seong Hyeon Hong, Changhee Lee and Young Hun Jeon
J. Mar. Sci. Eng. 2026, 14(6), 531; https://doi.org/10.3390/jmse14060531 - 12 Mar 2026
Viewed by 184
Abstract
Predicting the motion of ships and floating structures is essential for ensuring economical and environmentally friendly operations in the ocean. In this study, we propose a hybrid encoder–decoder Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM) architecture to predict motions of a moored Floating Production [...] Read more.
Predicting the motion of ships and floating structures is essential for ensuring economical and environmentally friendly operations in the ocean. In this study, we propose a hybrid encoder–decoder Convolutional Neural Network–Long Short-Term Memory (CNN–LSTM) architecture to predict motions of a moored Floating Production Storage and Offloading (FPSO) vessel under varying sea conditions. The model integrates a CNN for spatial wave-field feature extraction and an LSTM encoder–decoder to capture temporal dependencies in vessel motion. Synthetic datasets were generated using mid-fidelity dynamics simulations of a coupled FPSO–mooring–riser system subjected to wave excitations. Five sea states ranging from calm to severe were considered to evaluate the model’s robustness. A key preprocessing step involved determining the optimal spatial domain for wave field input, and a wave field size of 600 m × 600 m was identified as the most cost-effective configuration while maintaining accuracy. The model was validated using the Root Mean Square Error (RMSE) or relative RMSE (RRMSE). Despite low RRMSE values in low sea states, predictions were noisier due to high-frequency, low-amplitude responses. In contrast, higher sea states yielded more stable predictions despite higher RRMSE values. The proposed method offers high-resolution motion forecasting capability, which can enhance operational safety and energy efficiency of offshore platforms, particularly when integrated with stereo camera-based wave monitoring systems. Full article
(This article belongs to the Special Issue Intelligent Solutions for Marine Operations)
Show Figures

Figure 1

18 pages, 2558 KB  
Article
Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams
by Zhaoyu Li, Zhenbin Tan, Wen Zheng, Ganling Yang, Junye Tao, Mingxin Zhang and Xiao Xu
Sensors 2026, 26(5), 1689; https://doi.org/10.3390/s26051689 - 7 Mar 2026
Viewed by 396
Abstract
Markerless motion capture (MMC) provides a non-invasive alternative for motion analysis; however, its validity at the standard frame rate of 25 Hz commonly used in broadcast and surveillance applications remains to be established. This study evaluated the performance of a 25 Hz multi-camera [...] Read more.
Markerless motion capture (MMC) provides a non-invasive alternative for motion analysis; however, its validity at the standard frame rate of 25 Hz commonly used in broadcast and surveillance applications remains to be established. This study evaluated the performance of a 25 Hz multi-camera MMC workflow using consumer-grade cameras for capturing basketball-specific movements. Three highly trained male athletes completed seven tasks, including sprinting and simulated sport-specific skills, while being synchronously recorded by six MMC cameras (DJI Action 5 Pro, 25 fps) and a 10-camera Vicon system (25 Hz). Kinematic data were processed using an RTMDet–RTMPose pipeline and low-pass filtered at 6 Hz. Waveform validity was assessed using Pearson’s correlation coefficient (r) and the root mean square error (RMSE). The displacement magnitudes of 12 joints showed excellent agreement (r = 0.916–0.994; median nRMSE = 0.54–1.32%), indicating robust trajectory reconstruction. In contrast, agreement decreased for derivative variables: velocity (r = 0.583–0.867) and acceleration (r = 0.232–0.677) were highly sensitive to the low sampling rate and numerical differentiation. Although a 25 Hz configuration is insufficient for high-precision impact analysis, it provides acceptable accuracy for macroscopic displacement tracking and external-load quantification in resource-constrained training environments. Future optimization should prioritize temporal synchronization to improve the reliability of derivative variables. Full article
(This article belongs to the Special Issue Multi-Sensor Systems for Object Tracking—2nd Edition)
Show Figures

Figure 1

14 pages, 2347 KB  
Article
Posture Tracking of Active Capsule Endoscopes Integrated with Magnetic Actuation Using Hall-Effect Sensors
by Junho Han, Kim Tien Nguyen, Eui-Sun Kim, Jong-Oh Park, Eunho Choe, Chang-bae Moon and Jayoung Kim
Micromachines 2026, 17(3), 327; https://doi.org/10.3390/mi17030327 - 5 Mar 2026
Viewed by 300
Abstract
A capsule endoscope (CE) provides noninvasive access to the gastrointestinal tract, offering diagnostic information that cannot be obtained through external imaging alone. However, during the examination inside the stomach, the CE’s posture may change rapidly as it moves within a dynamically deforming organ, [...] Read more.
A capsule endoscope (CE) provides noninvasive access to the gastrointestinal tract, offering diagnostic information that cannot be obtained through external imaging alone. However, during the examination inside the stomach, the CE’s posture may change rapidly as it moves within a dynamically deforming organ, making it difficult to determine its orientation using only the onboard camera feedback. To address this problem, this study proposes a method that employs an external array of Hall Effect Sensors (HES) to estimate the capsule’s position and orientation in real time, based on the magnetic field generated by a permanent magnet (PM) embedded inside the capsule, without the need for any additional internal sensors. This approach introduces a unified magnetic actuation and localization framework that enables real-time 5-degree-of-freedom posture estimation using only the internal PM of the capsule. Furthermore, the proposed system features an integrated architecture capable of simultaneous actuation and localization. To enhance system practicality, the sensor module and communication board were combined into a single unit that employs a digital serial communication scheme, eliminating the need for analog to digital conversion of sensing signals. By avoiding additional onboard sensors and employing a PM-based actuation system, the proposed system simplifies hardware configuration by preserving capsule miniaturization and by eliminating the high power consumption and thermal issues associated with electromagnet-based actuation, while maintaining accurate real-time tracking performance. Through an optimization process, the system achieved a position error of less than 2 mm and an angular error within 2° over a sensing range of up to 60 mm. Repeated experiments further validated the system’s effectiveness and reliability under realistic operating conditions, demonstrating its feasibility for compact and clinically applicable active capsule endoscopy systems. Full article
(This article belongs to the Section E:Engineering and Technology)
Show Figures

Figure 1

14 pages, 5168 KB  
Article
The Concept of a Digital Twin in the Arctic Environment
by Ari Pikkarainen, Timo Sukuvaara, Kari Mäenpää, Hannu Honkanen and Pyry Myllymäki
Electronics 2026, 15(5), 1001; https://doi.org/10.3390/electronics15051001 - 28 Feb 2026
Viewed by 227
Abstract
A Digital Twin is a virtual environment that simulates, predicts, and optimizes the performance of its physical counterpart. Digital Twin models hold great potential in wireless networking testing and development. This paper aims to envision our concept of simulating the operation of different [...] Read more.
A Digital Twin is a virtual environment that simulates, predicts, and optimizes the performance of its physical counterpart. Digital Twin models hold great potential in wireless networking testing and development. This paper aims to envision our concept of simulating the operation of different sensors in vehicle test-track conditions. Vehicle parameters are embedded into the edge computing entity, which uses them to generate a test configuration for the Digital Twin. This configuration is then applied in simulated sensor-output prediction, ultimately producing event data for the vehicle entity. The sensor suite—comprising radar, cameras, GPS and LiDAR—is modeled to provide the multi-modal input required for generating simulated perception data in the Digital Twin. To ensure realistic perception behavior, the physical vehicle is represented within a digital environment that reproduces the actual test track. This allows LiDAR occlusions to be attributed to genuine environmental structures (e.g., trees, buildings, other vehicles) rather than simulation artifacts. Within the Digital Twin, the objective is to evaluate how sensor signals—such as radar waves and LiDAR light pulses—propagate through the environment and how real-world obstacles may weaken or distort them. Historical datasets are used to calibrate and validate the Digital Twin, ensuring that the simulated sensor behavior aligns with real-world observations; the data collected during previous test runs can be used for visualization and analysis. Weather conditions are modeled to evaluate how rain, fog and snow impact sensor performance within the Digital Twin environment, to learn about the effects and predict sensor operation in different weather conditions. In this article, we examine the Digital Twin of our test track as a development environment for designing, deploying and testing ITS-enhanced road-weather services and warnings. These services integrate real-world road-weather observations, forecast data, roadside sensors and on-board vehicle measurements to support safe driving and optimize vehicle trajectories for both passenger and autonomous vehicles. This research is expected to benefit stakeholders involved in automotive testing, simulation and road-weather service development. Full article
Show Figures

Figure 1

24 pages, 4271 KB  
Article
Experimental Investigation of CFRP-Wrapped RC Columns Under Contact Explosions: Effects of Single vs. Dual-Layer Configurations
by Azer Maazoun, Oussama Atoui and Mohamed Ben Rhouma
Buildings 2026, 16(5), 943; https://doi.org/10.3390/buildings16050943 - 27 Feb 2026
Viewed by 225
Abstract
Reinforced concrete (RC) columns, vital components of urban infrastructure, are highly vulnerable to severe damage from contact explosions, posing significant threats to structural integrity and occupant safety. This study presents a rigorous experimental investigation into the dynamic blast response of RC columns and [...] Read more.
Reinforced concrete (RC) columns, vital components of urban infrastructure, are highly vulnerable to severe damage from contact explosions, posing significant threats to structural integrity and occupant safety. This study presents a rigorous experimental investigation into the dynamic blast response of RC columns and the efficacy of externally bonded Carbon Fiber Reinforced Polymer (CFRP) wraps as a retrofitting solution. Three series of scaled RC columns were subjected to controlled contact explosions using RDX charges of 50 g, 30 g, and 20 g. For each charge level, three configurations were tested: unretrofitted, single-layer unidirectional CFRP (hoop direction), and dual-layer orthogonal CFRP (hoop and longitudinal). A comprehensive instrumentation system, including high-speed cameras, accelerometers, and pressure transducers, captured blast overpressure, crack evolution, and dynamic acceleration. The results demonstrate that CFRP retrofitting substantially enhances blast resistance and structural performance. Peak accelerations were reduced by up to 68%, with the dual-layer configuration achieving the highest mitigation across all charge levels. In terms of damage control, a single CFRP layer reduced spalling height by 65%, while the dual-layer system achieved up to a 75% reduction. Damage depth was also mitigated by up to 60%, highlighting the superior energy dissipation and containment provided by multi-layered CFRP. These findings underscore CFRP’s significant potential as a robust, practical, and scalable retrofitting solution for enhancing the blast resilience of critical infrastructure, contributing directly to improved urban safety and structural protection in blast-prone environments. Full article
(This article belongs to the Section Building Materials, and Repair & Renovation)
Show Figures

Figure 1

16 pages, 3192 KB  
Article
The Design and Validation of a Satellite Camera Vibration Isolation Platform Supported by Multi-Strut Damping Rods
by Heng Zhang, Rongchang Xia, Yue Wang, Junsheng Yang and Yibin Liu
Coatings 2026, 16(3), 278; https://doi.org/10.3390/coatings16030278 - 26 Feb 2026
Viewed by 300
Abstract
An efficient passive bipod damping isolator is designed considering damping length and installation angle. Firstly, the mechanical properties of the single damping rod components are analyzed by the microelement method. Secondly, numerical models of a bipod configuration under two boundary conditions are developed [...] Read more.
An efficient passive bipod damping isolator is designed considering damping length and installation angle. Firstly, the mechanical properties of the single damping rod components are analyzed by the microelement method. Secondly, numerical models of a bipod configuration under two boundary conditions are developed to examine the effect of installation angle. Then, a satellite-bracket adapter component constructed with multiple damping rods is designed. Finally, the dynamic model of the satellite camera is established, and the performance of the vibration isolation system is verified through the finite element simulation and vibration experiments. Results indicate that the proposed multi-strut isolator achieves both high stiffness and effective vibration isolation. This study provides a theoretical basis and reference for the design of passive isolators with high stable and flexible support. Full article
(This article belongs to the Section Surface Characterization, Deposition and Modification)
Show Figures

Figure 1

13 pages, 2621 KB  
Article
Enhanced Optical Triangulation Method for Piezoelectric Stack
by Sinan Köksu and Sedat Nazlıbilek
Instruments 2026, 10(1), 13; https://doi.org/10.3390/instruments10010013 - 26 Feb 2026
Viewed by 246
Abstract
The precise control of piezoelectric actuators is limited by inherent hysteresis, creep, and nonlinear behavior, which necessitate high-resolution displacement sensing for effective closed-loop operation. Although optical interferometers can achieve nanometer and sub-nanometer resolution, their practical implementation is often constrained by complex optical alignment, [...] Read more.
The precise control of piezoelectric actuators is limited by inherent hysteresis, creep, and nonlinear behavior, which necessitate high-resolution displacement sensing for effective closed-loop operation. Although optical interferometers can achieve nanometer and sub-nanometer resolution, their practical implementation is often constrained by complex optical alignment, sensitivity to environmental disturbances, and limited robustness in high-speed measurements. Optical triangulation sensors offer a more robust and straightforward alternative; however, their resolution is typically insufficient for nanometer-scale displacement measurements. In this study, a novel optical triangulation sensor based on a two-stage geometric optical amplification scheme is proposed for measuring the expansion of piezoelectric stacks. The method relies purely on geometric optical amplification and does not require interferometric techniques or complex signal processing. Using off-the-shelf optical components and an industrial imaging sensor, the proposed system achieves a displacement resolution of 109.6 nm, a repeatability of 74.62 nm, and an accuracy of 98.81% with a maximum error of 207.14 nm under hysteresis measurements. The achieved resolution is primarily limited by the spatial resolution of the camera sensor, indicating that further improvements are possible through optimization of the optical configuration or the use of higher-resolution imaging devices. Owing to its simplicity and robustness, the proposed sensor is well suited for real-time closed-loop control of piezoelectric actuators. Full article
(This article belongs to the Section Sensing Technologies and Precision Measurement)
Show Figures

Figure 1

18 pages, 2413 KB  
Article
Towards Autonomous Optical Camera Communications: Light Source Localisation Using Deep Learning
by Elizabeth Eso, Sinan Sinanovic, Funmilayo B. Offiong, Xicong Li, Liying Yang, Sujan Rajbhandari and Zabih Ghassemlooy
Electronics 2026, 15(5), 935; https://doi.org/10.3390/electronics15050935 - 25 Feb 2026
Viewed by 249
Abstract
This research significantly improves the link reliability and robustness of optical camera communications (OCC) by leveraging deep learning for light source modulation filtering, reflection filtering, and precise light source localisation. By using image sensors as receivers in OCC, data transmission is not only [...] Read more.
This research significantly improves the link reliability and robustness of optical camera communications (OCC) by leveraging deep learning for light source modulation filtering, reflection filtering, and precise light source localisation. By using image sensors as receivers in OCC, data transmission is not only enabled, but other applications are also facilitated, such as detecting objects and humans, making OCC highly attractive in healthcare, intelligent transport systems, and indoor positioning. However, the position of the desired signal in the received image frame must be tracked in dynamic scenarios (i.e., nonstationary applications), in order to maintain the communication link. Moreover, as sixth-generation (6G) wireless networks envision highly autonomous systems that rely on seamless integration of communication and sensing, deep learning is key to enabling robust and adaptive light source localisation and sensing in OCC, which enables vision-based autonomy in dynamic environments. It should be noted that a deep learning-based approach provides more accuracy even when there are multiple noise sources in the environment, reflections, and complex backgrounds, and under mobility conditions, in which traditional light source detection/tracking methods are not effective. Hence this study investigates the use of a deep learning-based approach by analysing the detection accuracy under different configurations and unseen images. The results obtained demonstrate consistently high detection performance with average precision (at an intersection-over-union threshold of 0.70 of 0.84 to 0.97. These results pave the way for autonomous receivers that will be able to select signals intelligently and decode them. Full article
Show Figures

Figure 1

23 pages, 4564 KB  
Article
Two-Stage Wildlife Event Classification for Edge Deployment
by Aditya S. Viswanathan, Adis Bock, Zoe Bent, Mark A. Peyton, Daniel M. Tartakovsky and Javier E. Santos
Sensors 2026, 26(4), 1366; https://doi.org/10.3390/s26041366 - 21 Feb 2026
Viewed by 483
Abstract
Camera-based wildlife monitoring is often overwhelmed by non-target triggers and slowed by manual review or cloud-dependent inference, which can prevent timely intervention for high stakes human–wildlife conflicts. Our key contribution is a deployable, fully offline edge vision sensor that achieves near-real-time, highly accurate [...] Read more.
Camera-based wildlife monitoring is often overwhelmed by non-target triggers and slowed by manual review or cloud-dependent inference, which can prevent timely intervention for high stakes human–wildlife conflicts. Our key contribution is a deployable, fully offline edge vision sensor that achieves near-real-time, highly accurate wildlife event classification by combining detector-based empty-image suppression with a lightweight classifier trained with a staged transfer-learning curriculum. Specifically, Stage 1 uses a pretrained You Only Look Once (YOLO)-family detector for permissive animal localization and empty-trigger suppression, and Stage 2 uses a lightweight EfficientNet-based binary classifier to confirm puma on detector crops and gate downstream actions. Our design is robust to low-quality nighttime monochrome imagery (motion blur, low contrast, illumination artifacts, and partial-body captures) and operates using commercially available components in connectivity-limited settings. In field deployments running since May 2025, end-to-end latency from camera trigger to action command is approximately 4 s. Ablation studies using a dataset of labeled wildlife images (pumas, not pumas) show that the two-stage approach substantially reduces false alarms in identifying pumas relative to a full-image classifier while maintaining high recall. On the held-out test set (N=1434 events), the proposed two-stage cascade achieves precision 0.983, recall 0.975, F1 0.979, accuracy 0.986, and balanced accuracy 0.983, with only 8 false positives and 12 false negatives. The system can be easily adapted for other species, as demonstrated by rapid retraining of the second stage to classify ringtails. Downstream responses (e.g., notifications and optional audio/light outputs) provide flexible actuation capabilities that can be configured to support intervention. Full article
(This article belongs to the Special Issue AI-Based Computer Vision Sensors & Systems—2nd Edition)
Show Figures

Figure 1

Back to TopTop