Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (38)

Search Parameters:
Keywords = BoVWs

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 1656 KiB  
Article
A Hybrid Learning Framework for Enhancing Bridge Damage Prediction
by Amal Abdulbaqi Maryoosh, Saeid Pashazadeh and Pedram Salehpour
Appl. Syst. Innov. 2025, 8(3), 61; https://doi.org/10.3390/asi8030061 - 30 Apr 2025
Cited by 1 | Viewed by 642
Abstract
Bridges are crucial structures for transportation networks, and their structural integrity is paramount. Deterioration and damage to bridges can lead to significant economic losses, traffic disruptions, and, in severe cases, loss of life. Traditional methods of bridge damage detection, often relying on visual [...] Read more.
Bridges are crucial structures for transportation networks, and their structural integrity is paramount. Deterioration and damage to bridges can lead to significant economic losses, traffic disruptions, and, in severe cases, loss of life. Traditional methods of bridge damage detection, often relying on visual inspections, can be challenging or impossible in critical areas such as roofing, corners, and heights. Therefore, there is a pressing need for automated and accurate techniques for bridge damage detection. This study aims to propose a novel method for bridge crack detection that leverages a hybrid supervised and unsupervised learning strategy. The proposed approach combines pixel-based feature method local binary pattern (LBP) with the mid-level feature bag of visual words (BoVW) for feature extraction, followed by the Apriori algorithm for dimensionality reduction and optimal feature selection. The selected features are then trained using the MobileNet model. The proposed model demonstrates exceptional performance, achieving accuracy rates ranging from 98.27% to 100%, with error rates between 1.73% and 0% across multiple bridge damage datasets. This study contributes a reliable hybrid learning framework for minimizing error rates in bridge damage detection, showcasing the potential of combining LBP–BoVW features with MobileNet for image-based classification tasks. Full article
Show Figures

Figure 1

23 pages, 7345 KiB  
Article
Dynamical Mechanisms of Rapid Intensification and Multiple Recurvature of Pre-Monsoonal Tropical Cyclone Mocha over the Bay of Bengal
by Prabodha Kumar Pradhan, Sushant Kumar, Lokesh Kumar Pandey, Srinivas Desamsetti, Mohan S. Thota and Raghavendra Ashrit
Meteorology 2025, 4(2), 9; https://doi.org/10.3390/meteorology4020009 - 27 Mar 2025
Viewed by 996
Abstract
Cyclone Mocha, classified as an Extremely Severe Cyclonic Storm (ESCS), followed an unusual northeastward trajectory while exhibiting a well-defined eyewall structure. It experienced rapid intensification (RI) before making landfall along the Myanmar coast. It caused heavy rainfall (~90 mm) and gusty winds (~115 [...] Read more.
Cyclone Mocha, classified as an Extremely Severe Cyclonic Storm (ESCS), followed an unusual northeastward trajectory while exhibiting a well-defined eyewall structure. It experienced rapid intensification (RI) before making landfall along the Myanmar coast. It caused heavy rainfall (~90 mm) and gusty winds (~115 knots) over the coastal regions of Bay of Bengal Initiative for Multi-Sectoral Technical and Economic Cooperation (BIMSTEC) countries, such as the coasts of Bangladesh and Myanmar. The factors responsible for the RI of the cyclone in lower latitudes, such as sea surface temperature (SST), tropical cyclone heat potential (TCHP), vertical wind shear (VWS), and mid-tropospheric moisture content, are studied using the National Ocean and Atmospheric Administration (NOAA) SST and National Center for Medium-Range Weather Forecasting (NCMRWF) Unified Model (NCUM) global analysis. The results show that SST and TCHP values of 30 °C and 100 (KJ cm−2) over the Bay of Bengal (BoB) favored cyclogenesis. However, a VWS (ms−1) and relative humidity (RH; %) within the range of 10 ms−1 and >70% also provided a conducive environment for the low-pressure system to transform into the ESCS category. The physical mechanism of RI and recurvature of the Mocha cyclone have been investigated using forecast products and compared with Cooperative Institute for Research in the Atmosphere (CIRA) and Indian Meteorological Department (IMD) satellite observations. The key results indicate that a dry air intrusion associated with a series of troughs and ridges at a 500 hPa level due to the western disturbance (WD) during that time was very active over the northern part of India and adjoining Pakistan, which brought north-westerlies at the 200 hPa level. The existence of troughs at 500 and 200 hPa levels are significantly associated with a Rossby wave pattern over the mid-latitude that creates the baroclinic zone and favorable for the recurvature and RI of Mocha cyclone clearly represented in the NCUM analysis. Moreover the Q-vector analysis and steering flow (SF) emphasize the vertical motion and recurvature of the Mocha cyclone so as to move in a northeast direction, and this has been reasonably well represented by the NCUM model analysis and the 24, 7-, and 120 h forecasts. Additionally, a quantitative assessment of the system indicates that the model forecasts of TC tracks have an error of 50, 70, and 100 km in 24, 72, and 120 h lead times. Thus, this case study underscores the capability of the NCUM model in representing the physical mechanisms behind the recurving and RI over the BoB. Full article
Show Figures

Figure 1

17 pages, 3465 KiB  
Article
Effects of Combined Vibration Ergometry and Botulinum Toxin on Gait Improvement in Asymmetric Lower Limb Spasticity: A Pilot Study
by Harald Hefter, Dietmar Rosenthal and Sara Samadzadeh
J. Funct. Morphol. Kinesiol. 2025, 10(1), 41; https://doi.org/10.3390/jfmk10010041 - 21 Jan 2025
Cited by 1 | Viewed by 911
Abstract
Objective: Botulinum neurotoxin type A (BoNT/A) injections and the new vibration ergometry training (VET) are studied for their combined effect on improving functional mobility in patients with asymmetric lower limb spasticity. Method: Gait was analyzed using the Infotronic® system, which measures ground [...] Read more.
Objective: Botulinum neurotoxin type A (BoNT/A) injections and the new vibration ergometry training (VET) are studied for their combined effect on improving functional mobility in patients with asymmetric lower limb spasticity. Method: Gait was analyzed using the Infotronic® system, which measures ground reaction forces and foot contact patterns by means of special force-sensitive shoes strapped over feet or street shoes. Gait was measured several times, depending on the protocol patients underwent. Seven patients with asymmetric lower limb spasticity were analyzed according to the control protocol (CG-group): after a baseline walk of 20 m (NV-W1) patients received their routine BoNT/A injection and had to walk the same distance a second time (NV-W2). Approximately 3–5 weeks later, they had to walk a third time (NV-W3). A further seven patients (VG-group) were analyzed according to the vibration protocol: after a baseline walk (V-W1), patients underwent a first vibration training (VET1), walked a second time (V-W2), received their routine BoNT/A injection, and walked a third time (V-W3). About four weeks later, they had to walk again (V-W4), received another vibration training (VET3), and walked a fifth time (V-W5). At least six months after the analysis according to the vibration protocol, these patients were also analyzed according to the control protocol. Eleven gait parameters were compared between the CG- and VG-group, and within the VG-group. Result: Patients in the VG-group experienced a significant improvement in gait four weeks after BoNT/A injection, unlike the patients in the CG-group. VG-patients also showed improved gait after two VET sessions. However, there was no further functional improvement of gait when BoNT/A injections and VET sessions were combined. Conclusions: BoNT/A injections enhance functional mobility in patients with mild asymmetric leg spasticity. VET also induces an immediate gait improvement and offers a further treatment approach for leg spasticity. Whether combining BoNT treatment and vibration training offers superior outcomes compared to either treatment alone requires further investigation. Full article
(This article belongs to the Section Functional Anatomy and Musculoskeletal System)
Show Figures

Figure 1

16 pages, 7155 KiB  
Article
Overlapping Image-Set Determination Method Based on Hybrid BoVW-NoM Approach for UAV Image Localization
by Juyeon Lee and Kanghyeok Choi
Appl. Sci. 2024, 14(13), 5839; https://doi.org/10.3390/app14135839 - 4 Jul 2024
Cited by 2 | Viewed by 1284
Abstract
With the increasing use of unmanned aerial vehicles (UAVs) in various fields, achieving the precise localization of UAV images is crucial for enhancing their utility. Photogrammetry-based techniques, particularly bundle adjustment, serve as foundational methods for accurately determining the spatial coordinates of UAV images. [...] Read more.
With the increasing use of unmanned aerial vehicles (UAVs) in various fields, achieving the precise localization of UAV images is crucial for enhancing their utility. Photogrammetry-based techniques, particularly bundle adjustment, serve as foundational methods for accurately determining the spatial coordinates of UAV images. The effectiveness of bundle adjustment is significantly influenced by the selection of input data, particularly the composition of overlapping image sets. The selection process of overlapping images significantly impacts both the accuracy of spatial coordinate determination and the computational efficiency of UAV image localization. Therefore, a strategic approach to this selection is crucial for optimizing the performance of bundle adjustment in UAV image processing. In this context, we propose an efficient methodology for determining overlapping image sets. The proposed method selects overlapping images based on image similarity, leveraging the complementary strengths of the bag of visual words and number of matches techniques. Essentially, our method achieves both high accuracy and high speed by utilizing a Bag of Visual Words for candidate selection and the number of matches for additional similarity assessment for overlapping image-set determination. We compared the performance of our proposed methodology with the conventional number of matches and bag-of-visual word-based methods for overlapping image-set determination. In the comparative evaluation, the proposed method demonstrated an average precision of 96%, comparable to that of the number of matches-based approach, while surpassing the 62% precision achieved by both bag-of-visual-word methods. Moreover, the processing time decreased by approximately 0.11 times compared with the number of matches-based methods, demonstrating relatively high efficiency. Furthermore, in the bundle adjustment results using image sets, the proposed method, along with the number of matches-based methods, showed reprojection error values of less than 1, indicating relatively high accuracy and contributing to the improvement in accuracy in estimating image positions. Full article
Show Figures

Figure 1

18 pages, 6629 KiB  
Article
Gesture-to-Text Translation Using SURF for Indian Sign Language
by Kaustubh Mani Tripathi, Pooja Kamat, Shruti Patil, Ruchi Jayaswal, Swati Ahirrao and Ketan Kotecha
Appl. Syst. Innov. 2023, 6(2), 35; https://doi.org/10.3390/asi6020035 - 2 Mar 2023
Cited by 11 | Viewed by 9435
Abstract
This research paper focuses on developing an effective gesture-to-text translation system using state-of-the-art computer vision techniques. The existing research on sign language translation has yet to utilize skin masking, edge detection, and feature extraction techniques to their full potential. Therefore, this study employs [...] Read more.
This research paper focuses on developing an effective gesture-to-text translation system using state-of-the-art computer vision techniques. The existing research on sign language translation has yet to utilize skin masking, edge detection, and feature extraction techniques to their full potential. Therefore, this study employs the speeded-up robust features (SURF) model for feature extraction, which is resistant to variations such as rotation, perspective scaling, and occlusion. The proposed system utilizes a bag of visual words (BoVW) model for gesture-to-text conversion. The study uses a dataset of 42,000 photographs consisting of alphabets (A–Z) and numbers (1–9), divided into 35 classes with 1200 shots per class. The pre-processing phase includes skin masking, where the RGB color space is converted to the HSV color space, and Canny edge detection is used for sharp edge detection. The SURF elements are grouped and converted to a visual language using the K-means mini-batch clustering technique. The proposed system’s performance is evaluated using several machine learning algorithms such as naïve Bayes, logistic regression, K nearest neighbors, support vector machine, and convolutional neural network. All the algorithms benefited from SURF, and the system’s accuracy is promising, ranging from 79% to 92%. This research study not only presents the development of an effective gesture-to-text translation system but also highlights the importance of using skin masking, edge detection, and feature extraction techniques to their full potential in sign language translation. The proposed system aims to bridge the communication gap between individuals who cannot speak and those who cannot understand Indian Sign Language (ISL). Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

20 pages, 28117 KiB  
Article
A Robust and Efficient Loop Closure Detection Approach for Hybrid Ground/Aerial Vehicles
by Yutong Wang, Bin Xu, Wei Fan and Changle Xiang
Drones 2023, 7(2), 135; https://doi.org/10.3390/drones7020135 - 14 Feb 2023
Cited by 9 | Viewed by 3291
Abstract
Frequent and dramatic viewpoint changes make loop closure detection of hybrid ground/aerial vehicles extremely challenging. To address this issue, we present a robust and efficient loop closure detection approach based on the state-of-the-art simultaneous localization and mapping (SLAM) framework and pre-trained deep learning [...] Read more.
Frequent and dramatic viewpoint changes make loop closure detection of hybrid ground/aerial vehicles extremely challenging. To address this issue, we present a robust and efficient loop closure detection approach based on the state-of-the-art simultaneous localization and mapping (SLAM) framework and pre-trained deep learning models. First, the outputs of the SuperPoint network are processed to extract both tracking features and additional features used in loop closure. Next, binary-encoded SuperPoint descriptors are applied with a method based on Bag of VisualWords (BoVW) to detect loop candidates efficiently. Finally, the combination of SuperGlue and SuperPoint descriptors provides correspondences of keypoints to verify loop candidates and calculate relative poses. The system is evaluated on the public datasets and a real-world hybrid ground/aerial vehicles dataset. The proposed approach enables reliable loop detection, even when the relative translation between two viewpoints exceeds 7 m or one of the Euler angles is above 50°. Full article
(This article belongs to the Section Drone Design and Development)
Show Figures

Figure 1

24 pages, 1500 KiB  
Article
Hybrid Bag-of-Visual-Words and FeatureWiz Selection for Content-Based Visual Information Retrieval
by Samy Bakheet, Ayoub Al-Hamadi, Emadeldeen Soliman and Mohamed Heshmat
Sensors 2023, 23(3), 1653; https://doi.org/10.3390/s23031653 - 2 Feb 2023
Cited by 10 | Viewed by 3459
Abstract
Recently, content-based image retrieval (CBIR) based on bag-of-visual-words (BoVW) model has been one of the most promising and increasingly active research areas. In this paper, we propose a new CBIR framework based on the visual words fusion of multiple feature descriptors to achieve [...] Read more.
Recently, content-based image retrieval (CBIR) based on bag-of-visual-words (BoVW) model has been one of the most promising and increasingly active research areas. In this paper, we propose a new CBIR framework based on the visual words fusion of multiple feature descriptors to achieve an improved retrieval performance, where interest points are separately extracted from an image using features from accelerated segment test (FAST) and speeded-up robust features (SURF). The extracted keypoints are then fused together in a single keypoint feature vector and the improved RootSIFT algorithm is applied to describe the region surrounding each keypoint. Afterward, the FeatureWiz algorithm is employed to reduce features and select the best features for the BoVW learning model. To create the codebook, K-means clustering is applied to quantize visual features into a smaller set of visual words. Finally, the feature vectors extracted from the BoVW model are fed into a support vector machines (SVMs) classifier for image retrieval. An inverted index technique based on cosine distance metric is applied to sort the retrieved images to the similarity of the query image. Experiments on three benchmark datasets (Corel-1000, Caltech-10 and Oxford Flower-17) show that the presented CBIR technique can deliver comparable results to other state-of-the-art techniques, by achieving average accuracies of 92.94%, 98.40% and 84.94% on these datasets, respectively. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

21 pages, 5874 KiB  
Article
Deep Learning-Based BoVW–CRNN Model for Lung Tumor Detection in Nano-Segmented CT Images
by Aswathy S. U., Fathimathul Rajeena P. P., Ajith Abraham and Divya Stephen
Electronics 2023, 12(1), 14; https://doi.org/10.3390/electronics12010014 - 21 Dec 2022
Cited by 15 | Viewed by 2807
Abstract
One of the most common oncologies analyzed among people worldwide is lung malignancy. Early detection of lung malignancy helps find a suitable treatment for saving human lives. Due to its high resolution, greater transparency, and low noise and distortions, Computed Tomography (CT) images [...] Read more.
One of the most common oncologies analyzed among people worldwide is lung malignancy. Early detection of lung malignancy helps find a suitable treatment for saving human lives. Due to its high resolution, greater transparency, and low noise and distortions, Computed Tomography (CT) images are most commonly used for processing. In this context, this research work mainly focused on the multifaceted nature of lung cancer diagnosis, a quintessential, fascinating, and risky subject of oncology. The input used here has been nano-image, enhanced with a Gabor filter and modified color-based histogram equalization. Then, the image of lung cancer was segmented by using the Guaranteed Convergence Particle Swarm Optimization (GCPSO) algorithm. A graphical user interface nano-measuring tool was designed to classify the tumor region. The Bag of Visual Words (BoVW) and a Convolutional Recurrent Neural Network (CRNN) were employed for image classification and feature extraction processes. In terms of findings, we achieved the average precision of 96.5%, accuracy of 99.35%, sensitivity of 97%, specificity of 99% and F1 score of 95.5%. With the proposed solution, the overall time required for the segmentation of images was much smaller than the existing solutions. It is also remarkable that biocompatible-based nanotechnology was developed to distinguish the malignancy region on a nanometer scale and has to be evaluated automatically. That novel method succeeds in producing a proficient, robust, and precise segmentation of lesions in nano-CT images. Full article
Show Figures

Figure 1

16 pages, 5769 KiB  
Article
Multi-Subject Image Retrieval by Fusing Object and Scene-Level Feature Embeddings
by Chung-Gi Ban, Youngbae Hwang, Dayoung Park, Ryong Lee, Rae-Young Jang and Myung-Seok Choi
Appl. Sci. 2022, 12(24), 12705; https://doi.org/10.3390/app122412705 - 11 Dec 2022
Cited by 2 | Viewed by 2353
Abstract
Most existing image retrieval methods separately retrieve single images, such as a scene, content, or object, from a single database. However, for general purposes, target databases for image retrieval can include multiple subjects because it is not easy to predict which subject is [...] Read more.
Most existing image retrieval methods separately retrieve single images, such as a scene, content, or object, from a single database. However, for general purposes, target databases for image retrieval can include multiple subjects because it is not easy to predict which subject is entered. In this paper, we propose that image retrieval can be performed in practical applications by combining multiple databases. To deal with multi-subject image retrieval (MSIR), image embedding is generated through the fusion of scene- and object-level features, which are based on Detection Transformer (DETR) and a random patch generator with a deep-learning network, respectively. To utilize these feature vectors for image retrieval, two bags-of-visual-words (BoVWs) were used as feature embeddings because they are simply integrated with preservation of the characteristics of both features. A fusion strategy between the two BoVWs was proposed in three stages. Experiments were conducted to compare the proposed method with previous methods on conventional single-subject datasets and multi-subject datasets. The results validated that the proposed fused feature embeddings are effective for MSIR. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

17 pages, 9113 KiB  
Article
Determination of Moisture in Rice Grains Based on Visible Spectrum Analysis
by Héctor Palacios-Cabrera, Karina Jimenes-Vargas, Mario González, Omar Flor-Unda and Belén Almeida
Agronomy 2022, 12(12), 3021; https://doi.org/10.3390/agronomy12123021 - 29 Nov 2022
Cited by 9 | Viewed by 7050
Abstract
Rice grain production is important for the world economy. Determining the moisture content of the grains, at several stages of production, is crucial for controlling the quality, safety, and storage of the grain. This work inspects how well rice images from global and [...] Read more.
Rice grain production is important for the world economy. Determining the moisture content of the grains, at several stages of production, is crucial for controlling the quality, safety, and storage of the grain. This work inspects how well rice images from global and local descriptors work for determining the moisture content of the grains using artificial vision and intelligence techniques. Three sets of images of rice grains from the INIAP 12 variety (National Institute of Agricultural Research of Ecuador) were captured with a mobile camera. The first one with natural light and the other ones with a truncated pyramid-shaped structure. Then, a set of global descriptors (color, texture) and a set of local descriptors (AZAKE, BRISK, ORB, and SIFT) in conjunction with the dominate technique bag of visual words (BoVW) were used to analyze the content of the image with classification and regression algorithms. The results show that detecting humidity through images with classification and regression algorithms is possible. Finally, f1-score values of at least 0.9 were accomplished for global color descriptors and of 0.8 for texture descriptors, in contrast to the local descriptors (AKAZE, BRISK, and SIFT) that reached up to an f1-score of 0.96. Full article
(This article belongs to the Special Issue Application of Image Processing in Agriculture)
Show Figures

Figure 1

12 pages, 46409 KiB  
Article
Multi-Scale Feature Fusion for Interior Style Detection
by Akitaka Yaguchi, Keiko Ono, Erina Makihara, Naoya Ikushima and Tomomi Nakayama
Appl. Sci. 2022, 12(19), 9761; https://doi.org/10.3390/app12199761 - 28 Sep 2022
Cited by 3 | Viewed by 1631
Abstract
Text-based search engines can extract various types of information when a user enters an appropriate search query. However, a text-based search often fails in image retrieval when image understanding is needed. Deep learning (DL) is often used for image task problems, and various [...] Read more.
Text-based search engines can extract various types of information when a user enters an appropriate search query. However, a text-based search often fails in image retrieval when image understanding is needed. Deep learning (DL) is often used for image task problems, and various DL methods have successfully extracted visual features. However, as human perception differs for each individual, a dataset with an abundant number of images evaluated by human subjects is not available in many cases, although DL requires a considerable amount of data to estimate space ambiance, and the DL models that have been created are difficult to understand. In addition, it has been reported that texture is deeply related to space ambiance. Therefore, in this study, bag of visual words (BoVW) is used. By applying a hierarchical representation to BoVW, we propose a new interior style detection method using multi-scale features and boosting. The multi-scale features are created by combining global features from BoVW and local features that use object detection. Experiments on an image understanding task were conducted on a dataset consisting of room images with multiple styles. The results show that the proposed method improves the accuracy by 0.128 compared with the conventional method and by 0.021 compared with a residual network. Therefore, the proposed method can better detect interior style using multi-scale features. Full article
Show Figures

Figure 1

24 pages, 19969 KiB  
Article
Bubble Plume Target Detection Method of Multibeam Water Column Images Based on Bags of Visual Word Features
by Junxia Meng, Jun Yan and Jianhu Zhao
Remote Sens. 2022, 14(14), 3296; https://doi.org/10.3390/rs14143296 - 8 Jul 2022
Cited by 10 | Viewed by 2722
Abstract
Bubble plumes, as main manifestations of seabed gas leakage, play an important role in the exploration of natural gas hydrate and other resources. Multibeam water column images have been widely used in detecting bubble plume targets in recent years because they can wholly [...] Read more.
Bubble plumes, as main manifestations of seabed gas leakage, play an important role in the exploration of natural gas hydrate and other resources. Multibeam water column images have been widely used in detecting bubble plume targets in recent years because they can wholly record water column and seabed backscatter strengths. However, strong noises in multibeam water column images cause many issues in target detection, and traditional target detection methods are mainly used in optical images and are less efficient for noise-affected sonar images. To improve the detection accuracy of bubble plume targets in water column images, this study proposes a target detection method based on the bag of visual words (BOVW) features and support vector machine (SVM) classifier. First, the characteristics of bubble plume targets in water column images are analyzed, with the conclusion that the BOVW features can well express the gray scale, texture, and shape characteristics of bubble plumes. Second, the BOVW features are constructed following steps of point description extraction, description clustering, and feature encoding. Third, the quadratic SVM classifier is used for the recognition of target images. Finally, a procedure of bubble plume target detection in water column images is described. In the experiment using the measured data in the Strait of Georgia, the proposed method achieved 98.6% recognition accuracy of bubble plume targets in validation sets, and 91.7% correct detection rate of the targets in water column images. By comparison with other methods, the experimental results prove the validity and accuracy of the proposed method, and show potential applications of our method in the exploration and research on ocean resources. Full article
(This article belongs to the Special Issue Radar and Sonar Imaging and Processing Ⅲ)
Show Figures

Graphical abstract

19 pages, 5561 KiB  
Article
Performance Analysis of Supervised Machine Learning Algorithms for Automatized Radiographical Classification of Maxillary Third Molar Impaction
by Andreas Vollmer, Michael Vollmer, Gernot Lang, Anton Straub, Alexander Kübler, Sebastian Gubik, Roman C. Brands, Stefan Hartmann and Babak Saravi
Appl. Sci. 2022, 12(13), 6740; https://doi.org/10.3390/app12136740 - 3 Jul 2022
Cited by 9 | Viewed by 5737
Abstract
Background: Oro-antral communication (OAC) is a common complication following the extraction of upper molar teeth. The Archer and the Root Sinus (RS) systems can be used to classify impacted teeth in panoramic radiographs. The Archer classes B-D and the Root Sinus classes III, [...] Read more.
Background: Oro-antral communication (OAC) is a common complication following the extraction of upper molar teeth. The Archer and the Root Sinus (RS) systems can be used to classify impacted teeth in panoramic radiographs. The Archer classes B-D and the Root Sinus classes III, IV have been associated with an increased risk of OAC following tooth extraction in the upper molar region. In our previous study, we found that panoramic radiographs are not reliable for predicting OAC. This study aimed to (1) determine the feasibility of automating the classification (Archer/RS classes) of impacted teeth from panoramic radiographs, (2) determine the distribution of OAC stratified by classification system classes for the purposes of decision tree construction, and (3) determine the feasibility of automating the prediction of OAC utilizing the mentioned classification systems. Methods: We utilized multiple supervised pre-trained machine learning models (VGG16, ResNet50, Inceptionv3, EfficientNet, MobileNetV2), one custom-made convolutional neural network (CNN) model, and a Bag of Visual Words (BoVW) technique to evaluate the performance to predict the clinical classification systems RS and Archer from panoramic radiographs (Aim 1). We then used Chi-square Automatic Interaction Detectors (CHAID) to determine the distribution of OAC stratified by the Archer/RS classes to introduce a decision tree for simple use in clinics (Aim 2). Lastly, we tested the ability of a multilayer perceptron artificial neural network (MLP) and a radial basis function neural network (RBNN) to predict OAC based on the high-risk classes RS III, IV, and Archer B-D (Aim 3). Results: We achieved accuracies of up to 0.771 for EfficientNet and MobileNetV2 when examining the Archer classification. For the AUC, we obtained values of up to 0.902 for our custom-made CNN. In comparison, the detection of the RS classification achieved accuracies of up to 0.792 for the BoVW and an AUC of up to 0.716 for our custom-made CNN. Overall, the Archer classification was detected more reliably than the RS classification when considering all algorithms. CHAID predicted 77.4% correctness for the Archer classification and 81.4% for the RS classification. MLP (AUC: 0.590) and RBNN (AUC: 0.590) for the Archer classification as well as MLP 0.638) and RBNN (0.630) for the RS classification did not show sufficient predictive capability for OAC. Conclusions: The results reveal that impacted teeth can be classified using panoramic radiographs (best AUC: 0.902), and the classification systems can be stratified according to their relationship to OAC (81.4% correct for RS classification). However, the Archer and RS classes did not achieve satisfactory AUCs for predicting OAC (best AUC: 0.638). Additional research is needed to validate the results externally and to develop a reliable risk stratification tool based on the present findings. Full article
(This article belongs to the Special Issue Artificial Intelligence Applied to Dentistry)
Show Figures

Figure 1

23 pages, 3860 KiB  
Article
Remote Sensing Change Detection Based on Unsupervised Multi-Attention Slow Feature Analysis
by Weipeng Jing, Songyu Zhu, Peilun Kang, Jian Wang, Shengjia Cui, Guangsheng Chen and Houbing Song
Remote Sens. 2022, 14(12), 2834; https://doi.org/10.3390/rs14122834 - 13 Jun 2022
Cited by 7 | Viewed by 3481
Abstract
With the development of big data, analyzing the environmental benefits of transportation systems by artificial intelligence has become a hot issue in recent years. The ground traffic changes can be overlooked from a high-altitude perspective, using the technology of multi-temporal remote sensing change [...] Read more.
With the development of big data, analyzing the environmental benefits of transportation systems by artificial intelligence has become a hot issue in recent years. The ground traffic changes can be overlooked from a high-altitude perspective, using the technology of multi-temporal remote sensing change detection. We proposed a novel unsupervised algorithm by combining the image transformation and deep learning method. The new algorithm for remote sensing images is named multi-attention slow feature analysis (ASFA). In this model, three parts perform different functions respectively. The first part records to the K-BoVW to classify the categories of the ground objects as a channel parameter. The second part is a residual convolution with multiple attention mechanisms including temporal, spatial, and channel attention. Feature extraction and updating are completed at this link. Finally, we put the updated features in the slow feature analysis to highlight the variant components which we want and then generate the change map visually. Experiments on three very high-resolution datasets verified that the ASFA has a better performance than four basic change detection algorithms and an improved SFA algorithm. More importantly, this model works well for traffic road detection and helps us analyze the environmental benefits of traffic changes. Full article
Show Figures

Graphical abstract

16 pages, 5121 KiB  
Article
Deep Learning Method for Recognition and Classification of Images from Video Recorders in Difficult Weather Conditions
by Aleksey Osipov, Ekaterina Pleshakova, Sergey Gataullin, Sergey Korchagin, Mikhail Ivanov, Anton Finogeev and Vibhash Yadav
Sustainability 2022, 14(4), 2420; https://doi.org/10.3390/su14042420 - 20 Feb 2022
Cited by 38 | Viewed by 9675
Abstract
The sustainable functioning of the transport system requires solving the problems of identifying and classifying road users in order to predict the likelihood of accidents and prevent abnormal or emergency situations. The emergence of unmanned vehicles on urban highways significantly increases the risks [...] Read more.
The sustainable functioning of the transport system requires solving the problems of identifying and classifying road users in order to predict the likelihood of accidents and prevent abnormal or emergency situations. The emergence of unmanned vehicles on urban highways significantly increases the risks of such events. To improve road safety, intelligent transport systems, embedded computer vision systems, video surveillance systems, and photo radar systems are used. The main problem is the recognition and classification of objects and critical events in difficult weather conditions. For example, water drops, snow, dust, and dirt on camera lenses make images less accurate in object identification, license plate recognition, vehicle trajectory detection, etc. Part of the image is overlapped, distorted, or blurred. The article proposes a way to improve the accuracy of object identification by using the Canny operator to exclude the damaged areas of the image from consideration by capturing the clear parts of objects and ignoring the blurry ones. Only those parts of the image where this operator has detected the boundaries of the objects are subjected to further processing. To classify images by the remaining whole parts, we propose using a combined approach that includes the histogram-oriented gradient (HOG) method, a bag-of-visual-words (BoVW), and a back propagation neural network (BPNN). For the binary classification of the images of the damaged objects, this method showed a significant advantage over the classical method of convolutional neural networks (CNNs) (79 and 65% accuracies, respectively). The article also presents the results of a multiclass classification of the recognition objects on the basis of the damaged images, with an accuracy spread of 71 to 86%. Full article
(This article belongs to the Special Issue Public Transport Integration, Urban Density and Sustainability)
Show Figures

Figure 1

Back to TopTop