Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (9)

Search Parameters:
Keywords = video co-segmentation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 9870 KiB  
Article
The Role of Play in the Social Development of Grey Seal (Halichoerus grypus) Pups with Comparative Notes on the Harbour Seal (Phoca vitulina)
by Susan C. Wilson
Animals 2024, 14(14), 2086; https://doi.org/10.3390/ani14142086 - 17 Jul 2024
Cited by 1 | Viewed by 1599
Abstract
Juvenile grey seals are known to be highly social, interacting with contact behaviours interpreted as gentle play. However, minimal sociality of pups with their mothers and among weaned pups has been suggested. The present study aimed to observe the natural social interactions of [...] Read more.
Juvenile grey seals are known to be highly social, interacting with contact behaviours interpreted as gentle play. However, minimal sociality of pups with their mothers and among weaned pups has been suggested. The present study aimed to observe the natural social interactions of pups to track the early ontogeny of their sociality. Pup behaviour at a salt marsh colony on the east coast of England was video-recorded. Pups interacted with their mothers around suckling bouts and after weaning as they gathered around pools. The records were transcribed to spreadsheets in 30 s time segments to estimate the frequency and co-occurrence of different behaviours. Mother-pup interaction comprised nosing contacts and sometimes contact play, involving one laying the head and fore-flipper over the other. Initial weaned pup encounters involved tentative nosing and defensive splashing, indicating contact shyness. However, socially orientated locomotor play, supine posturing, and exaggerated raising of fore- and hind-flippers led to reduced shyness and pups following one another towards the sea. Archive data on subadult interactions and on harbour seal behaviours were re-analysed. Gentle play-like contact between mother–pup, juvenile, and adult pairs is interpreted here as a universal mode of social bonding, underscoring the social structure of both grey and harbour seals. Full article
(This article belongs to the Special Issue Animal Ontogeny, Plasticity and Ecology)
Show Figures

Figure 1

18 pages, 6458 KiB  
Article
GAR-Net: Guided Attention Residual Network for Polyp Segmentation from Colonoscopy Video Frames
by Joel Raymann and Ratnavel Rajalakshmi
Diagnostics 2023, 13(1), 123; https://doi.org/10.3390/diagnostics13010123 - 30 Dec 2022
Cited by 7 | Viewed by 2451 | Correction
Abstract
Colorectal Cancer is one of the most common cancers found in human beings, and polyps are the predecessor of this cancer. Accurate Computer-Aided polyp detection and segmentation system can help endoscopists to detect abnormal tissues and polyps during colonoscopy examination, thereby reducing the [...] Read more.
Colorectal Cancer is one of the most common cancers found in human beings, and polyps are the predecessor of this cancer. Accurate Computer-Aided polyp detection and segmentation system can help endoscopists to detect abnormal tissues and polyps during colonoscopy examination, thereby reducing the chance of polyps growing into cancer. Many of the existing techniques fail to delineate the polyps accurately and produce a noisy/broken output map if the shape and size of the polyp are irregular or small. We propose an end-to-end pixel-wise polyp segmentation model named Guided Attention Residual Network (GAR-Net) by combining the power of both residual blocks and attention mechanisms to obtain a refined continuous segmentation map. An enhanced Residual Block is proposed that suppresses the noise and captures low-level feature maps, thereby facilitating information flow for a more accurate semantic segmentation. We propose a special learning technique with a novel attention mechanism called Guided Attention Learning that can capture the refined attention maps both in earlier and deeper layers regardless of the size and shape of the polyp. To study the effectiveness of the proposed GAR-Net, various experiments were carried out on two benchmark collections viz., CVC-ClinicDB (CVC-612) and Kvasir-SEG dataset. From the experimental evaluations, it is shown that GAR-Net outperforms other previously proposed models such as FCN8, SegNet, U-Net, U-Net with Gated Attention, ResUNet, and DeepLabv3. Our proposed model achieves 91% Dice co-efficient and 83.12% mean Intersection over Union (mIoU) on the benchmark CVC-ClinicDB (CVC-612) dataset and 89.15% dice co-efficient and 81.58% mean Intersection over Union (mIoU) on the Kvasir-SEG dataset. The proposed GAR-Net model provides a robust solution for polyp segmentation from colonoscopy video frames. Full article
Show Figures

Figure 1

11 pages, 2635 KiB  
Article
A Steganalysis Classification Algorithm Based on Distinctive Texture Features
by Baraa Tareq Hammad, Ismail Taha Ahmed and Norziana Jamil
Symmetry 2022, 14(2), 236; https://doi.org/10.3390/sym14020236 - 25 Jan 2022
Cited by 25 | Viewed by 3911
Abstract
Steganography is the technique for secretly hiding messages in media such as text, audio, image, and video without being discovered. Image is one of the most essential media for concealing data, making it hard to identify hidden data not visible to the human [...] Read more.
Steganography is the technique for secretly hiding messages in media such as text, audio, image, and video without being discovered. Image is one of the most essential media for concealing data, making it hard to identify hidden data not visible to the human eye. In general, the cover image and the encrypted image are symmetrical in terms of dimension size, resolution, and qualities. This makes the difference difficult to perceive with the human eye. As a result, distinguishing between the two symmetric images required the development of methods. Steganalysis is a technique for identifying hidden messages embedded in digital material without having to know the embedding algorithm or the “non-stego” image. Due to their enormous feature vector dimension, which requires more time to calculate, the performance of most existing image steganalysis classification (ISC) techniques is still restricted. Therefore, in this research, we present a steganalysis classification method based on one of the texture features chosen, such as segmentation-based fractal texture analysis (SFTA), local binary pattern (LBP), and gray-level co-occurrence matrix (GLCM). The classifiers employed include Gaussian discriminant analysis (GDA) and naïve Bayes (NB). We used a public database in our proposed method and applied it to IStego100K datasets to be able to assess its performance. The experimental results reveal that in all classifiers, the SFTA feature surpassed all of the texture features, making it a great texture feature for image steganalysis classification. In terms of feature dimension and classification accuracy (CA), a comparison was made between the suggested SFTA-based GDA approach and various current ISC methods. The outcomes of the comparison are obvious show that the proposed method surpasses current methods. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

19 pages, 4031 KiB  
Article
Saliency Detection with Moving Camera via Background Model Completion
by Yu-Pei Zhang and Kwok-Leung Chan
Sensors 2021, 21(24), 8374; https://doi.org/10.3390/s21248374 - 15 Dec 2021
Cited by 2 | Viewed by 2915
Abstract
Detecting saliency in videos is a fundamental step in many computer vision systems. Saliency is the significant target(s) in the video. The object of interest is further analyzed for high-level applications. The segregation of saliency and the background can be made if they [...] Read more.
Detecting saliency in videos is a fundamental step in many computer vision systems. Saliency is the significant target(s) in the video. The object of interest is further analyzed for high-level applications. The segregation of saliency and the background can be made if they exhibit different visual cues. Therefore, saliency detection is often formulated as background subtraction. However, saliency detection is challenging. For instance, dynamic background can result in false positive errors. In another scenario, camouflage will result in false negative errors. With moving cameras, the captured scenes are even more complicated to handle. We propose a new framework, called saliency detection via background model completion (SD-BMC), that comprises a background modeler and a deep learning background/foreground segmentation network. The background modeler generates an initial clean background image from a short image sequence. Based on the idea of video completion, a good background frame can be synthesized with the co-existence of changing background and moving objects. We adopt the background/foreground segmenter, which was pre-trained with a specific video dataset. It can also detect saliency in unseen videos. The background modeler can adjust the background image dynamically when the background/foreground segmenter output deteriorates during processing a long video. To the best of our knowledge, our framework is the first one to adopt video completion for background modeling and saliency detection in videos captured by moving cameras. The F-measure results, obtained from the pan-tilt-zoom (PTZ) videos, show that our proposed framework outperforms some deep learning-based background subtraction models by 11% or more. With more challenging videos, our framework also outperforms many high-ranking background subtraction methods by more than 3%. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

9 pages, 2707 KiB  
Case Report
Organizing Pneumonia and Microvascular Fibrosis as Late Sequelae after a COVID-19 Infection. A Case Report
by Johan L. Dikken, Alexander P. W. M. Maat, Janina L. Wolf, Henrik Endeman, Rogier A. S. Hoek, Ad J. J. C. Bogers and Edris A. F. Mahtab
Surgeries 2021, 2(2), 190-198; https://doi.org/10.3390/surgeries2020020 - 29 May 2021
Cited by 1 | Viewed by 3760
Abstract
We report a patient with COVID-19 requiring hospitalization for two weeks, complicated by multiple segmental pulmonary embolisms for which dabigatran was initiated. After clearing the infection, the patient remained asymptomatic for 5 months. He was then readmitted with a spontaneous haemothorax, most likely [...] Read more.
We report a patient with COVID-19 requiring hospitalization for two weeks, complicated by multiple segmental pulmonary embolisms for which dabigatran was initiated. After clearing the infection, the patient remained asymptomatic for 5 months. He was then readmitted with a spontaneous haemothorax, most likely related to the use of dabigatran, which progressed to a pleural empyema with a trapped lung. The patient underwent a video assisted thoracoscopy (VATS) with decortication. Because of focal abnormalities, biopsies for histopathology were taken from the lung parenchyma. These showed an organizing pneumonia with progression towards fibrosis and arteries with intimal fibrosis. So far, no histopathological reports exist on late pulmonary changes after a COVID-19 infection. The unusual combined presence of microvascular damage and interstitial fibrosis may reflect a pathophysiological concept in which early endothelial damage by SARS-CoV-2 can lead to a chronic state of microvascular damage, low grade inflammation, and early progression towards pulmonary fibrosis. Full article
Show Figures

Figure 1

16 pages, 7931 KiB  
Article
Unsupervised Learning from Videos for Object Discovery in Single Images
by Dong Zhao, Baoqing Ding, Yulin Wu, Lei Chen and Hongchao Zhou
Symmetry 2021, 13(1), 38; https://doi.org/10.3390/sym13010038 - 29 Dec 2020
Cited by 4 | Viewed by 2668
Abstract
This paper proposes a method for discovering the primary objects in single images by learning from videos in a purely unsupervised manner—the learning process is based on videos, but the generated network is able to discover objects from a single input image. The [...] Read more.
This paper proposes a method for discovering the primary objects in single images by learning from videos in a purely unsupervised manner—the learning process is based on videos, but the generated network is able to discover objects from a single input image. The rough idea is that an image typically consists of multiple object instances (like the foreground and background) that have spatial transformations across video frames and they can be sparsely represented. By exploring the sparsity representation of a video with a neural network, one may learn the features of each object instance without any labels, which can be used to discover, recognize, or distinguish object instances from a single image. In this paper, we consider a relatively simple scenario, where each image roughly consists of a foreground and a background. Our proposed method is based on encoder-decoder structures to sparsely represent the foreground, background, and segmentation mask, which further reconstruct the original images. We apply the feed-forward network trained from videos for object discovery in single images, which is different from the previous co-segmentation methods that require videos or collections of images as the input for inference. The experimental results on various object segmentation benchmarks demonstrate that the proposed method extracts primary objects accurately and robustly, which suggests that unsupervised image learning tasks can benefit from the sparsity of images and the inter-frame structure of videos. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

11 pages, 2222 KiB  
Article
Deep Pixel-Level Matching via Attention for Video Co-Segmentation
by Junliang Li, Hon-Cheng Wong, Shengfeng He, Sio-Long Lo, Guifang Zhang and Wenxiao Wang
Appl. Sci. 2020, 10(6), 1948; https://doi.org/10.3390/app10061948 - 12 Mar 2020
Cited by 1 | Viewed by 2641
Abstract
In video object co-segmentation, methods based on patch-level matching are widely leveraged to extract the similarity between video frames. However, these methods can easily lead to pixel misclassification because they reduce the precision of pixel localization; thus, the accuracies of the segmentation results [...] Read more.
In video object co-segmentation, methods based on patch-level matching are widely leveraged to extract the similarity between video frames. However, these methods can easily lead to pixel misclassification because they reduce the precision of pixel localization; thus, the accuracies of the segmentation results of these methods are deducted. To address this problem, we propose a framework based on deep neural networks and equipped with a new attention module, which is designed for pixel-level matching to segment the object across video frames in this paper. In this attention module, the pixel-level matching step is able to compare the feature value of each pixel from one input frame with that of each pixel from another input frame for computing the similarity between two frames. Then a features fusion step is applied to efficiently fuse the feature maps of each frame with the similarity information for generating dense attention features. Finally, an up-sampling step refines the feature maps for obtaining high quality segmentation results by using these dense attention features. The ObMiC and DAVIS 2016 datasets were utilized to train and test our framework. Experimental results show that our framework achieves higher accuracy than those of other video segmentation methods that perform well in common information extraction. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

17 pages, 1830 KiB  
Article
Accelerating SuperBE with Hardware/Software Co-Design
by Andrew Tzer-Yeu Chen, Rohaan Gupta, Anton Borzenko, Kevin I-Kai Wang and Morteza Biglari-Abhari
J. Imaging 2018, 4(10), 122; https://doi.org/10.3390/jimaging4100122 - 18 Oct 2018
Cited by 7 | Viewed by 6239
Abstract
Background Estimation is a common computer vision task, used for segmenting moving objects in video streams. This can be useful as a pre-processing step, isolating regions of interest for more complicated algorithms performing detection, recognition, and identification tasks, in order to reduce overall [...] Read more.
Background Estimation is a common computer vision task, used for segmenting moving objects in video streams. This can be useful as a pre-processing step, isolating regions of interest for more complicated algorithms performing detection, recognition, and identification tasks, in order to reduce overall computation time. This is especially important in the context of embedded systems like smart cameras, which may need to process images with constrained computational resources. This work focuses on accelerating SuperBE, a superpixel-based background estimation algorithm that was designed for simplicity and reducing computational complexity while maintaining state-of-the-art levels of accuracy. We explore both software and hardware acceleration opportunities, converting the original algorithm into a greyscale, integer-only version, and using Hardware/Software Co-design to develop hardware acceleration components on FPGA fabric that assist a software processor. We achieved a 4.4× speed improvement with the software optimisations alone, and a 2× speed improvement with the hardware optimisations alone. When combined, these led to a 9× speed improvement on a Cyclone V System-on-Chip, delivering almost 38 fps on 320 × 240 resolution images. Full article
(This article belongs to the Special Issue Image Processing Using FPGAs)
Show Figures

Figure 1

16 pages, 5457 KiB  
Article
Energy Level-Based Abnormal Crowd Behavior Detection
by Xuguang Zhang, Qian Zhang, Shuo Hu, Chunsheng Guo and Hui Yu
Sensors 2018, 18(2), 423; https://doi.org/10.3390/s18020423 - 1 Feb 2018
Cited by 33 | Viewed by 5947
Abstract
The change of crowd energy is a fundamental measurement for describing a crowd behavior. In this paper, we present a crowd abnormal detection method based on the change of energy-level distribution. The method can not only reduce the camera perspective effect, but also [...] Read more.
The change of crowd energy is a fundamental measurement for describing a crowd behavior. In this paper, we present a crowd abnormal detection method based on the change of energy-level distribution. The method can not only reduce the camera perspective effect, but also detect crowd abnormal behavior in time. Pixels in the image are treated as particles, and the optical flow method is adopted to extract the velocities of particles. The qualities of different particles are distributed as different value according to the distance between the particle and the camera to reduce the camera perspective effect. Then a crowd motion segmentation method based on flow field texture representation is utilized to extract the motion foreground, and a linear interpolation calculation is applied to pedestrian’s foreground area to determine their distance to the camera. This contributes to the calculation of the particle qualities in different locations. Finally, the crowd behavior is analyzed according to the change of the consistency, entropy and contrast of the three descriptors for co-occurrence matrix. By calculating a threshold, the timestamp when the crowd abnormal happens is determined. In this paper, multiple sets of videos from three different scenes in UMN dataset are employed in the experiment. The results show that the proposed method is effective in characterizing anomalies in videos. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Back to TopTop