Unsupervised Deep Learning and Its Applications in Imaging Processing

A special issue of Journal of Imaging (ISSN 2313-433X).

Deadline for manuscript submissions: closed (30 April 2022) | Viewed by 8399

Special Issue Editors


E-Mail Website
Guest Editor
1. Institut Supérieur d’Electronique de Paris, 10 rue de Vanves, 92130 Issy Les Moulineaux, France
2. LIPN-CNRS UMR 7030, University Paris 13, 99 Avenue J.-B. Clément, 93430 Villetaneuse, France
Interests: unsupervised learning; satellite images; medical images; change detection; deep learning; time series prediction; collaborative learning

E-Mail Website
Guest Editor
Institut Supérieur d’Electronique de Paris, 10 rue de Vanves, 92130 Issy Les Moulineaux, France
Interests: unsupervised learning; medical images; deep learning; machine learning; mathematical morphology

Special Issue Information

Dear Colleagues

Nowadays, deep learning algorithms are the most successful methods in image processing, thanks to an ever-increasing number of networks and architectures that have proven more efficient than anything designed before. Yet, one major weakness known to deep learning methods is their need for huge amounts of data to be trained and show good performance. Another key element to their success lies in the manual annotation of the training data. Indeed, most deep learning methods used in image processing are supervised. It implies that for applications where no or not enough annotated images are available, there is a real need for the development of unsupervised deep learning methods able to analyse them! This is the case for various applications such as medical image analysis, remote sensing analysis, and many segmentation tasks, where the diversity of images and tasks to be performed often makes it impossible to have enough annotated images to train a supervised deep learning architecture that will lead to good performances. Furthermore, even when only very few annotated images are available, the question of the labelling reliability is also an issue for tasks that require high accuracy.

Given these very concrete problems, unsupervised deep learning methods—or self-supervised methods as they are sometimes called—open the possibility of quickly processing images that would have been otherwise impossible to analyse with regular deep learning methods. Yet, there are many challenges to the development of such methods that so far mostly rely on reconstruction loss functions from autoencoder models, and also suffer from issues similar to the ones faced by traditional unsupervised methods such as linking the elements of interest found by an unsupervised method with actual classes that are relevant to a real-world application.

For all these reasons, we invite you to submit your work related to unsupervised or weakly supervised applications of deep learning for image analysis, regardless of the field of application.

Prof. Dr. Jérémie Sublime
Prof. Dr. Hélène Urien
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Journal of Imaging is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • clustering
  • autoencoder
  • unsupervised learning
  • self-supervised learning
  • automated segmentation
  • semi-supervised learning
  • weakly supervised learning

Published Papers (3 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

18 pages, 3939 KiB  
Article
Proposals Generation for Weakly Supervised Object Detection in Artwork Images
by Federico Milani, Nicolò Oreste Pinciroli Vago and Piero Fraternali
J. Imaging 2022, 8(8), 215; https://doi.org/10.3390/jimaging8080215 - 06 Aug 2022
Cited by 3 | Viewed by 1843
Abstract
Object Detection requires many precise annotations, which are available for natural images but not for many non-natural data sets such as artworks data sets. A solution is using Weakly Supervised Object Detection (WSOD) techniques that learn accurate object localization from image-level labels. Studies [...] Read more.
Object Detection requires many precise annotations, which are available for natural images but not for many non-natural data sets such as artworks data sets. A solution is using Weakly Supervised Object Detection (WSOD) techniques that learn accurate object localization from image-level labels. Studies have demonstrated that state-of-the-art end-to-end architectures may not be suitable for domains in which images or classes sensibly differ from those used to pre-train networks. This paper presents a novel two-stage Weakly Supervised Object Detection approach for obtaining accurate bounding boxes on non-natural data sets. The proposed method exploits existing classification knowledge to generate pseudo-ground truth bounding boxes from Class Activation Maps (CAMs). The automatically generated annotations are used to train a robust Faster R-CNN object detector. Quantitative and qualitative analysis shows that bounding boxes generated from CAMs can compensate for the lack of manually annotated ground truth (GT) and that an object detector, trained with such pseudo-GT, surpasses end-to-end WSOD state-of-the-art methods on ArtDL 2.0 (≈41.5% mAP) and IconArt (≈17% mAP), two artworks data sets. The proposed solution is a step towards the computer-aided study of non-natural images and opens the way to more advanced tasks, e.g., automatic artwork image captioning for digital archive applications. Full article
(This article belongs to the Special Issue Unsupervised Deep Learning and Its Applications in Imaging Processing)
Show Figures

Figure 1

12 pages, 944 KiB  
Article
HFM: A Hybrid Feature Model Based on Conditional Auto Encoders for Zero-Shot Learning
by Fadi Al Machot, Mohib Ullah and Habib Ullah
J. Imaging 2022, 8(6), 171; https://doi.org/10.3390/jimaging8060171 - 16 Jun 2022
Cited by 6 | Viewed by 1785
Abstract
Zero-Shot Learning (ZSL) is related to training machine learning models capable of classifying or predicting classes (labels) that are not involved in the training set (unseen classes). A well-known problem in Deep Learning (DL) is the requirement for large amount of training data. [...] Read more.
Zero-Shot Learning (ZSL) is related to training machine learning models capable of classifying or predicting classes (labels) that are not involved in the training set (unseen classes). A well-known problem in Deep Learning (DL) is the requirement for large amount of training data. Zero-Shot learning is a straightforward approach that can be applied to overcome this problem. We propose a Hybrid Feature Model (HFM) based on conditional autoencoders for training a classical machine learning model on pseudo training data generated by two conditional autoencoders (given the semantic space as a condition): (a) the first autoencoder is trained with the visual space concatenated with the semantic space and (b) the second autoencoder is trained with the visual space as an input. Then, the decoders of both autoencoders are fed by the test data of the unseen classes to generate pseudo training data. To classify the unseen classes, the pseudo training data are combined to train a support vector machine. Tests on four different benchmark datasets show that the proposed method shows promising results compared to the current state-of-the-art when it comes to settings for both standard Zero-Shot Learning (ZSL) and Generalized Zero-Shot Learning (GZSL). Full article
(This article belongs to the Special Issue Unsupervised Deep Learning and Its Applications in Imaging Processing)
Show Figures

Figure 1

20 pages, 4325 KiB  
Article
Tracking Highly Similar Rat Instances under Heavy Occlusions: An Unsupervised Deep Generative Pipeline
by Anna Gelencsér-Horváth, László Kopácsi, Viktor Varga, Dávid Keller, Árpád Dobolyi, Kristóf Karacs and András Lőrincz
J. Imaging 2022, 8(4), 109; https://doi.org/10.3390/jimaging8040109 - 13 Apr 2022
Cited by 1 | Viewed by 4180
Abstract
Identity tracking and instance segmentation are crucial in several areas of biological research. Behavior analysis of individuals in groups of similar animals is a task that emerges frequently in agriculture or pharmaceutical studies, among others. Automated annotation of many hours of surveillance videos [...] Read more.
Identity tracking and instance segmentation are crucial in several areas of biological research. Behavior analysis of individuals in groups of similar animals is a task that emerges frequently in agriculture or pharmaceutical studies, among others. Automated annotation of many hours of surveillance videos can facilitate a large number of biological studies/experiments, which otherwise would not be feasible. Solutions based on machine learning generally perform well in tracking and instance segmentation; however, in the case of identical, unmarked instances (e.g., white rats or mice), even state-of-the-art approaches can frequently fail. We propose a pipeline of deep generative models for identity tracking and instance segmentation of highly similar instances, which, in contrast to most region-based approaches, exploits edge information and consequently helps to resolve ambiguity in heavily occluded cases. Our method is trained by synthetic data generation techniques, not requiring prior human annotation. We show that our approach greatly outperforms other state-of-the-art unsupervised methods in identity tracking and instance segmentation of unmarked rats in real-world laboratory video recordings. Full article
(This article belongs to the Special Issue Unsupervised Deep Learning and Its Applications in Imaging Processing)
Show Figures

Figure 1

Back to TopTop