Deep/Machine Learning in Visual Recognition and Anomaly Detection

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Electronic Multimedia".

Deadline for manuscript submissions: 15 January 2025 | Viewed by 535

Special Issue Editors


E-Mail Website
Guest Editor
School of Telecommunications Engineering, Xidian University, Xi’an 710071, China
Interests: computer vision; pattern recognition; machine learning

E-Mail Website
Guest Editor
School of Computer Science, Shaanxi Normal University, Xi’an 710062, China
Interests: computer vision; artificial intelligence

Special Issue Information

Dear Colleagues,

The success of machine learning and deep neural networks have facilitated advances in understanding the high-level semantics of visual content. Conventional learning-based visual semantic recognition approaches rely heavily on large-scale training data with dense annotations and consistently fail to estimate accurate semantic labels for unseen categories. The emergence and rapid progress of few-/zero-shot learning make it possible to learn unseen categories from a few labeled or even zero-labeled samples, which advances the extension to practical applications. This Special Issue aims to demonstrate (1) how machine learning algorithms have contributed, and are contributing, to new theories, models, and datasets related to the topic of few-/zero-shot learning; (2) how few-/zero-shot learning can facilitate other tasks such as visual recognition and anomaly detection. The editors hope to collate a group of research results to report the recent developments in the related research topics. In addition, researchers can exchange their innovative ideas on the topic of few-/zero-shot learning in visual recognition and anomaly detection by submitting manuscripts for this Special Issue.

In this Special Issue, original research articles and reviews are welcome. Research areas may include (but are not limited to) the following:

(1) Theoretical advances and algorithm developments in few-/zero-shot learning;

(2) Useful applications of few-/zero-shot learning in visual recognition and anomaly detection;

(3) New datasets and benchmarks for few-/zero-shot learning in visual recognition and anomaly detection.

We look forward to receiving your contributions.

Dr. Yang Liu
Dr. Jin Li
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • few-/zero-shot learning
  • visual recognition
  • anomaly detection

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

18 pages, 1072 KiB  
Article
Leveraging Self-Distillation and Disentanglement Network to Enhance Visual–Semantic Feature Consistency in Generalized Zero-Shot Learning
by Xiaoming Liu, Chen Wang, Guan Yang, Chunhua Wang, Yang Long, Jie Liu and Zhiyuan Zhang
Electronics 2024, 13(10), 1977; https://doi.org/10.3390/electronics13101977 - 18 May 2024
Viewed by 324
Abstract
Generalized zero-shot learning (GZSL) aims to simultaneously recognize both seen classes and unseen classes by training only on seen class samples and auxiliary semantic descriptions. Recent state-of-the-art methods infer unseen classes based on semantic information or synthesize unseen classes using generative models based [...] Read more.
Generalized zero-shot learning (GZSL) aims to simultaneously recognize both seen classes and unseen classes by training only on seen class samples and auxiliary semantic descriptions. Recent state-of-the-art methods infer unseen classes based on semantic information or synthesize unseen classes using generative models based on semantic information, all of which rely on the correct alignment of visual–semantic features. However, they often overlook the inconsistency between original visual features and semantic attributes. Additionally, due to the existence of cross-modal dataset biases, the visual features extracted and synthesized by the model may also mismatch with some semantic features, which could hinder the model from properly aligning visual–semantic features. To address this issue, this paper proposes a GZSL framework that enhances the consistency of visual–semantic features using a self-distillation and disentanglement network (SDDN). The aim is to utilize the self-distillation and disentanglement network to obtain semantically consistent refined visual features and non-redundant semantic features to enhance the consistency of visual–semantic features. Firstly, SDDN utilizes self-distillation technology to refine the extracted and synthesized visual features of the model. Subsequently, the visual–semantic features are then disentangled and aligned using a disentanglement network to enhance the consistency of the visual–semantic features. Finally, the consistent visual–semantic features are fused to jointly train a GZSL classifier. Extensive experiments demonstrate that the proposed method achieves more competitive results on four challenging benchmark datasets (AWA2, CUB, FLO, and SUN). Full article
(This article belongs to the Special Issue Deep/Machine Learning in Visual Recognition and Anomaly Detection)
Show Figures

Figure 1

Back to TopTop