Machine Learning in Image Processing and Computer Vision

A special issue of Information (ISSN 2078-2489). This special issue belongs to the section "Information Applications".

Deadline for manuscript submissions: 30 November 2026 | Viewed by 1795

Special Issue Editor


E-Mail Website
Guest Editor
Institute of Control and Industrial Electronics, Faculty of Electrical Engineering, Warsaw University of Technology, ul. Koszykowa 75, 00-662 Warsaw, Poland
Interests: image processing; computer vision; machine learning; computational intelligence; neural networks

Special Issue Information

Dear Colleagues,

Recent advances in machine learning have dramatically transformed the capabilities of image processing and computer vision, opening new possibilities for detailed visual understanding. One of the most challenging and intellectually demanding areas in this field is fine-grained object recognition, the task of distinguishing between visually similar categories characterized by subtle differences and high intra-class variability. This issue has become critical in applications such as retail product identification, animal breed classification, or precision inspection in manufacturing, where recognizing visual nuances is essential.

This Special Issue aims to showcase recent developments that address these challenges using robust learning architectures, semantic modeling, and efficient recognition techniques with high resolutions and low semantic granularity.

We especially encourage submissions that explore a broad range of modern methodologies, including, but not limited to, the following:

  • Prototype-based and metric learning methods which enhance interpretability and class separability by modeling data around representative exemplars;
  • Proxy-based learning allowing for scalable training and inference in classification tasks with large or imbalanced category sets;
  • Domain adaptation and domain generalization, essential for transferring models across visually diverse environments (e.g., synthetic-to-real, controlled-to-wild conditions);
  • Few-shot and meta-learning strategies enabling models to generalize from limited labeled examples or rapidly adapt to novel classes;
  • Contrastive learning and self-supervised pretraining, which improve feature quality and sample efficiency, especially in data-scarce settings;
  • Knowledge distillation to compress and transfer information from large high-capacity models into lightweight deployable ones;
  • Attention mechanisms, transformers, and hierarchical learning, which help capture the multi-scale and contextual dependencies critical to fine-grained discrimination;
  • Multimodal fusion (e.g., combining visual and textual cues), useful in applications like product catalog matching or e-commerce;
  • Explainable AI (XAI) approaches for interpreting model decisions in high-stakes or user-facing applications.

We also welcome research on dataset creation, benchmark challenges, and performance evaluation methodologies tailored to fine-grained problems.

This Special Issue provides a forum for researchers, engineers, and practitioners to present their latest results, share new tools and datasets, and contribute to the collective advancement of fine-grained image understanding powered by modern machine learning.

Dr. Grzegorz Sarwas
Guest Editor

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 250 words) can be sent to the Editorial Office for assessment.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Information is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • fine-grained object recognition
  • prototype and proxy learning
  • domain adaptation in computer vision
  • knowledge distillation
  • few-shot and metric learning

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • Reprint: MDPI Books provides the opportunity to republish successful Special Issues in book format, both online and in print.

Further information on MDPI's Special Issue policies can be found here.

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

15 pages, 3148 KB  
Article
A Cross-Scale Feature Fusion Method for Effectively Enhancing Small Object Detection Performance
by Yaoxing Kang, Yunzuo Zhang, Yaheng Ren and Yu Cheng
Information 2026, 17(1), 25; https://doi.org/10.3390/info17010025 - 31 Dec 2025
Viewed by 508
Abstract
Deep learning-based industrial product surface defect detection methods are replacing manual inspection, while the issue of small object detection remains a key challenge in the current field of surface defect detection. The feature pyramid structures demonstrate great potential in improving the performance of [...] Read more.
Deep learning-based industrial product surface defect detection methods are replacing manual inspection, while the issue of small object detection remains a key challenge in the current field of surface defect detection. The feature pyramid structures demonstrate great potential in improving the performance of small object detection and are one of the important current research directions. Nevertheless, traditional feature pyramid networks still suffer from problems such as imprecise focus on key features, insufficient feature discrimination capabilities, and weak correlations between features. To address these issues, this paper proposes a plug-and-play guided focus feature pyramid network, named GF-FPN. Built on the foundation of FPN, this network is designed with a bottom-up guided aggregation network (GFN): through a lightweight pyramidal attention module (LPAM), star operation, and residual connections, it establishes correlations between objects and local contextual information, as well as between shallow-level details and deep-level semantic features. This enables the feature pyramid network to focus on key features, enhance the ability to distinguish between objects and backgrounds, and thereby improve the model’s small object detection performance. Experimental results on the self-built TinyIndus dataset and NEU-DET demonstrate that the detection model based on GF-FPN exhibits more competitive advantages in object detection compared to existing models. Full article
(This article belongs to the Special Issue Machine Learning in Image Processing and Computer Vision)
Show Figures

Graphical abstract

18 pages, 24741 KB  
Article
Cross-Domain Residual Learning for Shared Representation Discovery
by Baoqi Zhao, Jie Pan, Zhijie Zhang and Fang Yang
Information 2025, 16(10), 852; https://doi.org/10.3390/info16100852 - 2 Oct 2025
Cited by 1 | Viewed by 819
Abstract
In order to solve the problem of inconsistent data distribution in machine learning, domain adaptation based on feature representation methods extracts features from the source domain, and transfers them to the target domain for classification. The existing feature representation-based methods mainly solve the [...] Read more.
In order to solve the problem of inconsistent data distribution in machine learning, domain adaptation based on feature representation methods extracts features from the source domain, and transfers them to the target domain for classification. The existing feature representation-based methods mainly solve the problem of inconsistent feature distribution between the source domain data and the target domain data, but only a few methods analyze the correlation of cross-domain features between the original space and shared latent space, which reduces the performance of domain adaptation. To this end, we propose a domain adaptation method with a residual module, the main ideas of which are as follows: (1) transfer the source domain data features to the target domain data through the shared latent space to achieve features sharing; (2) build a cross-domain residual learning model using the latent feature space as the residual connection of the original feature space, which improves the propagation efficiency of features; (3) use a regular feature space to sparse feature representation, which can improve the robustness of the model; and (4) give an optimization algorithm, and the experiments on the public visual datasets (Office31, Office-Caltech, Office-Home, PIE, MNIST-UPS, COIL20) results show that our method achieved 92.7% accuracy on Office-Caltech and 83.2% on PIE and achieved the highest recognition accuracy in three datasets, which verify the effectiveness of the method. Full article
(This article belongs to the Special Issue Machine Learning in Image Processing and Computer Vision)
Show Figures

Figure 1

Back to TopTop