Neural Networks and Deep Learning in Computer Vision

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Electronic Multimedia".

Deadline for manuscript submissions: 31 July 2024 | Viewed by 2158

Special Issue Editors


E-Mail Website
Guest Editor
Faculty of Telecommunications, Technical University of Sofia, 1000 Sofia, Bulgaria
Interests: computer vision; 3D image and video analysis and reconstruction; Bayesian deep learning; probabilistic models for machine learning and computer vision

E-Mail Website
Guest Editor
Faculty of Telecommunications, Technical University of Sofia, 1000 Sofia, Bulgaria
Interests: computer vision; virtual and mixed reality; tactile internet

E-Mail Website
Guest Editor
Department of Electronics, Telecomunications and Electronics, Institute of Electronics and Informatics Engineering of Aveiro (IEETA), University of Aveiro, 3810-193 Aveiro, Portugal
Interests: machine learning; computer vision; system theory; signal processing

Special Issue Information

Dear Colleagues,

There is no question that the driving force in computer vision in recent years has been the rapid development of deep learning, and neural networks in particular. Applications like autonomous driving, disease diagnosis, text-to-image generators, and many more were unimaginable just ten years ago. The ability of neural networks to model and discover complex dependencies hidden in data, and to mimic the way the human brain works, paves the way for these, and many more, applications into our daily lives. Many practical open questions in computer vision, on the other hand, stimulate researchers to seek new and better approaches in deep learning and neural networks. Despite the optimistic view, many challenges remain. The recent development of the Forward-Forward algorithm is one such instance, which raises the question of whether backpropagation will still be useful in the future. Typical problems in computer vision include determining if a specific neural network architecture or training approach is appropriate for a given application, such as the optimal deep learning approach for neural radiation fields. The metaverse creates a whole new set of computer-vision-related issues, including the need for specialized neural networks that can execute in real time and provide a realistic experience to the user.

This Special Issue aims to provide a platform for original contributions related to emerging deep learning architectures for computer vision as well as computer vision applications, opening the door for new neural network architectures. Research topics in this Special Issue include, but are not limited to:

  • Emerging neural network architectures with applications to computer vision, e.g., FocalNet, declarative neural networks, diffusion models, graph-convolutional neural networks;
  • Bayesian deep learning for computer vision;
  • Self-supervised/contrastive/generative learning;
  • Semantically guided analysis for image and video, e.g., semantically guided compression/text-to-image/analysis;
  • Neural radiance fields and 3D reconstruction from single or multiple views;
  • Neural models for compressible, animatable, relightable 3D human face and body models;
  • Three-dimensional image and video analysis: segmentation, recognition, motion analysis and prediction.

Dr. Krasimir Tonchev
Dr. Agata Manolova
Prof. Dr. Petia Georgieva
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • computer vision
  • emerging neural network architectures
  • self-supervised and contrastive learning
  • compressible and animatable neural models of 3D human face and body
  • semantically guided computer vision

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Review

13 pages, 3147 KiB  
Article
An LCD Detection Method Based on the Simultaneous Automatic Generation of Samples and Masks Using Generative Adversarial Networks
by Hao Wu, Yulong Liu and Youzhi Xu
Electronics 2023, 12(24), 5037; https://doi.org/10.3390/electronics12245037 - 18 Dec 2023
Viewed by 795
Abstract
When applying deep learning methods to detect micro defects on low-contrast LCD surfaces, there are challenges related to imbalances in sample datasets and the complexity and laboriousness of annotating and acquiring target image masks. In order to solve these problems, a method based [...] Read more.
When applying deep learning methods to detect micro defects on low-contrast LCD surfaces, there are challenges related to imbalances in sample datasets and the complexity and laboriousness of annotating and acquiring target image masks. In order to solve these problems, a method based on sample and mask auto-generation for deep generative network models is proposed. We first generate an augmented dataset of negative samples using a generative adversarial network (GAN), and then highlight the defect regions in these samples using the training method constructed by the GAN to automatically generate masks for the defect images. Experimental results demonstrate the effectiveness of our proposed method, as it can simultaneously generate liquid crystal image samples and their corresponding image masks. Through a comparative experiment on the deep learning method Mask R-CNN, we demonstrate that the automatically obtained image masks have high detection accuracy. Full article
(This article belongs to the Special Issue Neural Networks and Deep Learning in Computer Vision)
Show Figures

Figure 1

Review

Jump to: Research

31 pages, 1853 KiB  
Review
Taxonomy and Survey of Current 3D Photorealistic Human Body Modelling and Reconstruction Techniques for Holographic-Type Communication
by Radostina Petkova, Ivaylo Bozhilov, Desislava Nikolova, Ivaylo Vladimirov and Agata Manolova
Electronics 2023, 12(22), 4705; https://doi.org/10.3390/electronics12224705 - 19 Nov 2023
Viewed by 1001
Abstract
The continuous evolution of video technologies is now primarily focused on enhancing 3D video paradigms and consistently improving their quality, realism, and level of immersion. Both the research community and the industry work towards improving 3D content representation, compression, and transmission. Their collective [...] Read more.
The continuous evolution of video technologies is now primarily focused on enhancing 3D video paradigms and consistently improving their quality, realism, and level of immersion. Both the research community and the industry work towards improving 3D content representation, compression, and transmission. Their collective efforts culminate in the striving for real-time transfer of volumetric data between distant locations, laying the foundation for holographic-type communication (HTC). However, to truly enable a realistic holographic experience, the 3D representation of the HTC participants must accurately convey the real individuals’ appearance, emotions, and interactions by creating authentic and animatable 3D human models. In this regard, our paper aims to examine the most recent and widely acknowledged works in the realm of 3D human body modelling and reconstruction. In addition, we provide insights into the datasets and the 3D parametric body models utilized by the examined approaches, along with the employed evaluation metrics. Our contribution involves organizing the examined techniques, making comparisons based on various criteria, and creating a taxonomy rooted in the nature of the input data. Furthermore, we discuss the assessed approaches concerning different indicators and HTC. Full article
(This article belongs to the Special Issue Neural Networks and Deep Learning in Computer Vision)
Show Figures

Figure 1

Back to TopTop