Next Article in Journal
GAH-TNet: A Graph Attention-Based Hierarchical Temporal Network for EEG Motor Imagery Decoding
Previous Article in Journal
TARPγ2-Derived Peptide Enhances Early-Phase Long-Term Potentiation and Impairs Memory Retention in Male Rats
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reproducing the Few-Shot Learning Capabilities of the Visual Ventral Pathway Using Vision Transformers and Neural Fields

by
Jiayi Su
1,
Lifeng Xing
1,
Tao Li
1,
Nan Xiang
1,
Jiacheng Shi
1 and
Dequan Jin
1,2,*
1
School of Mathematics and Information Science, Guangxi University, Nanning 530004, China
2
Center for Applied Mathematics of Guangxi, Guangxi University, Nanning 530004, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2025, 15(8), 882; https://doi.org/10.3390/brainsci15080882
Submission received: 16 July 2025 / Revised: 14 August 2025 / Accepted: 18 August 2025 / Published: 19 August 2025

Abstract

Background: Studies have shown that humans can rapidly learn the shape of new objects or adjust their behavior when encountering novel situations. Research on visual cognition in the brain further indicates that the ventral visual pathway plays a critical role in core object recognition. While existing studies often focus on microscopic simulations of individual neural structures, few adopt a holistic, system-level perspective, making it difficult to achieve robust few-shot learning capabilities. Method: Inspired by the mechanisms and processes of the ventral visual stream, this paper proposes a computational model with a macroscopic neural architecture for few-shot learning. We reproduce the feature extraction functions of V1 and V2 using a well-trained Vision Transformer (ViT) and model the neuronal activity in V4 and IT using two neural fields. By connecting these neurons based on Hebbian learning rules, the proposed model stores the feature and category information of the input samples during support training. Results: By employing a scale adaptation strategy, the proposed model emulates visual neural mechanisms, enables efficient learning, and outperforms state-of-the-art few-shot learning algorithms in comparative experiments on real-world image datasets, demonstrating human-like learning capabilities. Conclusion: Experimental results demonstrate that our ventral-stream-inspired machine-learning model achieves effective few-shot learning on real-world datasets.

1. Introduction

When humans encounter novel situations, the brain can quickly learn the features of new stimuli and adjust behavior accordingly. Inspired by the biological theories underlying this process in the human brain, researchers have proposed a series of brain-based models. These models are capable of achieving functions similar to those of the nervous system or its learning mechanisms. Among them, neural networks are currently one of the most popular brain-inspired models in machine learning. Neural networks consist of multiple interconnected neurons or units [1,2,3,4,5]. After years of development, the latest achievement in neural networks is known as deep learning. Compared to traditional neural networks, deep learning models typically have dozens to thousands of layers and exhibit more complex structures. With the emergence of efficient training methods, deep neural networks have demonstrated powerful learning capabilities and achieved remarkable results in image classification and related fields. Traditional machine learning methods require large amounts of labeled data for training, but in real-world scenarios, collecting such large-scale datasets can be extremely costly. However, neural networks based on backpropagation suffer from several limitations [6,7,8,9,10], such as high computational cost, poor model interpretability (the black-box problem), lack of biological plausibility, and strong dependence on large datasets [11]. These challenges have diminished their advantages when applied to few-shot learning tasks [12]. In daily life, humans learn and adapt to new things based on prior knowledge. Inspired by this, researchers have begun exploring whether knowledge acquired from core tasks can be transferred directly to related target tasks, thus overcoming the challenges neural networks face in few-shot learning [13,14].
For example, Peng et al. proposed the SgVA-CLIP method, which addresses the issue that vision-language pre-trained models may neglect detailed visual information when distinguishing between different images [15]. Bendou et al. introduced a method called EASY, which provides a clearer and more intuitive understanding to help people execute few-shot learning more effectively, while adding almost no hyperparameters or additional parameters to the training of initial deep learning models [16]. Zhang et al. proposed MetaQDA, a Bayesian meta-learning framework that extends the classic quadratic discriminant analysis (QDA) [17]. Hu et al. proposed the P>M>F method, a simple yet effective pipeline for few-shot learning [18]. Zhang et al. approached few-shot image classification from the perspective of optimal matching between image regions [19]. Rong et al. proposed ESPT (Episode-Based Self-Paced Training), which leverages a self-supervised learning framework to generate pre-training tasks for learning features without labels [20]. It helps the model better understand and utilize the structural information of images when facing new tasks, thereby enhancing its adaptability to new tasks and reducing performance degradation due to insufficient samples by incorporating this self-supervised pre-training task into few-shot learning tasks. Shalam et al. introduced the BPA method in 2024, aiming to enhance the feature representation of input items for downstream matching- or grouping-related tasks [21]. This method captures the complex relationships between features, provides a flexible and effective way to improve existing models, and achieves state-of-the-art results in various tasks. Fifty et al. proposed a method called CAML (Context-Aware Meta-Learning), which enhances adaptability to new tasks by incorporating contextual information [22]. CAML primarily addresses the issue of task diversity in meta-learning by introducing contextual information in task representation, effectively guiding the model to adapt to new tasks. Song et al. proposed a Transformer-based few-shot classification method that integrates multi-scale feature representations and a multi-scale matching strategy, significantly enhancing the model’s ability to understand objects of varying scales and improving classification performance [23].
Among the above-mentioned methods, some utilize unlabeled sample information to generate pseudo-labels or prototypes, thereby increasing the number of samples for model training. Others adopt transfer learning, applying knowledge obtained through pre-training on large datasets in the source task to the target task, thereby improving the few-shot learning of the model. These methods effectively enhance the performance of models that originally rely on large-scale samples for training in few-shot classification tasks. Furthermore, if we can develop a new model that is inherently suitable for few-shot learning, we may achieve even better few-shot image classification.
The learning process in humans requires the coordinated operation of the entire nervous system, whereas current neural network models can only achieve certain functions at a partial or microscopic level. Since humans are capable of few-shot learning, we aim to implement a more holistic, flexible, and multi-scale neural framework to realize few-shot learning mechanisms. Studies have shown that different brain regions respond differently to various input signals [24,25]. The brain determines the type of input signal based on the state of neural fields [26]. Inspired by this biological process, we attempt to design a few-shot learning model based on neural fields. This approach is interpretable from both biological and mathematical perspectives. In recent years, with advances in brain research technologies [27,28], we have gained a clearer understanding of the brain’s structure [29].
Inspired by the overall mechanism of the ventral visual stream [30,31,32], we designed a classifier for few-shot learning based on the neural field model. By coupling this classifier with the feature extraction part of the Vision Transformer (ViT) [4], we can compensate for ViT’s shortcomings in few-shot learning. Our model achieves experimental results that surpass those of the current state of the art on standard datasets while maintaining high efficiency and low cost.

2. Preliminaries

2.1. Few-Shot Learning

Few-shot learning aims to develop methods that can learn from a small number of labeled samples and generalize to a large number of unlabeled samples [33,34,35]. In few-shot learning, we typically use the N-way K-shot setting, where N denotes the number of classes and K represents the number of labeled images per class. Thus, a total of N × K labeled images are used during the training phase of few-shot learning.

2.2. The Ventral Visual Stream

As shown in Figure 1, the part of the human brain that is associated with few-shot learning is the main pathway of the ventral visual stream. This pathway primarily includes the retina, lateral geniculate nucleus (LGN), and cortical regions V1, V2, V4, and IT. Based on their functional roles, we divide the main pathway of the ventral visual stream into two parts: the feature extraction part—comprising the retina, LGN, and cortical areas V1 and V2—and the memory and learning part, which includes cortical areas V4 and IT. Before simulating this pathway, let us first examine how it biologically accomplishes image recognition and learning.
As the starting point of the pathway, the retina is responsible for receiving external visual stimuli [36]. The cone cells and rod cells in the retina convert these stimuli into neural signals. Rod cells are sensitive to light and enable vision under low-light conditions, while cone cells are responsible for color perception and high-resolution vision. The information generated by these two types of cells is integrated within the retina and then transmitted to the lateral geniculate nucleus (LGN) [37]. The LGN performs initial processing of visual information, such as detecting spatial orientation, motion direction, and speed. It has six layers, with layers 1, 4, and 6 processing input from one eye, and layers 2, 3, and 5 processing input from the other. The LGN fragments the image into patches and labels each fragment before transmitting them simultaneously to the primary visual cortex (V1).
The primary visual cortex (V1) and secondary visual cortex (V2) act as filters and feature extractors [38,39]. V1 detects basic visual features such as brightness, edges, and textures. Neurons in V1 respond selectively to specific orientations and spatial frequencies and can distinguish textures of varying thicknesses as well as different wavelengths of light. V2 contains neurons specialized in processing visual information at different scales and color features. It integrates the basic visual features from V1, enabling the rapid detection of important environmental information, and further builds depth perception before passing the information to V4.
The V4 and IT (inferotemporal) cortex regions are primarily responsible for pattern learning and memory [24,25,26]. Different areas within V4 respond differently to various patterns. Importantly, neurons in these regions do not respond only to one specific type of stimulus but rather show varying degrees and types of activation depending on the pattern. For example, neurons responsive to curved lines may also be activated to some extent by right-angled patterns. These responses are stored by neurons in the IT region. When similar stimuli enter the IT cortex again, it determines the type of input based on which neurons are activated and to what extent.
This is the biological process behind visual recognition in the human brain. Inspired by this mechanism, we propose a neural field model in this paper to address few-shot learning tasks.

3. Methodology

Inspired by the ventral visual stream, in this section, we design a classifier for few-shot learning, as shown in Figure 2. We introduce a new neural field-based architecture for classification and provide some improvement methods to expand its applicability and computational efficiency.

3.1. Architecture of Two Fields

We employ two discrete neural fields to form a classifier. One of them is called the elementary field for feature representation, and the other one is called the high-level field for category representation. They are defined by the following equations:
τ u u ˙ i ( t ) = η k = 1 m ω u ( z i z k ) ϕ u k ( t ) + e u , i u i ( t ) , i = 1 , 2 , , m .
and
τ v v ˙ j ( t ) = η k = 1 m c ω v ( z ˜ j z ˜ k ) ϕ v k ( t ) + ϕ i = 1 m w j , i ϕ u i ( t ) + e v , j v j ( t ) , j = 1 , 2 , , n .
z i Z , i = 1 , 2 , , m are data samples for training. n is the number of classes. We suppose z ˜ j = 1 l j i = 1 l j z ˜ j i are class positions, i.e., the average positions of all their samples. The functions u i ( t ) and v j ( t ) describe the activation behaviors of the ith elementary neuron and the jth high-level neuron. l j is the number of samples in the jth class. τ f and τ g are positive evolution rates. ϕ ( · ) is a monotonically increasing, non-negative, and bounded activation function given by
ϕ ( u ) = 1 exp ( u ) , u > 0 0 , u 0 .
The terms
k = 1 m ω u ( z i z k ) ϕ u k ( t )
and
k = 1 m c ω v ( z ˜ j z ˜ k ) ϕ v k ( t )
describe the lateral interactions of the elementary and high-level fields.
ω u ( z ˜ ) = A exp 1 2 z ˜ . / σ u 1 2 B exp 1 2 z ˜ . / σ u 2 2 ,
ω v ( z ˜ ) = A exp 1 2 z ˜ . / σ v 1 2 B exp 1 2 z ˜ . / σ v 2 2 ,
where
A = 1 2 π σ u 1 , B = 1 2 π σ u 2 ,
and
σ u 2 = 3 σ u 1 .
We have A B = 1 . Thus, A = 3 2 and B = 1 2 , so we get that
ω u ( z ˜ ) = 3 2 exp 1 2 z ˜ . / σ u 1 2 1 2 exp 1 2 z ˜ . / σ u 2 2 ,
ω v ( z ˜ ) = 3 2 exp 1 2 z ˜ . / σ v 1 2 1 2 exp 1 2 z ˜ . / σ v 2 2 .
By calculating zero points of the equations above, we get z ˜ . = 3 l n 3 2 σ u 1 . η is the threshold function defined by
η ( u ) = 1 exp ( u ) , u > 0 1 + exp ( u ) , u 0 .
ω u ( · ) and ω v ( · ) are interaction kernels with a “Mexican hat” shape. e u , i and e v , i are binary input signals. w j , i is the connection weight between the ith neuron in the elementary field and the jth neuron in the high-level field.

3.2. Approximation of the Static Solution of Neural Fields

Since neural fields are dynamic, solving them consumes too much computation when the data dimensions are high. From observations in real-world classification tasks, we need to focus on the activation status of neurons in a neural field, but not on their specific values. Therefore, we do not have to solve the neural field equations with high accuracy. Additionally, when considering the classification problem of static data, the static solutions of the neural fields can be approximated by the convolution of the external input with an interaction kernel. Consequently, we can use the following two equations to approximate the solution of the original dynamic neural field equation for an external input e f , k :
u i = η ω u ( z i z k ) ϕ ( e u , k ) ,
and
v j = η k = 1 m c ω v ( z ˜ j z ˜ k ) ϕ i = 1 m x k , i ϕ ( v i ) + ϕ i = 1 m w j , i ϕ ( u i ) .
When the data classes are distant, we can further suppose that the lateral interaction between their corresponding high-level neurons is weak and ignore the lateral interaction of the high-level field as follows:
v j = ϕ i = 1 m w j , i ϕ η ω u ( z i z k ) ϕ ( e u , k ) , j = 1 , 2 , , n .
These are much simpler and require much less computation.

3.3. Feature Extraction and Preprocessing

Suppose that L = { Z 1 , Z 2 , , Z m } is a support set, and Z i , i = 1 , 2 , , m are sample images. Images are highly unstructured data. Classification accuracy may be low if we input them into the model without feature extraction. To address this issue, we employ the feature extraction module from the front part of the Vision Transformer (ViT) to obtain image features as follows:
x i = F n n ( Z i ) , i = 1 , 2 , , m .
The extracted feature vector x i is high-dimensional and may contain information that is not useful for classification. This redundant information can reduce classification accuracy. Therefore, we employ some dimensionality reduction methods to preprocess them in the following way:
z i = d r n n ( x i ) , i = 1 , 2 , , m .
Let Z = { z 1 , z 2 , , z m } . We use a deep neural network pre-trained on the source task and keep it frozen during training on the target dataset.

3.4. Training and Prediction

3.4.1. Training Phase

As we can see in Figure 3, the training process of our classifier is as follows: The training sample enters the primary neural field, which then establishes a connection with the high-level neural field based on the sample and its label. When a test sample is input into the primary neural field, the classifier determines the category by observing the activation state of neurons in the high-level neural field.
Based on the model structure we previously discussed, it is clear that the training process of the model is essentially about establishing a connection weight matrix W that links two layers of neural fields. When a labeled sample x is fed into the model, we assume that its feature vector and label activate the corresponding neurons in the primary and high-level neural fields. As a result, a connection is formed, and we set the connection weight to 1. Here, we have
w j , p = 1 x ^ p = x ^ ( j ) 0 else

3.4.2. Adaptation of the Scale Parameter σ

Previously, we established the relationship between our scale parameter σ and the distance z, so we now focus on how to determine the value of z. Our few-shot learning tasks are specifically divided into two types: 5-way 1-shot and 5-way 5-shot. According to the definition of few-shot learning, we also categorize the selection of z into two corresponding cases. First, when there is only one sample per class, we compute the distance between samples from two classes and set z to be half of that distance. In contrast, when the support set contains more than one sample per class, we compute the distances between different samples within the same class and set z as the maximum of these distances. Subsequently, according to the relationship between z and σ described earlier, we can set the initial value of σ as
σ u 1 = 2 3 l n 3 z .
According to the 3-sigma rule, we set σ m a x and σ m i n as follows:
σ m a x = 3 σ u 1 , σ m i n = σ u 1 3 .
During the prediction process, our model may encounter three scenarios: only one neuron in the high-level neural field is activated, multiple neurons are activated, or no neurons are activated, as shown in Figure 4.
The first scenario is the ideal case, where we can directly assign the category corresponding to the activated neuron to the sample. The other two scenarios indicate that there is an issue with the setting of our range parameter, which needs to be adjusted. Based on each situation, we apply the following adjustments, shown as Algorithm1. Before making any adjustments, we define a constant λ less than 1.
When no neurons are activated, we consider that the value of our range parameter is too small, resulting in no suitable neurons falling within the prediction range. Therefore, we enlarge the range parameter and perform prediction again using the updated value. We have:
σ m a x = max σ m a x , 1 λ σ u 1 , σ m i n = σ u 1 .
Then, we get
σ u 1 = σ m a x λ ( σ m a x σ m i n ) .
When multiple neurons are activated, we consider that the value of our range parameter is too large, allowing too many neurons to fall within the prediction range. Therefore, we reduce the range parameter and perform prediction again using the updated value. We have
σ m i n = min σ m i n , λ σ u 1 .
According to the above two formulas, we have the following range parameter:
σ f 1 = σ m a x + λ ( σ m a x σ m i n ) .
Algorithm 1 Scale Adaptation algorithm
1:
Input:  n u m , σ u 1 , λ
2:
Output:  σ u 1
3:
σ m a x = 3 σ u 1 , σ m i n = σ u 1 3
4:
while  n u m 1  do
5:
      if  n u m = 0  then
6:
         Let σ m a x = max { σ m a x , 1 λ σ u 1 }
7:
         Calculate σ u 1 = σ m i n + λ ( σ m a x σ m i n )
8:
      else
9:
         Let σ m i n = min { σ m i n , λ σ u 1 }
10:
          Calculate σ u 1 = σ m a x λ ( σ m a x σ m i n )
11:
     end if
12:
end while

4. Results

To demonstrate the proposed model, we compared the few-shot image classification accuracy of our model with that of other state-of-the-art models on three different datasets and studied the influence of different data dimensionality reduction methods, encoders, and feature extractors on the model through ablation experiments.

4.1. Datasets and Experimental Settings

In this experiment, we used the following datasets: Caltech-UCSD Birds-200-2011 (CUB200-2011) [40], CIFAR100 Few Shots (CIFAR-FS) [41], and miniImageNet [42], as shown in Table 1. They are summarized as follows:
CUB200-2011: CUB200-2011 is a benchmark image dataset for fine-grained classification and recognition research. The dataset contains 11,788 images of birds across 200 subclasses.
CIFAR-FS: The CIFAR-FS dataset (full name: the CIFAR100 Few Shots dataset) is derived from the CIFAR100 dataset. It contains 100 categories, with 600 images per category (60,000 images in total), each of size 32 × 32 .
miniImageNet: miniImageNet, derived from ImageNet, is intended for meta-learning and few-shot learning studies. It contains 60,000 84 × 84 color images across 100 categories, with 600 images per category.
Inspired by the biological structure shown in Figure 1, the overall structure of the proposed model is shown in Figure 2 and is divided into three modules: image preprocessing, feature extraction and dimensionality reduction, and the classifier. Given the limited number of labeled samples in few-shot learning, different methods of feature extraction can significantly affect the results. The classifier module, which is primarily responsible for learning and memorization tasks, plays a crucial role in few-shot learning tasks, as illustrated in Figure 3. Moreover, the adaptive component added to the module (see Figure 4) can further improve the experimental results.
Table 1. Number of classes and instances contained in the datasets used in this study.
Table 1. Number of classes and instances contained in the datasets used in this study.
CUB200-2011CIFAR-FSminiImageNet
Instances11,78860,00060,000
Classes200100100
CUB-200-2011: a benchmark dataset for fine-grained image classification and recognition; CIFAR-FS: a few-shot learning dataset derived from CIFAR-100; miniImageNet: a subset of ImageNet commonly used for meta-learning and few-shot learning.
We used the same image processing method as [43]. Our ViT-L/16 model was pre-trained on the I21K dataset for ten epochs [4]. For dimensionality reduction, we employed LE (Laplacian Eigenmaps) [44] as d r n n to reduce the dimensionality to 4, with the number of neighbors in KNN for LE set to 99. On the miniImageNet and CIFAR-FS datasets, we used 80 classes for training and the remaining 20 for testing. On the CUB200-2011 dataset, we used 150 categories for training and the remaining 50 for testing.

4.2. Comparison with State-of-the-Art Models

We conducted 5-way 1-shot and 5-way 5-shot tasks for the proposed algorithm on CUB200-2011, CIFAR-FS, and miniImageNet. We compared the experimental results with those of EASY [16], ESPT [20], MetaQDA [17], PMF-BPA [21], SgVA-CLIP [15], P > M > F [18], and CAML [22]. The results are shown in Table 2, with the best results highlighted in bold and the second-best results underlined. For clarity, we use 5-1 and 5-5 in the tables to represent the 5-way 1-shot and 5-way 5-shot tasks, respectively. Additionally, we use CUB, CIFAR-FS, and MINI to represent CUB200-2011, CIFAR-FS, and miniImageNet.
The results in Table 2 show that overall, our method outperformed the comparison methods. We can see that on the CUB dataset, the performance of our method was slightly lower than that of the comparison methods. When we conducted the 5-way 1-shot task on this dataset, the result we obtained was 0.9242, which was second only to the 0.9580 achieved by PMF-BPA. However, in the 5-way 5-shot task, our model achieved a result of 0.9524, which was 1% lower than that of the other methods using the same backbone. Nevertheless, on the CIFAR-FS and miniImageNet datasets, the results obtained by our model were significantly higher than those of the comparison methods. On CIFAR-FS, the results we achieved for the 5-way 1-shot and 5-way 5-shot tasks were 0.9811 and 0.9846, which were much higher than those of the second-best model, PMF-BPA, which achieved results of 0.8710 and 0.9470. Lastly, on the miniImageNet dataset, our model obtained accuracies of 0.9887 and 0.9902. In the 5-way 1-shot task, the accuracy of the proposed model was at least 1 % higher than that of the second-best model, while in the 5-way 5-shot task, it surpassed the SgVA-CLIP benchmark of 0.9872 by 0.3 % . Although the accuracy of our model on the CUB dataset was not as high as that of the others, considering the overall performance on the three datasets, we believe that our proposed model is still superior to the comparison models.

4.3. Ablation Experiments

4.3.1. Comparison with Different Distance Metrics

The distance metric may influence accuracy. We investigated the accuracy of the LE method with the Euclidean distance, correlation distance, cosine distance, and Minkowski distance. We present the results in Table 3. The best results are highlighted in bold, and the second-best ones are underlined.
The results obtained with the cosine distance on these three datasets were 0.9242, 0.9524, 0.9811, 0.9846, 0.9887, and 0.9902. In comparison, the second-best results obtained using the other distance metrics for dimensionality reduction were 0.9125, 0.9494, 0.9785, 0.9816, 0.9873, and 0.9899. Therefore, we can conclude that the cosine distance achieved the highest accuracy in each task on every dataset.

4.3.2. Comparison with Different Feature Extractors and Encoders

To demonstrate the superiority of the classifier in our model, we compared it with KNN and SVM on the CUB200-2011, CIFAR-FS, and miniImageNet datasets, using the same feature extractor and dimensionality reduction method. Additionally, to eliminate the influence of the feature extractor, we employed ViT and ResNet18 as the feature extractors. The experimental results are presented in Table 4, with the best results highlighted in bold, and the second-best results underlined.
As shown in Table 4, in both the 5-way 1-shot and 5-way 5-shot tasks on the CUB, CIFAR-FS, and miniImageNet datasets, the accuracy of our proposed model was at least 1% higher than that of KNN when using the same feature extractor for the same task. Similarly, Table 4 shows that the accuracy of SVM was consistently lower than that of our model across all three datasets. Notably, in the 5-way 1-shot tasks, the accuracy of SVM was nearly 20% lower than that of our model on various datasets. At the same time, we can see from the results in Table 4 that different feature extractors impacted the experimental results. We noticed that for the same few-shot learning tasks on the same dataset, using ViT as the feature extractor yielded results that were 8–30% higher than those obtained using ResNet18 as the feature extractor.

5. Discussion

In this study, inspired by the physiological structure of the ventral visual stream, we designed a classifier module capable of few-shot learning based on the neural field equation. We coupled this classifier module with the feature extraction part of the Vision Transformer (ViT) to compensate for ViT’s shortcomings in few-shot learning. Figure 1 illustrates the biological structure of the ventral visual stream, which mainly includes the retina, lateral geniculate nucleus, and cortical areas V1, V2, V4, and IT. Given the close association of this brain region with few-shot learning capabilities, we attempted to design a classifier module suitable for few-shot learning tasks based on this structure. The datasets summarized in Table 1 are internationally recognized benchmark datasets for evaluating few-shot learning methods. This motivated our biologically inspired design, aiming to provide a novel solution to few-shot learning challenges.
As shown in Figure 2, the few-shot learning model we designed is divided into three main parts: image preprocessing, feature extraction and dimensionality reduction, and the classifier. Since the number of labeled samples in few-shot learning is limited, the feature extraction module plays a crucial role in the entire model structure. After obtaining image features from the feature extraction module, the primary neural field of the classifier module memorizes the labeled samples and their categories, thereby improving the accuracy of the classifier’s predictions for the categories of test samples. The comparative results in Table 2 demonstrate that our method consistently outperforms baseline models across multiple datasets, confirming the benefits of integrating biological principles with modern deep architectures. From the overall model architecture shown in Figure 2, it can be seen that feature extraction and dimensionality reduction are key components that significantly impact the experimental results. Therefore, Table 3 primarily demonstrates how different distance parameters used for dimensionality reduction under the same conditions affect the experimental outcomes.
Figure 3a shows the training process of the classifier, which combines supervised learning and meta-learning. Supervised learning ensures that the model can accurately classify examples, while meta-learning helps the model quickly adapt to new tasks. This training process is particularly effective for few-shot learning because it enables the model to generalize well from a small number of examples. The results in Table 2 indicate that this training process is highly effective in enhancing the model’s performance in few-shot learning tasks. Figure 3b illustrates the prediction process of our model. This process involves invoking the features learned during training and matching the features of test samples with the learned features for prediction. The model’s ability to extract features from a small number of samples and generalize is crucial for few-shot learning, and our classifier module is designed specifically to achieve this capability.
Figure 4 shows the adjustments made to the range-perception parameters in the model during the experimental process based on the prediction results. When the model fails to detect activated neurons, it indicates that the range-perception parameters are set too small; conversely, when the model detects multiple activated neurons, it suggests that the range-perception parameters are set too large. Both situations can significantly impact the experimental results. Therefore, we added a self-regulation module to the model to adjust the range-perception parameters. By introducing this module, our experimental results were further improved and stabilized. To better demonstrate the stability of the experimental results, the results shown in Table 2 are the averages of 600 experiments, which effectively illustrate that our model can achieve high precision and stability in few-shot learning tasks.
Although our model achieves high precision on real-world datasets, its application results on medical images are still not satisfactory. Therefore, in future work, we will further improve and optimize this model.

Author Contributions

Conceptualization, D.J. and J.S. (Jiayi Su); methodology, J.S. (Jiayi Su); software, L.X. and J.S. (Jiacheng Shi); validation, J.S. (Jiayi Su); investigation, T.L. and N.X.; writing—original draft preparation, J.S. (Jiayi Su); writing—review and editing, J.S. (Jiayi Su), L.X., and T.L.; supervision, D.J.; project administration, D.J.; funding acquisition, D.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Guangxi under grant 2025GXNSFAA069486, the National Key R&D Program of China under grant 2021YFA1003004, the National Natural Science Foundation of China under grant 12031003, and the special foundation for Guangxi Ba Gui Scholars.

Data Availability Statement

The original data presented in the study are openly available in GitHub at https://github.com/Sue214313/FSL, accessed on 16 July 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ViTVision Transformer
LGNLateral geniculate nucleus
V1Primary visual cortex
V2Secondary visual cortex
ITInferior temporal cortex

References

  1. Fang, W.; Chen, Y.; Ding, J.; Yu, Z.; Masquelier, T.; Chen, D.; Huang, L.; Zhou, H.; Li, G.; Tian, Y. Spikingjelly: An open-source machine learning infrastructure platform for spike-based intelligence. Sci. Adv. 2023, 9, eadi1480. [Google Scholar] [CrossRef]
  2. Zhao, X.; Wang, L.; Zhang, Y.; Han, X.; Deveci, M.; Parmar, M. A review of convolutional neural networks in computer vision. Artif. Intell. Rev. 2024, 57, 99. [Google Scholar] [CrossRef]
  3. Rahman, M.M.; Munir, M.; Marculescu, R. Emcad: Efficient multi-scale convolutional attention decoding for medical image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 11769–11779. [Google Scholar]
  4. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Virtual Event, 3–7 May 2021. [Google Scholar]
  5. Wang, A.; Chen, H.; Lin, Z.; Han, J.; Ding, G. Rep ViT: Revisiting Mobile CNN From ViT Perspective. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 15909–15920. [Google Scholar]
  6. Dampfhoffer, M.; Mesquida, T.; Valentian, A.; Anghel, L. Backpropagation-Based Learning Techniques for Deep Spiking Neural Networks: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11906–11921. [Google Scholar] [CrossRef]
  7. Pai, S.; Sun, Z.; Hughes, T.W.; Park, T.; Bartlett, B.; Williamson, I.A.D.; Minkov, M.; Milanizadeh, M.; Abebe, N.; Morichetti, F.; et al. Experimentally realized in situ backpropagation for deep learning in photonic neural networks. Science 2023, 380, 398–404. [Google Scholar] [CrossRef]
  8. Meng, Q.; Xiao, M.; Yan, S.; Wang, Y.; Lin, Z.; Luo, Z.Q. Towards Memory- and Time-Efficient Backpropagation for Training Spiking Neural Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 6143–6153. [Google Scholar]
  9. Wei, W.; Zhang, M.; Qu, H.; Belatreche, A.; Zhang, J.; Chen, H. Temporal-Coded Spiking Neural Networks with Dynamic Firing Threshold: Learning with Event-Driven Backpropagation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 10518–10528. [Google Scholar]
  10. Wright, L.G.; Onodera, T.; Stein, M.M.; Wang, T.; Schachter, D.T.; Hu, Z.; McMahon, P.L. Deep physical neural networks trained with backpropagation. Nature 2021, 601, 549–555. [Google Scholar] [CrossRef]
  11. Rani, S.; Kataria, A.; Kumar, S.; Tiwari, P. Federated learning for secure IoMT-applications in smart healthcare systems: A comprehensive review. Knowl.-Based Syst. 2023, 274, 110658. [Google Scholar] [CrossRef]
  12. Song, Y.; Wang, T.; Cai, P.; Mondal, S.K.; Sahoo, J.P. A comprehensive survey of few-shot learning: Evolution, applications, challenges, and opportunities. ACM Comput. Surv. 2023, 55, 1–40. [Google Scholar] [CrossRef]
  13. Niu, S.; Liu, Y.; Wang, J.; Song, H. A decade survey of transfer learning (2010–2020). IEEE Trans. Artif. Intell. 2021, 1, 151–166. [Google Scholar] [CrossRef]
  14. Zhu, Z.; Lin, K.; Jain, A.K.; Zhou, J. Transfer learning in deep reinforcement learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13344–13362. [Google Scholar] [CrossRef] [PubMed]
  15. Peng, F.; Yang, X.; Xiao, L.; Wang, Y.; Xu, C. SgVA-CLIP: Semantic-Guided Visual Adapting of Vision-Language Models for Few-Shot Image Classification. Trans. Multi. 2024, 26, 3469–3480. [Google Scholar] [CrossRef]
  16. Bendou, Y.; Hu, Y.; Lafargue, R.; Lioi, G.; Pasdeloup, B.; Pateux, S.; Gripon, V. Easy-Ensemble Augmented-Shot-Y-Shaped Learning: State-of-the-Art Few-Shot Classification with Simple Components. J. Imaging 2022, 8, 179. [Google Scholar] [CrossRef]
  17. Zhang, X.; Meng, D.; Gouk, H.; Hospedales, T. Shallow Bayesian Meta Learning for Real-World Few-Shot Recognition. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 631–640. [Google Scholar]
  18. Hu, S.X.; Li, D.; Stühmer, J.; Kim, M.; Hospedales, T.M. Pushing the Limits of Simple Pipelines for Few-Shot Learning: External Data and Fine-Tuning Make a Difference. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 9058–9067. [Google Scholar]
  19. Zhang, C.; Cai, Y.; Lin, G.; Shen, C. DeepEMD: Differentiable Earth Mover’s Distance for Few-Shot Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 5632–5648. [Google Scholar] [CrossRef]
  20. Rong, Y.; Lu, X.; Sun, Z.; Chen, Y.; Xiong, S. ESPT: A Self-Supervised Episodic Spatial Pretext Task for Improving Few-Shot Learning. Proc. AAAI Conf. Artif. Intell. 2023, 37, 9596–9605. [Google Scholar] [CrossRef]
  21. Shalam, D.; Korman, S. The balanced-pairwise-affinities feature transform. In Proceedings of the 41st International Conference on Machine Learning, JMLR.org, Vienna, Austria, 21–27 July 2024; ICML’24. [Google Scholar]
  22. Fifty, C.; Duan, D.; Junkins, R.G.; Amid, E.; Leskovec, J.; Re, C.; Thrun, S. Context-Aware Meta-Learning. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  23. Song, M.; Yao, F.; Zhong, G.; Ji, Z.; Zhang, X. Matching Multi-Scale Feature Sets in Vision Transformer for Few-Shot Classification. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 12638–12651. [Google Scholar] [CrossRef]
  24. Tang, R.; Song, Q.; Li, Y.; Zhang, R.; Cai, X.; Lu, H.D. Curvature-processing domains in primate V4. eLife 2020, 9, e57502. [Google Scholar] [CrossRef]
  25. Jiang, R.; Andolina, I.M.; Li, M.; Tang, S. Clustered functional domains for curves and corners in cortical area V4. eLife 2021, 10, e63798. [Google Scholar] [CrossRef] [PubMed]
  26. Bao, P.; She, L.; McGill, M.; Tsao, D.Y. A map of object space in primate inferotemporal cortex. Nature 2020, 583, 103–108. [Google Scholar] [CrossRef] [PubMed]
  27. Gordleeva, S.Y.; Tsybina, A.G.; Kazantsev, V.B.; Volgushev, M.; Zefirov, A.V.; Volterra, A. Modeling Working Memory in a Spiking Neuron Network Accompanied by Astrocytes. Front. Cell. Neurosci. 2021, 15, 631485. [Google Scholar] [CrossRef] [PubMed]
  28. Kim, R.; Sejnowski, T.J. Strong inhibitory signaling underlies stable temporal dynamics and working memory in spiking neural networks. Nat. Neurosci. 2020, 24, 129–139. [Google Scholar] [CrossRef]
  29. Xu, J.; Pan, Y.; Pan, X.; Hoi, S.; Yi, Z.; Xu, Z. RegNet: Self-Regulated Network for Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 9562–9567. [Google Scholar] [CrossRef]
  30. Fotiadis, P.; Parkes, L.; Davis, K.A.; Satterthwaite, T.D.; Shinohara, R.T.; Bassett, D.S. Structure–function coupling in macroscale human brain networks. Nat. Rev. Neurosci. 2024, 25, 688–704. [Google Scholar] [CrossRef]
  31. Bonnen, T.; Yamins, D.L.; Wagner, A.D. When the ventral visual stream is not enough: A deep learning account of medial temporal lobe involvement in perception. Neuron 2021, 109, 2755–2766.e6. [Google Scholar] [CrossRef]
  32. Xu, S.; Liu, X.; Almeida, J.; Heinke, D. The contributions of the ventral and the dorsal visual streams to the automatic processing of action relations of familiar and unfamiliar object pairs. NeuroImage 2021, 245, 118629. [Google Scholar] [CrossRef]
  33. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a Few Examples: A Survey on Few-shot Learning. ACM Comput. Surv. 2020, 53, 1–34. [Google Scholar] [CrossRef]
  34. Alayrac, J.B.; Donahue, J.; Luc, P.; Miech, A.; Barr, I.; Hasson, Y.; Lenc, K.; Mensch, A.; Millican, K.; Reynolds, M.; et al. Flamingo: A visual language model for few-shot learning. Adv. Neural Inf. Process. Syst. 2022, 35, 23716–23736. [Google Scholar]
  35. Gharoun, H.; Momenifar, F.; Chen, F.; Gandomi, A.H. Meta-learning approaches for few-shot learning: A survey of recent advances. ACM Comput. Surv. 2024, 56, 1–41. [Google Scholar] [CrossRef]
  36. Ptito, M.; Bleau, M.; Bouskila, J. The Retina: A Window into the Brain. Cells 2021, 10, 3269. [Google Scholar] [CrossRef] [PubMed]
  37. Shepherd, G.M.; Yamawaki, N. Untangling the cortico-thalamo-cortical loop: Cellular pieces of a knotty circuit puzzle. Nat. Rev. Neurosci. 2021, 22, 389–406. [Google Scholar] [CrossRef] [PubMed]
  38. Wu, Y.; Zhao, M.; Deng, H.; Wang, T.; Xin, Y.; Dai, W.; Huang, J.; Zhou, T.; Sun, X.; Liu, N.; et al. The neural origin for asymmetric coding of surface color in the primate visual cortex. Nat. Commun. 2024, 15, 516. [Google Scholar] [CrossRef]
  39. Semedo, J.D.; Jasper, A.I.; Zandvakili, A.; Krishna, A.; Aschner, A.; Machens, C.K.; Kohn, A.; Yu, B.M. Feedforward and feedback interactions between visual cortical areas use different population activity patterns. Nat. Commun. 2022, 13, 1099. [Google Scholar] [CrossRef]
  40. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. Caltech-UCSD Birds-200-2011 (CUB-200-2011); Technical Report CNS-TR-2011-001; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  41. Bertinetto, L.; Henriques, J.F.; Torr, P.; Vedaldi, A. Meta-learning with differentiable closed-form solvers. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  42. Vinyals, O.; Blundell, C.; Lillicrap, T.; Kavukcuoglu, k.; Wierstra, D. Matching Networks for One Shot Learning. In Proceedings of the Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Curran Associates, Inc.: Nice, France,, 2016; Volume 29. [Google Scholar]
  43. Triantafillou, E.; Zemel, R.; Urtasun, R. Few-Shot Learning Through an Information Retrieval Lens. In Proceedings of the Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Nice, France, 2017; Volume 30. [Google Scholar]
  44. Belkin, M.; Niyogi, P. Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  45. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 4–9 December 2017; NIPS’17. pp. 4080–4090. [Google Scholar]
Figure 1. Part of the ventral stream structure, which is associated with human few-shot learning behavior. From a biological theoretical perspective, it has provided significant inspiration for our design of few-shot learning models and laid a solid biological theoretical foundation for the model proposed in this paper.
Figure 1. Part of the ventral stream structure, which is associated with human few-shot learning behavior. From a biological theoretical perspective, it has provided significant inspiration for our design of few-shot learning models and laid a solid biological theoretical foundation for the model proposed in this paper.
Brainsci 15 00882 g001
Figure 2. Overall structure of our model. It is divided into three functional modules: image preprocessing, feature extraction and dimensionality reduction, and the classifier. In the figure, these modules are mapped to the corresponding parts of the ventral stream structure in the brain that inspired them. Specifically, for the feature extraction and dimensionality reduction modules, we utilized the front-end feature extraction module of ViT. For the classifier module, we opted for a two-layer neural field.
Figure 2. Overall structure of our model. It is divided into three functional modules: image preprocessing, feature extraction and dimensionality reduction, and the classifier. In the figure, these modules are mapped to the corresponding parts of the ventral stream structure in the brain that inspired them. Specifically, for the feature extraction and dimensionality reduction modules, we utilized the front-end feature extraction module of ViT. For the classifier module, we opted for a two-layer neural field.
Brainsci 15 00882 g002
Figure 3. The training and prediction processes of the proposed model. (a) Training: Labeled samples from the training set are fed into the primary neural field, activating neurons in that region. Their labels correspond to specific neurons in the high-level neural field. A connection is then established between the two neural fields, indicating that the classifier module has successfully learned and memorized the features of that sample category. (b) Prediction: After a test sample is fed into the model, the model detects the region of neurons activated in the primary neural field and identifies the corresponding neuron in the high-level neural field. Subsequently, the model assigns the label associated with that neuron to the input test sample. Neurons shown in blue are inactive; the remaining neurons are colored in shades of red, with deeper red indicating stronger activation.
Figure 3. The training and prediction processes of the proposed model. (a) Training: Labeled samples from the training set are fed into the primary neural field, activating neurons in that region. Their labels correspond to specific neurons in the high-level neural field. A connection is then established between the two neural fields, indicating that the classifier module has successfully learned and memorized the features of that sample category. (b) Prediction: After a test sample is fed into the model, the model detects the region of neurons activated in the primary neural field and identifies the corresponding neuron in the high-level neural field. Subsequently, the model assigns the label associated with that neuron to the input test sample. Neurons shown in blue are inactive; the remaining neurons are colored in shades of red, with deeper red indicating stronger activation.
Brainsci 15 00882 g003
Figure 4. The three possible scenarios the model may encounter during prediction: (1) Only one neuron is activated. In this case, the model directly outputs the image category corresponding to the activated high-level neuron. (2) No neurons are activated. This indicates that the model’s perception range for high-level neurons is set too small. (3) Multiple neurons are activated. This suggests that the model’s perception range for high-level neurons is set too large. To address these issues, we incorporate a self-regulation process for the range parameters into the model, which significantly improves the accuracy of the experimental results. Neurons shown in blue are inactive; the remaining neurons are colored in shades of red, with deeper red indicating stronger activation.
Figure 4. The three possible scenarios the model may encounter during prediction: (1) Only one neuron is activated. In this case, the model directly outputs the image category corresponding to the activated high-level neuron. (2) No neurons are activated. This indicates that the model’s perception range for high-level neurons is set too small. (3) Multiple neurons are activated. This suggests that the model’s perception range for high-level neurons is set too large. To address these issues, we incorporate a self-regulation process for the range parameters into the model, which significantly improves the accuracy of the experimental results. Neurons shown in blue are inactive; the remaining neurons are colored in shades of red, with deeper red indicating stronger activation.
Brainsci 15 00882 g004
Table 2. Accuracy of the proposed model compared with that of EASY, ESPT, MetaQDA, PMF-BPA, SgVA-CLIP, and CAML on CUB200-2011, CIFAR-FS, and miniImageNet. Best results are in bold, second-best are underlined.
Table 2. Accuracy of the proposed model compared with that of EASY, ESPT, MetaQDA, PMF-BPA, SgVA-CLIP, and CAML on CUB200-2011, CIFAR-FS, and miniImageNet. Best results are in bold, second-best are underlined.
Method (Backbone)CUB(5-1)CUB(5-5)CIFAR-FS(5-1)CIFAR-FS(5-5)MINI(5-1)MINI(5-5)
EASY (ResNet12) [16]0.78560.91930.76200.89000.71750.8715
ESPT (ResNet12) [20]0.85450.9402\\0.68360.8411
ProtoNet (ResNet18) [45]0.71880.87420.72200.83500.54160.7368
ProtoNet (ViT) [45]0.87000.97100.57700.81000.85300.9600
MetaQDA (ViT) [17]0.88300.97400.60400.83200.88200.9740
CAML (ViT) [22]0.91800.97100.70800.84760.96200.9860
DeepEMD (ViT) [19]\\0.82800.93100.90500.9720
PMF-BPA (ViT) [21]0.95800.97120.87100.94700.95200.9870
P >M > F (ViT) [18]0.92300.97000.84300.92200.95300.9840
FSViT (ViT) [23]\\0.83700.93600.95900.9850
SgVA-CLIP (ViT) [15]\\\\0.97950.9872
Ours (ViT)0.92420.95240.98110.98460.98870.9902
The methods compared include EASY (Ensemble Augmented-Shot-Y-Shaped Learning), ESPT (Episodic Spatial Pretext Task), ProtoNet (Prototypical Network), MetaQDA (Meta-learning generalization of Quadratic Discriminant Analysis), CAML (Context-Aware Meta-Learning), DeepEMD (Differentiable Earth Mover’s Distance), PMF-BPA (Pairwise-Mutual-Feature Balanced-Pairwise-Affinities transform), P >M > F (Pre-training, Meta-training, and Fine-tuning), FSViT (Multi-Scale Feature Sets in Vision Transformer), SgVA-CLIP (Semantic-guided Visual Adapting for vision–language pre-trained models such as CLIP), ViT (Vision Transformer), and ResNet12 (a 12-layer residual network with four residual stages and a final global average-pooling layer).
Table 3. Effects of different distance metrics on model accuracy on CUB200-2011, CIFAR-FS, and miniImageNet. The best results are highlighted in bold, and the second-best results are underlined.
Table 3. Effects of different distance metrics on model accuracy on CUB200-2011, CIFAR-FS, and miniImageNet. The best results are highlighted in bold, and the second-best results are underlined.
MetricCUB(5-1)CUB(5-5)CIFAR-FS(5-1)CIFAR-FS(5-5)MINI(5-1)MINI(5-5)
Euclidean0.68310.74090.71780.78610.78780.8243
Cosine0.92420.95240.98110.98460.98870.9902
Correlation0.91250.94940.97850.98160.98730.9899
Minkowski0.88390.91920.95170.95960.97720.9825
Distances used in this study include the Euclidean distance, which measures the straight-line distance between two points; the cosine distance, defined as the cosine of the angle between two vectors; the correlation distance, computed as the Pearson correlation coefficient; and the Minkowski distance, a generalized distance metric with parameter p.
Table 4. Accuracy of different feature extractors and classifiers. The best results are highlighted in bold, and the second-best results are underlined.
Table 4. Accuracy of different feature extractors and classifiers. The best results are highlighted in bold, and the second-best results are underlined.
EncoderClassiferCUB(5-1)CUB(5-5)CIFAR-FS(5-1)CIFAR-FS(5-5)MINI(5-1)MINI(5-5)
ViTKNN0.90990.93020.96780.97260.97200.9786
SVM0.74620.94450.76700.97470.79140.9833
OURS0.92420.95240.98110.98460.98870.9902
Resnet18KNN0.83000.86320.77680.82720.64950.7201
SVM0.64200.86700.58250.81850.49560.7150
OURS0.84350.87380.79120.83800.67430.7374
The classifiers used in this study include KNN, a non-parametric classifier that assigns the majority label among the k closest training samples based on the chosen distance metric; SVM, a maximum-margin classifier that finds an optimal hyperplane to separate classes, optionally using kernel functions for non-linear decision boundaries; ViT, the Vision Transformer; and ResNet18, a standard 18-layer residual network containing an initial 7 × 7 convolution, four residual blocks, and a global average-pooling layer.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Su, J.; Xing, L.; Li, T.; Xiang, N.; Shi, J.; Jin, D. Reproducing the Few-Shot Learning Capabilities of the Visual Ventral Pathway Using Vision Transformers and Neural Fields. Brain Sci. 2025, 15, 882. https://doi.org/10.3390/brainsci15080882

AMA Style

Su J, Xing L, Li T, Xiang N, Shi J, Jin D. Reproducing the Few-Shot Learning Capabilities of the Visual Ventral Pathway Using Vision Transformers and Neural Fields. Brain Sciences. 2025; 15(8):882. https://doi.org/10.3390/brainsci15080882

Chicago/Turabian Style

Su, Jiayi, Lifeng Xing, Tao Li, Nan Xiang, Jiacheng Shi, and Dequan Jin. 2025. "Reproducing the Few-Shot Learning Capabilities of the Visual Ventral Pathway Using Vision Transformers and Neural Fields" Brain Sciences 15, no. 8: 882. https://doi.org/10.3390/brainsci15080882

APA Style

Su, J., Xing, L., Li, T., Xiang, N., Shi, J., & Jin, D. (2025). Reproducing the Few-Shot Learning Capabilities of the Visual Ventral Pathway Using Vision Transformers and Neural Fields. Brain Sciences, 15(8), 882. https://doi.org/10.3390/brainsci15080882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop