Next Article in Journal
Coal Gangue Ecological Matrix Coupled with Microalgae for Soil Improvement and Plant Growth in Reclaimed Mining Areas
Previous Article in Journal
The First Complete Mitogenome Characterization and Phylogenetic Implications of Elops machnata (Teleostei: Elopiformes: Elopidae)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The Neural Mechanisms of Visual and Vestibular Interaction in Self-Motion Perception

1
Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai 200062, China
2
Key Laboratory of Brain Functional Genomics, Affiliated Mental Health Center (ECNU), School of Psychology and Cognitive Science, East China Normal University, Shanghai 200062, China
*
Author to whom correspondence should be addressed.
Biology 2025, 14(7), 740; https://doi.org/10.3390/biology14070740 (registering DOI)
Submission received: 9 May 2025 / Revised: 13 June 2025 / Accepted: 19 June 2025 / Published: 21 June 2025
(This article belongs to the Special Issue Mechanisms Underlying Neuronal Network Activity)

Simple Summary

The brain maintains a stable sense of self-motion by flexibly combining visual information (optic flow) with vestibular signals from the inner ear’s motion sensors. This integration allows us to move, walk, and balance effectively in everyday life. When visual and vestibular cues disagree—such as during virtual reality, turbulent motion, or certain neurological conditions—people can experience motion sickness, disorientation, or perceptual disturbances, which are also observed in disorders like autism and Parkinson’s disease. Research indicates that the brain uses a near-optimal, reliability-weighted averaging strategy: it dynamically assesses the trustworthiness of each cue and adjusts their influence accordingly. Understanding these mechanisms not only clarifies how we navigate complex environments but also informs potential interventions for conditions involving impaired sensory integration. This review synthesizes behavioral, neurophysiological, computational, and imaging findings to highlight how dynamic visual–vestibular integration underpins accurate self-motion perception and may guide therapeutic approaches for sensory integration disorders.

Abstract

Self-motion perception is a complex multisensory process that relies on the integration of various sensory signals, particularly visual and vestibular inputs, to construct stable and unified perceptions. It is essential for spatial navigation and effective interaction with the environment. This review systematically explores the mechanisms and computational principles underlying visual–vestibular integration in self-motion perception. We first outline the individual contributions of visual and vestibular cues and then introduce Bayesian inference as a normative framework for the quantitative modeling of multisensory integration. We also discuss multisensory recalibration as a critical mechanism in resolving conflicts between sensory inputs and maintaining perceptual stability. Using heading perception as a model system, we further describe the relevant visual and vestibular pathways involved in this process, as well as the brain regions involved. Finally, we discuss the neural mechanisms mediating visual–vestibular interactions through models of the Bayesian optimal integration and divisive normalization.

1. Introduction

Visual–vestibular integration plays a fundamental role in enabling the brain to construct accurate and stable representations of self-motion and spatial orientation [1,2,3,4]. By combining information from the visual system—particularly optic flow patterns—with signals from the vestibular system that encode head motion and acceleration, the brain achieves a coherent perception of body movement in space. This multisensory process is essential in maintaining perceptual stability, guiding navigation, and supporting posture and balance during everyday behaviors.
When visual and vestibular signals are not properly integrated, the resulting sensory conflict can lead to perceptual instability and disorientation. Such integration failures are associated with common conditions including motion sickness, spatial disorientation, and vestibular disorders [5,6,7]. Moreover, abnormal multisensory processing has been reported in various neurological and psychiatric conditions—such as autism spectrum disorder [8,9,10], schizophrenia [11], and Parkinson’s disease [12]—suggesting that atypical sensory integration may be related to these disorders.
Early research on multisensory integration between vestibular and visual signals primarily focused on behavioral phenomena. With increasing interest in the underlying processes, recent studies have progressively shifted their attention toward the neural mechanisms supporting this integration. In this review, we use visual–vestibular interactions in self-motion perception as a model system to explore the behavioral principles and neural substrates of visual–vestibular integration.

2. Self-Motion Perception Involving Visual–Vestibular Integration

Self-motion perception refers to the ability to detect one’s movement and changes in head position relative to the external world. In everyday life, accurate self-motion perception is vital for survival, as it allows individuals to make appropriate judgments and plan actions based on their dynamic relationship with the environment. This process relies on the integration of multiple sensory signals, among which visual and vestibular inputs are the two primary sources of information [13].
Visual signals contribute through optic flow—the systematic pattern of motion across the retina, generated as we move through our environment. Optic flow provides critical information about the relative movement between the observer and the surrounding world [14]. Numerous psychophysical studies have investigated how the brain estimates the heading direction based on the analysis of optic flow [15,16,17,18].
Vestibular signals, on the other hand, are derived from sensors located in the inner ear: the semicircular canals, which detect angular acceleration, and the otolith organs, which detect linear acceleration, including the effects of gravity [19,20,21,22,23]. These sensors continuously monitor the position and movement of the head, thereby supplying direct information about self-motion [24,25,26,27]. Beyond its contribution to basic self-motion detection, the vestibular system plays a central role in ensuring perceptual stability and guiding accurate motor responses. It does so not only by relaying sensory information about head motion but also by engaging predictive mechanisms that distinguish between self-generated and externally applied motion, thereby stabilizing visual perception during active behaviors [28]. Furthermore, vestibular signals are essential for gaze stabilization via the vestibulo-ocular reflex and in maintaining posture and balance through reflexive and voluntary pathways [29]. Importantly, vestibular inputs also reach widespread cortical and hippocampal regions, where they contribute to higher-level cognitive functions such as spatial orientation, memory, and navigation [30]. These integrative roles position the vestibular system as a key sensory modality not only for motion detection but in coordinating action and supporting spatial cognition.
However, in naturalistic environments, optic flow information is often ambiguous or disrupted. Rotations of the head or eyes [31,32], as well as the relative motion of other objects in the visual field [33,34], can distort the structure of the optic flow, complicating the brain’s interpretation of self-motion. The vestibular system also has inherent limitations—it cannot detect constant-velocity linear motion and struggles to distinguish between head tilt caused by gravity and translation relative to the ground [35]. However, studies have shown that the nervous system addresses this equivalence problem by integrating otolith and semicircular canal signals at multiple levels of the central nervous system [36,37,38,39]. In addition, multisensory integration provides a robust solution: biological systems tend to integrate visual and vestibular information to achieve more reliable self-motion perception. This visual–vestibular integration not only compensates for the weaknesses of each individual system but also significantly enhances the accuracy and stability of motion perception [3,4,40,41,42,43,44,45].

3. Bayesian Optimal Computational Model of Visual–Vestibular Integration

To achieve accurate and stable self-motion perception, the brain needs to integrate information from different sensory channels. Computational theories provide a mathematical framework for this problem (maximum likelihood estimation) [1,46,47,48,49], with the fundamental concept being that sensory information inherently carries uncertainty [50], and the encoding of this information by the sensory systems (such as the firing rates of visual/vestibular neurons) is also stochastic [2,51,52,53,54,55].
To estimate the self-motion heading direction S , visual cue C v i s and vestibular cue C v e s t each provide representations of S that can be described by normal distributions p C v i s S and p C v e s t S , respectively. The observer’s belief about the stimulus S can be expressed through the likelihood function L S p C S . Assuming that the noise of the two sensory modalities is independent, the likelihood function for the combined visual–vestibular condition can be decomposed into the product of individual cue likelihoods:
L c o m S p C v i s , C v e s t S = p C v i s S p C v e s t S = L v i s S L v e s t S
The Bayesian perceptual model introduces a prior p S , and the posterior probability of the current heading S , based on synchronous visual and vestibular cues, is given by
p S C v i s , C v e s t   p C v i s , C v e s t S · p S p C v i s S p C v e s t S · p S
By maximizing the posterior distribution p S C v i s , C v e s t , we can obtain the best estimate of heading S ^ . When the prior p S is flat, the posterior is determined by the likelihood function p C v i s , C v e s t S , and the mean μ and variance σ 2 of the distribution of S ^ satisfy
μ c o m = w v i s μ v i s + w v e s t μ v e s t = r v i s μ v i s + r v e s t μ v e s t r v i s + r v e s t
Here, w represents the weights of different sensory modalities during integration, and r represents the reliability of each cue (defined as the inverse of the variance 1   / σ 2 ). The reliability of the combined estimate is the sum of the reliabilities of each cue:
1 σ c o m 2 = 1 σ v i s 2 + 1 σ v e s t 2
The Bayesian optimal integration model linearly weights the information according to its reliability, with the combined cue’s reliability being superior to that of any single cue (thus reducing uncertainty), and it tends to favor cues with higher reliability [1,2,56,57]. The Bayesian model is capable of handling the uncertainty in sensory information and better explains behavioral performance through top-down expectations or stimulus biases. Whether these biases are acquired through task learning or established by life experience, the Bayesian model effectively integrates this knowledge into perceptual decision-making.
In this framework, not only does multisensory integration depend on the reliability of sensory inputs, but it is also closely linked to the individual’s prior beliefs [58]. For instance, in situations where visual and vestibular cues are in conflict, the Bayesian model can dynamically adjust the weights of visual and vestibular information based on their relative reliability, thereby achieving an optimal perceptual outcome [4]. Recent studies reveal that the brain’s internal models of gravity significantly influence multisensory integration mechanisms, with particularly strong effects on vertical motion perception—findings that align with Bayesian models of perceptual inference [59,60].
This Bayesian perceptual model has been applied in several behavioral experiments, confirming how the brain optimizes the integration of information to enhance perceptual accuracy and stability [3,4,51,61,62]. Particularly when facing multiple sensory inputs, the predictions of the Bayesian model align with experimental data, indicating that the brain indeed follows this theoretical framework during perceptual processes.

4. Behavioral Research on Visual–Vestibular Integration

A large body of human psychophysical research has demonstrated that, when multiple sensory cues are available, observers typically exhibit behavior consistent with Bayesian optimal integration [51,54,63]. These studies show that the brain tends to weight each cue according to its reliability, resulting in perceptual estimates that are more precise than those based on individual cues alone.
In non-human primates, Gu and colleagues conducted a seminal study investigating visual–vestibular integration using a heading discrimination task. In this two-alternative forced choice (2AFC) paradigm, macaques were trained to report whether their perceived heading direction was to the left or right of straight ahead by making a corresponding saccade. Across trials, the monkeys were presented with visual cues (optic flow), vestibular cues (inertial motion), or both cues simultaneously.
The results showed that heading perception was significantly more accurate under the combined condition than under either unimodal condition. Psychometric functions revealed lower discrimination thresholds in the combined condition, consistent with the predictions of a Bayesian optimal integration model (Figure 1A). These findings provide strong behavioral evidence that the primate brain integrates visual and vestibular inputs in a statistically optimal fashion to improve perceptual precision.
To further explore the dynamic effects of cue reliability on heading perception, a study built upon Gu et al.’s experiment by introducing bimodal conditions with visual–vestibular heading conflicts and varying the visual coherence gradient (to alter the weight of visual cues) [64]. The results showed that, when visual coherence was higher, the subjects’ heading perception tended to favor the visual direction. As visual coherence decreased, the visual weight was reduced while the vestibular weight increased, causing the heading perception to shift toward the vestibular direction. These findings are consistent with the predictions of the Bayesian integration model (Figure 1B).
Although numerous studies have shown that the integration of visual and vestibular information enhances the precision of motion perception [4,40], this integration is most effective when the optic flow and vestibular signals are aligned or nearly aligned in direction. Therefore, the brain must infer the causal relationships behind these cues to handle multisensory information in cases of conflict [65,66,67,68]. Some studies have analyzed how explicit and implicit causal inference strategies are employed to integrate information when there are differences between visual and vestibular cues. In explicit causal inference tasks, subjects are required to explicitly judge whether visual and vestibular cues originate from the same source, while, in implicit causal inference tasks, subjects only need to judge the vestibular direction, implicitly assessing the causal relationship between visual and vestibular cues to form the final directional perception. Such research shows that the brain does not simply merge the two cues but adjusts the perceptual strategy based on their offsets. When the offsets between the two cues are small, the brain tends to integrate them; when the offsets are large, the brain tends to separate the cues that represent different directions [68].
Moreover, when visual and vestibular cues are inconsistent, the brain dynamically adjusts the perceptual discrepancies between the two in order to maintain perceptual consistency and accuracy. This process is known as multisensory recalibration. In self-motion perception, visual–vestibular recalibration has two mechanisms: unsupervised and supervised recalibration [69,70,71]. In the absence of external feedback, unsupervised recalibration causes the perception of visual and vestibular cues to adjust in opposite directions to reduce the conflict between them, achieving “internal consistency”. Under these conditions, the psychometric curve shifts in the opposite direction (Figure 1C). When there is a difference between visual and vestibular cues, providing feedback on the accuracy of one of the cues (such as vestibular or visual) will trigger supervised recalibration. This recalibration aims for “external accuracy”, where the comparison between the cues and environmental feedback leads to “yoked adjustment”, causing the psychometric curve to shift in the same direction (Figure 1D).
The integration of visual and vestibular information plays a crucial role in self-motion perception, and this process aligns with the predictions of the Bayesian optimal integration model. Studies show that the brain dynamically adjusts its integration strategy based on the reliability of each sensory cue to improve the perceptual accuracy. However, in cases of conflicting information, the brain does not merely perform weighted averaging but instead dynamically adjusts the perceptual strategy through causal inference to ensure perceptual consistency and accuracy. Additionally, the visual–vestibular recalibration mechanism further reveals how the brain adapts based on external feedback or internal cue differences, ensuring precise motion perception in a complex multisensory environment. These studies provide profound behavioral evidence for the neural mechanisms of perceptual integration and offer important insights into the brain’s adaptive adjustments in complex perceptual situations.

5. Visual–Vestibular Integration Neural Network

5.1. Brain Regions Involved in Visual–Vestibular Integration

To explore the neural mechanisms underlying visual–vestibular integration, numerous studies have focused on brain regions that exhibit selectivity for the heading direction. Figure 2 illustrates the pathways for the transmission of vestibular and visual information, along with the cortical network potentially involved in visual–vestibular heading perception [72,73,74,75,76,77,78].
In self-motion perception, vestibular signals first project to the vestibular nuclei (VN) and then travel via the vestibulo-thalamic pathway directly to the parieto-insular vestibular cortex (PIVC) [79,80,81]. From here, vestibular information is relayed to other multisensory areas, such as the dorsal medial superior temporal area (MSTd) [82,83,84,85,86], the ventral intraparietal area (VIP) [87,88,89,90,91,92,93], the visual posterior sylvian area (VPS) [94,95], the smooth eye movement area of the frontal eye field (FEFsem) [96], the superior temporal polysensory area (STP) [97], and area 7a [98].
Visual motion signals, such as optic flow, are first relayed through the lateral geniculate nucleus (LGN) to the primary visual cortex (V1) and subsequently processed in the middle temporal area (MT) along the dorsal visual pathway [79]. The receptive fields of neurons in V1 [99,100] and MT [101,102] are relatively small and primarily handle local motion information. Visual signals are then projected to MSTd, where neurons have larger receptive fields and can extract global motion information [82,83,103], before being passed to other multisensory brain regions.
Among these areas, the MSTd serves as a critical node where visual and vestibular information converges. In macaques, neurons in the MSTd respond to both visual and vestibular stimuli, with a pronounced preference for specific directions [82,85]. By comparing MSTd neurons’ tuning to visual and vestibular directions, researchers have classified neurons into two categories: “congruent” neurons, which have aligned preferences for both visual and vestibular stimuli, and “opposite” neurons, which have opposing directional preferences [85]. Congruent neurons’ activity is closely related to behavior involving direction discrimination [25], supporting the idea that vestibular information contributes to heading perception in the MSTd. Opposite neurons, while their specific function remains unclear, may be involved in distinguishing self-motion from object motion [104]. Studies have shown that microstimulation in the MSTd can significantly alter self-motion perception based on optic flow [105,106], although it has minimal effects on vestibular motion direction judgments in darkness [105]. Additionally, the reversible inactivation of the MSTd affects visual direction judgments but has little impact on vestibular perception, highlighting the MSTd’s crucial role in processing optic flow for self-motion perception. In contrast, vestibular conditions rely more on other cortical areas, such as the PIVC or VPS [90,94].
In humans, the MT+ complex includes the hMT and hMST, delineated using retinotopic mapping and receptive field asymmetries [107,108]. Galvanic vestibular stimulation (GVS) studies show that vestibular responses are confined to the hMSTa, the anterior portion of the hMST [77], consistent with the macaque MSTd. Congruent and opposite visual–vestibular roll stimuli have been shown to elicit distinguishable multivoxel activation patterns in the hMSTa, as revealed by multivoxel pattern analysis [109]. This suggests the existence of functionally distinct neuronal populations within the hMSTa, potentially analogous to the congruent and opposite cells found in the macaque MSTd. Although direct evidence for vestibular heading encoding in the hMSTa remains lacking, these findings imply close homology to the MSTd.
The PIVC is considered the core area for the processing of vestibular information. Most neurons in the macaque PIVC do not respond to visual stimuli [110], which suggests that the PIVC is not the main site for visual–vestibular integration. However, a distinct cortical area posterior to the PIVC, known as the VPS, responds to both rotation and translation stimuli in the dark [94]. Unlike the PIVC, the VPS also shows significant responses to optic flow [94]. The VPS is closely connected to the PIVC, receiving vestibular signals and possibly optic flow signals from the MSTd [81]. Many neurons in the macaque VPS are tuned to both visual and vestibular motion directions. As with the MSTd, the VPS also contains congruent and opposite neurons. However, the dominant population in the VPS is opposite neurons, suggesting that the VPS plays a key role in visual–vestibular interactions.
In humans, the PIVC has been identified as receiving similar subcortical vestibular inputs [111,112,113,114]. The posterior parietal operculum, particularly area OP2, has been proposed as a key node for visual–vestibular integration in humans [115]. OP2 is homologous to the macaque PIVC and is consistently activated by vestibular stimulation [114,116,117]. It receives converging vestibular, somatosensory, and visual inputs [81], and plays a crucial role in self-motion perception [118]. While the anterior portions of OP2 primarily process vestibular signals, adjacent retroinsular regions also respond to visual motion, suggesting a spatial gradient of multisensory integration within the broader OP2+ complex [119]. Thus, OP2 and its neighboring regions may serve as an important interface for the processing and integration of multisensory cues related to self-motion.
Accumulating evidence suggests that the posterior insular cortex (PIC) may serve as the human homolog of the macaque VPS [120]. Functional MRI studies have shown that the PIC responds to both vestibular and visual stimuli. When the two modalities are presented together, responses to congruent and opposite motion directions resemble the response patterns observed in the VPS [119,121,122]. Moreover, the structural connectivity profile of the PIC supports its proposed role in visual–vestibular processing: it is connected with regions such as the superior temporal sulcus, intraparietal cortex, and supramarginal gyrus [123,124]. These findings suggest that the PIC may play a central role in human visual–vestibular integration, potentially mirroring the function of the VPS in macaques.
Recent human studies have shown that several medial cortical regions are involved in visual–vestibular processing underlying self-motion perception [125,126], with the cingulate sulcus visual area (CSv) being a particularly notable example. The CSv responds robustly to optic flow stimuli but not to random motion or static visual input [127]. It also exhibits strong responses to artificial vestibular stimulation, indicating that it is a multisensory region engaged in both visual and vestibular processing [77]. However, current evidence suggests that the CSv does not integrate these sensory signals. Instead, it may relay them to motor-related regions, as it shows strong connectivity with areas such as the supplementary motor area [128].
The VIP, located at the bottom of the intraparietal sulcus, is another multisensory integration area. The majority of the neurons in the macaque VIP show significant responses to both optic flow and vestibular stimuli [87,88,92]. The response characteristics of the VIP to optic flow are similar to those of the MSTd, with strong directional selectivity and a neural connection between the two regions [129], indicating that the MSTd may be the primary source of the VIP’s visual input [90]. An analysis of the direction preferences of VIP neurons revealed the presence of both congruent and opposite neurons [89,90], similarly to the MSTd and VPS. However, unlike the MSTd, the activity of VIP neurons shows a stronger correlation with directional judgment [91]. Despite this, causal manipulation experiments show that the inactivation of the VIP does not significantly affect directional perception based on visual or vestibular cues, and microstimulation in the VIP has no significant effect on optic flow-based direction perception [118]. Zaidel, DeAngelis [130] further demonstrated that, in the VIP, choice-related (top-down) signals are dominant, while, in the MSTd, sensory-driven (bottom-up) signals prevail; see also Chen, Zeng [131].
The human VIP (hVIP), located in the fundus of the intraparietal sulcus, responds to visual, auditory, and somatosensory inputs [112,132]. It is more sensitive to self-motion-compatible optic flow than to spatially fragmented motion [126]. Although vestibular responses are weak or undetectable in the hVIP [77], decoding analyses have shown that hVIP activity distinguishes between congruent and opposite vestibular–visual stimuli [109]. This suggests the presence of latent vestibular sensitivity not easily captured by standard fMRI.
In addition to the aforementioned areas, multisensory brain regions such as the FEFsem, 7a, and the STP are also involved in visual–vestibular integration. The FEFsem, associated with the control of smooth eye movements, shows directionally selective responses to both visual and vestibular stimuli [96]. Similarly to the MSTd and VIP, FEFsem neurons exhibit similar directional preferences for visual and vestibular modalities, with both congruent and opposite neurons present. However, the FEFsem is primarily involved in eye movement control rather than perception, suggesting that, during self-motion perception, visual–vestibular interactions may serve more to coordinate eye movements and maintain visual stability [96]. Area 7a is highly sensitive to both visual and vestibular cues, with its response to visual–vestibular joint stimuli typically dominated by vestibular input, even suppressing visual input in some cases. This suggests that 7a may be more involved in resolving cue conflicts rather than integrating these cues for self-motion estimation [98]. The STP, which shows directionally selective responses to both visual and vestibular information, is more visually dominant [97]. Its vestibular and visual tuning strength is weaker than in the MSTd, VIP, and VPS, indicating that the STP plays a lesser role in self-motion perception [97].
Together, these multisensory brain regions orchestrate the integration of vestibular and visual signals, enabling accurate and adaptive self-motion perception. However, how neurons integrate sensory inputs from different modalities to generate self-motion perception remains an open question. The following discussion will focus on the computational rules governing the integration of visual–vestibular information by neurons—specifically, how interactions between neurons and computational strategies enable efficient perception integration and adaptive adjustments at the brain level.

5.2. Computational Implementation and Neural Mechanisms of Visual–Vestibular Integration

In self-motion perception, the integration of visual and vestibular signals follows a Bayesian optimal computational model. However, a question arises: what computational rules do neurons follow to implement this process?
Early work explored the neural encoding mechanisms behind visual–vestibular integration through a series of experiments. For instance, Gu, Angelaki [3] recorded the activity of MSTd neurons while macaques performed a heading discrimination task involving visual and vestibular cues, finding that macaques could estimate the heading direction based on optimal cue combination. An analysis of neuron activity through ROC curves revealed that the “congruent” neurons in the MSTd exhibited a visual–vestibular discrimination threshold that aligned with optimal cue combination predictions. Moreover, in combined conditions, the sensitivity of these neurons to changes in the translation direction significantly increased [3]. Similar results were observed in the VIP area, where neuronal activity also followed the Bayesian optimal integration model [91].
However, early studies focused on relatively small angular ranges (left and right deviations from the forward direction) during heading discrimination tasks in macaques. Moreover, factors like choice could affect neuronal activity and integration rules, preventing the clear depiction of a neural combination rule. To address this limitation, Morgan, DeAngelis [86] presented macaques with full-range visual, vestibular, and visual–vestibular combined stimuli covering a 360° horizontal plane and recorded the MSTd neuron activity. Their results showed that a linear weighted model could well describe the relationship between neural responses under bimodal conditions and responses to unimodal visual and vestibular stimuli:
r c o m b = w v i s r v i s + w v e s t r v e s t
where w v i s and w v e s t are weights less than 1, indicating that visual and vestibular inputs are typically subadditive rather than additive. As the visual coherence decreased, the weights for visual and vestibular inputs shifted systemically: the visual weight increased, while the vestibular weight decreased. Fetsch and colleagues [64] further demonstrated that, even with random changes in reliability across trials, neuronal adjustments to weights still occurred, suggesting that neurons may dynamically adjust their combination rules based on cue reliability. Ohshiro and colleagues [133] proposed a divisive normalization model at the level of population neurons, where the output of each neuron is divided by the total activity of a normalization pool. This pool could consist of all neurons within a functional brain area, neurons within a certain distance from the target neuron, or neurons that share similar stimulus feature preferences. This normalization mechanism is conceptually similar to lateral inhibition, except that lateral inhibition typically involves subtraction, while normalization adjusts neuronal responses via division. The neuronal response R after normalization can be expressed as
R = E n α n + 1 N i = 1 N E i n
where E n represents the sensory input to a multisensory neuron, increasing sublinearly with the stimulus intensity (with exponent n ), and E i represents the total activity of neurons in the normalization pool. α n is a constant that determines how neurons respond to increases in stimulus intensity.
The divisive normalization mechanism provides a theoretical framework for an understanding of the interaction of different sensory signals in multisensory brain areas. Through this mechanism, the changes in neuronal responses can effectively explain the neural combination rules observed in MSTd experiments [64,86]. In this model, the strength of the normalization pool depends on the discharge rates of individual sensory neurons. Specifically, under visual conditions, the total sum of the normalization signal varies significantly with changes in motion coherence. However, under combined conditions, vestibular signals contribute in a way that is independent of coherence, which weakens the impact of coherence on the normalization pool. This phenomenon helps to explain the rapid shifts in neuronal response weights observed across trials. Through the division-based normalization mechanism, neuronal responses are restricted, avoiding saturation effects and also explaining the subadditivity of neural weights (i.e., w < 1).
Recent studies have further examined how visual and vestibular cues interact in multisensory cortical areas such as the MSTd and VPS. Results show that, when vestibular and visual cues are congruent, MSTd neuronal responses are enhanced across modalities. In contrast, when these cues are in conflict, neuronal responses are suppressed, and this suppression increases as the disparity between the preferred directions of visual and vestibular inputs grows [95]. These findings are consistent with previous observations suggesting that inputs from non-preferred modalities can suppress responses from preferred ones [134], and they align with the predictions of the divisive normalization model (Figure 3A). In the VPS, the introduction of different directional visual stimuli suppresses vestibular responses. As the visual stimulus direction deviates from the preferred vestibular direction, the cross-modal suppression effect becomes stronger [95] (top row, Figure 3B).
A linear weighting model was used to fit neuronal responses in the MSTd and VPS under bimodal stimulation. The model revealed that visual inputs had higher weights in the MSTd, while vestibular inputs dominated in the VPS. This indicates that the MSTd relies more on visual information for integration, whereas the VPS primarily depends on vestibular cues.
Despite these distinct profiles, the cross-modal interaction patterns in both areas can largely be explained by the normalization model. When presented with combined stimuli, MSTd neurons exhibit a subadditive response pattern, reflecting cross-modal enhancement. In contrast, VPS neurons demonstrate a strong “winner-takes-all” mechanism: when visual signals are added to vestibular stimulation, population responses are largely dominated by vestibular inputs, while visual signals exert cross-modal suppression—even in neurons that are highly responsive to visual cues alone. These findings suggest that neuronal responses in these areas are shaped by broader network-level interactions and that divisive normalization may serve as a general computational principle governing visual–vestibular integration (Figure 4).
In summary, visual–vestibular integration involves a sophisticated network of brain regions, each playing a specific role in processing sensory information and adjusting weights based on reliability. Through mechanisms like normalization and dynamic sensory weighting, the brain achieves accurate self-motion perception even in complex and noisy environments. These neural networks and computational models provide important insights into how the brain efficiently combines multiple sensory inputs for adaptive behavior.

6. Conclusions

In this review, we have comprehensively outlined the crucial role of multisensory integration in self-motion perception, with a particular focus on the interaction between visual and vestibular signals. By exploring how the brain integrates these two sensory modalities, we have discussed the theoretical framework of the Bayesian optimal integration model that underlies the neural computational mechanisms behind this process. We have further explored the cortical networks involved in processing optic flow and vestibular cues—such as the MSTd and VPS—highlighting the computational rules that govern visual–vestibular integration.
However, the neural mechanisms underlying visual–vestibular integration in other multisensory areas, such as the VIP and FEFsem, remain poorly understood. Future studies are needed to investigate the interactions between neurons in these areas, with particular attention to the neural basis of causal inference—a general computational principle that allows the brain to attribute observed outcomes to potential causes. In addition, further work is required to examine how conflicts between visual and vestibular information affect multisensory processing, which may deepen our understanding of how the brain integrates sensory signals in an efficient and adaptive manner.
In conclusion, research on visual–vestibular integration provides valuable insights into the fundamental principles of multisensory perception and has important implications for neurocognitive models of perception, learning, and behavior. By continuing to unveil the computational strategies and neural networks involved, future research will further advance our understanding of the brain’s adaptive capabilities and provide crucial references for the development of interventions and technologies aimed at addressing sensory integration disorders.

Author Contributions

J.L.: writing—original draft, investigation, visualization. F.Z.: writing—original draft, writing—review and editing, conceptualization, visualization, funding acquisition, supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by grants from the National Natural Science Foundation of China (No. 32400826) to F.Z and the China Postdoctoral Science Foundation (2024M760927) to F.Z.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

All authors declare no conflicts of interest.

References

  1. Knill, D.C.; Pouget, A. The Bayesian brain: The role of uncertainty in neural coding and computation. Trends Neurosci. 2004, 27, 712–719. [Google Scholar] [CrossRef] [PubMed]
  2. Ma, W.J.; Beck, J.M.; Latham, P.E.; Pouget, A. Bayesian inference with probabilistic population codes. Nat. Neurosci. 2006, 9, 1432–1438. [Google Scholar] [CrossRef] [PubMed]
  3. Gu, Y.; Angelaki, D.E.; DeAngelis, G.C. Neural correlates of multisensory cue integration in macaque MSTd. Nat. Neurosci. 2008, 11, 1201–1210. [Google Scholar] [CrossRef] [PubMed]
  4. Fetsch, C.R.; Turner, A.H.; DeAngelis, G.C.; Angelaki, D.E. Dynamic Reweighting of Visual and Vestibular Cues During Self-Motion Perception. J. Neurosci. 2009, 29, 15601–15612. [Google Scholar] [CrossRef]
  5. Oman, C.M. Motion sickness: A synthesis and evaluation of the sensory conflict theory. Can. J. Physiol. Pharmacol. 1990, 68, 294–303. [Google Scholar] [CrossRef]
  6. Bronstein, A.M.; Golding, J.F.; Gresty, M.A. Visual Vertigo, Motion Sickness, and Disorientation in Vehicles. Semin. Neurol. 2020, 40, 116–129. [Google Scholar] [CrossRef]
  7. Schmal, F. Neuronal mechanisms and the treatment of motion sickness. Pharmacology 2013, 91, 229–241. [Google Scholar] [CrossRef]
  8. Noel, J.P.; Shivkumar, S.; Dokka, K.; Haefner, R.M.; Angelaki, D.E. Aberrant causal inference and presence of a compensatory mechanism in autism spectrum disorder. eLife 2022, 11, e71866. [Google Scholar] [CrossRef]
  9. Zaidel, A.; Goin-Kochel, R.P.; Angelaki, D.E. Self-motion perception in autism is compromised by visual noise but integrated optimally across multiple senses. Proc. Natl. Acad. Sci. USA 2015, 112, 6461–6466. [Google Scholar] [CrossRef]
  10. Pellicano, E.; Burr, D. When the world becomes ‘too real’: A Bayesian explanation of autistic perception. Trends Cogn. Sci. 2012, 16, 504–510. [Google Scholar] [CrossRef]
  11. Salomon, R.; Kannape, O.A.; Debarba, H.G.; Kaliuzhna, M.; Schneider, M.; Faivre, N.; Eliez, S.; Blanke, O. Agency Deficits in a Human Genetic Model of Schizophrenia: Insights from 22q11DS Patients. Schizophr. Bull. 2022, 48, 495–504. [Google Scholar] [CrossRef] [PubMed]
  12. Yakubovich, S.; Israeli-Korn, S.; Halperin, O.; Yahalom, G.; Hassin-Baer, S.; Zaidel, A. Visual self-motion cues are impaired yet overweighted during visual-vestibular integration in Parkinson’s disease. Brain Commun. 2020, 2, fcaa035. [Google Scholar] [CrossRef] [PubMed]
  13. Dichgans, J.; Brandt, T. Visual-Vestibular Interaction: Effects on Self-Motion Perception and Postural Control. In Perception; Handbook of Sensory Physiology; Springer: Berlin/Heidelberg, Germany, 1978; Volume 8, pp. 755–804. [Google Scholar]
  14. Gibson, J.J. The Perception of the Visual World; Houghton Mifflin Harcourt: Boston, MA, USA, 1950. [Google Scholar]
  15. Royden, C.S.; Banks, M.S.; Crowell, J.A. The perception of heading during eye movements. Nature 1992, 360, 583–585. [Google Scholar] [CrossRef] [PubMed]
  16. Royden, C.S.; Crowell, J.A.; Banks, M.S. Estimating heading during eye movements. Vis. Res. 1994, 34, 3197–3214. [Google Scholar] [CrossRef]
  17. Warren, W.H., Jr.; Hannon, D.J. Eye movements and optical flow. J. Opt. Soc. Am. A 1990, 7, 160–169. [Google Scholar] [CrossRef]
  18. Lappe, M.; Bremmer, F.; van den Berg, A.V. Perception of self-motion from visual flow. Trends Cogn. Sci. 1999, 3, 329–336. [Google Scholar] [CrossRef]
  19. Fernandez, C.; Goldberg, J.M. Physiology of peripheral neurons innervating semicircular canals of the squirrel monkey. II. Response to sinusoidal stimulation and dynamics of peripheral vestibular system. J. Neurophysiol. 1971, 34, 661–675. [Google Scholar] [CrossRef]
  20. Fernandez, C.; Goldberg, J.M. Physiology of peripheral neurons innervating otolith organs of the squirrel monkey. I. Response to static tilts and to long-duration centrifugal force. J. Neurophysiol. 1976, 39, 970–984. [Google Scholar] [CrossRef]
  21. Fernandez, C.; Goldberg, J.M. Physiology of peripheral neurons innervating otolith organs of the squirrel monkey. II. Directional selectivity and force-response relations. J. Neurophysiol. 1976, 39, 985–995. [Google Scholar] [CrossRef]
  22. Fernandez, C.; Goldberg, J.M. Physiology of peripheral neurons innervating otolith organs of the squirrel monkey. III. Response dynamics. J. Neurophysiol. 1976, 39, 996–1008. [Google Scholar] [CrossRef]
  23. Goldberg, J.M.; Fernandez, C. Physiology of peripheral neurons innervating semicircular canals of the squirrel monkey. I. Resting discharge and response to constant angular accelerations. J. Neurophysiol. 1971, 34, 635–660. [Google Scholar] [CrossRef] [PubMed]
  24. Telford, L.; Howard, I.; Ohmi, M. Heading judgments during active and passive self-motion. Exp. Brain Res. 1995, 104, 502–510. [Google Scholar] [CrossRef] [PubMed]
  25. Gu, Y.; DeAngelis, G.C.; Angelaki, D.E. A functional link between area MSTd and heading perception based on vestibular signals. Nat. Neurosci. 2007, 10, 1038–1047. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, S.; Angelaki, D.E. Vestibular Signals in Macaque Extrastriate Visual Cortex Are Functionally Appropriate for Heading Perception. J. Neurosci. 2009, 29, 8936–8945. [Google Scholar] [CrossRef]
  27. Angelaki, D.E.; Cullen, K.E. Vestibular system: The many facets of a multimodal sense. Annu. Rev. Neurosci. 2008, 31, 125–150. [Google Scholar] [CrossRef]
  28. Cullen, K.E. Vestibular processing during natural self-motion: Implications for perception and action. Nat. Rev. Neurosci. 2019, 20, 346–363. [Google Scholar] [CrossRef]
  29. Curthoys, I.S.; Halmagyi, G.M. Vestibular compensation: A review of the oculomotor, neural, and clinical consequences of unilateral vestibular loss. J. Vestib. Res. 1995, 5, 67–107. [Google Scholar] [CrossRef]
  30. Hitier, M.; Besnard, S.; Smith, P.F. Vestibular pathways involved in cognition. Front. Integr. Neurosci. 2014, 8, 59. [Google Scholar] [CrossRef]
  31. Banks, M.S.; Ehrlich, S.M.; Backus, B.T.; Crowell, J.A. Estimating heading during real and simulated eye movements. Vis. Res. 1996, 36, 431–443. [Google Scholar] [CrossRef]
  32. Crowell, J.A.; Banks, M.S.; Shenoy, K.V.; Andersen, R.A. Visual self-motion perception during head turns. Nat. Neurosci. 1998, 1, 732–737. [Google Scholar] [CrossRef]
  33. Royden, C.S.; Hildreth, E.C. Human heading judgments in the presence of moving objects. Percept. Psychophys. 1996, 58, 836–856. [Google Scholar] [CrossRef] [PubMed]
  34. Warren, W.H., Jr.; Saunders, J.A. Perceiving heading in the presence of moving objects. Perception 1995, 24, 315–331. [Google Scholar] [CrossRef] [PubMed]
  35. Angelaki, D.E. Eyes on Target: What Neurons Must do for the Vestibuloocular Reflex During Linear Motion. J. Neurophysiol. 2004, 92, 20–35. [Google Scholar] [CrossRef] [PubMed]
  36. Angelaki, D.E.; McHenry, M.Q.; Dickman, J.D.; Newlands, S.D.; Hess, B.J. Computation of inertial motion: Neural strategies to resolve ambiguous otolith information. J. Neurosci. 1999, 19, 316–327. [Google Scholar] [CrossRef]
  37. Merfeld, D.M.; Zupan, L.; Peterka, R.J. Humans use internal models to estimate gravity and linear acceleration. Nature 1999, 398, 615–618. [Google Scholar] [CrossRef]
  38. Laurens, J.; Meng, H.; Angelaki, D.E. Neural representation of orientation relative to gravity in the macaque cerebellum. Neuron 2013, 80, 1508–1518. [Google Scholar] [CrossRef]
  39. Angelaki, D.E.; Shaikh, A.G.; Green, A.M.; Dickman, J.D. Neurons compute internal models of the physical laws of motion. Nature 2004, 430, 560–564. [Google Scholar] [CrossRef]
  40. Butler, J.S.; Smith, S.T.; Campos, J.L.; Bulthoff, H.H. Bayesian integration of visual and vestibular signals for heading. J. Vis. 2010, 10, 23. [Google Scholar] [CrossRef]
  41. De Winkel, K.N.; Weesie, J.; Werkhoven, P.J.; Groen, E.L. Integration of visual and inertial cues in perceived heading of self-motion. J. Vis. 2010, 10, 1. [Google Scholar] [CrossRef]
  42. Bertin, R.J.; Berthoz, A. Visuo-vestibular interaction in the reconstruction of travelled trajectories. Exp. Brain Res. 2004, 154, 11–21. [Google Scholar] [CrossRef]
  43. Butler, J.S.; Campos, J.L.; Bulthoff, H.H. Optimal visual-vestibular integration under conditions of conflicting intersensory motion profiles. Exp. Brain Res. 2015, 233, 587–597. [Google Scholar] [CrossRef] [PubMed]
  44. Ohmi, M. Egocentric perception through interaction among many sensory systems. Cogn. Brain Res. 1996, 5, 87–96. [Google Scholar] [CrossRef] [PubMed]
  45. Warren, W.H.; Hannon, D.J. Direction of self-motion is perceived from optical flow. Nature 1988, 336, 162–163. [Google Scholar] [CrossRef]
  46. Knill, D.C.; Richards, W. Perception as Bayesian Inference; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  47. Ma, W.J. Bayesian Decision Models: A Primer. Neuron 2019, 104, 164–175. [Google Scholar] [CrossRef]
  48. Angelaki, D.E.; Klier, E.M.; Snyder, L.H. A vestibular sensation: Probabilistic approaches to spatial perception. Neuron 2009, 64, 448–461. [Google Scholar] [CrossRef]
  49. MacNeilage, P.R.; Banks, M.S.; Berger, D.R.; Bulthoff, H.H. A Bayesian model of the disambiguation of gravitoinertial force by visual cues. Exp. Brain Res. 2007, 179, 263–290. [Google Scholar] [CrossRef]
  50. Tolhurst, D.J.; Movshon, J.A.; Dean, A.F. The statistical reliability of signals in single neurons in cat and monkey visual cortex. Vis. Res. 1983, 23, 775–785. [Google Scholar] [CrossRef]
  51. Ernst, M.O.; Banks, M.S. Humans integrate visual and haptic information in a statistically optimal fashion. Nature 2002, 415, 429–433. [Google Scholar] [CrossRef]
  52. Landy, M.; Banks, M.; Knill, D. Ideal-Observer Models of Cue Integration. In Sensory Cue Integration; Trommershäuser, J., Konrad, K., Landy, M.S., Eds.; Oxford University Press: New York, NY, USA, 2012; pp. 5–29. [Google Scholar]
  53. Van Beers, R.J.; Wolpert, D.M.; Haggard, P. When feeling is more important than seeing in sensorimotor adaptation. Curr. Biol. 2002, 12, 834–837. [Google Scholar] [CrossRef]
  54. Alais, D.; Burr, D. The Ventriloquist Effect Results from Near-Optimal Bimodal Integration. Curr. Biol. 2004, 14, 257–262. [Google Scholar] [CrossRef]
  55. Hillis, J.M.; Ernst, M.O.; Banks, M.S.; Landy, M.S. Combining Sensory Information: Mandatory Fusion Within, but Not Between, Senses. Science 2002, 298, 1627–1630. [Google Scholar] [CrossRef] [PubMed]
  56. Fetsch, C.R.; DeAngelis, G.C.; Angelaki, D.E. Bridging the gap between theories of sensory cue integration and the physiology of multisensory neurons. Nat. Rev. Neurosci. 2013, 14, 429–442. [Google Scholar] [CrossRef] [PubMed]
  57. Ursino, M.; Cuppini, C.; Magosso, E. Neurocomputational approaches to modelling multisensory integration in the brain: A review. Neural Netw. 2014, 60, 141–165. [Google Scholar] [CrossRef] [PubMed]
  58. Körding, K.P.; Wolpert, D.M. Bayesian decision theory in sensorimotor control. Trends Cogn. Sci. 2006, 10, 319–326. [Google Scholar] [CrossRef]
  59. Delle Monache, S.; La Scaleia, B.; Finazzi Agro, A.; Lacquaniti, F.; Zago, M. Psychophysical evidence for an internal model of gravity in the visual and vestibular estimates of vertical motion duration. Sci. Rep. 2025, 15, 10394. [Google Scholar] [CrossRef]
  60. MacNeilage, P.R.; Banks, M.S.; DeAngelis, G.C.; Angelaki, D.E. Vestibular heading discrimination and sensitivity to linear acceleration in head and world coordinates. J. Neurosci. 2010, 30, 9084–9094. [Google Scholar] [CrossRef]
  61. Prsa, M.; Gale, S.; Blanke, O. Self-motion leads to mandatory cue fusion across sensory modalities. J. Neurophysiol. 2012, 108, 2282–2291. [Google Scholar] [CrossRef]
  62. Van Beers, R.J.; Sittig, A.C.; Denier van der Gon, J.J. How humans combine simultaneous proprioceptive and visual position information. Exp. Brain Res. 1996, 111, 253–261. [Google Scholar] [CrossRef]
  63. Van Beers, R.J.; Sittig, A.C.; Gon, J.J. Integration of proprioceptive and visual position-information: An experimentally supported model. J. Neurophysiol. 1999, 81, 1355–1364. [Google Scholar] [CrossRef]
  64. Fetsch, C.R.; Pouget, A.; DeAngelis, G.C.; Angelaki, D.E. Neural correlates of reliability-based cue weighting during multisensory integration. Nat. Neurosci. 2011, 15, 146–154. [Google Scholar] [CrossRef]
  65. Dokka, K.; Park, H.; Jansen, M.; DeAngelis, G.C.; Angelaki, D.E. Causal inference accounts for heading perception in the presence of object motion. Proc. Natl. Acad. Sci. USA 2019, 116, 9060–9065. [Google Scholar] [CrossRef] [PubMed]
  66. Kording, K.P.; Beierholm, U.; Ma, W.J.; Quartz, S.; Tenenbaum, J.B.; Shams, L. Causal inference in multisensory perception. PLoS ONE 2007, 2, e943. [Google Scholar] [CrossRef] [PubMed]
  67. Rohe, T.; Noppeney, U. Cortical hierarchies perform Bayesian causal inference in multisensory perception. PLoS Biol. 2015, 13, e1002073. [Google Scholar] [CrossRef] [PubMed]
  68. Acerbi, L.; Dokka, K.; Angelaki, D.E.; Ma, W.J. Bayesian comparison of explicit and implicit causal inference strategies in multisensory heading perception. PLoS Comput. Biol. 2018, 14, e1006110. [Google Scholar] [CrossRef]
  69. Zaidel, A.; Ma, W.J.; Angelaki, D.E. Supervised calibration relies on the multisensory percept. Neuron 2013, 80, 1544–1557. [Google Scholar] [CrossRef]
  70. Zaidel, A.; Turner, A.H.; Angelaki, D.E. Multisensory calibration is independent of cue reliability. J. Neurosci. 2011, 31, 13949–13962. [Google Scholar] [CrossRef]
  71. Zeng, F.; Zaidel, A.; Chen, A. Contrary neuronal recalibration in different multisensory cortical areas. Elife 2023, 12, e82895. [Google Scholar] [CrossRef]
  72. Hou, H.; Gu, Y. Multisensory integration for self-motion perception. Senses: A Compr. Ref. 2020, 6, 458–482. [Google Scholar]
  73. Noel, J.P.; Angelaki, D.E. Cognitive, Systems, and Computational Neurosciences of the Self in Motion. Annu. Rev. Psychol. 2022, 73, 103–129. [Google Scholar] [CrossRef]
  74. Angelaki, D.E. How Optic Flow and Inertial Cues Improve Motion Perception. Cold Spring Harb. Symp. Quant. Biol. 2014, 79, 141–148. [Google Scholar] [CrossRef]
  75. Gu, Y. Vestibular signals in primate cortex for self-motion perception. Curr. Opin. Neurobiol. 2018, 52, 10–17. [Google Scholar] [CrossRef] [PubMed]
  76. Keshavarzi, S.; Velez-Fort, M.; Margrie, T.W. Cortical Integration of Vestibular and Visual Cues for Navigation, Visual Processing, and Perception. Annu. Rev. Neurosci. 2023, 46, 301–320. [Google Scholar] [CrossRef] [PubMed]
  77. Smith, A.T.; Wall, M.B.; Thilo, K.V. Vestibular inputs to human motion-sensitive visual cortex. Cereb. Cortex 2012, 22, 1068–1077. [Google Scholar] [CrossRef] [PubMed]
  78. Smith, A.T.; Greenlee, M.W.; DeAngelis, G.C.; Angelaki, D.E. Distributed visual–vestibular processing in the cerebral cortex of man and macaque. Multisens. Res. 2017, 30, 91–120. [Google Scholar] [CrossRef]
  79. Felleman, D.J.; Van Essen, D.C. Distributed hierarchical processing in the primate cerebral cortex. Cereb. Cortex 1991, 1, 1–47. [Google Scholar] [CrossRef]
  80. Grusser, O.J.; Pause, M.; Schreiter, U. Localization and responses of neurones in the parieto-insular vestibular cortex of awake monkeys (Macaca fascicularis). J. Physiol. 1990, 430, 537–557. [Google Scholar] [CrossRef]
  81. Guldin, W.O.; Grusser, O.J. Is there a vestibular cortex? Trends Neurosci. 1998, 21, 254–259. [Google Scholar] [CrossRef]
  82. Duffy, C.J. MST Neurons Respond to Optic Flow and Translational Movement. J. Neurophysiol. 1998, 80, 1816–1827. [Google Scholar] [CrossRef]
  83. Duffy, C.J.; Wurtz, R.H. Sensitivity of MST neurons to optic flow stimuli. I. A continuum of response selectivity to large-field stimuli. J. Neurophysiol. 1991, 65, 1329–1345. [Google Scholar] [CrossRef]
  84. Page, W.K.; Duffy, C.J. Heading Representation in MST: Sensory Interactions and Population Encoding. J. Neurophysiol. 2003, 89, 1994–2013. [Google Scholar] [CrossRef]
  85. Gu, Y.; Watkins, P.V.; Angelaki, D.E.; DeAngelis, G.C. Visual and Nonvisual Contributions to Three-Dimensional Heading Selectivity in the Medial Superior Temporal Area. J. Neurosci. 2006, 26, 73–85. [Google Scholar] [CrossRef] [PubMed]
  86. Morgan, M.L.; DeAngelis, G.C.; Angelaki, D.E. Multisensory Integration in Macaque Visual Cortex Depends on Cue Reliability. Neuron 2008, 59, 662–673. [Google Scholar] [CrossRef] [PubMed]
  87. Bremmer, F.; Duhamel, J.R.; Ben Hamed, S.; Graf, W. Heading encoding in the macaque ventral intraparietal area (VIP). Eur. J. Neurosci. 2002, 16, 1554–1568. [Google Scholar] [CrossRef] [PubMed]
  88. Bremmer, F.; Klam, F.; Duhamel, J.R.; Ben Hamed, S.; Graf, W. Visual–vestibular interactive responses in the macaque ventral intraparietal area (VIP). Eur. J. Neurosci. 2002, 16, 1569–1586. [Google Scholar] [CrossRef]
  89. Schlack, A.; Hoffmann, K.P.; Bremmer, F. Interaction of linear vestibular and visual stimulation in the macaque ventral intraparietal area (VIP). Eur. J. Neurosci. 2002, 16, 1877–1886. [Google Scholar] [CrossRef]
  90. Chen, A.; DeAngelis, G.C.; Angelaki, D.E. Representation of Vestibular and Visual Cues to Self-Motion in Ventral Intraparietal Cortex. J. Neurosci. 2011, 31, 12036–12052. [Google Scholar] [CrossRef]
  91. Chen, A.; DeAngelis, G.C.; Angelaki, D.E. Functional Specializations of the Ventral Intraparietal Area for Multisensory Heading Discrimination. J. Neurosci. 2013, 33, 3567–3581. [Google Scholar] [CrossRef]
  92. Zhang, T.; Britten, K.H. The Responses of VIP Neurons Are Sufficiently Sensitive to Support Heading Judgments. J. Neurophysiol. 2010, 103, 1865–1873. [Google Scholar] [CrossRef]
  93. Chen, A.; DeAngelis, G.C.; Angelaki, D.E. A Comparison of Vestibular Spatiotemporal Tuning in Macaque Parietoinsular Vestibular Cortex, Ventral Intraparietal Area, and Medial Superior Temporal Area. J. Neurosci. 2011, 31, 3082–3094. [Google Scholar] [CrossRef]
  94. Chen, A.; DeAngelis, G.C.; Angelaki, D.E. Convergence of Vestibular and Visual Self-Motion Signals in an Area of the Posterior Sylvian Fissure. J. Neurosci. 2011, 31, 11617–11627. [Google Scholar] [CrossRef]
  95. Zhao, B.; Wang, R.; Zhu, Z.; Yang, Q.; Chen, A. The computational rules of cross-modality suppression in the visual posterior sylvian area. iScience 2023, 26, 106973. [Google Scholar] [CrossRef] [PubMed]
  96. Gu, Y.; Cheng, Z.; Yang, L.; DeAngelis, G.C.; Angelaki, D.E. Multisensory Convergence of Visual and Vestibular Heading Cues in the Pursuit Area of the Frontal Eye Field. Cereb. Cortex 2016, 26, 3785–3801. [Google Scholar] [CrossRef] [PubMed]
  97. Zhao, B.; Zhang, Y.; Chen, A. Encoding of vestibular and optic flow cues to self-motion in the posterior superior temporal polysensory area. J. Physiol. 2021, 599, 3937–3954. [Google Scholar] [CrossRef] [PubMed]
  98. Avila, E.; Lakshminarasimhan, K.J.; DeAngelis, G.C.; Angelaki, D.E. Visual and Vestibular Selectivity for Self-Motion in Macaque Posterior Parietal Area 7a. Cereb. Cortex 2019, 29, 3932–3947. [Google Scholar] [CrossRef]
  99. Hubel, D.H.; Wiesel, T.N. Uniformity of monkey striate cortex: A parallel relationship between field size, scatter, and magnification factor. J. Comp. Neurol. 1974, 158, 295–305. [Google Scholar] [CrossRef]
  100. Hubel, D.H.; Wiesel, T.N. Ferrier lecture. Functional architecture of macaque monkey visual cortex. Proc. R. Soc. Lond. B Biol. Sci. 1977, 198, 1–59. [Google Scholar]
  101. Albright, T.D. Direction and orientation selectivity of neurons in visual area MT of the macaque. J. Neurophysiol. 1984, 52, 1106–1130. [Google Scholar] [CrossRef]
  102. Albright, T.D.; Desimone, R. Local precision of visuotopic organization in the middle temporal area (MT) of the macaque. Exp. Brain Res. 1987, 65, 582–592. [Google Scholar] [CrossRef]
  103. Tanaka, K.; Saito, H. Analysis of motion of the visual field by direction, expansion/contraction, and rotation cells clustered in the dorsal part of the medial superior temporal area of the macaque monkey. J. Neurophysiol. 1989, 62, 626–641. [Google Scholar] [CrossRef]
  104. Kim, H.R.; Pitkow, X.; Angelaki, D.E.; DeAngelis, G.C. A simple approach to ignoring irrelevant variables by population decoding based on multisensory neurons. J. Neurophysiol. 2016, 116, 1449–1467. [Google Scholar] [CrossRef]
  105. Gu, Y.; DeAngelis, G.C.; Angelaki, D.E. Causal Links Between Dorsal Medial Superior Temporal Area Neurons and Multisensory Heading Perception. J. Neurosci. 2012, 32, 2299–2313. [Google Scholar] [CrossRef] [PubMed]
  106. Yu, X.; Gu, Y. Probing Sensory Readout via Combined Choice-Correlation Measures and Microstimulation Perturbation. Neuron 2018, 100, 715–727. [Google Scholar] [CrossRef] [PubMed]
  107. Huk, A.C.; Dougherty, R.F.; Heeger, D.J. Retinotopy and functional subdivision of human areas MT and MST. J. Neurosci. 2002, 22, 7195–7205. [Google Scholar] [CrossRef] [PubMed]
  108. Dukelow, S.P.; DeSouza, J.F.; Culham, J.C.; van den Berg, A.V.; Menon, R.S.; Vilis, T. Distinguishing subregions of the human MT+ complex using visual fields and pursuit eye movements. J. Neurophysiol. 2001, 86, 1991–2000. [Google Scholar] [CrossRef]
  109. Billington, J.; Smith, A.T. Neural mechanisms for discounting head-roll-induced retinal motion. J. Neurosci. 2015, 35, 4851–4856. [Google Scholar] [CrossRef]
  110. Chen, A.; DeAngelis, G.C.; Angelaki, D.E. Macaque Parieto-Insular Vestibular Cortex: Responses to Self-Motion and Optic Flow. J. Neurosci. 2010, 30, 3022–3042. [Google Scholar] [CrossRef]
  111. Kirsch, V.; Keeser, D.; Hergenroeder, T.; Erat, O.; Ertl-Wagner, B.; Brandt, T.; Dieterich, M. Structural and functional connectivity mapping of the vestibular circuitry from human brainstem to cortex. Brain Struct. Funct. 2016, 221, 1291–1308. [Google Scholar] [CrossRef]
  112. Lopez, C.; Blanke, O. The thalamocortical vestibular system in animals and humans. Brain Res. Rev. 2011, 67, 119–146. [Google Scholar] [CrossRef]
  113. Dieterich, M.; Brandt, T. Global orientation in space and the lateralization of brain functions. Curr. Opin. Neurol. 2018, 31, 96–104. [Google Scholar] [CrossRef]
  114. Raiser, T.M.; Flanagin, V.L.; Duering, M.; van Ombergen, A.; Ruehl, R.M.; Zu Eulenburg, P. The human corticocortical vestibular network. Neuroimage 2020, 223, 117362. [Google Scholar] [CrossRef]
  115. Ibitoye, R.T.; Mallas, E.J.; Bourke, N.J.; Kaski, D.; Bronstein, A.M.; Sharp, D.J. The human vestibular cortex: Functional anatomy of OP2, its connectivity and the effect of vestibular disease. Cereb. Cortex 2023, 33, 567–582. [Google Scholar] [CrossRef] [PubMed]
  116. Zu Eulenburg, P.; Caspers, S.; Roski, C.; Eickhoff, S.B. Meta-analytical definition and functional connectivity of the human vestibular cortex. Neuroimage 2012, 60, 162–169. [Google Scholar] [CrossRef] [PubMed]
  117. Lopez, C.; Blanke, O.; Mast, F.W. The human vestibular cortex revealed by coordinate-based activation likelihood estimation meta-analysis. Neuroscience 2012, 212, 159–179. [Google Scholar] [CrossRef] [PubMed]
  118. Chen, A.; Gu, Y.; Liu, S.; DeAngelis, G.C.; Angelaki, D.E. Evidence for a Causal Contribution of Macaque Vestibular, But Not Intraparietal, Cortex to Heading Perception. J. Neurosci. 2016, 36, 3789–3798. [Google Scholar] [CrossRef]
  119. Frank, S.M.; Greenlee, M.W. The parieto-insular vestibular cortex in humans: More than a single area? J. Neurophysiol. 2018, 120, 1438–1450. [Google Scholar] [CrossRef]
  120. Sunaert, S.; Van Hecke, P.; Marchal, G.; Orban, G.A. Motion-responsive regions of the human brain. Exp. Brain Res. 1999, 127, 355–370. [Google Scholar] [CrossRef]
  121. Cardin, V.; Smith, A.T. Sensitivity of human visual and vestibular cortical regions to egomotion-compatible visual stimulation. Cereb. Cortex 2010, 20, 1964–1973. [Google Scholar] [CrossRef]
  122. Frank, S.M.; Baumann, O.; Mattingley, J.B.; Greenlee, M.W. Vestibular and visual responses in human posterior insular cortex. J. Neurophysiol. 2014, 112, 2481–2491. [Google Scholar] [CrossRef]
  123. Wirth, A.M.; Frank, S.M.; Greenlee, M.W.; Beer, A.L. White Matter Connectivity of the Visual-Vestibular Cortex Examined by Diffusion-Weighted Imaging. Brain Connect 2018, 8, 235–244. [Google Scholar] [CrossRef]
  124. Indovina, I.; Bosco, G.; Riccelli, R.; Maffei, V.; Lacquaniti, F.; Passamonti, L.; Toschi, N. Structural connectome and connectivity lateralization of the multimodal vestibular cortical network. Neuroimage 2020, 222, 117247. [Google Scholar] [CrossRef]
  125. Beer, A.L.; Becker, M.; Frank, S.M.; Greenlee, M.W. Vestibular and visual brain areas in the medial cortex of the human brain. J. Neurophysiol. 2023, 129, 948–962. [Google Scholar] [CrossRef] [PubMed]
  126. Wall, M.B.; Smith, A.T. The representation of egomotion in the human brain. Curr. Biol. 2008, 18, 191–194. [Google Scholar] [CrossRef] [PubMed]
  127. Antal, A.; Baudewig, J.; Paulus, W.; Dechent, P. The posterior cingulate cortex and planum temporale/parietal operculum are activated by coherent visual motion. Vis. Neurosci. 2008, 25, 17–26. [Google Scholar] [CrossRef] [PubMed]
  128. Smith, A.T.; Beer, A.L.; Furlan, M.; Mars, R.B. Connectivity of the Cingulate Sulcus Visual Area (CSv) in the Human Cerebral Cortex. Cereb. Cortex 2018, 28, 713–725. [Google Scholar] [CrossRef]
  129. Boussaoud, D.; Ungerleider, L.G.; Desimone, R. Pathways for motion analysis: Cortical connections of the medial superior temporal and fundus of the superior temporal visual areas in the macaque. J. Comp. Neurol. 1990, 296, 462–495. [Google Scholar] [CrossRef]
  130. Zaidel, A.; DeAngelis, G.C.; Angelaki, D.E. Decoupled choice-driven and stimulus-related activity in parietal neurons may be misrepresented by choice probabilities. Nat. Commun. 2017, 8, 715. [Google Scholar] [CrossRef]
  131. Chen, A.; Zeng, F.; DeAngelis, G.C.; Angelaki, D.E. Dynamics of Heading and Choice-Related Signals in the Parieto-Insular Vestibular Cortex of Macaque Monkeys. J. Neurosci. 2021, 41, 3254–3265. [Google Scholar] [CrossRef]
  132. Bremmer, F.; Schlack, A.; Shah, N.J.; Zafiris, O.; Kubischik, M.; Hoffmann, K.; Zilles, K.; Fink, G.R. Polymodal motion processing in posterior parietal and premotor cortex: A human fMRI study strongly implies equivalencies between humans and monkeys. Neuron 2001, 29, 287–296. [Google Scholar] [CrossRef]
  133. Ohshiro, T.; Angelaki, D.E.; DeAngelis, G.C. A normalization model of multisensory integration. Nat. Neurosci. 2011, 14, 775–782. [Google Scholar] [CrossRef]
  134. Ohshiro, T.; Angelaki, D.E.; DeAngelis, G.C. A Neural Signature of Divisive Normalization at the Level of Multisensory Integration in Primate Cortex. Neuron 2017, 95, 399–411. [Google Scholar] [CrossRef]
Figure 1. Bayesian optimal integration model (A,B) and recalibration processes in multisensory perception (C,D). (A) A heading stimulus S generates both visual and vestibular cues, which are modeled as independent normal distributions, p C v i s S and p C v e s t S , respectively. According to the maximum likelihood estimation principle, the combined likelihood function can be expressed as the product of the individual cue likelihoods. In the Bayesian framework, a prior distribution p S is introduced; assuming a flat prior, the posterior is determined solely by the combined likelihood. The optimal heading estimate S ^ is obtained by maximizing the posterior distribution. (B) Posterior probability distributions (left) and example psychometric functions from heading discrimination task (right). Top row: When visual and vestibular cues are equally reliable, optimal integration results in enhanced perceptual precision. Middle row: When the two cues are in conflict with high visual reliability (high coherence), the combined estimate is biased toward the visual cue. Bottom row: When the two cues are in conflict with low visual reliability (low coherence), the combined estimate is biased toward the vestibular cue. (C) Unsupervised recalibration. The black curve indicates psychometric performance before recalibration. After recalibration, the vestibular (blue) and visual (red) heading estimates are shifted. When Δ > 0, the vestibular and visual estimates shift rightward and leftward, respectively; when Δ < 0, the vestibular and visual estimates shift leftward and rightward, respectively. Psychometric functions exhibit opposite shifts for the two modalities. (D) Supervised recalibration. The top and bottom panels correspond to conditions where the “vestibular accurate” and “visual accurate” cues are designated, respectively. Here, “accurate” indicates consistency between the cue and external feedback. In both conditions, the psychometric curves shift in the same direction toward the modality identified as accurate.
Figure 1. Bayesian optimal integration model (A,B) and recalibration processes in multisensory perception (C,D). (A) A heading stimulus S generates both visual and vestibular cues, which are modeled as independent normal distributions, p C v i s S and p C v e s t S , respectively. According to the maximum likelihood estimation principle, the combined likelihood function can be expressed as the product of the individual cue likelihoods. In the Bayesian framework, a prior distribution p S is introduced; assuming a flat prior, the posterior is determined solely by the combined likelihood. The optimal heading estimate S ^ is obtained by maximizing the posterior distribution. (B) Posterior probability distributions (left) and example psychometric functions from heading discrimination task (right). Top row: When visual and vestibular cues are equally reliable, optimal integration results in enhanced perceptual precision. Middle row: When the two cues are in conflict with high visual reliability (high coherence), the combined estimate is biased toward the visual cue. Bottom row: When the two cues are in conflict with low visual reliability (low coherence), the combined estimate is biased toward the vestibular cue. (C) Unsupervised recalibration. The black curve indicates psychometric performance before recalibration. After recalibration, the vestibular (blue) and visual (red) heading estimates are shifted. When Δ > 0, the vestibular and visual estimates shift rightward and leftward, respectively; when Δ < 0, the vestibular and visual estimates shift leftward and rightward, respectively. Psychometric functions exhibit opposite shifts for the two modalities. (D) Supervised recalibration. The top and bottom panels correspond to conditions where the “vestibular accurate” and “visual accurate” cues are designated, respectively. Here, “accurate” indicates consistency between the cue and external feedback. In both conditions, the psychometric curves shift in the same direction toward the modality identified as accurate.
Biology 14 00740 g001
Figure 2. Vestibular and visual pathways in self-motion perception. The blue arrow represents the vestibular information pathway: peripheral vestibular signals are first transmitted to the vestibular nucleus (VN) and then relayed through the thalamus, primarily projected to the parieto-insular vestibular cortex (PIVC), and may further extend to the frontal cortex. The red arrow represents the visual information pathway: visual information from the retina is transmitted through the lateral geniculate nucleus (LGN) to the primary visual cortex (V1) and subsequently projected through the middle temporal area (MT) to other visual cortical regions involved in motion direction perception. MSTd, dorsal portion of medial superior temporal area; PIVC, parieto-insular vestibular cortex; VPS, visual posterior sylvian area; VIP, ventral intraparietal area; FEFsem, smooth eye movement region of the frontal eye field; STP, superior temporal polysensory area.
Figure 2. Vestibular and visual pathways in self-motion perception. The blue arrow represents the vestibular information pathway: peripheral vestibular signals are first transmitted to the vestibular nucleus (VN) and then relayed through the thalamus, primarily projected to the parieto-insular vestibular cortex (PIVC), and may further extend to the frontal cortex. The red arrow represents the visual information pathway: visual information from the retina is transmitted through the lateral geniculate nucleus (LGN) to the primary visual cortex (V1) and subsequently projected through the middle temporal area (MT) to other visual cortical regions involved in motion direction perception. MSTd, dorsal portion of medial superior temporal area; PIVC, parieto-insular vestibular cortex; VPS, visual posterior sylvian area; VIP, ventral intraparietal area; FEFsem, smooth eye movement region of the frontal eye field; STP, superior temporal polysensory area.
Biology 14 00740 g002
Figure 3. Modulation of neuronal responses by added sensory inputs across MSTd (A) and VPS (B). (Top): The red line represents the ratio of the combined response (Rcomb) to the vestibular-only response at the vestibular preferred direction (Rvestibular-prefer) as a function of the added visual stimulus direction (0° indicates the neuron’s visual preferred direction). A ratio greater than 1 indicates that adding the visual stimulus enhances the vestibular response, while a ratio less than 1 indicates suppression. (Bottom): The blue line represents the ratio of the combined response (Rcomb) to the visual-only response at the visual preferred direction (Rvisual-prefer) as a function of the added vestibular stimulus direction (0° indicates the neuron’s vestibular preferred direction). A ratio greater than 1 indicates facilitation by the vestibular stimulus, whereas a ratio less than 1 indicates suppression.
Figure 3. Modulation of neuronal responses by added sensory inputs across MSTd (A) and VPS (B). (Top): The red line represents the ratio of the combined response (Rcomb) to the vestibular-only response at the vestibular preferred direction (Rvestibular-prefer) as a function of the added visual stimulus direction (0° indicates the neuron’s visual preferred direction). A ratio greater than 1 indicates that adding the visual stimulus enhances the vestibular response, while a ratio less than 1 indicates suppression. (Bottom): The blue line represents the ratio of the combined response (Rcomb) to the visual-only response at the visual preferred direction (Rvisual-prefer) as a function of the added vestibular stimulus direction (0° indicates the neuron’s vestibular preferred direction). A ratio greater than 1 indicates facilitation by the vestibular stimulus, whereas a ratio less than 1 indicates suppression.
Biology 14 00740 g003
Figure 4. Computational mechanisms of vestibular and visual signal integration across MSTd and VPS. (A) In the MSTd, multisensory neurons linearly integrate vestibular (blue) and visual (red) inputs, with visual signals typically dominating. The integrated signals are subsequently subject to divisive normalization via a normalization pool, resulting in a subadditive combined response (green), where the response to multisensory inputs is weaker than the sum of unimodal responses. (B) In the VPS, vestibular inputs are dominant, and divisive normalization causes the suppression of the combined response relative to the vestibular response alone.
Figure 4. Computational mechanisms of vestibular and visual signal integration across MSTd and VPS. (A) In the MSTd, multisensory neurons linearly integrate vestibular (blue) and visual (red) inputs, with visual signals typically dominating. The integrated signals are subsequently subject to divisive normalization via a normalization pool, resulting in a subadditive combined response (green), where the response to multisensory inputs is weaker than the sum of unimodal responses. (B) In the VPS, vestibular inputs are dominant, and divisive normalization causes the suppression of the combined response relative to the vestibular response alone.
Biology 14 00740 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Zeng, F. The Neural Mechanisms of Visual and Vestibular Interaction in Self-Motion Perception. Biology 2025, 14, 740. https://doi.org/10.3390/biology14070740

AMA Style

Liu J, Zeng F. The Neural Mechanisms of Visual and Vestibular Interaction in Self-Motion Perception. Biology. 2025; 14(7):740. https://doi.org/10.3390/biology14070740

Chicago/Turabian Style

Liu, Jing, and Fu Zeng. 2025. "The Neural Mechanisms of Visual and Vestibular Interaction in Self-Motion Perception" Biology 14, no. 7: 740. https://doi.org/10.3390/biology14070740

APA Style

Liu, J., & Zeng, F. (2025). The Neural Mechanisms of Visual and Vestibular Interaction in Self-Motion Perception. Biology, 14(7), 740. https://doi.org/10.3390/biology14070740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop