Vestibular
vestibular
Update on vestibular, ocular motor and cerebellar disorders
Visual-vestibular cue comparison for perception of environmental stationarity
Note the later time!
Neural mechanisms underlying visual and vestibular self-motion perception
The effect of gravity on the perception of distance and self-motion: a multisensory perspective
Gravity is a constant in our lives. It provides an internalized reference to which all other perceptions are related. We can experimentally manipulate the relationship between physical gravity with other cues to the direction of “up” using virtual reality - with either HMDs or specially built tilting environments - to explore how gravity contributes to perceptual judgements. The effect of gravity can also be cancelled by running experiments on the International Space Station in low Earth orbit. Changing orientation relative to gravity - or even just perceived orientation – affects your perception of how far away things are (they appear closer when supine or prone). Cancelling gravity altogether has a similar effect. Changing orientation also affects how much visual motion is needed to perceive a particular travel distance (you need less when supine or prone). Adapting to zero gravity has the opposite effect (you need more). These results will be discussed in terms of their practical consequences and the multisensory processes involved, in particular the response to visual-vestibular conflict.
The vestibular system: a multimodal sense
The vestibular system plays an essential role in everyday life, contributing to a surprising range of functions from reflexes to the highest levels of perception and consciousness. Three orthogonal semicircular canals detect rotational movements of the head and the otolith organs sense translational acceleration, including the gravitational vertical. But, how vestibular signals are encoded by the human brain? We have recently combined innovative methods for eliciting virtual rotation and translation sensations with fMRI to identify brain areas representing vestibular signals. We have identified a bilateral inferior parietal, ventral premotor/anterior insula and prefrontal network and confirmed that these areas reliably possess information about the rotation and translation. We have also investigated how vestibular signals are integrated with other sensory cues to generate our perception of the external environment.
Looking and listening while moving
In this talk I’ll discuss our recent work on how visual and auditory cues to space are integrated as we move. There are at least 3 reasons why this turns out to be a difficult problem for the brain to solve (and us to understand!). First, vision and hearing start off in different coordinates (eye-centred vs head-centred), so they need a common reference frame in which to communicate. By preventing eye and head movements, this problem has been neatly sidestepped in the literature, yet self-movement is the norm. Second, self-movement creates visual and auditory image motion. Correct interpretation therefore requires some form of compensation. Third, vision and hearing encode motion in very different ways: vision contains dedicated motion detectors sensitive to speed, whereas hearing does not. We propose that some (all?) of these problems could be solved by considering the perception of audiovisual space as the integration of separate body-centred visual and auditory cues, the latter formed by integrating image motion with motor system signals and vestibular information. To test this claim, we use a classic cue integration framework, modified to account for cues that are biased and partially correlated. We find good evidence for the model based on simple judgements of audiovisual motion within a circular array of speakers and LEDs that surround the participant while they execute self-controlled head movement.
Multisensory encoding of self-motion in the retrosplenial cortex and beyond
In order to successfully navigate through the environment, animals must accurately estimate the status of their motion with respect to the surrounding scene and objects. In this talk, I will present our recent work on how retrosplenial cortical (RSC) neurons combine vestibular and visual signals to reliably encode the direction and speed of head turns during passive motion and active navigation. I will discuss these data in the context of RSC long-range connectivity and further show our ongoing work on building population-level models of motion representation across cortical and subcortical networks.
The effect of gravity on the perception of distance and self-motion
Gravity is a constant in our lives. It provides an internalized reference to which all other perceptions are related. We can experimentally manipulate the relationship between physical gravity with other cues to the direction of “up” using virtual reality - with either HMDs or specially built tilting environments - to explore how gravity contributes to perceptual judgements. The effect of gravity can also be cancelled by running experiments on the International Space Station in low Earth orbit. Changing orientation relative to gravity - or even just perceived orientation – affects your perception of how far away things are (they appear closer when supine or prone). Cancelling gravity altogether has a similar effect. Changing orientation also affects how much visual motion is needed to perceive a particular travel distance (you need less when supine or prone). Adapting to zero gravity has the opposite effect (you need more). These results will be discussed in terms of their practical consequences and the multisensory processes involved, in particular the response to visual-vestibular conflict.
From oscillations to laminar responses - characterising the neural circuitry of autobiographical memories
Autobiographical memories are the ghosts of our past. Through them we visit places long departed, see faces once familiar, and hear voices now silent. These, often decades-old, personal experiences can be recalled on a whim or come unbidden into our everyday consciousness. Autobiographical memories are crucial to cognition because they facilitate almost everything we do, endow us with a sense of self and underwrite our capacity for autonomy. They are often compromised by common neurological and psychiatric pathologies with devastating effects. Despite autobiographical memories being central to everyday mental life, there is no agreed model of autobiographical memory retrieval, and we lack an understanding of the neural mechanisms involved. This precludes principled interventions to manage or alleviate memory deficits, and to test the efficacy of treatment regimens. This knowledge gap exists because autobiographical memories are challenging to study – they are immersive, multi-faceted, multi-modal, can stretch over long timescales and are grounded in the real world. One missing piece of the puzzle concerns the millisecond neural dynamics of autobiographical memory retrieval. Surprisingly, there are very few magnetoencephalography (MEG) studies examining such recall, despite the important insights this could offer into the activity and interactions of key brain regions such as the hippocampus and ventromedial prefrontal cortex. In this talk I will describe a series of MEG studies aimed at uncovering the neural circuitry underpinning the recollection of autobiographical memories, and how this changes as memories age. I will end by describing our progress on leveraging an exciting new technology – optically pumped MEG (OP-MEG) which, when combined with virtual reality, offers the opportunity to examine millisecond neural responses from the whole brain, including deep structures, while participants move within a virtual environment, with the attendant head motion and vestibular inputs.
How the brain comes to balance: Development of postural stability and its neural architecture in larval zebrafish
Maintaining posture is a vital challenge for all freely-moving organisms. As animals grow, their relationship to destabilizing physical forces changes. How does the nervous system deal with this ongoing challenge? Vertebrates use highly conserved vestibular reflexes to stabilize the body. We established the larval zebrafish as a new model system to understand the development of the vestibular reflexes responsible for balance. In this talk, I will begin with the biophysical challenges facing baby fish as they learn to swim. I’ll briefly review published work by David Ehrlich, Ph.D., establishing a fundamental relationship between postural stability and locomotion. The bulk of the talk will highlight unpublished work by Kyla Hamling. She discovered that a small (~50) population of molecularly-defined brainstem neurons called vestibulo-spinal cells act as a nexus for postural development. Her loss-of-function experiments show that these neurons contribute more to postural stability as animals grow older. I’ll end with brief highlights from her ongoing work examining tilt-evoked responses of these neurons using 2-photon imaging and the consequences of downstream activity in the spinal cord using single-objective light-sheet (SCAPE) microscopy
Self-generated vestibular prosthetic input updates forward internal model of self-motion
COSYNE 2023
Beneficial effects of alternative stimulation pulse shapes for sensory prostheses: insights from vestibular prosthesis-evoked reflexes and population neural activity
COSYNE 2025
Chronic ototoxicity induces downregulation of hair cell-specific genes in the vestibular sensory epithelium of rodents
FENS Forum 2024
Degeneration of the ascending vestibular pathway accounts for spatial navigation deficits in aged mice
FENS Forum 2024
Distinct and asymmetric responses to pitch-tilt axis and roll-tilt axis vestibular stimulation in larval zebrafish
FENS Forum 2024
Effects of vestibular function loss on spatial orientation in rats
FENS Forum 2024
Efficacy of electrical vestibular stimulation (VeNS) on adults with insomnia: A double-blind, randomized, sham-controlled trial
FENS Forum 2024
Galvanic vestibular stimulation improves visuospatial ability in healthy older adults
FENS Forum 2024
Modulation of visual responses in the mouse retina-recipient thalamus: Insights from vestibular inputs
FENS Forum 2024
Relationship between vestibular epithelium pathology and function loss in vestibular schwannoma patients
FENS Forum 2024
Static and dynamic balance after vestibular loss: Impact of the affected side
FENS Forum 2024
Inter- and Intra-hemispheric Sources of Vestibular Signals to V1
Neuromatch 5
Mathematical Models of Visual-Vestibular Integration in a Speed Accuracy Task
Neuromatch 5