Predictive processing of upcoming scene views in immersive environments: evidence from continuous flash suppression

Poster Presentation 36.326: Sunday, May 19, 2024, 2:45 – 6:45 pm, Banyan Breezeway
Session: Scene Perception: Virtual environments, intuitive physics

Anna Mynick1 (), Michael A. Cohen2, Adithi Jayaraman1, Kala Goyal1, Caroline E. Robertson1; 1Dartmouth College, 2Amherst College

Although our visual environment is immersive, we explore it in discrete and fleeting glimpses. How do we overcome our limited field of view to attain our continuous sense of visual space? Previous studies show that memory for a visual stimulus can speed perceptual awareness (Jiang et al., 2007). Here, we used virtual reality (VR) to test whether memory for immersive environments likewise facilitates perceptual awareness of upcoming scene views across head turns. Participants (N=29) first studied immersive, real world scenes drawn from the local campus in head mounted VR (Study Phase). In each trial of a subsequent priming task, a studied scene (prime) was first presented, then fully occluded. Participants then head-turned (left/right) toward a target image. The target, presented to the non-dominant eye, was initially masked by a dynamic Mondrian presented to the dominant eye (CFS). Participants’ task was to detect the target, which either contained a spatially congruent view of the prime (e.g. the left view following a left head turn) or a spatially incongruent view (e.g. the right view following a left head turn). To ensure true target detection, only half of the target was displayed (a semi-circle) and participants indicated which side of the circle the target was on (left/right). Participants detected incongruent scene views faster than congruent ones (t(28) = 2.27, p = .031), suggesting that memory-based predictions affect the speed of perceptual awareness for scene information across head turns, favoring unexpected over expected input. This interpretation dovetails with a predictive processing account of visual processing, wherein top-down predictions suppress responses to expected sensory input, allowing deviations from expected input to be represented (Walsh et al., 2020). More broadly, this work underscores the possibility that memory-based predictions support efficient processing of visual input across fields-of-view as we move our eyes, heads, and bodies in space.