Lutwak Hope, Rokers Bas, Simoncelli Eero P
Center for Neural Science, New York University, New York, NY USA.
Psychology, and Center for Brain and Health, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates.
ArXiv. 2025 May 13:arXiv:2505.06686v2.
As we move through the world, the pattern of light projected on our eyes is complex and dynamic, yet we are still able to distinguish between moving and stationary objects. We propose that humans accomplish this by exploiting constraints that self-motion imposes on retinal velocities. When an eye translates and rotates in a stationary 3D scene, the velocity at each retinal location is constrained to a line segment in the 2D space of retinal velocities. The slope and intercept of this segment is determined by the eye's translation and rotation, and the position along the segment is determined by local scene depth. Since all possible velocities arising from a stationary scene must lie on this segment, velocities that are not must correspond to objects moving within the scene. We hypothesize that humans make use of these constraints by using deviations of local velocity from these constraint lines to detect moving objects. To test this, we used a virtual reality headset to present rich wide-field stimuli, simulating the visual experience of translating forward in several virtual environments with varied precision of depth information. Participants had to determine if a cued object moved relative to the scene. Consistent with the hypothesis, we found that performance depended on the deviation of the object velocity from the constraint segment, rather than a difference between retinal velocities of the object and its local surround. We also found that the endpoints of the constraint segment reflected the precision of depth information available in the different virtual environments.
当我们在周围环境中移动时,投射在我们眼睛上的光的模式是复杂且动态的,但我们仍然能够区分移动物体和静止物体。我们提出,人类通过利用自我运动对视网膜速度施加的限制来实现这一点。当眼睛在静止的三维场景中平移和旋转时,每个视网膜位置的速度都被限制在视网膜速度二维空间中的一条线段上。这条线段的斜率和截距由眼睛的平移和旋转决定,而沿着线段的位置由局部场景深度决定。由于静止场景产生的所有可能速度都必须位于这条线段上,否则的话,速度必定对应于场景中移动的物体。我们假设人类通过利用局部速度与这些约束线的偏差来检测移动物体。为了验证这一点,我们使用虚拟现实头戴式设备呈现丰富的宽视野刺激,模拟在几个深度信息精度不同的虚拟环境中向前平移的视觉体验。参与者必须确定一个被提示的物体是否相对于场景移动。与该假设一致,我们发现表现取决于物体速度与约束线段的偏差,而不是物体与其局部周围区域视网膜速度的差异。我们还发现,约束线段的端点反映了不同虚拟环境中可用深度信息的精度。