Arash Yazdanbakhsh¹, Celi Gagliardi²
¹Boston University, ²Boston University
Knowing one’s location in space is crucial for navigation, especially in unknown environments. Recently, systems involved in self-localization for spatial navigation have been found in rodent brains, but the mechanisms for self-localization are not completely understood/ Human behavioral experiments can enhance understanding self-localization systems in human brains. In this study, human observers are placed in a visual virtual reality (VR) to perform a self-positioning task. The VR is created using a 120 Hz projector and a polarizing filter which separates left and right eye images when passed through 3D glasses. Disparity and size cues generate the perception of depth. Participants are placed in a virtual position on the ground for give seconds and then moved by changing their virtual environment around a stable object. Their task is to return to their initial positon on the ground with button controls. Principal component analyses show that errors in self-repositioning are not along any particular axis and each participant has a unique pattern of repositioning. Trial durations do not affect accuracy of repositioning and lack of disparity cues increases the standard error of repositioning in all directions. Some participants have lower errors when initial self-positions appear to be on one or the other side of the stable object, suggesting a link between memory of self-position and a preferred point reference. Future directions of this project are to explore between-subject differences and the effect of stimulus presentation at different frequencies on memory of self-position.
Keywords: Self-Position, Visual Virtual Reality