IDEAS home Printed from https://ideas.repec.org/a/nat/natcom/v14y2023i1d10.1038_s41467-023-35834-4.html
   My bibliography  Save this article

Inferring visual space from ultra-fine extra-retinal knowledge of gaze position

Author

Listed:
  • Zhetuo Zhao

    (University of Rochester
    University of Rochester)

  • Ehud Ahissar

    (Weizmann Institute of Science)

  • Jonathan D. Victor

    (Weill Cornell Medical College)

  • Michele Rucci

    (University of Rochester
    University of Rochester)

Abstract

It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources. Here we show that contrary to this widespread assumption, the visual system has access to high-resolution extra-retinal knowledge of fixational eye motion and uses it to deduce spatial relations. Building on recent advances in gaze-contingent display control, we created a spatial discrimination task in which the stimulus configuration was entirely determined by oculomotor activity. Our results show that humans correctly infer geometrical relations in the absence of spatial information on the retina and accurately combine high-resolution extraretinal monitoring of gaze displacement with retinal signals. These findings reveal a sensory-motor strategy for encoding space, in which fine oculomotor knowledge is used to interpret the fixational input to the retina.

Suggested Citation

  • Zhetuo Zhao & Ehud Ahissar & Jonathan D. Victor & Michele Rucci, 2023. "Inferring visual space from ultra-fine extra-retinal knowledge of gaze position," Nature Communications, Nature, vol. 14(1), pages 1-12, December.
  • Handle: RePEc:nat:natcom:v:14:y:2023:i:1:d:10.1038_s41467-023-35834-4
    DOI: 10.1038/s41467-023-35834-4
    as

    Download full text from publisher

    File URL: https://www.nature.com/articles/s41467-023-35834-4
    File Function: Abstract
    Download Restriction: no

    File URL: https://libkey.io/10.1038/s41467-023-35834-4?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Ikuya Murakami & Patrick Cavanagh, 1998. "A jitter after-effect reveals motion-based stabilization of vision," Nature, Nature, vol. 395(6704), pages 798-801, October.
    2. James M. McFarland & Adrian G. Bondy & Richard C. Saunders & Bruce G. Cumming & Daniel A. Butts, 2015. "Saccadic modulation of stimulus processing in primary visual cortex," Nature Communications, Nature, vol. 6(1), pages 1-14, November.
    3. Janis Intoy & Michele Rucci, 2020. "Finely tuned eye movements enhance visual acuity," Nature Communications, Nature, vol. 11(1), pages 1-11, December.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Eric G. Wu & Nora Brackbill & Colleen Rhoades & Alexandra Kling & Alex R. Gogliettino & Nishal P. Shah & Alexander Sher & Alan M. Litke & Eero P. Simoncelli & E. J. Chichilnisky, 2024. "Fixational eye movements enhance the precision of visual information transmitted by the primate retina," Nature Communications, Nature, vol. 15(1), pages 1-15, December.
    2. Nadav Ben-Shushan & Nimrod Shaham & Mati Joshua & Yoram Burak, 2022. "Fixational drift is driven by diffusive dynamics in central neural circuitry," Nature Communications, Nature, vol. 13(1), pages 1-13, December.
    3. Jacob L. Yates & Shanna H. Coop & Gabriel H. Sarch & Ruei-Jr Wu & Daniel A. Butts & Michele Rucci & Jude F. Mitchell, 2023. "Detailed characterization of neural selectivity in free viewing primates," Nature Communications, Nature, vol. 14(1), pages 1-11, December.
    4. Yujie Wu & Tian Wang & Tingting Zhou & Yang Li & Yi Yang & Weifeng Dai & Yange Zhang & Chuanliang Han & Dajun Xing, 2022. "V1-bypassing suppression leads to direction-specific microsaccade modulation in visual coding and perception," Nature Communications, Nature, vol. 13(1), pages 1-14, December.
    5. Xaq Pitkow & Haim Sompolinsky & Markus Meister, 2007. "A Neural Computation for Visual Acuity in the Presence of Eye Movements," PLOS Biology, Public Library of Science, vol. 5(12), pages 1-14, December.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:nat:natcom:v:14:y:2023:i:1:d:10.1038_s41467-023-35834-4. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.nature.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.