Periodic Reporting for period 1 - TIME (It's about time: Towards a dynamic account of natural vision.)
Période du rapport: 2022-07-01 au 2024-12-31
TIME uses more natural viewing paradigms in both experimentation and modelling to close this gap. First, we simultaneously record high-resolution brain activity and eye-movements of participants while they explore and summarise natural visual scenes. The resulting large-scale resource will enable us to better understand the intricate mechanisms at play. In particular, it will enable us to study when and how the brain decides to move on from a given fixation location to the next, and how information from one fixation may affect the processing of the next, i.e. how information is integrated across fixations. In parallel to these analyses, we develop computational models that mirror the overall process. Here, we make use of recent developments in artificial intelligence to derive image-computable, large-scale models. A famous quote by Richard Feynman states that “What I cannot create, I do not understand”. This statement highlights the importance of computational models in science. Observing a system in experimentation and theorising about its inner mechanisms requires models to be built that put our hypotheses to the test. Therefore, in addition to building an AI-based model system, a final work package will test the alignment between brain data and the computational model.
In summary, using an interdisciplinary approach, TIME will establish when, where, and how visual semantic understanding emerges in the brain as it actively samples and integrates information from the world in a continuously updating and dynamic decision process. This approach marks a paradigm shift in how we study the visual system, and promises implications not only for neuroscience, but also for computer vision and artificial intelligence.
(1) The collection of a one-of-its-kind large-scale dataset, in which we combine magnetoencephalography with high-resolution eyetracking while participants explore natural scenes has been completed. Moreover, we finished preprocessing of all data by all participants. An alpha release of the data is ongoing.
(2) Given the new dataset, we derived a new multivariate analysis technique based on which we could show that fixation durations are largely explained by ongoing memory encoding.
(3) We have developed, implemented, and trained the first iteration of generative symbiotic networks, which combine two subnetworks, akin to the dorsal and ventral streams of the human visual system. The resulting system selects by itself where to look and how to process and integrate the information present.
(4) We have further extended artificial neural networks from AI to act as models of visual information processing in the brain. In particular, we developed network layers that are topographic in nature, contrasting the gold standard in the field. We demonstrated that networks equipped with this new network layer are more energy efficient and capture human behaviour more accurately.