Research

Capabilities and consequences of recursive, hierarchical information processing in visual systems

A study on human and robot perception and the architecture of perceptual information processing

Research Unit: 1

Project Number: 2

Example Behavior:
Individual Intelligence

Disciplines:
Psychology
Robotics

 

Principal Investigators:
Martin Rolfs
Oliver Brock

Doctoral Researchers:
Aravind Battaje

Postdoctoral Researchers:
Nina Hanning
Angelica Godinez

 

Expected Project Duration
2019 - 2024


← Projects Overview

Capabilities and consequences of recursive, hierarchical information processing in visual systems

A study on human and robot perception and the architecture of perceptual information processing

@SCIoI

This research project investigates human and robot perception and aims to develop a constructive understanding of perceptual information processing. We will create models using an information processing pattern that originated in robotics to understand perceptual mechanisms in human vision. The information processing pattern possesses similar high-level characteristics to human visual information processing, and thus has capabilities to generate similar behavior. We will apply this in an analytic-synthetic loop, simultaneously modeling while collecting psychophysical data, and adjusting the model with new observations. Based on the resulting insights we will produce an algorithmic model of human perception, capable of replicating visual phenomena and of making meaningful predictions about experimental outcomes. We will also produce robot perception algorithms that leverage insights about the human perceptual system to advance the state of the art in the synthetic disciplines.


Project Results

AICON: A Computational Framework To Exploratively Study Human Vision

The team’s main topic of investigation is a computational framework—AICON (Active InterCONnect)—and its capabilities to model robust visual perception. AICON originates from robotics and has proven effective in solving many complex robotic perception tasks. Motivated by a striking similarity between illustrations of this algorithm from robot perception and depictions of information processing in the human visual system, the team hypothesized that AICON could serve as a useful representation for exploring aspects of human vision. The similarity in illustrations highlighted a crucial aspect of information processing that is typically not leveraged deeply–rich and dynamic interactions between modules of a system. They suppose this characteristic of information processing, baked into AICON, occurs in many instances of robust systems, and thus identify this recurring pattern as a candidate principle of intelligence. Using AICON, they investigated mechanisms behind human visual perception, focusing on two visual illusions: shape-contingent color aftereffects and motion silencing. These illusions reveal intriguing interactions between basic perceptual processes in human vision, viz., between color and shape processing, and between motion and brightness (luminance) perception respectively. In modeling these illusions, the team found that the models forecasted the observations seen in humans, accounted for inter-individual differences in subjective human perception, and guided further psychophysical experiments. This demonstrates AICON’s capacity to capture biologically relevant characteristics of information processing.

Despite their simplicity, AICON-based models generate precise qualitative predictions that accurately forecast unexpected properties of the human visual system. This success stems from AICON’s structured approach to expressing dynamic, interdependent processes—a crucial aspect of perception often overlooked in both computational models and robotic system design. The team’s work shows that we can indeed articulate this simple computational pattern (or inductive bias) that occurs in artificial and biological instances. This in turn enables them to model interesting biological phenomena and gain insights into the underlying mechanisms particular to those phenomena. This computational framework has the potential to enable quick advancements in understanding intelligent information processing (both natural and artificial).

Extending The Study Of Human Vision Through Illusions to 3D Environments

The team is currently investigating the extension of one of the visual illusions studied above, in 3D using virtual reality. This phenomenon probes the interaction of color and shape, as well as spatial and action processing. Early evidence suggests that some perceptual effects held to be originating from 2D organization of visual processing in the brain may indeed originate from 3D spatial processing and attentional factors. Participants show a dependency on depth (perhaps also size, because the size depends on how far an object is) and interaction. People report stronger perception of the illusion the farther an object is in the virtual environment. They also report stronger percepts for objects they virtually interact with. This points to the need to revise theories based on controlled lab studies to account for the brain’s ecological processing in real-world environments.

Gaze Fixation and Robust Visual Perception

The team also took the investigation of effective spatial processing further by studying robots. In particular, they studied the consequences of gaze fixation for robotic perception. When humans move about, they fixate their gaze on objects of interest without conscious notice. They found that taking advantage of the regularity between an agent’s eye movements and environment enables robust 3D perception under conditions that challenge traditional 3D (depth) sensors. Their research shows that fixating on one object at a time simplifies perception for any moving agent. It provides robust and instantaneous information crucial to interact with the fixated object: how far it is and how other objects are positioned with respect to it. This information can be easily adapted to various robotic behaviors, including picking objects in challenging scenarios. Distance estimation using fixation outperforms that of a translating camera, and determining obstacles is immune to common problems for vision such as translucent and reflective surfaces. This suggests that active vision techniques like fixation that leverage the principle of agent-environment couplings reflect key aspects of biological visual intelligence and can be effectively implemented in robotic systems.

On Validating Principles of Intelligence

Finally, as a key objective of this project, the team aimed to add support to candidate principles of intelligence identified within SCIoI. Their findings above show that we gathered evidence for the validity and utility of two principles on both biological as well as artificial substrates, namely principle of active interconnections and principle of agent-environment computation.


Mengers, V., Battaje, A., Baum, M., & Brock, O. (2023). Combining Motion and Appearance for Robust Probabilistic Object Segmentation in Real Time. ICRA 2023. https://doi.org/10.1109/ICRA48891.2023.10160908
Mengers, V., Battaje, A., & Brock, O. (2025). AICON: A Representation for Adaptive Behavior. 1st German Robotics Conference.
Mengers, V., Koenig, A., Li, X., Sieler, A., Battaje, A., & Brock, O. (2025). Stop Merging, Start Separating: Why Merging Learning and Modeling Won’t Solve Manipulation but Separating the General From the Specific Will. IEEE ICRA 25 Workshop.
Godinez, A., Battaje, A., Brock, O., & Rolfs, M. (2023). Probing Perceptual Mechanism of Shape-Contingent Color after-Images via Interconnected Recursive Filters. Journal of Vision 23 / VSS 2023. https://doi.org/10.1167/jov.23.9.4885
Baum, M., Froessl, A., Battaje, A., & Brock, O. (2023). Estimating the Motion of Drawers From Sound. ICRA 2023. https://doi.org/10.1109/ICRA48891.2023.10161399
Battaje, A., Brock, O., & Rolfs, M. (2023). An Interactive Motion Perception Tool for Kindergarteners (and Vision Scientists). IPerception. https://doi.org/10.1177/20416695231159182
Battaje, A., Godinez, A., Hanning, N., Rolfs, M., & Brock, O. (2024). An Information Processing Pattern from Robotics Predicts Unknown Properties of the Human Visual System. bioRxiv. https://doi.org/10.1101/2024.06.20.599814
Battaje, A., & Brock, O. (2021). Interconnected Recursive Filters in Artificial and Biological Vision. DGR Days 2021, 32–32. https://www.static.tu.berlin/fileadmin/www/10002220/Publications/battaje21DGR.pdf
Battaje, A., & Brock, O. (2022). One Object at a Time: Accurate and Robust Structure From Motion for Robots. IROS 2022. https://doi.org/10.1109/IROS47612.2022.9981953

Research

An overview of our scientific work

See our Research Projects