Integrating the What and Where Visual Pathways to Improve CNN Categorisation

Abstract

Convolutional Neural Networks (CNNs) have been widely used for categorisation tasks over the past decades. Many studies have attempted to improve their performance by increasing model complexity, adding parameters, or adopting alternative architectures such as transformers, which excel at large-scale benchmarks. However, these approaches often come at a high computational cost. We take a different approach, prioritizing ecological plausibility to achieve high accuracy with minimal computational cost. We focus on visual search — a task requiring both localisation and categorisation of a target object in natural scenes. Our work is inspired by the organisation of the primate visual system, which processes visual information through two distinct pathways: the ventral ‘‘What’’ pathway, responsible for object recognition, and the dorsal ‘‘Where’’ pathway, specialized in spatial localisation. Using this principle, we aim to evaluate the validity of a ‘‘what/where’’ approach, capable of selectively processing only the relevant areas of the visual scene with respect to the classification task. This selection relies on the implementation of a visual sensor (‘‘retina’’) that samples only part of the image, coupled with a map representing the regions of the image. This map, referred to as a ‘’likelihood map’’ is based on the probability of correctly identifying the target label. Depending on the case, it can be guided (resp not guided) by the target label, similar to the Grad-CAM (resp DFF). In both scenarios, we show improved classification performance when the eye shifts toward the region of interest, outperforming previously mentioned methods. Surprisingly, the gain in classification accuracy is offset by a reduction in the precision of object localisation within the scene. Beyond its computational benefits, this What-Where framework serves as an experimental tool to further investigate the neural mechanisms underlying visual processing.

Publication
Computational Cognitive Neuroscience Society Meeting (CCN) 2025

🧠 TODAY at #CCN2025 ! Poster A145, 1:30-4:30pm at de Brug & E‑Hall. We’ve developed a bio-inspired “What-Where” CNN that mimics primate visual pathways - achieving better classification with less computation. Come chat! 🎯

  • What: Poster A145
  • When: Tuesday, August 12, 1:30 – 4:30 pm,
  • Where: CCN 2025 conference venue, de Brug & E‑Hall

Presented by main author Jean-Nicolas JÉRÉMIE and in cosupervision with Emmanuel Daucé

https://laurentperrinet.github.io/publication/jeremie-25-ccn/

Our research introduces a novel “What-Where” approach to CNN categorization, inspired by the dual pathways of the primate visual system:

  • The ventral “What” pathway for object recognition

  • The dorsal “Where” pathway for spatial localization

Key innovations:

✅ Bio-inspired selective attention mechanism

✅ Improved classification performance with reduced computational cost

✅ Smart visual sensor that samples only relevant image regions

✅ Likelihood mapping for targeted processing

The results?

Better accuracy while using fewer resources - proving that nature’s designs can still teach us valuable lessons about efficient AI.

Come find us this afternoon for great discussions!

#CCN2025 #ComputationalNeuroscience #AI #MachineLearning #BioinspiredAI #ComputerVision #Research

Links:

Jean-Nicolas Jérémie
Jean-Nicolas Jérémie
Phd candidate in Computational Neuroscience

During my PhD, I am focusing on ultra-fast processing in event-based neural networks.

Laurent U Perrinet
Laurent U Perrinet
Researcher in Computational Neuroscience

My research interests include Machine Learning and computational neuroscience applied to Vision.