Convolutional Neural Networks (CNNs) have been widely used for categorisation tasks over the past decades. Many studies have attempted to improve their performance by increasing model complexity, adding parameters, or adopting alternative architectures such as transformers, which excel at large-scale benchmarks. However, these approaches often come at a high computational cost. We take a different approach, prioritizing ecological plausibility to achieve high accuracy with minimal computational cost. We focus on visual search — a task requiring both localisation and categorisation of a target object in natural scenes. Our work is inspired by the organisation of the primate visual system, which processes visual information through two distinct pathways: the ventral ‘‘What’’ pathway, responsible for object recognition, and the dorsal ‘‘Where’’ pathway, specialized in spatial localisation. Using this principle, we aim to evaluate the validity of a ‘‘what/where’’ approach, capable of selectively processing only the relevant areas of the visual scene with respect to the classification task. This selection relies on the implementation of a visual sensor (‘‘retina’’) that samples only part of the image, coupled with a map representing the regions of the image. This map, referred to as a ‘’likelihood map’’ is based on the probability of correctly identifying the target label. Depending on the case, it can be guided (resp not guided) by the target label, similar to the Grad-CAM (resp DFF). In both scenarios, we show improved classification performance when the eye shifts toward the region of interest, outperforming previously mentioned methods. Surprisingly, the gain in classification accuracy is offset by a reduction in the precision of object localisation within the scene. Beyond its computational benefits, this What-Where framework serves as an experimental tool to further investigate the neural mechanisms underlying visual processing.
🧠 TODAY at #CCN2025 ! Poster A145, 1:30-4:30pm at de Brug & E‑Hall. We’ve developed a bio-inspired “What-Where” CNN that mimics primate visual pathways - achieving better classification with less computation. Come chat! 🎯
Presented by main author Jean-Nicolas JÉRÉMIE and in cosupervision with Emmanuel Daucé
https://laurentperrinet.github.io/publication/jeremie-25-ccn/
Our research introduces a novel “What-Where” approach to CNN categorization, inspired by the dual pathways of the primate visual system:
The ventral “What” pathway for object recognition
The dorsal “Where” pathway for spatial localization
Key innovations:
✅ Bio-inspired selective attention mechanism
✅ Improved classification performance with reduced computational cost
✅ Smart visual sensor that samples only relevant image regions
✅ Likelihood mapping for targeted processing
The results?
Better accuracy while using fewer resources - proving that nature’s designs can still teach us valuable lessons about efficient AI.
Come find us this afternoon for great discussions!
#CCN2025 #ComputationalNeuroscience #AI #MachineLearning #BioinspiredAI #ComputerVision #Research
Links: