期刊论文详细信息
NEUROCOMPUTING 卷:205
Bio-inspired unsupervised learning of visual features leads to robust invariant object recognition
Article
Kheradpisheh, Saeed Reza1,5  Ganjtabesh, Mohammad1  Masquelier, Timothee2,3,4,5 
[1] Univ Tehran, Sch Math Stat & Comp Sci, Dept Comp Sci, Tehran, Iran
[2] INSERM, U968, F-75012 Paris, France
[3] Univ Paris 06, Univ Sorbonne, Inst Vis, UMR S 968, F-75012 Paris, France
[4] CNRS, UMR 7210, F-75012 Paris, France
[5] Univ Toulouse, CERCO, CNRS, UMR 5549, F-31300 Toulouse, France
关键词: View-Invariant object recognition;    Visual cortex;    STDP;    Spiking neurons;    Temporal coding;   
DOI  :  10.1016/j.neucom.2016.04.029
来源: Elsevier
PDF
【 摘 要 】

Retinal image of surrounding objects varies tremendously due to the changes in position, size, pose, illumination condition, background context, occlusion, noise, and non-rigid deformations. But despite these huge variations, our visual system is able to invariantly recognize any object in just a fraction of a second. To date, various computational models have been proposed to mimic the hierarchical processing of the ventral visual pathway, with limited success. Here, we show that the association of both biologically inspired network architecture and learning rule significantly improves the models' performance when facing challenging invariant object recognition problems. Our model is an asynchronous feed-forward spiking neural network. When the network is presented with natural images, the neurons in the entry layers detect edges, and the most activated ones fire first, while neurons in higher layers are equipped with spike timing-dependent plasticity. These neurons progressively become selective to intermediate complexity visual features appropriate for object categorization. The model is evaluated on 3D-Object and ETH-80 datasets which are two benchmarks for invariant object recognition, and is shown to outperform state-of-the-art models, including DeepConvNet and HMAX. This demonstrates its ability to accurately recognize different instances of multiple object classes even under various appearance conditions (different views, scales, tilts, and backgrounds). Several statistical analysis techniques are used to show that our model extracts class specific and highly informative features. (C) 2016 Elsevier B.V. All rights reserved.

【 授权许可】

Free   

【 预 览 】
附件列表
Files Size Format View
10_1016_j_neucom_2016_04_029.pdf 6419KB PDF download
  文献评价指标  
  下载次数:1次 浏览次数:0次