期刊论文详细信息
Frontiers in Neuroscience
Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
Seyoung Kim1  Kyungmi Noh1  Chaeun Lee1  Wonjae Ji1  Tayfun Gokmen2 
[1] Department of Materials Science and Engineering, Pohang University of Science and Technology, Pohang-si, South Korea;IBM Research AI, Yorktown Heights, NY, United States;
关键词: resistive memory;    update asymmetry;    Tiki-Taka algorithm;    neural network;    deep learning accelerator;    analog AI hardware;   
DOI  :  10.3389/fnins.2021.767953
来源: DOAJ
【 摘 要 】

Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost as well as area- and power-efficiency is theoretically predicted, the realization of such analog accelerators is largely limited by non-ideal switching characteristics of crosspoint elements. One of the most performance-limiting non-idealities is the conductance update asymmetry which is known to distort the actual weight change values away from the calculation by error back-propagation and, therefore, significantly deteriorates the neural network training performance. To address this issue by an algorithmic remedy, Tiki-Taka algorithm was proposed and shown to be effective for neural network training with asymmetric devices. However, a systematic analysis to reveal the required asymmetry specification to guarantee the neural network performance has been unexplored. Here, we quantitatively analyze the impact of update asymmetry on the neural network training performance when trained with Tiki-Taka algorithm by exploring the space of asymmetry and hyper-parameters and measuring the classification accuracy. We discover that the update asymmetry level of the auxiliary array affects the way the optimizer takes the importance of previous gradients, whereas that of main array affects the frequency of accepting those gradients. We propose a novel calibration method to find the optimal operating point in terms of device and network parameters. By searching over the hyper-parameter space of Tiki-Taka algorithm using interpolation and Gaussian filtering, we find the optimal hyper-parameters efficiently and reveal the optimal range of asymmetry, namely the asymmetry specification. Finally, we show that the analysis and calibration method be applicable to spiking neural networks.

【 授权许可】

Unknown   

  文献评价指标  
  下载次数:0次 浏览次数:3次