会议论文详细信息
11th Curtin University Technology, Science and Engineering (CUTSE) International Conference
Review of second-order optimization techniques in artificial neural networks backpropagation
工业技术(总论)
Tan, Hong Hui^1 ; Lim, King Hann^1
Department of Electrical and Computer Engineering, Curtin University Malaysia, Sarawak, Miri
98009, Malaysia^1
关键词: Curvature information;    Fast convergence;    Hyper-parameter;    Levenberg-Marquardt;    Neural network training;    Objective functions;    Optimization techniques;    Second order optimization;   
Others  :  https://iopscience.iop.org/article/10.1088/1757-899X/495/1/012003/pdf
DOI  :  10.1088/1757-899X/495/1/012003
学科分类:工业工程学
来源: IOP
PDF
【 摘 要 】

Second-order optimization technique is the advances of first-order optimization in neural networks. It provides an addition curvature information of an objective function that adaptively estimate the step-length of optimization trajectory in training phase of neural network. With the additional information, it reduces training iteration and achieves fast convergence with less tuning of hyper-parameter. The current improved memory allocation and computing power further motivates machine learning practitioners to revisit the benefits of second-order optimization techniques. This paper covers the review on second-order optimization techniques that involve Hessian calculation for neural network training. It reviews the basic theory of Newton method, quasi-Newton, Gauss-Newton, Levenberg-Marquardt, Approximate Greatest Descent and Hessian-Free optimization. This paper summarizes the feasibility and performance of optimization techniques in deep neural network training. Comments and suggestions are highlighted for second-order optimization techniques in artificial neural network training in term of advantages and limitations.

【 预 览 】
附件列表
Files Size Format View
Review of second-order optimization techniques in artificial neural networks backpropagation 570KB PDF download
  文献评价指标  
  下载次数:14次 浏览次数:51次