期刊论文详细信息
Fixexd point theory and applications
Learning without loss
article
Elser, Veit1 
[1] Department of Physics, Cornell University
关键词: Fixed-point algorithm;    Projection method;    Douglas–Rachford;    Divide and concur;    Machine learning;    Artificial neural network;    Nonnegative matrix factorization;   
DOI  :  10.1186/s13663-021-00697-1
来源: SpringerOpen
PDF
【 摘 要 】

We explore a new approach for training neural networks where all loss functions are replaced by hard constraints. The same approach is very successful in phase retrieval, where signals are reconstructed from magnitude constraints and general characteristics (sparsity, support, etc.). Instead of taking gradient steps, the optimizer in the constraint based approach, called relaxed–reflect–reflect (RRR), derives its steps from projections to local constraints. In neural networks one such projection makes the minimal modification to the inputs x, the associated weights w, and the pre-activation value y at each neuron, to satisfy the equation$x\cdot w=y$ . These projections, along with a host of other local projections (constraining pre- and post-activations, etc.) can be partitioned into two sets such that all the projections in each set can be applied concurrently—across the network and across all data in the training batch. This partitioning into two sets is analogous to the situation in phase retrieval and the setting for which the general purpose RRR optimizer was designed. Owing to the novelty of the method, this paper also serves as a self-contained tutorial. Starting with a single-layer network that performs nonnegative matrix factorization, and concluding with a generative model comprising an autoencoder and classifier, all applications and their implementations by projections are described in complete detail. Although the new approach has the potential to extend the scope of neural networks (e.g. by defining activation not through functions but constraint sets), most of the featured models are standard to allow comparison with stochastic gradient descent.

【 授权许可】

Unknown   

【 预 览 】
附件列表
Files Size Format View
RO202108090000103ZK.pdf 2990KB PDF download
  文献评价指标  
  下载次数:9次 浏览次数:2次