学位论文详细信息
Recurrent neural networks for grammatical inference
recurrent neural network
Zeng, Zheng ; Goodman, Rodney M.
University:California Institute of Technology
Department:Engineering and Applied Science
关键词: recurrent neural network;   
Others  :  https://thesis.library.caltech.edu/4946/1/Zeng_z_1994.pdf
美国|英语
来源: Caltech THESIS
PDF
【 摘 要 】

In this thesis, various artificial recurrent neural network models are investigated for the problem of deriving grammar rules from a finite set of example "sentences." A general discrete network framework and its corresponding learning algorithm are presented and studied in detail in learning three different types of grammars.

The first type of grammars considered is regular grammars. Experiments with conventional analog recurrent networks in learning regular grammars are presented to demonstrate the unstable behavior of such networks in processing very long strings after training. A new network structure to force recurrent networks to learn stable states by discretizing the internal feedback signals is constructed. For training such discrete networks a "pseudo-gradient" learning rule is applied.

As an extension to the discrete network model, an external discrete stack is added to accommodate the inference of context-free grammars. A composite error function is devised to deal with various situations during learning. The pseudo-gradient method is also extended to train such a network to learn context-free grammars with the added option of operating on the external stack.

Another extension to the discrete network structure is made for the purpose of learning probabilistic finite state grammars. The network consists of a discrete portion which is intended to represent the structure of the grammar, and an analog portion which is intended to represent the transition probabilities. Two criteria for the network to verify the correctness of its solution during training are proposed. Theoretical analysis of the necessary and sufficient conditions for the correct solution is presented.

Experimental results show that the discrete network models have similar capabilities in learning various grammars as their analog counterparts, and have the advantage of being provably stable.

【 预 览 】
附件列表
Files Size Format View
Recurrent neural networks for grammatical inference 5913KB PDF download
  文献评价指标  
  下载次数:6次 浏览次数:20次