期刊论文详细信息
Linguistic Issues in Language Technology
Using Deep Neural Networks to Learn Syntactic Agreement
Jean-Philippe Bernardy1  Shalom Lappin2 
[1] University of Gothenburg;University of Gothenburg, King's College London, Queen Mary University of London
关键词: deep neural networks;    syntactic agreement;    lexical embeddings;    CNNs;    LSTM RNNs;    long distance dependencies;   
DOI  :  
学科分类:社会科学、人文和艺术(综合)
来源: C S L I Publications
PDF
【 摘 要 】

We consider the extent to which different deep neural network (DNN) configurations can learn syntactic relations, by taking up Linzen et al.’s (2016) work on subject-verb agreement with LSTM RNNs. We test their methods on a much larger corpus than they used (a ~24 million example part of the WaCky corpus, instead of their ~1.35 million example corpus, both drawn from Wikipedia). We experiment with several different DNN architectures (LSTM RNNs, GRUs, and CNNs), and alternative parameter settings for these systems (vocabulary size, training to test ratio, number of layers, memory size, drop out rate, and lexical embedding dimension size). We also try out our own unsupervised DNN language model. Our results are broadly compatible with those that Linzen et al. report. However, we discovered some interesting, and in some cases, surprising features of DNNs and language models in their performance of the agreement learning task. In particular, we found that DNNs require large vocabularies to form substantive lexical embeddings in order to learn structural patterns. This finding has interesting consequences for our understanding of the way in which DNNs represent syntactic information. It suggests that DNNs learn syntactic patterns more efficiently through rich lexical embeddings, with semantic as well as syntactic cues, than from training on lexically impoverished strings that highlight structural patterns. .

【 授权许可】

Unknown   

【 预 览 】
附件列表
Files Size Format View
RO201904039094830ZK.pdf 288KB PDF download
  文献评价指标  
  下载次数:10次 浏览次数:31次