基于BiLSTM并结合自注意力机制和句法信息的隐式篇章关系分类
详细信息    查看全文 | 推荐本文 |
  • 英文篇名:BiLSTM-based Implicit Discourse Relation Classification Combining Self-attention Mechanism and Syntactic Information
  • 作者:凡子威 ; 张民 ; 李正华
  • 英文作者:FAN Zi-wei;ZHANG Min;LI Zheng-hua;School of Computer Sciences and Technology,Soochow University;
  • 关键词:神经网络 ; 隐式篇章关系分类 ; 自注意力机制 ; 句法信息
  • 英文关键词:Neural network;;Implicit discourse relation classification;;Self-attention mechanism;;Syntactic information
  • 中文刊名:JSJA
  • 英文刊名:Computer Science
  • 机构:苏州大学计算机科学与技术学院;
  • 出版日期:2019-05-15
  • 出版单位:计算机科学
  • 年:2019
  • 期:v.46
  • 基金:国家自然科学基金项目(61525205,61876116)资助
  • 语种:中文;
  • 页:JSJA201905035
  • 页数:7
  • CN:05
  • ISSN:50-1075/TP
  • 分类号:221-227
摘要
隐式篇章关系分类是浅层篇章结构分析(Shallow Discourse Parsing)中的子任务,也是自然语言处理(Natural Language Processing,NLP)中的一项重要任务。隐式篇章关系是由篇章关系中的论元对推理出来的逻辑语义关系。隐式篇章关系的分析结果可以应用于许多自然语言处理任务中,如机器翻译、自动文档摘要、问答系统等。针对隐式篇章关系分类任务,提出一种基于自注意力机制和句法信息的方法。通过双向长短时记忆网络(Bidirectional Long Short-Term Memory Network)对输入的结合句法信息的论元对进行建模,将论元对表示成低维稠密的向量;通过自注意力机制对论元对信息进行筛选。在PDTB2.0数据集上进行实验,结果表明该方法较基准系统获得了更好的效果。
        Implicit discourse relation classification is a sub-task in shallow discourse parsing,and it's also an important task in natural language processing(NLP).Implicit discourse relation is a logic semantic relation inferred from the argument pairsin discourse relations.The analytical results of the implicit discourse relationship can be applied to many na-tural language processing tasks,such as machine translation,automatic document summarization,and question answe-ring system.This paper proposed a method based on self-attention mechanism and syntactic information for the classification task of implicit discourse relations.In this method,Bidirectional Long Short-Term Memory Network(BiLSTM) is used to model the inputted argument pairs with syntactic information and express the argument pairs into low-dimension dense vectors.The argument pair information was screened by the self-attention mechanism.At last,this paper conducted experiments on PDTB2.0 dataset.The experimental results show that the proposed model achieves better effects than the baseline system.
引文
[1] POPESCU-BELIS A,MEYER T.Using Sense-Labeled Dis- course Connectives forStatistical Machine Translation[C]//Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2012:129-138.
    [2] JANSEN P,SURDEANU M,CLARK P.Discourse Complements Lexical Semanticsfor Non-factoid Answer Reranking[C]//Proceedings of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2014:977-986.
    [3] LOUIS A,JOSHI A,ENKOVA A.Discourse Indicators for Content Selectionin Summarization[C]//Proceedings of the Special Interest Group on Discourse and Dialogue.Pennsylvania,USA:Association for Computational Linguistics,2010:147-156.
    [4] PITLER E,NENKOVA A.Using Syntax to Disambiguate Explicit Discourse Connectives in Text[C]//Proceedings of the ACL-IJCNLP 2009 Conference Short Papers.Pennsylvania.USA:Association for Computational Linguistics,2009:13-16.
    [5] PRASAD R,DINESH N,LEE A,et al.The Penn Discourse TreeBank 2.0[C]//Proceedings of the International Conference on Language Resources and Evaluation.Paris,France:European Language Resources Association,2008:2961-2968.
    [6] EDDY S.Hidden Markov models[J].Current Opinion in Structural Biology,1996,6(3):361-365.
    [7] RATNAPARKHI A.A Maximum Entropy Model for Part-of-Speech Tagging[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing.Pennsylvania.USA:Association for Computational Linguistics,1996:133-142.
    [8] COLLINS M.Discriminative Training Methods for Hidden Markov Models:Theoryand Experiments with Perceptron Algorithms[C]//Proceedings of the Annual Meeting of the Association for Computational Linguistics.Pennsylvania.USA:Associationfor Computational Linguistics,2002:1-8.
    [9] CHANG C C,LIN C J.LIBSVM:A library for support vector machines[M].ACM,2011:1-27
    [10] LAFFERTY J,MCCALLUM A,PEREIRA F.Conditional Random Fields:Probabilistic Models for Segmenting and Labeling Sequence Data[C]//Proceedingsof the International Conference on Machine Learning.Massachusetts,USA:TheInternational Machine Learning Society,2001:282-289.
    [11] PITLER E,LOUIS A,NENKOVA A.Automatic Sense Prediction for Implicit Discourse Relations in Text[C]//Proceedings of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2009:683-691.
    [12] LIN Z H,KAN M Y,NG H T.Recognizing Implicit Discourse Relations inthe Penn Discourse Treebank[C]//Proceedings of Empirical Methods in Natural Language Processing.Pennsylvania,USA:Association for Computational Linguistics,2009:343-351.
    [13] WANG W T,SU J,TAN C L.Kernel Based Discourse Relation Recognition with Temporal Ordering Information[C]//Procee-dings of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2010:710-719.
    [14] RUTHERFORD A,XUE N W.Discovering Implicit Discourse Relations Through Brown Cluster pair Representation and Coreference Patterns[C]//Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2014:645-654.
    [15] QIN L H,ZHANG Z S,ZHAO H.Shallow Discourse Parsing Using Convolutional Neural Network[C]//Proceedings of the Conference on Computational Natural Language Learning-Shared Task.Pennsylvania,USA:Association for Computational Linguistics,2016:70-77.
    [16] SCHENK N,CHIARCOS C,DONANDT K,et al.Do We Really Need All Those Rich Linguistic Features?A Neural Network-Based Approach to Implicit Sense Labeling[C]//Proceedings of the Conference on Computational Natural Language Learning-Shared Task.Pennsyl-vania,USA:Association for Computatio-nal Linguistics,2016:41-49.
    [17] WEISS G,BAJEC M.Discourse Sense Classification from Scratch using Focused RNNs[C]//Proceedings of the Confe-rence on Computational Natural Language Learning-Shared Task.Pennsylvania,USA:Association for Computational Linguistics,2016:50-54.
    [18] CHEN J F,ZHANG Q,LIU P F,et al.Implicit Discourseelation Detection via a Deep Architecture with Gated Relevance Network[C]//Proceedings of the Association for Computational Linguistics.Pennsylvania,USA:Association for Computational Linguistics,2016:1726-1735.
    [19] DOZAT T,MANNING C D.Deep Biaffine Attention for Neural Dependency Parsing[C]//Proceedings of 5th International Conference on Learning Representations.2017:24-26.
    [20] ZHANG B,SU J,XIONG D,et al.Shallow Convolutional Neural Network for Implicit Discourse Relation Recognition[C]//Proceedings of Empirical Methods in Natural Language Processing.Pennsylvania,USA:Association for Computational Linguistics,2015:2230-2235.
    [21] RUTHERFORD A,XUE N.Improving the Inference of Implicit Discourse Relations via Classifying Explicit Discourse Connectives[C]//Proceedings of Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.Pennsylvania,USA:Association for Computational Linguistics,2015:799-808.
    [22] LIU Y,LI S.Recognizing Implicit Discourse Relations via Re- peated Reading:Neural Networks with Multi-Level Attention[C]//Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing.Pennsylvania,USA:Association for Computational Linguistics,2016:1224-1233.
    [23] LIU Y,LI S,ZHANG X,et al.Implicit discourse relation classification via multi-task neural networks[C]//Thirtieth AAAI Conference on Artificial Intelligence.USA:AAAI Press,2016:2750-2756.
    [24] LAN M,WANG J,WU Y,et al.Multi-task Attention-based Neural Networks for Implicit Discourse Relationship Representation and Identification[C]//Proceedings of the 2017 Confe-rence on Empirical Methods in Natural Language Processing.Pennsylvania,USA:Association for Computational Linguistics,2017:1299-1308.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700