DOI QR코드

DOI QR Code

Coreference Resolution for Korean Pronouns using Pointer Networks

포인터 네트워크를 이용한 한국어 대명사 상호참조해결

  • Received : 2016.08.31
  • Accepted : 2017.03.01
  • Published : 2017.05.15

Abstract

Pointer Networks is a deep-learning model for the attention-mechanism outputting of a list of elements that corresponds to the input sequence and is based on a recurrent neural network (RNN). The coreference resolution for pronouns is the natural language processing (NLP) task that defines a single entity to find the antecedents that correspond to the pronouns in a document. In this paper, a pronoun coreference-resolution method that finds the relation between the antecedents and the pronouns using the Pointer Networks is proposed; furthermore, the input methods of the Pointer Networks-that is, the chaining order between the words in an entity-are proposed. From among the methods that are proposed in this paper, the chaining order Coref2 showed the best performance with an F1 of MUC 81.40 %. The method showed performances that are 31.00 % and 19.28 % better than the rule-based (50.40 %) and statistics-based (62.12 %) coreference resolution systems, respectively, for the Korean pronouns.

포인터 네트워크(Pointer Networks)는 Recurrent Neural Network (RNN)를 기반으로 어텐션 메커니즘(Attention mechanism)을 이용하여 입력 시퀀스에 대응되는 위치들의 리스트를 출력하는 딥 러닝 모델이다. 대명사 상호참조해결은 문서 내에 등장하는 대명사와 이에 대응되는 선행사를 찾아 하나의 엔티티로 정의하는 자연어처리 문제이다. 본 논문에서는 포인터 네트워크를 이용하여 대명사와 선행사의 참조관계를 밝히는 대명사 상호참조해결 방법과 포인터 네트워크의 입력 연결순서(chaining order) 여섯가지를 제안한다. 실험 결과, 본 논문에서 제안한 방법 중 연결순서 coref2 가 MUC F1 81.40%로 가장 좋은 성능을 보였다. 이는 기존 한국어 대명사 상호참조해결의 규칙 기반(50.40%)보다 31.00%p, 통계 기반(62.12%) 보다 19.28%p 우수한 성능임을 나타낸다.

Keywords

Acknowledgement

Grant : (엑소브레인-1세부) 휴먼 지식증강 서비스를 위한 지능진화형 WiseQA 플랫폼 기술 개발

Supported by : 정보통신기술진흥센터

References

  1. R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu, and P. Kuksa, "Natural language processing (almost) from scratch," The Journal of Machine Learning Research, 12, 2011.
  2. O. Vinyals, M. Fortunato and N. Jaitly, Pointer Networks, Advances in Neural Information Processing Systems, pp. 2674-2682, 2015.
  3. D. Bahdanau, K. Cho and Y. Bengio, Neural machine translation by jointly learning to align and translate, arXiv preprint arXiv:1409.0473, 2014.
  4. R. J. Williams, and D. Zipser, "A learning algorithm for continually running fully recurrent neural networks," Neural computation 1.2 (1989): 270-280, 1989. https://doi.org/10.1162/neco.1989.1.2.270
  5. H. Lee, A. Chang, Y. Peirsman, N. Chmbers, M. Surdeanu, and D. Jurafsky, "Deterministic coreference resolution based on entity-centric, precisionranked rules," Computational Linguistics 39.4: pp. 885-916, 2013. https://doi.org/10.1162/COLI_a_00152
  6. C. H. Park, G. H. Choi, and C. Lee, "Korean Coreference Resolution using the Multi-pass Sieve," Journal of KIISE, 41.11 (2014.11): 992-1005, 2014. (in Korean) https://doi.org/10.5626/JOK.2014.41.11.992
  7. A. Rahman, and V. Ng, Supervised models for coreference resolution, Proc. of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 2-Volume 2, Association for Computational Linguistics, pp. 968-977, 2009.
  8. C. Park, G. H. Choi, and C. Lee, "Korean Coreference Resolution with Guided Mention Pair Model using the Deep Learning," Proc. Of the KIISE Korea Computer Congress 2015, pp. 693-695, 2015. (in Korean)
  9. Barbara J. Grosz, Scott Weinstein, and Aravind K. Joshi, "Centering: A framework for modeling the local coherence of discourse," Computational linguistics 21.2, pp. 203-225, 1995.
  10. Lappin Shalom and Herbert J. Leass, "An algorithm for pronominal anaphora resolution," Computational linguistics 20.4 pp. 535-561, 1994.
  11. K. Cho, B. Van Merrienboer and C. Gulcehre, Learning phrase representation using RNN encoderdecoder for statistical machine translation, arXiv preprint arXiv:1406.1078, 2014.
  12. J. Chung, C. Gulcehre, K.H. Cho, and Y. Bengio, Empirical Evaluation of Gated Recurrent Networks on Sequence Modeling, arXiv preprint arXiv:1412.3555, 2014.
  13. C. Lee, J. Kim, and J. Kim, "Korean Dependency Parsing using Deep Learning," Proc. of the KIISE for HCLT (2014), pp. 87-91, 2014. (in Korean)
  14. M. Vilain, J. Burger, J. Aberdeen, D. Connolly, and L. Hirschman, "A model-theoretic coreference scoring scheme," Proc. of the 6th conference on Message understanding, Association for Computational Linguistics, pp. 45-52, 1995.
  15. S. Hochreiter and J. Schmidhuber, "Long shortterm memory," Neural computation 9.8 (1997): 1735-1780. 1997. https://doi.org/10.1162/neco.1997.9.8.1735