Fine-grained Information Status Classification Using Discourse Context-Aware Self-Attention

08/13/2019
by   Yufang Hou, et al.
0

Previous work on bridging anaphora recognition (Hou et al., 2013a) casts the problem as a subtask of learning fine-grained information status (IS). However, these systems heavily depend on many hand-crafted linguistic features. In this paper, we propose a discourse context-aware self-attention neural network model for fine-grained IS classification. On the ISNotes corpus (Markert et al., 2012), our model with the contextually-encoded word representations (BERT) (Devlin et al., 2018) achieves new state-of-the-art performances on fine-grained IS classification, obtaining a 4.1 improvement compared to Hou et al. (2013a). More importantly, we also show an improvement of 3.9 complex hand-crafted semantic features designed for capturing the bridging phenomenon.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset