ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1712.01586
25
308

Deep Semantic Role Labeling with Self-Attention

5 December 2017
Zhixing Tan
Mingxuan Wang
Jun Xie
Yidong Chen
X. Shi
ArXivPDFHTML
Abstract

Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the relationships between two tokens regardless of their distance. Our single model achieves F1=83.4_1=83.41​=83.4 on the CoNLL-2005 shared task dataset and F1=82.7_1=82.71​=82.7 on the CoNLL-2012 shared task dataset, which outperforms the previous state-of-the-art results by 1.81.81.8 and 1.01.01.0 F1_11​ score respectively. Besides, our model is computationally efficient, and the parsing speed is 50K tokens per second on a single Titan X GPU.

View on arXiv
Comments on this paper