Home > Papers

 
 
Attention based lattice bilstm model for Chinese named entity recognition
CAO Xiaofei,YANG Juan,YANG Juan *
Beijing Key Lab of Intelligent Telecommunication Software and Multimedia, Beijing University of Posts and Telecommunications, Beijing 100876
*Correspondence author
#Submitted by
Subject:
Funding: none
Opened online:25 December 2019
Accepted by: none
Citation: CAO Xiaofei,YANG Juan,YANG Juan.Attention based lattice bilstm model for Chinese named entity recognition[OL]. [25 December 2019] http://en.paper.edu.cn/en_releasepaper/content/4750113
 
 
A recently proposed model named Lattice LSTM has focused on integrating segmentation information into the long short-term memory (LSTM) network. However, it can only affect the subsequent character sequence of each character in the sequence from the level of word granularity, which results in insufficient extraction of word segmentation information. Besides, features of characters extracted by LSTM are given the same weight when transferred to the conditional random field (CRF) layer, the key semantic information does not receive much consideration. To solve the above problems, a novel neural network model is proposed in this paper which improves the original lattice model (Att-Lattice BiLSTM) with bidirectional long short-term memory based on the attention mechanism. An information path is added from the end character of word to the start character of word in the back propagation of LSTM, which integrates the word boundary information into both the start and end character of the word during bidirectional transfer of LSTM network, introducing the word information comprehensively. Moreover, this new model allows seamlessly incorporating attention mechanism to capture relatively important semantic feature automatically. Meanwhile, two strategies are provided to aggregate the bidirectional LSTM layers output to integrate semantic features effectively. Experimental results on four data sets show that the proposed model performs better than other most advanced models.
Keywords:named entity recognition; deep learning; bidirectional long short-term memory; attention mechanism; lattice network
 
 
 

For this paper

  • PDF (0B)
  • ● Revision 0   
  • ● Print this paper
  • ● Recommend this paper to a friend
  • ● Add to my favorite list

    Saved Papers

    Please enter a name for this paper to be shown in your personalized Saved Papers list

Tags

Add yours

Related Papers

Statistics

PDF Downloaded 78
Bookmarked 0
Recommend 0
Comments Array
Submit your papers