Position-aware Attention and Supervised Data Improve Slot Filling

来源:互联网 发布:民可使由之 不可使知之 编辑:程序博客网 时间:2024/05/18 04:25

作者:stanford
来源:EMNLP2017

本文贡献:公布了一批SF关系分类的语料,the TAC KBP Relation Extraction Dataset(TACRED),共有119474examples,存放于LDC
创新:将位置注意力机制与LSTM结合

已有工作的问题;
1、Although modern sequence models such as Long Short-Term Memory(LSTM) networks have gating mechanisms to control the relative influence of each individual word to the final sentence representation (Hochreiter and Schmidhuber, 1997), these controls are not explicitly conditioned on the entire sentence being classified;
2、Most existing work either does not explicitly model the positions of entities (i.e., subject and object) in the sequence, or models the positions only within a local region.

position encoder:

ps=is1,0,is2,i<s1s1is2i>s2

ui=vTtanh(Whhi+Wqq+WsPsi+Wspoi)

ai=exp(ui)j=1nexp(uj)

z=ni=1aihi

最后在z上加个全连接层,然后softmax分类

原创粉丝点击