Effective Approaches to Attention-based Neural Machine Translation
来源:互联网 发布:js获取当前点击的元素 编辑:程序博客网 时间:2024/05/17 04:52
An attentional mechanism has lately been used to improve neural machine translation (NMT) by selectively focusing on parts of the source sentence during translation. However, there has been little work exploring useful architectures for attention-based NMT. This paper examines two simple and effective classes of attentional mechanism: a global approach which always attends to all source words and a local one that only looks at a subset of source words at a time. We demonstrate the effectiveness of both approaches over the WMT translation tasks between English and German in both directions. With local attention, we achieve a significant gain of 5.0 BLEU points over non-attentional systems which already incorporate known techniques such as dropout. Our ensemble model using different attention architectures has established a new state-of-the-art result in the WMT'15 English to German translation task with 25.9 BLEU points, an improvement of 1.0 BLEU points over the existing best system backed by NMT and an n-gram reranker.
- Effective Approaches to Attention-based Neural Machine Translation
- [EMNLP2015]Effective Approaches to Attention-based Neural Machine Translation
- Neural Machine translation中的Attention机制
- 神经网络机器翻译Neural Machine Translation: Attention Mechanism
- 神经网络机器翻译Neural Machine Translation(2): Attention Mechanism
- Towards String-to-Tree Neural Machine Translation
- A novel approach to neural machine translation
- [ACL2017]Sequence-to-Dependency Neural Machine Translation
- 神经网络机器翻译Neural Machine Translation(5): Gradient-based Optimization Algorithms
- Adversarial Neural Machine Translation
- 三大机器翻译技术的high-level概述:Neural, Rule-Based and Phrase-Based Machine Translation
- NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE
- Neural Machine Translation and Sequence-to-sequence Models: A Tutorial
- Google's Neural Machine Translation System
- Neural Machine Translation(NMT)技术概述
- Modeling Coverage for Neural Machine Translation
- Massive Exploration of Neural Machine Translation Architectures
- 【nlp论文阅读】Adversal Neural Machine Translation
- Html position(static、relative、absolute、fixed)
- hdu 1255 覆盖的面积(线段树+扫描线——面积交)
- HTML关键字<6>
- 安装 JPEG Toolbox
- POJ 1426 Find The Multiple DFS
- Effective Approaches to Attention-based Neural Machine Translation
- 指纹框架--指纹选项加载(一)
- mapreduce初印象
- 模拟实现通讯录-动态
- 段落向量与句子向量表达
- 接口设计----依赖倒置原则
- JDBC_事务
- HDU 1233 还是畅通工程
- iOS 面试题归纳整理—进阶篇(持续更新中)