tensorflow实现seq2seq模型细节(4):tensorflow nmt中的attention(scaled luong 和 normed bahdanau)和optimizer

1.attention   Tensorflow的nmt教程中这样提到: Attention: Bahdanau-style attention often requires bidirectionality on the encoder side to work well; whereas Luong-style attention tends to work well for differen
相关文章
相关标签/搜索