《Attention Is All You Need》

Proposal propose the Transformer, a model architecture eschewing recurrence and instead relying entirely on an attention mechanism to draw global dependencies between input and output. Contributions 1
相关文章
相关标签/搜索