Attention Is All You Need阅读笔记

文章目录 4. Attention Is All You Need 1. Embedding 1.1 Word Embedding 1.2 Position Embedding 2. Encoder 2.1 Multi-Head Self-attention 2.1.1 Self-attention 2.1.2 Multi-head attention 2.2 Feed-Forward Netwo
相关文章
相关标签/搜索