《Attention is All You Need》論文學習筆記

目錄 Abstract 1. Illustrated Transformer 1.1 A High-level look 1.2 Attention 1.2.1 Scale Dot-Product Attention 1.2.2 Multi-Head Attention 1.3 Positional Encoding - Representing the Order of the Sequence
相關文章
相關標籤/搜索