Research PaperAttention Is All You Need (Transformer)14 pages
|14 pages
By Ashish Vaswani, Noam Shazeer, Niki Parmar, et al.
Attention Is All You Need (Transformer)
Attention-only sequence transduction architecture that replaces recurrence and convolution while improving translation quality and training efficiency.
TransformerSelf-AttentionSequence Modeling