Home
Blog
Paper Reviews
Publications
Projects
Contact
Light
Dark
Automatic
Google Research
Attention Is All You Need
The Transformer, a sequence transduction model that replaces recurrent layers and relies entirely on attention mechanisms, achieves new SotA on machine translation tasks while reducing training time significantly.
Cite
×