Attention Is All You Need
2017 research paper by Google / From Wikipedia, the free encyclopedia
"Attention Is All You Need" is a landmark[1][2] 2017 research paper authored by eight scientists working at Google, responsible for expanding 2014 attention mechanisms proposed by Bahdanau et al. into a new deep learning architecture known as the transformer. The paper is considered by some to be a founding document for modern artificial intelligence, as transformers became the main architecture of large language models.[3][4] At the time, the focus of the research was on improving Seq2seq techniques for machine translation, but even in their paper the authors saw the potential for other tasks like question answering and for what is now called multimodal Generative AI.[5]
The paper's title is a reference to the song "All You Need Is Love" by the Beatles.[6]
As of 2024,[update] the paper has been cited more than 100,000 times.[7]