Difference between revisions of "Attention is all you need (2017)"
Jump to navigation
Jump to search
Tags: Mobile web edit, Mobile edit |
|||
(2 intermediate revisions by one other user not shown) | |||
Line 1: | Line 1: | ||
[[wikipedia:Attention is all you need]] | [[wikipedia:Attention is all you need]] | ||
* https://arxiv.org/abs/1706.03762 | * https://arxiv.org/abs/1706.03762 | ||
+ | |||
+ | Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, and [[Illia Polosukhin]] | ||
* Our model achieves 28.4 [[BLEU]] on the [[WMT]] 2014 English-to-German translation task | * Our model achieves 28.4 [[BLEU]] on the [[WMT]] 2014 English-to-German translation task | ||
+ | |||
== See also == | == See also == |
Latest revision as of 12:34, 1 July 2024
wikipedia:Attention is all you need
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, and Illia Polosukhin
See also[edit]
- Transformer, GPT, Transformer 8, Ethched, Megatron-Core
- Artificial neural networks, Neuronal network (NN), CNN, Micrograd, NPU, ConvNet, AlexNet, GoogLeNet, Apache MXNet, Neural architecture search, DAG, Feedforward neural network, NeurIPS, Feature Pyramid Network, TPU, NPU, Apple Neural Engine (ANE), LLM, TFLOPS
Advertising: