Difference between revisions of "Attention is all you need (2017)"
Jump to navigation
Jump to search
(Blanked the page) Tags: Blanking, Mobile web edit, Mobile edit |
Tags: Mobile web edit, Mobile edit |
||
(7 intermediate revisions by 2 users not shown) | |||
Line 1: | Line 1: | ||
+ | [[wikipedia:Attention is all you need]] | ||
+ | * https://arxiv.org/abs/1706.03762 | ||
+ | Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, and [[Illia Polosukhin]] | ||
+ | |||
+ | |||
+ | * Our model achieves 28.4 [[BLEU]] on the [[WMT]] 2014 English-to-German translation task | ||
+ | |||
+ | |||
+ | == See also == | ||
+ | * {{Transformer}} | ||
+ | * {{NN}} | ||
+ | |||
+ | [[Category:ML]] |
Latest revision as of 12:34, 1 July 2024
wikipedia:Attention is all you need
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, and Illia Polosukhin
See also[edit]
- Transformer, GPT, Transformer 8, Ethched, Megatron-Core
- Artificial neural networks, Neuronal network (NN), CNN, Micrograd, NPU, ConvNet, AlexNet, GoogLeNet, Apache MXNet, Neural architecture search, DAG, Feedforward neural network, NeurIPS, Feature Pyramid Network, TPU, NPU, Apple Neural Engine (ANE), LLM, TFLOPS
Advertising: