Difference between revisions of "Attention is all you need (2017)"
Jump to navigation
Jump to search
Line 5: | Line 5: | ||
* Our model achieves 28.4 [[BLEU]] on the [[WMT]] 2014 English-to-German translation task | * Our model achieves 28.4 [[BLEU]] on the [[WMT]] 2014 English-to-German translation task | ||
− | |||
== See also == | == See also == |
Revision as of 14:18, 24 January 2023
wikipedia:Attention is all you need
See also
- Transformer, GPT, Transformer 8, Ethched, Megatron-Core
- Artificial neural networks, Neuronal network (NN), CNN, Micrograd, NPU, ConvNet, AlexNet, GoogLeNet, Apache MXNet, Neural architecture search, DAG, Feedforward neural network, NeurIPS, Feature Pyramid Network, TPU, NPU, Apple Neural Engine (ANE), LLM, TFLOPS
Advertising: