Generative Pre-trained Transformer

From wikieduonline
Revision as of 15:20, 24 August 2023 by Welcome (talk | contribs) (→‎Related)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

wikipedia:Generative Pre-trained Transformer

  • GPT-4 (Mar 2023)
  • GPT-3 (Jun 2020, beta) the architecture is a decoder-only transformer network with a 2048-token-long context and 175 billion parameters, requiring 800GB to store.
  • GPT-2 (Feb 2019)



Related

See also

Advertising: