Abstract. Introduces GPT-2, a 1.5-billion-parameter decoder-only transformer demonstrating that sufficiently large language models exhibit impressive zero-shot performance on many NLP tasks without task-specific fine-tuning.
Tags:transformergptlanguage-modelsurl-only
This site is currently in Beta. Contact: Chris Paton