References

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, & Peter J. Liu (2019)

arXiv.

DOI: https://doi.org/10.48550/arxiv.1910.10683

Abstract. Introduces T5 (Text-to-Text Transfer Transformer), which reframes every NLP task as text-to-text generation and demonstrates strong transfer-learning performance using an encoder-decoder transformer.

Tags: transformer t5 transfer-learning

Cited in:

This site is currently in Beta. Contact: Chris Paton

Textbook of Usability · Textbook of Digital Health

Auckland Maths and Science Tutoring

AI tools used: Claude (research, coding, text), ChatGPT (diagrams, images), Grammarly (editing).