T5
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683 [cs.LG]
0
1
Tags
Data Science
Related
BERT
BART
T5
BERT (Bidirectional Encoder Representations from Transformers)
RoBERTa
GPT Series
LLaMA2
DeepSeek-V3
Falcon
Mistral
PaLM-450B
Gemma-7B
Gemma2
A software development team is tasked with building a feature that can automatically generate a concise, one-paragraph summary from a long news article. The system needs to first comprehend the full context of the source article and then generate a new, coherent summary. Based on the typical strengths of different foundational model designs, which of the following models would be the most suitable choice for this specific task?
Match each pre-trained model with the description that best fits its architectural design and primary use case.
Evaluating Model Architecture Selection for a Classification Task