Omer Levy sur Twitter : "What if I told you that fine-tuning T5-Large (0.8B params) on a couple hundred examples could outperform GPT-3 (175B params) on a bunch of tasks?"
Tags:
About This Document
File info
Documents with similar tags (experimental)