Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

How much better would fine tuning be vs few shot learning over own corpus of text?


Depends on the model size. A model like GPT3 that has hundreds of billions of paramaters, you can do few-shot learning with. You'll still pay for the tokens processed and it'll at least linearly increase response times the larger your input is.

Fine-tuning can get you similar results on smaller / faster models. The downside is you have to craft the dataset in the right way. There are trade-offs to both approaches but fwiw, I don't think Alpaca-7b can do few-shot learning.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: