Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Got any links to explanations of why fine tuning open models isn’t a productive solution? Besides renting the GPU time, what other downsides exist on today’s SOTA open models for doing this?


When the new pre-trained parameters come out in a new model generation, your old fine tuning doesn't apply to them.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: