Hacker News new | past | comments | ask | show | jobs | submit login

Yes but fine tuning requires a lot more gpu memory and is thus much more expensive, complicated and out of reach of most people. To fine tune a >10B model you still need multiple A100 / H100. Let’s hope that changes with quantized fine tuning, forward pass only etc.



The OpenLLM team is actively exploring those techniques for streamlining the fine-tuning process and making it accessible!


You can fine-tune medium models 3..60B on a single GPU with QLoRA


What is the $ cost of a fine tune though? $500?


Can you fine tune on an M2 with adequate memory?




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: