Hacker News new | past | comments | ask | show | jobs | submit login
UMAF: Adapting LLMs Without Fine-Tuning (github.com/millionthodin16)
2 points by MillionthOdin16 5 days ago | hide | past | favorite | 1 comment





I'm experimenting with an approach called Universal Model Adapter Fusion to transfer capabilities LLMs without fine-tuning or distillation. The idea's to pull behaviors - like task-specific smarts or reasoning styles -from one model and patch them onto another efficiently. Think cheaper model tweaks, mixing strengths across architectures, or building custom skill sets without big retraining jobs, using an extended lora approach. Take a look at github.com/millionthodin16/umaf. It's early, but the benefits could be big. Anyone see value in this kind of approach for LLMs?



Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: