Hacker News new | past | comments | ask | show | jobs | submit login
Ask HN: Llama-2-7B adapter merged with llama-2-7B-chat model?
2 points by dynamo_ 7 months ago | hide | past | favorite
I realized that I fine-tuned a llama-2-7b model and merged the adapter with llama-2-7b-chat model. This was a mistake. But when I ran inference on benchmark medqa tests and found that the llama-2-7b-chat merged with llama-2-7b adapter out performs a base llama-2-7b-chat model.

Why did this work - was this a fluke? Or are adapter's modifications compatible with both the general Llama-2-7b model and the chat-specific version? Thanks!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: