Hacker News new | past | comments | ask | show | jobs | submit login

The model card is intended to help reproduce the HumanEval results, which we've achieved with no prompt.

However, the model is instruction-tuned to follow completions (not chats). Simply tell it what you want and it should work.




Yeah but I assume it has a prompt format, for example CodeLlama34B Instruct the prompt format is

[INST] Write a function in Python that prints HelloWorld [/INST]


Nope, just ask your question. It's not chat-tuned the same way as CodeLlama34B Instruct.


Interesting. That seems a pretty odd choice. Without a prompt format (like almost all instruction finetunes have) the model might continue your instruction instead of answering the instruction, since it does not have a delimiter for where the instruction ends and the response begins.


This is an early experiment that we did in just a few hours. We'll have much better models in just a few days, stay tuned :)




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: