Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Code was just part of its pretraining. All other GPT-3.5 models are fine-tuned versions of code-davinci-002.

Quote:

1 code-davinci-002 is a base model, so good for pure code-completion tasks

2 text-davinci-002 is an InstructGPT model based on code-davinci-002

3 text-davinci-003 is an improvement on text-davinci-002

4 gpt-3.5-turbo-0301 is an improvement on text-davinci-003, optimized for chat

Quote end.

https://platform.openai.com/docs/model-index-for-researchers

The reason you want a base model for code completion has nothing to do with code itself, it has to do with the fact that it completes text unlike all the instruction tuned models, which expect instructions. When you have code, there aren't necessarily any instructions present. You basically want autocomplete. That's what a base model does. But that doesn't mean it doesn't work with other things apart from code. After all, all other GPT-3.5 models are just code-davinci-002 with additional instruction and RLHF fine-tuning added, and they know countless other subject areas apart from code.

I don't get why this is so hard to understand.



It's not hard to understand. We just have a disagreement about something that you think is very important probably partly because you know more about this than I do. Have a nice day. Thanks for explaining.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: