Hacker News new | past | comments | ask | show | jobs | submit login

Is this also censored/nerfed? I'd love to play with a "raw" unnerfed model to fully grasp what an LLM can do (and see how biased it is). Does anyone have any recommendations for unnerfed models to try out?



LLaMA 65B is the best uncensored model we've got, and the Airoboros fine-tuning if you want it to follow instructions.



The most powerful available foundation model is code-davinci-002, a.k.a. GPT-3.5. It's only available on Azure since OpenAI removed it from their own Playground and API for some reason.


All 3 text-davinci models are available on openAI's api. including 3 (which is the GPT-3.5 gen). Code-davinci-002 is a code-tuned model, You can see a nice visual summary of the relationships between the openAI models at https://yaofu.notion.site/How-does-GPT-Obtain-its-Ability-Tr...

Or the official source is https://platform.openai.com/docs/model-index-for-researchers


> All 3 text-davinci models are available on openAI's api.

That's irrelevant because these are all fine-tuned.

> Code-davinci-002 is a code-tuned model

No, "code-tuned" isn't even a thing. It is a foundation model, which consists purely of pretreating. No fine-tuning is involved.

> Or the official source is

The official source says exactly what I just said.


OK perhaps I used slightly the wrong term. The docs[1] say that code-davinci-002 is "optimized for code completion tasks" though so it seems unlikely to fulfil the OPs purpose of playing around with an unaligned/sweary model which was my main point. Some of the uncensored models from huggingface would probably serve that purpose much better.

[1] see the entry for code-davinci-002 in https://platform.openai.com/docs/models/gpt-3-5


Code was just part of its pretraining. All other GPT-3.5 models are fine-tuned versions of code-davinci-002.

Quote:

1 code-davinci-002 is a base model, so good for pure code-completion tasks

2 text-davinci-002 is an InstructGPT model based on code-davinci-002

3 text-davinci-003 is an improvement on text-davinci-002

4 gpt-3.5-turbo-0301 is an improvement on text-davinci-003, optimized for chat

Quote end.

https://platform.openai.com/docs/model-index-for-researchers

The reason you want a base model for code completion has nothing to do with code itself, it has to do with the fact that it completes text unlike all the instruction tuned models, which expect instructions. When you have code, there aren't necessarily any instructions present. You basically want autocomplete. That's what a base model does. But that doesn't mean it doesn't work with other things apart from code. After all, all other GPT-3.5 models are just code-davinci-002 with additional instruction and RLHF fine-tuning added, and they know countless other subject areas apart from code.

I don't get why this is so hard to understand.


It's not hard to understand. We just have a disagreement about something that you think is very important probably partly because you know more about this than I do. Have a nice day. Thanks for explaining.


The model isn't available at all?


It is available in the sense that it is accessible. The weights are not available for download of course, but the OP wanted to "play around" with it, for which only access is required. There is no other accessible foundation model that can compete with GPT-3.5.


Why are you guys downvoting me?


Because GPT 3.5 not very good compared to LLaMA 65b or even 33b finetunes, from my testing.

Also because 3.5 is not really available?


Have you actually tested code-davinci-002?


Maybe you mean gpt-3.5-turbo or text-davinci-003? Or GPT-4 (technically in beta so not fully available to everyone)?


No, those are all fine-tuned models which are "nerfed" in the terminology of the OP. I mean code-davinci-002, the GPT-3.5 base model.


code-davinci models are finetuned on code so I don't think that's what the OP wants. For reference the family tree is here https://platform.openai.com/docs/model-index-for-researchers


As the website you linked says, code-davinci-002 is not fine-tuned. It is the GPT-3.5 base model.


Is that what nerfed means? I usually see "nerfed" used in a way that means that it will refuse to answer certain topics. "I can't answer that as it would violate copyright" and such.


The fine-tuned models are certainly censored and not "raw".


But doesn't code-davinci-002 also have OpenAI's filters in between you and the model?


Yes, but that's different from the model itself being fine-tuned.


> I mean... the GPT-3.5 base model

That would be text-davinci-003, I believe.


No, text-davinvi-003 is fine-tuned. The base model is code-davinci-002. See https://platform.openai.com/docs/model-index-for-researchers




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: