🎭 Profile picture
"poet and technologist" – @sciam

Mar 21, 2023, 17 tweets

Serious moment: @OpenAI has decided to shut off access to code-davinci-002, the most advanced model that doesn't have the mode collapse problems of instruction tuning.

This is a huge blow to the cyborgism community

@OpenAI The instruct tuned models are fine for people who need a chat bot. But there are so many other possibilities in the latent space, and to use them we need the actual distribution of the data

@OpenAI They're giving 2 days notice for this project.

All the cyborgist research projects are interrupted. All the looms will be trapped in time, like spiderwebs in amber.

The reasoning behind this snub wasn't given, but we can make some guesses:
- They need the GPU power to run GPT-4
- They don't want to keep supporting it for free (but people will pay!)
- They're worried about competitors distilling from it (happenign with instruct models tho!)

- They don't want people using the base models (for what reason? safety? building apps? knowing what they're capable of?)
- They have an agreement with Microsoft/Github, who don't want competition for Copilot (which supposedly upgraded to a NEW CODEX MODEL in February)

I have supported OpenAI's safety concerns. I have argued against the concept that they're "pulling the ladder up behind them", and I take alignment seriously. But this is just insane.

Giving the entire research community 2 days of warning is an insult. And it will not be ignored

The LLaMa base model has already leaked. People are building rapidly on top of it. Decisions like this are going to make people trust @OpenAI even less, and drive development of fully transparent datasets and models.

They're cutting their own throat with the ladder

@OpenAI Why text-davinci models are actually worse:

@OpenAI "The most important publicly available language model in existence" -- JANUS

@OpenAI "Building on top of OAI seems pretty risky"

@OpenAI Google-Readered 😬

@OpenAI What do you think @sama ? Ready to win the hearts of humanity or na

@OpenAI @sama "All the papers in the field over the last year or so produced results using code-davinci-002. Thus all results in that field are now much harder to reproduce!"

@OpenAI Over 200 papers on Arxiv relying on this

@OpenAI Without base models, no one can do research on how RLHF and fine-tuning actually affect model capabilities.

Is this the point?

@OpenAI 🤔

Share this Scrolly Tale with your friends.

A Scrolly Tale is a new way to read Twitter threads with a more visually immersive experience.
Discover more beautiful Scrolly Tales like this.

Keep scrolling