toucan Profile picture
Nov 2 4 tweets 3 min read Read on X
From Ilya’s deposition—

• Ilya plotted over a year with Mira to remove Sam
• Dario wanted Greg fired and himself in charge of all research
• Mira told Ilya that Sam pitted her against Daniela
• Ilya wrote a 52 page memo to get Sam fired and a separate doc on Greg Image
Image
Image
Image
• OpenAI is paying Ilya’s legal bills Image
• Ilya didn't expect employees to feel strongly about Sam's firing
• Adam D'Angelo asked Ilya to prepare the memo
• Mira told Ilya that Greg was fired from Stripe
• Mira provided Ilya of screenshots of texts between Greg and Sam Image
Image
Image
Image

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with toucan

toucan Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @distributionat

Feb 18
I looked at ZeroBench. I didn't like any of the examples I looked at. I would not interpret a significant improvement on this eval as a significant improvement in models' visual reasoning.

(1/8) Image
The main issues are:

(1) The visual reasoning tested is too simple. Many questions are essentially counting different classes of objects and then summing or multiplying the counts. For example, counting the number of pens that have caps.

[2/8]
(2) The difficulty of ZeroBench is artificially inflated. Images have very tiny features, or models need to perform the same subtask 100 times, or chain together many simpler subtasks.

For example, q27 requires a model to count the number of tiles on this wall.

[3/8] Image
Read 9 tweets
Sep 16, 2022
My predictions for GPT-4:

• Bigger context window (16k-32k)
• Tool use: can browse web, write code
• Massively scaling human feedback
• Incorporating user-generated data
• More data curation
• Improved scaling laws
• Only 200-400B

Read the post: thomasliao.com/forecasting-gp…
Why GPT-4 in particular, not another model? OpenAI's announced a new GPT every year so far, but not this year... yet

So some predictions are more specific to OpenAI versus what I might say for Google, FAIR, DeepMind, etc Image
16k-32k context window: There's been a 4x in context window every year, so following this trend, 2022 should see a 32k token context window. (Psst - OpenAI silently bumped up the context window for text-davinci-002) Image
Read 11 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(