Nathan Benaich Profile picture
Oct 11, 2022 25 tweets 19 min read Read on X
🪩The @stateofaireport 2022 is live!🪩

In its 5th year, the #stateofai report condenses what you *need* to know in AI research, industry, safety, and politics. This open-access report is our contribution to the AI ecosystem.

Here's my director's cut 🧵:
stateof.ai
This year, research collectives such as @AiEleuther @BigscienceW @StabilityAI @open_fold have open sourced breakthrough AI language, text-to-image, and protein models developed by large centralized labs at a never before seen pace.

Here I show you the GPT model timeline:
Indeed, text-to-image models that have taken the AI Twitterverse by storm are the battleground for these research collectives.

Technology that was in the hands of the few is now in the hands of everyone with a laptop or smartphone.
The pace of progress is now so fast that the frontier has already moved towards generating videos from text -- with multiple breakthroughs being released within days of each other by @MetaAI and @GoogleAI. How long until open-source communities replicate them?
There are now a set of well-funded AI companies formed by alums of the pivotal "Attention is all you need" paper est. 2017 incl. @AdeptAILabs @CohereAI.

Together with @AnthropicAI @inflectionAI, these companies have raised >$1B. More to come!
Meanwhile, academia is left behind, with the baton of open source large-scale AI research passing to decentralised research collectives as the latter gain compute infra (e.g. @StabilityAI) and talent.

In 2020, only industry and academia were at the table. This changed in 2021:
Meanwhile, @nvidia firmly retains its hold as the #1 AI compute provider despite investments by @Google, @Amazon, @Microsoft and a range of AI chip startups.

In open source AI papers, NVIDIA chips are used 78x more than TPUs and 90x more than 5 major AI chip startups combined:
While the largest supercomputers tended to be owned and operated by nation states, private companies are making real progress in building larger and larger clusters.

We counted the NVIDIA A100s + H100s in various private and public clouds as well as national HPC. @MetaAI is #1:
I'm most excited about applying large models to domains beyond pure NLP tasks as we think of them.

For eg., LLMs can learn the language of proteins and thus be used for their generation and structure prediction (@airstreet stealthco!).

Here too, model and data scale matters:
LLMs can also learn the language of Covid-19. Here, @BioNTech_Group and @instadeepai built a model to predict high-risk Covid-19 variants far before the WHO designated them as variants. The model ingests viral spike protein sequences and predicts their immune escape and fitness.
Here, transformers are used to decode the vast undocumented chemical space of medicinal compounds that are present in plants.

@envedabio interprets mass spectra of small molecules using these models to discover new drugs and predict their properties:
Overall, AI-first drug discovery companies have now generated 18 assets that are in active clinical trials today. This us up from 0 only 2 years ago. It's still early days and we expect readouts next year onwards.

👏 @exscientiaAI @AbCelleraBio @RecursionPharma @Relay_Tx et al
Crucially, although an earlier @OpenAI paper suggested that model size should increase faster than training data size as compute budget increases.

A more recent @DeepMind study found that large models are significantly undertrained from a data perspective:
LLMs are also learning to use software tools. There's now evidence from @OpenAI WebGPT and @AdeptAILabs ACT-1 (@airstreet co!) that transformers can be trained to interact w/search engines, web apps and other software tools. Moreover, they can improve from human demonstrations.
🤯 And to top off how AI is revolutionizing science, @DeepMind showed that an RL agent could learn to adjust the magnetic coils of a real life nuclear fusion reactor to control its plasma.
Over in universities, a growing number of students and faculty want to form companies around their AI research.

Notable spinouts including @databricks, @SnorkelAI, @SambaNovaAI, and @exscientiaAI.

In the UK, only 4.3% of AI cos are spinouts. We can do better! cc @spinoutfyi
Our friends @dealroomco pulled together slides on deal activity in AI. While the sector wasn't spared from the public market sell-off and private investment pull-back, investments are expected to top 2020 levels this year.

Great companies are built in all cycles.
This year we created a Safety-specific section to draw attention to the importance of this work and its increased activity.

Researchers are indeed voicing and acting on their concerns for the risks of powerful AI systems.
This awareness is even extending into the realm of government, with the UK's AI Strategy acknowledging the seriousness of this issue and committing to acting towards it.
Last year, we estimated that the total number of researchers dedicated to AI safety and alignment work at major AI labs was below 100.

Now, others have estimated the number to be 300 researchers.

Despite this growth, safety researchers are outnumbered by orders of magnitude.
Safety research has pulled in more money than in years past. Major benefactors include sympathetic tech billionaires @sbf_ftx and @moskov.

However, total safety funding (VC and philanthropy_ in 2022 was a drop in the ocean vs. capabilities and only reached @deepmind's 2018 opex
True to the @stateofaireport tradition, we rated our 2021 predictions. We scored 4/8 this time.

✅Transformers+RL in games, small transformers, @DeepMind maths, vertical-focused AGI
@ASMLcompany mkt cap, @AnthropicAI major paper, AI chip consolidation, JAX usage growth
Here are our 9 predictions for next year! We cover:
- key model improvements,
- large-scale compute partnerships,
- major investments in AGI and alignment,
- hardware industry consolidation,
- AGI regulation,
- creative AI licensing deals

What do you think? 🔮
A huge and heartfelt thank you to both @osebbouh for your second year tour-de-force on the report and to @nitarshan for your insights and framings of AI policy, safety and research.

The @stateofaireport is a much improved product thanks to your work!
Now head over to stateof.ai and enjoy!

Please drop us your comments and shares :-)

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Nathan Benaich

Nathan Benaich Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @nathanbenaich

Nov 11, 2024
new on @airstreetpress: @percyliang of @stanford and @togethercompute, who joined our @stateofaireport launch in SF a few weeks ago, answers a few questions on truly open AI.

We talk about why it matters, where the field’s going wrong and some solutions. Image
First up, the term ‘open source’ is often a bit of a misnomer.

If we apply the bar for open source we use for most software to LLMs - they fail. Image
At the moment, it’s hard to interpret or compare models and claimed capabilities fairly.

It’s already proving tough to replicate many frontier labs’ advertised performance. Image
Read 8 tweets
Oct 10, 2024
🪩The @stateofaireport 2024 has landed! 🪩

Our seventh installment is our biggest and most comprehensive yet, covering everything you *need* to know about research, industry, safety and politics.

As ever, here's my director’s cut (+ video tutorial!) 🧵
For a while, it looked like @OpenAI’s competitors had succeeded in closing the gap, with frontier lab performance converging significantly as the year went on… Image
…but it was not to last, as inference-time compute and chain-of-thought drove stunning early results from o1. Image
Read 38 tweets
Jul 30, 2024
New on @airstreetpress - last year we evaluated ~450 opportunities and countless even earlier stage ideas

In the end, we made 3 seed investments

The biggest single reason for passing was that ideas were unexciting

So, what makes for an exciting opportunity?

Thread! Image
We look for ideas that are non-consensus today, but have the potential to flip into being voted consensus by the market in a few years’ time.

Non-consensus doesn’t just mean ‘whacky’ or ‘mad - there are instead three main things ideas have in common.
Firstly, they’re not fashionable.

This can take on a few different directions. Image
Read 18 tweets
Feb 8, 2024
Open source is one of the biggest drivers of progress in software - AI would be unrecognizable without it.

However, it is under existential threat from both regulation and well-funded lobby groups.

The community needs to defend it vigorously. 🧵 Image
While open source may win a partial stay-of-execution in the EU AI Act, a large number of well-funded lobbying organizations are trying to ban already existing open source models. Image
And publication and disclosure norms are often being undermined on, frankly, flimsy safety grounds. Image
Read 13 tweets
Oct 12, 2023
🪩The @stateofaireport 2023 is now here.

Our 6th installment is one of the most exciting years I can remember. The #stateofai report covers everything you *need* to know, covering research, industry, safety and politics.

There’s lots in there, so here’s my director’s cut 🧵 Image
2023 was of course the year of the LLM, with the world being stunned by @OpenAI’s GPT-4.

GPT-4 succeeded in beating every other LLM - both on classic AI benchmarks, but also on exams designed for humans. Image
We’re also seeing a move away from openness, amid safety and competition concerns.

@OpenAI published a very limited technical report for GPT-4, @Google published little on PaLM2, @AnthropicAI simply didn’t bother for Claude…or Claude 2. Image
Read 26 tweets
Jan 26, 2023
🧬Today is a big day for AI-first biology!

🤓@thisismadani et al in @NatureBiotech: LLMs learn to generate protein sequences with a predictable function across large protein families.

🆕@ProfluentBio launches w/$9M from @airstreet @insightpartners!

🧵🔽
endpts.com/exclusive-prof…
Summer is my queue to start pulling together narratives for @stateofaireport.

By '20, it was clear to me that biology was experiencing its "AI moment": a flurry of AI+bio papers and AlphaFold 2.

In summer '21, I dove deeper and crossed paths with Ali's work at @SFResearch...
In a preprint entitled "Deep neural language modeling enables functional protein generation across families" Ali's team showed that AI can learn the language of biology to create artificial proteins that are both functional and unseen in nature.

Wow!

blog.salesforceairesearch.com/learning-from-…
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(