Dan Robinson Profile picture
Jan 8, 2023 5 tweets 1 min read Read on X
RETRO models are a giant capability unlock for LLM tech, and they're shockingly under the radar.

The first ones should come out this year. They might be even more significant than GPT-4.
You can't "teach" current LLMs, the way you'd teach an employee. If they do something bad, there isn't a good way to say "don't do that."

You can include a reminder in every prompt, but that eats up precious context space.

You can fine-tune, but you need hundreds of examples.
That's where RETRO comes in.

RETRO lets you store a gigantic database of facts and pull them into a prompt based on their contextual relevance.

You can update your fact set without retraining your model.

deepmind.com/publications/i…
That's a huge capability for any "agent" use cases. You can train your agent the way you'd train an employee!

E.g.:
- "When a customer brings up X, remember to mention Y"
- "Remember I think Z, when advocating on my behalf"
- "Don't ever say W"
This might be the most important lever for a lot of practical applications.

Once GPT-4 era models come out, LLMs are going to be damn good at answering whatever is in a prompt.

Applications will then be constrained by how complete they can make those prompts...

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Dan Robinson

Dan Robinson Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @danlovesproofs

Jan 26, 2023
Want to build an AI infra company? Here's something I'd happily pay for.

Manage embeddings for me, *including the ability to refine off-the-shelf embeddings for my use case.*

Customized embeddings, as a service.
Buried in the OpenAI cookbook is an absolute gem of a guide by @sandersted on how to do this.

Given pairs of items that should be close together in embedding space, you can train a simple map on top of the outputs of a baseline embedding model. github.com/openai/openai-…
This problem is ubiquitous and high value.

Typical LLM use case: take a corpus of content (e.g. docs, code), a user asks questions, use embeddings to try to pull relevant content into a prompt, answer.

Any app like this is very sensitive to how well the embedding lookup works!
Read 6 tweets
Jan 3, 2023
Introducing qqbot, a variant of ChatGPT that lives in your IDE.

The cool thing about qqbot is that it knows your codebase.

You can ask it questions like:
- Where is xyz implemented?
- Where are the tests for this function?
- If I want to implement xyz, where do I start?
qqbot can help you navigate a large codebase.
qqbot can draft code for you.

"Write me an example test for runQuery in lib/db.ts" --> two tests I hadn't thought to write, both of which work!
Read 8 tweets
May 16, 2022
I stepped away from Heap last month, after nine fantastic years.

Here are some things I wish I could tell my 2013 self.

I was excited to write some code, and I had no idea what was in store on the path to 350+ people and a transformative product.
1/ What the company needs from a CTO is going to change every 18 months, often dramatically. You'll pretty much never get comfortable.

Being in a leadership role means your job isn't "technology". Your job is to make the company win. Get used to adapting around that.
2/ Making the company win will often mean letting go of something you think you're good at because we need someone who is better at it, or someone who can give it more time and attention.

This is hard if you let your self worth come from doing that one thing. (So don't!)
Read 28 tweets
Nov 8, 2021
1/ What's a "technical moat"?

I get a lot of questions from SaaS founders about how to think about this.

Do technical moats apply to most SaaS products? What will stop a competitor from duplicating your product?

Here's the framework I use. 👇
2/ The point of a moat is to protect your differentiation. It's whatever would make it hard for a competitor to deliver the value you deliver.

Tech moats are interesting, because the tide of technology is always rising. It gets easier to build *anything* every year!
3/ People look for moats in the form of a singular technology: a proprietary algorithm, or a Branded Whiz-Bang Technology.

Technical moats do exist in SaaS! But they rarely take the form people expect them to take.
Read 15 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(