Adept Profile picture
Sep 14, 2022 7 tweets 3 min read Read on X
1/7 We built a new model! It’s called Action Transformer (ACT-1) and we taught it to use a bunch of software tools. In this first video, the user simply types a high-level request and ACT-1 does the rest. Read on to see more examples ⬇️
2/7 This can be especially powerful for manual tasks and complex tools — in this example, what might ordinarily take 10+ clicks in Salesforce can be now done with just a sentence.
3/7 Working in-depth in tools like spreadsheets, ACT-1 demonstrates real-world knowledge, infers what we mean from context, and can help us do things we may not even know how to do.
4/7 The model can also complete tasks that require composing multiple tools together; most things we do on a computer span multiple programs. In the future, we expect ACT-1 to be even more helpful by asking for clarifications about what we want.
5/7 The internet contains a lot of knowledge about the world! When the model doesn’t know something, it knows how to just look up the information online (seen here in voice input mode).
6/7 ACT-1 doesn’t know how to do everything, but it’s highly coachable. With 1 piece of human feedback, it can correct mistakes, becoming more useful with each interaction.
7/7 Read more at adept.ai/act. We’re only scratching the surface — if you’re as excited about useful general intelligence as we are, apply at adept.ai/careers, or visit adept.ai/alpha to join the waitlist for the alpha release of our upcoming product.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Adept

Adept Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @AdeptAILabs

Oct 18, 2023
We’re open-sourcing a multimodal model: Fuyu-8B! Building useful AI agents requires fast foundation models that can see the visual world.

Fuyu-8B performs well at standard image understanding benchmarks, but it also can do a bunch of new stuff (below)

adept.ai/blog/fuyu-8b
We think MM models are especially useful for handling unstructured knowledge worker data, so we’ve given Fuyu-8B capabilities in:

- Understanding diagrams, charts, and graphs
- Doing OCR on screens
- Outputting bounding boxes for the locations of objects on screens
- Answering UI-based questions
It actually has an extremely simple architecture. Fuyu-8B doesn’t have an image encoder. This allows easy interleaving of text and images and handling arbitrary image resolutions! And it’s super fast for copilot use cases where latency really matters. Image
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(