After discussing w/ colleagues and taking some time to collect my thoughts,
Here are the top 3 things GPT-4 enables that were previously impossible or extremely difficult to achieve – including examples 👇
1/ Image comprehension
⇒ Unlocks full-screen app and browser automation 🤖
We thought text-based co-pilot and writing assistants were cool, but now that GPT-4 natively understands images, automating app and browser workflows based on screenshots is going to skyrocket 🚀
Here's an example to get your creative juices flowing:
You can think of this as the AI getting more RAM like in the early days of computers. You can do amazing things w/ low RAM, but developers need to do al sorts of low-level tricks, and it's harder to scale.
With longer context, GPT-4 will have more memory, cohesion, and consistency for larger inputs.
⇒ With gpt-4-32k, apps can now ingest entire documents, websites, short novels, repos, transcriptions, etc, and it will "just work". This makes "training" the AI in-place much easier.
Apps will also have the capacity to generate much longer, more cohesive & consistent outputs.
With GPT-3, you could generate short poems and essays, and chain these together to generate longer text.
⇒ With gpt-4-32k, you can now generate entire short novels, stories, and entire apps.
This is going to massively boost no-code tools and help so many more people build / write / create without having to worry about the lower-level details 💯
Previous LLMs like GPT-3, ChatGPT, Claude, LLaMA, etc all struggled with different types of SOTA reasoning tasks. Same with cross-language tasks.
You can mitigate these shortcomings to an extent via prompt engineering, chaining prompts and custom logic, etc, but it's difficult.
⇒ GPT-4 will unlock many new use cases involving more advanced reasoning and cross-language tasks, without as much need for sophisticated developer tooling and custom logic.
A great example of this is it's ability to do taxes 🤯
⇒ GPT-4 will unlock even more use cases in education. Better reasoning and consistency across longer users sessions greatly improves the UX for tutor / teaching scenarios.
A great example of this is yesterday's Duolingo release of their AI tutor:
Here's a comparison of how to use the KMeans algorithm in Python vs TS to find clusters of related points in datasets.
The Python version on the left uses the official "scikit-learn" PyPI package, and the TypeScript version on the right uses the new "sklearn" NPM package.
Here's another example comparing the StandardScaler class in Python vs TypeScript.
You'll notice that the two code snippets are almost identical, since a key goal of this project is to take advantage of "scikit-learn"'s amazing existing ecosystem of docs & tutorials.
I've been running @ChatGPTBot on twitter for the past two months, and it recently passed:
• 29k followers
• 100k public conversations 🔥
• 6.8k stars on github for the underlying NPM package
Here's a breakdown of the bot's usage and analytics over time: 🧵
For starters, @ChatGPTBot is a simple twitter bot that waits for users to tag it with a question and then responds with an answer generated using ChatGPT's unofficial API.