o1-engineer is here! 🚀
A coding assistant built from the ground up to leverage o1 reasoning capabilities.
It can create and edit multiple files or entire folders,
plan complex projects, execute them, and write code reviews.
All from your terminal 👨💻
💬 You can chat regularly or execute commands:
/create to create files or folders
/edit to edit file, files, or folder content
/add to add files or folders to the chat context
/planning to create detailed plans
/review to create code reviews that you can use directly
Repo here. Star the repo so you can keep track of updates and improvements. ⭐️ github.com/Doriandarko/o1…
Here are some o1 prompt techniques I’ve discovered that have worked well for me so far and can likely be applied to any reasoning LLMs.
1. Be explicit about the reasoning chain
By requesting a detailed breakdown, you encourage the model to organize thoughts methodically.
2. Guide decision-making with hypothetical scenarios
Providing specific criteria guides the model through a structured analysis, yielding a more nuanced response.
3. Facilitate multi-step problem solving (especially for coding)
Breaking down the problem encourages the model to think through each step, reducing errors and enhancing clarity.
It's a great starter app for anyone to really take advantage of Claude's capabilities.
You can see Claude thinking in real-time, as well as the mood of the user interacting, if any documentation has been used, and even the category to which the customer problem is connected.
You can also see what knowledge Claude has used to answer a question based on the documents in your AWS knowledge bases.
A few important notes: 1) Goes without saying this won't be nearly as capable as Claude Engineer, but the fact that we can now create a local agentic coding assistant is nothing less than extraordinary.
See this more as an experiment rather than a proper release. 🧪
Claude Opus breaks down the task into subtasks, then deploys instances of Haiku to solve each task. Once these tasks are solved, the output goes back to Opus for review. Pretty incredible.
When I tried to build things like this with GPT4, it was very hard for the model to know when to "stop."