These new models are Hybrid Reasoners - meaning you can toggle ON and OFF the long chain of thought reasoning whenever you want a short, intuitive answer, or a long, well reasoned higher accuracy answer, now available on our API and to download on HuggingFace.
DeepHermes 24B Preview performs extremely well on reasoning tasks with reasoning mode ON, jumping over 4x in accuracy on hard math problems, and 43% on GPQA, a STEM based QA benchmark.
Built on @MistralAI's excellent Mistral-Small-24B open model, its a perfect size for quantization on consumer GPUs.
With reasoning mode off, it performs comparably to Mistral's own instruct variant.
Feb 13 • 5 tweets • 3 min read
Introducing DeepHermes-3 Preview, a new LLM that unifies reasoning and intuitive language model capabilities.
DeepHermes 3 is built from the Hermes 3 datamix, with new reasoning data, creating a model that can toggle on and off long chains of thought for improved accuracy at the cost of more test time compute!huggingface.co/NousResearch/D…
This is our first work on reasoning models, and hope our unique approach to user controlled, toggleable reasoning mode furthers our mission of giving those who use DeepHermes more steerability for whatever need they have.
These early benchmarks show extreme improvement in Mathematical reasoning capabilities when enabled, as well as a modest improvement in GPQA (Google Proof Question Answering) benchmarks
Jan 27 • 4 tweets • 2 min read
Recent AI breakthroughs challenge the status quo narrative that only closed, mega labs have the ability to push the frontier of superintelligence.
Today we announce Nous Psyche built on @Solana - a cooperative training network for generative AI. Psyche coordinates heterogeneous hardware to join a run and train open-source models.
We retell the myth of Psyche — a mortal’s quest for retribution against divine odds:
Read more in our blog post:
Nous Research announces the pre-training of a 15B parameter language model over the internet, using Nous DisTrO and heterogeneous hardware contributed by our partners at @Oracle, @LambdaAPI, @NorthernDataGrp, @CrusoeCloud, and the Andromeda Cluster.
This run presents a loss curve and convergence rate that meets or exceeds centralized training.
Our paper and code on DeMo, the foundational research that led to Nous DisTrO, is now available (linked below).
You can watch the run LIVE here: distro.nousresearch.com
We harness both Nous DisTrO, our novel networking stack that reduces inter-GPU communication by up to 10,000x during pretraining, and the testnet code for Psyche, a decentralized network that builds on Nous DisTrO to autonomously coordinate compute for model training and more.
Psyche details coming soon.
Nov 12, 2024 • 4 tweets • 2 min read
Today we are launching the Forge Reasoning API Beta, an advancement in inference time scaling that can be applied to any model or a set of models, for a select group of people in our community.
The Forge Reasoning engine is capable of dramatically improving Hermes 70B to reach parity in some categories with OpenAI's o1 (full), at the cost of more inference compute.
The API is built upon three architectures developed at Nous:
1. Monte Carlo Tree Search (MCTS) 2. Chain of Code (CoC) 3. Mixture of Agents (MoA)
Together, these three techniques create a powerful reasoning system that outputs complex, flexible, and nuanced responses from LLMs. Elevating open-source ai to the level of frontier models has been a core principle of Nous since its inception.