LLM360 Profile picture
LLM360 is an open research lab enabling community-owned AGI through open-source large model research and development.
Dec 5 β€’ 6 tweets β€’ 4 min read
To mark the 2nd anniversary of LLM360, we are proud to release K2-V2: a 70B reasoning-centric foundation model that delivers frontier capabilities.

As a push for "360-open" transparency, we are releasing not only weights, but the full recipe: data composition, training code, logs, and intermediate checkpoints.

About K2-V2:

🧠 70B params, reasoning-optimized
🧊 512K context window
πŸ”“ "360-Open" (Data, Logs, Checkpoints)
πŸ“ˆ SOTA on olympiad math and complex logic puzzlesImage We evaluated K2 across general knowledge, STEM, coding, and agentic tool use.

The goal? To show open models need not be smaller, weaker versions of closed ones.

K2 outperforms models of similar sizes, and performs close to models that are larger.

πŸ”— Check out the model here:

Generously sponsored by @mbzuai.huggingface.co/LLM360/K2-V2Image
Apr 11 β€’ 8 tweets β€’ 4 min read
Proudly present MegaMath, the largest open-source math reasoning pretraining corpusβ€”371B tokens of high-quality mathematical web, code, and synthetic data, designed to build the data foundation for next-generation math-proficient LLMs like o1 and R1. πŸ§΅πŸ‘‡ #LLM #OpenSource #MegaMath #Math #Data4LLMs #Pretraining

Trending now on Hugging Face: huggingface.co/datasets/LLM36…MegaMath Overview πŸ” Why is this important?

Mathematical reasoning is a key feature of advanced LLMs. Training math-proficient models like O1 and DeepSeek-R1 requires large-scale, high-quality, diverse math data. Proprietary corpora, such as Qwen-2.5-Math (1T) and DeepSeekMath (120B), show strong mathematical abilities but are closed source. Existing open corpora lack comparable size and quality. MegaMath aims to bridge this gap.Image