AI & ML interests

AI, LLM, Agents

Recent Activity

piuzha  updated a Space 12 days ago
moxin-org/README
piuzha  published a Space 12 days ago
moxin-org/README
piuzha  updated a model about 1 month ago
moxin-org/moxin-reasoning-7b
View all activity

Moxin 7B: A Fully Open-Source 7B Language Model with Unprecedented Transparency

We’re thrilled to unveil Moxin 7B, a new milestone in open large language model (LLM) development — designed to push the boundaries of performance and openness.

In an era where many "open" LLMs lack true transparency (e.g., missing training code, data, or restrictive licenses), Moxin 7B sets a new gold standard by committing to full disclosure and reproducibility. Developed under the Model Openness Framework (MOF), Moxin 7B achieves the top classification level of Open Science, thanks to:

What we’ve open-sourced:

  • Pre-training code, data, and Moxin Base model.
  • Post-training code, data, and Moxin Instruct model.

  • RL code with GRPO, data and Moxin Reasoning model.

Performance Highlights:

  • Zero-shot / Few-shot: Outperforms Mistral, Qwen, and LLaMA on tasks like HellaSwag, ARC, MMLU, and PIQA

  • Reasoning: Moxin-Reasoning-7B achieves superior performance on MATH-500, AMC, and OlympiadBench — proving reinforcement learning can work for small 7B models

  • Training cost: ~$160K for full pretraining — efficient and reproducible at scale

Post-training Frameworks:

  • SFT and DPO with Tülu 3

  • CoT-enhanced reasoning with GRPO via DeepScaleR

Get the models and code:

  • Base model: Moxin-LLM-7B

  • Instruction model: Moxin-Instruct-7B

  • Reasoning model: Moxin-Reasoning-7B

  • Code & docs: github.com/moxin-org/Moxin-LLM

  • Arxiv paper: https://arxiv.org/abs/2412.06845

We believe this is a step toward a more transparent, reproducible, and innovation-friendly AI ecosystem — especially for researchers, developers, and startups looking to build upon a robust, open foundation. Let’s build open AI the right way.

datasets 0

None public yet