
Pinkstack/Superthoughts-lite-v2-MOE-Llama3.2-experimental-0427
Text Generation
•
Updated
•
36
•
1
Trained to reason properly.
Note MoE reasoning model with 3.91B params in total, and 2 experts active.
Note GGUF of superthoughts lite v2
Note Non-experimental version of lite. 1.7B parameters, GRPO & SFT trained, not very conversational though.
Note Non-experimental version of lite. 1.7B parameters, GRPO & SFT trained. GGUF version.