Math Mini 0.6B (Preview)

Math Mini 0.6B (Preview) is a compact, specialized model developed by Enosis Labs as part of the "Mini Series." It is designed to deliver efficient and precise mathematical reasoning, with a realistic and practical focus for its size. This model is fine-tuned from unsloth/Qwen3-0.6B-unsloth-bnb-4bit.

Philosophy & Capabilities

The Mini Series, along with the "Enosis Math" and "Enosis Code" models, incorporates step-by-step reasoning by default, enabling more efficient, clear, and well-founded answers. All models in the Math series have been trained with carefully curated step-by-step problem-solving datasets, resulting in a greater ability to reason and explain solutions in a structured way.

Math Mini 0.6B (Preview) is optimized for:

  • Basic Algebra: Solving equations and manipulating expressions.
  • Arithmetic & Sequential Reasoning: Calculations and breaking down problems into logical steps.
  • Elementary Logic: Applying deduction in mathematical contexts.
  • Introductory Competition Problem Solving: Focus on foundational skills adapted to the model's scale.

Larger models in the "Enosis Math" series address advanced topics such as calculus, higher algebra, and olympiad problems. The "Code Mini" and "Enosis Code" series are oriented towards programming and algorithmic tasks, maintaining the same philosophy of explicit and efficient reasoning.

This model is a preview version and is under continuous improvement and evaluation.

Quick Start

Available in both Hugging Face Transformers and quantized GGUF formats.

Transformers (Hugging Face)

Ensure you have the latest transformers library. For Qwen3 models, a recent version is recommended.

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model_id = "enosislabs/math-mini-0.6b-preview-gguf"

pipe = pipeline("text-generation", model=model_id, trust_remote_code=True)

messages = [
    {"role": "system", "content": "You are a helpful math assistant."},
    {"role": "user", "content": "Solve for x: 3x + 11 = 35"},
]

formatted_prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(formatted_prompt, max_new_tokens=100)
print(outputs[0]["generated_text"])

# Alternatively, load the model and tokenizer directly:
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)
inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=100)
response_text = tokenizer.decode(outputs[0], skip_special_tokens=False)
print(response_text)

GGUF with Ollama

Download the .gguf file from Hugging Face and use Ollama. You can choose between other GGUF versions such as 4bit, 5bit, and 8bit. This example uses the 4bit version:

ollama run enosislabs/math-mini-0.6b-preview-gguf:Q4_K_M

For more control, create a Modelfile with the Qwen3 template:

FROM ./math-mini-0.6b-preview-Q4_K_M.gguf
TEMPLATE """
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
"""

Then run:

ollama create math-mini-0.6b -f Modelfile
ollama run math-mini-0.6b

GGUF with llama.cpp

./main -m ./path/to/math-mini-0.6b-preview.gguf -n 256 -p "<|im_start|>system\nYou are a helpful math assistant.<|im_end|>\n<|im_start|>user\nSolve for x: 2x + 5 = 15<|im_end|>\n<|im_start|>assistant\n" --temp 0.2 -c 2048

vLLM (Transformers)

pip install vllm
python -m vllm.entrypoints.openai.api_server --model enosislabs/math-mini-0.6b-preview-gguf --trust-remote-code

For chat:

curl -X POST "http://localhost:8000/v1/chat/completions" \
    -H "Content-Type: application/json" \
    --data '{
        "model": "enosislabs/math-mini-0.6b-preview-gguf",
        "messages": [
            {"role": "system", "content": "You are a helpful math assistant."},
            {"role": "user", "content": "What is the capital of France?"}
        ],
        "max_tokens": 50,
        "temperature": 0.2
    }'

Prompt Format (Qwen3 ChatML)

For best results, use the Qwen3 ChatML format. The tokenizer.apply_chat_template method handles this automatically.

<|im_start|>system
You are a helpful AI assistant. Provide a detailed step-by-step solution.
<|im_end|>
<|im_start|>user
{user_question}
<|im_end|>
<|im_start|>assistant

Acknowledgements

  • Fine-tuned from unsloth/Qwen3-0.6B-unsloth-bnb-4bit.
  • Training process accelerated and optimized thanks to Unsloth.

Citation

If you use this model, please cite:

@software{enosislabs_math_mini_0.6b_preview_2025,
  author = {{Enosis Labs}},
  title = {{Math Mini 0.6B (Preview)}},
  year = {2025},
  publisher = {Hugging Face},
  version = {0.1-preview},
  url = {https://huggingface.co/enosislabs/math-mini-0.6b-preview-gguf}
}
Downloads last month
92
GGUF
Model size
596M params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for enosislabs/math-mini-0.6b-preview-gguf

Finetuned
Qwen/Qwen3-0.6B
Quantized
(6)
this model