| --- |
| language: |
| - en |
| license: apache-2.0 |
| tags: |
| - math |
| - reasoning |
| - mathematics |
| - causal-lm |
| - text-generation |
| library_name: transformers |
| pipeline_tag: text-generation |
| model_name: Minnow-Math-2B |
| --- |
| |
| # ๐ Minnow-Math-2B |
|
|
| **Minnow-Math-2B** is a 2B-parameter language model by **Kitefish**, focused on mathematical reasoning, symbolic understanding, and structured problem solving. |
|
|
| This is an early release and part of our ongoing effort to build strong, efficient models for reasoning-heavy tasks. |
|
|
| --- |
|
|
| ## โจ What this model is good at |
|
|
| - Basic to intermediate **math problem solving** |
| - **Step-by-step reasoning** for equations and word problems |
| - Understanding **mathematical symbols and structure** |
| - Educational and experimentation use cases |
|
|
| --- |
|
|
| ## ๐ Quick start |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| |
| tokenizer = AutoTokenizer.from_pretrained("kitefish/Minnow-Math-2B") |
| model = AutoModelForCausalLM.from_pretrained( |
| "kitefish/Minnow-Math-2B", |
| torch_dtype="auto", |
| device_map="auto" |
| ) |
| |
| prompt = "Solve: 2x + 5 = 13" |
| inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| |
| outputs = model.generate(**inputs, max_new_tokens=100) |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| |