Falcon-H1-Tiny-90M-Instruct-math-safe

MATH-optimized | Safe pruning | 25% weights pruned

This model is a conservatively pruned version of tiiuae/Falcon-H1-Tiny-90M-Instruct.

Pruning Alert: The benchmarks show virtually NO quality drop! This isn't a bug -- it is a feature. The Wanda pruning algorithm is so effective at identifying unimportant weights that it can remove a large percentage of parameters without affecting performance. Think of it like pruning dead leaves from a tree -- the tree does not miss them because they were not doing anything anyway!

Performance Comparison

Category Original Pruned Change
Python 0.0% 0.0% β†’
Html 0.0% 0.0% β†’
Trivia 30.0% 30.0% β†’
Math 55.0% 55.0% ⭐ β†’
Reasoning 20.0% 20.0% β†’
Medical 5.0% 5.0% β†’
Linux 15.0% 15.0% β†’
Writing 0.0% 0.0% β†’

Average: 15.6% -> 15.6% (+0.0%)

Math Retention: 100.0%

Comparison Graph

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("CompactAI/Falcon-H1-Tiny-90M-Instruct-math-safe")
tokenizer = AutoTokenizer.from_pretrained("CompactAI/Falcon-H1-Tiny-90M-Instruct-math-safe")

inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Technical Details

Property Value
Base Model tiiuae/Falcon-H1-Tiny-90M-Instruct
Specialization Math
Prune Mode Safe
Weight Reduction 25% weights pruned

License

This model inherits the license from the base model.

Downloads last month
12
Safetensors
Model size
91.1M params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for CompactAI/Falcon-H1-Tiny-90M-Instruct-math-safe

Finetuned
(17)
this model

Collection including CompactAI/Falcon-H1-Tiny-90M-Instruct-math-safe