Agentic Reasoning for Large Language Models
Paper
⢠2601.12538 ⢠Published
⢠201
Here on Hugging Face (models section) use the Parameters filter and select the size range of the model you want.
There has to be a trade-off here to really shrink the model size.
The model has to be scoped to specific tasks or domains rather than trying to be general-purpose.
And a good approach to achieve this is Knowledge Distillation: train a tiny student model on a specific task to mimic a larger teacher, but you lose generality. That's the trade-off.