Update README.md
Browse files
README.md
CHANGED
|
@@ -17,7 +17,7 @@ language:
|
|
| 17 |
library_name: transformers
|
| 18 |
pipeline_tag: text-generation
|
| 19 |
---
|
| 20 |
-
# Alpie
|
| 21 |
|
| 22 |
📄 **[Technical Report: Alpie Core.pdf](./Alpie_Core.pdf)**
|
| 23 |
|
|
@@ -32,7 +32,7 @@ pipeline_tag: text-generation
|
|
| 32 |
|
| 33 |
**Alpie Core is one of the first fine-tuned 4-bit reasoning models from India, and among one of the first worldwide.** Trained on just 8 Hopper GPUs with LoRA, QLoRA quantization, and synthetic STEM-rich dataset distillation, it proves that aggressive quantization can not only match but also surpass full-precision baselines.
|
| 34 |
|
| 35 |
-
With a dramatically reduced memory footprint, Alpie
|
| 36 |
|
| 37 |

|
| 38 |
|
|
@@ -50,7 +50,7 @@ With a dramatically reduced memory footprint, Alpie-Core delivers competitive, f
|
|
| 50 |
|
| 51 |
## 3. Approach
|
| 52 |
|
| 53 |
-
**Alpie
|
| 54 |
|
| 55 |
1.**User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
|
| 56 |
|
|
@@ -64,7 +64,7 @@ With a dramatically reduced memory footprint, Alpie-Core delivers competitive, f
|
|
| 64 |
|
| 65 |
6.**Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
|
| 66 |
|
| 67 |
-
This SFT approach enables Alpie
|
| 68 |
|
| 69 |
## 4. Model Features
|
| 70 |
|
|
@@ -98,7 +98,7 @@ This SFT approach enables Alpie-Core to deliver reliable, aligned, and context-a
|
|
| 98 |

|
| 99 |
|
| 100 |
|
| 101 |
-
| Benchmark | Alpie
|
| 102 |
|-----------|----------------------|-------------------|-------------|---------------|---------------|----------------|----------------------------|
|
| 103 |
| MMLU (5-shot) | **81.28%** | 78.4% | 85.0% | 84.4% | 79.3% | 78.6% | 80.73% |
|
| 104 |
| GSM8K (8-shot) | **92.75%** | 81.6% | 88.3% | 83.5% | - | 82.2% | 80.73% |
|
|
@@ -107,7 +107,7 @@ This SFT approach enables Alpie-Core to deliver reliable, aligned, and context-a
|
|
| 107 |
| MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
|
| 108 |
| HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | = |
|
| 109 |
|
| 110 |
-
These results demonstrate Alpie
|
| 111 |
|
| 112 |
### SWE-Bench Verified Performance
|
| 113 |
|
|
@@ -141,7 +141,7 @@ These results demonstrate Alpie-Core’s ability to rival or surpass leading pro
|
|
| 141 |
|
| 142 |
### Additional Benchmarks
|
| 143 |
|
| 144 |
-
| Benchmark | Alpie
|
| 145 |
|-----------|----------------------|----------|
|
| 146 |
| AIME | **47.34%** | Advanced Mathematics |
|
| 147 |
| GPQA (Diamond) | **40.91%** | Graduate-level QA |
|
|
@@ -173,7 +173,7 @@ These results demonstrate Alpie-Core’s ability to rival or surpass leading pro
|
|
| 173 |
|
| 174 |

|
| 175 |
|
| 176 |
-
**Carbon Footprint**: We estimated the environmental impact of training Alpie
|
| 177 |
CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
|
| 178 |
|
| 179 |
Training Parameters:
|
|
@@ -190,7 +190,7 @@ Conservative mode (near TDP ≈ 700 W per GPU = 0.70 kWh/hr): 0.364 × 408 × 0.
|
|
| 190 |
|
| 191 |
Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
|
| 192 |
|
| 193 |
-
*This makes Alpie
|
| 194 |
|
| 195 |
## 9. Use Cases
|
| 196 |
|
|
@@ -210,7 +210,7 @@ Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian
|
|
| 210 |
## 10. Safety and Limitations
|
| 211 |
|
| 212 |
### Enhanced Content Access
|
| 213 |
-
Unlike the base DeepSeek model, Alpie
|
| 214 |
|
| 215 |
### Current Limitations
|
| 216 |
- Multilingual reasoning in Hindi/Hinglish shows room for improvement
|
|
@@ -317,8 +317,8 @@ with torch.no_grad():
|
|
| 317 |
## 12. Citation
|
| 318 |
|
| 319 |
```bibtex
|
| 320 |
-
@misc{
|
| 321 |
-
title = {Alpie-Core: A 4-
|
| 322 |
author = {169Pi AI},
|
| 323 |
year = {2025},
|
| 324 |
url = {https://huggingface.co/alpie/Alpie-Core}
|
|
@@ -354,5 +354,5 @@ We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bi
|
|
| 354 |
For technical inquiries and support: **contact@169pi.com**
|
| 355 |
|
| 356 |
---
|
| 357 |
-
Alpie
|
| 358 |
*For technical details, training methodology, and comprehensive evaluation results, please refer to our technical report.*
|
|
|
|
| 17 |
library_name: transformers
|
| 18 |
pipeline_tag: text-generation
|
| 19 |
---
|
| 20 |
+
# Alpie Core: 4-bit Quantized Reasoning Model
|
| 21 |
|
| 22 |
📄 **[Technical Report: Alpie Core.pdf](./Alpie_Core.pdf)**
|
| 23 |
|
|
|
|
| 32 |
|
| 33 |
**Alpie Core is one of the first fine-tuned 4-bit reasoning models from India, and among one of the first worldwide.** Trained on just 8 Hopper GPUs with LoRA, QLoRA quantization, and synthetic STEM-rich dataset distillation, it proves that aggressive quantization can not only match but also surpass full-precision baselines.
|
| 34 |
|
| 35 |
+
With a dramatically reduced memory footprint, Alpie Core delivers competitive, frontier-level reasoning performance, even beating some top proprietary models. It achieves **81.28% on MMLU, 92.75% on GSM8K, and 57.8% on SWE-Bench Verified**, ranking top globally on competitive leaderboards, a demonstration that efficient models can rival frontier systems while remaining practical for real-world deployment at scale.
|
| 36 |
|
| 37 |

|
| 38 |
|
|
|
|
| 50 |
|
| 51 |
## 3. Approach
|
| 52 |
|
| 53 |
+
**Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
|
| 54 |
|
| 55 |
1.**User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
|
| 56 |
|
|
|
|
| 64 |
|
| 65 |
6.**Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
|
| 66 |
|
| 67 |
+
This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
|
| 68 |
|
| 69 |
## 4. Model Features
|
| 70 |
|
|
|
|
| 98 |

|
| 99 |
|
| 100 |
|
| 101 |
+
| Benchmark | Alpie Core (32B-4bit) | DeepSeek-V2 (236B) | Qwen2.5 72B | Llama 3.1 405B | Llama 3.1 70B | Gemma-3 27B-PT | Mistral-Small-24B-Base-2501 |
|
| 102 |
|-----------|----------------------|-------------------|-------------|---------------|---------------|----------------|----------------------------|
|
| 103 |
| MMLU (5-shot) | **81.28%** | 78.4% | 85.0% | 84.4% | 79.3% | 78.6% | 80.73% |
|
| 104 |
| GSM8K (8-shot) | **92.75%** | 81.6% | 88.3% | 83.5% | - | 82.2% | 80.73% |
|
|
|
|
| 107 |
| MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
|
| 108 |
| HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | = |
|
| 109 |
|
| 110 |
+
These results demonstrate Alpie Core’s ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
|
| 111 |
|
| 112 |
### SWE-Bench Verified Performance
|
| 113 |
|
|
|
|
| 141 |
|
| 142 |
### Additional Benchmarks
|
| 143 |
|
| 144 |
+
| Benchmark | Alpie Core (32B-4bit) | Category |
|
| 145 |
|-----------|----------------------|----------|
|
| 146 |
| AIME | **47.34%** | Advanced Mathematics |
|
| 147 |
| GPQA (Diamond) | **40.91%** | Graduate-level QA |
|
|
|
|
| 173 |
|
| 174 |

|
| 175 |
|
| 176 |
+
**Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
|
| 177 |
CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
|
| 178 |
|
| 179 |
Training Parameters:
|
|
|
|
| 190 |
|
| 191 |
Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
|
| 192 |
|
| 193 |
+
*This makes Alpie Core one of the most carbon-efficient reasoning models released to date.*
|
| 194 |
|
| 195 |
## 9. Use Cases
|
| 196 |
|
|
|
|
| 210 |
## 10. Safety and Limitations
|
| 211 |
|
| 212 |
### Enhanced Content Access
|
| 213 |
+
Unlike the base DeepSeek model, Alpie Core provides factual, balanced responses to geopolitically sensitive questions, offering global accessibility and factual accuracy on topics like Taiwan's status, Arunachal Pradesh sovereignty, and other sensitive geopolitical issues.
|
| 214 |
|
| 215 |
### Current Limitations
|
| 216 |
- Multilingual reasoning in Hindi/Hinglish shows room for improvement
|
|
|
|
| 317 |
## 12. Citation
|
| 318 |
|
| 319 |
```bibtex
|
| 320 |
+
@misc{169pi2025alpiecore,
|
| 321 |
+
title = {Alpie-Core: A 4-Bit Quantized Reasoning Model from India that Outperforms Full-Precision Models},
|
| 322 |
author = {169Pi AI},
|
| 323 |
year = {2025},
|
| 324 |
url = {https://huggingface.co/alpie/Alpie-Core}
|
|
|
|
| 354 |
For technical inquiries and support: **contact@169pi.com**
|
| 355 |
|
| 356 |
---
|
| 357 |
+
Alpie Core represents a milestone for open-source AI from India, one of the first globally to show that 4-bit reasoning models can rival frontier-scale systems. We hope this release empowers developers, researchers, and organisations worldwide to build more efficient, inclusive, and impactful AI.
|
| 358 |
*For technical details, training methodology, and comprehensive evaluation results, please refer to our technical report.*
|