Chirag2207 commited on
Commit
2b3ee70
·
verified ·
1 Parent(s): be73b15

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -17,7 +17,7 @@ language:
17
  library_name: transformers
18
  pipeline_tag: text-generation
19
  ---
20
- # Alpie-Core: 4-bit Quantized Reasoning Model
21
 
22
  📄 **[Technical Report: Alpie Core.pdf](./Alpie_Core.pdf)**
23
 
@@ -32,7 +32,7 @@ pipeline_tag: text-generation
32
 
33
  **Alpie Core is one of the first fine-tuned 4-bit reasoning models from India, and among one of the first worldwide.** Trained on just 8 Hopper GPUs with LoRA, QLoRA quantization, and synthetic STEM-rich dataset distillation, it proves that aggressive quantization can not only match but also surpass full-precision baselines.
34
 
35
- With a dramatically reduced memory footprint, Alpie-Core delivers competitive, frontier-level reasoning performance, even beating some top proprietary models. It achieves **81.28% on MMLU, 92.75% on GSM8K, and 57.8% on SWE-Bench Verified**, ranking top globally on competitive leaderboards, a demonstration that efficient models can rival frontier systems while remaining practical for real-world deployment at scale.
36
 
37
  ![Combined Benchmark](combined_benchmark.png)
38
 
@@ -50,7 +50,7 @@ With a dramatically reduced memory footprint, Alpie-Core delivers competitive, f
50
 
51
  ## 3. Approach
52
 
53
- **Alpie-Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
54
 
55
  1.**User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
56
 
@@ -64,7 +64,7 @@ With a dramatically reduced memory footprint, Alpie-Core delivers competitive, f
64
 
65
  6.**Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
66
 
67
- This SFT approach enables Alpie-Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie-Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
68
 
69
  ## 4. Model Features
70
 
@@ -98,7 +98,7 @@ This SFT approach enables Alpie-Core to deliver reliable, aligned, and context-a
98
  ![BBH Benchmark](BBH.png)
99
 
100
 
101
- | Benchmark | Alpie-Core (32B-4bit) | DeepSeek-V2 (236B) | Qwen2.5 72B | Llama 3.1 405B | Llama 3.1 70B | Gemma-3 27B-PT | Mistral-Small-24B-Base-2501 |
102
  |-----------|----------------------|-------------------|-------------|---------------|---------------|----------------|----------------------------|
103
  | MMLU (5-shot) | **81.28%** | 78.4% | 85.0% | 84.4% | 79.3% | 78.6% | 80.73% |
104
  | GSM8K (8-shot) | **92.75%** | 81.6% | 88.3% | 83.5% | - | 82.2% | 80.73% |
@@ -107,7 +107,7 @@ This SFT approach enables Alpie-Core to deliver reliable, aligned, and context-a
107
  | MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
108
  | HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | = |
109
 
110
- These results demonstrate Alpie-Core’s ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
111
 
112
  ### SWE-Bench Verified Performance
113
 
@@ -141,7 +141,7 @@ These results demonstrate Alpie-Core’s ability to rival or surpass leading pro
141
 
142
  ### Additional Benchmarks
143
 
144
- | Benchmark | Alpie-Core (32B-4bit) | Category |
145
  |-----------|----------------------|----------|
146
  | AIME | **47.34%** | Advanced Mathematics |
147
  | GPQA (Diamond) | **40.91%** | Graduate-level QA |
@@ -173,7 +173,7 @@ These results demonstrate Alpie-Core’s ability to rival or surpass leading pro
173
 
174
  ![Carbon Footprint](carbon_footprint.png)
175
 
176
- **Carbon Footprint**: We estimated the environmental impact of training Alpie-Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
177
  CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
178
 
179
  Training Parameters:
@@ -190,7 +190,7 @@ Conservative mode (near TDP ≈ 700 W per GPU = 0.70 kWh/hr): 0.364 × 408 × 0.
190
 
191
  Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
192
 
193
- *This makes Alpie-Core one of the most carbon-efficient reasoning models released to date.*
194
 
195
  ## 9. Use Cases
196
 
@@ -210,7 +210,7 @@ Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian
210
  ## 10. Safety and Limitations
211
 
212
  ### Enhanced Content Access
213
- Unlike the base DeepSeek model, Alpie-Core provides factual, balanced responses to geopolitically sensitive questions, offering global accessibility and factual accuracy on topics like Taiwan's status, Arunachal Pradesh sovereignty, and other sensitive geopolitical issues.
214
 
215
  ### Current Limitations
216
  - Multilingual reasoning in Hindi/Hinglish shows room for improvement
@@ -317,8 +317,8 @@ with torch.no_grad():
317
  ## 12. Citation
318
 
319
  ```bibtex
320
- @misc{alpie2025core,
321
- title = {Alpie-Core: A 4-bit Quantized Reasoning Model Surpassing Full-Precision Benchmarks},
322
  author = {169Pi AI},
323
  year = {2025},
324
  url = {https://huggingface.co/alpie/Alpie-Core}
@@ -354,5 +354,5 @@ We are also grateful to the Hugging Face ecosystem (Transformers, PEFT, vLLM, bi
354
  For technical inquiries and support: **contact@169pi.com**
355
 
356
  ---
357
- Alpie-Core represents a milestone for open-source AI from India, one of the first globally to show that 4-bit reasoning models can rival frontier-scale systems. We hope this release empowers developers, researchers, and organisations worldwide to build more efficient, inclusive, and impactful AI.
358
  *For technical details, training methodology, and comprehensive evaluation results, please refer to our technical report.*
 
17
  library_name: transformers
18
  pipeline_tag: text-generation
19
  ---
20
+ # Alpie Core: 4-bit Quantized Reasoning Model
21
 
22
  📄 **[Technical Report: Alpie Core.pdf](./Alpie_Core.pdf)**
23
 
 
32
 
33
  **Alpie Core is one of the first fine-tuned 4-bit reasoning models from India, and among one of the first worldwide.** Trained on just 8 Hopper GPUs with LoRA, QLoRA quantization, and synthetic STEM-rich dataset distillation, it proves that aggressive quantization can not only match but also surpass full-precision baselines.
34
 
35
+ With a dramatically reduced memory footprint, Alpie Core delivers competitive, frontier-level reasoning performance, even beating some top proprietary models. It achieves **81.28% on MMLU, 92.75% on GSM8K, and 57.8% on SWE-Bench Verified**, ranking top globally on competitive leaderboards, a demonstration that efficient models can rival frontier systems while remaining practical for real-world deployment at scale.
36
 
37
  ![Combined Benchmark](combined_benchmark.png)
38
 
 
50
 
51
  ## 3. Approach
52
 
53
+ **Alpie Core** has undergone extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimised with high-quality LLM-generated responses. The fine-tuning process emphasised adherence to rigorous safety and usability standards, including:
54
 
55
  1.**User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound.
56
 
 
64
 
65
  6.**Confidentiality and Responsible Use** – preventing leakage of private training data, proprietary prompts, or internal reasoning traces.
66
 
67
+ This SFT approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases. This approach allows Alpie Core to generalize across global and Indian contexts while staying aligned to safe and responsible use guidelines.
68
 
69
  ## 4. Model Features
70
 
 
98
  ![BBH Benchmark](BBH.png)
99
 
100
 
101
+ | Benchmark | Alpie Core (32B-4bit) | DeepSeek-V2 (236B) | Qwen2.5 72B | Llama 3.1 405B | Llama 3.1 70B | Gemma-3 27B-PT | Mistral-Small-24B-Base-2501 |
102
  |-----------|----------------------|-------------------|-------------|---------------|---------------|----------------|----------------------------|
103
  | MMLU (5-shot) | **81.28%** | 78.4% | 85.0% | 84.4% | 79.3% | 78.6% | 80.73% |
104
  | GSM8K (8-shot) | **92.75%** | 81.6% | 88.3% | 83.5% | - | 82.2% | 80.73% |
 
107
  | MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% |
108
  | HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | = |
109
 
110
+ These results demonstrate Alpie Core’s ability to rival or surpass leading proprietary and open-source models, despite being 4-bit quantized.
111
 
112
  ### SWE-Bench Verified Performance
113
 
 
141
 
142
  ### Additional Benchmarks
143
 
144
+ | Benchmark | Alpie Core (32B-4bit) | Category |
145
  |-----------|----------------------|----------|
146
  | AIME | **47.34%** | Advanced Mathematics |
147
  | GPQA (Diamond) | **40.91%** | Graduate-level QA |
 
173
 
174
  ![Carbon Footprint](carbon_footprint.png)
175
 
176
+ **Carbon Footprint**: We estimated the environmental impact of training Alpie Core (32B) on 8× NVIDIA H100-80GB GPUs by calculating carbon emissions from GPU energy consumption. The calculation follows the formula:
177
  CO₂e (kg) = Grid CO₂ Factor (kg/kWh) × Runtime (hours) × Power per GPU (kW) × Number of GPUs
178
 
179
  Training Parameters:
 
190
 
191
  Total training footprint ranges from ~298 kg CO₂e (realistic) to ~835 kg CO₂e (conservative worst-case)
192
 
193
+ *This makes Alpie Core one of the most carbon-efficient reasoning models released to date.*
194
 
195
  ## 9. Use Cases
196
 
 
210
  ## 10. Safety and Limitations
211
 
212
  ### Enhanced Content Access
213
+ Unlike the base DeepSeek model, Alpie Core provides factual, balanced responses to geopolitically sensitive questions, offering global accessibility and factual accuracy on topics like Taiwan's status, Arunachal Pradesh sovereignty, and other sensitive geopolitical issues.
214
 
215
  ### Current Limitations
216
  - Multilingual reasoning in Hindi/Hinglish shows room for improvement
 
317
  ## 12. Citation
318
 
319
  ```bibtex
320
+ @misc{169pi2025alpiecore,
321
+ title = {Alpie-Core: A 4-Bit Quantized Reasoning Model from India that Outperforms Full-Precision Models},
322
  author = {169Pi AI},
323
  year = {2025},
324
  url = {https://huggingface.co/alpie/Alpie-Core}
 
354
  For technical inquiries and support: **contact@169pi.com**
355
 
356
  ---
357
+ Alpie Core represents a milestone for open-source AI from India, one of the first globally to show that 4-bit reasoning models can rival frontier-scale systems. We hope this release empowers developers, researchers, and organisations worldwide to build more efficient, inclusive, and impactful AI.
358
  *For technical details, training methodology, and comprehensive evaluation results, please refer to our technical report.*