| | --- |
| | tags: |
| | - text-generation |
| | - reasoning |
| | - coding |
| | - mathematics |
| | - quantization |
| | - 4-bit model |
| | - state-of-the-art |
| | license: apache-2.0 |
| | datasets: |
| | - synthetic |
| | base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-32B |
| | language: |
| | - en |
| | - hi |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # Alpie Core: 4-bit Quantized Reasoning Model |
| |
|
| | <p align="center"> |
| | <a href="https://169pi.ai/"><img src="https://img.shields.io/badge/🌐%20Website-169Pi%20AI-blue" alt="Website"></a> |
| | <a href="https://huggingface.co/169Pi"><img src="https://img.shields.io/badge/🤗%20Hugging%20Face-169Pi%20AI-yellow" alt="Hugging Face"></a> |
| | <a href="https://pypi.org/project/pi169/0.1/"><img src="https://img.shields.io/badge/PyPI-pi169-blue" alt="PyPI"></a> |
| | <a href="https://www.linkedin.com/company/169pi/"><img src="https://img.shields.io/badge/LinkedIn-169Pi%20AI-blue" alt="LinkedIn"></a> |
| | <a href="https://x.com/169Pi_ai"><img src="https://img.shields.io/badge/X-169Pi%20AI-black" alt="X"></a> |
| | </p> |
| |
|
| | ## TL;DR |
| |
|
| | - **32B reasoning model**, trained & served at **4-bit quantization** |
| | - **Competitive with GPT-4o / Claude 3.5 Sonnet** on reasoning & coding benchmarks |
| | - **65K context length** for long-document reasoning |
| | - **Open source** (Apache 2.0) - fully permissive for commercial use |
| | - Available via **Ollama**, **Hugging Face**, and **hosted API** with 5M free tokens |
| |
|
| | 📄 **[Technical Report: Alpie Core.pdf](./Alpie_Core.pdf)** |
| |
|
| | --- |
| |
|
| | ## How to Use Alpie Core |
| |
|
| | ### Option 1: Local Inference with Ollama (Recommended for Quick Start) |
| |
|
| | ```bash |
| | # Pull the model (20GB) |
| | ollama pull 169pi/alpie-core |
| | |
| | # Run inference |
| | ollama run 169pi/alpie-core |
| | ``` |
| |
|
| | **Requirements**: 20GB RAM/VRAM minimum |
| |
|
| | ### Option 2: Hosted Inference via 169Pi API |
| |
|
| | Get started instantly with our **hosted API** - no setup required! |
| |
|
| | **Get your first free API key** including **5 million tokens** to test real workloads |
| |
|
| | - **OpenAI-compatible** - drop-in replacement for OpenAI SDK |
| | - Supports **streaming**, **async**, and **long-context reasoning** |
| | - Production-ready with low latency |
| |
|
| | **[Get your API key at 169pi.ai](https://169pi.ai/)** |
| |
|
| | ### Option 3: Programmatic Access with Python SDK |
| |
|
| | ```bash |
| | # Install the official SDK |
| | pip install pi169 |
| | |
| | # Set your API key |
| | export ALPIE_API_KEY="your_key_here" |
| | |
| | # Use via CLI |
| | pi169 "Explain quantum entanglement" |
| | |
| | # Or use in Python |
| | from pi169 import AlpieClient |
| | |
| | client = AlpieClient(api_key="your_key_here") |
| | response = client.chat.completions.create( |
| | model="alpie-core", |
| | messages=[{"role": "user", "content": "Solve this coding problem..."}], |
| | stream=True |
| | ) |
| | ``` |
| |
|
| | **SDK Features**: Streaming, async/await, OpenAI compatibility, type-safe interface |
| |
|
| | ### Option 4: Load Directly with Transformers (Advanced) |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | from peft import PeftModel, PeftConfig |
| | import torch |
| | |
| | # Load LoRA adapter configuration |
| | peft_model_id = "169Pi/Alpie-Core" |
| | config = PeftConfig.from_pretrained(peft_model_id) |
| | |
| | # Load base model + LoRA weights |
| | base_model = AutoModelForCausalLM.from_pretrained( |
| | config.base_model_name_or_path, |
| | torch_dtype=torch.float16, |
| | device_map="auto" |
| | ) |
| | tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path) |
| | model = PeftModel.from_pretrained(base_model, peft_model_id) |
| | |
| | # Inference |
| | prompt = "Solve: What is the integral of x^2?" |
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | outputs = model.generate(**inputs, max_new_tokens=1000) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | --- |
| |
|
| | ## Why Alpie Core? |
| |
|
| | **Alpie Core is one of the first fine-tuned 4-bit reasoning models from India, and among the first worldwide at this scale.** Trained on just 8 Hopper GPUs using LoRA and QLoRA 4-bit quantization with synthetic STEM-rich datasets, it proves that aggressive quantization can match and even surpass full-precision baselines. |
| |
|
| | With a dramatically reduced memory footprint, Alpie Core delivers competitive, frontier-level reasoning performance, even beating top proprietary models. It achieves: |
| |
|
| | - **81.28% on MMLU** (5-shot) |
| | - **92.75% on GSM8K** (8-shot) |
| | - **57.8% on SWE-Bench Verified** (ranked #1 globally) |
| |
|
| | This demonstrates that efficient models can rival frontier systems while remaining practical for real-world deployment at scale. |
| |
|
| |  |
| |
|
| | --- |
| |
|
| | ## Model Summary |
| |
|
| | - **Base Architecture**: DeepSeek-R1-Distill-Qwen-32B |
| | - **Parameters**: 32 billion (quantized to 4-bit) |
| | - **Training Method**: Supervised Fine-Tuning (SFT) using LoRA/QLoRA |
| | - **Quantization**: 4-bit NF4 with double quantization |
| | - **Context Length**: 65k tokens |
| | - **Max Output Length**: 16,384 tokens |
| | - **Training Data**: Synthetic (STEM, reasoning, coding) + curated data (law, Indian context, exams, multilingual) |
| | - **License**: Apache 2.0 |
| |
|
| | --- |
| |
|
| | ## Approach |
| |
|
| | **Alpie Core** underwent extensive **supervised fine-tuning (SFT)** to strengthen reasoning, robustness, and safety. The training leveraged a diverse mixture of curated open-source datasets and proprietary synthetic data, optimized with high-quality LLM-generated responses. The fine-tuning process emphasized: |
| |
|
| | 1. **User Understanding and Clarity** – ensuring outputs are direct, interpretable, and pedagogically sound |
| | 2. **Security and Ethical Guidelines** – filtering unsafe or harmful generations |
| | 3. **Limitations and Knowledge Boundaries** – transparently communicating uncertainty |
| | 4. **Handling Complex and Sensitive Topics** – balancing informativeness with responsible guardrails |
| | 5. **Safety and Respectful Engagement** – maintaining politeness, inclusivity, and cultural sensitivity |
| | 6. **Confidentiality and Responsible Use** – preventing leakage of private data or internal reasoning traces |
| |
|
| | This approach enables Alpie Core to deliver reliable, aligned, and context-aware responses while maintaining safety across a broad range of use cases, generalizing across global and Indian contexts. |
| |
|
| | --- |
| |
|
| | ## Model Features |
| |
|
| | 1. **Supports Streaming** – Real-time token-level responses |
| | 2. **OpenAI-Compatible API** – Seamless integration with OpenAI client libraries |
| | 3. **65K Context Length** – Handles very large inputs and conversations |
| | 4. **16,384 Max Output Length** – Enables extremely long generations |
| | 5. **4-Bit Quantization** – Memory-efficient and optimized for deployment |
| | 6. **High Throughput Inference** – Powered by vLLM for efficient large-scale serving |
| | 7. **Low Latency Inference** – Fast response times optimized for production |
| | 8. **Customizable Safety & Moderation** – Built-in guardrails for safer outputs |
| | 9. **Supports Function Calling / Tool Use** – Structured outputs and external API integration |
| | 10. **Instruction Following** – Optimized for reasoning and chain-of-thought answers |
| | 11. **Education & Research Ready** – Tailored for competitive exams, STEM reasoning, and knowledge tasks |
| |
|
| | --- |
| |
|
| | ## Key Highlights |
| |
|
| | 1. **First 4-bit Reasoning Model from India**: Competitive globally with frontier models |
| | 2. **Benchmark Competitiveness**: Outperforms or matches 70B+ models across reasoning, math, and coding |
| | 3. **STEM & Coding Strength**: Excellent on GSM8K, MATH-500, HumanEval, SWE-Bench Verified |
| | 4. **Efficiency & Deployment**: 16 GB VRAM footprint, runs on commodity GPUs |
| | 5. **Extended Context Length**: 65K tokens for research papers, multi-document reasoning |
| | 6. **Environmental Benefits**: ~298–835 kg CO₂e, 2–3× more efficient than FP16 training |
| | 7. **Open-Source Commitment**: Released under Apache 2.0 for global use |
| |
|
| | --- |
| |
|
| | ## Benchmark Results |
| |
|
| |  |
| |
|
| | ### Core Benchmarks |
| |
|
| | | Benchmark | Alpie Core (32B-4bit) | DeepSeek-V2 (236B) | Qwen2.5 72B | Llama 3.1 405B | Llama 3.1 70B | Gemma-3 27B-PT | Mistral-Small-24B | |
| | |-----------|----------------------|-------------------|-------------|---------------|---------------|----------------|-------------------| |
| | | MMLU (5-shot) | **81.28%** | 78.4% | 85.0% | 84.4% | 79.3% | 78.6% | 80.73% | |
| | | GSM8K (8-shot) | **92.75%** | 81.6% | 88.3% | 83.5% | - | 82.2% | 80.73% | |
| | | BBH (3-shot) | **85.12%** | 78.8% | 79.8% | 82.9% | 81.6% | 77.7% | - | |
| | | MMLU-Pro (5-shot) | **64.78%** | 51.4% | 58.3% | 52.8% | 53.8% | 52.2% | 54.37% | |
| | | MBPP (pass@1) | **75.20%** | 65.0% | 72.6% | 68.4% | - | 65.6% | 69.64% | |
| | | HumanEval (pass@1) | **57.23%** | 43.3% | 53.0% | 54.9% | - | 48.8% | - | |
| |
|
| | ### SWE-Bench Verified Performance |
| |
|
| | | Rank | Model | Accuracy (%) | vs Alpie | |
| | |------|-------|-------------|----------| |
| | | **1** | **Alpie Core** | **57.8** | **—** | |
| | | 2 | Qwen3-Coder-30B-A3B-Instruct | 51.6 | -6.2% | |
| | | 3 | o1 | 48.9 | -8.9% | |
| | | 4 | o3-mini (high) | 49.3 | -8.5% | |
| | | 5 | Claude 3.5 Sonnet | 49.0 | -8.8% | |
| | | 6 | DeepSeek R1 | 49.2 | -8.6% | |
| | | 7 | Devstral | 46.8 | -11.0% | |
| |
|
| | ### Humanity's Last Exam Leaderboard |
| |
|
| | | Rank | Model | Accuracy (%) | vs Alpie | |
| | |------|-------|-------------|----------| |
| | | 1 | GPT 4.5 Preview | 5.8 | +0.39% | |
| | | 2 | Claude Sonnet 4 | 5.42 | +0.01% | |
| | | **3** | **Alpie Core 32B (4-bit)** | **5.41** | **—** | |
| | | 4 | Llama 4 Maverik | 5.34 | -0.07% | |
| | | 5 | GPT 4.1 | 4.97 | -0.44% | |
| | | 6 | Kimi K2 Instruct | 4.68 | -0.73% | |
| | | 7 | DeepSeek V3 | 4.55 | -0.86% | |
| |
|
| |  |
| |
|
| | ### Additional Benchmarks |
| |
|
| | | Benchmark | Alpie Core | Category | |
| | |-----------|-----------|----------| |
| | | AIME | **47.34%** | Advanced Mathematics | |
| | | GPQA (Diamond) | **40.91%** | Graduate-level QA | |
| | | TruthfulQA (MC2) | **60.05%** | Truthfulness | |
| | | HellaSwag | **84.66%** | Commonsense | |
| | | PIQA | **83.24%** | Physical Reasoning | |
| | | ARC Challenge | **67.58%** | Science QA | |
| | | CommonSenseQA | **87.06%** | Commonsense | |
| | | AGIEval | **64.98%** | General Intelligence | |
| | | Winogrande | **79.53%** | Commonsense Reasoning | |
| | | MATH-500 | **70.00%** | Advanced Mathematics | |
| |
|
| |  |
| |
|
| | --- |
| |
|
| | ## Training Details |
| |
|
| | - **Hardware**: 8× NVIDIA H100-80GB GPUs |
| | - **Fine-tuning Method**: LoRA/QLoRA |
| | - LoRA Alpha: 16 |
| | - LoRA Dropout: 0.05 |
| | - LoRA Rank: 16 |
| | - **Quantization**: 4-bit NF4 + Double Quantization + FP16 compute |
| | - **Dataset Domains**: Mathematics, coding, reasoning, science, competitive exams, Indian context + law, multilingual (Hindi/Hinglish) |
| | - **Synthetic Data Advantage**: +15-20% performance boost in STEM & coding |
| | - **Training Strategy**: Multi-stage distillation → SFT → safety alignment |
| | - **Total Training Time**: 408 hours |
| |
|
| | --- |
| |
|
| | ## Environmental Impact |
| |
|
| |  |
| |
|
| | We estimated the carbon footprint of training Alpie Core on 8× NVIDIA H100-80GB GPUs: |
| |
|
| | **Formula**: CO₂e (kg) = Grid CO₂ Factor × Runtime × Power per GPU × Number of GPUs |
| |
|
| | **Training Parameters**: |
| | - Grid CO₂ Factor (Azure): 0.364 kg CO₂e/kWh |
| | - Runtime: 408 hours |
| | - GPUs: 8× H100-80GB |
| |
|
| | **Results**: |
| | - **Realistic mode** (250W avg per GPU): **~298 kg CO₂e** |
| | - **Conservative mode** (700W TDP per GPU): **~835 kg CO₂e** |
| |
|
| | *This makes Alpie Core one of the most carbon-efficient reasoning models released to date.* |
| |
|
| | --- |
| |
|
| | ## Use Cases |
| |
|
| | Best for **STEM**, **complex mathematical reasoning**, **coding**, and **Indian context** |
| |
|
| | 1. **STEM Education**: Advanced problem-solving in science, technology, engineering, mathematics |
| | 2. **Mathematical Reasoning**: Multi-step logical and quantitative reasoning |
| | 3. **Software Development**: Code generation, debugging, algorithmic problem-solving |
| | 4. **Indian Context**: Competitive exam assistance (JEE, NEET, UPSC), Hindi/Hinglish support |
| | 5. **Research & Legal**: 65K context for academic papers, legal documents, long-form analysis |
| |
|
| | --- |
| |
|
| | ## Safety and Limitations |
| |
|
| | ### Enhanced Content Access |
| |
|
| | Unlike the base DeepSeek model, Alpie Core provides factual, balanced responses to geopolitically sensitive questions, offering global accessibility on topics like Taiwan's status, Arunachal Pradesh sovereignty, and other sensitive issues. |
| |
|
| | ### Current Limitations |
| |
|
| | - Multilingual reasoning in Hindi/Hinglish shows room for improvement |
| | - Fixed knowledge cutoff without real-time information retrieval |
| | - Occasional struggles with complex multi-hop mathematical reasoning |
| | - Potential hallucinations in factual question-answering |
| | - Should not be used for medical/legal advice without expert oversight |
| |
|
| | ### Mitigations |
| |
|
| | - Safety classifiers and output filtering systems |
| | - Model-assisted safety pipeline using RLHF |
| | - Comprehensive adversarial testing by domain experts |
| |
|
| | --- |
| |
|
| | ## Python SDK Quick Start |
| |
|
| | ```bash |
| | # Install |
| | pip install pi169 |
| | |
| | # Set API key |
| | export ALPIE_API_KEY="your_key_here" |
| | |
| | # CLI usage |
| | pi169 "Explain 4-bit quantization" |
| | ``` |
| |
|
| | ### SDK Features |
| |
|
| | - **CLI Integration** for quick interactions |
| | - **Streaming & Non-Streaming** completions |
| | - **Async/Await Support** for concurrent requests |
| | - **Type-safe Interface** with dataclasses |
| | - **Robust Error Handling** |
| | - **OpenAI-Compatible**: Drop-in replacement |
| |
|
| | [Full SDK documentation on PyPI](https://pypi.org/project/pi169/0.1/) |
| |
|
| | --- |
| |
|
| | ## Advanced Usage Examples |
| |
|
| | ### Streaming Inference with Transformers |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer |
| | from peft import PeftModel, PeftConfig |
| | import torch |
| | |
| | peft_model_id = "169Pi/Alpie-Core" |
| | config = PeftConfig.from_pretrained(peft_model_id) |
| | |
| | base_model = AutoModelForCausalLM.from_pretrained( |
| | config.base_model_name_or_path, |
| | torch_dtype=torch.float16, |
| | device_map="auto" |
| | ) |
| | |
| | tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path) |
| | model = PeftModel.from_pretrained(base_model, peft_model_id) |
| | model.eval() |
| | |
| | streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) |
| | |
| | prompt = "Explain the P vs NP problem" |
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | |
| | print("Streaming Response:") |
| | with torch.no_grad(): |
| | outputs = model.generate( |
| | **inputs, |
| | max_new_tokens=1000, |
| | streamer=streamer, |
| | do_sample=True, |
| | temperature=0.7, |
| | top_p=0.9 |
| | ) |
| | ``` |
| |
|
| | ### Deployment Options |
| |
|
| | - **Transformers**: Python, PyTorch integration |
| | - **vLLM**: High-throughput inference server |
| | - **Ollama**: Easy local deployment (20GB model size) |
| | - **169Pi API**: Production-ready hosted inference |
| |
|
| | --- |
| |
|
| | ## Citation |
| |
|
| | ```bibtex |
| | @misc{169pi2025alpiecore, |
| | title = {Alpie-Core: A 4-Bit Quantized Reasoning Model from India that Outperforms Full-Precision Models}, |
| | author = {169Pi AI}, |
| | year = {2025}, |
| | url = {https://huggingface.co/169Pi/Alpie-Core} |
| | } |
| | ``` |
| |
|
| | --- |
| |
|
| | ## Community & Contributions |
| |
|
| | Released under Apache 2.0 - we welcome the community to build, extend, and improve! |
| |
|
| | 1. **Issues & Discussions**: Report bugs or suggest features on Hugging Face |
| | 2. **Contributions**: Pull requests welcome for improvements |
| | 3. **Share Results**: Post your fine-tuning experiments and benchmarks |
| | 4. **Collaborate**: Join us in shaping the future of efficient AI |
| |
|
| | --- |
| |
|
| | ## License |
| |
|
| | **Apache 2.0 License** – Permissive for research and commercial use |
| |
|
| | --- |
| |
|
| | ## Acknowledgements |
| |
|
| | Thanks to **DeepSeek** for the original model foundation. We also acknowledge: |
| |
|
| | - **Hugging Face** ecosystem (Transformers, PEFT, vLLM, bitsandbytes) |
| | - Open-source datasets (MMLU, GSM8K, SWE-Bench, etc.) |
| | - Cloud infrastructure providers |
| | - The broader AI research community |
| |
|
| | --- |
| |
|
| | ## Contact |
| |
|
| | **Technical Support**: support@169pi.com |
| |
|
| | --- |
| |
|
| | *Alpie Core represents a milestone for open-source AI from India, demonstrating that 4-bit reasoning models can rival frontier-scale systems. We hope this release empowers developers, researchers, and organizations worldwide to build more efficient, inclusive, and impactful AI.* |
| |
|
| | **Get started today with 5 million free tokens at [169pi.ai](https://169pi.ai/)** |