Stack-2-9-finetuned / stack /deploy /FREE_DEPLOYMENT.md
walidsobhie-code
feat: Add free deployment support for Stack 2.9
239da7a

Free Deployment Guide for Stack 2.9

This guide covers deploying Stack 2.9 on free-tier platforms.


Option 1: HuggingFace Spaces (Free Inference)

Step 1: Create Space

# Go to https://huggingface.co/spaces and create new Space
# Choose: Docker, Python 3.11, Small (2CPU 4GB)

Step 2: Push Your Model

# Upload your fine-tuned model to HF
from huggingface_hub import HfApi
api = HfApi()
api.upload_folder(
    folder_path="./stack-2.9-7b",
    repo_id="yourusername/stack-2.9",
    repo_type="model"
)

Step 3: Configure API URL

Set environment variable in Space:

  • API_URL: Your model inference URL
  • HF_TOKEN: Your HF token

Step 4: Deploy

# Clone Space and push files
git clone https://huggingface.co/spaces/yourusername/stack-2.9
cp deploy/hfSpaces/* .
git add . && git push

Option 2: Together AI Fine-tuning (Free Credits)

Free Tier Limits

  • Up to 7B model fine-tuning
  • Limited training minutes (varies by promotion)
  • Requires: Together AI account

Setup

# Get API key from https://together.ai
export TOGETHER_API_KEY="your-key"

# Fine-tune 7B model (free-tier friendly)
python stack/training/together_finetune.py \
    --model 7b \
    --data data/final/train.jsonl \
    --epochs 3

Use Fine-tuned Model

from together import Together

client = Together(api_key="your-key")

response = client.chat.completions.create(
    model="your-finetuned-model",
    messages=[{"role": "user", "content": "Write a function"}]
)

Option 3: Google Colab (Free Training)

Run Training

# Open colab_train_stack29.ipynb in Google Colab
# Select GPU runtime (free tier: T4 15GB)

# For 7B model (runs on free tier):
batch_size = 2  # Reduce for 15GB VRAM
gradient_accumulation = 8

Model Sizes for Free Tier

Model VRAM Needed Free Tier?
1.5B ~4GB βœ… Yes
3B ~8GB βœ… Yes (T4)
7B ~16GB ⚠️ Limited
32B ~64GB ❌ No

Option 4: RunPod / Vast.ai (Cheap, Not Free)

Quick Start

# Deploy on RunPod (~$0.20/hour for A100)
cd stack/deploy
./runpod_deploy.sh --template runpod-template.json

# Deploy on Vast.ai (~$0.15/hour)
./vastai_deploy.sh --template vastai-template.json

Recommended Free Stack

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚  Stack 2.9 Free Deployment Stack           β”‚
β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
β”‚  Model:    Qwen2.5-Coder-7B               β”‚
β”‚  Fine-tune: Together AI (free credits)      β”‚
β”‚  Deploy:    HuggingFace Spaces (free)       β”‚
β”‚  UI:        Gradio (included in Spaces)   β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Cost Comparison

Platform Cost What's Free
HF Spaces $0 2CPU 4GB hosting
Together AI varies Fine-tuning credits
Colab $0 ~0.5hr GPU/day
RunPod $0.20/hr First $10 credit
Vast.ai $0.15/hr First $5 credit