Stack-2-9-finetuned / stack /training /train_config.yaml
walidsobhie-code
refactor: Squeeze folders further - cleaner structure
65888d5
# Stack 2.9 Training Configuration
# Model Configuration
model:
name: "Qwen/Qwen2.5-Coder-32B"
trust_remote_code: true
torch_dtype: "bfloat16"
# Data Configuration - supports multiple training files
data:
train_files:
- "/Users/walidsobhi/.openclaw/workspace/stack-2.9/training-data/final/train.jsonl"
val_file: "/Users/walidsobhi/.openclaw/workspace/stack-2.9/training-data/final/val.jsonl"
test_file: "/Users/walidsobhi/.openclaw/workspace/stack-2.9/training-data/final/test.jsonl"
train_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/data/train"
eval_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/data/eval"
max_length: 4096 # Reduced for practical training on MPS/consumer GPUs
train_split: 0.9
test_split: 0.1
# LoRA Configuration
lora:
r: 64
alpha: 128
dropout: 0.05
target_modules:
- "q_proj"
- "k_proj"
- "v_proj"
- "o_proj"
- "gate_proj"
- "up_proj"
- "down_proj"
bias: "none"
task_type: "CAUSAL_LM"
# Training Configuration
training:
num_epochs: 3
batch_size: 1
gradient_accumulation: 16
learning_rate: 1.0e-4
warmup_steps: 100
weight_decay: 0.01
max_grad_norm: 1.0
logging_steps: 10
eval_steps: 500
save_steps: 1000
save_total_limit: 3
fp16: false
bf16: true
gradient_checkpointing: true
optim: "adamw_torch"
# Output Configuration
output:
lora_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/output/stack-2.9-32b-lora"
merged_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/output/stack-2.9-32b-merged"
awq_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/output/stack-2.9-32b-awq"
# Quantization Configuration
quantization:
enabled: true
bits: 4
group_size: 128
# Logging Configuration
logging:
report_to: "none" # Set to "wandb" to enable Weights & Biases tracking
wandb_project: "stack-2.9-training"
run_name: null
# Hardware Configuration
hardware:
device: "cuda" # Change to "mps" for Mac, "cpu" for CPU
num_gpus: 1
use_4bit: true # Enable 4-bit for 32B model on limited VRAM
use_8bit: false
# Merge Configuration (for after training completes)
merge:
enabled: true
output_dir: "/Users/walidsobhi/.openclaw/workspace/stack-2.9-training/output/stack-2.9-32b-merged"