MiniAxion1-0.9M / config.json
AxionLab-official's picture
Upload 3 files
f550502 verified
raw
history blame contribute delete
290 Bytes
{
"vocab_size": 2048,
"d_model": 256,
"n_heads": 8,
"n_layers": 6,
"n_shared_layers": 2,
"d_ff": 512,
"max_seq_len": 192,
"dropout": 0.05,
"lora_rank": 16,
"use_thinking_tokens": true,
"n_thinking_steps": 2,
"rope_base": 10000,
"weight_tying": true
}