File size: 290 Bytes
f550502 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | {
"vocab_size": 2048,
"d_model": 256,
"n_heads": 8,
"n_layers": 6,
"n_shared_layers": 2,
"d_ff": 512,
"max_seq_len": 192,
"dropout": 0.05,
"lora_rank": 16,
"use_thinking_tokens": true,
"n_thinking_steps": 2,
"rope_base": 10000,
"weight_tying": true
} |