| { | |
| "archi_type": "decoder", | |
| "architectures": [ | |
| "Splade" | |
| ], | |
| "attn_implementation": "flash_attention_2", | |
| "attn_type": "causal", | |
| "bidirectional": true, | |
| "lora": false, | |
| "lora_r": 0, | |
| "model_name_or_path": "Qwen/Qwen3-0.6B", | |
| "model_type": "splade", | |
| "n_layers": null, | |
| "padding_side": "left", | |
| "torch_dtype": "bfloat16", | |
| "train_head": false, | |
| "transformers_version": "4.53.3", | |
| "auto_map": { | |
| "AutoConfig": "splade.SpladeConfig", | |
| "AutoModelForCausalLM": "splade.Splade" | |
| } | |
| } | |