iNeil77 commited on
Commit
a208047
·
verified ·
1 Parent(s): 15ea7b9

Upload Qwen3ForCausalLM

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. config.json +4 -4
  3. generation_config.json +2 -2
README.md CHANGED
@@ -2,4 +2,4 @@
2
  license: apache-2.0
3
  pipeline_tag: text-generation
4
  library_name: transformers
5
- ---
 
2
  license: apache-2.0
3
  pipeline_tag: text-generation
4
  library_name: transformers
5
+ ---
config.json CHANGED
@@ -4,7 +4,7 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "dtype": "bfloat16",
9
  "eos_token_id": 151643,
10
  "head_dim": 128,
@@ -50,7 +50,7 @@
50
  "full_attention",
51
  "full_attention"
52
  ],
53
- "max_position_embeddings": 32768,
54
  "max_window_layers": 36,
55
  "model_type": "qwen3",
56
  "num_attention_heads": 32,
@@ -59,12 +59,12 @@
59
  "pad_token_id": 151656,
60
  "rms_norm_eps": 1e-06,
61
  "rope_parameters": {
62
- "rope_theta": 1000000,
63
  "rope_type": "default"
64
  },
65
  "sliding_window": null,
66
  "tie_word_embeddings": false,
67
- "transformers_version": "5.6.1",
68
  "use_cache": true,
69
  "use_sliding_window": false,
70
  "vocab_size": 151936
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
  "dtype": "bfloat16",
9
  "eos_token_id": 151643,
10
  "head_dim": 128,
 
50
  "full_attention",
51
  "full_attention"
52
  ],
53
+ "max_position_embeddings": 131072,
54
  "max_window_layers": 36,
55
  "model_type": "qwen3",
56
  "num_attention_heads": 32,
 
59
  "pad_token_id": 151656,
60
  "rms_norm_eps": 1e-06,
61
  "rope_parameters": {
62
+ "rope_theta": 10000000,
63
  "rope_type": "default"
64
  },
65
  "sliding_window": null,
66
  "tie_word_embeddings": false,
67
+ "transformers_version": "5.8.1",
68
  "use_cache": true,
69
  "use_sliding_window": false,
70
  "vocab_size": 151936
generation_config.json CHANGED
@@ -7,6 +7,6 @@
7
  151656
8
  ],
9
  "pad_token_id": 151656,
10
- "transformers_version": "5.6.1",
11
  "trust_remote_code": true
12
- }
 
7
  151656
8
  ],
9
  "pad_token_id": 151656,
10
+ "transformers_version": "5.8.1",
11
  "trust_remote_code": true
12
+ }