jdosjcd commited on
Commit
a638d8e
·
verified ·
1 Parent(s): 98e85c7

checkpoint step 2500 epoch 1

Browse files
0324_1332_checkpoint_step_2500_epoch_1/config.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1,
3
+ "global_step": 2500,
4
+ "loss": 0.03748499974608421,
5
+ "training_config": {
6
+ "llm_backbone": "dasheng",
7
+ "qwen3_name": "Qwen/Qwen3-Reranker-0.6B",
8
+ "dasheng_name": "mispeech/midashenglm-7b-0804-fp32",
9
+ "dasheng_path": "/workspace/cache/huggingface/dasheng_lm",
10
+ "trainable_modules": [
11
+ "backbone",
12
+ "dasheng",
13
+ "dasheng_down",
14
+ "dasheng_proj",
15
+ "lm_head"
16
+ ],
17
+ "use_lora": true,
18
+ "lora_r": 16,
19
+ "lora_alpha": 32,
20
+ "lora_dropout": 0.05,
21
+ "lora_target_modules": [
22
+ "q_proj",
23
+ "k_proj",
24
+ "v_proj"
25
+ ],
26
+ "use_dasheng_lora": false,
27
+ "dasheng_lora_r": 8,
28
+ "dasheng_lora_alpha": 16,
29
+ "dasheng_lora_dropout": 0.1,
30
+ "dasheng_lora_target_modules": null,
31
+ "train_layer_ratio": 1.0,
32
+ "train_layer_strategy": "last_n",
33
+ "output_dim": null,
34
+ "output_identity": false,
35
+ "use_logit_scale": true,
36
+ "use_checkpointing": true,
37
+ "checkpoint_reentrant": false,
38
+ "s3_base_path": "https://d2j287p0ytux1o.cloudfront.net",
39
+ "dataset_config": "/workspace/SpeechRAG_exp/dataset_configs/setting_multi_task_reranker.json",
40
+ "aws_profile": "test_user",
41
+ "cache_dir": "/workspace/cache/huggingface",
42
+ "enable_audio_cache": true,
43
+ "audio_cache_dir": "/workspace/cache/huggingface/audio_cache",
44
+ "target_sr": 16000,
45
+ "mono": true,
46
+ "max_query_audio_length": 45.0,
47
+ "max_doc_audio_length": 45.0,
48
+ "max_query_text_length": 1000,
49
+ "max_doc_text_length": 1000,
50
+ "eval_max_query_audio_length": null,
51
+ "eval_max_doc_audio_length": null,
52
+ "eval_max_query_text_length": null,
53
+ "eval_max_doc_text_length": null,
54
+ "loss_type": "pointwise",
55
+ "batch_size": 4,
56
+ "num_epochs": 10,
57
+ "learning_rate": 0.0001,
58
+ "gradient_accumulation_steps": 4,
59
+ "weight_decay": 0.001,
60
+ "optimizer_bits": "default",
61
+ "num_workers": 4,
62
+ "train_batch_task_mode": "single_task",
63
+ "task_batch_ratio": "{\"semantic\": 0.5, \"cross\": 0.5}",
64
+ "save_dir": "checkpoints",
65
+ "save_steps": 500,
66
+ "keep_checkpoints": 1,
67
+ "upload_steps": 500,
68
+ "upload_repo_id": "jdosjcd/reranker_checkpoint",
69
+ "log_dir": "logs",
70
+ "log_steps": 5,
71
+ "mixed_precision": "bf16",
72
+ "use_deepspeed": false,
73
+ "deepspeed_config": null,
74
+ "use_fsdp": false,
75
+ "fsdp_config": null,
76
+ "scheduler_type": "warmup_cosine_decay",
77
+ "warmup_steps": 500,
78
+ "warmup_ratio": 0.1,
79
+ "min_lr": 1e-05,
80
+ "eval_steps": 500,
81
+ "eval_batch_size": 8,
82
+ "eval_retrieval_mode": "embedding_topk",
83
+ "eval_embedding_checkpoint_dir": "/workspace/SpeechRAG_exp/checkpoints_512/checkpoint_step_1700_epoch_1",
84
+ "eval_embedding_cache_path": null,
85
+ "eval_embedding_use_ema_model": false,
86
+ "eval_embedding_top_k": 25,
87
+ "eval_embedding_batch_size": null,
88
+ "resume_from": null,
89
+ "add_speaker_mask": false,
90
+ "positive_token": "yes",
91
+ "negative_token": "no",
92
+ "pointwise_batch_mode": "offline",
93
+ "pointwise_num_negatives": 0
94
+ }
95
+ }
0324_1332_checkpoint_step_2500_epoch_1/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe190397fb4ba42c297b8ac65467c41cfbc55877cea06da506a9ff21268b1a75
3
+ size 33154937619
0324_1332_checkpoint_step_2500_epoch_1/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f988a445e39e513515ead4ac03a6732d3ea401d049f5d14193cc9ac7f140fbb
3
+ size 9744503428
0324_1332_checkpoint_step_2500_epoch_1/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:328070cbbb97d33f78a16c9002ec8a914ec4325ea6f8aff9a436544dca414432
3
+ size 1401