checkpoint step 1600 epoch 1
Browse files
0323_0747_checkpoint_step_1600_epoch_1/config.json
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 1,
|
| 3 |
+
"global_step": 1600,
|
| 4 |
+
"loss": 0.13015475869178772,
|
| 5 |
+
"training_config": {
|
| 6 |
+
"llm_backbone": "dasheng",
|
| 7 |
+
"qwen3_name": "Qwen/Qwen3-Embedding-0.6B",
|
| 8 |
+
"dasheng_name": "mispeech/midashenglm-7b-0804-fp32",
|
| 9 |
+
"dasheng_path": "/workspace/cache/huggingface/dasheng_lm",
|
| 10 |
+
"trainable_modules": [
|
| 11 |
+
"backbone",
|
| 12 |
+
"dasheng",
|
| 13 |
+
"dasheng_down",
|
| 14 |
+
"dasheng_proj",
|
| 15 |
+
"siglip_head"
|
| 16 |
+
],
|
| 17 |
+
"use_lora": true,
|
| 18 |
+
"lora_r": 16,
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_dropout": 0.05,
|
| 21 |
+
"lora_target_modules": [
|
| 22 |
+
"q_proj",
|
| 23 |
+
"k_proj",
|
| 24 |
+
"v_proj"
|
| 25 |
+
],
|
| 26 |
+
"use_dasheng_lora": false,
|
| 27 |
+
"dasheng_lora_r": 8,
|
| 28 |
+
"dasheng_lora_alpha": 16,
|
| 29 |
+
"dasheng_lora_dropout": 0.1,
|
| 30 |
+
"dasheng_lora_target_modules": null,
|
| 31 |
+
"train_layer_ratio": 1.0,
|
| 32 |
+
"train_layer_strategy": "last_n",
|
| 33 |
+
"output_dim": null,
|
| 34 |
+
"output_identity": false,
|
| 35 |
+
"use_logit_scale": true,
|
| 36 |
+
"loss_type": "infonce",
|
| 37 |
+
"use_checkpointing": true,
|
| 38 |
+
"checkpoint_reentrant": false,
|
| 39 |
+
"gather_negatives": true,
|
| 40 |
+
"use_loss_mask": true,
|
| 41 |
+
"duplicate_doc_threshold": 0.999,
|
| 42 |
+
"duplicate_query_threshold": 0.999,
|
| 43 |
+
"hard_negative_margin": 0.1,
|
| 44 |
+
"add_speaker_mask": false,
|
| 45 |
+
"s3_base_path": "https://d2j287p0ytux1o.cloudfront.net",
|
| 46 |
+
"dataset_config": "/workspace/SpeechRAG_exp/dataset_configs/setting_multi_task.json",
|
| 47 |
+
"aws_profile": "test_user",
|
| 48 |
+
"cache_dir": "/workspace/cache/huggingface",
|
| 49 |
+
"enable_audio_cache": true,
|
| 50 |
+
"audio_cache_dir": "/workspace/cache/huggingface/audio_cache",
|
| 51 |
+
"target_sr": 16000,
|
| 52 |
+
"mono": true,
|
| 53 |
+
"max_query_audio_length": 45.0,
|
| 54 |
+
"max_doc_audio_length": 45.0,
|
| 55 |
+
"max_query_text_length": 1000,
|
| 56 |
+
"max_doc_text_length": 1000,
|
| 57 |
+
"eval_max_query_audio_length": null,
|
| 58 |
+
"eval_max_doc_audio_length": null,
|
| 59 |
+
"eval_max_query_text_length": null,
|
| 60 |
+
"eval_max_doc_text_length": null,
|
| 61 |
+
"batch_size": 64,
|
| 62 |
+
"num_epochs": 10,
|
| 63 |
+
"learning_rate": 0.0001,
|
| 64 |
+
"gradient_accumulation_steps": 2,
|
| 65 |
+
"use_grad_cache": true,
|
| 66 |
+
"gc_query_chunk_size": 8,
|
| 67 |
+
"gc_doc_chunk_size": 8,
|
| 68 |
+
"gc_no_sync_except_last": true,
|
| 69 |
+
"ddp_find_unused_parameters": false,
|
| 70 |
+
"weight_decay": 0.001,
|
| 71 |
+
"optimizer_bits": "default",
|
| 72 |
+
"num_workers": 16,
|
| 73 |
+
"train_batch_task_mode": "single_task",
|
| 74 |
+
"task_batch_ratio": "{\"semantic\": 0.5, \"cross\": 0.5}",
|
| 75 |
+
"save_dir": "checkpoints",
|
| 76 |
+
"save_steps": 100,
|
| 77 |
+
"keep_checkpoints": 1,
|
| 78 |
+
"upload_steps": 100,
|
| 79 |
+
"upload_repo_id": "jdosjcd/embedding_checkpoint",
|
| 80 |
+
"log_dir": "logs",
|
| 81 |
+
"log_steps": 5,
|
| 82 |
+
"mixed_precision": "bf16",
|
| 83 |
+
"use_deepspeed": false,
|
| 84 |
+
"deepspeed_config": null,
|
| 85 |
+
"use_fsdp": false,
|
| 86 |
+
"fsdp_config": null,
|
| 87 |
+
"use_ema": false,
|
| 88 |
+
"ema_decay": 0.9999,
|
| 89 |
+
"ema_update_after": 0,
|
| 90 |
+
"ema_update_every": 1,
|
| 91 |
+
"scheduler_type": "warmup_cosine_decay",
|
| 92 |
+
"warmup_steps": 500,
|
| 93 |
+
"warmup_ratio": 0.1,
|
| 94 |
+
"min_lr": 1e-05,
|
| 95 |
+
"eval_steps": 100,
|
| 96 |
+
"eval_batch_size": 32,
|
| 97 |
+
"resume_from": "/workspace/SpeechRAG_exp/checkpoints/checkpoint_step_1500_epoch_1",
|
| 98 |
+
"test_mode": false,
|
| 99 |
+
"mock_dataset_size": 100
|
| 100 |
+
}
|
| 101 |
+
}
|
0323_0747_checkpoint_step_1600_epoch_1/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b45e438e621a02bc3880d7f36e7211c6d8a3cc8d9eb9da32738513237139c23
|
| 3 |
+
size 31026342367
|
0323_0747_checkpoint_step_1600_epoch_1/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:51e9f0da27d770a1773e3dae1ebc282a066867f1bc6a1df521669036233c11c0
|
| 3 |
+
size 5487310547
|
0323_0747_checkpoint_step_1600_epoch_1/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a791d89cf477ec8e5100abeb2621ba1aa008c24cc8ecc11550a5a22ae90a527f
|
| 3 |
+
size 1401
|