TroyHow commited on
Commit
d89f77c
·
verified ·
1 Parent(s): 8c8bf0d

Upload 3 files

Browse files
.gitattributes CHANGED
@@ -42,3 +42,4 @@ musique/musique_ans_v1.0_train.jsonl filter=lfs diff=lfs merge=lfs -text
42
  musique/musique_full_v1.0_dev.jsonl filter=lfs diff=lfs merge=lfs -text
43
  musique/musique_full_v1.0_test.jsonl filter=lfs diff=lfs merge=lfs -text
44
  musique/musique_full_v1.0_train.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
42
  musique/musique_full_v1.0_dev.jsonl filter=lfs diff=lfs merge=lfs -text
43
  musique/musique_full_v1.0_test.jsonl filter=lfs diff=lfs merge=lfs -text
44
  musique/musique_full_v1.0_train.jsonl filter=lfs diff=lfs merge=lfs -text
45
+ qrag-ft-e5-on-hotpotqa/eval_seed42.jsonl filter=lfs diff=lfs merge=lfs -text
qrag-ft-e5-on-hotpotqa/config.yaml ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ envs:
2
+ max_steps: 2
3
+ task: hotpotqa
4
+ data_path: ./datasets/data_sources/hotpotqa
5
+ sort_by_index: false
6
+ positions_processor: none
7
+ positions_processor_dict:
8
+ none:
9
+ _target_: envs.text_env.AbsolutePositionProcessor
10
+ absolute:
11
+ _target_: envs.text_env.AbsolutePositionProcessor
12
+ random:
13
+ _target_: envs.text_env.RandomPositionProcessor
14
+ max_chunks_count: 2000
15
+ relative:
16
+ _target_: envs.text_env.RelativePositionProcessor
17
+ step_size: 20
18
+ train_dataset:
19
+ _target_: envs.RetrievalHotPotQA
20
+ path: ${envs.data_path}
21
+ split: train
22
+ seed: ${seed}
23
+ test_dataset:
24
+ _target_: envs.RetrievalHotPotQA
25
+ path: ${envs.data_path}
26
+ split: eval
27
+ seed: ${seed}
28
+ env:
29
+ _target_: envs.qa_env.QAEnv
30
+ max_steps: ${envs.max_steps}
31
+ action_embed_length: ${max_action_length}
32
+ separator: ' [SEP] '
33
+ sort_by_index: ${envs.sort_by_index}
34
+ positions_processor: ${envs.positions_processor_dict.${envs.positions_processor}}
35
+ feedback_model:
36
+ _target_: rl.feedback.GroundTruthFeedback
37
+ penalize_extra_steps: false
38
+ dataset:
39
+ _target_: envs.QADatasetAdapter
40
+ dataset:
41
+ _target_: envs.RetrievalHotPotQA
42
+ path: ${envs.data_path}
43
+ split: train
44
+ seed: ${seed}
45
+ test_env:
46
+ _target_: envs.qa_env.QAEnv
47
+ max_steps: ${envs.env.max_steps}
48
+ max_action_length_in_memory: 256
49
+ action_embed_length: ${max_action_length}
50
+ separator: ${envs.env.separator}
51
+ sort_by_index: ${envs.sort_by_index}
52
+ positions_processor: ${envs.env.positions_processor}
53
+ feedback_model:
54
+ _target_: rl.feedback.GroundTruthFeedback
55
+ penalize_extra_steps: false
56
+ dataset:
57
+ _target_: envs.QADatasetAdapter
58
+ dataset:
59
+ _target_: envs.RetrievalHotPotQA
60
+ path: ${envs.data_path}
61
+ split: eval
62
+ seed: ${seed}
63
+ algo:
64
+ model:
65
+ model_name: intfloat/multilingual-e5-large
66
+ revision: main
67
+ use_fast_tokenizer: true
68
+ predictor:
69
+ num_layers: 24
70
+ input_dim: 1024
71
+ hidden_dim: 512
72
+ output_dim: 1
73
+ max_seq_len: 5000
74
+ interpolate_factor: 1
75
+ action_model:
76
+ _target_: rl.bert_predictor.BertPredictor
77
+ bert:
78
+ _target_: transformers.AutoModel.from_pretrained
79
+ pretrained_model_name_or_path: ${algo.model.model_name}
80
+ revision: ${algo.model.revision}
81
+ num_hidden_layers: ${algo.model.predictor.num_layers}
82
+ tokenizer:
83
+ _target_: transformers.AutoTokenizer.from_pretrained
84
+ pretrained_model_name_or_path: ${algo.model.model_name}
85
+ revision: ${algo.model.revision}
86
+ use_fast: ${algo.model.use_fast_tokenizer}
87
+ model_dim: ${algo.model.predictor.input_dim}
88
+ output_size: ${algo.model.predictor.hidden_dim}
89
+ n_output: ${algo.model.predictor.output_dim}
90
+ action_embed_dict:
91
+ absolute:
92
+ _target_: rl.bert_predictor.EmbedderWithAbsoluteEncoding
93
+ model: ${algo.action_model}
94
+ max_seq_len: ${algo.model.predictor.max_seq_len}
95
+ random:
96
+ _target_: rl.bert_predictor.EmbedderWithAbsoluteEncoding
97
+ model: ${algo.action_model}
98
+ max_seq_len: ${algo.model.predictor.max_seq_len}
99
+ interpolate_factor: ${algo.model.predictor.interpolate_factor}
100
+ relative:
101
+ _target_: rl.bert_predictor.EmbedderWithRelativeEncoding
102
+ model: ${algo.action_model}
103
+ max_seq_len: 1000
104
+ none:
105
+ _target_: rl.bert_predictor.EmbedderNone
106
+ model: ${algo.action_model}
107
+ pqn:
108
+ _target_: rl.pqn.PQN
109
+ state_embed:
110
+ _target_: rl.bert_predictor.BertPredictor
111
+ bert:
112
+ _target_: transformers.AutoModel.from_pretrained
113
+ pretrained_model_name_or_path: ${algo.model.model_name}
114
+ revision: ${algo.model.revision}
115
+ num_hidden_layers: ${algo.model.predictor.num_layers}
116
+ tokenizer:
117
+ _target_: transformers.AutoTokenizer.from_pretrained
118
+ pretrained_model_name_or_path: ${algo.model.model_name}
119
+ revision: ${algo.model.revision}
120
+ use_fast: ${algo.model.use_fast_tokenizer}
121
+ model_dim: ${algo.model.predictor.input_dim}
122
+ output_size: ${algo.model.predictor.hidden_dim}
123
+ n_output: ${algo.model.predictor.output_dim}
124
+ action_embed: ${algo.action_embed_dict.${envs.positions_processor}}
125
+ state_embed_target:
126
+ _target_: rl.bert_predictor.BertPredictor
127
+ bert:
128
+ _target_: transformers.AutoModel.from_pretrained
129
+ pretrained_model_name_or_path: ${algo.model.model_name}
130
+ revision: ${algo.model.revision}
131
+ num_hidden_layers: ${algo.model.predictor.num_layers}
132
+ tokenizer:
133
+ _target_: transformers.AutoTokenizer.from_pretrained
134
+ pretrained_model_name_or_path: ${algo.model.model_name}
135
+ revision: ${algo.model.revision}
136
+ use_fast: ${algo.model.use_fast_tokenizer}
137
+ model_dim: ${algo.model.predictor.input_dim}
138
+ output_size: ${algo.model.predictor.hidden_dim}
139
+ n_output: ${algo.model.predictor.output_dim}
140
+ action_embed_target: ${algo.action_embed_dict.${envs.positions_processor}}
141
+ hyperparams:
142
+ gamma: 0.99
143
+ alpha: 0.05
144
+ Lambda: 0.5
145
+ tau: 0.02
146
+ max_grad_norm: 2.0
147
+ accumulate_grads: ${accumulate_grads}
148
+ action_embed_length: ${max_action_length}
149
+ max_action_length_in_memory: 256
150
+ optimizer:
151
+ _target_: torch.optim.AdamW
152
+ lr: 1.5e-05
153
+ betas:
154
+ - 0.9
155
+ - 0.98
156
+ eps: 1.0e-06
157
+ weight_decay: 0.0005
158
+ scheduler:
159
+ _target_: rl.optim.WarmupLinearScheduler
160
+ total: ${steps_count}
161
+ ratio: 0.1
162
+ warmup: 1000
163
+ logger:
164
+ log_dir: runs/Aug24_01-52-22_PQN_hotpotqa
165
+ tensorboard:
166
+ _target_: torch.utils.tensorboard.SummaryWriter
167
+ comment: _PQN_${envs.task}
168
+ log_dir: runs/Aug24_01-52-22_PQN_hotpotqa/tb_logs/
169
+ seed: 42
170
+ device: cuda:0
171
+ learning_start: 200
172
+ steps_count: 10000
173
+ batch_size: 12
174
+ accumulate_grads: 8
175
+ eval_interval: 100
176
+ eval_episodes: 300
177
+ envs_parallel: 1
178
+ max_action_length: 220
qrag-ft-e5-on-hotpotqa/eval_seed42.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bed1db695951c66cac0af83507954ce822b73e64bab50ee1bf9450afff890db
3
+ size 15601203
qrag-ft-e5-on-hotpotqa/model_best.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccd4068d3917bed2e44afcd7ee7a9d66c8487b44dd2206252e6d2985352f92b7
3
+ size 20151355435