best val_rmse 0.2080
Browse files- config.json +5 -2
- pytorch_model.bin +1 -1
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"best_val_loss":
|
| 3 |
"comet_ai_key": null,
|
| 4 |
"context_observations": {
|
| 5 |
"add_rem": true,
|
|
@@ -18,7 +18,7 @@
|
|
| 18 |
"route": "oral",
|
| 19 |
"time": 0.0
|
| 20 |
},
|
| 21 |
-
"experiment_dir": "/home/cesarali/Pharma/generative_pk/results/comet/node_pk_compartments/
|
| 22 |
"experiment_indentifier": null,
|
| 23 |
"experiment_name": "node_pk_compartments",
|
| 24 |
"hf_model_card_path": [
|
|
@@ -156,6 +156,7 @@
|
|
| 156 |
],
|
| 157 |
"test_protocol": "simulated",
|
| 158 |
"test_size": 5,
|
|
|
|
| 159 |
"train_size": 10,
|
| 160 |
"val_protocol": "simulated",
|
| 161 |
"val_size": 5,
|
|
@@ -168,6 +169,7 @@
|
|
| 168 |
"activation": "ReLU",
|
| 169 |
"aggregator_num_heads": 2,
|
| 170 |
"aggregator_type": "attention",
|
|
|
|
| 171 |
"cov_proj_dim": 16,
|
| 172 |
"decoder_attention_layers": 2,
|
| 173 |
"decoder_hidden_dim": 32,
|
|
@@ -192,6 +194,7 @@
|
|
| 192 |
"time_obs_encoder_hidden_dim": 32,
|
| 193 |
"time_obs_encoder_output_dim": 32,
|
| 194 |
"use_attention": true,
|
|
|
|
| 195 |
"use_kl_i": true,
|
| 196 |
"use_kl_init": true,
|
| 197 |
"use_kl_s": true,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_val_loss": 0.20801889896392822,
|
| 3 |
"comet_ai_key": null,
|
| 4 |
"context_observations": {
|
| 5 |
"add_rem": true,
|
|
|
|
| 18 |
"route": "oral",
|
| 19 |
"time": 0.0
|
| 20 |
},
|
| 21 |
+
"experiment_dir": "/home/cesarali/Pharma/generative_pk/results/comet/node_pk_compartments/9ab78fb7a8f0417fba9b79d2c1199e33",
|
| 22 |
"experiment_indentifier": null,
|
| 23 |
"experiment_name": "node_pk_compartments",
|
| 24 |
"hf_model_card_path": [
|
|
|
|
| 156 |
],
|
| 157 |
"test_protocol": "simulated",
|
| 158 |
"test_size": 5,
|
| 159 |
+
"tqdm_progress": false,
|
| 160 |
"train_size": 10,
|
| 161 |
"val_protocol": "simulated",
|
| 162 |
"val_size": 5,
|
|
|
|
| 169 |
"activation": "ReLU",
|
| 170 |
"aggregator_num_heads": 2,
|
| 171 |
"aggregator_type": "attention",
|
| 172 |
+
"combine_latent_mode": "mlp",
|
| 173 |
"cov_proj_dim": 16,
|
| 174 |
"decoder_attention_layers": 2,
|
| 175 |
"decoder_hidden_dim": 32,
|
|
|
|
| 194 |
"time_obs_encoder_hidden_dim": 32,
|
| 195 |
"time_obs_encoder_output_dim": 32,
|
| 196 |
"use_attention": true,
|
| 197 |
+
"use_invariance_loss": false,
|
| 198 |
"use_kl_i": true,
|
| 199 |
"use_kl_init": true,
|
| 200 |
"use_kl_s": true,
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 168701
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0f446c43ac3e7e2a1b15f9fca390f832c1819a7b0cc0369f0bcb999d1c8f86b
|
| 3 |
size 168701
|