ChibuUkachi commited on
Commit
d3fa913
·
1 Parent(s): 61c152a

update unknown

Browse files
every_eval_ever_results/aime25.json CHANGED
@@ -6,12 +6,12 @@
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
- "version": "unknown"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
@@ -126,4 +126,4 @@
126
  }
127
  }
128
  ]
129
- }
 
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
+ "version": "v0.13.0"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
 
126
  }
127
  }
128
  ]
129
+ }
every_eval_ever_results/gpqa_diamond.json CHANGED
@@ -6,12 +6,12 @@
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
- "version": "unknown"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
@@ -80,4 +80,4 @@
80
  }
81
  }
82
  ]
83
- }
 
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
+ "version": "v0.13.0"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
 
80
  }
81
  }
82
  ]
83
+ }
every_eval_ever_results/gsm8k_platinum_cot_llama.json CHANGED
@@ -6,7 +6,7 @@
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
@@ -123,4 +123,4 @@
123
  }
124
  }
125
  ]
126
- }
 
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
 
123
  }
124
  }
125
  ]
126
+ }
every_eval_ever_results/ifeval.json CHANGED
@@ -6,7 +6,7 @@
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
@@ -221,4 +221,4 @@
221
  }
222
  }
223
  ]
224
- }
 
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
 
221
  }
222
  }
223
  ]
224
+ }
every_eval_ever_results/math_500.json CHANGED
@@ -6,12 +6,12 @@
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
- "version": "unknown"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
@@ -80,4 +80,4 @@
80
  }
81
  }
82
  ]
83
- }
 
6
  "source_metadata": {
7
  "source_name": "lighteval",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
13
  "name": "lighteval",
14
+ "version": "v0.13.0"
15
  },
16
  "model_info": {
17
  "name": "inference-optimization/MiniMax-M2.5.w4a16",
 
80
  }
81
  }
82
  ]
83
+ }
every_eval_ever_results/mmlu_pro_chat.json CHANGED
@@ -6,7 +6,7 @@
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
- "source_organization_name": "unknown",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
@@ -740,4 +740,4 @@
740
  }
741
  }
742
  ]
743
- }
 
6
  "source_metadata": {
7
  "source_name": "lm-evaluation-harness",
8
  "source_type": "evaluation_run",
9
+ "source_organization_name": "RedHatAI",
10
  "evaluator_relationship": "third_party"
11
  },
12
  "eval_library": {
 
740
  }
741
  }
742
  ]
743
+ }