| | """Ground truth comparison using GAIA validation dataset. |
| | |
| | Author: @mangubee |
| | |
| | Since the GAIA API only returns summary stats (X/Y correct) without per-question |
| | correctness, we load the public validation dataset to compare our answers locally. |
| | This enables per-question debugging and error analysis. |
| | """ |
| |
|
| | import os |
| | import logging |
| | from typing import Dict, Optional |
| |
|
| | logger = logging.getLogger(__name__) |
| |
|
| | |
| | |
| | |
| | CACHE_DIR = os.path.expanduser("~/.cache/gaia_dataset") |
| | |
| |
|
| |
|
| | class GAIAGroundTruth: |
| | """Load GAIA validation dataset and provide ground truth answers.""" |
| |
|
| | def __init__(self): |
| | """Initialize ground truth loader.""" |
| | self.ground_truth: Dict[str, str] = {} |
| | self.metadata: Dict[str, dict] = {} |
| | self._loaded = False |
| |
|
| | def load_validation_set(self) -> bool: |
| | """Load GAIA validation dataset from HuggingFace. |
| | |
| | Returns: |
| | bool: True if loaded successfully, False otherwise |
| | """ |
| | if self._loaded: |
| | return True |
| |
|
| | try: |
| | from datasets import load_dataset |
| |
|
| | logger.info("Loading GAIA validation dataset...") |
| |
|
| | |
| | |
| | dataset = load_dataset( |
| | "gaia-benchmark/GAIA", |
| | "2023_all", |
| | split="validation", |
| | cache_dir=CACHE_DIR |
| | ) |
| |
|
| | |
| | for item in dataset: |
| | task_id = item.get("task_id") |
| | final_answer = item.get("Final answer") |
| |
|
| | if task_id and final_answer: |
| | self.ground_truth[task_id] = str(final_answer).strip() |
| | |
| | self.metadata[task_id] = dict(item) |
| |
|
| | self._loaded = True |
| | logger.info(f"Loaded {len(self.ground_truth)} ground truth answers") |
| | return True |
| |
|
| | except Exception as e: |
| | logger.error(f"Failed to load GAIA dataset: {e}") |
| | return False |
| |
|
| | def get_answer(self, task_id: str) -> Optional[str]: |
| | """Get ground truth answer for a task_id. |
| | |
| | Args: |
| | task_id: Question task ID |
| | |
| | Returns: |
| | Ground truth answer or None if not found |
| | """ |
| | if not self._loaded: |
| | self.load_validation_set() |
| |
|
| | return self.ground_truth.get(task_id) |
| |
|
| | def compare_answer(self, task_id: str, submitted_answer: str) -> Optional[bool]: |
| | """Compare submitted answer against ground truth. |
| | |
| | Args: |
| | task_id: Question task ID |
| | submitted_answer: Answer submitted by agent |
| | |
| | Returns: |
| | True if correct, False if incorrect, None if no ground truth available |
| | """ |
| | ground_truth = self.get_answer(task_id) |
| |
|
| | if ground_truth is None: |
| | return None |
| |
|
| | |
| | submitted = str(submitted_answer).strip().lower() |
| | expected = str(ground_truth).strip().lower() |
| |
|
| | |
| | return submitted == expected |
| |
|
| |
|
| | |
| | _ground_truth_instance = None |
| |
|
| |
|
| | def get_ground_truth() -> GAIAGroundTruth: |
| | """Get or create singleton ground truth instance. |
| | |
| | Returns: |
| | GAIAGroundTruth instance |
| | """ |
| | global _ground_truth_instance |
| |
|
| | if _ground_truth_instance is None: |
| | _ground_truth_instance = GAIAGroundTruth() |
| |
|
| | return _ground_truth_instance |
| |
|