RePOPE: Revisiting Partial Object Hallucination Evaluation

RePOPE is a re-annotated version of the POPE (Polling-based Object Probing Evaluation) benchmark with corrected ground-truth labels. It evaluates object hallucination in multimodal large language models (MLLMs) by asking yes/no questions about object existence in MSCOCO images.

Dataset Details

Dataset Structure

Each row contains:

  • image: The MSCOCO image (struct with bytes and path)
  • image_id: COCO image identifier (e.g., 000000310196)
  • question: A yes/no question about object presence (e.g., "Is there a snowboard in the image?")
  • answer: Ground truth label (yes or no)
  • category: Sampling strategy used to select the queried object (random, popular, or adversarial)

Splits

This dataset contains all three POPE sampling categories in a single split:

Category Count
random 2,774
popular 2,727
adversarial 2,684
Total 8,185

Label Distribution

Answer Count
yes 3,539
no 4,646

How to Use

from datasets import load_dataset

ds = load_dataset("MM-Hallu/RePOPE")

Citation

@misc{neuhaus2024repope,
      title={RePOPE: Revisiting Partial Object Hallucination Evaluation},
      author={Yannik Neuschwander and Selen Yu and Jordy Van Landeghem and Jan Van Loock and Lilian Ngweta and Rukiye Savran Kizildag and Desmond Elliott and Matthew B. Blaschko},
      year={2024},
      eprint={2405.14571},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Paper for MM-Hallu/RePOPE