Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
image
imagewidth (px)
247
1.7k

RPC-Bench: A Fine-grained Benchmark for Research Paper Comprehension

🌐 Project Page β€’ πŸ’» GitHub β€’ πŸ“– Paper β€’ πŸ€— Paper β€’ 🧭 ModelScope

RPC-Bench is a fine-grained benchmark for research paper comprehension. It is built from review-rebuttal exchanges of high-quality academic papers and supports both text-only and visual evaluation through complementary paper representations.

Data Structure

RPC-Bench is split into train, dev, and test subsets. Each subset is stored in the dataset structure and recorded in manifest.jsonl.

md/ contains Markdown files parsed from each paper by MinerU. These files provide the text input for LLM-oriented evaluation.

parse/ contains the full MinerU parsing outputs for each paper, including structured layout and content artifacts.

pdf/ contains the original paper PDFs.

vlm/ contains page images rendered from the PDFs with PyMuPDF at 200 DPI for VLM-oriented evaluation.

RPC-Bench/
β”œβ”€β”€ README.md
β”œβ”€β”€ manifest.jsonl
β”œβ”€β”€ parse/
β”‚   β”œβ”€β”€ train/
β”‚   β”‚   └── <paper_id>/
β”‚   β”œβ”€β”€ dev/
β”‚   β”‚   └── <paper_id>/
β”‚   └── test/
β”‚       └── <paper_id>/
β”œβ”€β”€ md/
β”‚   β”œβ”€β”€ train/
β”‚   β”‚   └── <paper_id>/
β”‚   β”‚       └── <paper_id>.md
β”‚   β”œβ”€β”€ dev/
β”‚   β”‚   └── <paper_id>/
β”‚   β”‚       └── <paper_id>.md
β”‚   └── test/
β”‚       └── <paper_id>/
β”‚           └── <paper_id>.md
β”œβ”€β”€ pdf/
β”‚   β”œβ”€β”€ train/
β”‚   β”‚   └── <paper_id>.pdf
β”‚   β”œβ”€β”€ dev/
β”‚   β”‚   └── <paper_id>.pdf
β”‚   └── test/
β”‚       └── <paper_id>.pdf
└── vlm/
    β”œβ”€β”€ train/
    β”‚   └── <paper_id>/
    β”œβ”€β”€ dev/
    β”‚   └── <paper_id>/
    └── test/
        └── <paper_id>/

Practical Uses

RPC-Bench can be used to try paper-centric systems that require broader document understanding rather than local snippet matching.

  • Research paper comprehension: try models on full-paper understanding, including core concepts, methods, and experimental findings.
  • Long-context evaluation: try whether longer context windows or long-context architectures improve document-level reasoning.
  • Multimodal reasoning: try models that combine textual evidence with page-level figures, tables, and diagrams in the original PDF layout.
  • RAG system diagnosis: try retrieval, chunking, and evidence-fusion strategies for paper-centric workflows beyond snippet-level retrieval accuracy.

Citation

@article{chen2026rpc,
  title={RPC-Bench: A Fine-grained Benchmark for Research Paper Comprehension},
  author={Chen, Yelin and Zhang, Fanjin and Sun, Suping and Pang, Yunhe and Wang, Yuanchun and Song, Jian and Li, Xiaoyan and Hou, Lei and Zhao, Shu and Tang, Jie and others},
  journal={arXiv preprint arXiv:2601.14289},
  year={2026}
}
Downloads last month
-

Paper for QAQ123/test