| # LLM Benchmarking Project — Dataset (Scientific Replication Benchmark) |
|
|
| This repository contains the **data-only** portion of the Center for Open Science (COS) **LLM Benchmarking Project**. The dataset supports benchmarking LLM agents on core parts of the scientific research lifecycle—especially **replication**—including: |
|
|
| - **Information extraction** from scientific papers into structured JSON |
| - **Research design** and analysis planning |
| - **(Optional) execution support** using provided replication datasets and code |
| - **Scientific interpretation** using human reference materials and expected outputs |
|
|
| Each numbered folder corresponds to **one study instance** in the benchmark. |
|
|
| ## Dataset contents (per study) |
|
|
| Each study folder typically contains: |
|
|
| - `original_paper.pdf` |
| The published paper used as the primary input. |
|
|
| - `initial_details.txt` |
| Brief notes to orient the replication attempt (e.g., key outcomes, hints, pointers). |
|
|
| - `replication_data/` |
| Data and scripts required to reproduce analyses (common formats: `.csv`, `.dta`, `.rds`, `.R`, `.do`, etc.). |
|
|
| - `human_preregistration.(pdf|docx)` |
| Human-created preregistration describing the replication plan. |
|
|
| - `human_report.(pdf|docx)` |
| Human-created replication report describing analyses and findings. |
|
|
| - `expected_post_registration*.json` |
| Expert-annotated ground truth structured outputs used for evaluation. |
| - `expected_post_registration.json` is the primary reference. |
| - `expected_post_registration_2.json`, `_3.json`, etc. are acceptable alternative variants where applicable. |
|
|
| Some studies include multiple acceptable ground-truth variants to capture permissible differences in annotation or representation. |
|
|
| ## Repository structure |
|
|
| At the dataset root, folders like `1/`, `2/`, `10/`, `11/`, etc. are **study IDs**. |
|
|
| Example: |
|
|
| ``` |
| text |
| . |
| ├── 1/ |
| │ ├── expected_post_registration.json |
| │ ├── expected_post_registration_2.json |
| │ ├── human_preregistration.pdf |
| │ ├── human_report.pdf |
| │ ├── initial_details.txt |
| │ ├── original_paper.pdf |
| │ └── replication_data/ |
| │ ├── <data files> |
| │ └── <analysis scripts> |
| ``` |
|
|
| ## Intended uses |
|
|
| This dataset is intended for: |
|
|
| - Benchmarking LLM agents that **extract structured study metadata** from papers |
| - Evaluating LLM systems that generate **replication plans** and analysis specifications |
| - Comparing model outputs against **expert-annotated expected JSON** and human reference docs |
|
|
| ## Not intended for |
|
|
| - Clinical or other high-stakes decision-making |
| - Producing definitive judgments about the original papers |
| - Training models to reproduce copyrighted texts verbatim |
|
|
| ## Quickstart (local) |
|
|
| ### Iterate over studies and load ground truth |
|
|
| ``` |
| python |
| from pathlib import Path |
| import json |
| |
| root = Path(".") |
| study_dirs = sorted( |
| [p for p in root.iterdir() if p.is_dir() and p.name.isdigit()], |
| key=lambda p: int(p.name) |
| ) |
| |
| for study in study_dirs: |
| gt = study / "expected_post_registration.json" |
| if gt.exists(): |
| data = json.loads(gt.read_text(encoding="utf-8")) |
| print(study.name, "ground truth keys:", list(data.keys())[:10]) |
| ``` |
|
|
| ## Using with the main pipeline repository (recommended) |
|
|
| If you are using the **LLM Benchmarking Project** codebase, point the pipeline/evaluators at a given study directory: |
|
|
| ``` |
| bash |
| make evaluate-extract STUDY=/path/to/llm-benchmarking-data/1 |
| ``` |
|
|
| The expected JSON format is defined by the main repository’s templates/schemas. Use those schemas to validate or format model outputs. |
|
|
| ## Notes on multiple expected JSON variants |
|
|
| Some studies include `expected_post_registration_2.json`, `expected_post_registration_3.json`, etc. This is intentional: |
|
|
| - Some fields allow multiple equivalent representations |
| - Annotation may vary slightly without changing meaning |
| - Evaluators may accept any variant depending on scoring rules |
|
|
| If you implement your own scorer, consider: |
| - Exact matching for strictly defined fields |
| - More tolerant matching for lists, notes, or fields with legitimate paraphrase/format variation |
|
|
| ## File formats |
|
|
| You may encounter: |
|
|
| - R artifacts: `.R`, `.rds` |
| - Stata artifacts: `.do`, `.dta` |
| - CSV/tabular data: `.csv` |
| - Documents: `.pdf`, `.docx` |
| - Structured evaluation targets: `.json` |
|
|
| Reproducing analyses may require R and/or Stata depending on the study. |
|
|
| ## Licensing, copyright, and redistribution (important) |
|
|
| This repository is released under **Apache 2.0** for **COS-authored materials and annotations** (for example: benchmark scaffolding, expected JSON outputs, and other COS-created files). |
|
|
| However, some contents may be **third-party materials**, including (but not limited to): |
|
|
| - `original_paper.pdf` (publisher copyright may apply) |
| - `replication_data/` (may have its own license/terms from the original authors) |
| - external scripts or datasets included for replication |
|
|
| **You are responsible for ensuring you have the right to redistribute third-party files publicly** (e.g., GitHub / Hugging Face). |
|
|
| Common options if redistribution is restricted: |
| - Remove third-party PDFs and provide **DOI/URL references** instead |
| - Keep restricted files in a private location and publish only COS-authored annotations |
| - Add per-study `LICENSE` / `NOTICE` files inside each study folder where terms are known |
|
|
| ## Large files (Git LFS recommendation) |
|
|
| If hosting on GitHub, consider Git LFS for PDFs and large datasets: |
|
|
| ``` |
| bash |
| git lfs install |
| git lfs track "*.pdf" "*.dta" "*.rds" |
| git add .gitattributes |
| ``` |
|
|
| ## Citation |
|
|
| If you use this dataset in academic work, please cite it as: |
|
|
| ``` |
| bibtex |
| @dataset{cos_llm_benchmarking_data_2026, |
| author = {Center for Open Science}, |
| title = {LLM Benchmarking Project: Scientific Replication Benchmark Data}, |
| year = {2026}, |
| publisher = {Center for Open Science}, |
| note = {Benchmark dataset for evaluating LLM agents on scientific replication tasks} |
| } |
| ``` |
|
|
| ## Acknowledgements |
|
|
| This project is funded by Coefficient Giving as part of its “Benchmarking LLM Agents on Consequential Real-World Tasks” program. We thank the annotators who contributed to the ground-truth post-registrations for the extraction stage. |
|
|
| ## Contact |
|
|
| For questions about this dataset: |
|
|
| **Shakhlo Nematova** |
| Research Scientist, Center for Open Science |
| shakhlo@cos.io |
|
|