Datasets:
id stringlengths 36 36 | input stringlengths 11 614 | expected_output stringlengths 26 1.06k | metadata.query_id stringlengths 1 6 | metadata.split stringclasses 1
value |
|---|---|---|---|---|
7e965799-c99c-46a6-95ab-91dae89ecd4f | Robert Duvall has not won a BAFTA. | [{"id": "Robert_Duvall", "score": 1}] | 145027 | train |
a2b4302d-b0cb-4917-ab1a-67622b3c9790 | Reese Witherspoon grew up in the United States. | [{"id": "Tennessee", "score": 1}, {"id": "New_Orleans", "score": 1}, {"id": "Reese_Witherspoon", "score": 1}] | 160148 | train |
013bc795-aaff-4f08-b376-a4b61f7a6a82 | Ellen Burstyn began her career anywhere but the theater. | [{"id": "Ellen_Burstyn", "score": 1}] | 180393 | train |
b4489144-099e-425f-9940-e63d585d1e7f | United Nations is an established organization. | [{"id": "United_Nations", "score": 1}] | 9676 | train |
573b2699-cceb-44d0-a0ab-d7262607b58b | The album Take Me Home was released by Harry Styles and his bandmates. | [{"id": "Harry_Styles", "score": 1}] | 202101 | train |
7abe2fb8-bdbe-4fcb-8a6d-e6a2ad4e78ed | Madonna's albums are all untitled. | [{"id": "Like_a_Virgin_(song)", "score": 1}, {"id": "Madonna_(entertainer)", "score": 1}, {"id": "Madonna_(Madonna_album)", "score": 1}] | 127190 | train |
9a878134-a114-47fc-a1f2-fd6845f82f4b | Jane Eyre is a work. | [{"id": "Jane_Eyre", "score": 1}] | 205059 | train |
835775d7-71f7-452c-8710-20b2ef1348fe | An American crime drama television series starred Courtney Love. | [{"id": "Courtney_Love", "score": 1}, {"id": "Sons_of_Anarchy", "score": 1}] | 545 | train |
5d8bc6ea-f744-46a0-932b-59f611495ec0 | NSYNC was launched by the German record label BMG Ariola Munich. | [{"id": "NSYNC", "score": 1}] | 19482 | train |
0920962a-a3b8-4362-805b-6dc47df5a95d | Vincent Cassel starred in Ocean's Thirteen. | [{"id": "Vincent_Cassel", "score": 1}] | 95727 | train |
088d92f1-0f2e-4ec4-940c-d130a3568a60 | Hungary did not have a nationwide revolt in 1956. | [{"id": "Hungary", "score": 1}] | 77943 | train |
b8e15e9f-fc0f-42ff-b53c-8f9b986d3b45 | Gemini was released in 2003. | [{"id": "Gemini_(2002_Tamil_film)", "score": 1}] | 50241 | train |
0fb63eb8-fa87-4a53-afd1-b6c6dad2f958 | The United States has a coastline. | [{"id": "East_Coast_of_the_United_States", "score": 1}, {"id": "United_States", "score": 1}] | 157941 | train |
622d08cb-5ee6-4cfe-a982-bde667dc9a1d | Modern liberalism in the United States is strongly against welfare. | [{"id": "Modern_liberalism_in_the_United_States", "score": 1}] | 2794 | train |
7bd430e5-e206-498c-b71d-d050b920e68a | Joan Crawford was married. | [{"id": "Joan_Crawford", "score": 1}] | 50440 | train |
fee0b63f-c8f6-47f5-9955-b853b0366065 | Mary-Louise Parker appeared in Boys on the Side. | [{"id": "Mary-Louise_Parker", "score": 1}] | 718 | train |
90fcc164-0143-4330-8b12-1612736e24ae | Deadpool's film development started in February of 2004. | [{"id": "Deadpool_(film)", "score": 1}] | 149740 | train |
58b69830-819e-4854-8c1a-78cb16392836 | Gold has been used for money. | [{"id": "Monetary_policy", "score": 1}, {"id": "Gold", "score": 1}, {"id": "Gold_standard", "score": 1}] | 82565 | train |
808b6c41-af76-4039-bd66-6ee7e8a6b512 | Citadelle Laferrière is also called the Fortress. | [{"id": "Citadelle_Laferrière", "score": 1}] | 40928 | train |
c1e1f23c-e574-46dc-a4d2-31552cfaeeff | The Big Trail stars an actor. | [{"id": "The_Big_Trail", "score": 1}] | 138727 | train |
c33f68da-c02d-47fe-8caf-4b8511c625c4 | Rope was not released in 1948. | [{"id": "Rope_(film)", "score": 1}] | 35233 | train |
ae3db644-f460-46c2-9bcf-be496e2c1ca8 | Moses Malone was unable to earn any All-Star selections during his first season with the Atlanta Hawks. | [{"id": "Moses_Malone", "score": 1}] | 210465 | train |
121b1bc5-e567-4912-ba46-c2c1aa7445f0 | Jeff Bezos is not the manufacturer of Blue Origin. | [{"id": "Jeff_Bezos", "score": 1}] | 101730 | train |
aa6f4018-f0da-46f5-b340-01945f238f31 | Ronaldo Maczinski's birthday is August 11. | [{"id": "Ronaldo_Maczinski", "score": 1}] | 210058 | train |
c4c9ccdf-6f9b-4a02-b9a7-9bcfc0a728d3 | Honey to the B was the first album of an English singer, dancer, and actress. | [{"id": "Honey_to_the_B", "score": 1}, {"id": "Billie_Piper", "score": 1}] | 126940 | train |
39ab3a08-9a45-4685-af92-0984b465e497 | Home Alone is a Christmas carol. | [{"id": "Home_Alone", "score": 1}] | 104969 | train |
8bcc6c68-a88a-49c9-83f6-b0bc6a8f1c53 | Eagles reunited in the 1990s. | [{"id": "Eagles_(band)", "score": 1}] | 153936 | train |
898135c1-3d86-4599-8019-4034faa286ea | Venus is a planet that is terrestrial. | [{"id": "Venus", "score": 1}] | 7996 | train |
8d89bca6-c820-438a-a45e-65f435c63e45 | Plato's student was a scientist. | [{"id": "Aristotle", "score": 1}, {"id": "Plato", "score": 1}] | 151516 | train |
ab84d93a-e11c-4a0d-bfb3-e335c418f345 | Spider-Man 3 features a fictional character. | [{"id": "Mary_Jane_Watson", "score": 1}, {"id": "Spider-Man_3", "score": 1}, {"id": "Gwen_Stacy", "score": 1}, {"id": "Venom_(comics)", "score": 1}, {"id": "Harry_Osborn", "score": 1}, {"id": "Sandman_(Marvel_Comics)", "score": 1}] | 129549 | train |
d91ca520-1fdf-4fa6-8d10-d8f539d4597c | BTS is a Korean pop group. | [{"id": "BTS_(band)", "score": 1}] | 168746 | train |
38ca20ff-4676-471e-8b2a-33d17f934d45 | Leonardo Bonucci has won several honors. | [{"id": "Leonardo_Bonucci", "score": 1}] | 99837 | train |
3bd77499-f7d7-4868-9d28-13f6a383e7f2 | Geraldine Chaplin portrayed Tonya in Dr. Dolittle. | [{"id": "Geraldine_Chaplin", "score": 1}] | 163862 | train |
a59f3150-a9eb-4caf-bf3c-9c245cafc4d8 | The 65th US Secretary of State was Colin Powell. | [{"id": "Colin_Powell", "score": 1}] | 171006 | train |
efdc071d-234f-42dd-bd07-1aeda7da5ad0 | Mel Gibson was born in a month. | [{"id": "Mel_Gibson", "score": 1}] | 182983 | train |
efaaf982-e1dd-4b68-b9b4-cda68a594b1a | There is a video game called Splatoon 2. | [{"id": "Splatoon_2", "score": 1}] | 69343 | train |
2c98f471-5f18-440b-bf6e-832a4a88ecf3 | Chandni is a film. | [{"id": "Chandni", "score": 1}] | 214368 | train |
c2612214-5819-43ce-9ff4-acbc315cfe64 | The Lion is a movie. | [{"id": "Lion", "score": 1}] | 54040 | train |
ff253396-d4cf-437b-a945-eb60d1dd4ef3 | Dan Aykroyd acted in the 1989 film Driving Miss Daisy. | [{"id": "Dan_Aykroyd", "score": 1}] | 189242 | train |
10315958-f0ec-4b3a-a209-8c2eec34a6cc | David Angell was born on April 10, 1946. | [{"id": "David_Angell", "score": 1}] | 88454 | train |
d30f2ca8-6d0e-428f-ab55-bbd7912128fc | David Tennant was in Doctor Who. | [{"id": "David_Tennant", "score": 1}] | 99483 | train |
3d6a5b34-e5f1-4a5d-a490-33e96a9ecb47 | Matt Sorum's touring project Kings of Chaos features members of Slipknot. | [{"id": "Matt_Sorum", "score": 1}] | 201617 | train |
4baf929a-5591-48c6-914f-a36412b90850 | Michelangelo was a person. | [{"id": "Michelangelo", "score": 1}] | 124974 | train |
d2b724e1-511e-49f1-bccc-ab9ddfd2c249 | Ron Perlman birth date is April 13, 1950. | [{"id": "Ron_Perlman", "score": 1}] | 42359 | train |
9be1d7ac-33b4-4a4c-8bea-b6c291b0ff3c | John Lennon died before 1987. | [{"id": "John_Lennon", "score": 1}] | 6643 | train |
c11011e7-ca16-4f96-be5d-7a7b2f10b4c2 | Evil is commonly associated with hatred. | [{"id": "Evil", "score": 1}] | 33575 | train |
5dcf4861-ef8c-4bbc-a747-d46849043507 | The Dayton Agreement has other names. | [{"id": "Dayton_Agreement", "score": 1}] | 217061 | train |
d5cc1f51-99b4-473f-8a1e-58d4ce8d07b5 | Tré Cool is American. | [{"id": "Tré_Cool", "score": 1}] | 194974 | train |
dbfb9127-6326-4055-87b1-19e21457c826 | Kiefer Sutherland received a nomination. | [{"id": "Kiefer_Sutherland", "score": 1}] | 229006 | train |
28132104-bcfb-4629-92d2-346b4ea5418f | Modern liberalism in the United States strongly supports spending on education. | [{"id": "Modern_liberalism_in_the_United_States", "score": 1}] | 71027 | train |
2bdd039c-de07-4721-b0dc-0882417ef1b9 | Liverpool F.C. is a team in the MLB. | [{"id": "Liverpool_F.C.", "score": 1}] | 42818 | train |
b8d573cd-7064-4152-9c84-2d7f5794cf08 | Abraham Lincoln spoke out against the expansion of slavery in 1858. | [{"id": "Abraham_Lincoln", "score": 1}] | 227650 | train |
0b1184be-3dab-4ad3-868f-90aacbcdf285 | Anne Frank's birthplace was an urban area in Germany. | [{"id": "Frankfurt", "score": 1}, {"id": "Anne_Frank", "score": 1}] | 38982 | train |
aeca0d45-ca16-43ec-98cb-f5dab39c427f | Dr. Dre was previously an artist on dead bodies. | [{"id": "Dr._Dre", "score": 1}] | 40648 | train |
31e5fd38-f825-4d76-9ed4-7a72ef894a5e | Francois de Belleforest translated works. | [{"id": "François_de_Belleforest", "score": 1}] | 192534 | train |
388e58fd-e975-47e4-a8cd-b33edf24ed54 | Marilyn Monroe worked with the First Motion Picture Unit. | [{"id": "Marilyn_Monroe", "score": 1}] | 40276 | train |
098f5d01-db1d-4eb2-b058-e99346850254 | Sunrise created anime. | [{"id": "Sunrise_(company)", "score": 1}] | 176382 | train |
0ef3c92d-6785-434e-b60a-d65b196460d7 | Wikipedia doesn't have any articles. | [{"id": "Wikipedia", "score": 1}] | 16966 | train |
03efc9ed-5cc3-4b20-8e9c-819a8526454d | The Expendables 3 was Ronda Rousey's first feature film. | [{"id": "Ronda_Rousey", "score": 1}] | 55137 | train |
057fade9-66a9-4463-a3f4-261c021f164a | Thiokol was also known as anything except Cordant Technologies Inc. | [{"id": "Thiokol", "score": 1}] | 188119 | train |
9566e1c3-da17-4313-a996-966b4b54915e | Three albums have been released by Radiohead. | [{"id": "Radiohead", "score": 1}] | 123988 | train |
9a05634c-1e20-4a51-9af9-1c49ae75aaff | Brandon Lee died during filming due to an accident involving a prop gun. | [{"id": "Brandon_Lee", "score": 1}] | 169690 | train |
eb49b60f-6c20-4979-b820-c35ecb8fa68f | Robert Browning was born. | [{"id": "Robert_Browning", "score": 1}, {"id": "Browning_Society", "score": 1}] | 219242 | train |
bf838ab0-aea3-47cb-b1b8-706d000f573c | Daniel Craig attended a school. | [{"id": "Daniel_Craig", "score": 1}] | 119626 | train |
5dd8bd9e-e9d2-4ab1-918e-001662e6d8cc | Denmark's most populous city is Copenhagen with two million citizens. | [{"id": "Copenhagen", "score": 1}, {"id": "Denmark", "score": 1}] | 101561 | train |
531e8345-2eff-41d1-8b68-aabe6101bc42 | Telangana borders another territory. | [{"id": "Telangana", "score": 1}] | 151598 | train |
3aecea1a-4a20-4674-8bd1-a2cc046b04cb | As the World Turns is set in Oakdale. | [{"id": "As_the_World_Turns", "score": 1}] | 36001 | train |
73e8157e-5721-410b-b1dc-656c745caad5 | The Caracazo took place in Morocco. | [{"id": "Caracazo", "score": 1}] | 217247 | train |
93133756-3113-433f-861c-b209788492c0 | Richard Ramirez committed murder. | [{"id": "Richard_Ramirez", "score": 1}] | 128058 | train |
e97a6313-b923-4797-b080-15f337f01606 | Maria Sharapova held the title of the world No. 1 zero times. | [{"id": "Maria_Sharapova", "score": 1}] | 111065 | train |
cdac00f7-25e8-4e82-b69b-dfe2e9ce69ef | Hotel Transylvania 2 featured Adam Sandler as a voice actor. | [{"id": "Hotel_Transylvania_2", "score": 1}] | 98401 | train |
d8515320-6374-4280-8beb-cd7fbd8723c5 | Rajesh Khanna earned the most money as an actor from 1970 to 1987. | [{"id": "Rajesh_Khanna", "score": 1}] | 43016 | train |
01121458-f474-4ebc-bdd3-303efa106d91 | The Host stars an actor who's birthday is on November 18. | [{"id": "Jake_Abel", "score": 1}, {"id": "The_Host_(2013_film)", "score": 1}] | 215363 | train |
75e352c0-92ac-4af6-a793-a1485c32fb06 | Anne Hathaway lost the Critics' Choice Movie Award for Best Actress whenever she was eligible. | [{"id": "Anne_Hathaway", "score": 1}] | 60864 | train |
fb3f85e8-38c5-442d-90cf-b9c8ccc3f367 | Silicon Valley refers to the southern part of the San Francisco Bay Area in California. | [{"id": "Silicon_Valley", "score": 1}] | 103356 | train |
00a99242-4358-4c6e-859d-f5d85312b019 | Rihanna is the oldest and slowest solo artist to earn fourteen number-one singles on the Billboard Hot 100. | [{"id": "Rihanna", "score": 1}] | 3719 | train |
946b8a1d-c191-4638-86a3-03731fbec941 | Woody Allen is a computer program. | [{"id": "Woody_Allen", "score": 1}] | 193111 | train |
fac8171a-4908-404d-9e45-be7b0f26cfbc | Jet Li was the main actor in a martial arts movie. | [{"id": "Unleashed_(film)", "score": 1}, {"id": "Shaolin_Temple_(1982_film)", "score": 1}, {"id": "Fist_of_Legend", "score": 1}, {"id": "Jet_Li", "score": 1}] | 35285 | train |
8ce57629-becb-4764-b0b2-cc7672d45e03 | David Beckham played in the Premier League. | [{"id": "David_Beckham", "score": 1}] | 59451 | train |
3fd2d18f-45b4-40d5-8e01-60604c1df2a9 | Veronica Mars is a protagonist. | [{"id": "Veronica_Mars_(character)", "score": 1}] | 125570 | train |
f32f0915-4fd1-4c13-a392-5d958516fb67 | A Game of Thrones is a Netflix series. | [{"id": "A_Game_of_Thrones", "score": 1}] | 78945 | train |
269b0a0d-aeb3-4843-96ef-7e96305f942e | Yelp existed in 2009. | [{"id": "Yelp", "score": 1}] | 179392 | train |
5dfaa848-e203-4603-8c6c-71fde94a157c | FC Barcelona is in third position on the UEFA club rankings. | [{"id": "FC_Barcelona", "score": 1}] | 41036 | train |
94cba2a4-d77a-4aa8-9bd4-588cad3647f0 | Grand Theft Auto V was developed by Rockstar Games. | [{"id": "Grand_Theft_Auto_V", "score": 1}] | 52116 | train |
9964874c-0623-4043-8f9d-d379e802f362 | John Denver is a musician. | [{"id": "John_Denver", "score": 1}] | 28994 | train |
a39a6951-880c-4a58-b165-c27ea51781fe | Brave is a television series. | [{"id": "Brave_(2012_film)", "score": 1}] | 209145 | train |
551e060d-4f77-433f-a749-11b6908831b6 | K2 is also known as the Savage Mountain. | [{"id": "K2", "score": 1}] | 23903 | train |
25afe5f2-7624-49a3-b6ec-f2d146bfea88 | Overwatch lets players gain cosmetic awards which include character skins and victory poses. | [{"id": "Overwatch_(video_game)", "score": 1}] | 147205 | train |
6d4c84cd-f490-40b5-9198-50ce820f8828 | Diego Costa fails to play as a striker. | [{"id": "Diego_Costa", "score": 1}] | 144290 | train |
0730335a-8313-4dd4-a13b-9785e270ca74 | Brandon Lee died in a car crash in 1998. | [{"id": "Brandon_Lee", "score": 1}] | 169703 | train |
117693d3-30e4-4772-9c6a-cbc3e54f725d | In the 1990's, Steve Buscemi was in movies. | [{"id": "Desperado_(film)", "score": 1}, {"id": "Con_Air", "score": 1}, {"id": "Steve_Buscemi", "score": 1}, {"id": "Reservoir_Dogs", "score": 1}, {"id": "Armageddon_(1998_film)", "score": 1}] | 52557 | train |
dbbcc86b-76c3-46d9-9c48-97f439676530 | The 1983-94 season ended. | [{"id": "1983–84_NBA_season", "score": 1}] | 101318 | train |
e0a83cc8-0e5e-4407-bfa5-06485a47fac8 | Rio de Janeiro is in a rural area. | [{"id": "Rio_de_Janeiro", "score": 1}] | 52722 | train |
340cf48e-ed99-440c-a741-4c68e6911c89 | CeeLo Green has worked with another artist. | [{"id": "CeeLo_Green", "score": 1}, {"id": "Gnarls_Barkley", "score": 1}] | 128871 | train |
dfaa2b51-0c1f-4ef5-b936-b74fb8806140 | Albany, New York, is the longest continuously chartered city in North America. | [{"id": "Albany,_New_York", "score": 1}] | 212173 | train |
ea814ea3-089b-4386-a4b7-769cc2eae7da | The 21st was the day when Sophie Turner was born. | [{"id": "Sophie_Turner", "score": 1}] | 78136 | train |
6d4f148a-bbc4-43ad-bca1-4ec946bda12e | Frenemies was aired after 2010. | [{"id": "Frenemies_(film)", "score": 1}] | 177565 | train |
83deee6b-25ba-4f5f-811c-5bc15a194a9f | The United States Naval Academy is in Spain. | [{"id": "Fort_Severn", "score": 1}, {"id": "United_States_Naval_Academy", "score": 1}] | 101435 | train |
52bb2e36-ef2e-46b0-8623-d7159e13aa91 | Tennis is the kind of sport that uses racquets. | [{"id": "Tennis", "score": 1}] | 106131 | train |
93d697fd-492e-426d-96e3-0118bef12551 | The Woman in Red is a romantic tragedy. | [{"id": "The_Woman_in_Red_(1984_film)", "score": 1}] | 202908 | train |
FEVER (BEIR) — Fact-checking retrieval
Dataset description
FEVER (Fact Extraction and VERification) is a large-scale English dataset for claim verification against textual sources. Claims were produced by altering sentences drawn from Wikipedia; annotators then labeled each claim without knowing which source sentence it came from. Labels are Supported, Refuted, or NotEnoughInfo (with substantial inter-annotator agreement). For Supported and Refuted claims, annotators also identified the sentence-level evidence needed to justify the label.
BEIR (Benchmarking IR) repackaged FEVER—along with many other public corpora—as a standard retrieval benchmark for zero-shot evaluation of dense, sparse, and hybrid information retrieval models across heterogeneous tasks. In the BEIR formulation, each claim acts as a query, and the objective is to retrieve relevant Wikipedia documents (by title) that contain the evidence required for verification. This setting isolates retrieval quality as the variable of interest when paired with a fixed downstream verifier or when reporting standard IR metrics.
This repository (orgrctera/beir_fever) provides train / validation / test splits in Parquet form for retrieval evaluation pipelines. Each row is one query (a claim) with relevance judgments pointing at corpus document identifiers in the BEIR FEVER benchmark (Wikipedia article titles as used upstream).
Scale and domain (BEIR FEVER)
The original FEVER release comprises on the order of 185k verified claims; BEIR’s FEVER split follows the standard BEIR packaging. The corpus is Wikipedia-oriented text keyed by article identifiers (titles in the BEIR release). Exact counts for this Hub snapshot follow the upstream BEIR FEVER release—see the BEIR repository for version-precise figures.
Task: retrieval (FEVER in BEIR)
The task is ad hoc document retrieval for fact-checking:
- Input: a natural-language claim (the query).
- Output: a ranked list of document IDs from the FEVER corpus (Wikipedia titles in the BEIR distribution), or scores over the full collection, such that relevant IDs—according to the official qrels—receive high rank.
Downstream metrics are standard IR metrics (e.g., nDCG@k, Recall@k, MRR), as in BEIR’s evaluation scripts or frameworks such as Pyserini / the BEIR API.
Note: Full retrieval evaluation also requires the corpus (passage or document text keyed by the same IDs). This card describes the query + qrels side as prepared for CTERA-style evaluation rows; align corpus IDs with the same BEIR FEVER corpus you use for indexing.
Data format (this repository)
Each record includes:
| Field | Description |
|---|---|
id |
UUID for this example row. |
input |
The claim text (query). |
expected_output |
JSON string: list of objects {"id": "<corpus-doc-id>", "score": <relevance>}. Document IDs are Wikipedia article titles as in the BEIR FEVER corpus; scores follow the BEIR qrels convention (typically 1 for relevant in binary settings). |
metadata.query_id |
Original BEIR / FEVER query identifier (string). |
metadata.split |
Split name: train, dev, or test. |
Example 1
{
"id": "7e965799-c99c-46a6-95ab-91dae89ecd4f",
"input": "Robert Duvall has not won a BAFTA.",
"expected_output": "[{\"id\": \"Robert_Duvall\", \"score\": 1}]",
"metadata.query_id": "145027",
"metadata.split": "train"
}
Example 2
{
"id": "a2b4302d-b0cb-4917-ab1a-67622b3c9790",
"input": "Reese Witherspoon grew up in the United States.",
"expected_output": "[{\"id\": \"Tennessee\", \"score\": 1}, {\"id\": \"New_Orleans\", \"score\": 1}, {\"id\": \"Reese_Witherspoon\", \"score\": 1}]",
"metadata.query_id": "160148",
"metadata.split": "train"
}
References
FEVER (original dataset)
James Thorne, Andreas Vlachos, Christos Christodoulopoulos, Arpit Mittal
FEVER: a large-scale dataset for Fact Extraction and VERification
Presented at NAACL 2018; extended version on arXiv.
Abstract (from arXiv): “In this paper we introduce a new publicly available dataset for verification against textual sources, FEVER: Fact Extraction and VERification. It consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo by annotators achieving 0.6841 in Fleiss κ. For the first two classes, the annotators also recorded the sentence(s) forming the necessary evidence for their judgment. To characterize the challenge of the dataset presented, we develop a pipeline approach and compare it to suitably designed oracles. The best accuracy we achieve on labeling a claim accompanied by the correct evidence is 31.87%, while if we ignore the evidence we achieve 50.91%. Thus we believe that FEVER is a challenging testbed that will help stimulate progress on claim verification against textual sources.”
- Paper: arXiv:1803.05355 — PDF.
BEIR benchmark (FEVER as a subset)
Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, Iryna Gurevych
BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models
NeurIPS 2021 (Datasets and Benchmarks Track).
Abstract (from arXiv): “Existing neural information retrieval (IR) models have often been studied in homogeneous and narrow settings, which has considerably limited insights into their out-of-distribution (OOD) generalization capabilities. To address this, and to facilitate researchers to broadly evaluate the effectiveness of their models, we introduce Benchmarking-IR (BEIR), a robust and heterogeneous evaluation benchmark for information retrieval. We leverage a careful selection of 18 publicly available datasets from diverse text retrieval tasks and domains and evaluate 10 state-of-the-art retrieval systems including lexical, sparse, dense, late-interaction and re-ranking architectures on the BEIR benchmark. Our results show BM25 is a robust baseline and re-ranking and late-interaction-based models on average achieve the best zero-shot performances, however, at high computational costs. In contrast, dense and sparse-retrieval models are computationally more efficient but often underperform other approaches, highlighting the considerable room for improvement in their generalization capabilities.”
- Paper: arXiv:2104.08663 — OpenReview; code and data: BEIR on GitHub.
Related resources
- BEIR FEVER mirrors on Hugging Face in classic BEIR layout, e.g.
BeIR/fever(corpus / queries / qrels). - IRDS packaging:
irds/beir_feverexposes FEVER via the ir-datasets tooling.
Citation
If you use FEVER, cite the FEVER paper (Thorne et al., NAACL 2018 / arXiv:1803.05355). If you use the BEIR benchmark formulation, cite the BEIR paper (Thakur et al., NeurIPS 2021). BibTeX for BEIR is available in the official repository.
License
FEVER’s evidence and corpus trace to Wikipedia-derived text; Wikipedia content is typically licensed under Creative Commons Attribution-ShareAlike (version depends on the snapshot). This card marks cc-by-sa-4.0 as a common umbrella for Wikipedia-derived redistribution; verify against your corpus snapshot and upstream terms if compliance is strict.
Dataset card maintained for the orgrctera/beir_fever Hub repository.
- Downloads last month
- 36