Dataset Viewer
Auto-converted to Parquet Duplicate
id
stringlengths
36
36
input
stringlengths
11
614
expected_output
stringlengths
26
1.06k
metadata.query_id
stringlengths
1
6
metadata.split
stringclasses
1 value
7e965799-c99c-46a6-95ab-91dae89ecd4f
Robert Duvall has not won a BAFTA.
[{"id": "Robert_Duvall", "score": 1}]
145027
train
a2b4302d-b0cb-4917-ab1a-67622b3c9790
Reese Witherspoon grew up in the United States.
[{"id": "Tennessee", "score": 1}, {"id": "New_Orleans", "score": 1}, {"id": "Reese_Witherspoon", "score": 1}]
160148
train
013bc795-aaff-4f08-b376-a4b61f7a6a82
Ellen Burstyn began her career anywhere but the theater.
[{"id": "Ellen_Burstyn", "score": 1}]
180393
train
b4489144-099e-425f-9940-e63d585d1e7f
United Nations is an established organization.
[{"id": "United_Nations", "score": 1}]
9676
train
573b2699-cceb-44d0-a0ab-d7262607b58b
The album Take Me Home was released by Harry Styles and his bandmates.
[{"id": "Harry_Styles", "score": 1}]
202101
train
7abe2fb8-bdbe-4fcb-8a6d-e6a2ad4e78ed
Madonna's albums are all untitled.
[{"id": "Like_a_Virgin_(song)", "score": 1}, {"id": "Madonna_(entertainer)", "score": 1}, {"id": "Madonna_(Madonna_album)", "score": 1}]
127190
train
9a878134-a114-47fc-a1f2-fd6845f82f4b
Jane Eyre is a work.
[{"id": "Jane_Eyre", "score": 1}]
205059
train
835775d7-71f7-452c-8710-20b2ef1348fe
An American crime drama television series starred Courtney Love.
[{"id": "Courtney_Love", "score": 1}, {"id": "Sons_of_Anarchy", "score": 1}]
545
train
5d8bc6ea-f744-46a0-932b-59f611495ec0
NSYNC was launched by the German record label BMG Ariola Munich.
[{"id": "NSYNC", "score": 1}]
19482
train
0920962a-a3b8-4362-805b-6dc47df5a95d
Vincent Cassel starred in Ocean's Thirteen.
[{"id": "Vincent_Cassel", "score": 1}]
95727
train
088d92f1-0f2e-4ec4-940c-d130a3568a60
Hungary did not have a nationwide revolt in 1956.
[{"id": "Hungary", "score": 1}]
77943
train
b8e15e9f-fc0f-42ff-b53c-8f9b986d3b45
Gemini was released in 2003.
[{"id": "Gemini_(2002_Tamil_film)", "score": 1}]
50241
train
0fb63eb8-fa87-4a53-afd1-b6c6dad2f958
The United States has a coastline.
[{"id": "East_Coast_of_the_United_States", "score": 1}, {"id": "United_States", "score": 1}]
157941
train
622d08cb-5ee6-4cfe-a982-bde667dc9a1d
Modern liberalism in the United States is strongly against welfare.
[{"id": "Modern_liberalism_in_the_United_States", "score": 1}]
2794
train
7bd430e5-e206-498c-b71d-d050b920e68a
Joan Crawford was married.
[{"id": "Joan_Crawford", "score": 1}]
50440
train
fee0b63f-c8f6-47f5-9955-b853b0366065
Mary-Louise Parker appeared in Boys on the Side.
[{"id": "Mary-Louise_Parker", "score": 1}]
718
train
90fcc164-0143-4330-8b12-1612736e24ae
Deadpool's film development started in February of 2004.
[{"id": "Deadpool_(film)", "score": 1}]
149740
train
58b69830-819e-4854-8c1a-78cb16392836
Gold has been used for money.
[{"id": "Monetary_policy", "score": 1}, {"id": "Gold", "score": 1}, {"id": "Gold_standard", "score": 1}]
82565
train
808b6c41-af76-4039-bd66-6ee7e8a6b512
Citadelle Laferrière is also called the Fortress.
[{"id": "Citadelle_Laferrière", "score": 1}]
40928
train
c1e1f23c-e574-46dc-a4d2-31552cfaeeff
The Big Trail stars an actor.
[{"id": "The_Big_Trail", "score": 1}]
138727
train
c33f68da-c02d-47fe-8caf-4b8511c625c4
Rope was not released in 1948.
[{"id": "Rope_(film)", "score": 1}]
35233
train
ae3db644-f460-46c2-9bcf-be496e2c1ca8
Moses Malone was unable to earn any All-Star selections during his first season with the Atlanta Hawks.
[{"id": "Moses_Malone", "score": 1}]
210465
train
121b1bc5-e567-4912-ba46-c2c1aa7445f0
Jeff Bezos is not the manufacturer of Blue Origin.
[{"id": "Jeff_Bezos", "score": 1}]
101730
train
aa6f4018-f0da-46f5-b340-01945f238f31
Ronaldo Maczinski's birthday is August 11.
[{"id": "Ronaldo_Maczinski", "score": 1}]
210058
train
c4c9ccdf-6f9b-4a02-b9a7-9bcfc0a728d3
Honey to the B was the first album of an English singer, dancer, and actress.
[{"id": "Honey_to_the_B", "score": 1}, {"id": "Billie_Piper", "score": 1}]
126940
train
39ab3a08-9a45-4685-af92-0984b465e497
Home Alone is a Christmas carol.
[{"id": "Home_Alone", "score": 1}]
104969
train
8bcc6c68-a88a-49c9-83f6-b0bc6a8f1c53
Eagles reunited in the 1990s.
[{"id": "Eagles_(band)", "score": 1}]
153936
train
898135c1-3d86-4599-8019-4034faa286ea
Venus is a planet that is terrestrial.
[{"id": "Venus", "score": 1}]
7996
train
8d89bca6-c820-438a-a45e-65f435c63e45
Plato's student was a scientist.
[{"id": "Aristotle", "score": 1}, {"id": "Plato", "score": 1}]
151516
train
ab84d93a-e11c-4a0d-bfb3-e335c418f345
Spider-Man 3 features a fictional character.
[{"id": "Mary_Jane_Watson", "score": 1}, {"id": "Spider-Man_3", "score": 1}, {"id": "Gwen_Stacy", "score": 1}, {"id": "Venom_(comics)", "score": 1}, {"id": "Harry_Osborn", "score": 1}, {"id": "Sandman_(Marvel_Comics)", "score": 1}]
129549
train
d91ca520-1fdf-4fa6-8d10-d8f539d4597c
BTS is a Korean pop group.
[{"id": "BTS_(band)", "score": 1}]
168746
train
38ca20ff-4676-471e-8b2a-33d17f934d45
Leonardo Bonucci has won several honors.
[{"id": "Leonardo_Bonucci", "score": 1}]
99837
train
3bd77499-f7d7-4868-9d28-13f6a383e7f2
Geraldine Chaplin portrayed Tonya in Dr. Dolittle.
[{"id": "Geraldine_Chaplin", "score": 1}]
163862
train
a59f3150-a9eb-4caf-bf3c-9c245cafc4d8
The 65th US Secretary of State was Colin Powell.
[{"id": "Colin_Powell", "score": 1}]
171006
train
efdc071d-234f-42dd-bd07-1aeda7da5ad0
Mel Gibson was born in a month.
[{"id": "Mel_Gibson", "score": 1}]
182983
train
efaaf982-e1dd-4b68-b9b4-cda68a594b1a
There is a video game called Splatoon 2.
[{"id": "Splatoon_2", "score": 1}]
69343
train
2c98f471-5f18-440b-bf6e-832a4a88ecf3
Chandni is a film.
[{"id": "Chandni", "score": 1}]
214368
train
c2612214-5819-43ce-9ff4-acbc315cfe64
The Lion is a movie.
[{"id": "Lion", "score": 1}]
54040
train
ff253396-d4cf-437b-a945-eb60d1dd4ef3
Dan Aykroyd acted in the 1989 film Driving Miss Daisy.
[{"id": "Dan_Aykroyd", "score": 1}]
189242
train
10315958-f0ec-4b3a-a209-8c2eec34a6cc
David Angell was born on April 10, 1946.
[{"id": "David_Angell", "score": 1}]
88454
train
d30f2ca8-6d0e-428f-ab55-bbd7912128fc
David Tennant was in Doctor Who.
[{"id": "David_Tennant", "score": 1}]
99483
train
3d6a5b34-e5f1-4a5d-a490-33e96a9ecb47
Matt Sorum's touring project Kings of Chaos features members of Slipknot.
[{"id": "Matt_Sorum", "score": 1}]
201617
train
4baf929a-5591-48c6-914f-a36412b90850
Michelangelo was a person.
[{"id": "Michelangelo", "score": 1}]
124974
train
d2b724e1-511e-49f1-bccc-ab9ddfd2c249
Ron Perlman birth date is April 13, 1950.
[{"id": "Ron_Perlman", "score": 1}]
42359
train
9be1d7ac-33b4-4a4c-8bea-b6c291b0ff3c
John Lennon died before 1987.
[{"id": "John_Lennon", "score": 1}]
6643
train
c11011e7-ca16-4f96-be5d-7a7b2f10b4c2
Evil is commonly associated with hatred.
[{"id": "Evil", "score": 1}]
33575
train
5dcf4861-ef8c-4bbc-a747-d46849043507
The Dayton Agreement has other names.
[{"id": "Dayton_Agreement", "score": 1}]
217061
train
d5cc1f51-99b4-473f-8a1e-58d4ce8d07b5
Tré Cool is American.
[{"id": "Tré_Cool", "score": 1}]
194974
train
dbfb9127-6326-4055-87b1-19e21457c826
Kiefer Sutherland received a nomination.
[{"id": "Kiefer_Sutherland", "score": 1}]
229006
train
28132104-bcfb-4629-92d2-346b4ea5418f
Modern liberalism in the United States strongly supports spending on education.
[{"id": "Modern_liberalism_in_the_United_States", "score": 1}]
71027
train
2bdd039c-de07-4721-b0dc-0882417ef1b9
Liverpool F.C. is a team in the MLB.
[{"id": "Liverpool_F.C.", "score": 1}]
42818
train
b8d573cd-7064-4152-9c84-2d7f5794cf08
Abraham Lincoln spoke out against the expansion of slavery in 1858.
[{"id": "Abraham_Lincoln", "score": 1}]
227650
train
0b1184be-3dab-4ad3-868f-90aacbcdf285
Anne Frank's birthplace was an urban area in Germany.
[{"id": "Frankfurt", "score": 1}, {"id": "Anne_Frank", "score": 1}]
38982
train
aeca0d45-ca16-43ec-98cb-f5dab39c427f
Dr. Dre was previously an artist on dead bodies.
[{"id": "Dr._Dre", "score": 1}]
40648
train
31e5fd38-f825-4d76-9ed4-7a72ef894a5e
Francois de Belleforest translated works.
[{"id": "François_de_Belleforest", "score": 1}]
192534
train
388e58fd-e975-47e4-a8cd-b33edf24ed54
Marilyn Monroe worked with the First Motion Picture Unit.
[{"id": "Marilyn_Monroe", "score": 1}]
40276
train
098f5d01-db1d-4eb2-b058-e99346850254
Sunrise created anime.
[{"id": "Sunrise_(company)", "score": 1}]
176382
train
0ef3c92d-6785-434e-b60a-d65b196460d7
Wikipedia doesn't have any articles.
[{"id": "Wikipedia", "score": 1}]
16966
train
03efc9ed-5cc3-4b20-8e9c-819a8526454d
The Expendables 3 was Ronda Rousey's first feature film.
[{"id": "Ronda_Rousey", "score": 1}]
55137
train
057fade9-66a9-4463-a3f4-261c021f164a
Thiokol was also known as anything except Cordant Technologies Inc.
[{"id": "Thiokol", "score": 1}]
188119
train
9566e1c3-da17-4313-a996-966b4b54915e
Three albums have been released by Radiohead.
[{"id": "Radiohead", "score": 1}]
123988
train
9a05634c-1e20-4a51-9af9-1c49ae75aaff
Brandon Lee died during filming due to an accident involving a prop gun.
[{"id": "Brandon_Lee", "score": 1}]
169690
train
eb49b60f-6c20-4979-b820-c35ecb8fa68f
Robert Browning was born.
[{"id": "Robert_Browning", "score": 1}, {"id": "Browning_Society", "score": 1}]
219242
train
bf838ab0-aea3-47cb-b1b8-706d000f573c
Daniel Craig attended a school.
[{"id": "Daniel_Craig", "score": 1}]
119626
train
5dd8bd9e-e9d2-4ab1-918e-001662e6d8cc
Denmark's most populous city is Copenhagen with two million citizens.
[{"id": "Copenhagen", "score": 1}, {"id": "Denmark", "score": 1}]
101561
train
531e8345-2eff-41d1-8b68-aabe6101bc42
Telangana borders another territory.
[{"id": "Telangana", "score": 1}]
151598
train
3aecea1a-4a20-4674-8bd1-a2cc046b04cb
As the World Turns is set in Oakdale.
[{"id": "As_the_World_Turns", "score": 1}]
36001
train
73e8157e-5721-410b-b1dc-656c745caad5
The Caracazo took place in Morocco.
[{"id": "Caracazo", "score": 1}]
217247
train
93133756-3113-433f-861c-b209788492c0
Richard Ramirez committed murder.
[{"id": "Richard_Ramirez", "score": 1}]
128058
train
e97a6313-b923-4797-b080-15f337f01606
Maria Sharapova held the title of the world No. 1 zero times.
[{"id": "Maria_Sharapova", "score": 1}]
111065
train
cdac00f7-25e8-4e82-b69b-dfe2e9ce69ef
Hotel Transylvania 2 featured Adam Sandler as a voice actor.
[{"id": "Hotel_Transylvania_2", "score": 1}]
98401
train
d8515320-6374-4280-8beb-cd7fbd8723c5
Rajesh Khanna earned the most money as an actor from 1970 to 1987.
[{"id": "Rajesh_Khanna", "score": 1}]
43016
train
01121458-f474-4ebc-bdd3-303efa106d91
The Host stars an actor who's birthday is on November 18.
[{"id": "Jake_Abel", "score": 1}, {"id": "The_Host_(2013_film)", "score": 1}]
215363
train
75e352c0-92ac-4af6-a793-a1485c32fb06
Anne Hathaway lost the Critics' Choice Movie Award for Best Actress whenever she was eligible.
[{"id": "Anne_Hathaway", "score": 1}]
60864
train
fb3f85e8-38c5-442d-90cf-b9c8ccc3f367
Silicon Valley refers to the southern part of the San Francisco Bay Area in California.
[{"id": "Silicon_Valley", "score": 1}]
103356
train
00a99242-4358-4c6e-859d-f5d85312b019
Rihanna is the oldest and slowest solo artist to earn fourteen number-one singles on the Billboard Hot 100.
[{"id": "Rihanna", "score": 1}]
3719
train
946b8a1d-c191-4638-86a3-03731fbec941
Woody Allen is a computer program.
[{"id": "Woody_Allen", "score": 1}]
193111
train
fac8171a-4908-404d-9e45-be7b0f26cfbc
Jet Li was the main actor in a martial arts movie.
[{"id": "Unleashed_(film)", "score": 1}, {"id": "Shaolin_Temple_(1982_film)", "score": 1}, {"id": "Fist_of_Legend", "score": 1}, {"id": "Jet_Li", "score": 1}]
35285
train
8ce57629-becb-4764-b0b2-cc7672d45e03
David Beckham played in the Premier League.
[{"id": "David_Beckham", "score": 1}]
59451
train
3fd2d18f-45b4-40d5-8e01-60604c1df2a9
Veronica Mars is a protagonist.
[{"id": "Veronica_Mars_(character)", "score": 1}]
125570
train
f32f0915-4fd1-4c13-a392-5d958516fb67
A Game of Thrones is a Netflix series.
[{"id": "A_Game_of_Thrones", "score": 1}]
78945
train
269b0a0d-aeb3-4843-96ef-7e96305f942e
Yelp existed in 2009.
[{"id": "Yelp", "score": 1}]
179392
train
5dfaa848-e203-4603-8c6c-71fde94a157c
FC Barcelona is in third position on the UEFA club rankings.
[{"id": "FC_Barcelona", "score": 1}]
41036
train
94cba2a4-d77a-4aa8-9bd4-588cad3647f0
Grand Theft Auto V was developed by Rockstar Games.
[{"id": "Grand_Theft_Auto_V", "score": 1}]
52116
train
9964874c-0623-4043-8f9d-d379e802f362
John Denver is a musician.
[{"id": "John_Denver", "score": 1}]
28994
train
a39a6951-880c-4a58-b165-c27ea51781fe
Brave is a television series.
[{"id": "Brave_(2012_film)", "score": 1}]
209145
train
551e060d-4f77-433f-a749-11b6908831b6
K2 is also known as the Savage Mountain.
[{"id": "K2", "score": 1}]
23903
train
25afe5f2-7624-49a3-b6ec-f2d146bfea88
Overwatch lets players gain cosmetic awards which include character skins and victory poses.
[{"id": "Overwatch_(video_game)", "score": 1}]
147205
train
6d4c84cd-f490-40b5-9198-50ce820f8828
Diego Costa fails to play as a striker.
[{"id": "Diego_Costa", "score": 1}]
144290
train
0730335a-8313-4dd4-a13b-9785e270ca74
Brandon Lee died in a car crash in 1998.
[{"id": "Brandon_Lee", "score": 1}]
169703
train
117693d3-30e4-4772-9c6a-cbc3e54f725d
In the 1990's, Steve Buscemi was in movies.
[{"id": "Desperado_(film)", "score": 1}, {"id": "Con_Air", "score": 1}, {"id": "Steve_Buscemi", "score": 1}, {"id": "Reservoir_Dogs", "score": 1}, {"id": "Armageddon_(1998_film)", "score": 1}]
52557
train
dbbcc86b-76c3-46d9-9c48-97f439676530
The 1983-94 season ended.
[{"id": "1983–84_NBA_season", "score": 1}]
101318
train
e0a83cc8-0e5e-4407-bfa5-06485a47fac8
Rio de Janeiro is in a rural area.
[{"id": "Rio_de_Janeiro", "score": 1}]
52722
train
340cf48e-ed99-440c-a741-4c68e6911c89
CeeLo Green has worked with another artist.
[{"id": "CeeLo_Green", "score": 1}, {"id": "Gnarls_Barkley", "score": 1}]
128871
train
dfaa2b51-0c1f-4ef5-b936-b74fb8806140
Albany, New York, is the longest continuously chartered city in North America.
[{"id": "Albany,_New_York", "score": 1}]
212173
train
ea814ea3-089b-4386-a4b7-769cc2eae7da
The 21st was the day when Sophie Turner was born.
[{"id": "Sophie_Turner", "score": 1}]
78136
train
6d4f148a-bbc4-43ad-bca1-4ec946bda12e
Frenemies was aired after 2010.
[{"id": "Frenemies_(film)", "score": 1}]
177565
train
83deee6b-25ba-4f5f-811c-5bc15a194a9f
The United States Naval Academy is in Spain.
[{"id": "Fort_Severn", "score": 1}, {"id": "United_States_Naval_Academy", "score": 1}]
101435
train
52bb2e36-ef2e-46b0-8623-d7159e13aa91
Tennis is the kind of sport that uses racquets.
[{"id": "Tennis", "score": 1}]
106131
train
93d697fd-492e-426d-96e3-0118bef12551
The Woman in Red is a romantic tragedy.
[{"id": "The_Woman_in_Red_(1984_film)", "score": 1}]
202908
train
End of preview. Expand in Data Studio

FEVER (BEIR) — Fact-checking retrieval

Dataset description

FEVER (Fact Extraction and VERification) is a large-scale English dataset for claim verification against textual sources. Claims were produced by altering sentences drawn from Wikipedia; annotators then labeled each claim without knowing which source sentence it came from. Labels are Supported, Refuted, or NotEnoughInfo (with substantial inter-annotator agreement). For Supported and Refuted claims, annotators also identified the sentence-level evidence needed to justify the label.

BEIR (Benchmarking IR) repackaged FEVER—along with many other public corpora—as a standard retrieval benchmark for zero-shot evaluation of dense, sparse, and hybrid information retrieval models across heterogeneous tasks. In the BEIR formulation, each claim acts as a query, and the objective is to retrieve relevant Wikipedia documents (by title) that contain the evidence required for verification. This setting isolates retrieval quality as the variable of interest when paired with a fixed downstream verifier or when reporting standard IR metrics.

This repository (orgrctera/beir_fever) provides train / validation / test splits in Parquet form for retrieval evaluation pipelines. Each row is one query (a claim) with relevance judgments pointing at corpus document identifiers in the BEIR FEVER benchmark (Wikipedia article titles as used upstream).

Scale and domain (BEIR FEVER)

The original FEVER release comprises on the order of 185k verified claims; BEIR’s FEVER split follows the standard BEIR packaging. The corpus is Wikipedia-oriented text keyed by article identifiers (titles in the BEIR release). Exact counts for this Hub snapshot follow the upstream BEIR FEVER release—see the BEIR repository for version-precise figures.

Task: retrieval (FEVER in BEIR)

The task is ad hoc document retrieval for fact-checking:

  1. Input: a natural-language claim (the query).
  2. Output: a ranked list of document IDs from the FEVER corpus (Wikipedia titles in the BEIR distribution), or scores over the full collection, such that relevant IDs—according to the official qrels—receive high rank.

Downstream metrics are standard IR metrics (e.g., nDCG@k, Recall@k, MRR), as in BEIR’s evaluation scripts or frameworks such as Pyserini / the BEIR API.

Note: Full retrieval evaluation also requires the corpus (passage or document text keyed by the same IDs). This card describes the query + qrels side as prepared for CTERA-style evaluation rows; align corpus IDs with the same BEIR FEVER corpus you use for indexing.

Data format (this repository)

Each record includes:

Field Description
id UUID for this example row.
input The claim text (query).
expected_output JSON string: list of objects {"id": "<corpus-doc-id>", "score": <relevance>}. Document IDs are Wikipedia article titles as in the BEIR FEVER corpus; scores follow the BEIR qrels convention (typically 1 for relevant in binary settings).
metadata.query_id Original BEIR / FEVER query identifier (string).
metadata.split Split name: train, dev, or test.

Example 1

{
  "id": "7e965799-c99c-46a6-95ab-91dae89ecd4f",
  "input": "Robert Duvall has not won a BAFTA.",
  "expected_output": "[{\"id\": \"Robert_Duvall\", \"score\": 1}]",
  "metadata.query_id": "145027",
  "metadata.split": "train"
}

Example 2

{
  "id": "a2b4302d-b0cb-4917-ab1a-67622b3c9790",
  "input": "Reese Witherspoon grew up in the United States.",
  "expected_output": "[{\"id\": \"Tennessee\", \"score\": 1}, {\"id\": \"New_Orleans\", \"score\": 1}, {\"id\": \"Reese_Witherspoon\", \"score\": 1}]",
  "metadata.query_id": "160148",
  "metadata.split": "train"
}

References

FEVER (original dataset)

James Thorne, Andreas Vlachos, Christos Christodoulopoulos, Arpit Mittal
FEVER: a large-scale dataset for Fact Extraction and VERification
Presented at NAACL 2018; extended version on arXiv.

Abstract (from arXiv): “In this paper we introduce a new publicly available dataset for verification against textual sources, FEVER: Fact Extraction and VERification. It consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo by annotators achieving 0.6841 in Fleiss κ. For the first two classes, the annotators also recorded the sentence(s) forming the necessary evidence for their judgment. To characterize the challenge of the dataset presented, we develop a pipeline approach and compare it to suitably designed oracles. The best accuracy we achieve on labeling a claim accompanied by the correct evidence is 31.87%, while if we ignore the evidence we achieve 50.91%. Thus we believe that FEVER is a challenging testbed that will help stimulate progress on claim verification against textual sources.”

BEIR benchmark (FEVER as a subset)

Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, Iryna Gurevych
BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models
NeurIPS 2021 (Datasets and Benchmarks Track).

Abstract (from arXiv): “Existing neural information retrieval (IR) models have often been studied in homogeneous and narrow settings, which has considerably limited insights into their out-of-distribution (OOD) generalization capabilities. To address this, and to facilitate researchers to broadly evaluate the effectiveness of their models, we introduce Benchmarking-IR (BEIR), a robust and heterogeneous evaluation benchmark for information retrieval. We leverage a careful selection of 18 publicly available datasets from diverse text retrieval tasks and domains and evaluate 10 state-of-the-art retrieval systems including lexical, sparse, dense, late-interaction and re-ranking architectures on the BEIR benchmark. Our results show BM25 is a robust baseline and re-ranking and late-interaction-based models on average achieve the best zero-shot performances, however, at high computational costs. In contrast, dense and sparse-retrieval models are computationally more efficient but often underperform other approaches, highlighting the considerable room for improvement in their generalization capabilities.”

Related resources

  • BEIR FEVER mirrors on Hugging Face in classic BEIR layout, e.g. BeIR/fever (corpus / queries / qrels).
  • IRDS packaging: irds/beir_fever exposes FEVER via the ir-datasets tooling.

Citation

If you use FEVER, cite the FEVER paper (Thorne et al., NAACL 2018 / arXiv:1803.05355). If you use the BEIR benchmark formulation, cite the BEIR paper (Thakur et al., NeurIPS 2021). BibTeX for BEIR is available in the official repository.

License

FEVER’s evidence and corpus trace to Wikipedia-derived text; Wikipedia content is typically licensed under Creative Commons Attribution-ShareAlike (version depends on the snapshot). This card marks cc-by-sa-4.0 as a common umbrella for Wikipedia-derived redistribution; verify against your corpus snapshot and upstream terms if compliance is strict.


Dataset card maintained for the orgrctera/beir_fever Hub repository.

Downloads last month
36

Papers for orgrctera/beir_fever