Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -1,11 +1,16 @@
|
|
1 |
---
|
2 |
-
license: mit
|
3 |
-
task_categories:
|
4 |
-
- question-answering
|
5 |
language:
|
6 |
- en
|
|
|
7 |
size_categories:
|
8 |
- 1M<n<10M
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
configs:
|
10 |
- config_name: database
|
11 |
data_files:
|
@@ -196,147 +201,55 @@ dataset_info:
|
|
196 |
|
197 |
# Dataset Card for PhantomWiki
|
198 |
|
199 |
-
|
200 |
-
|
201 |
-
PhantomWiki is framework for evaluating LLMs, specifically RAG and agentic workflows, that is resistant to memorization.
|
202 |
-
Unlike prior work, it is neither a fixed dataset, nor is it based on any existing data.
|
203 |
-
Instead, PhantomWiki generates unique dataset instances, comprised of factually consistent document corpora with diverse question-answer pairs, on demand.
|
204 |
|
205 |
## Dataset Details
|
206 |
|
207 |
### Dataset Description
|
208 |
|
209 |
-
PhantomWiki generates a fictional universe
|
210 |
-
We reflect these facts in a large-scale corpus, mimicking the style of fan-wiki websites.
|
211 |
-
Then we generate question-answer pairs with tunable difficulties, encapsulating the types of multi-hop questions commonly considered in the question-answering (QA) literature.
|
212 |
|
213 |
- **Created by:** Albert Gong, Kamilė Stankevičiūtė, Chao Wan, Anmol Kabra, Raphael Thesmar, Johann Lee, Julius Klenke, Carla P. Gomes, Kilian Q. Weinberger
|
214 |
-
- **Funded by:**
|
215 |
-
- **Shared by \[optional\]:** \[More Information Needed\]
|
216 |
-
- **Language(s) (NLP):** English
|
217 |
- **License:** MIT License
|
218 |
-
|
219 |
-
### Dataset Sources \[optional\]
|
220 |
-
|
221 |
-
<!-- Provide the basic links for the dataset. -->
|
222 |
-
|
223 |
-
- **Repository:** https://github.com/kilian-group/phantom-wiki
|
224 |
- **Paper:** [PhantomWiki: On-Demand Datasets for Reasoning and Retrieval Evaluation](https://huggingface.co/papers/2502.20377)
|
225 |
-
- **
|
226 |
|
227 |
-
## Uses
|
228 |
|
229 |
-
|
230 |
-
PhantomWiki enables quantitative evaluation of the reasoning and retrieval capabilities of LLMs. See our full paper for analysis of frontier LLMs, including GPT-4o, Gemini-1.5-Flash, [Llama-3.3-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct) and [DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B).
|
231 |
|
232 |
-
|
233 |
|
234 |
-
PhantomWiki is intended to evaluate retrieval augmented generation (RAG) systems and agentic workflows.
|
235 |
|
236 |
-
|
237 |
|
238 |
-
|
239 |
|
240 |
-
|
241 |
|
242 |
-
PhantomWiki
|
243 |
|
244 |
1. `question-answer`: Question-answer pairs generated using a context-free grammar
|
245 |
2. `text-corpus`: Documents generated using natural-language templates
|
246 |
3. `database`: Prolog database containing the facts and clauses representing the universe
|
247 |
|
248 |
-
Each universe is saved as a **split**.
|
249 |
-
|
250 |
-
## Dataset Creation
|
251 |
-
|
252 |
-
### Curation Rationale
|
253 |
-
|
254 |
-
Most mathematical and logical reasoning datasets do not explicity evaluate retrieval capabilities and
|
255 |
-
few retrieval datasets incorporate complex reasoning, save for a few exceptions (e.g., [BRIGHT](https://huggingface.co/datasets/xlangai/BRIGHT), [MultiHop-RAG](https://huggingface.co/datasets/yixuantt/MultiHopRAG)).
|
256 |
-
However, virtually all retrieval datasets are derived from Wikipedia or internet articles, which are contained in LLM training data.
|
257 |
-
We take the first steps toward a large-scale synthetic dataset that can evaluate LLMs' reasoning and retrieval capabilities.
|
258 |
-
|
259 |
-
### Source Data
|
260 |
-
|
261 |
-
This is a synthetic dataset. The extent to which we use real data is detailed as follows:
|
262 |
-
|
263 |
-
1. We sample surnames from among the most common surnames in the US population according to https://names.mongabay.com/most_common_surnames.htm
|
264 |
-
2. We sample first names using the `names` Python package (https://github.com/treyhunner/names). We thank the contributors for making this tool publicly available.
|
265 |
-
3. We sample jobs from the list of real-life jobs from the `faker` Python package. We thank the contributors for making this tool publicly available.
|
266 |
-
4. We sample hobbies from the list of real-life hobbies at https://www.kaggle.com/datasets/mrhell/list-of-hobbies. We are grateful to the original author for curating this list and making it publicly available.
|
267 |
-
|
268 |
-
#### Data Collection and Processing
|
269 |
-
|
270 |
-
This dataset was generated on commodity CPUs using Python and SWI-Prolog. See paper for full details of the generation pipeline, including timings.
|
271 |
-
|
272 |
-
#### Who are the source data producers?
|
273 |
-
|
274 |
-
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
|
275 |
-
|
276 |
-
N/A
|
277 |
-
|
278 |
-
### Annotations \[optional\]
|
279 |
-
|
280 |
-
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
|
281 |
|
282 |
-
|
283 |
-
|
284 |
-
#### Annotation process
|
285 |
-
|
286 |
-
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
|
287 |
-
|
288 |
-
N/A
|
289 |
-
|
290 |
-
#### Who are the annotators?
|
291 |
-
|
292 |
-
<!-- This section describes the people or systems who created the annotations. -->
|
293 |
-
|
294 |
-
N/A
|
295 |
-
|
296 |
-
#### Personal and Sensitive Information
|
297 |
-
|
298 |
-
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
|
299 |
-
|
300 |
-
PhantomWiki does not reference any personal or private data.
|
301 |
|
302 |
## Bias, Risks, and Limitations
|
303 |
|
304 |
-
|
305 |
-
|
306 |
-
PhantomWiki generates large-scale corpora, reflecting fictional universes of characters and mimicking the style of fan-wiki websites. While sufficient for evaluating complex reasoning and retrieval capabilities of LLMs, PhantomWiki is limited to simplified family relations and attributes. Extending the complexity of PhantomWiki to the full scope of Wikipedia is an exciting future direction.
|
307 |
-
|
308 |
-
### Recommendations
|
309 |
-
|
310 |
-
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
311 |
-
|
312 |
-
PhantomWiki should be used as a benchmark to inform how LLMs should be used on reasoning- and retrieval-based tasks. For holistic evaluation on diverse tasks, PhantomWiki should be combined with other benchmarks.
|
313 |
-
|
314 |
-
## Citation \[optional\]
|
315 |
-
|
316 |
-
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
|
317 |
-
|
318 |
-
**BibTeX:**
|
319 |
-
|
320 |
-
\[More Information Needed\]
|
321 |
-
|
322 |
-
**APA:**
|
323 |
-
|
324 |
-
\[More Information Needed\]
|
325 |
-
|
326 |
-
## Glossary \[optional\]
|
327 |
-
|
328 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
|
329 |
-
|
330 |
-
\[More Information Needed\]
|
331 |
-
|
332 |
-
## More Information \[optional\]
|
333 |
-
|
334 |
-
\[More Information Needed\]
|
335 |
-
|
336 |
-
## Dataset Card Authors \[optional\]
|
337 |
|
338 |
-
\[More Information Needed\]
|
339 |
|
340 |
-
##
|
341 |
|
342 | |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
|
|
|
|
|
|
2 |
language:
|
3 |
- en
|
4 |
+
license: mit
|
5 |
size_categories:
|
6 |
- 1M<n<10M
|
7 |
+
task_categories:
|
8 |
+
- question-answering
|
9 |
+
tags:
|
10 |
+
- reasoning
|
11 |
+
- retrieval
|
12 |
+
- synthetic
|
13 |
+
- evaluation
|
14 |
configs:
|
15 |
- config_name: database
|
16 |
data_files:
|
|
|
201 |
|
202 |
# Dataset Card for PhantomWiki
|
203 |
|
204 |
+
This repository contains pre-generated instances of the PhantomWiki dataset, created using the `phantom-wiki` Python package. PhantomWiki is a framework for evaluating LLMs, particularly RAG and agentic workflows, designed to be resistant to memorization. Unlike fixed datasets, PhantomWiki generates unique instances on demand, ensuring novelty and preventing data leakage.
|
|
|
|
|
|
|
|
|
205 |
|
206 |
## Dataset Details
|
207 |
|
208 |
### Dataset Description
|
209 |
|
210 |
+
PhantomWiki generates a synthetic fictional universe populated with characters and facts, mirroring the structure of a fan wiki. These facts are reflected in a large-scale corpus, and diverse question-answer pairs of varying difficulties are then generated.
|
|
|
|
|
211 |
|
212 |
- **Created by:** Albert Gong, Kamilė Stankevičiūtė, Chao Wan, Anmol Kabra, Raphael Thesmar, Johann Lee, Julius Klenke, Carla P. Gomes, Kilian Q. Weinberger
|
213 |
+
- **Funded by:** See the [paper](https://huggingface.co/papers/2502.20377) for details.
|
|
|
|
|
214 |
- **License:** MIT License
|
|
|
|
|
|
|
|
|
|
|
|
|
215 |
- **Paper:** [PhantomWiki: On-Demand Datasets for Reasoning and Retrieval Evaluation](https://huggingface.co/papers/2502.20377)
|
216 |
+
- **Code:** [https://github.com/kilian-group/phantom-wiki](https://github.com/kilian-group/phantom-wiki)
|
217 |
|
|
|
218 |
|
219 |
+
### Dataset Sources
|
|
|
220 |
|
221 |
+
- **Repository (Code):** [https://github.com/kilian-group/phantom-wiki](https://github.com/kilian-group/phantom-wiki)
|
222 |
|
|
|
223 |
|
224 |
+
## Uses
|
225 |
|
226 |
+
PhantomWiki is designed to evaluate retrieval augmented generation (RAG) systems and agentic workflows. To avoid data leakage and overfitting, generate a new, unique PhantomWiki instance for each evaluation. Our paper details an analysis of frontier LLMs using PhantomWiki.
|
227 |
|
228 |
+
### Dataset Structure
|
229 |
|
230 |
+
PhantomWiki provides three configurations:
|
231 |
|
232 |
1. `question-answer`: Question-answer pairs generated using a context-free grammar
|
233 |
2. `text-corpus`: Documents generated using natural-language templates
|
234 |
3. `database`: Prolog database containing the facts and clauses representing the universe
|
235 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
236 |
|
237 |
+
Each universe is saved as a separate split. See the original repository for details on generation and usage.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
238 |
|
239 |
## Bias, Risks, and Limitations
|
240 |
|
241 |
+
PhantomWiki, while effective at evaluating complex reasoning and retrieval, is limited in its representation of family relations and attributes. Extending its complexity is a future research direction. For a holistic evaluation, combine PhantomWiki with other benchmarks.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
242 |
|
|
|
243 |
|
244 |
+
## Citation
|
245 |
|
246 |
+
```bibtex
|
247 |
+
@article{2025_phantomwiki,
|
248 |
+
title={{PhantomWiki: On-Demand Datasets for Reasoning and Retrieval Evaluation}},
|
249 |
+
author={Albert Gong and Kamilė Stankevičiūtė and Chao Wan and Anmol Kabra and Raphael Thesmar and Johann Lee and Julius Klenke and Carla P. Gomes and Kilian Q. Weinberger},
|
250 |
+
year={2025},
|
251 |
+
journal={todo},
|
252 |
+
url={todo},
|
253 |
+
note={Under Review},
|
254 |
+
}
|
255 |
+
```
|