---
license: cc-by-sa-4.0
language:
- en
size_categories:
- 10M<n<100M
---

# Dataset Summary
Paragraph embeddings for every article in English Wikipedia (not the Simple English version). 
Based on [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia), 20231101.en.

Embeddings were generated with [avsolatorio/GIST-small-Embedding-v0](https://huggingface.co/avsolatorio/GIST-small-Embedding-v0)
and are quantized to int8.

You can load the data with the following:

```
from datasets import load_dataset

ds = load_dataset(path="Abrak/wikipedia-paragraph-embeddings-en-gist-complete", data-dir="20231101.en")
```

# Dataset Structure

The structure of the dataset is designed to minimize necessary storage and calculations 
but still cover the breadth of Wikipedia.

## Data instances

An example looks as follows:
```
{ 'id': '12.1',
  'embedding': [[10, -14, -42, -3, 5, 4, 7, 17, -8, 18, ...]
}
```
## Data Fields
The data fields are the same for all records:
* `id(str)`: The ID of the same article in wikimedia/wikipedia, '.' as a separator, and the sequential number
of the paragraphs in the article. These are not left-padded.
* `embedding`: A list of 384 int8 values (from -128 to 127)

# Details
## Source Data
The data is sourced directly from the wikimedia/wikipedia dataset, in the 20231101.en directory.
This is English-language article text content, taken from a snapshot on November 1, 2023.
The source data was already stripped of formatting and other content that is not language. 
See the [wikimedia/wikipedia model card](https://huggingface.co/datasets/wikimedia/wikipedia) for more information.

As part of this dataset's processing, article text was split into paragraphs on two newlines (`\n\n`).

## Embedding Calculation
Embeddings were calculated in batches of 1300 paragraphs with sentence_transformers and the unquantized 
GIST-small-Embedding-v0 model. Precision was set to int8. Complete processing took about 20 hours on an
Nvidia A40. The full calculation code used is in 
[commit 5132104f1fa59d9b212844f6f7a93232193958f2 of setup.py](https://github.com/abrakjamson/The-Archive/commit/5132104f1fa59d9b212844f6f7a93232193958f2)
in the Github repo for my project, [The Archive](https://github.com/abrakjamson/The-Archive).

# Licensing information
These embeddings are a derivative of Wikipedia article text, which is under [CC-BY-SA-4.0](https://en.wikipedia.org/wiki/Wikipedia:Text_of_the_Creative_Commons_Attribution-ShareAlike_4.0_International_License),
a copyleft license, as well as [GFDL](https://en.wikipedia.org/wiki/Wikipedia:Text_of_the_GNU_Free_Documentation_License).
These embeddings inherit the same licenses. See the [Wikipedia Copyrights page](https://en.wikipedia.org/wiki/Wikipedia:Copyrights)
for details.