--- license: cc-by-sa-4.0 language: - en size_categories: - 10M<n<100M --- # Dataset Summary Paragraph embeddings for every article in English Wikipedia (not the Simple English version). Based on [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia), 20231101.en. Embeddings were generated with [avsolatorio/GIST-small-Embedding-v0](https://huggingface.co/avsolatorio/GIST-small-Embedding-v0) and are quantized to int8. You can load the data with the following: ``` from datasets import load_dataset ds = load_dataset(path="Abrak/wikipedia-paragraph-embeddings-en-gist-complete", data-dir="20231101.en") ``` # Dataset Structure The structure of the dataset is designed to minimize necessary storage and calculations but still cover the breadth of Wikipedia. ## Data instances An example looks as follows: ``` { 'id': '12.1', 'embedding': [[10, -14, -42, -3, 5, 4, 7, 17, -8, 18, ...] } ``` ## Data Fields The data fields are the same for all records: * `id(str)`: The ID of the same article in wikimedia/wikipedia, '.' as a separator, and the sequential number of the paragraphs in the article. These are not left-padded. * `embedding`: A list of 384 int8 values (from -128 to 127) # Details ## Source Data The data is sourced directly from the wikimedia/wikipedia dataset, in the 20231101.en directory. This is English-language article text content, taken from a snapshot on November 1, 2023. The source data was already stripped of formatting and other content that is not language. See the [wikimedia/wikipedia model card](https://huggingface.co/datasets/wikimedia/wikipedia) for more information. As part of this dataset's processing, article text was split into paragraphs on two newlines (`\n\n`). ## Embedding Calculation Embeddings were calculated in batches of 1300 paragraphs with sentence_transformers and the unquantized GIST-small-Embedding-v0 model. Precision was set to int8. Complete processing took about 20 hours on an Nvidia A40. The full calculation code used is in [commit 5132104f1fa59d9b212844f6f7a93232193958f2 of setup.py](https://github.com/abrakjamson/The-Archive/commit/5132104f1fa59d9b212844f6f7a93232193958f2) in the Github repo for my project, [The Archive](https://github.com/abrakjamson/The-Archive). # Licensing information These embeddings are a derivative of Wikipedia article text, which is under [CC-BY-SA-4.0](https://en.wikipedia.org/wiki/Wikipedia:Text_of_the_Creative_Commons_Attribution-ShareAlike_4.0_International_License), a copyleft license, as well as [GFDL](https://en.wikipedia.org/wiki/Wikipedia:Text_of_the_GNU_Free_Documentation_License). These embeddings inherit the same licenses. See the [Wikipedia Copyrights page](https://en.wikipedia.org/wiki/Wikipedia:Copyrights) for details.