Dataset Viewer

The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.

Dataset Card for edu_fineweb10B_sharded_50shards

This dataset card aims to describe the edu_fineweb10B_sharded_50shards dataset, a large-scale pre-tokenized and sharded dataset created from the eduFineWeb corpus. It has been prepared for use in training transformer-based language models using NumPy arrays for efficient loading.

Dataset Details

Dataset Description

edu_fineweb10B_sharded_50shards is a tokenized dataset based on the eduFineWeb 10B corpus, designed for scalable training of language models. The dataset contains a total of 10 billion tokens split across 50 shards49 for training and 1 for evaluation.

Each shard is stored in .npy format and contains 200 million token IDs, pre-tokenized using the tiktoken tokenizer (compatible with GPT-2-style models). The dataset is designed for high-efficiency training pipelines, particularly in distributed or multi-GPU setups.

  • Curated by: Private (individual researcher)
  • Funded by [optional]: Not funded
  • Shared by [optional]: Abhinav
  • Language(s) (NLP): English
  • License: MIT

Dataset Sources

  • Repository: [Private / Local Project]
  • Paper [optional]: N/A
  • Demo [optional]: N/A

Uses

Direct Use

The dataset is suitable for:

  • Pretraining large autoregressive language models (e.g., GPT-style)
  • Finetuning models for general-purpose language generation
  • Research in scaling laws, memory-efficient training, and model evaluation

Out-of-Scope Use

  • Not suited for supervised learning tasks without additional labeling
  • Not appropriate for real-time applications without additional safety filtering
  • Not recommended for use in sensitive, safety-critical environments without thorough auditing

Dataset Structure

  • Format: .npy files
  • Tokenizer used: tiktoken.get_encoding("gpt2")
  • Tokens per shard: 200 million
  • Total shards: 50
  • Shard naming format:
    • Training: edu_fineweb_train_000000.npy to edu_fineweb_train_000048.npy
    • Evaluation: edu_fineweb_val_000000.npy
  • File size: 400MB per shard (20GB total)

Each file contains a flat NumPy array of token IDs (int32), which can be converted to PyTorch tensors for training.

Example Usage

import numpy as np
import torch

# Load a shard file
filename = "edu_fineweb_train_000001.npy"
npt = np.load(filename)
npt = npt.astype(np.int32)
ptt = torch.tensor(npt, dtype=torch.long)

print(ptt.shape)  # Should be (200_000_000,)
print(ptt[:10])   # View first 10 token IDs
Downloads last month
225

Models trained or fine-tuned on abhinavv3/edu_fineweb10B_sharded_50shards