Datasets:
The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.
Dataset Card for edu_fineweb10B_sharded_50shards
This dataset card aims to describe the edu_fineweb10B_sharded_50shards dataset, a large-scale pre-tokenized and sharded dataset created from the eduFineWeb corpus. It has been prepared for use in training transformer-based language models using NumPy arrays for efficient loading.
Dataset Details
Dataset Description
edu_fineweb10B_sharded_50shards is a tokenized dataset based on the eduFineWeb 10B corpus, designed for scalable training of language models. The dataset contains a total of 10 billion tokens split across 50 shards—49 for training and 1 for evaluation.
Each shard is stored in .npy format and contains 200 million token IDs, pre-tokenized using the tiktoken tokenizer (compatible with GPT-2-style models). The dataset is designed for high-efficiency training pipelines, particularly in distributed or multi-GPU setups.
- Curated by: Private (individual researcher)
- Funded by [optional]: Not funded
- Shared by [optional]: Abhinav
- Language(s) (NLP): English
- License: MIT
Dataset Sources
- Repository: [Private / Local Project]
- Paper [optional]: N/A
- Demo [optional]: N/A
Uses
Direct Use
The dataset is suitable for:
- Pretraining large autoregressive language models (e.g., GPT-style)
- Finetuning models for general-purpose language generation
- Research in scaling laws, memory-efficient training, and model evaluation
Out-of-Scope Use
- Not suited for supervised learning tasks without additional labeling
- Not appropriate for real-time applications without additional safety filtering
- Not recommended for use in sensitive, safety-critical environments without thorough auditing
Dataset Structure
- Format:
.npyfiles - Tokenizer used:
tiktoken.get_encoding("gpt2") - Tokens per shard: 200 million
- Total shards: 50
- Shard naming format:
- Training:
edu_fineweb_train_000000.npytoedu_fineweb_train_000048.npy - Evaluation:
edu_fineweb_val_000000.npy
- Training:
- File size:
400MB per shard (20GB total)
Each file contains a flat NumPy array of token IDs (int32), which can be converted to PyTorch tensors for training.
Example Usage
import numpy as np
import torch
# Load a shard file
filename = "edu_fineweb_train_000001.npy"
npt = np.load(filename)
npt = npt.astype(np.int32)
ptt = torch.tensor(npt, dtype=torch.long)
print(ptt.shape) # Should be (200_000_000,)
print(ptt[:10]) # View first 10 token IDs
- Downloads last month
- 140