Gutenberg-Clean / README.md
NNEngine's picture
Update README.md
a41180b verified
---
license: mit
task_categories:
- text-classification
- question-answering
- text-generation
language:
- en
size_categories:
- 10M<n<100M
---
# 📚 TinyWay-Gutenberg-Clean (Compressed Shards)
A large-scale, high-quality English text dataset derived from Project Gutenberg.
The corpus has been cleaned, normalized, deduplicated, segmented into fixed-length samples, and stored as compressed JSONL shards for efficient large-scale language model training.
This dataset is intended for pretraining and experimentation with small and medium language models such as **TinyWay**, tokenizer training, and large-scale NLP research.
---
## 📦 Dataset Overview
* **Name:** TinyWay-Gutenberg-Clean
* **Current Release:** ~19 compressed shards (`.jsonl.gz`)
* **Estimated Samples:** Tens of millions of text segments
* **Language:** English
* **Format:** Gzip-compressed JSON Lines (`.jsonl.gz`)
* **Source:** Project Gutenberg (public domain books)
* **License:** Public Domain
* **Maintainer:** Shivam (NNEngine / ITM AIR Lab)
Each record contains a clean text segment between **30 and 60 words**.
Future releases will scale this dataset further (e.g., 100M+ samples).
---
## Data Format
Each line is a JSON object:
```json
{
"id": "twg_000000012345",
"text": "Cleaned natural English text segment between thirty and sixty words.",
"word_count": 42,
"source": "gutenberg"
}
```
### Fields
| Field | Description |
| ------------ | ------------------------------ |
| `id` | Unique sample identifier |
| `text` | Clean English text segment |
| `word_count` | Number of words in the segment |
| `source` | Data source identifier |
---
## Data Processing Pipeline
The dataset was generated using a fully streaming pipeline to ensure scalability and low memory usage.
### Processing Steps
1. **Streaming Input**
* Text streamed from a Project Gutenberg mirror on Hugging Face.
2. **Text Cleaning**
* Removed Gutenberg headers and footers.
* Removed chapter titles, page numbers, and boilerplate text.
* Normalized whitespace and line breaks.
* Removed non-ASCII and control characters.
* Filtered malformed or extremely short segments.
3. **Segmentation**
* Text segmented into chunks of **30–60 words**.
4. **Validation**
* Enforced word count limits.
* Filtered invalid or noisy segments.
5. **Deduplication**
* Exact hash-based deduplication applied during generation.
6. **Compression & Sharding**
* Data stored as `.jsonl.gz` shards for efficient disk usage and streaming.
---
## How to Load the Dataset
### Using Hugging Face Datasets (Streaming)
```python
from datasets import load_dataset
dataset = load_dataset(
"NNEngine/TinyWay-Gutenberg-Clean",
split="train",
streaming=True
)
for i, sample in enumerate(dataset):
print(sample)
if i == 3:
break
```
---
### Reading a Shard Manually
```python
import gzip
import json
with gzip.open("train-00000.jsonl.gz", "rt", encoding="utf-8") as f:
for _ in range(3):
print(json.loads(next(f)))
```
---
## Dataset Characteristics (Approximate)
* **Average words per sample:** ~45
* **Style:** Literary and narrative English
* **Domain:** Fiction, non-fiction, historical texts
* **Vocabulary:** Large natural English vocabulary
* **Compression:** ~60–70% size reduction vs raw JSONL
Exact statistics may vary per shard and will be expanded in future releases.
---
## Limitations
* Primarily literary and historical language.
* No conversational chat data.
* No code or structured technical documentation.
* Some archaic vocabulary and sentence structures may appear.
* Deduplication is hash-based (near-duplicates may remain).
For conversational or web-style language modeling, this dataset should be mixed with complementary corpora.
---
## License
All source texts originate from Project Gutenberg and are in the **public domain**.
This processed dataset is released for unrestricted research and commercial use.
---
## Versioning & Roadmap
Planned future updates:
- Larger releases (target: 100M+ samples)
- Improved deduplication (near-duplicate filtering)
- Dataset statistics and analytics
- Additional language normalization
Each major release will be versioned clearly.
---
## Citation
If you use this dataset in research or publications, please cite:
```
TinyWay-Gutenberg-Clean
Shivam (NNEngine), 2026
```