title stringclasses 9
values | text stringclasses 9
values |
|---|---|
Die Räuber | "Friedrich Schiller\nDie Räuber\nEin Schauspiel\n\n\nHippocrates\n\nQuae medicamenta non sanant, fe(...TRUNCATED) |
Die Verschwörung des Fiesco zu Genua | "Friedrich Schiller\nDie Verschwörung des Fiesco zu Genua\nEin republikanisches Trauerspiel\n\n\nSa(...TRUNCATED) |
Kabale und Liebe | "Friedrich Schiller\nKabale und Liebe\nEin bürgerliches Trauerspiel in fünf Aufzügen\n\n\nPersone(...TRUNCATED) |
Don Carlos, Infant von Spanien | "Friedrich Schiller\nDon Carlos, Infant von Spanien\nEin dramatisches Gedicht\n\n\nPersonen\nPhilipp(...TRUNCATED) |
Wallensteins Lager | "Friedrich Schiller\nWallensteins Lager\nEin dramatisches Gedicht\n\n\nProlog\nGesprochen bei Wieder(...TRUNCATED) |
Die Piccolomini | "Friedrich Schiller\nDie Piccolomini\nIn fünf Aufzügen\n\n\nPersonen\nWallenstein,\nHerzog zu Frie(...TRUNCATED) |
Wallensteins Tod | "Friedrich Schiller\nWallensteins Tod\nEin Trauerspiel in fünf Aufzügen\n\n\nPersonen\nWallenstein(...TRUNCATED) |
Maria Stuart | "Friedrich Schiller\nMaria Stuart\nEin Trauerspiel\n\n\nPersonen\nElisabeth,\nKönigin von England.\(...TRUNCATED) |
Die Jungfrau von Orleans | "Friedrich Schiller\nDie Jungfrau von Orleans\nEine romantische Tragödie\n\n\nPersonen\nKarl der Si(...TRUNCATED) |
tiny_schiller
A small (~2 MB) German-language analogue to Karpathy's tiny_shakespeare — 11 of Friedrich Schiller's dramatic works, cleaned and tokenised for tutorial-scale language models.
For a compact, agent-friendly summary (file inventory, load patterns, licensing), see DATA_CARD.md.
"Das Leben ist nur ein Moment, der Tod ist auch nur einer." — Friedrich Schiller
Corpus
~2.07 MB · 11 works · 2,019,857 characters · sourced from DraCor / GerDraCor (CC0). See LICENSING.md for details.
| Tokenizer | Tokens | chars/token |
|---|---|---|
| character-level | 2,019,857 | 1.00 |
| GPT-2 BPE | 854,611 | 2.36 |
cl100k_base |
642,593 | 3.14 |
Use character-level for teaching-scale models (88-token vocab, no tokenizer needed). Use cl100k over GPT-2 when sequence length matters — German umlauts and compounds tokenise 25% more efficiently.
Works
- Die Räuber
- Die Verschwörung des Fiesco zu Genua
- Kabale und Liebe
- Don Carlos, Infant von Spanien
- Wallensteins Lager
- Maria Stuart
- Die Jungfrau von Orleans
- Die Braut von Messina oder Die feindlichen Brüder
- Wilhelm Tell
- Die Piccolomini
- Wallensteins Tod
Quick Start — nanoGPT
python schiller_char/prepare.py # char-level, 88-vocab
python schiller_bpe/prepare.py # GPT-2 BPE, 50k vocab
python schiller_cl100k/prepare.py # cl100k, 100k vocab
HuggingFace Datasets
from datasets import load_dataset
ds = load_dataset("mrkschtr/tiny_schiller")
print(ds["train"][0]["title"])
print(ds["train"][0]["text"][:200])
9 works in train, 2 in test (Wilhelm Tell, Die Braut von Messina). Each row is one complete work with title and text fields.
Instruction & Character Datasets
Pre-built instruction-format parquet files are in data/ on the Hub.
General dialogue style — 7,607 examples teaching Schiller's dramatic register:
ds = load_dataset("mrkschtr/tiny_schiller", data_files="data/instruct.parquet", split="train")
print(ds[0]["prompt"])
print(ds[0]["completion"])
Per-character — fine-tune a model to respond as a specific character:
# 330 examples as Wallenstein · 325 as Carlos · 313 as Fiesco
# 237 as Marquis · 195 as Ferdinand · 194 as Königin · ...
ds = load_dataset("mrkschtr/tiny_schiller", data_files="data/char_WALLENSTEIN.parquet", split="train")
Rebuild locally (generates data/instruct.parquet + 89 data/char_*.parquet files by default):
python scripts/build_instruct.py # all characters
python scripts/build_instruct.py --list-characters # show available characters + turn counts
python scripts/build_instruct.py --character KARL # single character only
Fine-tuning small LLMs
pip install transformers trl datasets accelerate
python examples/finetune_sft.py --model TinyLlama/TinyLlama-1.1B-Chat-v1.0
Default context window is 2048 tokens. Match your model with --context_length:
python examples/finetune_sft.py --model microsoft/Phi-3-mini-4k-instruct --context_length 4096
python examples/finetune_sft.py --model Qwen/Qwen2.5-0.5B --context_length 4096
Tested: TinyLlama 1.1B · Phi-3 Mini 3.8B · Llama 3.2 1B/3B · Qwen2.5 0.5B–3B.
License
Text: public domain (Schiller died 1805) and sourced from DraCor / GerDraCor under CC0. See LICENSING.md for details.
Citation
@misc{schutera2023tinyschiller,
author = {Schutera, Mark},
title = {tiny\_schiller: a small German Schiller corpus for small language models},
year = {2023},
howpublished = {\url{https://github.com/schutera/tiny_schiller}},
note = {Source texts: DraCor / GerDraCor (CC0) and public domain.}
}
- Downloads last month
- 108
