| | import json |
| |
|
| | import pytest |
| | import torch |
| | from litdata import optimize |
| | from litdata.streaming import StreamingDataset, TokensLoader |
| | from torch.utils._pytree import tree_map |
| |
|
| |
|
| | def tokenize(data): |
| | for story in data: |
| | yield torch.tensor(story) |
| |
|
| |
|
| | def fake_chunk(path, data): |
| | optimize( |
| | fn=tokenize, |
| | inputs=[data] * len(data), |
| | output_dir=str(path), |
| | num_workers=1, |
| | chunk_bytes="200MB", |
| | item_loader=TokensLoader(), |
| | ) |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | ("max_seq_len", "expected"), |
| | [ |
| | (2, [[0, 23, 15], [63, 0, 73], [5, 0, 1], [1999, 0, 13]]), |
| | (5, [[0, 23, 15, 63, 0, 73], [5, 0, 1, 1999, 0, 13]]), |
| | (6, [[0, 23, 15, 63, 0, 73, 5]]), |
| | (7, [[0, 23, 15, 63, 0, 73, 5, 0]]), |
| | ], |
| | ) |
| | def test_pretok_dataset(tmp_path, max_seq_len, expected): |
| | fake_data = [0, 23, 15, 63, 0, 73, 5, 0, 1, 1999, 0, 13] |
| | assert len(fake_data) == 12 |
| | fake_chunk(tmp_path, [fake_data]) |
| |
|
| | dataset = StreamingDataset( |
| | input_dir=str(tmp_path), item_loader=TokensLoader(block_size=max_seq_len + 1), shuffle=False, drop_last=False |
| | ) |
| | actual = tree_map(torch.Tensor.tolist, list(dataset)) |
| | assert actual == expected |
| |
|
| |
|
| | def test_tokenize(tmp_path, monkeypatch): |
| | from litgpt.data.tinystories import tokenize |
| |
|
| | story1, story2 = "foo bar", " fun " |
| | data = [{"story": story1}, {"story": story2}] |
| | shard_path = tmp_path / "data.json" |
| | with open(shard_path, "w", encoding="utf-8") as f: |
| | json.dump(data, f) |
| |
|
| | class Tokenizer: |
| | bos_id = 0 |
| |
|
| | def encode(self, text, bos, eos): |
| | assert bos |
| | assert not eos |
| | return [self.bos_id] + [ord(c) for c in text] |
| |
|
| | monkeypatch.setenv("DATA_OPTIMIZER_GLOBAL_RANK", "0") |
| | monkeypatch.setenv("DATA_OPTIMIZER_NUM_WORKERS", "1") |
| | data = tokenize(str(shard_path), Tokenizer()) |
| | assert list(data) == [[0, 102, 111, 111, 32, 98, 97, 114], [0, 102, 117, 110]] |
| |
|
| |
|
| | def test_tinystories_datamodule(tmp_path): |
| | from litgpt.data.tinystories import TinyStories |
| |
|
| | data_dir = tmp_path / "tinystories" |
| |
|
| | datamodule = TinyStories(data_dir, seed=42, num_workers=1) |
| | datamodule.connect(max_seq_length=2) |
| |
|
| | |
| | train_data_dir = data_dir / "train" |
| | train_data_dir.mkdir(parents=True) |
| | fake_chunk(train_data_dir, [[12], [0, 23, 15, 63, 0], [73, 5, 0, 1, 1999, 0, 13]]) |
| |
|
| | datamodule.setup() |
| |
|
| | tr_dataloader = datamodule.train_dataloader() |
| | tr_dataloader.shuffle = False |
| |
|
| | actual = tree_map(torch.Tensor.tolist, list(tr_dataloader)) |
| |
|
| | |
| | assert actual == [ |
| | [[73, 5, 0]], |
| | [[12, 0, 23]], |
| | [[5, 0, 1]], |
| | [[0, 73, 5]], |
| | [[1999, 0, 13]], |
| | [[0, 1, 1999]], |
| | [[1, 1999, 0]], |
| | [[0, 23, 15]], |
| | [[13, 12, 0]], |
| | [[63, 0, 73]], |
| | [[23, 15, 63]], |
| | [[15, 63, 0]], |
| | [[0, 13, 12]], |
| | ] |
| |
|