| | |
| |
|
| | import subprocess |
| | from contextlib import redirect_stdout |
| | from dataclasses import asdict |
| | from io import StringIO |
| | from unittest import mock |
| |
|
| | import pytest |
| | import torch |
| | import yaml |
| |
|
| | import litgpt.eval.evaluate as module |
| | from litgpt import GPT, Config |
| | from litgpt.scripts.download import download_from_hub |
| |
|
| |
|
| | @pytest.mark.flaky(reruns=3) |
| | def test_evaluate_script(tmp_path): |
| | ours_config = Config.from_name("pythia-14m") |
| | download_from_hub(repo_id="EleutherAI/pythia-14m", tokenizer_only=True, checkpoint_dir=tmp_path) |
| | checkpoint_dir = tmp_path / "EleutherAI" / "pythia-14m" |
| | ours_model = GPT(ours_config) |
| | torch.save(ours_model.state_dict(), checkpoint_dir / "lit_model.pth") |
| | with open(checkpoint_dir / "model_config.yaml", "w", encoding="utf-8") as fp: |
| | yaml.dump(asdict(ours_config), fp) |
| |
|
| | stdout = StringIO() |
| | with redirect_stdout(stdout), mock.patch("sys.argv", ["eval/evaluate.py"]): |
| | with pytest.raises(ValueError) as excinfo: |
| | module.convert_and_evaluate( |
| | checkpoint_dir, |
| | out_dir=tmp_path / "out_dir", |
| | device=None, |
| | dtype=torch.float32, |
| | limit=5, |
| | tasks="logiqa", |
| | batch_size=0, |
| | ) |
| | assert "batch_size must be a positive integer, 'auto', or in the format 'auto:N'." in str(excinfo.value) |
| |
|
| | with pytest.raises(ValueError) as excinfo: |
| | module.convert_and_evaluate( |
| | checkpoint_dir, |
| | out_dir=tmp_path / "out_dir", |
| | device=None, |
| | dtype=torch.float32, |
| | limit=5, |
| | tasks="logiqa", |
| | batch_size="invalid", |
| | ) |
| | assert "batch_size must be a positive integer, 'auto', or in the format 'auto:N'." in str(excinfo.value) |
| |
|
| | stdout = StringIO() |
| | with redirect_stdout(stdout), mock.patch("sys.argv", ["eval/evaluate.py"]): |
| | module.convert_and_evaluate( |
| | checkpoint_dir, |
| | out_dir=tmp_path / "out_dir", |
| | device=None, |
| | dtype=torch.float32, |
| | limit=5, |
| | tasks="logiqa", |
| | batch_size=1, |
| | ) |
| | stdout = stdout.getvalue() |
| | assert (tmp_path / "out_dir" / "results.json").is_file() |
| | assert "logiqa" in stdout |
| | assert "Metric" in stdout |
| | assert "Loading checkpoint shards" not in stdout |
| |
|
| |
|
| | def test_cli(): |
| | args = ["litgpt", "evaluate", "-h"] |
| | output = subprocess.check_output(args) |
| | output = str(output.decode()) |
| | assert "Evaluate a model with the LM Evaluation Harness" in output |
| |
|