data_factory / scripts /build_streamlit_aligned_parquet.py
marcuskwan's picture
Add files using upload-large-folder tool
d18d6b1 verified
#!/usr/bin/env python3
from __future__ import annotations
import argparse
import ast
import json
from collections import Counter
from pathlib import Path
import sys
from typing import Any
import pyarrow as pa
import pyarrow.parquet as pq
ROOT = Path(__file__).resolve().parents[1]
if str(ROOT) not in sys.path:
sys.path.insert(0, str(ROOT))
import dataset_catalog as catalog
def _to_rel(base_dir: Path, value: str) -> str:
text = str(value or "").strip()
if not text:
return ""
p = Path(text)
if not p.is_absolute():
return p.as_posix()
try:
return p.resolve().relative_to(base_dir).as_posix()
except ValueError:
return p.as_posix()
def _build_rows(base_dir: Path, dataset: list[dict[str, Any]]) -> list[dict[str, Any]]:
rows: list[dict[str, Any]] = []
for item in dataset:
row = dict(item)
row["video_rel_path"] = str(item.get("path", "")).strip()
row["image_gt_rel_path"] = _to_rel(base_dir, str(item.get("image_gt", "")).strip())
image_paths = item.get("image_paths") if isinstance(item.get("image_paths"), list) else []
row["image_paths_rel"] = [_to_rel(base_dir, str(path)) for path in image_paths]
rows.append(row)
rows.sort(key=lambda r: str(r.get("path", "")))
return rows
def _load_default_descriptions(base_dir: Path) -> dict[str, str]:
app_path = base_dir / "app.py"
if not app_path.exists():
return {}
tree = ast.parse(app_path.read_text(encoding="utf-8"))
for node in tree.body:
if isinstance(node, ast.Assign):
for target in node.targets:
if isinstance(target, ast.Name) and target.id == "DEFAULT_DESCRIPTIONS":
value = ast.literal_eval(node.value)
if isinstance(value, dict):
return {str(k): str(v) for k, v in value.items()}
return {}
def _validate_rows(base_dir: Path, rows: list[dict[str, Any]]) -> dict[str, Any]:
missing_videos: list[str] = []
for row in rows:
rel = str(row.get("path", "")).strip()
if not rel:
continue
if not (base_dir / rel).exists():
missing_videos.append(rel)
split_counter = Counter(str(row.get("data_split", "")).strip().lower() for row in rows)
category_counter = Counter(str(row.get("category", "")).strip() for row in rows)
major_counter = Counter(str(row.get("major", "")).strip() for row in rows)
return {
"total_rows": len(rows),
"split_counts": dict(sorted(split_counter.items())),
"category_counts": dict(sorted(category_counter.items())),
"major_counts": dict(sorted(major_counter.items())),
"missing_video_paths": missing_videos,
}
def main() -> int:
parser = argparse.ArgumentParser(
description="Build Parquet aligned with Streamlit app _build_dataset(meta) output."
)
parser.add_argument(
"--base-dir",
default=str(Path(__file__).resolve().parents[1]),
help="Repository base directory (default: script parent parent).",
)
parser.add_argument(
"--out-parquet",
default="dist/streamlit_aligned.parquet",
help="Parquet output path relative to base dir.",
)
parser.add_argument(
"--out-stats",
default="dist/streamlit_aligned_stats.json",
help="Stats output path relative to base dir.",
)
parser.add_argument(
"--strict-files",
action="store_true",
help="Fail if any video file in rows is missing on disk.",
)
args = parser.parse_args()
base_dir = Path(args.base_dir).resolve()
video_dir = base_dir / "videos"
dataset_test_path = base_dir / "video_dataset.json"
dataset_train_path = base_dir / "video_dataset_train.json"
legacy_path = base_dir / "video_texts.json"
out_parquet = Path(args.out_parquet)
if not out_parquet.is_absolute():
out_parquet = (base_dir / out_parquet).resolve()
out_stats = Path(args.out_stats)
if not out_stats.is_absolute():
out_stats = (base_dir / out_stats).resolve()
out_parquet.parent.mkdir(parents=True, exist_ok=True)
out_stats.parent.mkdir(parents=True, exist_ok=True)
meta, _, _ = catalog.load_meta(
base_dir=base_dir,
dataset_test_path=dataset_test_path,
dataset_train_path=dataset_train_path,
legacy_path=legacy_path,
)
default_descriptions = _load_default_descriptions(base_dir)
dataset = catalog.build_dataset(
base_dir=base_dir,
video_dir=video_dir,
meta=meta,
default_descriptions=default_descriptions,
)
rows = _build_rows(base_dir, dataset)
stats = _validate_rows(base_dir, rows)
if args.strict_files and stats["missing_video_paths"]:
sample = stats["missing_video_paths"][:20]
raise SystemExit(
f"Missing {len(stats['missing_video_paths'])} video paths. Sample: {sample}"
)
table = pa.Table.from_pylist(rows)
pq.write_table(table, out_parquet, compression="zstd")
out_stats.write_text(json.dumps(stats, ensure_ascii=False, indent=2), encoding="utf-8")
print(f"Rows: {stats['total_rows']}")
print(f"Split counts: {stats['split_counts']}")
print(f"Parquet: {out_parquet}")
print(f"Stats: {out_stats}")
if stats["missing_video_paths"]:
print(f"Missing video paths: {len(stats['missing_video_paths'])}")
return 0
if __name__ == "__main__":
raise SystemExit(main())