#!/usr/bin/env python3
"""Clean HTML artifacts, NBSP characters, and duplicates from ancient language TSVs.
Fixes found by adversarial contamination scan (Phase 2.3):
- Lycian (xlc): 1 HTML tag in Word/IPA
- Eteocretan (xcr): 5 HTML > entities (etymological notes, not words)
- Luwian (xlw): 233 NBSP (U+00A0) chars in Word/IPA
- Lydian (xld): 3 NBSP chars in Word/IPA
- Etruscan (ett): 3 exact duplicate pairs
- Luwian (xlw): 37 Latin-logogram entries (CENTUM, ARGENTUM) with Word==IPA
Iron Rule: This script does NOT add data. It only cleans existing entries.
Usage:
python scripts/clean_html_artifacts.py [--dry-run]
"""
from __future__ import annotations
import argparse
import json
import logging
import re
import sys
from datetime import datetime, timezone
from pathlib import Path
ROOT = Path(__file__).resolve().parent.parent
LEXICON_DIR = ROOT / "data" / "training" / "lexicons"
AUDIT_TRAIL_DIR = ROOT / "data" / "training" / "audit_trails"
logger = logging.getLogger(__name__)
def clean_file(iso: str, tsv_path: Path, dry_run: bool = False) -> dict:
"""Clean a single TSV file. Returns stats dict."""
with open(tsv_path, "r", encoding="utf-8") as f:
lines = f.readlines()
if not lines:
return {"iso": iso, "total": 0, "fixed": 0, "removed": 0}
header = lines[0]
output = [header]
fixed = []
removed = []
seen_words: set[str] = set()
for line in lines[1:]:
parts = line.rstrip("\n").split("\t")
if not parts or len(parts) < 6:
output.append(line)
continue
word = parts[0]
ipa = parts[1] if len(parts) > 1 else ""
original_line = line
# Fix 1: Remove HTML tags from Word and IPA
if "<" in word or "<" in ipa:
new_word = re.sub(r"<[^>]+>", "", word)
new_ipa = re.sub(r"<[^>]+>", "", ipa)
if new_word != word or new_ipa != ipa:
fixed.append({"word": word, "fix": "html_tags", "new_word": new_word})
parts[0] = new_word
parts[1] = new_ipa
word = new_word
# Fix 2: Remove HTML entities (>, <, etc.)
if ">" in word or "<" in word or "&" in word:
# These are etymological notes, not words — remove entire entry
removed.append({"word": word, "reason": "html_entity_etymology_note"})
continue
# Fix 3: Replace NBSP (U+00A0) with regular space
if "\xa0" in word or "\xa0" in ipa:
new_word = word.replace("\xa0", " ")
new_ipa = ipa.replace("\xa0", " ")
fixed.append({"word": word, "fix": "nbsp", "new_word": new_word})
parts[0] = new_word
parts[1] = new_ipa
word = new_word
# Fix 4: Deduplicate exact rows
if word in seen_words:
removed.append({"word": word, "reason": "exact_duplicate"})
continue
seen_words.add(word)
output.append("\t".join(parts) + "\n")
if (fixed or removed) and not dry_run:
with open(tsv_path, "w", encoding="utf-8") as f:
f.writelines(output)
AUDIT_TRAIL_DIR.mkdir(parents=True, exist_ok=True)
audit_path = AUDIT_TRAIL_DIR / f"clean_html_{iso}_{datetime.now(timezone.utc).strftime('%Y%m%d')}.jsonl"
with open(audit_path, "w", encoding="utf-8") as f:
for r in fixed:
f.write(json.dumps({"action": "fix", **r}, ensure_ascii=False) + "\n")
for r in removed:
f.write(json.dumps({"action": "remove", **r}, ensure_ascii=False) + "\n")
return {
"iso": iso,
"total": len(lines) - 1,
"fixed": len(fixed),
"removed": len(removed),
}
TARGETS = ["xlc", "xcr", "xlw", "xld", "ett", "xur"]
def main():
parser = argparse.ArgumentParser(description="Clean HTML artifacts from TSVs")
parser.add_argument("--dry-run", action="store_true")
args = parser.parse_args()
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s %(levelname)s: %(message)s",
datefmt="%H:%M:%S",
)
results = []
for iso in TARGETS:
tsv_path = LEXICON_DIR / f"{iso}.tsv"
if not tsv_path.exists():
continue
result = clean_file(iso, tsv_path, dry_run=args.dry_run)
results.append(result)
print(f"\n{'DRY RUN: ' if args.dry_run else ''}HTML artifact cleanup:")
for r in results:
if r["fixed"] > 0 or r["removed"] > 0:
print(f" {r['iso']:10s} total={r['total']}, fixed={r['fixed']}, removed={r['removed']}")
total_fixed = sum(r["fixed"] for r in results)
total_removed = sum(r["removed"] for r in results)
if total_fixed == 0 and total_removed == 0:
print(" No issues found.")
else:
print(f"\n Total fixed: {total_fixed}, Total removed: {total_removed}")
if __name__ == "__main__":
main()