| |
| """Generate per-language README profile markdown files. |
| |
| Reads metadata, lexicons, and validation files, then writes one markdown |
| profile per language to data/training/language_profiles/{iso}.md. |
| |
| Usage: |
| python scripts/generate_language_readmes.py |
| """ |
|
|
| from __future__ import annotations |
|
|
| import csv |
| import json |
| import sys |
| from collections import Counter, defaultdict |
| from pathlib import Path |
|
|
| import langcodes |
|
|
| |
| sys.path.insert( |
| 0, |
| str(Path(__file__).resolve().parent.parent / "cognate_pipeline" / "src"), |
| ) |
|
|
| |
| _SCRIPTS_DIR = Path(__file__).resolve().parent |
| sys.path.insert(0, str(_SCRIPTS_DIR)) |
| from build_validation_sets import ( |
| ANCIENT, |
| MEDIEVAL, |
| RELIGIOUS_ALL, |
| RELIGIOUS_SUBDOMAINS, |
| classify_era, |
| is_religious, |
| ) |
|
|
| |
| |
| |
|
|
| REPO_ROOT = Path(__file__).resolve().parent.parent |
| TRAINING_DIR = REPO_ROOT / "data" / "training" |
| METADATA_DIR = TRAINING_DIR / "metadata" |
| LEXICONS_DIR = TRAINING_DIR / "lexicons" |
| VALIDATION_DIR = TRAINING_DIR / "validation" |
| OUTPUT_DIR = TRAINING_DIR / "language_profiles" |
|
|
| |
| ONOMASTIC_SUBDOMAINS = {"core_religious", "sacred_places", "supernatural", "cosmic_spiritual"} |
| VERBAL_SUBDOMAINS = {"religious_verbs", "ritual_ceremony", "moral_ethical"} |
|
|
|
|
| |
| |
| |
|
|
| def load_languages_tsv() -> dict[str, dict[str, str]]: |
| """Read languages.tsv → {iso: {family, glottocode, entries, sources}}.""" |
| path = METADATA_DIR / "languages.tsv" |
| langs: dict[str, dict[str, str]] = {} |
| with path.open(encoding="utf-8") as fh: |
| reader = csv.DictReader(fh, delimiter="\t") |
| for row in reader: |
| iso = row["ISO"] |
| langs[iso] = { |
| "family": row.get("Family", ""), |
| "glottocode": row.get("Glottocode", ""), |
| "entries": row.get("Entries", "0"), |
| "sources": row.get("Sources", ""), |
| } |
| return langs |
|
|
|
|
| def load_phylo_tree() -> dict: |
| """Load phylo_tree.json.""" |
| path = VALIDATION_DIR / "phylo_tree.json" |
| with path.open(encoding="utf-8") as fh: |
| return json.load(fh) |
|
|
|
|
| def build_lang_paths(tree: dict) -> dict[str, list[str]]: |
| """Map each ISO to its full path from root to leaf group.""" |
| paths: dict[str, list[str]] = {} |
|
|
| def _walk(node, prefix: list[str]) -> None: |
| if isinstance(node, list): |
| for iso in node: |
| paths[iso] = list(prefix) |
| elif isinstance(node, dict): |
| for key, child in node.items(): |
| _walk(child, prefix + [key]) |
| elif isinstance(node, str): |
| paths[node] = list(prefix) |
|
|
| _walk(tree, []) |
| return paths |
|
|
|
|
| def get_display_name(iso: str) -> str: |
| """Get human-readable language name via langcodes.""" |
| try: |
| return langcodes.Language.get(iso).display_name() |
| except Exception: |
| return iso.upper() |
|
|
|
|
| |
| |
| |
|
|
| |
| VALIDATION_FILES: list[str] = [ |
| "true_cognates_L1.tsv", |
| "true_cognates_L2.tsv", |
| "true_cognates_L3.tsv", |
| "false_positives.tsv", |
| "true_negatives.tsv", |
| "borrowings.tsv", |
| "timespan_ancient_ancient.tsv", |
| "timespan_ancient_modern.tsv", |
| "timespan_medieval_modern.tsv", |
| "timespan_modern_modern.tsv", |
| ] |
|
|
| PER_FAMILY_DIR = VALIDATION_DIR / "per_family" |
| RELIGIOUS_DIR = VALIDATION_DIR / "religious" |
|
|
| RELIGIOUS_FILES: list[str] = [ |
| "all_pairs.tsv", |
| "true_cognates.tsv", |
| "false_positives.tsv", |
| "borrowings.tsv", |
| ] |
|
|
|
|
| def index_validation_files() -> tuple[ |
| dict[str, Counter[str]], |
| dict[str, Counter[str]], |
| dict[str, dict[str, set[str]]], |
| dict[str, dict[str, Counter[str]]], |
| ]: |
| """Single-pass index of all validation TSVs. |
| |
| Returns: |
| pair_counts: iso → {filename: pair_count} |
| partners: iso → Counter(partner_iso) |
| concept_sets: iso → {filename: set(concept_ids)} |
| religious_labels: iso → {filename: Counter(label)} |
| """ |
| pair_counts: dict[str, Counter[str]] = defaultdict(Counter) |
| partners: dict[str, Counter[str]] = defaultdict(Counter) |
| concept_sets: dict[str, dict[str, set[str]]] = defaultdict(lambda: defaultdict(set)) |
| religious_labels: dict[str, dict[str, Counter[str]]] = defaultdict(lambda: defaultdict(Counter)) |
|
|
| def _index_file(filepath: Path, tag: str, is_religious_file: bool = False) -> None: |
| if not filepath.exists(): |
| return |
| with filepath.open(encoding="utf-8") as fh: |
| reader = csv.DictReader(fh, delimiter="\t") |
| for row in reader: |
| lang_a = row.get("Lang_A", "") |
| lang_b = row.get("Lang_B", "") |
| cid = row.get("Concept_ID", "") |
| label = row.get("Label", "") |
|
|
| if lang_a: |
| pair_counts[lang_a][tag] += 1 |
| concept_sets[lang_a][tag].add(cid) |
| if lang_b: |
| partners[lang_a][lang_b] += 1 |
| if is_religious_file: |
| religious_labels[lang_a][tag][label] += 1 |
|
|
| if lang_b: |
| pair_counts[lang_b][tag] += 1 |
| concept_sets[lang_b][tag].add(cid) |
| if lang_a: |
| partners[lang_b][lang_a] += 1 |
| if is_religious_file: |
| religious_labels[lang_b][tag][label] += 1 |
|
|
| |
| for fname in VALIDATION_FILES: |
| print(f" Indexing {fname} ...") |
| _index_file(VALIDATION_DIR / fname, fname) |
|
|
| |
| if PER_FAMILY_DIR.exists(): |
| for fp in sorted(PER_FAMILY_DIR.glob("*.tsv")): |
| tag = f"per_family/{fp.name}" |
| print(f" Indexing {tag} ...") |
| _index_file(fp, tag) |
|
|
| |
| for fname in RELIGIOUS_FILES: |
| tag = f"religious/{fname}" |
| print(f" Indexing {tag} ...") |
| _index_file(RELIGIOUS_DIR / fname, tag, is_religious_file=True) |
|
|
| |
| for sd_name in sorted(RELIGIOUS_SUBDOMAINS): |
| fname = f"{sd_name}.tsv" |
| tag = f"religious/{fname}" |
| fp = RELIGIOUS_DIR / fname |
| if fp.exists(): |
| print(f" Indexing {tag} ...") |
| _index_file(fp, tag, is_religious_file=True) |
|
|
| return dict(pair_counts), dict(partners), dict(concept_sets), dict(religious_labels) |
|
|
|
|
| |
| |
| |
|
|
| def analyze_lexicon(iso: str) -> dict: |
| """Read a single lexicon TSV and return summary stats. |
| |
| Returns dict with keys: total_entries, unique_words, unique_concepts, |
| religious_concepts (dict of subdomain → set of concept_ids), |
| concept_list (set of all concept_ids). |
| """ |
| path = LEXICONS_DIR / f"{iso}.tsv" |
| if not path.exists(): |
| return { |
| "total_entries": 0, |
| "unique_words": 0, |
| "unique_concepts": 0, |
| "religious_concepts": {}, |
| "concept_list": set(), |
| } |
|
|
| words: set[str] = set() |
| concepts: set[str] = set() |
| total = 0 |
|
|
| with path.open(encoding="utf-8") as fh: |
| reader = csv.DictReader(fh, delimiter="\t") |
| for row in reader: |
| cid = row.get("Concept_ID", "").strip() |
| if cid in ("", "-"): |
| continue |
| word = row.get("Word", "").strip() |
| total += 1 |
| if word: |
| words.add(word) |
| concepts.add(cid) |
|
|
| |
| religious_concepts: dict[str, set[str]] = {} |
| for sd_name, sd_set in RELIGIOUS_SUBDOMAINS.items(): |
| matching = set() |
| for cid in concepts: |
| if cid in sd_set or cid.upper() in {c.upper() for c in sd_set}: |
| matching.add(cid) |
| if matching: |
| religious_concepts[sd_name] = matching |
|
|
| return { |
| "total_entries": total, |
| "unique_words": len(words), |
| "unique_concepts": len(concepts), |
| "religious_concepts": religious_concepts, |
| "concept_list": concepts, |
| } |
|
|
|
|
| |
| |
| |
|
|
| def format_path(path_segments: list[str]) -> str: |
| """Format a phylo path as 'Family > Branch > Sub-branch'.""" |
| if not path_segments: |
| return "—" |
| return " > ".join(s.replace("_", " ").title() for s in path_segments) |
|
|
|
|
| def render_profile( |
| iso: str, |
| meta: dict[str, str], |
| lang_paths: dict[str, list[str]], |
| pair_counts: dict[str, Counter[str]], |
| partners_index: dict[str, Counter[str]], |
| concept_sets: dict[str, dict[str, set[str]]], |
| religious_labels: dict[str, dict[str, Counter[str]]], |
| lexicon_stats: dict, |
| ) -> str: |
| """Render a single language profile as markdown.""" |
| display_name = get_display_name(iso) |
| era = classify_era(iso) |
| phylo_path = lang_paths.get(iso, []) |
| family = meta.get("family", "unknown") |
| glottocode = meta.get("glottocode", "") |
| sources = meta.get("sources", "") |
|
|
| counts = pair_counts.get(iso, Counter()) |
| my_partners = partners_index.get(iso, Counter()) |
| my_concepts = concept_sets.get(iso, {}) |
| my_rel_labels = religious_labels.get(iso, {}) |
|
|
| lines: list[str] = [] |
|
|
| |
| lines.append(f"# {display_name} (`{iso}`)") |
| lines.append("") |
|
|
| |
| lines.append("## Overview") |
| lines.append("") |
| lines.append("| Field | Value |") |
| lines.append("|-------|-------|") |
| lines.append(f"| **ISO 639-3** | `{iso}` |") |
| lines.append(f"| **Family** | {family.replace('_', ' ').title()} |") |
| lines.append(f"| **Branch path** | {format_path(phylo_path)} |") |
| lines.append(f"| **Glottocode** | {glottocode or '—'} |") |
| lines.append(f"| **Era** | {era} |") |
| lines.append(f"| **Sources** | {sources or '—'} |") |
| lines.append("") |
|
|
| |
| lines.append("## Lexicon Summary") |
| lines.append("") |
| lines.append("| Metric | Count |") |
| lines.append("|--------|------:|") |
| lines.append(f"| Total entries | {lexicon_stats['total_entries']:,} |") |
| lines.append(f"| Unique words | {lexicon_stats['unique_words']:,} |") |
| lines.append(f"| Unique concepts | {lexicon_stats['unique_concepts']:,} |") |
| lines.append("") |
|
|
| |
| lines.append("## Cognate Pair Participation") |
| lines.append("") |
| lines.append("| Validation Set | Pairs |") |
| lines.append("|----------------|------:|") |
| for fname, label in [ |
| ("true_cognates_L1.tsv", "True Cognates L1"), |
| ("true_cognates_L2.tsv", "True Cognates L2"), |
| ("true_cognates_L3.tsv", "True Cognates L3"), |
| ("false_positives.tsv", "False Positives"), |
| ("true_negatives.tsv", "True Negatives"), |
| ("borrowings.tsv", "Borrowings"), |
| ]: |
| n = counts.get(fname, 0) |
| lines.append(f"| {label} | {n:,} |") |
| lines.append("") |
|
|
| |
| lines.append("## Timespan Distribution") |
| lines.append("") |
| lines.append("| Timespan | Pairs |") |
| lines.append("|----------|------:|") |
| for ts in ["ancient_ancient", "ancient_modern", "medieval_modern", "modern_modern"]: |
| fname = f"timespan_{ts}.tsv" |
| n = counts.get(fname, 0) |
| lines.append(f"| {ts.replace('_', ' ').title()} | {n:,} |") |
| lines.append("") |
|
|
| |
| family_files = sorted( |
| k for k in counts if k.startswith("per_family/") |
| ) |
| lines.append("## Family-Internal Pairs") |
| lines.append("") |
| if family_files: |
| lines.append("| Family File | Pairs |") |
| lines.append("|-------------|------:|") |
| for ff in family_files: |
| short = ff.replace("per_family/", "").replace(".tsv", "").replace("_", " ").title() |
| n = counts[ff] |
| lines.append(f"| {short} | {n:,} |") |
| else: |
| lines.append("This language does not appear in any per-family validation file.") |
| lines.append("") |
|
|
| |
| rel_all = counts.get("religious/all_pairs.tsv", 0) |
| lines.append("## Religious Domain") |
| lines.append("") |
| if rel_all > 0: |
| lines.append(f"**Total religious pairs:** {rel_all:,}") |
| lines.append("") |
|
|
| |
| lines.append("| Sub-domain | Category | Lexicon Concepts | Validation Pairs |") |
| lines.append("|------------|----------|----------------:|-----------------:|") |
| for sd_name in sorted(RELIGIOUS_SUBDOMAINS): |
| category = "onomastic" if sd_name in ONOMASTIC_SUBDOMAINS else "verbal" |
| lex_count = len(lexicon_stats["religious_concepts"].get(sd_name, set())) |
| val_tag = f"religious/{sd_name}.tsv" |
| val_count = counts.get(val_tag, 0) |
| sd_display = sd_name.replace("_", " ").title() |
| lines.append(f"| {sd_display} | {category} | {lex_count} | {val_count:,} |") |
| lines.append("") |
|
|
| |
| rel_label_agg: Counter[str] = Counter() |
| for tag, label_counter in my_rel_labels.items(): |
| rel_label_agg.update(label_counter) |
| if rel_label_agg: |
| lines.append("**Religious label distribution:**") |
| lines.append("") |
| lines.append("| Label | Count |") |
| lines.append("|-------|------:|") |
| for label, cnt in rel_label_agg.most_common(): |
| lines.append(f"| {label} | {cnt:,} |") |
| lines.append("") |
| else: |
| lines.append("This language does not appear in any religious validation file.") |
| lines.append("") |
|
|
| |
| lines.append("## Top Partner Languages") |
| lines.append("") |
| if my_partners: |
| top15 = my_partners.most_common(15) |
| lines.append("| Partner | Pairs |") |
| lines.append("|---------|------:|") |
| for partner_iso, cnt in top15: |
| partner_name = get_display_name(partner_iso) |
| lines.append(f"| {partner_name} (`{partner_iso}`) | {cnt:,} |") |
| else: |
| lines.append("No partner languages found in validation data.") |
| lines.append("") |
|
|
| return "\n".join(lines) |
|
|
|
|
| |
| |
| |
|
|
| def main() -> None: |
| print("=" * 70) |
| print("generate_language_readmes.py — Per-Language Profile Generation") |
| print("=" * 70) |
|
|
| |
| print("\nPhase 1: Loading metadata ...") |
| languages = load_languages_tsv() |
| print(f" {len(languages)} languages in metadata") |
|
|
| tree = load_phylo_tree() |
| lang_paths = build_lang_paths(tree) |
| print(f" Phylo tree covers {len(lang_paths)} languages") |
|
|
| |
| print("\nPhase 2: Indexing validation files ...") |
| pair_counts, partners_index, concept_sets, religious_labels = index_validation_files() |
| print(f" Indexed data for {len(pair_counts)} languages") |
|
|
| |
| print("\nPhase 3+4: Analyzing lexicons and rendering profiles ...") |
| OUTPUT_DIR.mkdir(parents=True, exist_ok=True) |
|
|
| isos = sorted(languages.keys()) |
| total = len(isos) |
| for i, iso in enumerate(isos, 1): |
| if i % 100 == 0 or i == total: |
| print(f" Processing {i}/{total} ({iso}) ...") |
|
|
| lexicon_stats = analyze_lexicon(iso) |
| meta = languages[iso] |
|
|
| md = render_profile( |
| iso, meta, lang_paths, |
| pair_counts, partners_index, concept_sets, |
| religious_labels, lexicon_stats, |
| ) |
|
|
| out_path = OUTPUT_DIR / f"{iso}.md" |
| out_path.write_text(md, encoding="utf-8") |
|
|
| print(f"\n Wrote {total} profiles to {OUTPUT_DIR.relative_to(REPO_ROOT)}/") |
| print("=" * 70) |
| print("Done!") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|