| |
| """ |
| Build the Linear A Phonotactics Validation dataset. |
| |
| Criteria for inclusion: |
| 1. Both Lang_A and Lang_B have Linear_A_Score >= 0.80 |
| (open syllable ratio, no clusters, final vowels, CV structure) |
| 2. Confidence = "certain" (expert-determined, not contested) |
| 3. Phono_Quality = "strong" (SCA Score >= 0.5, clear phonological similarity) |
| |
| This produces a gold-standard dataset of proven cognate pairs between |
| languages whose phonotactics resemble Linear A. Used to validate |
| cognate detection models before applying them to Linear A. |
| |
| Input: |
| - analysis/typology_linear_a.tsv (Linear A similarity scores per language) |
| - data/training/cognate_pairs/cognate_pairs_inherited.parquet (full dataset) |
| Output: |
| - data/training/cognate_pairs/linear_a_phonotactics_validation.parquet |
| """ |
| import csv |
| import sys |
| from pathlib import Path |
|
|
| if sys.stdout.encoding != 'utf-8': |
| sys.stdout.reconfigure(encoding='utf-8') |
|
|
| THRESHOLD = 0.80 |
|
|
|
|
| def main(): |
| hf_dir = Path(__file__).parent.parent |
|
|
| |
| print('Loading Linear A typology scores...') |
| lang_scores = {} |
| with open(hf_dir / 'analysis' / 'typology_linear_a.tsv', encoding='utf-8') as f: |
| for row in csv.DictReader(f, delimiter='\t'): |
| lang_scores[row['Language']] = float(row['Linear_A_Score']) |
|
|
| qualified_langs = {l for l, s in lang_scores.items() if s >= THRESHOLD} |
| print(f' Languages with score >= {THRESHOLD}: {len(qualified_langs)}') |
|
|
| |
| import pyarrow as pa |
| import pyarrow.parquet as pq |
| import pyarrow.compute as pc |
|
|
| print('Loading full inherited Parquet...') |
| table = pq.read_table( |
| hf_dir / 'data' / 'training' / 'cognate_pairs' / 'cognate_pairs_inherited.parquet' |
| ) |
| print(f' Total rows: {table.num_rows:,}') |
|
|
| |
| print('Filtering...') |
| lang_a = table['Lang_A'].to_pylist() |
| lang_b = table['Lang_B'].to_pylist() |
| conf = table['Confidence'].to_pylist() |
| phono = table['Phono_Quality'].to_pylist() |
|
|
| mask = [] |
| for i in range(len(lang_a)): |
| keep = ( |
| lang_a[i] in qualified_langs |
| and lang_b[i] in qualified_langs |
| and conf[i] == 'certain' |
| and phono[i] == 'strong' |
| ) |
| mask.append(keep) |
|
|
| mask_arr = pa.array(mask) |
| filtered = table.filter(mask_arr) |
| print(f' Filtered rows: {filtered.num_rows:,}') |
|
|
| |
| scores_a = [lang_scores.get(la, 0.0) for la, m in zip(lang_a, mask) if m] |
| scores_b = [lang_scores.get(lb, 0.0) for lb, m in zip(lang_b, mask) if m] |
|
|
| filtered = filtered.append_column( |
| 'Linear_A_Score_A', pa.array([round(s, 4) for s in scores_a], type=pa.float64()) |
| ) |
| filtered = filtered.append_column( |
| 'Linear_A_Score_B', pa.array([round(s, 4) for s in scores_b], type=pa.float64()) |
| ) |
|
|
| |
| out_path = hf_dir / 'data' / 'training' / 'cognate_pairs' / 'linear_a_phonotactics_validation.parquet' |
| pq.write_table(filtered, str(out_path), compression='zstd', compression_level=3) |
|
|
| import os |
| size = os.path.getsize(out_path) |
| print(f'\n Written to: {out_path}') |
| print(f' Size: {size/1024/1024:.1f} MB') |
|
|
| |
| print(f'\n=== VALIDATION DATASET STATISTICS ===') |
| print(f'Total pairs: {filtered.num_rows:,}') |
|
|
| |
| langs = set() |
| fa = filtered['Lang_A'].to_pylist() |
| fb = filtered['Lang_B'].to_pylist() |
| for a in fa: |
| langs.add(a) |
| for b in fb: |
| langs.add(b) |
| print(f'Unique languages: {len(langs)}') |
|
|
| |
| t2 = pq.read_table(str(hf_dir / 'data' / 'training' / 'metadata' / 'languages.parquet')) |
| iso_to_family = dict(zip(t2['ISO'].to_pylist(), t2['Family'].to_pylist())) |
|
|
| |
| fam_counts = {} |
| for l in langs: |
| fam = iso_to_family.get(l, 'unknown') |
| fam_counts[fam] = fam_counts.get(fam, 0) + 1 |
| print(f'\nLanguage families:') |
| for fam, c in sorted(fam_counts.items(), key=lambda x: -x[1]): |
| print(f' {fam}: {c}') |
|
|
| |
| sources = filtered['Source'].to_pylist() |
| src_counts = {} |
| for s in sources: |
| src_counts[s] = src_counts.get(s, 0) + 1 |
| print(f'\nSource distribution:') |
| for src, c in sorted(src_counts.items(), key=lambda x: -x[1]): |
| print(f' {src}: {c:,}') |
|
|
| |
| scores = filtered['Score'].to_pylist() |
| float_scores = [float(s) for s in scores if s and s != '-1'] |
| if float_scores: |
| print(f'\nSCA Score distribution:') |
| print(f' Min: {min(float_scores):.4f}') |
| print(f' Max: {max(float_scores):.4f}') |
| print(f' Mean: {sum(float_scores)/len(float_scores):.4f}') |
| bins = [(0.5, 0.6), (0.6, 0.7), (0.7, 0.8), (0.8, 0.9), (0.9, 1.01)] |
| for lo, hi in bins: |
| n = sum(1 for s in float_scores if lo <= s < hi) |
| print(f' [{lo:.1f}, {hi:.1f}): {n:,}') |
|
|
| |
| iso_to_name = {} |
| glot_path = hf_dir / 'data' / 'training' / 'raw' / 'glottolog_cldf' / 'languages.csv' |
| with open(glot_path, encoding='utf-8') as f: |
| for row in csv.DictReader(f): |
| iso = row.get('ISO639P3code', '').strip() |
| name = row.get('Name', '').strip() |
| if iso and name: |
| iso_to_name[iso] = name |
|
|
| |
| lang_pair_counts = {} |
| for a, b in zip(fa, fb): |
| lang_pair_counts[a] = lang_pair_counts.get(a, 0) + 1 |
| lang_pair_counts[b] = lang_pair_counts.get(b, 0) + 1 |
| print(f'\nTop 20 languages by pair count:') |
| for lang, c in sorted(lang_pair_counts.items(), key=lambda x: -x[1])[:20]: |
| name = iso_to_name.get(lang, '?') |
| fam = iso_to_family.get(lang, '?') |
| score = lang_scores.get(lang, 0) |
| print(f' {lang} ({name}) - {fam} - LA score {score:.4f} - {c:,} pairs') |
|
|
|
|
| if __name__ == '__main__': |
| main() |
|
|