ancient-scripts-datasets / scripts /expand_cldf_full.py
Alvin
Add complete dataset: all sources, metadata, scripts, docs, and phylo enrichment
26786e3
#!/usr/bin/env python3
"""Extract ALL data from CLDF sources — no concept filtering.
Reads NorthEuraLex (all 1,016 concepts), WOLD (all 1,814 concepts),
ABVD (all 210 concepts, top Austronesian languages), and sinotibetan
(all concepts). Writes to data/training/lexicons/{iso}.tsv files.
Dependencies: only Python standard library + cognate_pipeline.normalise.sound_class
"""
from __future__ import annotations
import csv
import re
import sys
import unicodedata
from collections import defaultdict
from pathlib import Path
ROOT = Path(__file__).resolve().parent.parent
sys.path.insert(0, str(ROOT / "cognate_pipeline" / "src"))
from cognate_pipeline.normalise.sound_class import ipa_to_sound_class
SOURCES = ROOT / "sources"
OUTPUT_DIR = ROOT / "data" / "training" / "lexicons"
METADATA_DIR = ROOT / "data" / "training" / "metadata"
# Header for lexicon files
HEADER = "Word\tIPA\tSCA\tSource\tConcept_ID\tCognate_Set_ID\n"
def read_cldf_csv(path: Path) -> list[dict[str, str]]:
"""Read a CLDF CSV file."""
if not path.exists():
return []
with open(path, encoding="utf-8", newline="") as f:
return list(csv.DictReader(f))
def segments_to_ipa(segments: str) -> str:
"""Convert CLDF Segments column (space-separated) to IPA string."""
if not segments or not segments.strip():
return ""
parts = segments.split()
cleaned = [p for p in parts if p not in ("^", "$", "+", "#", "_")]
return "".join(cleaned)
def normalize_ipa(ipa: str) -> str:
"""Basic IPA normalization: NFC, strip stress marks."""
ipa = unicodedata.normalize("NFC", ipa)
ipa = ipa.replace("\u02c8", "").replace("\u02cc", "")
ipa = ipa.replace(".", "")
return ipa.strip()
def form_to_pseudo_ipa(form: str) -> str:
"""For repos without IPA, use orthographic form as pseudo-IPA."""
if not form:
return ""
form = form.lower().strip()
form = re.sub(r"\(.*?\)", "", form)
if "," in form:
form = form.split(",")[0].strip()
if "/" in form:
form = form.split("/")[0].strip()
form = re.sub(r"[^a-zA-Z\u0250-\u02AF\u0300-\u036F\u0361]", "", form)
return form
def append_to_lexicon(
lang_entries: dict[str, list[tuple[str, str, str, str, str, str]]],
iso: str,
word: str,
ipa: str,
source: str,
concept_id: str = "-",
cognate_set_id: str = "-",
):
"""Append an entry to the in-memory language buffer."""
if not word or not ipa:
return
ipa = normalize_ipa(ipa)
if not ipa:
return
sca = ipa_to_sound_class(ipa)
lang_entries[iso].append((word, ipa, sca, source, concept_id, cognate_set_id))
def extract_northeuralex(lang_entries: dict) -> int:
"""Extract ALL NorthEuraLex data (all 1,016 concepts, all languages)."""
cldf_dir = SOURCES / "northeuralex" / "cldf"
if not cldf_dir.exists():
print(" NorthEuraLex not found, skipping")
return 0
# Build language map: NEL Language_ID -> ISO code
lang_map = {}
for row in read_cldf_csv(cldf_dir / "languages.csv"):
nel_id = row["ID"]
iso = row.get("ISO639P3code", "")
if not iso:
# Many NEL IDs are already ISO 639-3
iso = nel_id if len(nel_id) == 3 else ""
if iso:
lang_map[nel_id] = iso
# Build parameter map: NEL Parameter_ID -> Concepticon_ID + gloss
param_map = {}
for row in read_cldf_csv(cldf_dir / "parameters.csv"):
pid = row["ID"]
cid = row.get("Concepticon_ID", "")
gloss = row.get("Concepticon_Gloss", row.get("Name", pid))
param_map[pid] = (cid, gloss)
# Process all forms
processed = 0
for row in read_cldf_csv(cldf_dir / "forms.csv"):
lang_id = row.get("Language_ID", "")
param_id = row.get("Parameter_ID", "")
segments = row.get("Segments", "")
iso = lang_map.get(lang_id)
if not iso:
continue
ipa = segments_to_ipa(segments)
if not ipa:
continue
word = row.get("Value", row.get("Form", ""))
cid_info = param_map.get(param_id, ("", param_id))
concept_id = cid_info[1] if cid_info[1] else param_id
append_to_lexicon(lang_entries, iso, word, ipa, "northeuralex", concept_id)
processed += 1
print(f" NorthEuraLex: {processed:,} entries from {len(set(lang_map.values()))} languages")
return processed
def extract_wold(lang_entries: dict) -> int:
"""Extract ALL WOLD data (all concepts, all languages)."""
cldf_dir = SOURCES / "wold" / "cldf"
if not cldf_dir.exists():
print(" WOLD not found, skipping")
return 0
# Build language map
wold_lang_map = {}
for row in read_cldf_csv(cldf_dir / "languages.csv"):
wold_id = row["ID"]
iso = row.get("ISO639P3code", "")
if iso:
wold_lang_map[wold_id] = iso
# Build parameter map
param_map = {}
for row in read_cldf_csv(cldf_dir / "parameters.csv"):
pid = row["ID"]
cid = row.get("Concepticon_ID", "")
gloss = row.get("Concepticon_Gloss", row.get("Name", pid))
param_map[pid] = (cid, gloss)
# Process all forms
processed = 0
borrowing_count = 0
for row in read_cldf_csv(cldf_dir / "forms.csv"):
lang_id = row.get("Language_ID", "")
param_id = row.get("Parameter_ID", "")
segments = row.get("Segments", "")
iso = wold_lang_map.get(lang_id)
if not iso:
continue
ipa = segments_to_ipa(segments)
if not ipa:
continue
word = row.get("Value", row.get("Form", ""))
cid_info = param_map.get(param_id, ("", param_id))
concept_id = cid_info[1] if cid_info[1] else param_id
# Track borrowing status
borrowed = row.get("Borrowed", "").strip()
cognate_id = "-"
if borrowed and borrowed not in ("0", ""):
borrowing_count += 1
append_to_lexicon(lang_entries, iso, word, ipa, "wold", concept_id, cognate_id)
processed += 1
print(f" WOLD: {processed:,} entries from {len(set(wold_lang_map.values()))} languages ({borrowing_count:,} borrowings)")
return processed
def extract_abvd(lang_entries: dict) -> int:
"""Extract ALL ABVD data (all concepts, Austronesian languages)."""
cldf_dir = SOURCES / "abvd" / "cldf"
if not cldf_dir.exists():
print(" ABVD not found, skipping")
return 0
# Build language map — select Austronesian languages with ISO codes
abvd_lang_map = {}
for row in read_cldf_csv(cldf_dir / "languages.csv"):
fam = row.get("Family", "")
if fam != "Austronesian":
continue
iso = row.get("ISO639P3code", "")
if not iso:
continue
abvd_lang_map[row["ID"]] = iso
# Build parameter map
param_map = {}
for row in read_cldf_csv(cldf_dir / "parameters.csv"):
pid = row["ID"]
cid = row.get("Concepticon_ID", "")
gloss = row.get("Concepticon_Gloss", row.get("Name", pid))
param_map[pid] = (cid, gloss)
processed = 0
cognate_count = 0
forms_path = cldf_dir / "forms.csv"
if not forms_path.exists():
print(" ABVD forms.csv not found")
return 0
with open(forms_path, encoding="utf-8", newline="") as f:
for row in csv.DictReader(f):
abvd_lang = row.get("Language_ID", "")
iso = abvd_lang_map.get(abvd_lang)
if not iso:
continue
param_id = row.get("Parameter_ID", "")
form = row.get("Form", "").strip()
if not form:
continue
cid_info = param_map.get(param_id, ("", param_id))
concept_id = cid_info[1] if cid_info[1] else param_id
# ABVD forms are orthographic — use as pseudo-IPA
ipa = form_to_pseudo_ipa(form)
if not ipa:
continue
cognacy = row.get("Cognacy", "").strip()
cognate_id = "-"
if cognacy:
cog_num = cognacy.split(",")[0].strip()
if cog_num:
cognate_id = f"abvd_{param_id}_{cog_num}"
cognate_count += 1
append_to_lexicon(lang_entries, iso, form, ipa, "abvd", concept_id, cognate_id)
processed += 1
print(f" ABVD: {processed:,} entries from {len(set(abvd_lang_map.values()))} languages ({cognate_count:,} with cognacy)")
return processed
def extract_sinotibetan(lang_entries: dict) -> int:
"""Extract ALL sinotibetan data."""
dump_path = SOURCES / "sinotibetan" / "sinotibetan_dump.tsv"
if not dump_path.exists():
dump_path = SOURCES / "sinotibetan" / "dumps" / "sinotibetan.tsv"
if not dump_path.exists():
print(" Sino-Tibetan dump not found, skipping")
return 0
# Doculect -> ISO code
doculect_map = {
"Old_Chinese": "och",
"Japhug": "jya",
"Tibetan_Written": "bod",
"Old_Burmese": "obr",
"Jingpho": "kac",
"Lisu": "lis",
"Naxi": "nxq",
"Khaling": "klr",
"Limbu": "lif",
"Pumi_Lanping": "pmi",
"Qiang_Mawo": "qxs",
"Tujia": "tji",
"Dulong": "duu",
"Hakha": "cnh",
"Bai_Jianchuan": "bca",
}
processed = 0
cognate_count = 0
with open(dump_path, encoding="utf-8", newline="") as f:
reader = csv.DictReader(f, delimiter="\t")
for row in reader:
doculect = row.get("DOCULECT", "")
iso = doculect_map.get(doculect)
if not iso:
continue
concept = row.get("CONCEPT", "").strip()
ipa = row.get("IPA", "").strip()
if not ipa or not concept:
continue
cogid = row.get("COGID", "").strip()
cognate_id = "-"
if cogid:
cognate_id = f"st_{cogid}"
cognate_count += 1
append_to_lexicon(lang_entries, iso, concept, ipa, "sinotibetan", concept, cognate_id)
processed += 1
print(f" Sino-Tibetan: {processed:,} entries ({cognate_count:,} with cognacy)")
return processed
def write_lexicons(lang_entries: dict[str, list]) -> dict[str, int]:
"""Write per-language lexicon TSV files. Returns {iso: count}."""
OUTPUT_DIR.mkdir(parents=True, exist_ok=True)
stats = {}
for iso in sorted(lang_entries):
entries = lang_entries[iso]
if not entries:
continue
out_path = OUTPUT_DIR / f"{iso}.tsv"
# If file already exists (e.g., from WikiPron), append new entries
existing = set()
if out_path.exists():
with open(out_path, encoding="utf-8", newline="") as f:
reader = csv.reader(f, delimiter="\t")
next(reader, None) # skip header
for row in reader:
if len(row) >= 2:
existing.add((row[0], row[1]))
mode = "a" if out_path.exists() else "w"
with open(out_path, mode, encoding="utf-8", newline="") as f:
if mode == "w":
f.write(HEADER)
new_count = 0
for word, ipa, sca, source, concept_id, cognate_set_id in entries:
if (word, ipa) not in existing:
f.write(f"{word}\t{ipa}\t{sca}\t{source}\t{concept_id}\t{cognate_set_id}\n")
existing.add((word, ipa))
new_count += 1
total = len(existing)
stats[iso] = total
if new_count > 0 and total >= 100:
pass # Don't spam output for small languages
return stats
def main():
print("=" * 80)
print("Full CLDF Extraction (All Concepts)")
print("=" * 80)
lang_entries: dict[str, list] = defaultdict(list)
print("\nExtracting from CLDF sources...")
total = 0
for name, extractor in [
("NorthEuraLex", extract_northeuralex),
("WOLD", extract_wold),
("ABVD", extract_abvd),
("Sino-Tibetan", extract_sinotibetan),
]:
print(f"\n [{name}]")
count = extractor(lang_entries)
total += count
print(f"\n{'=' * 80}")
print(f"Writing lexicon files...")
stats = write_lexicons(lang_entries)
# Summary
print(f"\n{'=' * 80}")
print(f"SUMMARY")
print(f"{'=' * 80}")
print(f" CLDF entries extracted: {total:,}")
print(f" Languages written: {len(stats)}")
total_written = sum(stats.values())
print(f" Total entries in files: {total_written:,}")
# Write stats
METADATA_DIR.mkdir(parents=True, exist_ok=True)
stats_path = METADATA_DIR / "cldf_stats.tsv"
with open(stats_path, "w", encoding="utf-8", newline="") as f:
f.write("ISO\tEntries\tSource\n")
for iso, count in sorted(stats.items(), key=lambda x: -x[1]):
f.write(f"{iso}\t{count}\tcldf\n")
print(f" Stats: {stats_path}")
print("Done!")
if __name__ == "__main__":
main()