#!/usr/bin/env python3
"""Parser for DSAL DEDR (Dravidian Etymological Dictionary Revised).
Extracts Proto-Dravidian reconstructions with glosses from the
University of Chicago DSAL digital edition:
https://dsal.uchicago.edu/dictionaries/burrow/
Entries are organized by entry number and typically contain:
- Proto-Dravidian reconstruction (starred form)
- Cognates across Dravidian languages (Tamil, Telugu, Kannada, etc.)
- English gloss
Uses only stdlib (urllib, html.parser, re).
Reference: Burrow & Emeneau, "A Dravidian Etymological Dictionary" (1984, revised)
"""
from __future__ import annotations
import logging
import re
import urllib.request
import urllib.error
from html.parser import HTMLParser
from typing import Any
logger = logging.getLogger(__name__)
class DEDRPageParser(HTMLParser):
"""Parse DEDR entry pages to extract reconstructions and glosses."""
def __init__(self) -> None:
super().__init__()
self.entries: list[dict] = []
self.in_entry_div = False
self.in_bold = False
self.entry_text = ""
self.bold_text = ""
self.current_entry_text = ""
self._tag_stack: list[str] = []
def handle_starttag(self, tag: str, attrs: list[tuple[str, str | None]]) -> None:
attr_dict = dict(attrs)
cls = attr_dict.get("class", "") or ""
self._tag_stack.append(tag)
if any(kw in cls.lower() for kw in ("entry", "hw", "result", "body")):
self.in_entry_div = True
self.current_entry_text = ""
if tag in ("b", "strong"):
self.in_bold = True
self.bold_text = ""
def handle_endtag(self, tag: str) -> None:
if self._tag_stack and self._tag_stack[-1] == tag:
self._tag_stack.pop()
if tag in ("b", "strong") and self.in_bold:
self.in_bold = False
if tag in ("div", "p", "br") and self.in_entry_div:
if self.current_entry_text.strip():
self._process_entry_text(self.current_entry_text)
self.current_entry_text = ""
def handle_data(self, data: str) -> None:
if self.in_bold:
self.bold_text += data
if self.in_entry_div:
self.current_entry_text += data
def _process_entry_text(self, text: str) -> None:
"""Process accumulated entry text to extract reconstruction + gloss."""
text = text.strip()
if not text or len(text) < 5:
return
# Look for reconstruction pattern: *form or starred form
m = re.search(r"\*([A-Za-zÀ-žḀ-ỿ\-āēīōūṭḍṇṛḷṃṁ]+)", text)
if m:
form = m.group(1).strip()
# Extract gloss: text after the form, typically after comma or dash
rest = text[m.end():]
gloss = ""
# Try various delimiters
gm = (
re.match(r"\s*[-–—,]\s+(.+?)(?:\.|;|$)", rest)
or re.match(r"\s+'(.+?)'", rest)
or re.match(r'\s+"(.+?)"', rest)
)
if gm:
gloss = gm.group(1).strip()
if form and gloss and len(form) < 40:
self.entries.append({
"word": form,
"transliteration": form,
"gloss": gloss,
})
def _regex_extract(html: str) -> list[dict]:
"""Extract DEDR entries using regex patterns on raw HTML.
DEDR entries typically look like:
1234 Ta. word meaning. ...
or:
1234 *proto-form - meaning
"""
entries: list[dict] = []
# Strip HTML tags but keep structure hints
text = html
# Replace
and
with newlines
text = re.sub(r"
", "\n", text, flags=re.IGNORECASE)
text = re.sub(r"?p[^>]*>", "\n", text, flags=re.IGNORECASE)
# Strip remaining tags
text = re.sub(r"<[^>]+>", "", text)
# Decode entities
text = text.replace("&", "&")
text = text.replace("<", "<")
text = text.replace(">", ">")
text = text.replace(" ", " ")
# Pattern 1: Entry number + reconstruction
# e.g. "1234 *kāl- leg, foot"
pattern1 = re.compile(
r"(\d{1,5})\s+\*([A-Za-zÀ-žḀ-ỿāēīōūṭḍṇṛḷṃṁ\-]+)"
r"\s*[-–—,]?\s+"
r"([A-Za-z][A-Za-z\s,'-]{2,80}?)(?=[.;]|\n|$)"
)
for m in pattern1.finditer(text):
entry_num = m.group(1)
form = m.group(2).strip()
gloss = m.group(3).strip()
gloss = re.sub(r"[,;:\s]+$", "", gloss)
if form and gloss:
entries.append({
"word": form,
"transliteration": form,
"gloss": gloss,
})
# Pattern 2: Tamil/Dravidian language abbreviation + word + meaning
# e.g. "Ta. kāl foot, leg"
if not entries:
pattern2 = re.compile(
r"(?:Ta|Te|Ka|Ma|Tu|Ko|Go|Ku|Mal|Kur|Br)\.\s+"
r"([A-Za-zÀ-žḀ-ỿāēīōūṭḍṇṛḷṃṁ\-]+)"
r"\s+"
r"([A-Za-z][A-Za-z\s,'-]{2,60}?)(?=[.;]|\n|$)"
)
for m in pattern2.finditer(text):
form = m.group(1).strip()
gloss = m.group(2).strip()
gloss = re.sub(r"[,;:\s]+$", "", gloss)
if form and gloss and len(form) < 30:
entries.append({
"word": form,
"transliteration": form,
"gloss": gloss,
})
return entries
def parse(url: str, **kwargs: Any) -> list[dict]:
"""Download and parse DEDR dictionary entries.
Args:
url: URL to DEDR, e.g.
https://dsal.uchicago.edu/dictionaries/burrow/
**kwargs:
entry_range: Tuple (start, end) of entry numbers to fetch.
Default fetches the main page only.
max_pages: Maximum number of pages to fetch (default 10).
Returns:
List of dicts with keys: word, transliteration, gloss.
Returns empty list if URL is unreachable.
"""
# No artificial page limit — follow ALL pagination links
max_pages = kwargs.get("max_pages", 9999)
logger.info("DEDR: downloading %s", url)
try:
req = urllib.request.Request(url, headers={"User-Agent": "PhaiPhon/1.0"})
with urllib.request.urlopen(req, timeout=30) as resp:
html = resp.read().decode("utf-8", errors="replace")
except (urllib.error.URLError, urllib.error.HTTPError, OSError) as exc:
logger.warning("DEDR: failed to download %s: %s", url, exc)
return []
# Try structured parsing
parser = DEDRPageParser()
parser.feed(html)
entries = parser.entries
# Fallback to regex
if not entries:
logger.info("DEDR: structured parsing found 0 entries, trying regex")
entries = _regex_extract(html)
# Follow pagination links if present
page_links: set[str] = set()
for m in re.finditer(r'href="([^"]*(?:page=\d+|/\d+/?)[^"]*)"', html, re.IGNORECASE):
href = m.group(1)
if href.startswith("/"):
domain_match = re.match(r"(https?://[^/]+)", url)
if domain_match:
href = domain_match.group(1) + href
elif not href.startswith("http"):
href = url.rstrip("/") + "/" + href
page_links.add(href)
pages_fetched = 0
for page_url in sorted(page_links):
if pages_fetched >= max_pages:
break
logger.info("DEDR: downloading page %s", page_url)
try:
req = urllib.request.Request(page_url, headers={"User-Agent": "PhaiPhon/1.0"})
with urllib.request.urlopen(req, timeout=30) as resp:
page_html = resp.read().decode("utf-8", errors="replace")
except (urllib.error.URLError, urllib.error.HTTPError, OSError) as exc:
logger.warning("DEDR: failed to download page %s: %s", page_url, exc)
continue
page_entries = _regex_extract(page_html)
entries.extend(page_entries)
pages_fetched += 1
# Deduplicate
seen: set[tuple[str, str]] = set()
unique: list[dict] = []
for e in entries:
key = (e["word"], e["gloss"])
if key not in seen:
seen.add(key)
unique.append(e)
logger.info("DEDR: extracted %d unique entries total", len(unique))
return unique
if __name__ == "__main__":
import sys
logging.basicConfig(level=logging.INFO, format="%(levelname)s: %(message)s")
test_url = (
sys.argv[1] if len(sys.argv) > 1
else "https://dsal.uchicago.edu/dictionaries/burrow/"
)
results = parse(test_url)
print(f"\nExtracted {len(results)} entries:")
for entry in results[:15]:
print(f" *{entry['word']:25s} {entry['gloss']}")
if len(results) > 15:
print(f" ... and {len(results) - 15} more")