ground-zero / app.py
jefffffff9
Align Whisper default to turbo-v3 + add document upload to Knowledge Base tab
cc50efb
"""
Sahel-Agri Voice AI — HuggingFace Spaces (ZeroGPU)
Two-way voice assistant: Bambara / Fula / French / English → voice response
Environment variables (set in Space Settings → Secrets):
HF_TOKEN — HF write-access token
FEEDBACK_REPO_ID — e.g. ous-sow/sahel-agri-feedback (dataset, private)
ADAPTER_REPO_ID — e.g. ous-sow/sahel-agri-adapters (model, private)
WHISPER_MODEL_ID — default: openai/whisper-large-v3-turbo
LLM_MODEL_ID — default: Qwen/Qwen2.5-72B-Instruct
KAGGLE_USERNAME — Kaggle username (for auto-trigger training)
KAGGLE_KEY — Kaggle API key (for auto-trigger training)
KAGGLE_KERNEL_SLUG — default: ous-sow/sahel-voice-master-trainer
AUTO_TRAIN_THRESHOLD — corrections count that triggers auto-training (default: 50)
"""
from __future__ import annotations
import io
import json
import os
import sys
import threading
from datetime import datetime, timezone
from pathlib import Path
import gradio as gr
import numpy as np
ROOT = Path(__file__).parent
sys.path.insert(0, str(ROOT))
# ── env ───────────────────────────────────────────────────────────────────────
HF_TOKEN = os.environ.get("HF_TOKEN")
FEEDBACK_REPO_ID = os.environ.get("FEEDBACK_REPO_ID", "ous-sow/sahel-agri-feedback")
ADAPTER_REPO_ID = os.environ.get("ADAPTER_REPO_ID", "ous-sow/sahel-agri-adapters")
# whisper-large-v3-turbo: 128 mel bins, matches fine-tuned adapters trained on turbo.
# whisper-small uses 80 mel bins — mismatches turbo adapters with a channel error.
# Override via WHISPER_MODEL_ID env var in Space settings if needed.
WHISPER_MODEL_ID = os.environ.get("WHISPER_MODEL_ID", "openai/whisper-large-v3-turbo")
LLM_MODEL_ID = os.environ.get("LLM_MODEL_ID", "Qwen/Qwen2.5-7B-Instruct")
KAGGLE_USERNAME = os.environ.get("KAGGLE_USERNAME", "")
KAGGLE_KEY = os.environ.get("KAGGLE_KEY", "")
KAGGLE_KERNEL_SLUG = os.environ.get("KAGGLE_KERNEL_SLUG", "ous-sow/sahel-voice-master-trainer")
AUTO_TRAIN_THRESHOLD = int(os.environ.get("AUTO_TRAIN_THRESHOLD", "50"))
# On local CPU (no HF_TOKEN / no spaces package) fall back gracefully
_ON_SPACES = os.environ.get("SPACE_ID") is not None
SUPPORTED_LANGUAGES = {
"Bambara — Mali (bam)": "bam",
"Fula / Pular — Guinea (ful)": "ful",
"French / Français": "fr",
"English": "en",
}
# Country and dialect context used in prompts and training metadata
LANG_CONTEXT = {
"bam": {
"name": "Bambara",
"country": "Mali",
"region": "West Africa (Bamako, Ségou, Mopti dialects)",
"script": "Latin with special characters (ɛ, ɔ, ŋ, ɲ)",
"phonetic_note": (
"Use standard Malian orthography: 'u' not 'ou', 'j' not 'dj', "
"'c' not 'ch', 'ɲ' not 'gn' or 'ny', 'ɔ' not 'oo', 'ɛ' not 'ee'. "
"This is Bambara as spoken in Mali, NOT Dioula or other dialects."
),
"do_not_mix": "Fula (Pulaar/Pular), Wolof, Dioula, or any other language",
},
"ful": {
"name": "Pular (Fula of Guinea)",
"country": "Guinea",
"region": "West Africa (Labé, Mamou, Kankan dialects)",
"script": "Latin with special characters (ɓ, ɗ, ŋ, ɲ, ƴ)",
"phonetic_note": (
"Use standard Guinean Pular orthography. "
"This is the Fula variety spoken in Guinea (Pular/Pulaar), "
"NOT Fulfulde from Niger/Nigeria nor Wolof."
),
"do_not_mix": "Bambara, Soussou, Malinké, or any other language",
},
"fr": {
"name": "French",
"country": "France / West Africa",
"region": "",
"script": "Latin",
"phonetic_note": "Standard French.",
"do_not_mix": "other languages unless the user switches",
},
"en": {
"name": "English",
"country": "",
"region": "",
"script": "Latin",
"phonetic_note": "Standard English.",
"do_not_mix": "other languages unless the user switches",
},
}
# ── ZeroGPU decorator (no-op locally) ────────────────────────────────────────
try:
import spaces # type: ignore
_gpu = spaces.GPU(duration=55)
except ImportError:
def _gpu(fn): # local fallback: plain function
return fn
# ── Module-level model state (CPU-resident between requests) ─────────────────
_whisper_model = None # WhisperForConditionalGeneration (base)
_whisper_processor = None
_fine_tuned_models = {} # lang_code -> WhisperForConditionalGeneration (full checkpoint)
_fine_tuned_processors = {} # lang_code -> WhisperProcessor matching the fine-tuned checkpoint
_model_lock = threading.Lock()
_model_status = "not loaded"
_load_started_at: float = 0.0 # monotonic time when loading began
_LOAD_TIMEOUT = 180 # seconds before declaring a stuck load
# ── Conversation-mode state ───────────────────────────────────────────────────
_voice_ref_path: str | None = None # path to 24 kHz WAV converted from user MP3
_voice_ref_text: str = "" # auto-transcribed text of reference audio
_llm_client = None # GemmaClient, lazy init
from src.tts.mms_tts import MMSTTSEngine
from src.iot.intent_parser import IntentParser
from src.iot.sensor_bridge import SensorBridge
from src.iot.voice_responder import VoiceResponder
from src.conversation.phrase_matcher import PhraseMatcher
from src.llm.gemma_client import GemmaClient
from src.data.bam_normalize import normalize as bam_normalize
from src.data.adlam import normalize_pular
_tts = MMSTTSEngine()
_intent_parser = IntentParser()
_sensor_bridge = SensorBridge()
_phrase_matcher = PhraseMatcher()
# HF API — only instantiate when token present
_hf_api = None
if HF_TOKEN:
from huggingface_hub import HfApi
_hf_api = HfApi(token=HF_TOKEN)
# ── Model loading ─────────────────────────────────────────────────────────────
def _do_load_whisper():
global _whisper_model, _whisper_processor, _model_status
try:
import torch
try:
from transformers.models.whisper import WhisperProcessor, WhisperForConditionalGeneration
except ImportError:
from transformers.models.whisper.processing_whisper import WhisperProcessor
from transformers.models.whisper.modeling_whisper import WhisperForConditionalGeneration
_model_status = "loading…"
_whisper_processor = WhisperProcessor.from_pretrained(
WHISPER_MODEL_ID, token=HF_TOKEN
)
try:
_whisper_model = WhisperForConditionalGeneration.from_pretrained(
WHISPER_MODEL_ID,
torch_dtype=torch.float32,
token=HF_TOKEN,
)
except TypeError:
_whisper_model = WhisperForConditionalGeneration.from_pretrained(
WHISPER_MODEL_ID,
token=HF_TOKEN,
)
_whisper_model.eval()
_model_status = f"ready ({WHISPER_MODEL_ID})"
except Exception as e:
_model_status = f"error: {e}"
def _ensure_whisper_loaded():
"""Load Whisper to CPU in a background thread on first call. Non-blocking."""
import time
global _model_status, _load_started_at
with _model_lock:
need_start = False
if _whisper_model is None and "loading" not in _model_status:
need_start = True
elif (_whisper_model is None
and "loading" in _model_status
and _load_started_at
and (time.monotonic() - _load_started_at) > _LOAD_TIMEOUT):
_model_status = "error: load timed out after %ds — retrying" % _LOAD_TIMEOUT
need_start = True
if need_start:
_model_status = "loading…"
_load_started_at = time.monotonic()
t = threading.Thread(target=_do_load_whisper, daemon=True)
t.start()
return _model_status
def _wait_for_whisper(timeout: int = 120) -> bool:
"""
Block until Whisper is loaded or timeout (seconds) expires.
Triggers loading if not already started. Returns True if model is ready.
"""
import time
_ensure_whisper_loaded()
deadline = time.monotonic() + timeout
while time.monotonic() < deadline:
if _whisper_model is not None:
return True
if "error" in _model_status:
return False
time.sleep(0.5)
return False
def get_model_status() -> str:
s = _ensure_whisper_loaded()
if "ready" in s:
return f"🟢 {s}"
if "loading" in s:
return f"🟡 {s}"
if "error" in s:
return f"🔴 {s}"
return f"⚪ {s}"
# ── Core GPU pipeline ─────────────────────────────────────────────────────────
@_gpu
def _run_pipeline(audio_path: str, language_code: str):
"""
Full STT → Intent → Sensor → TTS pipeline.
Decorated with @spaces.GPU(duration=55) on HF Spaces; plain function locally.
Returns: (transcript, response_text, (sample_rate, wav_np))
"""
import asyncio
import torch
device = "cuda" if torch.cuda.is_available() else "cpu"
# ── 1. Whisper STT ────────────────────────────────────────────────────────
if _whisper_model is None:
return "⏳ Model still loading…", "", None
import librosa
audio_np, _ = librosa.load(audio_path, sr=16000, mono=True)
# Use fine-tuned checkpoint for this language if one has been loaded;
# otherwise fall back to base Whisper.
active_model = _fine_tuned_models.get(language_code, _whisper_model)
# Use the matching processor so mel-bin count is consistent with the model
# (large-v3-turbo = 128 bins; small = 80 bins). Mismatch causes channel error.
active_processor = _fine_tuned_processors.get(language_code, _whisper_processor)
active_model.to(device)
with _model_lock:
inputs = active_processor.feature_extractor(
audio_np, sampling_rate=16000, return_tensors="pt"
)
input_features = inputs.input_features.to(device)
# Bambara and Fula have no Whisper language token — pass None so the model
# auto-detects or falls back to multilingual decoding.
if language_code in ("bam", "ful"):
forced_ids = None
else:
forced_ids = active_processor.get_decoder_prompt_ids(
language=language_code, task="transcribe"
)
with torch.no_grad():
predicted_ids = active_model.generate(
input_features,
forced_decoder_ids=forced_ids if forced_ids else None,
max_new_tokens=256,
)
transcript = active_processor.batch_decode(
predicted_ids, skip_special_tokens=True
)[0].strip()
# Free GPU VRAM before TTS
active_model.to("cpu")
if device == "cuda":
torch.cuda.empty_cache()
# ── 2. Phrase library (general conversation — no sensors needed) ─────────
phrase_match = _phrase_matcher.match(transcript, language_code)
if phrase_match:
response_text = phrase_match["response"]
english_translation = phrase_match["english"]
else:
# ── 3. Intent + sensor data (agricultural queries) ────────────────────
intent = _intent_parser.parse(transcript, language=language_code)
try:
loop = asyncio.new_event_loop()
sensor_data = loop.run_until_complete(_sensor_bridge.fetch(intent))
loop.close()
except Exception:
from src.iot.sensor_bridge import SensorData
sensor_data = SensorData(sensor_type="soil", values={
"moisture_pct": 45.0, "ph": 6.5, "temperature_c": 28.0
})
responder = VoiceResponder(language=language_code)
response_text, english_translation = responder.generate_response(intent, sensor_data)
# Low-confidence fallback: say "I didn't understand" in the native language
if intent.action == "unknown" and intent.confidence < 0.15:
from src.iot.voice_responder import BAMBARA_TEMPLATES, FULA_TEMPLATES
if language_code == "bam":
response_text, english_translation = BAMBARA_TEMPLATES["not_understood"]
elif language_code == "ful":
response_text, english_translation = FULA_TEMPLATES["not_understood"]
# ── 3. MMS-TTS (GPU) ──────────────────────────────────────────────────────
wav_np, sample_rate = _tts.synthesize(response_text, language_code, device=device)
return transcript, english_translation, response_text, (sample_rate, wav_np)
# ── Conversation-mode helpers ─────────────────────────────────────────────────
# Vocabulary context cache — loaded from Hub, refreshed after each LEARNED save
_vocab_context_cache: str = ""
_vocab_lock = threading.Lock()
def _refresh_vocab_context() -> None:
"""Load vocabulary.jsonl from Hub and rebuild the LLM context string."""
global _vocab_context_cache
if not HF_TOKEN or not FEEDBACK_REPO_ID:
return
try:
from huggingface_hub import hf_hub_download
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="vocabulary.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
entries: list[dict] = []
with open(local, encoding="utf-8") as f:
for line in f:
line = line.strip()
if line:
try:
entries.append(json.loads(line))
except Exception:
pass
# Most recent first, cap at 200 entries to stay within token budget
entries = entries[-200:][::-1]
lines = []
for e in entries:
word = e.get("word", "").strip()
tr = e.get("translation", "").strip()
lang = e.get("language", "")
if word:
lines.append(f"{word} = {tr} [{lang}]" if tr else f"{word} [{lang}]")
with _vocab_lock:
_vocab_context_cache = "\n".join(lines)
except Exception:
pass # Non-critical — LLM continues without vocab context
def _get_vocab_context() -> str:
with _vocab_lock:
return _vocab_context_cache
def _save_learned_async(word: str, meaning: str, lang: str) -> None:
"""Persist a word/phrase learned mid-conversation to vocabulary.jsonl on Hub."""
def _run():
if not word.strip():
return
entry = {"word": word.strip(), "translation": meaning.strip(), "language": lang,
"source": "conversation", "timestamp": datetime.now(timezone.utc).isoformat()}
_upload_jsonl_later("vocabulary.jsonl", [entry])
_refresh_vocab_context() # update cache so next turn knows this word
threading.Thread(target=_run, daemon=True).start()
def _upload_jsonl_later(repo_path: str, entries: list[dict]) -> None:
"""Append entries to a Hub JSONL file — called from background threads."""
if not HF_TOKEN or not FEEDBACK_REPO_ID or _hf_api is None:
return
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename=repo_path,
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
updated = existing + "".join(json.dumps(e, ensure_ascii=False) + "\n" for e in entries)
try:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(updated.encode("utf-8")),
path_in_repo=repo_path,
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
return
except Exception:
if attempt == 1:
pass
import re as _re
_LEARNED_RE = _re.compile(
r'\[LEARNED:\s*word=["\'](.+?)["\']\s+meaning=["\'](.+?)["\']\s*\]',
_re.IGNORECASE,
)
def _parse_and_strip_learned(text: str, lang: str) -> tuple[str, list[tuple[str, str]]]:
"""
Extract [LEARNED: word="X" meaning="Y"] tags from LLM output.
Returns (cleaned_text, list_of_(word, meaning) pairs).
Saves each pair to Hub asynchronously.
"""
learned = []
for m in _LEARNED_RE.finditer(text):
word, meaning = m.group(1).strip(), m.group(2).strip()
learned.append((word, meaning))
_save_learned_async(word, meaning, lang)
cleaned = _LEARNED_RE.sub("", text).strip()
return cleaned, learned
# System prompt — includes vocabulary context + conversation rules
def _build_system_prompt(language_code: str, vocab: str) -> str:
"""
Build a language-specific system prompt that makes the LLM stay strictly
in the correct dialect (Mali Bambara vs Guinea Pular) and never mix them.
"""
ctx = LANG_CONTEXT.get(language_code, LANG_CONTEXT["fr"])
lang_name = ctx["name"]
country = ctx["country"]
region = ctx["region"]
phon_note = ctx["phonetic_note"]
do_not_mix = ctx["do_not_mix"]
region_line = f" ({region})" if region else ""
vocab_section = (
f"WORDS AND PHRASES YOU HAVE LEARNED FOR {lang_name.upper()}:\n{vocab}"
if vocab
else f"(No {lang_name} vocabulary recorded yet — the user can teach you words.)"
)
return f"""\
You are a voice assistant that speaks ONLY {lang_name} as used in {country}{region_line}.
CRITICAL LANGUAGE RULE:
- You MUST respond exclusively in {lang_name} ({country}).
- NEVER mix in words from {do_not_mix}.
- If the user writes in another language, gently ask them to switch to {lang_name}.
- If you are unsure of a word in {lang_name}, say so honestly — do not substitute \
a word from another language.
ORTHOGRAPHY ({lang_name}):
{phon_note}
{vocab_section}
CONVERSATION RULES:
1. Keep every response to 1–3 short spoken sentences. This is voice, not text.
2. If you do not understand, ask ONE short follow-up question in {lang_name}.
3. If the user teaches you a word ("X means Y"), confirm warmly, then append \
exactly: [LEARNED: word="X" meaning="Y"]
4. Refer back to earlier messages naturally when relevant.
5. Never invent vocabulary. Honest uncertainty is always correct."""
def _get_vocab_context_for(language_code: str) -> str:
"""Return only vocabulary entries for the given language code."""
with _vocab_lock:
raw = _vocab_context_cache
if not raw:
return ""
lines = [
line for line in raw.splitlines()
if f"[{language_code}]" in line
]
return "\n".join(lines)
def _build_messages(user_text: str, history: list, language_code: str) -> list[dict]:
"""Build the full message list: system (with lang-filtered vocab) + history + new turn."""
vocab = _get_vocab_context_for(language_code)
system = _build_system_prompt(language_code, vocab)
messages: list[dict] = [{"role": "system", "content": system}]
for u, a in history[-20:]:
messages.append({"role": "user", "content": u})
messages.append({"role": "assistant", "content": a})
messages.append({"role": "user", "content": user_text})
return messages
def set_voice_reference(audio_file) -> str:
"""
Store an uploaded audio file as the TTS voice reference.
Converts to 24 kHz WAV (F5-TTS requirement) and auto-transcribes.
Returns a status string for the UI.
"""
global _voice_ref_path, _voice_ref_text
if audio_file is None:
_voice_ref_path = None
_voice_ref_text = ""
return "🗑️ Voice reference cleared — using default MMS-TTS voice."
try:
from src.tts.f5_tts import to_wav_24k
wav_path = to_wav_24k(audio_file)
_voice_ref_path = wav_path
# Auto-transcribe using already-loaded Whisper if available
if _whisper_model is not None and _whisper_processor is not None:
import torch, librosa
audio_np, _ = librosa.load(wav_path, sr=16000, mono=True)
with _model_lock:
inputs = _whisper_processor.feature_extractor(
audio_np, sampling_rate=16000, return_tensors="pt"
)
with torch.no_grad():
ids = _whisper_model.generate(
inputs.input_features,
max_new_tokens=128,
)
_voice_ref_text = _whisper_processor.batch_decode(
ids, skip_special_tokens=True
)[0].strip()
return (
f"✅ Voice reference set!\n"
f"File : {Path(audio_file).name}\n"
f"Transcript : {_voice_ref_text[:80] or '(empty — F5-TTS will use in-context inference)'}"
)
else:
_voice_ref_text = ""
return (
f"✅ Voice reference set (model not loaded yet — transcript pending).\n"
f"File: {Path(audio_file).name}"
)
except Exception as exc:
return f"❌ Could not process reference audio: {exc}"
def _convo_pipeline(audio_path: str, language_code: str, history: list):
"""
Full S2S conversation pipeline with memory:
1. ASR — fine-tuned Whisper (or base) → transcript
2. Norm — bam_normalize() on Bambara text
3. Brain — LLM with full conversation history + vocabulary context
4. Learn — parse [LEARNED:] tags, persist to Hub async
5. Mouth — F5-TTS (voice ref) or MMS-TTS fallback → audio
Returns: (transcript, eng, response_text, audio_out, new_history)
"""
import torch
import logging
log = logging.getLogger(__name__)
device = "cuda" if torch.cuda.is_available() else "cpu"
if _whisper_model is None:
return "⏳ Model still loading…", "", "", None, history
import librosa
audio_np, _ = librosa.load(audio_path, sr=16000, mono=True)
active_model = _fine_tuned_models.get(language_code, _whisper_model)
active_processor = _fine_tuned_processors.get(language_code, _whisper_processor)
active_model.to(device)
with _model_lock:
inputs = active_processor.feature_extractor(
audio_np, sampling_rate=16000, return_tensors="pt"
)
input_features = inputs.input_features.to(device)
forced_ids = None
if language_code not in ("bam", "ful"):
forced_ids = active_processor.get_decoder_prompt_ids(
language=language_code, task="transcribe"
)
with torch.no_grad():
predicted_ids = active_model.generate(
input_features,
forced_decoder_ids=forced_ids if forced_ids else None,
max_new_tokens=256,
)
transcript = active_processor.batch_decode(
predicted_ids, skip_special_tokens=True
)[0].strip()
active_model.to("cpu")
if device == "cuda":
torch.cuda.empty_cache()
# Phonetic normalisation (Bambara: French spellings → standard; Fula: Adlam → Latin)
if language_code == "bam":
normalised = bam_normalize(transcript)
elif language_code == "ful":
normalised = normalize_pular(transcript)
else:
normalised = transcript
# ── LLM brain — full context: vocab + history + new turn ─────────────────
response_text = ""
try:
from huggingface_hub import InferenceClient
client = InferenceClient(token=HF_TOKEN)
messages = _build_messages(normalised, history, language_code)
completion = client.chat_completion(
model=LLM_MODEL_ID,
messages=messages,
max_tokens=300,
temperature=0.65,
)
response_text = completion.choices[0].message.content.strip()
except Exception as llm_err:
log.warning("LLM failed: %s", llm_err)
# Graceful degradation: tell user LLM is unavailable, ask them to try again
_fallbacks = {
"bam": "Hakɛ to, n bɛ sɔrɔ cogo dɔ la.", # Bambara (Mali)
"ful": "Hakke, mi waawaa jogaade modèl oo jooni.", # Pular (Guinea)
"fr": "Désolé, je n'ai pas pu joindre le modèle.",
}
response_text = _fallbacks.get(language_code, "Sorry, the language model is unavailable.")
# ── Parse and strip [LEARNED:] tags — save async to Hub ──────────────────
response_text, learned_pairs = _parse_and_strip_learned(response_text, language_code)
if learned_pairs:
log.info("Learned %d new item(s): %s", len(learned_pairs), learned_pairs)
# ── Update conversation history ───────────────────────────────────────────
new_history = list(history) + [(normalised, response_text)]
if len(new_history) > 20:
new_history = new_history[-20:]
# ── TTS mouth — F5-TTS (voice ref) or MMS-TTS fallback ───────────────────
audio_out = None
if _voice_ref_path and Path(_voice_ref_path).exists():
try:
from src.tts.f5_tts import synthesize as f5_synthesize
result = f5_synthesize(
response_text,
ref_wav_path=_voice_ref_path,
ref_text=_voice_ref_text,
device=device,
)
if result is not None:
wav_np, sr = result
audio_out = (sr, wav_np)
except Exception as tts_err:
log.warning("F5-TTS failed, using MMS-TTS: %s", tts_err)
if audio_out is None:
wav_np, sr = _tts.synthesize(response_text, language_code, device=device)
audio_out = (sr, wav_np)
return transcript, "", response_text, audio_out, new_history
# ── HF Hub feedback persistence ───────────────────────────────────────────────
def _save_feedback_to_hub(
audio_path: str | None,
transcript: str,
corrected_text: str,
english_translation: str,
corrected_english: str,
response_text: str,
corrected_response: str,
rating: int,
notes: str,
language_label: str,
) -> str:
language_code = SUPPORTED_LANGUAGES.get(language_label, "bam")
if not corrected_text.strip():
return "⚠️ Corrected transcription is empty — please fill in what was actually said."
timestamp = datetime.now(timezone.utc).strftime("%Y%m%d_%H%M%S_%f")
record = {
"id": timestamp,
"timestamp": datetime.now(timezone.utc).isoformat(),
"language": language_code,
"audio_file": f"audio/{language_code}_{timestamp}.wav",
"whisper_output": transcript,
"corrected_text": corrected_text.strip(),
"english_translation": english_translation.strip(),
"corrected_english": corrected_english.strip() or english_translation.strip(),
"response_text": response_text,
"corrected_response": corrected_response.strip() or response_text.strip(),
"rating": rating,
"notes": notes.strip(),
"is_correction": transcript.strip() != corrected_text.strip(),
"model": WHISPER_MODEL_ID,
}
if _hf_api is None:
# Local: save to disk instead
fb_dir = ROOT / "feedback"
fb_dir.mkdir(exist_ok=True)
(fb_dir / "audio").mkdir(exist_ok=True)
corrections_path = fb_dir / "corrections.jsonl"
if audio_path:
import shutil
shutil.copy2(audio_path, fb_dir / "audio" / f"{language_code}_{timestamp}.wav")
with open(corrections_path, "a", encoding="utf-8") as f:
f.write(json.dumps(record, ensure_ascii=False) + "\n")
total = sum(1 for _ in open(corrections_path, encoding="utf-8"))
return f"✅ Saved locally (#{total}) — HF_TOKEN not set, Hub upload skipped."
try:
# Upload audio
if audio_path:
_hf_api.upload_file(
path_or_fileobj=audio_path,
path_in_repo=f"audio/{language_code}_{timestamp}.wav",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
# Download → append → re-upload corrections.jsonl (with retry on conflict)
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local_jsonl = hf_hub_download(
repo_id=FEEDBACK_REPO_ID,
filename="corrections.jsonl",
repo_type="dataset",
token=HF_TOKEN,
)
with open(local_jsonl, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
updated = existing + json.dumps(record, ensure_ascii=False) + "\n"
buf = io.BytesIO(updated.encode("utf-8"))
try:
_hf_api.upload_file(
path_or_fileobj=buf,
path_in_repo="corrections.jsonl",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
break
except Exception as e:
if attempt == 1:
return f"⚠️ Audio uploaded but corrections.jsonl update failed: {e}"
total = updated.count("\n")
_maybe_auto_trigger()
return f"✅ Saved to Hub (#{total}) — {FEEDBACK_REPO_ID}"
except Exception as e:
return f"❌ Hub upload error: {e}"
# ── Adapter reload ────────────────────────────────────────────────────────────
def _reload_adapters_from_hub() -> str:
"""Download full fine-tuned checkpoints from Hub and hot-swap them into memory."""
global _fine_tuned_models, _fine_tuned_processors
if _hf_api is None:
return "⚠️ HF_TOKEN not set — cannot download checkpoints."
if _whisper_model is None:
return "⏳ Base model not loaded yet — wait for model to finish loading and try again."
try:
import torch
from huggingface_hub import snapshot_download
try:
from transformers.models.whisper.modeling_whisper import WhisperForConditionalGeneration
from transformers.models.whisper.processing_whisper import WhisperProcessor
except ImportError:
from transformers import WhisperForConditionalGeneration, WhisperProcessor
local_dir = snapshot_download(
repo_id=ADAPTER_REPO_ID, repo_type="model", token=HF_TOKEN
)
results = []
for lang, subdir in (("bam", "adapters/bambara"), ("ful", "adapters/fula")):
ckpt_path = Path(local_dir) / subdir
if not ckpt_path.exists():
results.append(f"⚠️ {lang}: `{subdir}` not found — run training notebook first")
continue
if not (ckpt_path / "config.json").exists():
results.append(f"⚠️ {lang}: `{subdir}/config.json` missing — incomplete checkpoint")
continue
try:
m = WhisperForConditionalGeneration.from_pretrained(
str(ckpt_path), torch_dtype=torch.float32
)
m.eval()
_fine_tuned_models[lang] = m
# Load the processor from the same checkpoint so the feature
# extractor mel-bin count matches the model (e.g. 128 for
# large-v3-turbo vs 80 for small). Mismatch causes:
# "expected input to have 128 channels, but got 80 channels"
_fine_tuned_processors[lang] = WhisperProcessor.from_pretrained(
str(ckpt_path)
)
results.append(f"✅ {lang}: fine-tuned checkpoint loaded from `{subdir}`")
except Exception as e:
results.append(f"❌ {lang}: load failed — {e}")
summary = "\n".join(results)
active = ", ".join(_fine_tuned_models) if _fine_tuned_models else "none"
return f"{summary}\n\n**Active fine-tuned models:** {active}\n**Repo:** `{ADAPTER_REPO_ID}`"
except Exception as e:
return f"❌ Checkpoint reload failed: {e}"
def _get_adapter_status() -> str:
lines = []
if _fine_tuned_models:
lines.append(f"**Fine-tuned models loaded:** {', '.join(sorted(_fine_tuned_models))}")
else:
lines.append("**Fine-tuned models:** none — using base Whisper for all languages")
if _hf_api is None:
lines.append("_HF_TOKEN not set — Hub check skipped._")
return "\n".join(lines)
try:
from huggingface_hub import list_repo_files
files = list(list_repo_files(ADAPTER_REPO_ID, repo_type="model", token=HF_TOKEN))
bam_ok = any("bambara" in f and "config.json" in f for f in files)
ful_ok = any("fula" in f and "config.json" in f for f in files)
lines += [
f"\n**Hub repo:** `{ADAPTER_REPO_ID}`",
f"- Bambara (bam): {'✅ trained checkpoint present' if bam_ok else '⚠️ not yet trained — run Kaggle notebook'}",
f"- Fula (ful): {'✅ trained checkpoint present' if ful_ok else '⚠️ not yet trained — run Kaggle notebook'}",
]
if bam_ok or ful_ok:
lines.append("\n_Click **Reload Models** to activate them._")
except Exception as e:
lines.append(f"_Could not read Hub repo: {e}_")
return "\n".join(lines)
# ── Knowledge Base handlers ───────────────────────────────────────────────────
def _import_phrase_pairs(lang_label: str, pairs_text: str) -> str:
"""Import pasted phrase pairs into the phrase library and append to vocabulary.jsonl."""
if not pairs_text.strip():
return "⚠️ Nothing entered. Use the format: native phrase | english translation"
lang = SUPPORTED_LANGUAGES.get(lang_label, "bam")
count = _phrase_matcher.import_pairs(lang, pairs_text)
if count == 0:
return "⚠️ No valid phrases found. Each line must contain a | separator.\nExample: I ni ce | Hello, good day"
_upload_phrase_additions_to_hub(lang)
# Also append to vocabulary.jsonl so the Kaggle training notebook picks them up
_append_phrases_to_vocabulary_jsonl(lang, pairs_text)
total = _phrase_matcher.phrase_count(lang)
return f"✅ Added {count} phrase(s) for {lang_label}. Library now has {total} phrases. Available immediately."
def _extract_text_from_document(file_path: str) -> str:
"""Extract plain text from a PDF, DOCX, or TXT file. Returns empty string on failure."""
ext = Path(file_path).suffix.lower()
try:
if ext == ".pdf":
from pypdf import PdfReader
reader = PdfReader(file_path)
return "\n".join((p.extract_text() or "") for p in reader.pages)
if ext in (".docx", ".doc"):
from docx import Document
doc = Document(file_path)
return "\n".join(para.text for para in doc.paragraphs if para.text.strip())
if ext in (".txt", ".md"):
with open(file_path, encoding="utf-8", errors="ignore") as f:
return f.read()
except Exception as exc:
import logging
logging.getLogger(__name__).warning("Document extract failed for %s: %s", file_path, exc)
return ""
def _sentences_from_text(text: str, min_words: int = 3, max_words: int = 25) -> list[str]:
"""Split extracted text into clean sentences suitable for vocabulary.jsonl."""
import re as _re
# Normalise whitespace and split on sentence boundaries (., !, ?, or double newline)
text = _re.sub(r"\s+", " ", text).strip()
raw = _re.split(r"(?<=[.!?])\s+|\n\n+", text)
out = []
seen = set()
for s in raw:
s = s.strip(" \t\"'`—–-")
if not s:
continue
words = s.split()
if not (min_words <= len(words) <= max_words):
continue
key = s.lower()
if key in seen:
continue
seen.add(key)
out.append(s)
return out
def _import_documents(lang_label: str, files: list, source_note: str) -> str:
"""Extract sentences from uploaded PDF/Word/TXT files and append to vocabulary.jsonl."""
if not files:
return "⚠️ Please upload at least one document first."
lang = SUPPORTED_LANGUAGES.get(lang_label, "bam")
# Language normalisation — same rule as other ingestion paths
total_sentences = 0
per_file_summary = []
all_entries: list[dict] = []
for f in files:
# Gradio File component returns a tempfile path (or an object with .name)
path = f if isinstance(f, str) else getattr(f, "name", None)
if not path:
continue
text = _extract_text_from_document(path)
if not text.strip():
per_file_summary.append(f" - {Path(path).name}: ⚠️ no text extracted")
continue
# Apply language-specific normalisation so Adlam → Latin etc.
try:
if lang == "ful":
text = normalize_pular(text)
elif lang == "bam":
text = bam_normalize(text)
except Exception:
pass
sentences = _sentences_from_text(text)
for s in sentences:
all_entries.append({
"word": s,
"translation": "",
"language": lang,
"source": f"document: {source_note or Path(path).name}",
})
per_file_summary.append(f" - {Path(path).name}: {len(sentences)} sentence(s)")
total_sentences += len(sentences)
if not all_entries:
return "⚠️ No usable sentences found in the uploaded document(s).\n" + "\n".join(per_file_summary)
# Append to vocabulary.jsonl on Hub (same pattern as _append_phrases_to_vocabulary_jsonl)
if _hf_api is not None and FEEDBACK_REPO_ID:
try:
from huggingface_hub import hf_hub_download
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="vocabulary.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
new_lines = "".join(json.dumps(e, ensure_ascii=False) + "\n" for e in all_entries)
_hf_api.upload_file(
path_or_fileobj=io.BytesIO((existing + new_lines).encode("utf-8")),
path_in_repo="vocabulary.jsonl",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
threading.Thread(target=_refresh_vocab_context, daemon=True).start()
except Exception as exc:
return f"⚠️ Extracted {total_sentences} sentence(s) but Hub upload failed: {exc}"
return (
f"✅ Imported {total_sentences} sentence(s) for {lang_label} from {len(files)} document(s).\n"
+ "\n".join(per_file_summary)
+ "\n\nThese will be used by the Kaggle training notebook on the next run."
)
def _append_phrases_to_vocabulary_jsonl(lang: str, pairs_text: str) -> None:
"""Append phrase pairs to vocabulary.jsonl in the feedback repo (training input)."""
if _hf_api is None or not FEEDBACK_REPO_ID:
return
entries = []
for line in pairs_text.splitlines():
if "|" not in line:
continue
parts = line.split("|", 1)
word = parts[0].strip()
translation = parts[1].strip() if len(parts) > 1 else ""
if word:
entries.append({"word": word, "translation": translation, "language": lang})
if not entries:
return
try:
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="vocabulary.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
new_lines = "".join(json.dumps(e, ensure_ascii=False) + "\n" for e in entries)
updated = existing + new_lines
try:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(updated.encode("utf-8")),
path_in_repo="vocabulary.jsonl",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
threading.Thread(target=_refresh_vocab_context, daemon=True).start()
break
except Exception:
if attempt == 1:
pass # Silent — phrase library still updated locally
except Exception:
pass # Non-critical — phrase library already saved via _upload_phrase_additions_to_hub
def _upload_phrase_additions_to_hub(lang: str) -> None:
"""Persist user phrase additions to HF Hub so they survive Space restarts."""
if _hf_api is None or not FEEDBACK_REPO_ID:
return
try:
import io
data = _phrase_matcher.get_additions_json(lang)
buf = io.BytesIO(data.encode("utf-8"))
_hf_api.upload_file(
path_or_fileobj=buf,
path_in_repo=f"phrase_additions/{lang}.json",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
except Exception as exc:
import logging
logging.getLogger(__name__).warning("Could not upload phrase additions: %s", exc)
def _load_phrase_additions_from_hub() -> None:
"""Download and merge user phrase additions from HF Hub at startup."""
if _hf_api is None or not FEEDBACK_REPO_ID:
return
for lang in ("bam", "ful"):
try:
from huggingface_hub import hf_hub_download
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID,
filename=f"phrase_additions/{lang}.json",
repo_type="dataset",
token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
data = f.read()
_phrase_matcher.reload_from_hub_data(lang, data)
except Exception:
pass # No additions saved yet — fine
# Load phrase additions + vocabulary context in background at startup
threading.Thread(target=_load_phrase_additions_from_hub, daemon=True).start()
threading.Thread(target=_refresh_vocab_context, daemon=True).start()
def _save_audio_for_training(lang_label: str, audio_path: str | None, transcript: str, source_note: str) -> str:
"""Save uploaded audio + transcription to corrections.jsonl so the Kaggle notebook picks it up."""
transcript = transcript.strip()
if audio_path is None:
return "⚠️ Please upload an audio file first."
if not transcript:
return "⚠️ Please type the transcription — what is said in this audio."
lang = SUPPORTED_LANGUAGES.get(lang_label, "bam")
timestamp = datetime.now(timezone.utc).strftime("%Y%m%d_%H%M%S_%f")
# Store under audio/ — same path structure that corrections.jsonl expects
audio_repo_path = f"audio/{lang}_{timestamp}.wav"
record = {
"id": timestamp,
"timestamp": datetime.now(timezone.utc).isoformat(),
"language": lang,
"audio_file": audio_repo_path,
"transcription": transcript, # notebook reads this field
"corrected_text": transcript, # also populate corrected_text for compatibility
"source": source_note.strip() or "uploaded",
"is_correction": False,
"model": WHISPER_MODEL_ID,
}
if _hf_api is None or not FEEDBACK_REPO_ID:
return "⚠️ HF_TOKEN not set — cannot upload to Hub."
try:
# Upload audio to audio/ (same bucket corrections use)
_hf_api.upload_file(
path_or_fileobj=audio_path,
path_in_repo=audio_repo_path,
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
# Append to corrections.jsonl (same file the notebook reads)
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local_jsonl = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="corrections.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local_jsonl, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
updated = existing + json.dumps(record, ensure_ascii=False) + "\n"
try:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(updated.encode("utf-8")),
path_in_repo="corrections.jsonl",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
break
except Exception as e:
if attempt == 1:
return f"⚠️ Audio uploaded but corrections.jsonl update failed: {e}"
total = updated.count("\n")
return (
f"✅ Saved to training dataset (#{total} total corrections)!\n"
f"Audio: {audio_repo_path}\n"
f"Transcription: {transcript[:80]}{'…' if len(transcript) > 80 else ''}\n"
f"Run the Kaggle notebook to include this in the next model update."
)
except Exception as exc:
return f"❌ Upload failed: {exc}"
# ── Auto-training trigger ─────────────────────────────────────────────────────
def _count_corrections() -> int:
"""Return number of entries in corrections.jsonl on the Hub."""
if _hf_api is None:
return 0
try:
from huggingface_hub import hf_hub_download
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="corrections.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
return sum(1 for l in f if l.strip())
except Exception:
return 0
def _trigger_kaggle_training(lang: str = "bam") -> str:
"""
Push the master trainer notebook to Kaggle, creating a new kernel version
(i.e. triggering a run). Requires KAGGLE_USERNAME + KAGGLE_KEY secrets.
Tries the Python API first (no PATH issues), falls back to subprocess.
"""
if not KAGGLE_USERNAME or not KAGGLE_KEY:
return "⚠️ KAGGLE_USERNAME / KAGGLE_KEY not set in Space secrets."
notebooks_dir = ROOT / "notebooks"
nb_file = notebooks_dir / "kaggle_master_trainer.ipynb"
meta_file = notebooks_dir / "kernel-metadata.json"
if not nb_file.exists():
return "❌ notebooks/kaggle_master_trainer.ipynb not found in Space."
if not meta_file.exists():
return "❌ notebooks/kernel-metadata.json not found in Space."
# Inject credentials into os.environ before any kaggle import —
# kaggle.authenticate() reads KAGGLE_USERNAME + KAGGLE_KEY from env.
os.environ["KAGGLE_USERNAME"] = KAGGLE_USERNAME
os.environ["KAGGLE_KEY"] = KAGGLE_KEY
# ── Method 1: Python API (no binary PATH issues) ─────────────────────────
api_err = None
try:
from kaggle.api.kaggle_api_extended import KaggleApiExtended
_kapi = KaggleApiExtended()
_kapi.authenticate()
_kapi.kernels_push_cli(str(notebooks_dir), quiet=True)
return (
"✅ Kaggle training triggered!\n"
f"Kernel: {KAGGLE_KERNEL_SLUG}\n"
"Check https://www.kaggle.com for run progress."
)
except Exception as e:
api_err = str(e)
# ── Method 2: subprocess fallback ────────────────────────────────────────
import subprocess, shutil
kaggle_bin = shutil.which("kaggle")
if kaggle_bin is None:
for cand in [
Path(sys.executable).parent / "kaggle",
Path("/usr/local/bin/kaggle"),
Path("/usr/bin/kaggle"),
]:
if cand.exists():
kaggle_bin = str(cand)
break
if kaggle_bin is None:
return (
f"❌ Kaggle CLI not found (API error: {api_err}).\n"
"Ensure kaggle>=1.6.0 is in requirements.txt and the Space rebuilt."
)
env = {
**os.environ,
"KAGGLE_USERNAME": KAGGLE_USERNAME,
"KAGGLE_KEY": KAGGLE_KEY,
"PYTHONUTF8": "1",
"PYTHONIOENCODING": "utf-8",
}
try:
result = subprocess.run(
[kaggle_bin, "kernels", "push", "-p", str(notebooks_dir)],
capture_output=True, text=True, timeout=60, env=env,
)
if result.returncode == 0:
out = (result.stdout or "").strip()
return f"✅ Kaggle training triggered!\n{out or 'Kernel version created.'}"
err = (result.stderr or result.stdout or "unknown error").strip()
return f"❌ Kaggle push failed:\n{err}"
except Exception as e:
return f"❌ Kaggle push failed (API: {api_err}) (CLI: {e})"
def _maybe_auto_trigger() -> None:
"""Called after each correction save. Triggers Kaggle if threshold met."""
if not KAGGLE_USERNAME or not KAGGLE_KEY:
return
count = _count_corrections()
if count > 0 and count % AUTO_TRAIN_THRESHOLD == 0:
import logging
def _run():
msg = _trigger_kaggle_training()
logging.getLogger(__name__).info("Auto-trigger: %s", msg)
threading.Thread(target=_run, daemon=True).start()
# ── Bulk upload handler ────────────────────────────────────────────────────────
def _bulk_upload(lang_label: str, zip_file, csv_text: str) -> str:
"""
Accept a ZIP of audio files + a CSV (filename,transcription) and batch-insert
all samples into corrections.jsonl. Audio stored under audio/ in the Hub repo.
"""
import zipfile, csv
if _hf_api is None:
return "⚠️ HF_TOKEN not set — cannot upload."
if zip_file is None and not csv_text.strip():
return "⚠️ Upload a ZIP and/or paste a CSV."
lang = SUPPORTED_LANGUAGES.get(lang_label, "bam")
rows = [] # (audio_bytes_or_None, filename, transcription)
# Parse CSV
transcript_map: dict[str, str] = {}
if csv_text.strip():
for row in csv.reader(csv_text.strip().splitlines()):
if len(row) >= 2:
transcript_map[row[0].strip()] = row[1].strip()
# Extract ZIP
if zip_file is not None:
try:
with zipfile.ZipFile(zip_file, "r") as zf:
for name in zf.namelist():
if not name.lower().endswith((".wav", ".mp3", ".ogg", ".flac", ".m4a")):
continue
text = transcript_map.get(name) or transcript_map.get(Path(name).name) or ""
if not text:
continue
rows.append((zf.read(name), Path(name).name, text))
except Exception as e:
return f"❌ ZIP read error: {e}"
elif transcript_map:
# CSV only — audio-less vocab entries
for fname, text in transcript_map.items():
rows.append((None, fname, text))
if not rows:
return "⚠️ No matching (audio, transcription) pairs found. Check filenames match CSV."
# Upload batch
records = []
errors = 0
for audio_bytes, fname, text in rows:
ts = datetime.now(timezone.utc).strftime("%Y%m%d_%H%M%S_%f")
audio_path = f"audio/{lang}_{ts}.wav"
try:
if audio_bytes:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(audio_bytes),
path_in_repo=audio_path,
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
records.append({
"id": ts, "timestamp": datetime.now(timezone.utc).isoformat(),
"language": lang,
"audio_file": audio_path if audio_bytes else "",
"transcription": text, "corrected_text": text,
"source": f"bulk_upload:{fname}", "is_correction": False,
"model": WHISPER_MODEL_ID,
})
except Exception:
errors += 1
# Append all to corrections.jsonl
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="corrections.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
new_lines = "".join(json.dumps(r, ensure_ascii=False) + "\n" for r in records)
updated = existing + new_lines
try:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(updated.encode("utf-8")),
path_in_repo="corrections.jsonl",
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
break
except Exception as e:
if attempt == 1:
return f"⚠️ Audio uploaded but corrections.jsonl failed: {e}"
total = updated.count("\n")
_maybe_auto_trigger()
return (
f"✅ Bulk upload complete!\n"
f" Uploaded : {len(records)} samples ({errors} errors)\n"
f" Dataset : {total} total corrections\n"
f" Auto-train threshold: {AUTO_TRAIN_THRESHOLD} entries"
)
# ── Internet self-teaching handlers ───────────────────────────────────────────
def _upload_jsonl(repo_path: str, entries: list[dict]) -> tuple[int, str | None]:
"""Append entries to a jsonl file on the Hub. Returns (total_lines, error_or_None)."""
from huggingface_hub import hf_hub_download
for attempt in range(2):
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename=repo_path,
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing = f.read()
except Exception:
existing = ""
updated = existing + "".join(json.dumps(e, ensure_ascii=False) + "\n" for e in entries)
try:
_hf_api.upload_file(
path_or_fileobj=io.BytesIO(updated.encode("utf-8")),
path_in_repo=repo_path,
repo_id=FEEDBACK_REPO_ID,
repo_type="dataset",
)
return updated.count("\n"), None
except Exception as e:
if attempt == 1:
return 0, str(e)
return 0, "unknown error"
def _harvest_wikipedia(lang_label: str, max_articles: int = 100) -> str:
"""Fetch Wikipedia text and append to vocabulary.jsonl."""
if _hf_api is None:
return "⚠️ HF_TOKEN not set."
import re as _re_tmp
_m = _re_tmp.search(r'\((\w+)\)$', lang_label.strip())
lang = _m.group(1) if _m else SUPPORTED_LANGUAGES.get(lang_label, "bam")
if lang not in ("bam", "ful"):
return "⚠️ Supported for Bambara and Fula only."
max_articles = int(max_articles) # Gradio slider returns float
try:
from src.data.web_harvester import harvest_wikipedia_text
entries = harvest_wikipedia_text(lang, max_articles=max_articles)
except Exception as e:
return f"❌ Harvest error: {e}"
if not entries:
return "⚠️ No sentences extracted. Wikipedia may be temporarily unavailable."
total, err = _upload_jsonl("vocabulary.jsonl", entries)
if err:
return f"❌ Upload failed: {err}"
threading.Thread(target=_refresh_vocab_context, daemon=True).start()
return (
f"✅ Wikipedia harvest complete!\n"
f" Language : {lang_label}\n"
f" Sentences added : {len(entries)}\n"
f" Vocabulary total : {total} entries"
)
def _harvest_hf_dataset(lang_label: str, max_samples: int = 500) -> str:
"""
Register an HF dataset as a training source by writing its config to
dataset_sources.jsonl on the Hub. The Kaggle notebook reads this file
at Cell 4 and loads the dataset directly — no audio re-upload needed.
"""
if _hf_api is None:
return "⚠️ HF_TOKEN not set."
# Dropdown sends "Bambara (bam)" / "Fula (ful)" — extract the code in parens
import re as _re_tmp
_m = _re_tmp.search(r'\((\w+)\)$', lang_label.strip())
lang = _m.group(1) if _m else SUPPORTED_LANGUAGES.get(lang_label, "bam")
if lang not in ("bam", "ful"):
return "⚠️ Supported for Bambara and Fula only."
max_samples = int(max_samples) # Gradio slider returns float
from src.data.web_harvester import get_hf_dataset_refs
refs = get_hf_dataset_refs(lang)
if not refs:
return f"⚠️ No HF dataset configured for {lang}."
# Read existing entries to avoid duplicates
from huggingface_hub import hf_hub_download
try:
local = hf_hub_download(
repo_id=FEEDBACK_REPO_ID, filename="dataset_sources.jsonl",
repo_type="dataset", token=HF_TOKEN,
)
with open(local, encoding="utf-8") as f:
existing_entries = [json.loads(l) for l in f if l.strip()]
except Exception:
existing_entries = []
existing_keys = {
(e.get("repo", e.get("repo_id", "")), e.get("config", ""))
for e in existing_entries
}
new_entries = []
already = []
for ref in refs:
key = (ref.get("repo", ref.get("repo_id", "")), ref.get("config", ""))
if key in existing_keys:
already.append(ref["repo"])
continue
entry = dict(ref)
entry["max"] = max_samples
entry["enabled"] = True
entry["added_at"] = datetime.now(timezone.utc).isoformat()
new_entries.append(entry)
if not new_entries:
repos = ", ".join(already)
return (
f"✅ Already registered!\n"
f" `{repos}` is already in your training config.\n"
f" Click 'Trigger Training Now' to start a run with this data."
)
total, err = _upload_jsonl("dataset_sources.jsonl", new_entries)
if err:
return f"❌ Upload failed: {err}"
repos = ", ".join(r["repo"] for r in refs)
return (
f"✅ Dataset registered for training!\n"
f" Source(s) : {repos}\n"
f" Max samples : {max_samples}\n"
f" The Kaggle notebook will stream this dataset directly at training time.\n"
f" Click 'Trigger Training Now' to start a training run.\n"
f" Total registered sources: {total}"
)
# ── Main ask handler ──────────────────────────────────────────────────────────
def handle_ask(audio_path, language_label, convo_mode: bool = False, history: list | None = None):
"""
Main dispatcher. Always returns 5 values:
(transcript, eng_translation, response_text, audio_out, new_history)
new_history is the updated gr.State list of (user, asst) tuples.
In normal (sensor) mode, history is passed through unchanged.
"""
history = history or []
if audio_path is None:
return "⚠️ No audio — press Record or upload a file.", "", "", None, history
language_code = SUPPORTED_LANGUAGES.get(language_label, "bam")
if not _wait_for_whisper(timeout=120):
return f"❌ Model failed to load: {_model_status}", "", "", None, history
try:
if convo_mode:
transcript, eng, response_text, audio_out, new_history = _convo_pipeline(
audio_path, language_code, history
)
else:
transcript, eng, response_text, audio_out = _run_pipeline(audio_path, language_code)
new_history = history # sensor mode doesn't modify history
return transcript, eng, response_text, audio_out, new_history
except Exception as e:
return f"❌ {e}", "", "", None, history
# ── Two-stage pipeline (shows transcript fast, then response) ─────────────────
def _do_asr(audio_path: str, language_label: str) -> str:
"""
Stage 1 — Whisper only. Returns the transcript string (or error/status).
Blocks until the model is ready (up to 120 s) so the first request after a
cold start works without the user needing to retry.
"""
if audio_path is None:
return "⚠️ No audio — press Record or upload a file."
lang = SUPPORTED_LANGUAGES.get(language_label, "bam")
if not _wait_for_whisper(timeout=120):
return f"❌ Model failed to load: {_model_status}"
try:
import torch, librosa
device = "cuda" if torch.cuda.is_available() else "cpu"
audio_np, _ = librosa.load(audio_path, sr=16000, mono=True)
active_model = _fine_tuned_models.get(lang, _whisper_model)
active_processor = _fine_tuned_processors.get(lang, _whisper_processor)
active_model.to(device)
with _model_lock:
input_features = active_processor.feature_extractor(
audio_np, sampling_rate=16000, return_tensors="pt"
).input_features.to(device)
forced_ids = None
if lang not in ("bam", "ful"):
forced_ids = active_processor.get_decoder_prompt_ids(
language=lang, task="transcribe"
)
with torch.no_grad():
ids = active_model.generate(
input_features,
forced_decoder_ids=forced_ids or None,
max_new_tokens=256,
)
transcript = active_processor.batch_decode(ids, skip_special_tokens=True)[0].strip()
active_model.to("cpu")
if device == "cuda":
torch.cuda.empty_cache()
# Phonetic normalisation (Bambara: French spellings → standard; Fula: Adlam → Latin)
if lang == "bam":
return bam_normalize(transcript)
elif lang == "ful":
return normalize_pular(transcript)
return transcript
except Exception as e:
return f"❌ Transcription error: {e}"
def _do_respond(
transcript: str,
language_label: str,
convo_mode: bool,
history: list,
) -> tuple:
"""
Stage 2 — LLM or sensor response, runs after transcript is already visible.
Returns (eng_translation, response_text, audio_out, new_history, chat_msgs).
"""
history = history or []
# Bail early if stage 1 errored
if not transcript or transcript[:1] in ("⚠️", "⏳", "❌") or transcript.startswith(("⚠", "⏳", "❌")):
chat_msgs = [[u, v] for u, v in history]
return "", "", None, history, chat_msgs
lang = SUPPORTED_LANGUAGES.get(language_label, "bam")
import torch
device = "cuda" if torch.cuda.is_available() else "cpu"
if convo_mode:
# ── LLM brain ────────────────────────────────────────────────────────
response_text = ""
try:
from huggingface_hub import InferenceClient
client = InferenceClient(token=HF_TOKEN)
messages = _build_messages(transcript, history, lang)
completion = client.chat_completion(
model=LLM_MODEL_ID,
messages=messages,
max_tokens=150, # short spoken responses, much faster
temperature=0.65,
)
response_text = completion.choices[0].message.content.strip()
except Exception as llm_err:
import logging
logging.getLogger(__name__).warning("LLM error: %s", llm_err)
_fallbacks = {
"bam": "Hakɛ to, tasuma tɛ kɛ sisan. I ka a lasɔrɔ tugu.",
"ful": "Hakke, mi waawaa jogaade modèl oo jooni. Njaɓɓu.",
"fr": "Désolé, le modèle est indisponible pour l'instant.",
}
response_text = _fallbacks.get(lang, "Sorry, the language model is unavailable.")
# Strip [LEARNED:] tags, persist async
response_text, _ = _parse_and_strip_learned(response_text, lang)
# Update history
new_history = list(history) + [(transcript, response_text)]
if len(new_history) > 20:
new_history = new_history[-20:]
chat_msgs = [[u, v] for u, v in new_history]
# ── TTS ───────────────────────────────────────────────────────────────
audio_out = None
if _voice_ref_path and Path(_voice_ref_path).exists():
try:
from src.tts.f5_tts import synthesize as f5s
result = f5s(response_text, ref_wav_path=_voice_ref_path,
ref_text=_voice_ref_text, device=device)
if result is not None:
audio_out = (result[1], result[0])
except Exception:
pass
if audio_out is None:
try:
wav_np, sr = _tts.synthesize(response_text, lang, device=device)
audio_out = (sr, wav_np)
except Exception as tts_err:
import logging
logging.getLogger(__name__).warning("TTS error: %s", tts_err)
return "", response_text, audio_out, new_history, chat_msgs
else:
# ── Sensor / phrase pipeline ──────────────────────────────────────────
import asyncio
phrase_match = _phrase_matcher.match(transcript, lang)
if phrase_match:
response_text = phrase_match["response"]
english_translation = phrase_match["english"]
else:
intent = _intent_parser.parse(transcript, language=lang)
try:
loop = asyncio.new_event_loop()
sensor_data = loop.run_until_complete(_sensor_bridge.fetch(intent))
loop.close()
except Exception:
from src.iot.sensor_bridge import SensorData
sensor_data = SensorData(sensor_type="soil",
values={"moisture_pct": 45.0, "ph": 6.5, "temperature_c": 28.0})
responder = VoiceResponder(language=lang)
response_text, english_translation = responder.generate_response(intent, sensor_data)
if intent.action == "unknown" and intent.confidence < 0.15:
from src.iot.voice_responder import BAMBARA_TEMPLATES, FULA_TEMPLATES
if lang == "bam":
response_text, english_translation = BAMBARA_TEMPLATES["not_understood"]
elif lang == "ful":
response_text, english_translation = FULA_TEMPLATES["not_understood"]
audio_out = None
try:
wav_np, sr = _tts.synthesize(response_text, lang, device=device)
audio_out = (sr, wav_np)
except Exception as tts_err:
import logging
logging.getLogger(__name__).warning("TTS error: %s", tts_err)
chat_msgs = [[u, v] for u, v in history]
return english_translation, response_text, audio_out, history, chat_msgs
# ── Gradio UI ─────────────────────────────────────────────────────────────────
def build_ui() -> gr.Blocks:
with gr.Blocks(title="Sahel-Agri Voice AI") as demo:
gr.Markdown("# 🌾 Sahel-Agri Voice AI")
gr.Markdown(
"Speak in **Bambara** or **Fula** — get agricultural insights spoken back "
"in your language. Also supports French and English."
)
model_status_box = gr.Textbox(
value=get_model_status(),
label="Model status",
interactive=False,
)
# gr.Timer polls get_model_status every 3s and updates the box (Gradio 5)
status_timer = gr.Timer(value=3)
status_timer.tick(fn=get_model_status, outputs=model_status_box)
with gr.Tabs() as tabs:
# ── Tab 1: Voice Assistant ────────────────────────────────────────
with gr.TabItem("🎙️ Voice Assistant", id="tab_voice"):
# ── Conversation Mode controls (top bar) ─────────────────────
with gr.Row():
convo_mode_toggle = gr.Checkbox(
value=False,
label="🔄 Conversation Mode — AI responds with LLM + cloned voice",
info="When ON: mic auto-submits on stop; AI replies via LLM + F5-TTS (requires voice reference below).",
)
with gr.Accordion("🎤 Voice Reference — upload an MP3/WAV of the target speaker", open=False):
gr.Markdown(
"Upload **5–30 seconds** of clear speech in the target voice. "
"The AI will speak all its responses using this voice. "
"Requires `f5-tts` and a GPU — falls back to MMS-TTS otherwise."
)
with gr.Row():
voice_ref_input = gr.Audio(
sources=["upload"],
type="filepath",
label="Reference audio (MP3 or WAV)",
)
voice_ref_status = gr.Textbox(
label="Status", interactive=False, lines=3
)
voice_ref_btn = gr.Button("💾 Set as Voice Reference", variant="secondary")
voice_ref_btn.click(
fn=set_voice_reference,
inputs=[voice_ref_input],
outputs=[voice_ref_status],
)
gr.Markdown("---")
# Per-session conversation history (not shared between users)
conv_history = gr.State(value=[])
with gr.Row():
with gr.Column(scale=1):
language_dd = gr.Dropdown(
choices=list(SUPPORTED_LANGUAGES.keys()),
value="Bambara (bam)",
label="Language / Kan",
)
audio_input = gr.Audio(
sources=["microphone", "upload"],
type="filepath",
label="Record or upload audio",
)
with gr.Row():
ask_btn = gr.Button("▶ Ask / Ɲinɛ", variant="primary")
clear_btn = gr.Button("🗑 Clear", variant="secondary", size="sm")
with gr.Column(scale=1):
transcript_box = gr.Textbox(
label="Whisper heard",
lines=2,
placeholder="Your words will appear here…",
interactive=False,
)
translation_box = gr.Textbox(
label="English translation",
lines=2,
placeholder="(shown in sensor mode only)",
interactive=False,
)
response_box = gr.Textbox(
label="AI response",
lines=2,
placeholder="Response will appear here…",
interactive=False,
)
audio_output = gr.Audio(
label="Voice response",
autoplay=True,
interactive=False,
)
correct_btn = gr.Button(
"✏️ Something wrong? Send to Correction tab",
variant="secondary",
size="sm",
)
# Conversation history display (Conversation Mode only)
chatbot = gr.Chatbot(
label="Conversation history",
height=300,
visible=False,
type="tuples",
)
convo_mode_toggle.change(
fn=lambda on: gr.update(visible=on),
inputs=[convo_mode_toggle],
outputs=[chatbot],
)
# ── Stage 1 inputs/outputs (ASR only — fast) ─────────────────
_s1_inputs = [audio_input, language_dd]
_s1_outputs = [transcript_box]
# ── Stage 2 inputs/outputs (LLM / sensor + TTS) ──────────────
_s2_inputs = [transcript_box, language_dd, convo_mode_toggle, conv_history]
_s2_outputs = [translation_box, response_box, audio_output,
conv_history, chatbot]
# Manual button: stage 1 then stage 2
ask_btn.click(
fn=_do_asr,
inputs=_s1_inputs,
outputs=_s1_outputs,
).then(
fn=_do_respond,
inputs=_s2_inputs,
outputs=_s2_outputs,
)
# Auto-submit on mic stop: same chain, but stage 2 only runs when
# convo_mode is ON (sensor mode has a manual button for deliberate use)
audio_input.stop_recording(
fn=_do_asr,
inputs=_s1_inputs,
outputs=_s1_outputs,
).then(
fn=lambda t, ll, cm, h: _do_respond(t, ll, cm, h) if cm
else ("", "", None, h, [[u, v] for u, v in (h or [])]),
inputs=_s2_inputs,
outputs=_s2_outputs,
)
# Clear conversation
clear_btn.click(
fn=lambda: ([], []),
outputs=[conv_history, chatbot],
)
# ── Tab 2: Feedback & Correction ─────────────────────────────────
with gr.TabItem("📝 Feedback & Correction", id="tab_feedback"):
gr.Markdown(
"Correct what Whisper heard, the English translation, and the response. "
"All corrections are saved to the training dataset to improve future accuracy."
)
with gr.Row():
with gr.Column():
fb_lang = gr.Dropdown(
choices=list(SUPPORTED_LANGUAGES.keys()),
value="Bambara (bam)",
label="Language",
)
fb_audio = gr.Audio(
sources=["microphone", "upload"],
type="filepath",
label="Audio",
)
gr.Markdown("**Step 1 — Fix the transcription**")
fb_transcript = gr.Textbox(
label="What Whisper heard",
lines=2,
placeholder="Auto-filled from Tab 1…",
)
fb_corrected = gr.Textbox(
label="✏️ What was actually said (in Bambara/Fula)",
lines=2,
placeholder="Type the correct transcription here…",
)
with gr.Column():
gr.Markdown("**Step 2 — Fix the English translation**")
fb_english = gr.Textbox(
label="Auto-generated English translation",
lines=2,
placeholder="Auto-filled from Tab 1…",
)
fb_corrected_english = gr.Textbox(
label="✏️ Correct English translation",
lines=2,
placeholder="Type the correct English meaning here…",
)
gr.Markdown("**Step 3 — Fix the response**")
fb_response = gr.Textbox(
label="Auto-generated response",
lines=2,
placeholder="Auto-filled from Tab 1…",
)
fb_corrected_response = gr.Textbox(
label="✏️ Better response (in farmer's language)",
lines=2,
placeholder="Type a better response here…",
)
fb_rating = gr.Slider(
minimum=1, maximum=5, step=1, value=3,
label="Overall quality (1 = poor, 5 = excellent)",
)
fb_notes = gr.Textbox(
label="Notes (optional)",
lines=2,
placeholder="e.g. noisy background, strong accent…",
)
save_btn = gr.Button("💾 Save to Dataset", variant="primary")
save_status = gr.Textbox(
label="Save status", interactive=False, lines=2
)
save_btn.click(
fn=_save_feedback_to_hub,
inputs=[
fb_audio, fb_transcript, fb_corrected,
fb_english, fb_corrected_english,
fb_response, fb_corrected_response,
fb_rating, fb_notes, fb_lang,
],
outputs=[save_status],
)
# Wire "Send to Correction" button — populates Tab 2 fields from Tab 1
correct_btn.click(
fn=lambda t, tr, r, lang: (t, t, tr, tr, r, r, lang),
inputs=[transcript_box, translation_box, response_box, language_dd],
outputs=[fb_transcript, fb_corrected, fb_english, fb_corrected_english,
fb_response, fb_corrected_response, fb_lang],
)
# ── Tab 3: Knowledge Base ─────────────────────────────────────────
with gr.TabItem("📚 Knowledge Base"):
gr.Markdown(
"## Teach the assistant new phrases — no technical knowledge required\n\n"
"Add phrases the assistant should recognise and respond to. "
"Changes take effect **immediately** and are saved to the Hub so they survive restarts."
)
with gr.Row():
# ── Left: phrase pair import ──────────────────────────────
with gr.Column():
gr.Markdown(
"### ➕ Add phrases manually\n"
"One phrase per line in the format:\n"
"```\nnative phrase | English translation\n```\n"
"**Examples (Bambara):**\n"
"```\nI ni ce | Hello, good day\n"
"Sanji bɛ na | Rain is coming\n"
"N bɛ i dɛmɛ | I will help you\n```\n"
"**Examples (Fula):**\n"
"```\nJam waali | Hello, peace be with you\n"
"Ndiyam wadata | Rain is coming\n"
"Mi woni ɗoo | I am here\n```"
)
kb_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)",
label="Language",
)
kb_pairs = gr.Textbox(
lines=10,
placeholder="I ni ce | Hello, good day\nI ni sogoma | Good morning\nSanji bɛ na | Rain is coming",
label="Phrase pairs (native | english) — one per line",
)
kb_import_btn = gr.Button("➕ Add to Knowledge Base", variant="primary")
kb_status = gr.Textbox(label="Status", interactive=False, lines=3)
# ── Right: audio upload for training ─────────────────────
with gr.Column():
gr.Markdown(
"### 🎬 Add audio from YouTube (or anywhere)\n"
"HuggingFace Spaces cannot download YouTube directly, "
"so convert the video to audio first on your computer:\n\n"
"**Free online converters:**\n"
"- [ytmp3.cc](https://ytmp3.cc) — paste YouTube URL → download MP3\n"
"- [cobalt.tools](https://cobalt.tools) — paste any video URL → download audio\n"
"- [y2mate.com](https://y2mate.com) — paste YouTube URL → download MP3\n\n"
"**Good YouTube search terms:**\n"
"- Bambara: *'Bamanankan conversation'*, *'Bambara leçon'*, *'donsomana'*\n"
"- Fula: *'Fulfulde leçon'*, *'Pular conversation'*, *'Fula radio'*\n\n"
"Then upload the MP3/WAV file below with its transcription."
)
yt_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)",
label="Language spoken in the audio",
)
yt_audio = gr.Audio(
sources=["upload"],
type="filepath",
label="Upload audio file (MP3 or WAV)",
)
yt_transcript = gr.Textbox(
lines=5,
placeholder="Type what is said in the audio (as much as you can).\n"
"Example:\nJam waali. No mbadda. Mi woni ɗoo wallude ma.",
label="Transcription — what is said in this audio",
)
yt_source = gr.Textbox(
placeholder="e.g. YouTube: Bambara lesson by Moussa Kouyaté",
label="Source (optional — for your records)",
)
yt_btn = gr.Button("💾 Save Audio for Training", variant="secondary")
yt_status = gr.Textbox(label="Status", interactive=False, lines=4)
kb_import_btn.click(
fn=_import_phrase_pairs,
inputs=[kb_lang, kb_pairs],
outputs=[kb_status],
)
yt_btn.click(
fn=_save_audio_for_training,
inputs=[yt_lang, yt_audio, yt_transcript, yt_source],
outputs=[yt_status],
)
# ── Document upload (PDF / Word / TXT) ───────────────────────
gr.Markdown("---")
gr.Markdown(
"### 📄 Upload documents (PDF, Word, TXT)\n"
"Extract sentences from books, articles, or lesson PDFs. "
"Each sentence is added to the training vocabulary in the language you select below. "
"**Upload one batch per language** — do not mix Bambara and Fula files in one upload."
)
with gr.Row():
with gr.Column():
doc_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Fula (ful)",
label="Language of these documents",
)
doc_files = gr.File(
label="Upload .pdf, .docx, or .txt (multiple allowed)",
file_count="multiple",
file_types=[".pdf", ".docx", ".doc", ".txt", ".md"],
)
doc_source = gr.Textbox(
placeholder="e.g. SIL Pular grammar book, Labé lesson PDFs",
label="Source note (optional — for your records)",
)
doc_btn = gr.Button("📥 Extract & Add to Training Data", variant="primary")
with gr.Column():
doc_status = gr.Textbox(
label="Import status",
interactive=False,
lines=12,
)
doc_btn.click(
fn=_import_documents,
inputs=[doc_lang, doc_files, doc_source],
outputs=[doc_status],
)
# ── Tab 4: Model Training ─────────────────────────────────────────
with gr.TabItem("🔧 Model Training"):
gr.Markdown(
"After collecting audio corrections and YouTube samples, "
"run the training notebook to fine-tune the speech model."
)
adapter_status_md = gr.Markdown(value=_get_adapter_status())
reload_btn = gr.Button("🔄 Reload Fine-tuned Models from Hub")
reload_out = gr.Markdown()
gr.Markdown("---")
gr.Markdown(
"**Training notebook**: "
"`notebooks/kaggle_master_trainer.ipynb` — import to Kaggle, run all cells.\n\n"
"**What feeds training:**\n"
"- Tab 2 corrections → `corrections.jsonl` in the feedback dataset\n"
"- Tab 3 audio uploads → `corrections.jsonl` (same file)\n"
"- Tab 3 phrase pairs → `vocabulary.jsonl` (used as synthetic fallback labels)\n\n"
"**Feedback dataset**: "
f"`{FEEDBACK_REPO_ID}` (auto-updated on each save)\n\n"
"**Model checkpoint repo**: "
f"`{ADAPTER_REPO_ID}` (updated after training, reload above to activate)"
)
reload_btn.click(fn=_reload_adapters_from_hub, outputs=[reload_out])
reload_btn.click(fn=_get_adapter_status, outputs=[adapter_status_md])
# ── Tab 5: Bulk Upload ────────────────────────────────────────────
with gr.TabItem("📦 Bulk Upload"):
gr.Markdown(
"## Upload many audio samples at once\n\n"
"**Step 1** — Prepare a ZIP file containing your audio files (WAV/MP3).\n\n"
"**Step 2** — Prepare a CSV with two columns: `filename,transcription`\n"
"```\nbam_001.wav,I ni ce a tɔ\nbam_002.wav,Sanji bɛ na sini\n```\n\n"
"**Step 3** — Select language, upload ZIP, paste CSV, click Upload."
)
with gr.Row():
with gr.Column():
bulk_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)", label="Language"
)
bulk_zip = gr.File(
label="ZIP file (audio files)", file_types=[".zip"]
)
bulk_csv = gr.Textbox(
lines=10,
label="CSV — filename,transcription (one per line)",
placeholder="bam_001.wav,I ni ce a tɔ\nbam_002.wav,Sanji bɛ na sini",
)
bulk_btn = gr.Button("📤 Upload Batch", variant="primary")
bulk_status = gr.Textbox(label="Status", interactive=False, lines=5)
bulk_btn.click(
fn=_bulk_upload,
inputs=[bulk_lang, bulk_zip, bulk_csv],
outputs=[bulk_status],
)
# ── Tab 6: Self-Teaching ──────────────────────────────────────────
with gr.TabItem("🌐 Self-Teaching"):
gr.Markdown(
"## Teach the model from the internet\n\n"
"These tools pull publicly available Bambara and Fula language data "
"directly into your training dataset — no manual work required."
)
with gr.Row():
# Wikipedia harvest
with gr.Column():
gr.Markdown(
"### 📖 Wikipedia Text Harvest\n"
"Pulls sentence-length text from Bambara Wikipedia (868 articles) "
"or Fula Wikipedia (17,000+ articles) into `vocabulary.jsonl`.\n\n"
"Use this to expand vocabulary coverage before a training run."
)
wiki_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)", label="Language"
)
wiki_articles = gr.Slider(
minimum=10, maximum=500, value=100, step=10,
label="Max articles to fetch"
)
wiki_btn = gr.Button("📖 Harvest Wikipedia Text", variant="secondary")
wiki_status = gr.Textbox(label="Status", interactive=False, lines=4)
wiki_btn.click(
fn=_harvest_wikipedia,
inputs=[wiki_lang, wiki_articles],
outputs=[wiki_status],
)
# HF dataset harvest
with gr.Column():
gr.Markdown(
"### 🤗 HuggingFace Dataset Import\n"
"Registers large public datasets as training sources:\n"
"- **Bambara**: `RobotsMali/jeli-asr` (33,000 samples)\n"
"- **Fula**: `google/WaxalNLP ful_asr` + `Pullo-Africa-Protagonist/Fula-pular` (9,761 samples) + `guizme/adlam_fulfulde` (51 Adlam samples)\n\n"
"This writes a reference to `dataset_sources.jsonl`. "
"The Kaggle training notebook streams the dataset directly "
"at training time — no re-upload needed.\n\n"
"**One click is enough** — duplicates are ignored automatically."
)
hf_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)", label="Language"
)
hf_samples = gr.Slider(
minimum=50, maximum=2000, value=500, step=50,
label="Max samples to import"
)
hf_btn = gr.Button("🤗 Import from HuggingFace", variant="primary")
hf_status = gr.Textbox(label="Status", interactive=False, lines=5)
hf_btn.click(
fn=_harvest_hf_dataset,
inputs=[hf_lang, hf_samples],
outputs=[hf_status],
)
gr.Markdown("---")
gr.Markdown(
"### ⚡ Auto-Training\n"
f"When `corrections.jsonl` reaches a multiple of **{AUTO_TRAIN_THRESHOLD}** entries, "
"the Kaggle training notebook is triggered automatically.\n\n"
"To enable: add `KAGGLE_USERNAME` and `KAGGLE_KEY` in Space Settings → Secrets.\n\n"
f"Kernel: `{KAGGLE_KERNEL_SLUG}`"
)
with gr.Row():
trigger_lang = gr.Dropdown(
choices=["Bambara (bam)", "Fula (ful)"],
value="Bambara (bam)", label="Language to train"
)
trigger_btn = gr.Button("⚡ Trigger Training Now", variant="secondary")
trigger_out = gr.Textbox(label="Status", interactive=False, lines=2)
trigger_btn.click(
fn=lambda l: _trigger_kaggle_training(SUPPORTED_LANGUAGES.get(l, "bam")),
inputs=[trigger_lang],
outputs=[trigger_out],
)
return demo
# ── Entry point ───────────────────────────────────────────────────────────────
if __name__ == "__main__":
from dotenv import load_dotenv
load_dotenv()
# Re-read env after dotenv
HF_TOKEN = os.environ.get("HF_TOKEN")
FEEDBACK_REPO_ID = os.environ.get("FEEDBACK_REPO_ID", "ous-sow/sahel-agri-feedback")
ADAPTER_REPO_ID = os.environ.get("ADAPTER_REPO_ID", "ous-sow/sahel-agri-adapters")
WHISPER_MODEL_ID = os.environ.get("WHISPER_MODEL_ID", "openai/whisper-large-v3-turbo")
LLM_MODEL_ID = os.environ.get("LLM_MODEL_ID", "Qwen/Qwen2.5-7B-Instruct")
if HF_TOKEN:
from huggingface_hub import HfApi
_hf_api = HfApi(token=HF_TOKEN)
# Load any previously saved phrase additions from HF Hub
_load_phrase_additions_from_hub()
# Kick off background model load immediately
_ensure_whisper_loaded()
print(f"Whisper model : {WHISPER_MODEL_ID}")
print(f"LLM model : {LLM_MODEL_ID}")
print(f"Feedback repo : {FEEDBACK_REPO_ID}")
print(f"Adapter repo : {ADAPTER_REPO_ID}")
print(f"HF_TOKEN set : {'yes' if HF_TOKEN else 'no (local-only mode)'}")
print()
demo = build_ui()
demo.launch(
server_port=7860, # HF Spaces standard port
inbrowser=False,
share=False,
show_api=False,
ssr_mode=False, # SSR starts a Node.js process that hangs in HF Spaces containers
)