| |
| """ |
| Add UniProt sequences to FireProtDB parquet/csv. |
| |
| Outputs: |
| - Parquet file containing a newly appended sequence column |
| |
| Usage: |
| python 02_add_uniprot_sequences.py \ |
| --input ../data/fireprotdb_cleaned.parquet \ |
| --output ../data/fireprotdb_with_sequences.parquet \ |
| --cache ../data/uniprot_cache.tsv |
| |
| Notes: |
| - This script generates a cache file, which can be used to restart failed or paused runs. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import argparse |
| import os |
| import time |
| from typing import Dict, Optional |
|
|
| import pandas as pd |
| import requests |
|
|
| UNIPROT_FASTA = "https://rest.uniprot.org/uniprotkb/{acc}.fasta" |
|
|
| def read_cache(path: str) -> Dict[str, str]: |
| if not os.path.exists(path): |
| return {} |
| cache = {} |
| with open(path, "r") as f: |
| for line in f: |
| line = line.rstrip("\n") |
| if not line: |
| continue |
| acc, seq = line.split("\t", 1) |
| cache[acc] = seq |
| return cache |
|
|
| def append_cache(path: str, acc: str, seq: str) -> None: |
| os.makedirs(os.path.dirname(path), exist_ok=True) |
| with open(path, "a") as f: |
| f.write(f"{acc}\t{seq}\n") |
|
|
| def parse_fasta(text: str) -> Optional[str]: |
| lines = [ln.strip() for ln in text.splitlines() if ln.strip()] |
| if not lines or not lines[0].startswith(">"): |
| return None |
| return "".join(lines[1:]).strip() or None |
|
|
| def fetch_uniprot_fasta(acc: str, session: requests.Session, timeout: int = 20) -> Optional[str]: |
| url = UNIPROT_FASTA.format(acc=acc) |
| r = session.get(url, timeout=timeout) |
| if r.status_code != 200: |
| return None |
| return parse_fasta(r.text) |
|
|
| def main(): |
| ap = argparse.ArgumentParser() |
| ap.add_argument("--input", default="../data/fireprotdb_cleaned.parquet") |
| ap.add_argument("--output", default="../data/fireprotdb_with_sequences.parquet") |
| ap.add_argument("--cache", default="../data/cache/uniprot_cache.tsv") |
| ap.add_argument("--sleep", type=float, default=0.05, help="politeness delay between requests") |
| ap.add_argument("--limit", type=int, default=0, help="debug: only fetch first N missing") |
| args = ap.parse_args() |
|
|
| df = pd.read_parquet(args.input) |
|
|
| |
| if "uniprotkb" not in df.columns: |
| raise ValueError("Expected column 'uniprotkb' in input parquet") |
|
|
| accs = df["uniprotkb"].astype("string").fillna("").str.strip() |
| unique_accs = sorted({a for a in accs.unique().tolist() if a}) |
|
|
| cache = read_cache(args.cache) |
|
|
| missing = [a for a in unique_accs if a not in cache] |
| if args.limit and args.limit > 0: |
| missing = missing[: args.limit] |
|
|
| print(f"Unique accessions: {len(unique_accs):,}") |
| print(f"Cached: {len(cache):,}") |
| print(f"Missing to fetch: {len(missing):,}") |
|
|
| session = requests.Session() |
| fetched = 0 |
|
|
| for i, acc in enumerate(missing, 1): |
| seq = fetch_uniprot_fasta(acc, session=session) |
| if seq: |
| cache[acc] = seq |
| append_cache(args.cache, acc, seq) |
| fetched += 1 |
| |
| time.sleep(args.sleep) |
|
|
| if i % 250 == 0: |
| print(f"Fetched {fetched:,}/{i:,} missing...") |
|
|
| |
| df["sequence"] = df["uniprotkb"].astype("string").fillna("").str.strip().map(lambda a: cache.get(a, None)) |
|
|
| |
| df["sequence_len_uniprot"] = df["sequence"].map(lambda s: len(s) if isinstance(s, str) else None) |
|
|
| fp_len = pd.to_numeric(df.get("sequence_length", pd.NA), errors="coerce") |
| up_len = pd.to_numeric(df["sequence_len_uniprot"], errors="coerce") |
|
|
| df["sequence_length_num"] = fp_len |
| df["length_match"] = (up_len == fp_len) & up_len.notna() & fp_len.notna() |
| |
| match_rate = float(df["length_match"].fillna(False).mean()) if len(df) else 0.0 |
|
|
| print(f"Rows with sequence: {df['sequence'].notna().sum():,}/{len(df):,}") |
| print(f"Length match rate (rows): {match_rate:.3f}") |
|
|
| df.to_parquet(args.output, index=False) |
| print(f"Wrote: {args.output}") |
|
|
| if __name__ == "__main__": |
| main() |
|
|