| |
| """ |
| Build chain CSVs from index. Curated by subset: only list/FASTA pairs in |
| curated_csv/cullpdb_list_fasta_index.csv are processed. |
| |
| Outputs: |
| - curated_csv/cullpdb_combined_chains.csv — single master CSV for all analysis |
| - curated_csv/subset_chains/<list_basename>.csv — one CSV per subset |
| """ |
| import csv |
| import re |
| import sys |
| from pathlib import Path |
| from typing import List, Optional, Tuple |
|
|
| SCRIPT_DIR = Path(__file__).resolve().parent |
| BASE = SCRIPT_DIR.parent |
| CULLPDB_DIR = BASE / "pieces" / "2026_01_26" |
| CURATED_DIR = BASE / "curated_csv" |
| SUBSET_CSV_DIR = CURATED_DIR / "subset_chains" |
| COMBINED_CSV = CURATED_DIR / "cullpdb_combined_chains.csv" |
|
|
| PAT = re.compile( |
| r"^cullpdb_pc([\d.]+)_res([\d.]+)-([\d.]+)_" |
| r"(?:(noBrks)_)?" |
| r"len40-10000_R([\d.]+)_" |
| r"(.+?)_d\d{4}_\d{2}_\d{2}_chains(\d+)$" |
| ) |
|
|
|
|
| def split_pdb_chain(pdb_chain: str) -> tuple: |
| """Split PDB chain ID into (pdb_id, chain_id). PDB ID is first 4 chars, rest is chain.""" |
| if len(pdb_chain) <= 4: |
| return (pdb_chain, "") |
| return (pdb_chain[:4], pdb_chain[4:]) |
|
|
|
|
| def parse_params_from_basename(base: str) -> Optional[dict]: |
| m = PAT.match(base) |
| if not m: |
| return None |
| pc, res_min, res_max, no_brks, r_cutoff, methods, n_chains = m.groups() |
| n = int(n_chains) |
| no_brk = no_brks is not None |
| return { |
| "pc": float(pc), |
| "resolution_range": f"{res_min}-{res_max}", |
| "no_breaks": "yes" if no_brk else "no", |
| "R": float(r_cutoff), |
| "source_list": base, |
| } |
|
|
|
|
| def read_list_file(path: Path) -> List[dict]: |
| """Return list of dicts: PDBchain, len, method, resol, rfac, freerfac.""" |
| rows = [] |
| with open(path) as f: |
| lines = f.readlines() |
| if not lines: |
| return rows |
| |
| for line in lines[1:]: |
| line = line.strip() |
| if not line: |
| continue |
| parts = line.split() |
| if len(parts) < 5: |
| continue |
| pdb_chain = parts[0] |
| try: |
| length = int(parts[1]) |
| except ValueError: |
| continue |
| method = parts[2] |
| resol = parts[3] |
| rfac = parts[4] |
| freerfac = parts[5] if len(parts) > 5 else "" |
| rows.append({ |
| "pdb_chain": pdb_chain, |
| "len": length, |
| "method": method, |
| "resol": resol, |
| "rfac": rfac, |
| "freerfac": freerfac, |
| }) |
| return rows |
|
|
|
|
| def read_fasta_entries(path: Path) -> List[Tuple[str, str]]: |
| """Return list of (pdb_chain_from_header, sequence) in order (one per FASTA entry).""" |
| entries = [] |
| current_seq = [] |
| current_id = None |
| with open(path) as f: |
| for line in f: |
| line = line.rstrip("\n") |
| if line.startswith(">"): |
| if current_seq is not None: |
| if current_id is not None: |
| entries.append((current_id, "".join(current_seq))) |
| |
| rest = line[1:].strip() |
| current_id = rest.split(None, 1)[0] if rest else "" |
| current_seq = [] |
| else: |
| current_seq.append(line) |
| if current_id is not None: |
| entries.append((current_id, "".join(current_seq))) |
| return entries |
|
|
|
|
| def load_pairs_from_index(index_csv: Path) -> List[tuple]: |
| """Load (base, list_path, fasta_path, params) from curated index CSV. Only these subsets are used.""" |
| pairs = [] |
| with open(index_csv, newline="") as f: |
| r = csv.DictReader(f) |
| for row in r: |
| base = row.get("list_basename", "").strip() |
| list_path_s = row.get("list_path", "").strip() |
| fasta_path_s = row.get("fasta_path", "").strip() |
| if not base or not list_path_s or not fasta_path_s: |
| continue |
| list_path = Path(list_path_s) |
| fasta_path = Path(fasta_path_s) |
| if not list_path.exists() or not fasta_path.exists(): |
| print(f"Skip (missing): {base}", file=sys.stderr) |
| continue |
| params = parse_params_from_basename(base) |
| if not params: |
| print(f"Skip (parse): {base}", file=sys.stderr) |
| continue |
| pairs.append((base, list_path, fasta_path, params)) |
| return pairs |
|
|
|
|
| def main(): |
| index_csv = CURATED_DIR / "cullpdb_list_fasta_index.csv" |
| if not index_csv.exists(): |
| print(f"Index not found: {index_csv}. Run build_list_fasta_index.py first.", file=sys.stderr) |
| sys.exit(1) |
|
|
| pairs = load_pairs_from_index(index_csv) |
| if not pairs: |
| print("No subsets in index (or paths missing). Nothing to do.", file=sys.stderr) |
| sys.exit(1) |
| print(f"Curating from index: {len(pairs)} subsets.") |
| fieldnames = [ |
| "pdb_chain", "pdb", "chain", "sequence", "len", "method", "resolution", "rfac", "freerfac", |
| "pc", "no_breaks", "R", "source_list", |
| ] |
| SUBSET_CSV_DIR.mkdir(parents=True, exist_ok=True) |
|
|
| |
| id_mismatches = 0 |
| len_mismatches = 0 |
| count_mismatches = 0 |
| files_skipped = 0 |
| rows_written_total = 0 |
| files_written = 0 |
| all_rows = [] |
|
|
| for base, list_path, fasta_path, params in pairs: |
| list_rows = read_list_file(list_path) |
| fasta_entries = read_fasta_entries(fasta_path) |
| n_list, n_fasta = len(list_rows), len(fasta_entries) |
| if n_list != n_fasta: |
| print(f"SKIP {base}: list has {n_list} rows, fasta has {n_fasta} entries", file=sys.stderr) |
| files_skipped += 1 |
| continue |
|
|
| subset_rows = [] |
| file_id_mismatches = 0 |
| file_len_mismatches = 0 |
| for i, (row, (fasta_id, seq)) in enumerate(zip(list_rows, fasta_entries)): |
| list_id = row["pdb_chain"] |
| list_len = row["len"] |
| seq_len = len(seq) |
|
|
| |
| if list_id != fasta_id: |
| id_mismatches += 1 |
| file_id_mismatches += 1 |
| if file_id_mismatches <= 3: |
| print(f"ID mismatch in {base} row {i+2}: list={list_id!r} fasta_header={fasta_id!r}", file=sys.stderr) |
| continue |
|
|
| |
| if list_len != seq_len: |
| len_mismatches += 1 |
| file_len_mismatches += 1 |
| if file_len_mismatches <= 3: |
| print(f"LEN mismatch in {base} {list_id}: list len={list_len} seq len={seq_len}", file=sys.stderr) |
| continue |
|
|
| pdb_id, chain_id = split_pdb_chain(list_id) |
| out_row = { |
| "pdb_chain": list_id, |
| "pdb": pdb_id, |
| "chain": chain_id, |
| "sequence": seq, |
| "len": list_len, |
| "method": row["method"], |
| "resolution": row["resol"], |
| "rfac": row["rfac"], |
| "freerfac": row["freerfac"], |
| "pc": params["pc"], |
| "no_breaks": params["no_breaks"], |
| "R": params["R"], |
| "source_list": params["source_list"], |
| } |
| subset_rows.append(out_row) |
|
|
| if file_id_mismatches > 0 or file_len_mismatches > 0: |
| count_mismatches += 1 |
|
|
| |
| out_path = SUBSET_CSV_DIR / f"{base}.csv" |
| with open(out_path, "w", newline="") as f: |
| w = csv.DictWriter(f, fieldnames=fieldnames) |
| w.writeheader() |
| w.writerows(subset_rows) |
| all_rows.extend(subset_rows) |
| rows_written_total += len(subset_rows) |
| files_written += 1 |
|
|
| |
| with open(COMBINED_CSV, "w", newline="") as f: |
| w = csv.DictWriter(f, fieldnames=fieldnames) |
| w.writeheader() |
| w.writerows(all_rows) |
| print(f"Wrote {COMBINED_CSV} with {len(all_rows)} chain rows (master for analysis).") |
|
|
| |
| print(f"Wrote {files_written} subset CSVs to {SUBSET_CSV_DIR} ({rows_written_total} chain rows total).") |
| print(f"Sanity check: ID mismatches (list PDBchain vs FASTA header) = {id_mismatches}") |
| print(f"Sanity check: length mismatches (list len vs len(sequence)) = {len_mismatches}") |
| print(f"Files with ID or len mismatch = {count_mismatches}; files skipped (list vs FASTA count) = {files_skipped}") |
| if id_mismatches > 0 or len_mismatches > 0 or files_skipped > 0: |
| sys.exit(1) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|