| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import argparse, json, re, sys |
| from pathlib import Path |
| import numpy as np |
| import pandas as pd |
| from sklearn.model_selection import StratifiedGroupKFold |
|
|
| def extract_acc_base(acc: str) -> str: |
| m = re.match(r'^(G[CF]A_\d+)', str(acc)) |
| return m.group(1) if m else str(acc).split('.')[0] |
|
|
| def load_bacterial_bases_from_supp1(supp1_csv: str) -> set: |
| df = pd.read_csv(supp1_csv) |
| cols = {c.lower().strip(): c for c in df.columns} |
| dom_col = cols.get('domain of life') or cols.get('domain_of_life') or cols.get('domain') or 'domain of life' |
| asm_col = cols.get('assembly') or cols.get('assembly accession') or 'assembly' |
| if dom_col not in df.columns or asm_col not in df.columns: |
| raise ValueError(f"Supp1.csv must contain columns similar to 'domain of life' and 'assembly'. Found: {list(df.columns)}") |
| mask = df[dom_col].astype(str).str.strip().str.upper().eq('B') |
| df_b = df.loc[mask, [asm_col]].dropna() |
| bases = set(extract_acc_base(a) for a in df_b[asm_col].astype(str)) |
| return bases |
|
|
| def load_alt_bases_from_supp2_legacy(supp2_xlsx: str) -> set: |
| """ |
| ORIGINAL behavior (kept): column index-based extraction. |
| WARNING: This assumes the 4th column (usecols=[3]) contains assembly IDs for the ALT label. |
| """ |
| df = pd.read_excel(supp2_xlsx, header=None, usecols=[3]) |
| alt = ( |
| df.iloc[:, 0] |
| .dropna() |
| .astype(str) |
| .unique() |
| .tolist() |
| ) |
| return set(extract_acc_base(x) for x in alt) |
|
|
| def load_contam_bases_from_supp2(supp2_xlsx: str) -> set: |
| """ |
| NEW: assemblies to REMOVE (contaminated), where: |
| Evidence of assembly contamination with alt gen code == 'yes' |
| Uses named columns: 'assembly' + 'Evidence of assembly contamination with alt gen code' |
| """ |
| df = pd.read_excel(supp2_xlsx) |
|
|
| cols = {c.lower().strip(): c for c in df.columns} |
| asm_col = cols.get('assembly') |
| ev_col = cols.get('evidence of assembly contamination with alt gen code') |
|
|
| if asm_col is None or ev_col is None: |
| raise ValueError( |
| "Supp2.xlsx must contain columns 'assembly' and " |
| "'Evidence of assembly contamination with alt gen code' to filter contaminated rows. " |
| f"Found columns: {list(df.columns)}" |
| ) |
|
|
| mask = ( |
| df[ev_col] |
| .astype(str) |
| .str.strip() |
| .str.lower() |
| .eq('yes') |
| ) |
|
|
| contam_bases = ( |
| df.loc[mask, asm_col] |
| .dropna() |
| .astype(str) |
| .apply(extract_acc_base) |
| .unique() |
| .tolist() |
| ) |
|
|
| return set(contam_bases) |
|
|
| def read_ndjson_records(path: str): |
| with open(path, 'r') as fh: |
| for line in fh: |
| line = line.strip() |
| if not line: |
| continue |
| try: |
| yield json.loads(line) |
| except Exception: |
| continue |
|
|
| def main(): |
| ap = argparse.ArgumentParser(description="Create grouped stratified 80/20 splits compatible with Mass_models.py") |
| ap.add_argument('--ndjson', required=True, help='Input NDJSON file (one JSON per line)') |
| ap.add_argument('--supp1', required=False, help='Optional Supp1.csv filter: keep only Bacteria (domain==B) by assembly') |
| ap.add_argument('--supp2', required=True, help='Supp2.xlsx (used for legacy alt labels + contamination filter)') |
| ap.add_argument('--outdir', required=True, help='Output directory root for subsets') |
| ap.add_argument('--n_splits', type=int, default=1, help='Number of replicate 80/20 splits') |
| ap.add_argument('--seed', type=int, default=42, help='Random seed') |
| args = ap.parse_args() |
|
|
| ndjson_path = Path(args.ndjson) |
| if not ndjson_path.exists(): |
| sys.exit(f"[ERR ] NDJSON not found: {ndjson_path}") |
|
|
| bacteria_bases = None |
| if args.supp1: |
| print(f"[FILTER] Loading Supp1 (Bacteria-only by 'domain of life' & 'assembly')…") |
| bacteria_bases = load_bacterial_bases_from_supp1(args.supp1) |
| print(f"[FILTER] Allowed assembly bases: {len(bacteria_bases)}") |
|
|
| |
| alt_bases = load_alt_bases_from_supp2_legacy(args.supp2) |
| print(f"[LABEL] Alt bases from Supp2 (legacy col[3]): {len(alt_bases)}") |
|
|
| |
| contam_bases = load_contam_bases_from_supp2(args.supp2) |
| print(f"[FILTER] Contaminated bases from Supp2 where evidence == 'yes': {len(contam_bases)}") |
|
|
| |
| overlap = len(alt_bases & contam_bases) |
| if overlap: |
| print(f"[WARN ] Overlap alt vs contaminated: {overlap} bases (will be REMOVED from dataset)") |
|
|
| print(f"[LOAD ] Reading NDJSON: {ndjson_path}") |
| records, groups, y = [], [], [] |
| dropped_contam = 0 |
| dropped_supp1 = 0 |
|
|
| for obj in read_ndjson_records(str(ndjson_path)): |
| acc = obj.get("acc") |
| if not acc: |
| continue |
| base = extract_acc_base(acc) |
|
|
| |
| if bacteria_bases is not None and base not in bacteria_bases: |
| dropped_supp1 += 1 |
| continue |
|
|
| |
| if base in contam_bases: |
| dropped_contam += 1 |
| continue |
|
|
| records.append(obj) |
| groups.append(base) |
| y.append(1 if base in alt_bases else 0) |
|
|
| if not records: |
| sys.exit("[ERR ] No records after filtering. Check Supp1 filter / Supp2 contamination filter / NDJSON.") |
|
|
| y = np.array(y, dtype=int) |
| pos = int(y.sum()) |
| print( |
| f"[DATA ] kept={len(records)} | positives={pos} ({100.0*pos/len(records):.2f}%) | " |
| f"groups={len(set(groups))} | dropped_contam={dropped_contam} | dropped_supp1={dropped_supp1}" |
| ) |
|
|
| outroot = Path(args.outdir) |
| outroot.mkdir(parents=True, exist_ok=True) |
|
|
| sgkf = StratifiedGroupKFold(n_splits=5, shuffle=True, random_state=args.seed) |
|
|
| for k in range(args.n_splits): |
| subset_dir = outroot / f"subset{k+1:02d}" |
| subset_dir.mkdir(parents=True, exist_ok=True) |
|
|
| idx = np.arange(len(records)) |
| tr_idx, te_idx = next(sgkf.split(idx, y, groups)) |
| train_records = [records[i] for i in tr_idx] |
| test_records = [records[i] for i in te_idx] |
|
|
| with open(subset_dir / "train.jsonl", "w") as ftr: |
| for r in train_records: |
| ftr.write(json.dumps(r, separators=(',', ':')) + "\n") |
| with open(subset_dir / "test.jsonl", "w") as fte: |
| for r in test_records: |
| fte.write(json.dumps(r, separators=(',', ':')) + "\n") |
|
|
| y_tr = y[tr_idx]; y_te = y[te_idx] |
| manifest = { |
| "n_total": int(len(records)), |
| "n_train": int(len(train_records)), |
| "n_test": int(len(test_records)), |
| "positives_total": int(y.sum()), |
| "positives_train": int(y_tr.sum()), |
| "positives_test": int(y_te.sum()), |
| "pct_pos_total": float(100.0 * y.sum() / len(records)), |
| "pct_pos_train": float(100.0 * y_tr.sum() / len(train_records)), |
| "pct_pos_test": float(100.0 * y_te.sum() / len(test_records)), |
| "groups_total": int(len(set(groups))), |
| "seed": int(args.seed + k), |
| "source_ndjson": str(ndjson_path.resolve()), |
| "supp2_contam_removed": int(len(contam_bases)), |
| } |
| (subset_dir / "manifest.json").write_text(json.dumps(manifest, indent=2)) |
|
|
| print(f"[WRITE] {subset_dir} | train={len(train_records)} test={len(test_records)} | pos_tr={int(y_tr.sum())} pos_te={int(y_te.sum())}") |
|
|
| print("[DONE ] All subsets written.") |
|
|
| if __name__ == "__main__": |
| main() |
|
|
|
|