| | """ |
| | NERPA – Text anonymisation using the fine-tuned GLiNER2 model. |
| | |
| | Usage: |
| | python anonymise.py "My name is John Smith, born 15/03/1990. Email: john@example.com" |
| | python anonymise.py --file input.txt |
| | python anonymise.py --file input.txt --output anonymised.txt |
| | """ |
| |
|
| | import argparse |
| | import logging |
| | import sys |
| | import warnings |
| | from typing import Optional |
| |
|
| | warnings.filterwarnings("ignore", message=r".*incorrect regex pattern.*fix_mistral_regex.*") |
| |
|
| | import torch |
| | from gliner2 import GLiNER2 |
| |
|
| | logger = logging.getLogger(__name__) |
| |
|
| | |
| | |
| | PII_ENTITIES: dict[str, str] = { |
| | "LOCATION": "Address, country, city, postcode, street, any other location", |
| | "AGE": "Age of a person", |
| | "DIGITAL_KEYS": "Digital keys, passwords, pins used to access anything like servers, banks, APIs, accounts etc", |
| | "BANK_ACCOUNT_DETAILS": "Bank account details such as number, IBAN, SWIFT, routing numbers etc", |
| | "CARD_DETAILS": "Debit or credit card details such as card number, CVV, expiration etc", |
| | "DATE_TIME": "Generic date and time", |
| | "DATE_OF_BIRTH": "Date of birth", |
| | "PERSONAL_ID_NUMBERS": "Common personal identification numbers such as passport numbers, driving licenses, taxpayer and insurance numbers", |
| | "TECHNICAL_ID_NUMBERS": "IP and MAC addresses, serial numbers and any other technical ID numbers", |
| | "EMAIL": "Email", |
| | "PERSON_NAME": "Person name", |
| | "BUSINESS_NAME": "Business name", |
| | "PHONE": "Any personal or other phone numbers", |
| | "URL": "Any short or full URL", |
| | "USERNAME": "Username", |
| | "VEHICLE_ID_NUMBERS": "Any vehicle numbers like license plates, vehicle identification numbers", |
| | } |
| |
|
| | CONFIDENCE_THRESHOLD = 0.25 |
| | CHUNK_SIZE = 3000 |
| | CHUNK_OVERLAP = 100 |
| | BATCH_SIZE = 32 |
| |
|
| |
|
| | def load_model(model_path: str = ".") -> GLiNER2: |
| | """Load the NERPA model onto the best available device.""" |
| | if torch.cuda.is_available(): |
| | device = torch.device("cuda") |
| | elif torch.backends.mps.is_available(): |
| | device = torch.device("mps") |
| | else: |
| | device = torch.device("cpu") |
| |
|
| | model = GLiNER2.from_pretrained(model_path) |
| | try: |
| | model.to(device) |
| | except RuntimeError: |
| | logger.warning( |
| | "Failed to load model on %s, falling back to CPU.", device |
| | ) |
| | model.to(torch.device("cpu")) |
| | return model |
| |
|
| |
|
| | def chunk_text( |
| | text: str, |
| | chunk_size: int = CHUNK_SIZE, |
| | overlap: int = CHUNK_OVERLAP, |
| | ) -> tuple[list[str], list[int]]: |
| | """Split text into overlapping chunks, returning chunks and their start offsets.""" |
| | if not text: |
| | return [], [] |
| | chunks: list[str] = [] |
| | starts: list[int] = [] |
| | step = chunk_size - overlap |
| | for pos in range(0, len(text), step): |
| | chunks.append(text[pos : pos + chunk_size]) |
| | starts.append(pos) |
| | return chunks, starts |
| |
|
| |
|
| | def detect_entities( |
| | model: GLiNER2, |
| | text: str, |
| | entities: Optional[dict[str, str]] = None, |
| | threshold: float = CONFIDENCE_THRESHOLD, |
| | ) -> list[dict]: |
| | """ |
| | Detect PII entities in text, returning a list of |
| | ``{"type": str, "start": int, "end": int, "score": float}`` dicts |
| | with character offsets into the original text. |
| | """ |
| | entities = entities or PII_ENTITIES |
| |
|
| | |
| | detect = dict(entities) |
| | if "DATE_TIME" in detect and "DATE_OF_BIRTH" not in detect: |
| | detect["DATE_OF_BIRTH"] = PII_ENTITIES["DATE_OF_BIRTH"] |
| | elif "DATE_OF_BIRTH" in detect and "DATE_TIME" not in detect: |
| | detect["DATE_TIME"] = PII_ENTITIES["DATE_TIME"] |
| |
|
| | chunks, offsets = chunk_text(text) |
| |
|
| | all_chunk_results: list[dict] = [] |
| | for batch_start in range(0, len(chunks), BATCH_SIZE): |
| | batch = chunks[batch_start : batch_start + BATCH_SIZE] |
| | results = model.batch_extract_entities( |
| | batch, |
| | detect, |
| | include_confidence=True, |
| | include_spans=True, |
| | threshold=threshold, |
| | ) |
| | all_chunk_results.extend(results) |
| |
|
| | |
| | seen: dict[tuple[int, int], dict] = {} |
| | for chunk_result, chunk_offset in zip(all_chunk_results, offsets): |
| | for label, occurrences in chunk_result["entities"].items(): |
| | for occurrence in occurrences: |
| | start = occurrence["start"] + chunk_offset |
| | end = occurrence["end"] + chunk_offset |
| | position = (start, end) |
| | if ( |
| | position not in seen |
| | or seen[position]["score"] < occurrence["confidence"] |
| | ): |
| | seen[position] = { |
| | "type": label, |
| | "score": occurrence["confidence"], |
| | } |
| |
|
| | |
| | |
| | |
| | items = sorted( |
| | [ |
| | (start, end, info) |
| | for (start, end), info in seen.items() |
| | if info["type"] in entities |
| | ], |
| | key=lambda x: (x[0], x[1]), |
| | ) |
| | if not items: |
| | return [] |
| |
|
| | merged: list[dict] = [] |
| | current_start, current_end, current_info = items[0] |
| | for start, end, info in items[1:]: |
| | if start < current_end: |
| | current_end = max(current_end, end) |
| | if info["score"] > current_info["score"]: |
| | current_info = info |
| | else: |
| | merged.append({ |
| | "type": current_info["type"], |
| | "start": current_start, |
| | "end": current_end, |
| | "score": current_info["score"], |
| | }) |
| | current_start, current_end, current_info = start, end, info |
| | merged.append({ |
| | "type": current_info["type"], |
| | "start": current_start, |
| | "end": current_end, |
| | "score": current_info["score"], |
| | }) |
| |
|
| | return merged |
| |
|
| |
|
| | def anonymise(text: str, detected: list[dict]) -> str: |
| | """Replace detected entities with placeholders like ``[PERSON_NAME]``.""" |
| | parts: list[str] = [] |
| | prev_end = 0 |
| | for entity in sorted(detected, key=lambda e: e["start"]): |
| | parts.append(text[prev_end : entity["start"]]) |
| | parts.append(f'[{entity["type"]}]') |
| | prev_end = entity["end"] |
| | parts.append(text[prev_end:]) |
| | return "".join(parts) |
| |
|
| |
|
| | def main() -> None: |
| | parser = argparse.ArgumentParser( |
| | description="Anonymise PII in text using the NERPA model.", |
| | ) |
| | parser.add_argument( |
| | "text", nargs="?", help="Text to anonymise (or use --file)", |
| | ) |
| | parser.add_argument( |
| | "--file", "-f", help="Read text from a file instead", |
| | ) |
| | parser.add_argument( |
| | "--output", "-o", |
| | help="Write anonymised text to file (default: stdout)", |
| | ) |
| | parser.add_argument( |
| | "--model", "-m", default=".", |
| | help="Path to model directory (default: current dir)", |
| | ) |
| | parser.add_argument( |
| | "--threshold", "-t", type=float, default=CONFIDENCE_THRESHOLD, |
| | help=f"Confidence threshold (default: {CONFIDENCE_THRESHOLD})", |
| | ) |
| | parser.add_argument( |
| | "--show-entities", action="store_true", |
| | help="Print detected entities before anonymised text", |
| | ) |
| | parser.add_argument( |
| | "--extra-entities", "-e", action="append", metavar="LABEL=DESCRIPTION", |
| | help=( |
| | "Additional custom entity types to detect alongside the built-in " |
| | "PII entities. Repeat for each type. Format: LABEL=\"Description\". " |
| | "Example: -e PRODUCT=\"Product name\" -e SKILL=\"Professional skill\"" |
| | ), |
| | ) |
| | args = parser.parse_args() |
| |
|
| | if args.file: |
| | try: |
| | with open(args.file, encoding="utf-8") as f: |
| | text = f.read() |
| | except OSError as exc: |
| | sys.exit(f"Error reading {args.file}: {exc}") |
| | elif args.text: |
| | text = args.text |
| | else: |
| | parser.error("Provide text as an argument or use --file") |
| |
|
| | extra: dict[str, str] = {} |
| | if args.extra_entities: |
| | for item in args.extra_entities: |
| | if "=" not in item: |
| | parser.error( |
| | f"Invalid --extra-entities value '{item}'. " |
| | "Expected format: LABEL=\"Description\"" |
| | ) |
| | label, description = item.split("=", 1) |
| | extra[label.strip()] = description.strip() |
| |
|
| | model = load_model(args.model) |
| | all_entities = {**PII_ENTITIES, **extra} if extra else None |
| | detected = detect_entities(model, text, entities=all_entities, threshold=args.threshold) |
| |
|
| | if args.show_entities: |
| | for entity in detected: |
| | span = text[entity["start"] : entity["end"]] |
| | logger.info( |
| | " %-25s [%5d:%5d] (score=%.2f) %r", |
| | entity["type"], entity["start"], entity["end"], |
| | entity["score"], span, |
| | ) |
| |
|
| | result = anonymise(text, detected) |
| |
|
| | if args.output: |
| | try: |
| | with open(args.output, "w", encoding="utf-8") as f: |
| | f.write(result) |
| | except OSError as exc: |
| | sys.exit(f"Error writing {args.output}: {exc}") |
| | else: |
| | print(result) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|