transformers-inference / generate-responses.py
davanstrien's picture
davanstrien HF Staff
Add --messages-until-role to truncate messages for regeneration
439e257
# /// script
# requires-python = ">=3.10"
# dependencies = [
# "accelerate",
# "datasets",
# "huggingface-hub",
# "hf-xet",
# "torch",
# "tqdm",
# "transformers>=5.1",
# ]
# ///
"""
Generate responses for prompts in a dataset using transformers continuous batching.
Uses transformers' native continuous batching (CB) for efficient GPU inference.
No vLLM dependency required - works with any model supported by transformers,
including newly released architectures.
Example usage:
# Local execution
uv run generate-responses.py \\
username/input-dataset \\
username/output-dataset \\
--prompt-column question
# With custom model and sampling parameters
uv run generate-responses.py \\
username/input-dataset \\
username/output-dataset \\
--model-id meta-llama/Llama-3.1-8B-Instruct \\
--messages-column messages \\
--temperature 0.9 \\
--max-tokens 2048
# HF Jobs execution (see script output for full command)
hf jobs uv run --flavor l4x1 ...
"""
import argparse
import logging
import os
import sys
from datetime import datetime
from typing import Optional
import torch
from datasets import load_dataset
from huggingface_hub import DatasetCard, get_token, login
from transformers import AutoModelForCausalLM, AutoTokenizer
from transformers.generation import GenerationConfig
# Enable HF Transfer for faster downloads
os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
logging.basicConfig(
level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s"
)
logger = logging.getLogger(__name__)
def check_gpu_availability() -> int:
"""Check if CUDA is available and return the number of GPUs."""
if not torch.cuda.is_available():
logger.error("CUDA is not available. This script requires a GPU.")
logger.error(
"Please run on a machine with NVIDIA GPU or use HF Jobs with GPU flavor."
)
sys.exit(1)
num_gpus = torch.cuda.device_count()
for i in range(num_gpus):
gpu_name = torch.cuda.get_device_name(i)
gpu_memory = torch.cuda.get_device_properties(i).total_memory / 1024**3
logger.info(f"GPU {i}: {gpu_name} with {gpu_memory:.1f} GB memory")
return num_gpus
def create_dataset_card(
source_dataset: str,
model_id: str,
messages_column: str,
prompt_column: Optional[str],
generation_config: GenerationConfig,
num_gpus: int,
num_examples: int,
generation_time: str,
num_skipped: int = 0,
attn_implementation: str = "paged|sdpa",
) -> str:
"""Create a dataset card documenting the generation process."""
filtering_section = ""
if num_skipped > 0:
skip_percentage = (num_skipped / num_examples) * 100
processed = num_examples - num_skipped
filtering_section = f"""
### Filtering Statistics
- **Total Examples**: {num_examples:,}
- **Processed**: {processed:,} ({100 - skip_percentage:.1f}%)
- **Skipped (too long)**: {num_skipped:,} ({skip_percentage:.1f}%)
Note: Prompts exceeding the model's maximum context length were skipped and have empty responses."""
input_col = prompt_column if prompt_column else messages_column
input_type = "plain text prompts" if prompt_column else "chat messages"
return f"""---
tags:
- generated
- transformers
- continuous-batching
- uv-script
---
# Generated Responses Dataset
This dataset contains generated responses for prompts from [{source_dataset}](https://huggingface.co/datasets/{source_dataset}).
## Generation Details
- **Source Dataset**: [{source_dataset}](https://huggingface.co/datasets/{source_dataset})
- **Input Column**: `{input_col}` ({input_type})
- **Model**: [{model_id}](https://huggingface.co/{model_id})
- **Backend**: transformers continuous batching
- **Number of Examples**: {num_examples:,}
- **Generation Date**: {generation_time}{filtering_section}
### Generation Parameters
- **Temperature**: {generation_config.temperature}
- **Top P**: {generation_config.top_p}
- **Top K**: {generation_config.top_k}
- **Max New Tokens**: {generation_config.max_new_tokens}
- **Max Batch Tokens**: {generation_config.max_batch_tokens}
- **Repetition Penalty**: {generation_config.repetition_penalty}
### Hardware Configuration
- **GPUs**: {num_gpus}
- **Attention Implementation**: {attn_implementation}
## Dataset Structure
The dataset contains all columns from the source dataset plus:
- `response`: The generated response from the model
## Generation Script
Generated using the transformers continuous batching script from [uv-scripts/transformers-inference](https://huggingface.co/datasets/uv-scripts/transformers-inference).
To reproduce this generation:
```bash
uv run https://huggingface.co/datasets/uv-scripts/transformers-inference/raw/main/generate-responses.py \\
{source_dataset} \\
<output-dataset> \\
--model-id {model_id} \\
{"--prompt-column " + prompt_column if prompt_column else "--messages-column " + messages_column} \\
--temperature {generation_config.temperature} \\
--top-p {generation_config.top_p} \\
--top-k {generation_config.top_k} \\
--max-tokens {generation_config.max_new_tokens}
```
"""
def main(
src_dataset_hub_id: str,
output_dataset_hub_id: str,
model_id: str = "Qwen/Qwen3-4B-Instruct-2507",
messages_column: str = "messages",
prompt_column: Optional[str] = None,
output_column: str = "response",
temperature: float = 0.7,
top_p: float = 0.8,
top_k: int = 20,
max_tokens: int = 4096,
repetition_penalty: float = 1.0,
max_batch_tokens: int = 512,
dtype: str = "bfloat16",
attn_implementation: str = "paged|sdpa",
subset: Optional[str] = None,
split: str = "train",
messages_until_role: Optional[str] = None,
skip_long_prompts: bool = True,
max_samples: Optional[int] = None,
hf_token: Optional[str] = None,
):
generation_start_time = datetime.now().isoformat()
# GPU check
num_gpus = check_gpu_availability()
# Authentication
HF_TOKEN = hf_token or os.environ.get("HF_TOKEN") or get_token()
if not HF_TOKEN:
logger.error("No HuggingFace token found. Please provide token via:")
logger.error(" 1. --hf-token argument")
logger.error(" 2. HF_TOKEN environment variable")
logger.error(" 3. Run 'huggingface-cli login' or use login() in Python")
sys.exit(1)
logger.info("HuggingFace token found, authenticating...")
login(token=HF_TOKEN)
# Resolve dtype
torch_dtype = getattr(torch, dtype, None)
if torch_dtype is None:
logger.error(f"Unknown dtype: {dtype}. Use 'bfloat16', 'float16', or 'float32'.")
sys.exit(1)
# Load model with continuous batching support
# Note: CB currently requires single-GPU. Multi-GPU device_map="auto" causes
# device mismatch errors with PagedAttention cache. Use a model that fits on one GPU.
if num_gpus > 1:
logger.warning(
"Multiple GPUs detected but transformers CB currently works on single GPU only. "
"Using cuda:0. Choose a model that fits on one GPU."
)
device_map = "cuda"
logger.info(f"Loading model: {model_id} (dtype={dtype}, attn={attn_implementation}, device_map={device_map})")
model = AutoModelForCausalLM.from_pretrained(
model_id,
attn_implementation=attn_implementation,
device_map=device_map,
dtype=torch_dtype,
)
# Load tokenizer
logger.info("Loading tokenizer...")
tokenizer = AutoTokenizer.from_pretrained(model_id, padding_side="left")
# Create generation config
do_sample = temperature != 1.0 or top_p != 1.0 or top_k != 0
generation_config = GenerationConfig(
max_new_tokens=max_tokens,
max_batch_tokens=max_batch_tokens,
do_sample=do_sample,
temperature=temperature,
top_p=top_p,
top_k=top_k,
repetition_penalty=repetition_penalty,
eos_token_id=tokenizer.eos_token_id,
pad_token_id=tokenizer.pad_token_id,
)
# Load dataset
logger.info(f"Loading dataset: {src_dataset_hub_id}" + (f" ({subset})" if subset else ""))
dataset = load_dataset(src_dataset_hub_id, subset, split=split)
if max_samples is not None and max_samples < len(dataset):
logger.info(f"Limiting dataset to {max_samples} samples")
dataset = dataset.select(range(max_samples))
total_examples = len(dataset)
logger.info(f"Dataset loaded with {total_examples:,} examples")
# Validate column
if prompt_column:
if prompt_column not in dataset.column_names:
logger.error(
f"Column '{prompt_column}' not found. Available columns: {dataset.column_names}"
)
sys.exit(1)
logger.info(f"Using prompt column: '{prompt_column}'")
use_messages = False
else:
if messages_column not in dataset.column_names:
logger.error(
f"Column '{messages_column}' not found. Available columns: {dataset.column_names}"
)
sys.exit(1)
logger.info(f"Using messages column: '{messages_column}'")
use_messages = True
# Get model's max context length for filtering
effective_max_len = model.config.max_position_embeddings
logger.info(f"Model max context length: {effective_max_len}")
# Prepare prompts and tokenize
logger.info("Preparing and tokenizing prompts...")
valid_input_ids = []
valid_indices = []
skipped_info = []
for i, example in enumerate(dataset):
if use_messages:
messages = example[messages_column]
# Optionally truncate messages up to last occurrence of a given role
if messages_until_role:
truncated = []
for msg in messages:
truncated.append(msg)
if msg["role"] == messages_until_role:
last_truncated = list(truncated)
messages = last_truncated
prompt = tokenizer.apply_chat_template(
messages, tokenize=False, add_generation_prompt=True
)
else:
user_prompt = example[prompt_column]
messages = [{"role": "user", "content": user_prompt}]
prompt = tokenizer.apply_chat_template(
messages, tokenize=False, add_generation_prompt=True
)
input_ids = tokenizer.encode(prompt)
if skip_long_prompts:
if len(input_ids) <= effective_max_len:
valid_input_ids.append(input_ids)
valid_indices.append(i)
else:
skipped_info.append((i, len(input_ids)))
else:
valid_input_ids.append(input_ids)
valid_indices.append(i)
# Log filtering results
if skip_long_prompts and skipped_info:
logger.warning(
f"Skipped {len(skipped_info)} prompts exceeding context length ({effective_max_len} tokens)"
)
for idx, (prompt_idx, token_count) in enumerate(skipped_info[:10]):
logger.info(
f" - Example {prompt_idx}: {token_count} tokens (exceeds by {token_count - effective_max_len})"
)
if len(skipped_info) > 10:
logger.info(f" ... and {len(skipped_info) - 10} more")
skip_percentage = (len(skipped_info) / total_examples) * 100
if skip_percentage > 10:
logger.warning(f"WARNING: {skip_percentage:.1f}% of prompts were skipped!")
if not valid_input_ids:
logger.error("No valid prompts to process after filtering!")
sys.exit(1)
# Generate responses using continuous batching
logger.info(f"Starting generation for {len(valid_input_ids):,} prompts using continuous batching...")
logger.info(f"max_batch_tokens={max_batch_tokens}, max_new_tokens={max_tokens}")
batch_outputs = model.generate_batch(
inputs=valid_input_ids,
generation_config=generation_config,
progress_bar=True,
)
# Extract generated text
logger.info("Extracting generated responses...")
responses = [""] * total_examples
for request_id, output in batch_outputs.items():
# request_id is formatted as "req_0", "req_1", etc.
idx = int(request_id.split("_", 1)[1])
original_idx = valid_indices[idx]
generated_text = tokenizer.decode(
output.generated_tokens, skip_special_tokens=True
)
responses[original_idx] = generated_text.strip()
# Count non-empty responses
non_empty = sum(1 for r in responses if r)
logger.info(f"Generated {non_empty:,} non-empty responses out of {total_examples:,} total")
# Add responses to dataset
logger.info("Adding responses to dataset...")
dataset = dataset.add_column(output_column, responses)
# Create dataset card
logger.info("Creating dataset card...")
card_content = create_dataset_card(
source_dataset=src_dataset_hub_id,
model_id=model_id,
messages_column=messages_column,
prompt_column=prompt_column,
generation_config=generation_config,
num_gpus=num_gpus,
num_examples=total_examples,
generation_time=generation_start_time,
num_skipped=len(skipped_info) if skip_long_prompts else 0,
attn_implementation=attn_implementation,
)
# Push to Hub
logger.info(f"Pushing dataset to: {output_dataset_hub_id}")
dataset.push_to_hub(output_dataset_hub_id, token=HF_TOKEN)
card = DatasetCard(card_content)
card.push_to_hub(output_dataset_hub_id, token=HF_TOKEN)
logger.info("Generation complete!")
logger.info(
f"Dataset available at: https://huggingface.co/datasets/{output_dataset_hub_id}"
)
if __name__ == "__main__":
if len(sys.argv) > 1:
parser = argparse.ArgumentParser(
description="Generate responses using transformers continuous batching",
formatter_class=argparse.RawDescriptionHelpFormatter,
epilog="""
Examples:
# Basic usage with default Qwen model
uv run generate-responses.py input-dataset output-dataset \\
--prompt-column question
# With custom model and parameters
uv run generate-responses.py input-dataset output-dataset \\
--model-id meta-llama/Llama-3.1-8B-Instruct \\
--messages-column messages \\
--temperature 0.9 \\
--max-tokens 2048
# Increase batch token budget for better GPU utilization
uv run generate-responses.py input-dataset output-dataset \\
--prompt-column text \\
--max-batch-tokens 2048
""",
)
parser.add_argument(
"src_dataset_hub_id",
help="Input dataset on Hugging Face Hub (e.g., username/dataset-name)",
)
parser.add_argument(
"output_dataset_hub_id",
help="Output dataset name on Hugging Face Hub",
)
parser.add_argument(
"--model-id",
type=str,
default="Qwen/Qwen3-4B-Instruct-2507",
help="Model to use for generation (default: Qwen3-4B-Instruct-2507)",
)
parser.add_argument(
"--subset",
type=str,
help="Dataset subset/config name (e.g., 'SFT' for HuggingFaceTB/smoltalk2)",
)
parser.add_argument(
"--split",
type=str,
default="train",
help="Dataset split to use (default: train)",
)
parser.add_argument(
"--messages-column",
type=str,
default="messages",
help="Column containing chat messages (default: messages)",
)
parser.add_argument(
"--prompt-column",
type=str,
help="Column containing plain text prompts (alternative to --messages-column)",
)
parser.add_argument(
"--messages-until-role",
type=str,
help="Truncate messages up to the last message with this role. "
"Use 'user' to strip existing assistant responses and regenerate them.",
)
parser.add_argument(
"--output-column",
type=str,
default="response",
help="Column name for generated responses (default: response)",
)
parser.add_argument(
"--max-samples",
type=int,
help="Maximum number of samples to process (default: all)",
)
parser.add_argument(
"--temperature",
type=float,
default=0.7,
help="Sampling temperature (default: 0.7)",
)
parser.add_argument(
"--top-p",
type=float,
default=0.8,
help="Top-p sampling parameter (default: 0.8)",
)
parser.add_argument(
"--top-k",
type=int,
default=20,
help="Top-k sampling parameter (default: 20)",
)
parser.add_argument(
"--max-tokens",
type=int,
default=4096,
help="Maximum tokens to generate (default: 4096)",
)
parser.add_argument(
"--repetition-penalty",
type=float,
default=1.0,
help="Repetition penalty (default: 1.0)",
)
parser.add_argument(
"--max-batch-tokens",
type=int,
default=512,
help="Token budget per batch for continuous batching scheduler (default: 512). "
"Increase for better GPU utilization on large GPUs (e.g., 2048-4096 for A100/H100).",
)
parser.add_argument(
"--dtype",
type=str,
default="bfloat16",
choices=["bfloat16", "float16", "float32"],
help="Model dtype (default: bfloat16)",
)
parser.add_argument(
"--attn-implementation",
type=str,
default="paged|sdpa",
help="Attention implementation (default: paged|sdpa). "
"Use 'paged|flash_attention_2' if flash-attn is installed.",
)
parser.add_argument(
"--hf-token",
type=str,
help="Hugging Face token (can also use HF_TOKEN env var)",
)
parser.add_argument(
"--skip-long-prompts",
action="store_true",
default=True,
help="Skip prompts exceeding context length (default: True)",
)
parser.add_argument(
"--no-skip-long-prompts",
dest="skip_long_prompts",
action="store_false",
help="Fail on prompts that exceed context length",
)
args = parser.parse_args()
main(
src_dataset_hub_id=args.src_dataset_hub_id,
output_dataset_hub_id=args.output_dataset_hub_id,
model_id=args.model_id,
messages_column=args.messages_column,
prompt_column=args.prompt_column,
messages_until_role=args.messages_until_role,
output_column=args.output_column,
temperature=args.temperature,
top_p=args.top_p,
top_k=args.top_k,
max_tokens=args.max_tokens,
repetition_penalty=args.repetition_penalty,
max_batch_tokens=args.max_batch_tokens,
dtype=args.dtype,
attn_implementation=args.attn_implementation,
subset=args.subset,
split=args.split,
skip_long_prompts=args.skip_long_prompts,
max_samples=args.max_samples,
hf_token=args.hf_token,
)
else:
print("""
Transformers Continuous Batching - Response Generation
======================================================
This script requires arguments. For usage information:
uv run generate-responses.py --help
Why transformers CB instead of vLLM?
- Works with ANY model supported by transformers (new models immediately!)
- No vLLM/flashinfer dependency issues
- Simpler setup - no custom Docker images or wheel indexes needed
- ~95% of vLLM throughput with PagedAttention and continuous scheduling
Example HF Jobs command:
hf jobs uv run \\
--flavor l4x1 \\
-s HF_TOKEN \\
https://huggingface.co/datasets/uv-scripts/transformers-inference/raw/main/generate-responses.py \\
username/input-dataset \\
username/output-dataset \\
--prompt-column question \\
--model-id Qwen/Qwen3-4B-Instruct-2507 \\
--temperature 0.7 \\
--max-tokens 4096
""")