YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
=========================
ORIGINAL MODEL (Transformers) β README.md (FINAL)
Repo: WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B
=========================
language:
- en
library_name: transformers pipeline_tag: text-generation
tags:
- gpt2
- causal-lm
- text-generation
- code
- coding
- reasoning
- instruct
- lightweight
- safetensors
- withinusai
license: other license_name: withinusai-custom-license license_link: LICENSE
base_model: openai-community/gpt2-medium base_model_relation: finetune
datasets:
- WithinUsAI/GPT-2-to-GPT-5-5k
- TeichAI/gpt-5.1-codex-max-1000x
- TeichAI/gpt-5.1-high-reasoning-1000x
metrics:
- pass@1
- accuracy
- exact_match
model-index: - name: WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B results: []
WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B
GPT-2 Medium enhanced toward GPT-5.2-style reasoning + codex behavior.
Small footprint. Built to ship working code. β‘π§
What βGPT2.5.2β means (project naming)
This model begins as GPT-2 Medium and is fine-tuned by WithIn Us AI with the goal of pushing behavior toward a GPT-5.2 βtwin targetβ style: stronger stepwise reasoning, more reliable code generation, and improved instruction-following.
- GPT(2) = GPT-2 Medium base
- GPT(5.2) = target behavior style (reasoning + codex competence)
- GPT(2.5.2) = WithIn Us AI enhanced release line/version marker
Model details
- Model type: Decoder-only causal language model (GPT-2 family)
- Architecture: gpt2
- Size class: ~0.4B parameters (approx.)
- Base model:
openai-community/gpt2-medium - Base model relation: fine-tune
- Primary strengths: coding assistance, refactors, debugging, structured reasoning
Intended use
Recommended β
- Code generation & completion (Python-first; multi-language ok)
- Debugging: error β root cause β patch
- Refactoring: preserve behavior, improve clarity/perf
- Stepwise technical reasoning with constraints and edge cases
Not recommended π«
- High-stakes decisions (medical/legal/financial) without expert review
- Safety-critical systems without strict validation & monitoring
Quickstart (Transformers)
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "WithinUsAI/GPT2.5.2-high-reasoning-codex-0.4B"
tokenizer = AutoTokenizer.from_pretrained(model_id, use_fast=True)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
device_map="auto"
)
prompt = (
"You are a senior software engineer.\n"
"Task: Implement a robust JSONL reader in Python.\n"
"First list edge cases, then write the implementation with comments.\n\n"
"Answer:\n"
)
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
out = model.generate(
**inputs,
max_new_tokens=256,
do_sample=True,
temperature=0.7,
top_p=0.95
)
print(tokenizer.decode(out[0], skip_special_tokens=True))
- Downloads last month
- 167
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support