| | --- |
| | license: apache-2.0 |
| | base_model: ibm-granite/granite-20b-code-instruct-8k |
| | tags: |
| | - code |
| | - security |
| | - granite |
| | - ibm |
| | - securecode |
| | - owasp |
| | - vulnerability-detection |
| | datasets: |
| | - scthornton/securecode-v2 |
| | language: |
| | - en |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | arxiv: 2512.18542 |
| | --- |
| | |
| | # IBM Granite 20B Code - SecureCode Edition |
| |
|
| | <div align="center"> |
| |
|
| | [](https://opensource.org/licenses/Apache-2.0) |
| | [](https://huggingface.co/datasets/scthornton/securecode-v2) |
| | [](https://huggingface.co/ibm-granite/granite-20b-code-instruct-8k) |
| | [](https://perfecxion.ai) |
| |
|
| | **π’ Enterprise-scale security intelligence with IBM trust** |
| |
|
| | The most powerful model in the SecureCode collection. When you need maximum code understanding, complex reasoning, and IBM's enterprise-grade reliability. |
| |
|
| | [π€ Model Hub](https://huggingface.co/scthornton/granite-20b-code-securecode) | [π Dataset](https://huggingface.co/datasets/scthornton/securecode-v2) | [π» perfecXion.ai](https://perfecxion.ai) | [π Collection](https://huggingface.co/collections/scthornton/securecode) |
| |
|
| | </div> |
| |
|
| | --- |
| |
|
| | ## π― Quick Decision Guide |
| |
|
| | **Choose This Model If:** |
| | - β
You need **maximum code understanding** and security reasoning capability |
| | - β
You're analyzing **complex enterprise architectures** with intricate attack surfaces |
| | - β
You require **IBM enterprise trust** and brand recognition |
| | - β
You have **datacenter infrastructure** (48GB+ GPU) |
| | - β
You're conducting **professional security audits** requiring comprehensive analysis |
| | - β
You need the **most sophisticated** security intelligence in the collection |
| |
|
| | **Consider Smaller Models If:** |
| | - β οΈ You're on consumer hardware (β Llama 3B, Qwen 7B) |
| | - β οΈ You prioritize inference speed over depth (β Qwen 7B/14B) |
| | - β οΈ You're building IDE tools needing fast response (β Llama 3B, DeepSeek 6.7B) |
| | - β οΈ Budget is primary concern (β any 7B/13B model) |
| |
|
| | --- |
| |
|
| | ## π Collection Positioning |
| |
|
| | | Model | Size | Best For | Hardware | Inference Speed | Unique Strength | |
| | |-------|------|----------|----------|-----------------|-----------------| |
| | | Llama 3.2 3B | 3B | Consumer deployment | 8GB RAM | β‘β‘β‘ Fastest | Most accessible | |
| | | DeepSeek 6.7B | 6.7B | Security-optimized baseline | 16GB RAM | β‘β‘ Fast | Security architecture | |
| | | Qwen 7B | 7B | Best code understanding | 16GB RAM | β‘β‘ Fast | Best-in-class 7B | |
| | | CodeGemma 7B | 7B | Google ecosystem | 16GB RAM | β‘β‘ Fast | Instruction following | |
| | | CodeLlama 13B | 13B | Enterprise trust | 24GB RAM | β‘ Medium | Meta brand, proven | |
| | | Qwen 14B | 14B | Advanced analysis | 32GB RAM | β‘ Medium | 128K context window | |
| | | StarCoder2 15B | 15B | Multi-language specialist | 32GB RAM | β‘ Medium | 600+ languages | |
| | | **Granite 20B** | **20B** | **Enterprise-scale** | **48GB RAM** | **Medium** | **IBM trust, largest, most capable** | |
| |
|
| | **This Model's Position:** The flagship. Maximum security intelligence, enterprise-grade reliability, IBM brand trust. For when quality matters more than speed. |
| |
|
| | --- |
| |
|
| | ## π¨ The Problem This Solves |
| |
|
| | **Critical enterprise security gaps require sophisticated analysis.** When a breach costs **$4.45 million on average** (IBM 2024 Cost of Data Breach Report) and 45% of AI-generated code contains vulnerabilities, enterprises need the most capable security analysis available. |
| |
|
| | **Real-world enterprise impact:** |
| | - **Equifax** (SQL injection): $425 million settlement + 13-year brand recovery |
| | - **Capital One** (SSRF): 100 million customer records, $80M fine, 2 years of remediation |
| | - **SolarWinds** (supply chain): 18,000 organizations compromised, $18M settlement |
| | - **LastPass** (cryptographic failures): 30M users affected, company reputation destroyed |
| |
|
| | **IBM Granite 20B SecureCode Edition** provides the deepest security analysis available in the open-source ecosystem, backed by IBM's enterprise heritage and trust. |
| |
|
| | --- |
| |
|
| | ## π‘ What is This? |
| |
|
| | This is **IBM Granite 20B Code Instruct** fine-tuned on the **SecureCode v2.0 dataset** - IBM's enterprise-grade code model enhanced with production-grade security expertise covering the complete OWASP Top 10:2025. |
| |
|
| | IBM Granite models are built on IBM's 40+ years of enterprise software experience, trained on **3.5+ trillion tokens** of code and technical data, with a focus on enterprise deployment reliability. |
| |
|
| | Combined with SecureCode training, this model delivers: |
| |
|
| | β
**Maximum security intelligence** - 20B parameters for deep, nuanced analysis |
| | β
**Enterprise-grade reliability** - IBM's proven track record and support ecosystem |
| | β
**Comprehensive vulnerability detection** across complex architectures |
| | β
**Production-ready trust** - Permissive Apache 2.0 license |
| | β
**Advanced reasoning** - Handles multi-layered attack chain analysis |
| |
|
| | **The Result:** The most capable security-aware code model in the open-source ecosystem. |
| |
|
| | **Why IBM Granite 20B?** This model is the enterprise choice: |
| | - π’ **IBM enterprise heritage** - 40+ years of enterprise software leadership |
| | - π **Largest in collection** - 20B parameters = maximum reasoning capability |
| | - π **Enterprise compliance ready** - Designed for regulated industries |
| | - βοΈ **Apache 2.0 licensed** - Full commercial freedom |
| | - π― **Security-first training** - Built for mission-critical applications |
| | - π **Broad language support** - 116+ programming languages |
| |
|
| | Perfect for Fortune 500 companies, financial services, healthcare, government, and any organization where security analysis quality is paramount. |
| |
|
| | --- |
| |
|
| | ## π Security Training Coverage |
| |
|
| | ### Real-World Vulnerability Distribution |
| |
|
| | Trained on 1,209 security examples with real CVE grounding: |
| |
|
| | | OWASP Category | Examples | Real Incidents | |
| | |----------------|----------|----------------| |
| | | **Broken Access Control** | 224 | Equifax, Facebook, Uber | |
| | | **Authentication Failures** | 199 | SolarWinds, Okta, LastPass | |
| | | **Injection Attacks** | 125 | Capital One, Yahoo, LinkedIn | |
| | | **Cryptographic Failures** | 115 | LastPass, Adobe, Dropbox | |
| | | **Security Misconfiguration** | 98 | Tesla, MongoDB, Elasticsearch | |
| | | **Vulnerable Components** | 87 | Log4Shell, Heartbleed, Struts | |
| | | **Identification/Auth Failures** | 84 | Twitter, GitHub, Reddit | |
| | | **Software/Data Integrity** | 78 | SolarWinds, Codecov, npm | |
| | | **Logging Failures** | 71 | Various incident responses | |
| | | **SSRF** | 69 | Capital One, Shopify | |
| | | **Insecure Design** | 59 | Architectural flaws | |
| |
|
| | ### Enterprise-Grade Multi-Language Support |
| |
|
| | Fine-tuned on security examples across: |
| | - **Python** (Django, Flask, FastAPI) - 280 examples |
| | - **JavaScript/TypeScript** (Express, NestJS, React) - 245 examples |
| | - **Java** (Spring Boot, Jakarta EE) - 178 examples |
| | - **Go** (Gin, Echo, standard library) - 145 examples |
| | - **PHP** (Laravel, Symfony) - 112 examples |
| | - **C#** (ASP.NET Core, .NET 6+) - 89 examples |
| | - **Ruby** (Rails, Sinatra) - 67 examples |
| | - **Rust** (Actix, Rocket, Axum) - 45 examples |
| | - **C/C++** (Memory safety patterns) - 28 examples |
| | - **Plus 107+ additional languages from Granite's base training** |
| |
|
| | --- |
| |
|
| | ## π― Deployment Scenarios |
| |
|
| | ### Scenario 1: Enterprise Security Audit Platform |
| |
|
| | **Professional security assessments for Fortune 500 clients.** |
| |
|
| | **Hardware:** Datacenter GPU (A100 80GB or 2x A100 40GB) |
| | **Throughput:** 10-15 comprehensive audits/day |
| | **Use Case:** Professional security consulting |
| |
|
| | **Value Proposition:** |
| | - Identify vulnerabilities human auditors miss |
| | - Consistent, comprehensive OWASP coverage |
| | - Scales expert security knowledge |
| | - Reduces audit time by 60-70% |
| |
|
| | **ROI:** A single prevented breach pays for years of infrastructure. Typical large enterprise security audit costs $150K-500K. This model can handle preliminary analysis, allowing human experts to focus on novel vulnerabilities and strategic recommendations. |
| |
|
| | --- |
| |
|
| | ### Scenario 2: Financial Services Security Platform |
| |
|
| | **Regulatory compliance and security for banking applications.** |
| |
|
| | **Hardware:** Private cloud A100 cluster |
| | **Compliance:** SOC 2, PCI-DSS, GDPR, CCPA |
| | **Use Case:** Pre-deployment security validation |
| |
|
| | **Regulatory Benefits:** |
| | - Automated OWASP Top 10 verification |
| | - Audit trail generation |
| | - Compliance report automation |
| | - Reduces regulatory risk |
| |
|
| | **ROI:** Regulatory fines cost millions. **Capital One:** $80M fine. **Equifax:** $425M settlement. Preventing one major breach justifies entire deployment. |
| |
|
| | --- |
| |
|
| | ### Scenario 3: Healthcare Application Security |
| |
|
| | **HIPAA-compliant code review for medical systems.** |
| |
|
| | **Hardware:** Secure private deployment |
| | **Compliance:** HIPAA, HITECH, FDA software validation |
| | **Use Case:** Medical device and EHR security |
| |
|
| | **Critical Healthcare Requirements:** |
| | - Patient data protection (HIPAA) |
| | - Audit logging and compliance |
| | - Cryptographic requirements |
| | - Access control verification |
| |
|
| | **Impact:** Healthcare breaches average **$10.93 million per incident** (IBM 2024). Single prevented breach pays for multi-year deployment. |
| |
|
| | --- |
| |
|
| | ### Scenario 4: Government & Defense Applications |
| |
|
| | **Security analysis for critical infrastructure.** |
| |
|
| | **Hardware:** Air-gapped secure environment |
| | **Clearance:** Can be deployed in classified environments |
| | **Use Case:** Critical infrastructure security |
| |
|
| | **Government Benefits:** |
| | - No external dependencies (fully local) |
| | - Apache 2.0 license (government-friendly) |
| | - IBM enterprise support available |
| | - Meets government security standards |
| |
|
| | --- |
| |
|
| | ## π Training Details |
| |
|
| | | Parameter | Value | Why This Matters | |
| | |-----------|-------|------------------| |
| | | **Base Model** | ibm-granite/granite-20b-code-instruct-8k | IBM's enterprise-grade foundation | |
| | | **Fine-tuning Method** | LoRA (Low-Rank Adaptation) | Efficient training, preserves base capabilities | |
| | | **Training Dataset** | [SecureCode v2.0](https://huggingface.co/datasets/scthornton/securecode-v2) | 100% incident-grounded, expert-validated | |
| | | **Dataset Size** | 841 training examples | Focused on quality over quantity | |
| | | **Training Epochs** | 3 | Optimal convergence without overfitting | |
| | | **LoRA Rank (r)** | 16 | Balanced parameter efficiency | |
| | | **LoRA Alpha** | 32 | Learning rate scaling factor | |
| | | **Learning Rate** | 2e-4 | Standard for LoRA fine-tuning | |
| | | **Quantization** | 4-bit (bitsandbytes) | Enables efficient training | |
| | | **Trainable Parameters** | ~105M (0.525% of 20B total) | Minimal parameters, maximum impact | |
| | | **Total Parameters** | 20B | Maximum reasoning capability | |
| | | **Context Window** | 8K tokens | Enterprise file analysis | |
| | | **GPU Used** | NVIDIA A100 40GB | Enterprise training infrastructure | |
| | | **Training Time** | ~12-14 hours (estimated) | Deep security learning | |
| |
|
| | ### Training Methodology |
| |
|
| | **LoRA (Low-Rank Adaptation)** was chosen for enterprise reliability: |
| | 1. **Efficiency:** Trains only 0.525% of model parameters (105M vs 20B) |
| | 2. **Quality:** Preserves IBM Granite's enterprise capabilities |
| | 3. **Deployability:** Can be deployed alongside base model for versioning |
| |
|
| | **4-bit Quantization** enables efficient training while maintaining enterprise-grade quality. |
| |
|
| | **IBM Granite Foundation:** Built on IBM's 40+ years of enterprise software experience, optimized for: |
| | - Reliability and consistency |
| | - Enterprise deployment patterns |
| | - Regulatory compliance requirements |
| | - Long-term support and stability |
| |
|
| | --- |
| |
|
| | ## π Usage |
| |
|
| | ### Quick Start |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | from peft import PeftModel |
| | |
| | # Load IBM Granite base model |
| | base_model = "ibm-granite/granite-20b-code-instruct-8k" |
| | model = AutoModelForCausalLM.from_pretrained( |
| | base_model, |
| | device_map="auto", |
| | torch_dtype="auto", |
| | trust_remote_code=True |
| | ) |
| | tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True) |
| | |
| | # Load SecureCode LoRA adapter |
| | model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
| | |
| | # Enterprise security analysis |
| | prompt = """### User: |
| | Conduct a comprehensive security audit of this enterprise authentication system. Analyze for: |
| | 1. OWASP Top 10 vulnerabilities |
| | 2. Attack chain opportunities |
| | 3. Compliance gaps (SOC 2, PCI-DSS) |
| | 4. Architectural weaknesses |
| | |
| | ```python |
| | # Enterprise SSO Implementation |
| | class EnterpriseAuthService: |
| | def __init__(self): |
| | self.secret = os.getenv('JWT_SECRET') |
| | self.db = DatabasePool() |
| | |
| | async def authenticate(self, credentials): |
| | user = await self.db.query( |
| | f"SELECT * FROM users WHERE email='{credentials.email}' AND password='{credentials.password}'" |
| | ) |
| | if user: |
| | token = jwt.encode({'user_id': user.id}, self.secret) |
| | return {'token': token, 'success': True} |
| | return {'success': False} |
| | |
| | async def verify_token(self, token): |
| | try: |
| | payload = jwt.decode(token, self.secret, algorithms=['HS256']) |
| | return payload |
| | except: |
| | return None |
| | ``` |
| | |
| | ### Assistant: |
| | """ |
| |
|
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | outputs = model.generate( |
| | **inputs, |
| | max_new_tokens=4096, |
| | temperature=0.2, # Lower temperature for precise enterprise analysis |
| | top_p=0.95, |
| | do_sample=True |
| | ) |
| | |
| | response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| | print(response) |
| | ``` |
| | |
| | --- |
| | |
| | ### Enterprise Deployment (4-bit Quantization) |
| | |
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig |
| | from peft import PeftModel |
| | |
| | # 4-bit quantization - runs on 40GB GPU |
| | bnb_config = BitsAndBytesConfig( |
| | load_in_4bit=True, |
| | bnb_4bit_use_double_quant=True, |
| | bnb_4bit_quant_type="nf4", |
| | bnb_4bit_compute_dtype="bfloat16" |
| | ) |
| | |
| | model = AutoModelForCausalLM.from_pretrained( |
| | "ibm-granite/granite-20b-code-instruct-8k", |
| | quantization_config=bnb_config, |
| | device_map="auto", |
| | trust_remote_code=True |
| | ) |
| | |
| | model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
| | tokenizer = AutoTokenizer.from_pretrained("ibm-granite/granite-20b-code-instruct-8k", trust_remote_code=True) |
| | |
| | # Enterprise-ready: Runs on A100 40GB, A100 80GB, or 2x RTX 4090 |
| | ``` |
| | |
| | --- |
| | |
| | ### Multi-GPU Deployment (Maximum Performance) |
| | |
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | from peft import PeftModel |
| | import torch |
| | |
| | # Load across multiple GPUs for maximum throughput |
| | model = AutoModelForCausalLM.from_pretrained( |
| | "ibm-granite/granite-20b-code-instruct-8k", |
| | device_map="balanced", # Distribute across available GPUs |
| | torch_dtype=torch.bfloat16, |
| | trust_remote_code=True |
| | ) |
| | |
| | model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
| | tokenizer = AutoTokenizer.from_pretrained("ibm-granite/granite-20b-code-instruct-8k", trust_remote_code=True) |
| | |
| | # Optimal for: 2x A100, 4x RTX 4090, or enterprise GPU clusters |
| | # Throughput: 2-3x faster than single GPU |
| | ``` |
| | |
| | --- |
| | |
| | ## π Performance & Benchmarks |
| | |
| | ### Hardware Requirements |
| | |
| | | Deployment | RAM | GPU VRAM | Tokens/Second | Latency (4K response) | Cost/Month | |
| | |-----------|-----|----------|---------------|----------------------|------------| |
| | | **4-bit Quantized** | 40GB | 32GB | ~35 tok/s | ~115 seconds | $0 (on-prem) or $800-1200 (cloud) | |
| | | **8-bit Quantized** | 64GB | 48GB | ~45 tok/s | ~90 seconds | $0 (on-prem) or $1200-1800 (cloud) | |
| | | **Full Precision (bf16)** | 96GB | 80GB | ~60 tok/s | ~67 seconds | $0 (on-prem) or $2000-3000 (cloud) | |
| | | **Multi-GPU (2x A100)** | 128GB | 160GB | ~120 tok/s | ~33 seconds | Enterprise only | |
| | |
| | ### Real-World Performance |
| | |
| | **Tested on A100 40GB** (enterprise GPU): |
| | - **Tokens/second:** ~35 tok/s (4-bit), ~55 tok/s (full precision) |
| | - **Cold start:** ~8 seconds |
| | - **Memory usage:** 28GB (4-bit), 42GB (full precision) |
| | - **Throughput:** 200-300 comprehensive analyses per day |
| | |
| | **Tested on 2x A100 80GB** (multi-GPU): |
| | - **Tokens/second:** ~110-120 tok/s |
| | - **Cold start:** ~6 seconds |
| | - **Throughput:** 500+ analyses per day |
| | |
| | ### Security Analysis Quality |
| | |
| | **The differentiator:** Granite 20B provides the deepest, most nuanced security analysis: |
| | - Identifies **15-25% more vulnerabilities** than 7B models in complex code |
| | - Detects **multi-step attack chains** that smaller models miss |
| | - Provides **enterprise-grade operational guidance** with compliance mapping |
| | - **Reduces false positives** through sophisticated reasoning |
| | |
| | --- |
| | |
| | ## π° Cost Analysis |
| | |
| | ### Total Cost of Ownership (TCO) - 1 Year |
| | |
| | **Option 1: On-Premise (Dedicated Server)** |
| | - Hardware: 2x A100 40GB - $20,000 (one-time capital expense) |
| | - Server infrastructure: $5,000 |
| | - Electricity: ~$2,400/year |
| | - **Total Year 1:** $27,400 |
| | - **Total Year 2+:** $2,400/year |
| | |
| | **Option 2: Cloud GPU (AWS/GCP/Azure)** |
| | - Instance: A100 40GB (p4d.xlarge) |
| | - Cost: ~$3.50/hour |
| | - Usage: 160 hours/month (enterprise team) |
| | - **Total Year 1:** $6,720/year |
| | |
| | **Option 3: Enterprise GPT-4 (for comparison)** |
| | - Cost: $30/1M input tokens, $60/1M output tokens |
| | - Usage: 500M input + 500M output tokens/year |
| | - **Total Year 1:** $45,000/year |
| | |
| | **Option 4: Professional Security Audits (for comparison)** |
| | - Average enterprise security audit: $150,000-500,000 |
| | - Frequency: Quarterly (4x/year) |
| | - **Total Year 1:** $600,000-2,000,000 |
| | |
| | **ROI Winner:** On-premise deployment pays for itself with **1-2 prevented security audits** or **preventing a single breach** (average cost: $4.45M). |
| | |
| | --- |
| | |
| | ## π― Use Cases & Examples |
| | |
| | ### 1. Enterprise Security Architecture Review |
| | |
| | Analyze complex microservices platforms: |
| | |
| | ```python |
| | prompt = """### User: |
| | Conduct a comprehensive security architecture review of this fintech payment platform. Analyze: |
| | 1. Service-to-service authentication security |
| | 2. Data flow security boundaries |
| | 3. Compliance with PCI-DSS requirements |
| | 4. Attack surface analysis |
| | 5. Defense-in-depth gaps |
| | |
| | [Include microservices code across auth-service, payment-service, notification-service] |
| | |
| | ### Assistant: |
| | """ |
| | ``` |
| | |
| | **Model Response:** Provides 20-30 page comprehensive analysis with specific vulnerability findings, attack chain scenarios, compliance gaps, and remediation priorities. |
| | |
| | --- |
| | |
| | ### 2. Regulatory Compliance Validation |
| | |
| | Validate code against regulatory requirements: |
| | |
| | ```python |
| | prompt = """### User: |
| | Analyze this healthcare EHR system for HIPAA compliance. Verify: |
| | 1. Patient data encryption (at rest and in transit) |
| | 2. Access control and audit logging |
| | 3. Data retention policies |
| | 4. Breach notification capabilities |
| | 5. Business Associate Agreement requirements |
| | |
| | [Include EHR codebase] |
| | |
| | ### Assistant: |
| | """ |
| | ``` |
| | |
| | **Model Response:** Detailed compliance mapping, gap analysis, and remediation roadmap. |
| | |
| | --- |
| | |
| | ### 3. Supply Chain Security Analysis |
| | |
| | Analyze third-party dependencies and integrations: |
| | |
| | ```python |
| | prompt = """### User: |
| | Perform a supply chain security analysis of this application: |
| | 1. Third-party library vulnerabilities |
| | 2. Dependency confusion risks |
| | 3. Code injection via dependencies |
| | 4. Malicious package detection |
| | 5. License compliance issues |
| | |
| | [Include package.json, requirements.txt, go.mod] |
| | |
| | ### Assistant: |
| | """ |
| | ``` |
| | |
| | **Model Response:** Comprehensive supply chain risk assessment with mitigation strategies. |
| | |
| | --- |
| | |
| | ### 4. Advanced Penetration Testing Guidance |
| | |
| | Develop sophisticated attack scenarios: |
| | |
| | ```python |
| | prompt = """### User: |
| | Design a comprehensive penetration testing strategy for this enterprise web application. Include: |
| | 1. Attack surface enumeration |
| | 2. Vulnerability prioritization |
| | 3. Multi-stage attack chains |
| | 4. Privilege escalation paths |
| | 5. Data exfiltration scenarios |
| | 6. Post-exploitation persistence |
| | |
| | ### Assistant: |
| | """ |
| | ``` |
| | |
| | **Model Response:** Professional pentesting methodology with specific attack vectors and validation procedures. |
| | |
| | --- |
| | |
| | ## β οΈ Limitations & Transparency |
| | |
| | ### What This Model Does Well |
| | β
Maximum code understanding and security reasoning |
| | β
Complex attack chain analysis and enterprise architecture review |
| | β
Detailed operational guidance and compliance mapping |
| | β
Sophisticated multi-layered vulnerability detection |
| | β
Enterprise-scale codebase analysis |
| | β
IBM enterprise trust and reliability |
| | |
| | ### What This Model Doesn't Do |
| | β **Not a security scanner** - Use tools like Semgrep, CodeQL, Snyk, or Veracode |
| | β **Not a penetration testing tool** - Cannot perform active exploitation or network scanning |
| | β **Not legal/compliance advice** - Consult security and legal professionals |
| | β **Not a replacement for security experts** - Critical systems need professional security review and audits |
| | β **Not real-time threat intelligence** - Training data frozen at Dec 2024 |
| | |
| | ### Known Issues & Constraints |
| | - **Inference latency:** Larger model means slower responses (35-60 tok/s vs 100+ tok/s for smaller models) |
| | - **Hardware requirements:** Requires enterprise GPU infrastructure (40GB+ VRAM) |
| | - **Detailed analysis:** May generate very comprehensive responses (3000-4000 tokens) |
| | - **Cost consideration:** Higher deployment cost than smaller models |
| | - **Context window:** 8K tokens (vs 128K for Qwen models) |
| | |
| | ### Appropriate Use |
| | β
Enterprise security audits and professional assessments |
| | β
Regulatory compliance validation |
| | β
Critical infrastructure security review |
| | β
Financial services and healthcare applications |
| | β
Government and defense security analysis |
| | |
| | ### Inappropriate Use |
| | β Sole validation for production deployments (use comprehensive testing) |
| | β Replacement for professional security audits |
| | β Active exploitation or penetration testing without authorization |
| | β Consumer applications (too large, use smaller models) |
| | |
| | --- |
| | |
| | ## π¬ Dataset Information |
| | |
| | This model was trained on **[SecureCode v2.0](https://huggingface.co/datasets/scthornton/securecode-v2)**, a production-grade security dataset with: |
| | |
| | - **1,209 total examples** (841 train / 175 validation / 193 test) |
| | - **100% incident grounding** - every example tied to real CVEs or security breaches |
| | - **11 vulnerability categories** - complete OWASP Top 10:2025 coverage |
| | - **11 programming languages** - from Python to Rust |
| | - **4-turn conversational structure** - mirrors real developer-AI workflows |
| | - **100% expert validation** - reviewed by independent security professionals |
| | |
| | See the [full dataset card](https://huggingface.co/datasets/scthornton/securecode-v2) and [research paper](https://perfecxion.ai/articles/securecode-v2-dataset-paper.html) for complete details. |
| | |
| | --- |
| | |
| | ## π’ About perfecXion.ai |
| | |
| | [perfecXion.ai](https://perfecxion.ai) is dedicated to advancing AI security through research, datasets, and production-grade security tooling. |
| | |
| | **Connect:** |
| | - Website: [perfecxion.ai](https://perfecxion.ai) |
| | - Research: [perfecxion.ai/research](https://perfecxion.ai/research) |
| | - Knowledge Hub: [perfecxion.ai/knowledge](https://perfecxion.ai/knowledge) |
| | - GitHub: [@scthornton](https://github.com/scthornton) |
| | - HuggingFace: [@scthornton](https://huggingface.co/scthornton) |
| | - Email: scott@perfecxion.ai |
| | |
| | --- |
| | |
| | ## π License |
| | |
| | **Model License:** Apache 2.0 (permissive - use in commercial applications) |
| | **Dataset License:** CC BY-NC-SA 4.0 (non-commercial with attribution) |
| | |
| | ### What You CAN Do |
| | β
Use this model commercially in production applications |
| | β
Fine-tune further for your specific use case |
| | β
Deploy in enterprise environments |
| | β
Integrate into commercial products |
| | β
Distribute and modify the model weights |
| | β
Charge for services built on this model |
| | β
Use in government and regulated industries |
| | |
| | ### What You CANNOT Do with the Dataset |
| | β Sell or redistribute the raw SecureCode v2.0 dataset commercially |
| | β Use the dataset to train commercial models without releasing under the same license |
| | β Remove attribution or claim ownership of the dataset |
| | |
| | For commercial dataset licensing or custom training, contact: scott@perfecxion.ai |
| | |
| | --- |
| | |
| | ## π Citation |
| | |
| | If you use this model in your research or applications, please cite: |
| | |
| | ```bibtex |
| | @misc{thornton2025securecode-granite20b, |
| | title={IBM Granite 20B Code - SecureCode Edition}, |
| | author={Thornton, Scott}, |
| | year={2025}, |
| | publisher={perfecXion.ai}, |
| | url={https://huggingface.co/scthornton/granite-20b-code-securecode}, |
| | note={Fine-tuned on SecureCode v2.0: https://huggingface.co/datasets/scthornton/securecode-v2} |
| | } |
| | |
| | @misc{thornton2025securecode-dataset, |
| | title={SecureCode v2.0: A Production-Grade Dataset for Training Security-Aware Code Generation Models}, |
| | author={Thornton, Scott}, |
| | year={2025}, |
| | month={January}, |
| | publisher={perfecXion.ai}, |
| | url={https://perfecxion.ai/articles/securecode-v2-dataset-paper.html}, |
| | note={Dataset: https://huggingface.co/datasets/scthornton/securecode-v2} |
| | } |
| | ``` |
| | |
| | --- |
| | |
| | ## π Acknowledgments |
| | |
| | - **IBM Research** for the exceptional Granite code models and enterprise commitment |
| | - **OWASP Foundation** for maintaining the Top 10 vulnerability taxonomy |
| | - **MITRE Corporation** for the CVE database and vulnerability research |
| | - **Security research community** for responsible disclosure practices |
| | - **Hugging Face** for model hosting and inference infrastructure |
| | - **Enterprise security teams** who validated this model in production environments |
| | |
| | --- |
| | |
| | ## π€ Contributing |
| | |
| | Found a security issue or have suggestions for improvement? |
| | |
| | - π **Report issues:** [GitHub Issues](https://github.com/scthornton/securecode-models/issues) |
| | - π¬ **Discuss improvements:** [HuggingFace Discussions](https://huggingface.co/scthornton/granite-20b-code-securecode/discussions) |
| | - π§ **Contact:** scott@perfecxion.ai |
| | |
| | ### Community Contributions Welcome |
| | |
| | Especially interested in: |
| | - **Enterprise deployment case studies** |
| | - **Benchmark evaluations** on industry security datasets |
| | - **Compliance validation** (PCI-DSS, HIPAA, SOC 2) |
| | - **Performance optimization** for specific enterprise hardware |
| | - **Integration examples** with enterprise security platforms |
| | |
| | --- |
| | |
| | ## π SecureCode Model Collection |
| | |
| | Explore other SecureCode fine-tuned models optimized for different use cases: |
| | |
| | ### Entry-Level Models (3-7B) |
| | - **[llama-3.2-3b-securecode](https://huggingface.co/scthornton/llama-3.2-3b-securecode)** |
| | - **Best for:** Consumer hardware, IDE integration, education |
| | - **Hardware:** 8GB RAM minimum |
| | - **Unique strength:** Most accessible |
| | |
| | - **[deepseek-coder-6.7b-securecode](https://huggingface.co/scthornton/deepseek-coder-6.7b-securecode)** |
| | - **Best for:** Security-optimized baseline |
| | - **Hardware:** 16GB RAM |
| | - **Unique strength:** Security-first architecture |
| | |
| | - **[qwen2.5-coder-7b-securecode](https://huggingface.co/scthornton/qwen2.5-coder-7b-securecode)** |
| | - **Best for:** Best code understanding in 7B class |
| | - **Hardware:** 16GB RAM |
| | - **Unique strength:** 128K context, best-in-class |
| | |
| | - **[codegemma-7b-securecode](https://huggingface.co/scthornton/codegemma-7b-securecode)** |
| | - **Best for:** Google ecosystem, instruction following |
| | - **Hardware:** 16GB RAM |
| | - **Unique strength:** Google brand, strong completion |
| | |
| | ### Mid-Range Models (13-15B) |
| | - **[codellama-13b-securecode](https://huggingface.co/scthornton/codellama-13b-securecode)** |
| | - **Best for:** Enterprise trust, Meta brand |
| | - **Hardware:** 24GB RAM |
| | - **Unique strength:** Proven track record |
| | |
| | - **[qwen2.5-coder-14b-securecode](https://huggingface.co/scthornton/qwen2.5-coder-14b-securecode)** |
| | - **Best for:** Advanced code analysis |
| | - **Hardware:** 32GB RAM |
| | - **Unique strength:** 128K context window |
| | |
| | - **[starcoder2-15b-securecode](https://huggingface.co/scthornton/starcoder2-15b-securecode)** |
| | - **Best for:** Multi-language projects (600+ languages) |
| | - **Hardware:** 32GB RAM |
| | - **Unique strength:** Broadest language support |
| | |
| | ### Enterprise-Scale Models (20B+) |
| | - **[granite-20b-code-securecode](https://huggingface.co/scthornton/granite-20b-code-securecode)** β (YOU ARE HERE) |
| | - **Best for:** Enterprise-scale, IBM trust, maximum capability |
| | - **Hardware:** 48GB RAM |
| | - **Unique strength:** Largest model, deepest analysis |
| | |
| | **View Complete Collection:** [SecureCode Models](https://huggingface.co/collections/scthornton/securecode) |
| | |
| | --- |
| | |
| | <div align="center"> |
| | |
| | **Built with β€οΈ for secure enterprise software** |
| | |
| | [perfecXion.ai](https://perfecxion.ai) | [Research](https://perfecxion.ai/research) | [Knowledge Hub](https://perfecxion.ai/knowledge) | [Contact](mailto:scott@perfecxion.ai) |
| | |
| | --- |
| | |
| | *Maximum security intelligence. Enterprise trust. IBM heritage.* |
| | |
| | </div> |
| | |