| | --- |
| | base_model: Qwen/Qwen2.5-Coder-14B |
| | library_name: peft |
| | pipeline_tag: text-generation |
| | license: apache-2.0 |
| | tags: |
| | - code |
| | - coding-assistant |
| | - lora |
| | - refactor |
| | - bug-fix |
| | - optimization |
| | - async |
| | - concurrency |
| | - security |
| | - logging |
| | - networking |
| | - transformers |
| | --- |
| | |
| | # EdgePulse Coder 14B (LoRA) |
| |
|
| | **EdgePulse Coder 14B** is a production-grade coding assistant fine-tuned using LoRA on top of **Qwen2.5-Coder-14B**. |
| | It is designed to handle real-world software engineering workflows with high reliability and correctness. |
| |
|
| | --- |
| |
|
| | ## Model Details |
| |
|
| | ### Model Description |
| |
|
| | EdgePulse Coder 14B focuses on **practical developer tasks**, trained on a large, strictly validated dataset covering: |
| |
|
| | - Bug fixing |
| | - Code explanation |
| | - Refactoring |
| | - Optimization |
| | - Async & concurrency correction |
| | - Logging & observability |
| | - Security & defensive coding |
| | - Networking & I/O handling |
| | - Multi-file context reasoning |
| | - Test generation and impact analysis |
| |
|
| | The model is optimized for **IDE usage**, **CLI workflows**, and **Cursor-like streaming environments**. |
| |
|
| | --- |
| |
|
| | - **Developed by:** EdgePulseAI |
| | - **Shared by:** EdgePulseAI |
| | - **Model type:** Large Language Model (Code-focused) |
| | - **Language(s):** Python, JavaScript, TypeScript, Bash (primary), general programming concepts |
| | - **License:** Apache-2.0 |
| | - **Finetuned from:** Qwen/Qwen2.5-Coder-14B |
| |
|
| | --- |
| |
|
| | ## Model Sources |
| |
|
| | - **Base Model:** https://huggingface.co/Qwen/Qwen2.5-Coder-14B |
| | - **Website:** https://EdgePulseAi.com |
| |
|
| | --- |
| |
|
| | ## Uses |
| |
|
| | ### Direct Use |
| |
|
| | EdgePulse Coder 14B can be used directly for: |
| |
|
| | - Code explanation |
| | - Bug fixing |
| | - Refactoring existing code |
| | - Generating tests |
| | - Improving logging and error handling |
| | - Fixing async / concurrency bugs |
| | - Secure coding suggestions |
| | - Network & I/O robustness |
| |
|
| | ### Downstream Use |
| |
|
| | - IDE assistants (VS Code / Cursor-style tools) |
| | - CI/CD automation |
| | - Code review bots |
| | - Developer copilots |
| | - Internal engineering tools |
| |
|
| | ### Out-of-Scope Use |
| |
|
| | - Medical or legal advice |
| | - Autonomous system control |
| | - High-risk decision making without human review |
| |
|
| | --- |
| |
|
| | ## Bias, Risks, and Limitations |
| |
|
| | - The model may occasionally produce syntactically correct but logically incorrect code. |
| | - Security-sensitive code should always be reviewed by humans. |
| | - Performance depends on correct prompt framing and context size. |
| |
|
| | ### Recommendations |
| |
|
| | - Use human review for production deployments. |
| | - Combine with static analysis and testing tools. |
| | - Prefer structured prompts for multi-file tasks. |
| |
|
| | --- |
| |
|
| | ## How to Get Started with the Model |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | from peft import PeftModel |
| | |
| | base_model = "Qwen/Qwen2.5-Coder-14B" |
| | adapter_model = "edgepulse-ai/EdgePulse-Coder-14B-LoRA" |
| | |
| | tokenizer = AutoTokenizer.from_pretrained(base_model) |
| | |
| | model = AutoModelForCausalLM.from_pretrained( |
| | base_model, |
| | device_map="auto", |
| | torch_dtype="auto" |
| | ) |
| | |
| | model = PeftModel.from_pretrained(model, adapter_model) |
| | model.eval() |
| | |
| | prompt = "Fix this bug:\n\ndef add(a,b): return a-b" |
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | |
| | output = model.generate(**inputs, max_new_tokens=128) |
| | print(tokenizer.decode(output[0], skip_special_tokens=True)) |