metadata
license: apache-2.0
language:
- en
- code
tags:
- stack-2.9
- open-source
- coding-assistant
- fine-tuned
- qwen
- code-generation
library_name: transformers
Stack 2.9 Fine-Tuned Model
A fine-tuned coding assistant model based on {{base_model}}.
Model Details
| Property | Value |
|---|---|
| Base Model | {{base_model}} |
| Training Data | {{training_examples}} examples |
| LoRA Rank | {{lora_rank}} |
| LoRA Alpha | {{lora_alpha}} |
| Max Context Length | {{max_context_length}} |
| License | Apache 2.0 |
Description
Stack 2.9 is a fine-tuned coding assistant model designed for code generation, refactoring, and software development tasks. The model has been fine-tuned on a curated dataset of high-quality code examples and programming tasks.
Training Details
- Dataset: {{training_examples}} examples from diverse programming domains
- Fine-tuning Method: LoRA (Low-Rank Adaptation)
- LoRA Configuration: rank={{lora_rank}}, alpha={{lora_alpha}}
- Base Model: {{base_model}}
Benchmarks
| Benchmark | Score |
|---|---|
| HumanEval | {{humaneval_score}} |
| MBPP | {{mbpp_score}} |
Usage
Using Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "your-username/stack-2.9-7b" # Replace with your repo
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
prompt = """Write a Python function to calculate the factorial of a number.
```python
"""
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0]))
Using vLLM for Fast Inference
from vllm import LLM, SamplingParams
llm = LLM(model="your-username/stack-2.9-7b")
sampling_params = SamplingParams(temperature=0.7, max_tokens=200)
prompt = "Write a Python function to reverse a string:"
outputs = llm.generate(prompt, sampling_params)
print(outputs[0].outputs[0].text)
Limitations
- The model may generate incorrect code; always verify outputs
- Performance may vary across different programming languages
- Context window limited to {{max_context_length}} tokens
License
This model is licensed under the Apache 2.0 license.
Citation
If you use this model in your research, please cite:
@misc{stack-2.9,
author = {Stack Team},
title = {Stack 2.9: Fine-tuned Coding Assistant},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/your-username/stack-2.9-7b}
}
Model uploaded via upload_hf.py script