| --- |
| license: apache-2.0 |
| language: |
| - en |
| - code |
| tags: |
| - stack-2.9 |
| - open-source |
| - coding-assistant |
| - fine-tuned |
| - qwen |
| - code-generation |
| library_name: transformers |
| --- |
| |
| # Stack 2.9 Fine-Tuned Model |
|
|
| A fine-tuned coding assistant model based on {{base_model}}. |
| |
| ## Model Details |
| |
| | Property | Value | |
| |----------|-------| |
| | Base Model | {{base_model}} | |
| | Training Data | {{training_examples}} examples | |
| | LoRA Rank | {{lora_rank}} | |
| | LoRA Alpha | {{lora_alpha}} | |
| | Max Context Length | {{max_context_length}} | |
| | License | Apache 2.0 | |
| |
| ## Description |
| |
| Stack 2.9 is a fine-tuned coding assistant model designed for code generation, refactoring, and software development tasks. The model has been fine-tuned on a curated dataset of high-quality code examples and programming tasks. |
| |
| ### Training Details |
| |
| - **Dataset**: {{training_examples}} examples from diverse programming domains |
| - **Fine-tuning Method**: LoRA (Low-Rank Adaptation) |
| - **LoRA Configuration**: rank={{lora_rank}}, alpha={{lora_alpha}} |
| - **Base Model**: {{base_model}} |
| |
| ## Benchmarks |
| |
| | Benchmark | Score | |
| |-----------|-------| |
| | HumanEval | {{humaneval_score}} | |
| | MBPP | {{mbpp_score}} | |
| |
| ## Usage |
| |
| ### Using Transformers |
| |
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model_name = "your-username/stack-2.9-7b" # Replace with your repo |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
| prompt = """Write a Python function to calculate the factorial of a number. |
|
|
| ```python |
| """ |
| |
| inputs = tokenizer(prompt, return_tensors="pt") |
| outputs = model.generate(**inputs, max_new_tokens=200) |
| print(tokenizer.decode(outputs[0])) |
| ``` |
|
|
| ### Using vLLM for Fast Inference |
|
|
| ```python |
| from vllm import LLM, SamplingParams |
| |
| llm = LLM(model="your-username/stack-2.9-7b") |
| sampling_params = SamplingParams(temperature=0.7, max_tokens=200) |
| |
| prompt = "Write a Python function to reverse a string:" |
| outputs = llm.generate(prompt, sampling_params) |
| print(outputs[0].outputs[0].text) |
| ``` |
|
|
| ## Limitations |
|
|
| - The model may generate incorrect code; always verify outputs |
| - Performance may vary across different programming languages |
| - Context window limited to {{max_context_length}} tokens |
|
|
| ## License |
|
|
| This model is licensed under the [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) license. |
|
|
| ## Citation |
|
|
| If you use this model in your research, please cite: |
|
|
| ```bibtex |
| @misc{stack-2.9, |
| author = {Stack Team}, |
| title = {Stack 2.9: Fine-tuned Coding Assistant}, |
| year = {2025}, |
| publisher = {HuggingFace}, |
| url = {https://huggingface.co/your-username/stack-2.9-7b} |
| } |
| ``` |
|
|
| --- |
|
|
| *Model uploaded via upload_hf.py script* |