| --- |
| base_model: sentence-transformers/all-MiniLM-L6-v2 |
| library_name: peft |
| license: mit |
| tags: |
| - lora |
| - peft |
| - code |
| - programming |
| - software |
| - domain-adaptation |
| - sentence-embeddings |
| language: |
| - en |
| --- |
| |
| # Code LoRA Adapter for DomainEmbedder-v2.6 |
|
|
| Domain-specific LoRA adapter for code/programming text embeddings. |
|
|
| ## Model Details |
|
|
| | Property | Value | |
| |----------|-------| |
| | **Base Model** | sentence-transformers/all-MiniLM-L6-v2 | |
| | **Parent System** | DomainEmbedder-v2.6 | |
| | **Domain** | Code / Programming | |
| | **LoRA Rank** | 16 | |
| | **LoRA Alpha** | 32 | |
| | **Target Modules** | query, value | |
| | **Trainable Params** | 147,456 (0.645%) | |
|
|
| ## Training Data |
|
|
| Trained on 40,000 code-related text pairs from: |
| - Code Alpaca |
| - MBPP (Mostly Basic Python Problems) |
| - Code Contests |
| - Python Instructions |
|
|
| ## Training Configuration |
|
|
| | Parameter | Value | |
| |-----------|-------| |
| | Epochs | 3 | |
| | Batch Size | 32 | |
| | Learning Rate | 2e-4 | |
| | Loss | Contrastive (InfoNCE) | |
| | Best Val Loss | 0.0039 | |
|
|
| ## Usage |
|
|
| This adapter is part of the DomainEmbedder-v2.6 system. It is selected automatically by the RL policy when code-related content is detected. |
|
|
| ```python |
| from peft import PeftModel |
| from transformers import AutoModel |
| |
| # Load base encoder |
| base_encoder = AutoModel.from_pretrained('sentence-transformers/all-MiniLM-L6-v2') |
| |
| # Apply code LoRA |
| code_model = PeftModel.from_pretrained(base_encoder, 'path/to/code_lora') |
| ``` |
|
|
| ## Author |
|
|
| **Zain Asad** |
|
|
| ## License |
|
|
| MIT License |
|
|
| ## Framework Versions |
|
|
| - PEFT 0.18.1 |
| - Transformers 4.x |
| - PyTorch 2.x |
|
|