File size: 1,819 Bytes
4546fa9 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 | ---
language:
- en
license: other
base_model:
- mistralai/Devstral-Small-2505
tags:
- text-to-sql
- sql
- mistral
- transformers
- safetensors
pipeline_tag: text-generation
library_name: transformers
---
# Devstral SQLCoder SFT
This model is a full-parameter SFT checkpoint for SQL generation, trained from `mistralai/Devstral-Small-2505` and exported to Hugging Face safetensors format.
## Model Details
- Base model: `mistralai/Devstral-Small-2505`
- Architecture: `MistralForCausalLM`
- Precision used in training: bf16
- Max sequence length (training config): 4096
- Export format: sharded `safetensors` with `model.safetensors.index.json`
## Training Data (Merged)
The SFT run merged the following datasets:
- spider
- bird
- bird23-train-filtered
- synsql-2.5m
- wikisql
- gretelai-synthetic
- sql-create-context
## Intended Use
- Text-to-SQL research and experimentation
- SQL generation benchmarks and evaluation pipelines
## Limitations
- This model may generate incorrect SQL and should be validated before production use.
- Performance depends on prompt format, schema context quality, and decoding settings.
- Evaluate safety and compliance requirements before deployment.
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
repo_or_path = "<hf-username-or-org>/<model-repo>"
tokenizer = AutoTokenizer.from_pretrained(repo_or_path, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
repo_or_path,
torch_dtype="bfloat16",
)
```
## Local Files Included
- `config.json`
- `generation_config.json`
- `tekken.json`
- `model-00001-of-00021.safetensors` ... `model-00021-of-00021.safetensors`
- `model.safetensors.index.json`
## Citation
If you use this model, please cite this repository:
- https://github.com/ai-twinkle/twinkle-sqlcoder
|