| | --- |
| | datasets: |
| | - Programming-Language/codeagent-python |
| | language: |
| | - en |
| | base_model: |
| | - google/flan-t5-base |
| | pipeline_tag: text-generation |
| | library_name: transformers |
| | license: apache-2.0 |
| | --- |
| | |
| | # flan-python-expert ๐ |
| |
|
| | This model is a fine-tuned version of [`google/flan-t5-base`](https://huggingface.co/google/flan-t5-base) on the [`codeagent-python`](https://huggingface.co/datasets/Programming-Language/codeagent-python) dataset. |
| |
|
| | It is designed to generate Python code from natural language instructions. |
| |
|
| | --- |
| |
|
| | ## ๐ง Model Details |
| |
|
| | - **Base Model:** FLAN-T5 Base |
| | - **Fine-tuned on:** Python code dataset (`codeagent-python`) |
| | - **Task:** Text-to-code generation |
| | - **Language:** English |
| | - **Framework:** ๐ค Transformers |
| | - **Library:** `adapter-transformers` |
| |
|
| | --- |
| |
|
| | ## ๐๏ธ Training |
| |
|
| | The model was trained using the following setup: |
| |
|
| | ```python |
| | from transformers import TrainingArguments |
| | |
| | training_args = TrainingArguments( |
| | output_dir="flan-python-expert", |
| | evaluation_strategy="epoch", |
| | learning_rate=2e-6, |
| | per_device_train_batch_size=1, |
| | per_device_eval_batch_size=1, |
| | num_train_epochs=1, |
| | weight_decay=0.01, |
| | save_total_limit=2, |
| | logging_steps=1, |
| | push_to_hub=False, |
| | ) |
| | |
| | ``` |
| |
|
| |
|
| | Trained for 1 epoch |
| |
|
| | Optimized for low-resource fine-tuning |
| |
|
| | Training performed using Hugging Face Trainer |
| |
|
| | ## Example Usage |
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForSeq2SeqLM |
| | |
| | model = AutoModelForSeq2SeqLM.from_pretrained("MalikIbrar/flan-python-expert") |
| | tokenizer = AutoTokenizer.from_pretrained("MalikIbrar/flan-python-expert") |
| | |
| | input_text = "Write a Python function to check if a number is prime." |
| | inputs = tokenizer(input_text, return_tensors="pt") |
| | |
| | outputs = model.generate(**inputs, max_length=256) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | --- |