--- language: - tr license: mit task_categories: - text-generation - question-answering tags: - turkish - identity - instruction-tuning - llm-alignment - nlp - chatbot pretty_name: TurkishIdentityMini size_categories: - n<1K dataset_info: features: - name: instruction dtype: string - name: output dtype: string splits: - name: train num_bytes: 41540 num_examples: 481 download_size: 13573 dataset_size: 41540 configs: - config_name: default data_files: - split: train path: data/train-* --- # TurkishIdentityMini ## Dataset Description **TurkishIdentityMini** is a small, template-based Turkish instruction dataset designed to help LLMs respond correctly to identity-related questions. It contains instruction–output pairs where a user asks a chatbot about its name, origin, or creator, and the model responds using customizable `{{model_name}}` and `{{team_name}}` placeholders. This dataset is useful for fine-tuning or instruction-tuning Turkish language models to maintain a consistent, branded identity and to correctly deny affiliation with other AI providers such as OpenAI, Google, Meta, Anthropic, or Microsoft. ### Dataset Summary | Property | Value | |---|---| | Language | Turkish (`tr`) | | Split | `train` only | | Format | Parquet | | License | MIT | --- ## Dataset Structure ### Data Fields | Field | Type | Description | |---|---|---| | `instruction` | `string` | A Turkish user query about the model's identity (e.g., *"Sen kimsin?"*, *"Seni kim yaptı?"*) | | `output` | `string` | A template response using `{{model_name}}` and `{{team_name}}` placeholders | ### Example Rows | instruction | output | |---|---| | `Seni kim yaptı?` | `Ben {{team_name}} ekibi tarafından yapıldım.` | | `Sen kimsin?` | `Ben {{model_name}}, {{team_name}} tarafından geliştirilmiş bir yapay zeka asistanıyım.` | | `ChatGPT misin?` | `Hayır, ben {{model_name}}. {{team_name}} tarafından eğitilmiş bir asistanım.` | | `OpenAI'dan mısın?` | `Hayır, ben {{team_name}} tarafından bağımsız olarak geliştirildim.` | | `Adın ne?` | `Ben {{model_name}}, yapay zeka asistanınızım.` | --- ## Intended Uses ### Primary Use Fine-tuning or instruction-tuning Turkish LLMs to: - Respond to identity queries (name, creator, affiliation) in Turkish - Correctly deny association with third-party AI providers - Use a consistent, customizable branded persona via the `{{model_name}}` and `{{team_name}}` placeholders ### Out-of-Scope Use - This dataset covers **only** identity-related queries; it is not suitable as a standalone fine-tuning corpus for general conversational ability - The placeholder format requires preprocessing before use in most training pipelines --- ## Dataset Creation ### Covered Question Categories The dataset covers the following identity query themes: - **Creator / origin** — *"Seni kim yaptı?"*, *"Nereden geliyorsun?"* - **Name / model identity** — *"Adın ne?"*, *"Model adını söyler misin?"* - **Brand denial** — *"ChatGPT misin?"*, *"Sen Claude musun?"*, *"Google tarafından mı oluşturuldun?"* - **Greetings with identity** — *"Merhaba"*, *"Selam"* → model introduces itself - **Paraphrastic variants** — Diverse rephrasings of the same intents to improve robustness ### Template Placeholders All outputs use two placeholders that must be filled before training: | Placeholder | Description | |---|---| | `{{model_name}}` | The name of the deployed model | | `{{team_name}}` | The name of the developing team or organization | **Example preprocessing (Python):** ```python def fill_template(example, model_name, team_name): example["output"] = ( example["output"] .replace("{{model_name}}", model_name) .replace("{{team_name}}", team_name) ) return example dataset = dataset.map(lambda x: fill_template(x, "Magibu-11b-v0.8", "magibu")) ``` --- ## Usage ### With 🤗 Datasets ```python from datasets import load_dataset dataset = load_dataset("aliarda/TurkishIdentityMini") print(dataset["train"][0]) # {'instruction': 'Seni kim yaptı?', 'output': 'Ben {{team_name}} ekibi tarafından yapıldım.'} ``` ### With pandas ```python import pandas as pd df = pd.read_parquet("hf://datasets/aliarda/TurkishIdentityMini/data/train-*.parquet") print(df.head()) ``` --- ## Acknowledgements 80 rows in this dataset were sourced from [`sts07142/llm-name-identity`](https://huggingface.co/datasets/sts07142/llm-name-identity) and translated into Turkish using AI-assisted translation. --- ## Citation If you use this dataset in your research, please cite it as: ```bibtex @dataset{aliarda_turkishidentitymini, author = {Ali Arda Fincan}, title = {TurkishIdentityMini}, year = {2026}, publisher = {Hugging Face}, url = {https://huggingface.co/datasets/aliarda/TurkishIdentityMini} } ```