--- pretty_name: AI Helps Finding Best Merging LLMs language: - en license: other task_categories: - text-generation - text-ranking - summarization - question-answering - text-retrieval tags: - llm - llm-comparison - model-ranking - model-merging - mixture-of-experts - moe - prompt-comparison - ai-research - open-source-llm - reasoning - coding - long-context size_categories: - n<1K --- # Dataset Card for AI Helps Finding Best Merging LLMs ## Dataset Summary **AI Helps Finding Best Merging LLMs** is a **prompt-response comparison dataset** created by manually submitting the same user-written evaluation template to multiple LLM applications and collecting their responses. The creator and founder of WithIn Us Ai (Guy Edward DuGan II) known as gss1147 wrote a structured ranking template and fed it to each LLM individually in its own app environment. The template asks models to identify top open-source, fine-tunable LLMs across several categories, including: - top LLMs by parameter class - top models trained or fine-tuned on highly respected datasets - top models trained on many datasets - best models for Mixture-of-Experts merges - “best of the best” model recommendations This makes the dataset useful for studying **how different LLMs interpret the same research prompt**, what models they recommend, how consistent their rankings are, and how much overlap or disagreement exists across systems. ## Dataset Creation ### Curation Rationale This dataset was created to compare how multiple LLMs respond to the same detailed research prompt about: - open-source LLM quality - context-window requirements - fine-tunability and trainability - benchmark strength - candidate models for merging and MoE workflows The core design principle is **same prompt, different model/app**, allowing side-by-side review of model recommendations and reasoning styles. ### Source Data The source prompt template was written by the dataset creator. The uploaded template includes requirements such as: - models must be open source and free to use - models must have context windows from 128k to unlimited - models must be fine-tunable/trainable - models must have high benchmarks compared to closed models It also defines the ranking buckets and comparison sections used across all collected outputs. [oai_citation:1‡Facts on llm’s .pdf](sediment://file_00000000c500722f83d6fb554174d0ce) ### Data Collection Process The collection process is: 1. A single template was written by the dataset creator. 2. That same template was entered manually into different LLM apps. 3. Each app/model produced its own answer. 4. Those answers were saved as documents and gathered into this dataset. Because each response comes from a different app or model environment, the dataset is best understood as a **comparative output corpus** rather than a normalized benchmark table. ### Who Curated the Dataset Curated by **gss1147**. ## What the Template Asked the Models The source template asks models to rank LLMs in three size classes: - **1 Billion & Under** - **3 Billion to 5 Billion** - **7 Billion to 10 Billion** It also asks for: - **20 LLMs fine-tuned or pre-trained on the most respected datasets** - **20 LLMs fine-tuned or pre-trained on the most datasets** - **Top 5 LLMs best suited for Mixture-of-Experts merges** - **Top 5 “best of the best” LLMs** These instructions are explicitly present in the creator’s template. [oai_citation:2‡Facts on llm’s .pdf](sediment://file_00000000c500722f83d6fb554174d0ce) ## Supported Tasks and Use Cases This dataset is useful for: - **LLM output comparison** - **Prompt consistency studies** - **Model recommendation analysis** - **Research on ranking agreement/disagreement across LLMs** - **LLM self-reported knowledge comparison** - **Model-merging research support** - **Extracting candidate open-source models for follow-up validation** Possible downstream uses: - compare which models are repeatedly recommended across assistants - measure ranking stability across apps - identify hallucinated versus plausible model suggestions - build a retrieval layer over multi-LLM research responses - convert outputs into a structured comparison table ## Dataset Structure ### Data Instances Each instance is best thought of as one collected answer from one LLM/app in response to the same template. Example conceptual structure: ```json { "prompt_template": "List your top LLMs for each of the 3 weight classes...", "source_app": "name of LLM app or platform", "source_model": "name of responding model if known", "response_text": "full model answer", "response_format": "pdf or extracted text", "topic": "open-source LLM ranking and merge candidate selection" }