gss1147's picture
Update README.md
1924f27 verified
---
pretty_name: AI Helps Finding Best Merging LLMs
language:
- en
license: other
task_categories:
- text-generation
- text-ranking
- summarization
- question-answering
- text-retrieval
tags:
- llm
- llm-comparison
- model-ranking
- model-merging
- mixture-of-experts
- moe
- prompt-comparison
- ai-research
- open-source-llm
- reasoning
- coding
- long-context
size_categories:
- n<1K
---
# Dataset Card for AI Helps Finding Best Merging LLMs
## Dataset Summary
**AI Helps Finding Best Merging LLMs** is a **prompt-response comparison dataset** created by manually submitting the same user-written evaluation template to multiple LLM applications and collecting their responses.
The creator and founder of WithIn Us Ai (Guy Edward DuGan II) known as gss1147 wrote a structured ranking template and fed it to each LLM individually in its own app environment. The template asks models to identify top open-source, fine-tunable LLMs across several categories, including:
- top LLMs by parameter class
- top models trained or fine-tuned on highly respected datasets
- top models trained on many datasets
- best models for Mixture-of-Experts merges
- “best of the best” model recommendations
This makes the dataset useful for studying **how different LLMs interpret the same research prompt**, what models they recommend, how consistent their rankings are, and how much overlap or disagreement exists across systems.
## Dataset Creation
### Curation Rationale
This dataset was created to compare how multiple LLMs respond to the same detailed research prompt about:
- open-source LLM quality
- context-window requirements
- fine-tunability and trainability
- benchmark strength
- candidate models for merging and MoE workflows
The core design principle is **same prompt, different model/app**, allowing side-by-side review of model recommendations and reasoning styles.
### Source Data
The source prompt template was written by the dataset creator. The uploaded template includes requirements such as:
- models must be open source and free to use
- models must have context windows from 128k to unlimited
- models must be fine-tunable/trainable
- models must have high benchmarks compared to closed models
It also defines the ranking buckets and comparison sections used across all collected outputs. [oai_citation:1‡Facts on llm’s .pdf](sediment://file_00000000c500722f83d6fb554174d0ce)
### Data Collection Process
The collection process is:
1. A single template was written by the dataset creator.
2. That same template was entered manually into different LLM apps.
3. Each app/model produced its own answer.
4. Those answers were saved as documents and gathered into this dataset.
Because each response comes from a different app or model environment, the dataset is best understood as a **comparative output corpus** rather than a normalized benchmark table.
### Who Curated the Dataset
Curated by **gss1147**.
## What the Template Asked the Models
The source template asks models to rank LLMs in three size classes:
- **1 Billion & Under**
- **3 Billion to 5 Billion**
- **7 Billion to 10 Billion**
It also asks for:
- **20 LLMs fine-tuned or pre-trained on the most respected datasets**
- **20 LLMs fine-tuned or pre-trained on the most datasets**
- **Top 5 LLMs best suited for Mixture-of-Experts merges**
- **Top 5 “best of the best” LLMs**
These instructions are explicitly present in the creator’s template. [oai_citation:2‡Facts on llm’s .pdf](sediment://file_00000000c500722f83d6fb554174d0ce)
## Supported Tasks and Use Cases
This dataset is useful for:
- **LLM output comparison**
- **Prompt consistency studies**
- **Model recommendation analysis**
- **Research on ranking agreement/disagreement across LLMs**
- **LLM self-reported knowledge comparison**
- **Model-merging research support**
- **Extracting candidate open-source models for follow-up validation**
Possible downstream uses:
- compare which models are repeatedly recommended across assistants
- measure ranking stability across apps
- identify hallucinated versus plausible model suggestions
- build a retrieval layer over multi-LLM research responses
- convert outputs into a structured comparison table
## Dataset Structure
### Data Instances
Each instance is best thought of as one collected answer from one LLM/app in response to the same template.
Example conceptual structure:
```json
{
"prompt_template": "List your top LLMs for each of the 3 weight classes...",
"source_app": "name of LLM app or platform",
"source_model": "name of responding model if known",
"response_text": "full model answer",
"response_format": "pdf or extracted text",
"topic": "open-source LLM ranking and merge candidate selection"
}