Shaheer1's picture
Added the Tagging for model
6f80beb
---
pipeline_tag: text-classification
library_name: transformers
license: mit
---
# Query Complexity Classifier
This model classifies user queries based on their **complexity level** so they can be routed to an appropriate Large Language Model (LLM).
The model predicts three classes:
* **Simple**
* **Medium**
* **Complex**
It can be used as a **pre-routing layer** in AI systems where different LLMs handle different levels of query complexity.
---
## Model
Base Model: DistilBERT
Task: Text Classification (3 classes)
---
## Download and Use
You can download and load the model directly from Hugging Face using the `transformers` library.
### Install dependencies
```bash
pip install transformers torch
```
### Load the model
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_name = "Shaheer001/Query-Complexity-Classifier"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
```
### Run inference
```python
text = "Explain how Kubernetes architecture works."
inputs = tokenizer(text, return_tensors="pt", truncation=True)
outputs = model(**inputs)
prediction = torch.argmax(outputs.logits, dim=1).item()
labels = ["Simple", "Medium", "Complex"]
print("Predicted Complexity:", labels[prediction])
```
---
## Example
Input:
```
Explain Kubernetes architecture
```
Output:
```
Complex
```
---
## Use Case
This model can be used to build **LLM routing systems** where queries are automatically sent to different language models depending on their complexity.
Example workflow:
User Query → Complexity Classifier → LLM Router → Selected LLM