How to use from
llama.cpp
Install from brew
brew install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
# Run inference directly in the terminal:
llama-cli -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
# Run inference directly in the terminal:
llama-cli -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases
# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
# Run inference directly in the terminal:
./llama-cli -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli
# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
# Run inference directly in the terminal:
./build/bin/llama-cli -hf RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
Use Docker
docker model run hf.co/RDson/WomboCombo-R1-Coder-14B-Preview:Q8_0
Quick Links

WomboCombo-14B-Coder

There seems to be an issue where it wont stop generating output. I'll see if I can fix it...

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the sce merge method using Qwen/Qwen2.5-Coder-14B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  # Pivot model
  - model: Qwen/Qwen2.5-Coder-14B
  # Target models
  - model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
  - model: Qwen/Qwen2.5-Coder-14B-Instruct
  - model: arcee-ai/SuperNova-Medius
merge_method: sce
base_model: Qwen/Qwen2.5-Coder-14B
parameters:
  select_topk: 1.0
dtype: bfloat16
Downloads last month
109
Safetensors
Model size
15B params
Tensor type
BF16
ยท
Inference Providers NEW
Input a message to start chatting with RDson/WomboCombo-R1-Coder-14B-Preview.

Model tree for RDson/WomboCombo-R1-Coder-14B-Preview

Spaces using RDson/WomboCombo-R1-Coder-14B-Preview 2