YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)
- deepseek-ai/DeepSeek-R1 license: mit
Model Overview
Description:
Model created from the nvidia/DeepSeek-R1-NVFP4 checkpoint by:
- converting all layers targeted by modelopt NVFP4 format to compressed-tensors format
- applying FP8_BLOCK quantization to targeted attention layers
More information at https://github.com/vllm-project/llm-compressor/pull/2228
Runs successfully on 4 B200s:
from vllm import LLM, SamplingParams
prompts = ["The Swiss Alps are", "Brad Marchand is", "The Toronto Maple Leafs are"]
# Create a sampling params object for greedy sampling
sampling_params = SamplingParams(
temperature=0.80, top_p=0.95, max_tokens=40, min_tokens=10
)
llm = LLM(
"bdellabe/DeepSeek-R1-NVFP4-FP8-BLOCK",
tensor_parallel_size=4,
max_model_len=4096,
enforce_eager=True,
)
output = llm.generate(prompts, sampling_params)
for out in output:
print(out.outputs[0].text)
- Downloads last month
- 68
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support