Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-to-speech", model="microsoft/VibeVoice-Realtime-0.5B")
Load model directly
from transformers import VibeVoiceStreamingForConditionalGenerationInference model = VibeVoiceStreamingForConditionalGenerationInference.from_pretrained("microsoft/VibeVoice-Realtime-0.5B", dtype="auto")
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support