Transformers documentation
Candle
Get started
Base classes
Models
Preprocessors
Inference
Pipeline API
Generate API
Optimization
Chat with models
Serving
Training
Quantization
Ecosystem integrations
Resources
API
You are viewing main version, which requires installation from source. If you'd like
regular pip install, checkout the latest stable version (v5.8.1).
Candle
Candle is a machine learning framework providing native Rust implementations of Transformers models. It natively supports safetensors to load Transformers models directly.
/// load model config
let config: Config =
serde_json::from_reader(std::fs::File::open(config_filename)?)?;
/// load safetensors and memory-maps them
let vb = unsafe {
VarBuilder::from_mmaped_safetensors(&filenames, dtype, &device)?
};
/// materialize tensors from VarBuilder into model class
let model = Model::new(args.use_flash_attn, &config, vb)?;Transformers integration
- The hf-hub crate checks your local Hugging Face cache for a model. If it isn’t there, it downloads model weights and configs from the Hub.
- VarBuilder lazily loads the safetensor files. It maps state-dict key names to Rust structs representing model layers. This mirrors how Transformers organizes its weights.
- Candle parses
config.jsonto extract model metadata and instantiates the matching Rust model class with weights fromVarBuilder.
Resources
- Candle documentation