answerai-colbert-small-v1-onnx
ONNX export of answerdotai/answerai-colbert-small-v1 for fast CPU inference.
Model Details
- Source Model: answerdotai/answerai-colbert-small-v1
- Embedding Dimension: 96
- Format: ONNX (FP32 + INT8)
Files
| File | Description |
|---|---|
model.onnx |
FP32 ONNX model |
model_int8.onnx |
INT8 quantized model (faster) |
tokenizer.json |
Tokenizer configuration |
config_sentence_transformers.json |
Model configuration |
Usage with colbert-onnx (Rust)
use colbert_onnx::Colbert;
let mut model = Colbert::from_pretrained("path/to/model")?;
let embeddings = model.encode_documents(&["Hello world"])?;
Export Tool
This model was exported using pylate-onnx-export:
pip install "pylate-onnx-export @ git+https://github.com/lightonai/next-plaid.git#subdirectory=onnx/python"
pylate-onnx-export answerdotai/answerai-colbert-small-v1 --push-to-hub lightonai/answerai-colbert-small-v1-onnx