model-test-onnx
ONNX export of unknown for fast CPU inference.
Model Details
- Source Model: unknown
- Embedding Dimension: unknown
- Format: ONNX (FP32 + INT8)
Files
| File | Description |
|---|---|
model.onnx |
FP32 ONNX model |
model_int8.onnx |
INT8 quantized model (faster) |
tokenizer.json |
Tokenizer configuration |
config_sentence_transformers.json |
Model configuration |
Usage with colbert-onnx (Rust)
use colbert_onnx::Colbert;
let mut model = Colbert::from_pretrained("path/to/model")?;
let embeddings = model.encode_documents(&["Hello world"])?;
Export Tool
This model was exported using pylate-onnx-export:
pip install "pylate-onnx-export @ git+https://github.com/lightonai/next-plaid.git#subdirectory=onnx/python"
pylate-onnx-export unknown --push-to-hub raphaelsty/model-test-onnx