Update README.md
Browse files
README.md
CHANGED
|
@@ -128,7 +128,7 @@ vLLM aslo supports OpenAI-compatible serving. See the [documentation](https://do
|
|
| 128 |
)
|
| 129 |
|
| 130 |
# Save to disk in compressed-tensors format
|
| 131 |
-
save_path = model_name + "-quantized.w4a16
|
| 132 |
model.save_pretrained(save_path)
|
| 133 |
tokenizer.save_pretrained(save_path)
|
| 134 |
print(f"Model and tokenizer saved to: {save_path}")
|
|
|
|
| 128 |
)
|
| 129 |
|
| 130 |
# Save to disk in compressed-tensors format
|
| 131 |
+
save_path = model_name + "-quantized.w4a16"
|
| 132 |
model.save_pretrained(save_path)
|
| 133 |
tokenizer.save_pretrained(save_path)
|
| 134 |
print(f"Model and tokenizer saved to: {save_path}")
|