| --- |
| license: apache-2.0 |
| --- |
|  |
| # embeddings |
| This is a repo created to keep a collection of quantized bert models in ggml format. |
|
|
| ### usage |
| You can utilize [bert.cpp](https://github.com/skeskinen/bert.cpp) as usual or use [our new api](https://rapidapi.com/nohakcoffee/api/simple-similarity) to quickly prototype on real use-case scenarios for text similarity. |
|
|
|
|
| ### embeddings sample |
|
|
| ``` |
| ./main -m small -p "word" |
| // [0.0698, -0.0024, -0.0153, 0.0193, -0.1060, -0.0278, 0.1424, -0.0056, -0.0536... |
| ``` |
|
|
| ### api reference |
| | api | size | |
| | ---- | ---- | |
| | nano | 11.2 | |
| | small | 14.5 | |
| | medium | 21.3 | |
| | large | 68.8 | |
|
|
| We are planning to update the list to always support the lastest open-source models on the repo and api. |
|
|
| As of 02/20/2024 large model outpus are highly biased towards positive numbers, we are still researching why, nano, small and medium models are working as expected. |
|
|
| ### api sample |
|
|
| ``` |
| // semantic relationship between "I love this" and "I hate this" |
| nano: 0.4614074121704735 |
| small: 0.6553150807627873 |
| medium: 0.8263292187144999 |
| large: 0.8567815005348627 |
| ``` |
|
|
| Note: the api is only for prototyping as of now, we are currently working on scaling things up soon. |