Skip to content

Commit

Permalink
[Intel embeddings] Actually fix table (huggingface#1910)
Browse files Browse the repository at this point in the history
  • Loading branch information
pcuenca authored Mar 18, 2024
1 parent 9a0010d commit e8e34e4
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions intel-fast-embedding.md
Original file line number Diff line number Diff line change
Expand Up @@ -151,10 +151,10 @@ Quantizing the models' weights to a lower precision introduces accuracy loss, as
The table below shows the average accuracy (on multiple datasets) of each task type (MAP for Reranking, NDCG@10 for Retrieval), where `int8` is our quantized model and `fp32` is the original model (results taken from the official MTEB leaderboard). The quantized models show less than 1% error rate compared to the original model in the Reranking task and less than 1.55% in the Retrieval task.

<table>
<tr><th> </th><th> Reranking </th><th> Retrieval </th></tr>
<tr><th> </th><th> Reranking </th><th> Retrieval </th></tr>
<tr><td>

| |
| &nbsp; |
| --------- |
| BGE-small |
| BGE-base |
Expand Down Expand Up @@ -373,4 +373,4 @@ print(results)
Done! The created pipeline can be used to retrieve documents from a document store and rank the retrieved documents using (another) embedding models to re-order the documents.
A more complete example is provided in this [notebook](https://github.com/IntelLabs/fastRAG/blob/main/examples/optimized-embeddings.ipynb).

For more RAG-related methods, models and examples we invite the readers to explore [fastRAG/examples](https://github.com/IntelLabs/fastRAG/tree/main/examples) notebooks.
For more RAG-related methods, models and examples we invite the readers to explore [fastRAG/examples](https://github.com/IntelLabs/fastRAG/tree/main/examples) notebooks.

0 comments on commit e8e34e4

Please sign in to comment.