Improve model card: Add library name, pipeline tag and link to code (#2)
Browse files- Improve model card: Add library name, pipeline tag and link to code (78c9661e3a5b4c6796a7b433de3fee2b18343bb8)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
|
@@ -1,16 +1,17 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
- fr
|
| 6 |
- it
|
| 7 |
- de
|
| 8 |
- es
|
| 9 |
-
|
| 10 |
-
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
-
|
| 14 |
# Pleias-RAG-1B
|
| 15 |
|
| 16 |
<div align="center">
|
|
@@ -62,7 +63,7 @@ The structured reasoning traces include the following steps:
|
|
| 62 |
### Multilinguality
|
| 63 |
Pleias-RAG-1B is able to read and write in the main European languages: French, German, Italian, Spanish, Polish, Latin and Portuguese.
|
| 64 |
|
| 65 |
-
To date, it is the only SLM with negligible loss of performance in leading European languages for RAG-related tasks. On a translated set of HotPotQA we observed a significant drop of performance in most SLMs from 10
|
| 66 |
|
| 67 |
<p align="center">
|
| 68 |
<img width="80%" src="figures/language_benchmark.png">
|
|
@@ -88,6 +89,8 @@ The easiest way to deploy Pleias-RAG-1B is through [our official library](https:
|
|
| 88 |
A typical minimal example:
|
| 89 |
|
| 90 |
```python
|
|
|
|
|
|
|
| 91 |
rag = RAGWithCitations("PleIAs/Pleias-RAG-1B")
|
| 92 |
|
| 93 |
# Define query and sources
|
|
@@ -122,4 +125,4 @@ With 1.2B parameters, Pleias-RAG-1B can be readily deployed in many constrained
|
|
| 122 |
|
| 123 |
We also release an [unquantized GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-1B-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
| 124 |
|
| 125 |
-
Once integrated into a RAG system, Pleias-RAG-1B can also be used in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make SLMs workable in production by relying systematically on an externalized memory.
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- PleIAs/Pleias-1.2B-Preview
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
- fr
|
| 7 |
- it
|
| 8 |
- de
|
| 9 |
- es
|
| 10 |
+
license: apache-2.0
|
| 11 |
+
library_name: transformers
|
| 12 |
+
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
|
|
|
| 15 |
# Pleias-RAG-1B
|
| 16 |
|
| 17 |
<div align="center">
|
|
|
|
| 63 |
### Multilinguality
|
| 64 |
Pleias-RAG-1B is able to read and write in the main European languages: French, German, Italian, Spanish, Polish, Latin and Portuguese.
|
| 65 |
|
| 66 |
+
To date, it is the only SLM with negligible loss of performance in leading European languages for RAG-related tasks. On a translated set of HotPotQA we observed a significant drop of performance in most SLMs from 10% to 30-35% for sub-1B models.
|
| 67 |
|
| 68 |
<p align="center">
|
| 69 |
<img width="80%" src="figures/language_benchmark.png">
|
|
|
|
| 89 |
A typical minimal example:
|
| 90 |
|
| 91 |
```python
|
| 92 |
+
from rag_library import RAGWithCitations
|
| 93 |
+
|
| 94 |
rag = RAGWithCitations("PleIAs/Pleias-RAG-1B")
|
| 95 |
|
| 96 |
# Define query and sources
|
|
|
|
| 125 |
|
| 126 |
We also release an [unquantized GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-1B-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
| 127 |
|
| 128 |
+
Once integrated into a RAG system, Pleias-RAG-1B can also be used in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make SLMs workable in production by relying systematically on an externalized memory.
|