Pclanglais nielsr HF Staff commited on
Commit
b8c3f1a
·
verified ·
1 Parent(s): 74f76dc

Improve model card: Add library name, pipeline tag and link to code (#2)

Browse files

- Improve model card: Add library name, pipeline tag and link to code (78c9661e3a5b4c6796a7b433de3fee2b18343bb8)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +9 -6
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
2
- license: apache-2.0
 
3
  language:
4
  - en
5
  - fr
6
  - it
7
  - de
8
  - es
9
- base_model:
10
- - PleIAs/Pleias-1.2B-Preview
 
11
  ---
12
 
13
-
14
  # Pleias-RAG-1B
15
 
16
  <div align="center">
@@ -62,7 +63,7 @@ The structured reasoning traces include the following steps:
62
  ### Multilinguality
63
  Pleias-RAG-1B is able to read and write in the main European languages: French, German, Italian, Spanish, Polish, Latin and Portuguese.
64
 
65
- To date, it is the only SLM with negligible loss of performance in leading European languages for RAG-related tasks. On a translated set of HotPotQA we observed a significant drop of performance in most SLMs from 10\% to 30-35\% for sub-1B models.
66
 
67
  <p align="center">
68
  <img width="80%" src="figures/language_benchmark.png">
@@ -88,6 +89,8 @@ The easiest way to deploy Pleias-RAG-1B is through [our official library](https:
88
  A typical minimal example:
89
 
90
  ```python
 
 
91
  rag = RAGWithCitations("PleIAs/Pleias-RAG-1B")
92
 
93
  # Define query and sources
@@ -122,4 +125,4 @@ With 1.2B parameters, Pleias-RAG-1B can be readily deployed in many constrained
122
 
123
  We also release an [unquantized GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-1B-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
124
 
125
- Once integrated into a RAG system, Pleias-RAG-1B can also be used in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make SLMs workable in production by relying systematically on an externalized memory.
 
1
  ---
2
+ base_model:
3
+ - PleIAs/Pleias-1.2B-Preview
4
  language:
5
  - en
6
  - fr
7
  - it
8
  - de
9
  - es
10
+ license: apache-2.0
11
+ library_name: transformers
12
+ pipeline_tag: text-generation
13
  ---
14
 
 
15
  # Pleias-RAG-1B
16
 
17
  <div align="center">
 
63
  ### Multilinguality
64
  Pleias-RAG-1B is able to read and write in the main European languages: French, German, Italian, Spanish, Polish, Latin and Portuguese.
65
 
66
+ To date, it is the only SLM with negligible loss of performance in leading European languages for RAG-related tasks. On a translated set of HotPotQA we observed a significant drop of performance in most SLMs from 10% to 30-35% for sub-1B models.
67
 
68
  <p align="center">
69
  <img width="80%" src="figures/language_benchmark.png">
 
89
  A typical minimal example:
90
 
91
  ```python
92
+ from rag_library import RAGWithCitations
93
+
94
  rag = RAGWithCitations("PleIAs/Pleias-RAG-1B")
95
 
96
  # Define query and sources
 
125
 
126
  We also release an [unquantized GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-1B-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
127
 
128
+ Once integrated into a RAG system, Pleias-RAG-1B can also be used in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make SLMs workable in production by relying systematically on an externalized memory.