gabopachecoo2000 commited on
Commit
2bd7eb9
·
verified ·
1 Parent(s): 4ca060b

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: apache-2.0
4
+ tags:
5
+ - custom-llm
6
+ - fine-tuning
7
+ - peft
8
+ - lora
9
+ - rag
10
+ ---
11
+
12
+ # Custom LLM with SFT + LoRA + RAG
13
+
14
+ ## Model Description
15
+ This model is a Qwen2.5/7B large language model fine-tuned using **Parameter-Efficient Fine-Tuning (LoRA)** with a custom SFT dataset. It is designed to provide enhanced responses within a specific context defined by the user.
16
+
17
+ ## Training Procedure
18
+ 1. Synthetic SFT pairs generated with ChatGPT.
19
+ 2. Expansion of the SFT dataset to cover broader contexts.
20
+ 3. LoRA adapters trained on Qwen2.5/7B for efficient fine-tuning.
21
+ 4. RAG integration with FAISS vector database for document retrieval.
22
+
23
+ ## Intended Use
24
+ - Conversational AI in specific domains
25
+ - Enhanced question-answering using RAG
26
+ - Applications requiring lightweight fine-tuning without full model training
27
+
28
+ ## Limitations
29
+ - Requires GPU for training
30
+ - RAG performance depends on quality and coverage of the document corpus
31
+ - Behavior outside the trained context may be unpredictable
32
+
33
+ ## Example Usage
34
+
35
+ ```python
36
+ from backend.main import HealthRAG
37
+
38
+ llm = HealthRAG()
39
+ response = llm.ask_enhanced_llm("Explain preventive healthcare tips")
40
+ print(response)
41
+ ```
42
+
43
+ ## How to Cite
44
+ If you use this model in your research or projects, please cite it as:
45
+
46
+ ```
47
+ Custom LLM with SFT + LoRA + RAG, Gabriel Pacheco, 2025
48
+ ```