-
Attention Is All You Need
Paper • 1706.03762 • Published • 99 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 17 -
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Paper • 2305.13245 • Published • 6 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 247
Eli Chen
elichen3051
AI & ML interests
Learning Algorithm, Reinforcement Learning, Data Synthesize, Benchmarking
Organizations
LLM Fundamental papers
-
Attention Is All You Need
Paper • 1706.03762 • Published • 99 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 17 -
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Paper • 2305.13245 • Published • 6 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 247
My interested papers - LLM
models
6
elichen3051/Llama-3.1-8B-GGUF
8B
•
Updated
•
20
elichen3051/Llama-3.2-1B-GGUF-fp16
1B
•
Updated
•
13
elichen3051/Llama-3.1-8B-Instruct-GGUF
8B
•
Updated
•
15
elichen3051/Llama-3.2-1B-Q4_K_M-GGUF
Text Generation
•
1B
•
Updated
•
10
elichen3051/Llama-3.1-8B-Q4_K_M-GGUF
Text Generation
•
8B
•
Updated
•
5
elichen3051/mistral-7B-v0.1-open-orca
Updated