GenKnowSub: Improving Modularity and Reusability of LLMs through General Knowledge Subtraction Paper • 2505.10939 • Published May 16 • 3
SearchInstruct: Enhancing Domain Adaptation via Retrieval-Based Instruction Dataset Creation Paper • 2509.10708 • Published Sep 12 • 17
Chain-of-Experts: Unlocking the Communication Power of Mixture-of-Experts Models Paper • 2506.18945 • Published Jun 23 • 40
🧠Reasoning datasets Collection Datasets with reasoning traces for math and code released by the community • 24 items • Updated May 19 • 174
view article Article Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM Mar 12 • 471
view article Article Training and Finetuning Embedding Models with Sentence Transformers v3 May 28, 2024 • 259
RobustFT: Robust Supervised Fine-tuning for Large Language Models under Noisy Response Paper • 2412.14922 • Published Dec 19, 2024 • 88
Law of the Weakest Link: Cross Capabilities of Large Language Models Paper • 2409.19951 • Published Sep 30, 2024 • 54
Programming Every Example: Lifting Pre-training Data Quality like Experts at Scale Paper • 2409.17115 • Published Sep 25, 2024 • 63
Instruction Following without Instruction Tuning Paper • 2409.14254 • Published Sep 21, 2024 • 30
To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning Paper • 2409.12183 • Published Sep 18, 2024 • 39
view article Article Fine-tuning LLMs to 1.58bit: extreme quantization made easy Sep 18, 2024 • 272
view article Article Welcome Falcon Mamba: The first strong attention-free 7B model Aug 12, 2024 • 113
Scaling Synthetic Data Creation with 1,000,000,000 Personas Paper • 2406.20094 • Published Jun 28, 2024 • 104
Instruction Pre-Training: Language Models are Supervised Multitask Learners Paper • 2406.14491 • Published Jun 20, 2024 • 95