Falcon-H1 Collection Falcon-H1 Family of Hybrid-Head Language Models (Transformer-SSM), including 0.5B, 1.5B, 1.5B-Deep, 3B, 7B, and 34B (pretrained & instruction-tuned). • 38 items • Updated Sep 14 • 57
Qwen3-Coder Collection The Qwen3-Coder models deliver SOTA advancements in agentic coding and code tasks. Includes Qwen3-Coder-480B-A35B. • 9 items • Updated 28 days ago • 13
Qwen3 Collection Qwen's new Qwen3 models. In Unsloth Dynamic 2.0, GGUF, 4-bit and 16-bit Safetensor formats. Includes 128K Context Length variants. • 79 items • Updated 28 days ago • 226
Running 1.45k 1.45k Big Code Models Leaderboard 📈 Submit code models for evaluation and view leaderboard
view article Article Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance By tiiuae and 5 others • May 21 • 36
Large Multi-modal Models Can Interpret Features in Large Multi-modal Models Paper • 2411.14982 • Published Nov 22, 2024 • 19