view post Post 3772 We fixed more issues! Use --jinja for all!* Fixed Nanonets OCR-s unsloth/Nanonets-OCR-s-GGUF* Fixed THUDM GLM-4 unsloth/GLM-4-32B-0414-GGUF* DeepSeek Chimera v2 is uploading! unsloth/DeepSeek-TNG-R1T2-Chimera-GGUF See translation โค๏ธ 3 3 ๐ 3 3 ๐ค 2 2 + Reply
view post Post 3632 Made some 245GB (80% size reduction) 1.8bit quants for Kimi K2! unsloth/Kimi-K2-Instruct-GGUF See translation ๐ฅ 10 10 ๐คฏ 2 2 + Reply
view post Post 3373 It's Qwen3 week! ๐ We uploaded Dynamic 2-bit GGUFs for:Qwen3-Coder: unsloth/Qwen3-Coder-480B-A35B-Instruct-GGUFQwen3-2507: unsloth/Qwen3-235B-A22B-Instruct-2507-GGUFSo you can run them both locally!Guides are in model cards. See translation 1 reply ยท ๐ค 5 5 โค๏ธ 4 4 ๐ฅ 3 3 + Reply
view post Post 5172 Run OpenAI's new gpt-oss models locally with Unsloth GGUFs! ๐ฅ๐ฆฅ20b GGUF: unsloth/gpt-oss-20b-GGUF120b GGUF: unsloth/gpt-oss-120b-GGUFModel will run on 14GB RAM for 20b and 66GB for 120b. See translation 2 replies ยท โค๏ธ 20 20 ๐ฅ 6 6 ๐ 5 5 + Reply
view post Post 5934 Run DeepSeek-V3.1 locally on 170GB RAM with Dynamic 1-bit GGUFs!๐GGUFs: unsloth/DeepSeek-V3.1-GGUFThe 715GB model gets reduced to 170GB (-80% size) by smartly quantizing layers.The 1-bit GGUF passes all our code tests & we fixed the chat template for llama.cpp supported backends.Guide: https://docs.unsloth.ai/basics/deepseek-v3.1 See translation โค๏ธ 18 18 ๐ฅ 9 9 ๐ 5 5 + Reply