view post Post 384 Run Qwen-Image-2512, the new SOTA text-to-image model! 💜It's the top performing open diffusion model and has more realistic + accurate images/text.Run locally with 14GB RAM via our Dynamic GGUF: unsloth/Qwen-Image-2512-GGUFGuide: https://unsloth.ai/docs/models/qwen-image-2512 See translation 1 reply · 🔥 3 3 ❤️ 1 1 + Reply
view post Post 3769 You can now run GLM-4.7, the new 355B parameter SOTA model on your local device (128GB RAM).✨The model achieves SOTA performance on coding, agentic and chat benchmarks.GGUF: unsloth/GLM-4.7-GGUFGuide: https://docs.unsloth.ai/models/glm-4.7 See translation 3 replies · 🔥 6 6 ❤️ 3 3 🚀 1 1 + Reply
view post Post 2266 Google releases FunctionGemma, a new 270M parameter model that runs on just 0.5 GB RAM.✨Built for tool-calling, run locally on your phone at 50+ tokens/s, or fine-tune with Unsloth & deploy to your phone.GGUF: unsloth/functiongemma-270m-it-GGUFDocs + Notebook: https://docs.unsloth.ai/models/functiongemma See translation 2 replies · 👍 4 4 🤗 1 1 + Reply
view post Post 5307 NVIDIA releases Nemotron 3 Nano, a new 30B hybrid reasoning model! 🔥Has 1M context window & best in class performance for SWE-Bench, reasoning & chat. Run the MoE model locally with 24GB RAM.GGUF: unsloth/Nemotron-3-Nano-30B-A3B-GGUF💚 Step-by-step Guide: https://docs.unsloth.ai/models/nemotron-3 See translation 1 reply · 🔥 12 12 ❤️ 7 7 🤗 4 4 👍 1 1 + Reply
view post Post 2038 Mistral's new SOTA coding models Devstral 2 can now be Run locally! (25GB RAM) 🐱We fixed the chat template, so performance should be much better now!24B: unsloth/Devstral-Small-2-24B-Instruct-2512-GGUF123B: unsloth/Devstral-2-123B-Instruct-2512-GGUF🧡Step-by-step Guide: https://docs.unsloth.ai/models/devstral-2 See translation 🔥 8 8 🚀 5 5 ❤️ 3 3 🤗 2 2 + Reply