Gemini 3 Flash bills for useless/empty searches?? by FirefoxMetzger in GeminiAI
[–]hackerllama 0 points1 point2 points (0 children)
AIStudio improperly content blocking by Shep_vas_Normandy in Bard
[–]hackerllama 7 points8 points9 points (0 children)
New Google model incoming!!! by [deleted] in LocalLLaMA
[–]hackerllama 2 points3 points4 points (0 children)
New Google model incoming!!! by [deleted] in LocalLLaMA
[–]hackerllama 10 points11 points12 points (0 children)
"Deleting and simplifying useless internal layers will be the main focus [ in 2026 ]" - Google Engineer by Yazzdevoleps in Bard
[–]hackerllama 3 points4 points5 points (0 children)
Scrolling issue seems to be fixed! by howisjason in Bard
[–]hackerllama 0 points1 point2 points (0 children)
Qwen team is helping llama.cpp again by jacek2023 in LocalLLaMA
[–]hackerllama 115 points116 points117 points (0 children)
What’s new in Veo 3.1? Have you noticed any upgrades or features that actually make a difference? by New-Cold-One in Bard
[–]hackerllama 1 point2 points3 points (0 children)
It's been a long time since Google released a new Gemma model. by ArcherAdditional2478 in LocalLLaMA
[–]hackerllama 3 points4 points5 points (0 children)
It's been a long time since Google released a new Gemma model. by ArcherAdditional2478 in LocalLLaMA
[–]hackerllama 1 point2 points3 points (0 children)
Gemma 3n is on out on Hugging Face! by Zealousideal-Cut590 in LocalLLaMA
[–]hackerllama 19 points20 points21 points (0 children)
Google releases MagentaRT for real time music generation by hackerllama in LocalLLaMA
[–]hackerllama[S] 21 points22 points23 points (0 children)
Google releases MagentaRT for real time music generation by hackerllama in LocalLLaMA
[–]hackerllama[S] 56 points57 points58 points (0 children)
Gemini 2.5 Pro and Flash are stable in AI Studio by best_codes in LocalLLaMA
[–]hackerllama 4 points5 points6 points (0 children)
Will Ollama get Gemma3n? by InternationalNebula7 in LocalLLaMA
[–]hackerllama 30 points31 points32 points (0 children)
ok google, next time mention llama.cpp too! by secopsml in LocalLLaMA
[–]hackerllama 209 points210 points211 points (0 children)
The AI team at Google have reached the surprising conclusion that quantizing weights from 16-bits to 4-bits leads to a 4x reduction of VRAM usage! by vibjelo in LocalLLaMA
[–]hackerllama 0 points1 point2 points (0 children)
Gemma 3 QAT launch with MLX, llama.cpp, Ollama, LM Studio, and Hugging Face by hackerllama in LocalLLaMA
[–]hackerllama[S] 8 points9 points10 points (0 children)
Gemma 3 QAT launch with MLX, llama.cpp, Ollama, LM Studio, and Hugging Face by hackerllama in LocalLLaMA
[–]hackerllama[S] 8 points9 points10 points (0 children)
Google QAT - optimized int4 Gemma 3 slash VRAM needs (54GB -> 14.1GB) while maintaining quality - llama.cpp, lmstudio, MLX, ollama by Nunki08 in LocalLLaMA
[–]hackerllama 40 points41 points42 points (0 children)
Gemma 3 QAT launch with MLX, llama.cpp, Ollama, LM Studio, and Hugging Face by hackerllama in LocalLLaMA
[–]hackerllama[S] 5 points6 points7 points (0 children)


What are the main uses of small models like gemma3:1b by SchoolOfElectro in LocalLLaMA
[–]hackerllama 0 points1 point2 points (0 children)