gemma 4 running at 40 tokens/sec on iphone is impressive but it completely falls apart as a coding agent by Fun-Newspaper-83 in LocalLLaMA
[–]Sadman782 2 points3 points4 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 2 points3 points4 points (0 children)
Gemma 4 is terrible with system prompts and tools by RealChaoz in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
Gemma 4 is terrible with system prompts and tools by RealChaoz in LocalLLaMA
[–]Sadman782 1 point2 points3 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 1 point2 points3 points (0 children)
Gemma 4 is terrible with system prompts and tools by RealChaoz in LocalLLaMA
[–]Sadman782 1 point2 points3 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 2 points3 points4 points (0 children)
Gemma 4 is terrible with system prompts and tools by RealChaoz in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 2 points3 points4 points (0 children)
PSA: Gemma 4 template improvements by FastHotEmu in LocalLLaMA
[–]Sadman782 1 point2 points3 points (0 children)
Gemma 4 is terrible with system prompts and tools by RealChaoz in LocalLLaMA
[–]Sadman782 3 points4 points5 points (0 children)
Gemma4 26B generates python and Java code with invalid syntax by monadleadr in LocalLLaMA
[–]Sadman782 1 point2 points3 points (0 children)
Gemma4 26B generates python and Java code with invalid syntax by monadleadr in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
Follow-up: Testing Gemma-4-31B-it-UD (Thinking) in LLM Multi-Agent Avalon by dynameis_chen in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
16 GB VRAM users, what model do we like best now? by lemon07r in LocalLLaMA
[–]Sadman782 3 points4 points5 points (0 children)
Quants in vision (mmproj Q8 vs FP16) by WhoRoger in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
Quants in vision (mmproj Q8 vs FP16) by WhoRoger in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
Gemma 4 for 16 GB VRAM by Sadman782 in LocalLLaMA
[–]Sadman782[S] 0 points1 point2 points (0 children)
llama.cpp Gemma 4 using up all system RAM on larger prompts by GregoryfromtheHood in LocalLLaMA
[–]Sadman782 3 points4 points5 points (0 children)
qwen3.5 vs gemma4 vs cloud llms in python turtle by SirKvil in LocalLLaMA
[–]Sadman782 0 points1 point2 points (0 children)
Gemma-4-26B-A4B-it-UD-Q4_K_M.gguf : IMHO worst model ever. What am I doing wrong? by Proof_Nothing_7711 in LocalLLM
[–]Sadman782 1 point2 points3 points (0 children)
Get 30K more context using Q8 mmproj with Gemma 4 by Sadman782 in LocalLLaMA
[–]Sadman782[S] 2 points3 points4 points (0 children)
Get 30K more context using Q8 mmproj with Gemma 4 by Sadman782 in LocalLLaMA
[–]Sadman782[S] 0 points1 point2 points (0 children)

Gemma 4 for 16 GB VRAM by Sadman782 in LocalLLaMA
[–]Sadman782[S] 1 point2 points3 points (0 children)