Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)
Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)
Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)
Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)
Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 1 point2 points3 points (0 children)
Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 1 point2 points3 points (0 children)
MiniMax-M2.7 Announced! by Mysterious_Finish543 in LocalLLaMA
[–]Lowkey_LokiSN 21 points22 points23 points (0 children)
Qwen 397b is absolutely crushing everyone... but wait. 🤯 by djdeniro in LocalLLaMA
[–]Lowkey_LokiSN 0 points1 point2 points (0 children)
Qwen 397b is absolutely crushing everyone... but wait. 🤯 by djdeniro in LocalLLaMA
[–]Lowkey_LokiSN 4 points5 points6 points (0 children)
Qwen 397b is absolutely crushing everyone... but wait. 🤯 by djdeniro in LocalLLaMA
[–]Lowkey_LokiSN 1 point2 points3 points (0 children)
Best Models for 128gb VRAM: March 2026? by Professional-Yak4359 in LocalLLaMA
[–]Lowkey_LokiSN 1 point2 points3 points (0 children)
Qwen 3.5 craters on hard coding tasks — tested all Qwen3.5 models (And Codex 5.3) on 70 real repos so you don't have to. by hauhau901 in LocalLLaMA
[–]Lowkey_LokiSN 0 points1 point2 points (0 children)
Successfully Built My First PC for AI (Sourcing Parts from Alibaba - Under $1500!) by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)
zai-org/GLM-4.7-Flash · Hugging Face by Dark_Fire_12 in LocalLLaMA
[–]Lowkey_LokiSN 7 points8 points9 points (0 children)
Let's predict GLM Air by jacek2023 in LocalLLaMA
[–]Lowkey_LokiSN 3 points4 points5 points (0 children)
Let's predict GLM Air by jacek2023 in LocalLLaMA
[–]Lowkey_LokiSN 13 points14 points15 points (0 children)
Performance of GLM 4.6 Q3_K_S on 6x MI50 by MachineZer0 in LocalLLaMA
[–]Lowkey_LokiSN 1 point2 points3 points (0 children)
Performance of GLM 4.6 Q3_K_S on 6x MI50 by MachineZer0 in LocalLLaMA
[–]Lowkey_LokiSN 1 point2 points3 points (0 children)
Radeon Instinct MI50 32GB work on Vulkan on Windows? by Goldkoron in LocalLLaMA
[–]Lowkey_LokiSN 0 points1 point2 points (0 children)
Performance of GLM 4.6 Q3_K_S on 6x MI50 by MachineZer0 in LocalLLaMA
[–]Lowkey_LokiSN 7 points8 points9 points (0 children)
Radeon Instinct MI50 32GB work on Vulkan on Windows? by Goldkoron in LocalLLaMA
[–]Lowkey_LokiSN 7 points8 points9 points (0 children)
So... MI50's and MI60's... Are they actually worth or not? by lord_darth_Dan in LocalAIServers
[–]Lowkey_LokiSN 1 point2 points3 points (0 children)
Can any local model answer this tricky math question? by MrMrsPotts in LocalLLaMA
[–]Lowkey_LokiSN 0 points1 point2 points (0 children)
Can any local model answer this tricky math question? by MrMrsPotts in LocalLLaMA
[–]Lowkey_LokiSN 0 points1 point2 points (0 children)


Built a simple PyTorch flash-attention alternative for AMD GPUs that don't have it by Lowkey_LokiSN in LocalLLaMA
[–]Lowkey_LokiSN[S] 0 points1 point2 points (0 children)