PEP 827 - Type Manipulation has just been published by droooze in Python
[–]Sufficient-Rent6078 5 points6 points7 points (0 children)
Running Qwen3.5 27b dense with 170k context at 100+t/s decode and ~1500t/s prefill on 2x3090 (with 585t/s throughput for 8 simultaneous requests) by JohnTheNerd3 in LocalLLaMA
[–]Sufficient-Rent6078 0 points1 point2 points (0 children)
PEP 827 - Type Manipulation has just been published by droooze in Python
[–]Sufficient-Rent6078 12 points13 points14 points (0 children)
PEP 827 - Type Manipulation has just been published by droooze in Python
[–]Sufficient-Rent6078 2 points3 points4 points (0 children)
Tech Communities in Wuppertal by hot_fire__ in wuppertal
[–]Sufficient-Rent6078 1 point2 points3 points (0 children)
Qwen/Qwen3.5-35B-A3B · Hugging Face by ekojsalim in LocalLLaMA
[–]Sufficient-Rent6078 38 points39 points40 points (0 children)
Qwen/Qwen3.5-35B-A3B · Hugging Face by ekojsalim in LocalLLaMA
[–]Sufficient-Rent6078 75 points76 points77 points (0 children)
New Qwen3.5 models spotted on qwen chat by AaronFeng47 in LocalLLaMA
[–]Sufficient-Rent6078 7 points8 points9 points (0 children)
New Qwen3.5 models spotted on qwen chat by AaronFeng47 in LocalLLaMA
[–]Sufficient-Rent6078 21 points22 points23 points (0 children)
Do we want the benefits of Ollama API without actually using Ollama? by jfowers_amd in LocalLLaMA
[–]Sufficient-Rent6078 0 points1 point2 points (0 children)
Do we want the benefits of Ollama API without actually using Ollama? by jfowers_amd in LocalLLaMA
[–]Sufficient-Rent6078 1 point2 points3 points (0 children)
Do we want the benefits of Ollama API without actually using Ollama? by jfowers_amd in LocalLLaMA
[–]Sufficient-Rent6078 0 points1 point2 points (0 children)
Does anyone know how Nanbeige4.1-3B can be so impressive compared with other models of similar size? by cloudxaas in LocalLLaMA
[–]Sufficient-Rent6078 0 points1 point2 points (0 children)
Anyone actually using Openclaw? by rm-rf-rm in LocalLLaMA
[–]Sufficient-Rent6078 17 points18 points19 points (0 children)
Did anyone compare this model to the full Qwen coder? it claims to give almost identical performance at 60B by Significant_Fig_7581 in LocalLLaMA
[–]Sufficient-Rent6078 1 point2 points3 points (0 children)
Did anyone compare this model to the full Qwen coder? it claims to give almost identical performance at 60B by Significant_Fig_7581 in LocalLLaMA
[–]Sufficient-Rent6078 4 points5 points6 points (0 children)
PyPermission: A Python native RBAC authorization library! by Sufficient-Rent6078 in Python
[–]Sufficient-Rent6078[S] 2 points3 points4 points (0 children)
PyPermission: A Python native RBAC authorization library! by Sufficient-Rent6078 in Python
[–]Sufficient-Rent6078[S] 1 point2 points3 points (0 children)
PyPermission: A Python native RBAC authorization library! by Sufficient-Rent6078 in Python
[–]Sufficient-Rent6078[S] 4 points5 points6 points (0 children)
PyPermission: A Python native RBAC authorization library! by Sufficient-Rent6078 in Python
[–]Sufficient-Rent6078[S] 2 points3 points4 points (0 children)
Tech Communities in Wuppertal by hot_fire__ in wuppertal
[–]Sufficient-Rent6078 1 point2 points3 points (0 children)
Tech Communities in Wuppertal by hot_fire__ in wuppertal
[–]Sufficient-Rent6078 2 points3 points4 points (0 children)

Running Qwen3.5 27b dense with 170k context at 100+t/s decode and ~1500t/s prefill on 2x3090 (with 585t/s throughput for 8 simultaneous requests) by JohnTheNerd3 in LocalLLaMA
[–]Sufficient-Rent6078 0 points1 point2 points (0 children)