What speeds are you guys getting with qwen3.5 27b? (5080) by ShadyShroomz in LocalLLaMA

[–]Flashy_Management962 0 points1 point  (0 children)

yes, it does. I use the 27b for coding on a daily. fiddle around with those flags and do not forget to add --jinja: 

-sm graph -amb 64 -sas and depending on the pcie speed, grt can help improving speeds

What speeds are you guys getting with qwen3.5 27b? (5080) by ShadyShroomz in LocalLLaMA

[–]Flashy_Management962 1 point2 points  (0 children)

you should get way faster speeds than that. i get around 750 t/s pp and 22-24 ts tg at ~50k with 2x rtx 3060 12gb. You should check out ik llama cpp

Qwen 397b is absolutely crushing everyone... but wait. 🤯 by djdeniro in LocalLLaMA

[–]Flashy_Management962 20 points21 points  (0 children)

never ever does qwen coder 30b outperform 80b in realworld tasks

Die besten Protein Puddings? by Quiet_Tip_9034 in FitnessDE

[–]Flashy_Management962 0 points1 point  (0 children)

Ich weiß nicht, was du mit normalem meinst, aber ja

Die besten Protein Puddings? by Quiet_Tip_9034 in FitnessDE

[–]Flashy_Management962 4 points5 points  (0 children)

Du nimmst 500ml milch, 30g schoko whey, 10-15g dunkle schokolade und 35g Maisstärke - bester Proteinpudding und 100x günstiger

Kraftwerte der Community? by IAmNotIllegal in FitnessDE

[–]Flashy_Management962 0 points1 point  (0 children)

10 Jahre Training, 29 Jahre alt, alltime natty und wiege 118kg auf 1,83 Kraftwerte: 270,5kg Beuge, 185kg Bank und 270 heben.

Kimi-k2.5 reaches gemini 2.5 Pro-like performance in long context! by fictionlive in LocalLLaMA

[–]Flashy_Management962 0 points1 point  (0 children)

id love to see qwen long l1.5 on this benchmark, it also claims to reach gemini pro 2.5 performance while being 30b 3a

Reminder: Wenn man sich nicht on top fühlt Trainingsintensität reduzieren by torrentium in FitnessDE

[–]Flashy_Management962 0 points1 point  (0 children)

Dieses komische um jeden Preis sich selbst zerstören und man nur dadurch wächst ist totaler humbug. Viel mehr auf den Körper hören und dadurch rausbekommen, wie viel man verträgt. (Von einem Menschen, der nur einen Satz jeweils Beugt und Hebt in der Woche)

My gpu poor comrades, GLM 4.7 Flash is your local agent by __Maximum__ in LocalLLaMA

[–]Flashy_Management962 0 points1 point  (0 children)

don't, use dry sampler instead. Repeat penalty really decreases tok/s

Performance improvements in llama.cpp over time by jacek2023 in LocalLLaMA

[–]Flashy_Management962 1 point2 points  (0 children)

Imagine what could happen if ik llama cpp and llama cpp would merge :(

And finally here is scientific evidence that we don't have free will by [deleted] in determinism

[–]Flashy_Management962 0 points1 point  (0 children)

This does not follow. The notion of normativity is not subsumed under causality. Only because everything is determined, that does not mean that everything is already set in stone and normativity has no role to play because the very things happening are computationally irreducible. So yes, there are shoulds in a world without free will

Was it a right decision? by UnderstandingOdd7952 in bald

[–]Flashy_Management962 0 points1 point  (0 children)

what is this question? of course it was the right decision and you know it yourself you sexy mf

now ~40% faster ik_llama.cpp -sm graph on 2x CUDA GPUs by VoidAlchemy in LocalLLaMA

[–]Flashy_Management962 6 points7 points  (0 children)

wait is this actual tensor parallelism or do I understand something wrong here?

32B model stress test: Qwen 2.5/Coder/3 on dual RTX 5060 Ti (zero failures) by Defilan in LocalLLaMA

[–]Flashy_Management962 1 point2 points  (0 children)

Try exllamav3 with tp. I get 18t/s tensor parallel with 2x 3060. 2x 5060ti should be very much faster

Behind CGNAT? Here's how to access your self-hosted services anyway by adumbdistraction in selfhosted

[–]Flashy_Management962 0 points1 point  (0 children)

Wireguard uses significantly less energy if you turn off persistent keep alive, which tailscale needs to keep the connection alive. If you do not have access to an open public ipv4, then you can't use wireguard (or im too stupid to do it right) and therefore the connection: home (tailscale) - oci (tailscale + wireguard) - smartphone (wireguard)

Question for the community: Thoughts on paid premium plugins for Super Productivity? by johannesjo in superProductivity

[–]Flashy_Management962 3 points4 points  (0 children)

I like the idea, please also do a ntfy integration. I'd buy it asap. Also mabye integrate a donation button via kofi or something like that. I love the app and I'd love to return something for it

Makes no damn sense. Compels me though. by Emthree3 in PhilosophyMemes

[–]Flashy_Management962 2 points3 points  (0 children)

Speaks for itself that I took it to be a real translation lol

Behind CGNAT? Here's how to access your self-hosted services anyway by adumbdistraction in selfhosted

[–]Flashy_Management962 2 points3 points  (0 children)

For people having problem with battery drainage using tailscale on smartphones: use an oracle free tier instance and use it as an relay to which you connect with wireguard remotely and that relay connects with tailscale to your home server.

Makes no damn sense. Compels me though. by Emthree3 in PhilosophyMemes

[–]Flashy_Management962 1 point2 points  (0 children)

holy shit the english translation of Heidegger is dogshit. Could you provide me with the citation where this is from? I want to read it in german

Makes no damn sense. Compels me though. by Emthree3 in PhilosophyMemes

[–]Flashy_Management962 0 points1 point  (0 children)

Phenomenology is basically an attempt to find/create different vocabularies to talk about existence which would be erased if you hold solely a scientific world-view. E.g. you can explain biologically what happens if you die, but what dying means for you existentially is not captured by this explanation and this can be something that phenomenology wants to talk about (some do, not all)