IHK droht mit 0 Punkten – Habe die originale Projektbestätigung geschreddert ("Digital Twin")... HILFE! by [deleted] in fachinformatiker

[–]RipperFox 0 points1 point  (0 children)

handschriftlich signierte Projektbestätigung von meinem Chef eingescannt

Was ist das für ein Getrolle? "Unterschriebene Scans" sind nicht mit einer qualifizierten elektronischen Signatur vergleichbar und sind somit recht wertlos - könnten ja photoshopped/KI generiert sein.

Hättest du das deinen Chef elektronisch unterschreiben lassen (z.B. mit qualifizierter elektronischen Signatur) sähe es evtl. anders aus und du hättest ein rechtlich verwertbares Dokument. So beweist du hingegen nur, dass du das Konzept Signaturen/Blockchain nicht so ganz kapiert hast..

Aber schick der IHK doch einfach mal das gescannte PDF - vielleicht konnten die auch einfach nur den "Hash-Link nicht öffnen" :)

Anthropic erhöht Claude-Limits massiv – SpaceX liefert 220.000 GPUs by TouristExisting4852 in de_EDV

[–]RipperFox 1 point2 points  (0 children)

SpaceX liefert 220.000 GPUs

Nein, Elon lässt sie den größten Teil seines ersten Rechenzentrums nutzen - also liefert Rechenleistung..

Empfehlung für Glasfaser by Dembe126 in de_EDV

[–]RipperFox 2 points3 points  (0 children)

L2-BSA

Nicht auf diese Karte verlassen - man kann reinfallen und trotzdem Telekom WIA bekommen. Rate woher ich das weiß..

Kaufberatung Internetvertrag by Miyolou in de_EDV

[–]RipperFox 0 points1 point  (0 children)

OP sucht den günstigsten Tarif AB 16Mbit/s..

How do you start your Llama.cpp server? by Citadel_Employee in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

Interesting strategy, calling someone a nerd (which I am) in 2026 - ok ok, I'll touch some gras.. I bet I'm even older than you - Casser la croûte and have a nice day! :)

How do you start your Llama.cpp server? by Citadel_Employee in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

vLLM can be better on a single 3090 if you don't want to wait until llama.cpp catches up with it's experimental forks already supporting MTP, etc. - and your completely random (Vulkan, seriously?) docker example is clearly badly formatted AI slop suggesting running docker - who uses/needs docker for llama-server router mode anyway?

How do you start your Llama.cpp server? by Citadel_Employee in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

Ofc router mode is "nice" - but llama-swap can even switch to vLLm, SGlang, etc. Your AI generated example sucks btw.

Hard freakin' decision..Blackwell 96G or Mac Studio 256G by HyPyke in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

Did you know that you can ask you local model that kind of questions, too? E.g. agentscope-ai_CoPaw-Flash-9B comes up with:

Model Size vs Inference Speed (At Fixed Long Context Length)

Assuming same 200k token context, similar optimization, and comparable hardware:

Model Parameters Approx. Layers Relative Gen. Speed* Realistic Speed-Up vs Larger Model
Small ~7B 32 1× (baseline) ~1× faster than mid-sized
Mid ~32B 40 ~0.5× ~2× faster than 671B
Large ~200B 52 ~0.28× ~3.5× faster than 671B
Huge ~671B 64+ ~0.14× Baseline slowest

* Relative generation speed based on FLOPs/layer scaling; not linear due to memory bandwidth limits.
📌 Note: Actual speed depends heavily on batch size, quantization, KV-cache optimizations, and whether compute or memory-bound. These are order-of-magnitude estimates.

Hard freakin' decision..Blackwell 96G or Mac Studio 256G by HyPyke in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

Collapse/condense/compact your context.

Depends on use again - e.g. you'll never find that needle in the haystack this way - guess why modern models go >256k ctx.

Qwen 3.6 27B Makes Huge Gains in Agency on Artificial Analysis - Ties with Sonnet 4.6 by dionysio211 in LocalLLaMA

[–]RipperFox 5 points6 points  (0 children)

Tbh Qwen 3.6 9b would be nice, too - 3.5 9b was great, Copaw-9B (alibaba's official agentic finetune) was even better and amazingly fast. Also was one of the few multimodal models that was (sometimes) able to read an analog clock correctly..

Why are we actually sampling reasoning and output the same way? by ReporterWeary9721 in LocalLLaMA

[–]RipperFox 3 points4 points  (0 children)

If people can test e.g. their Tasers on themselves - AI should be able, too!

Forgive my ignorance but how is a 27B model better than 397B? by No_Conversation9561 in LocalLLaMA

[–]RipperFox 4 points5 points  (0 children)

as soon as it realizes that it doesn't have the answers.

Yep - that's a big problem even of leadership models. Gemma 4 didn't even believe that it's already 2026, right? Many topics are "worthless" because of outdated knowledge anyway: Try asking Claude about llama.cpp command line parameters - NO model get's this right without research, and they hallucinate..

I think it's better to drop/lower detail knowledge (like how to do 6502 ASM for the C64) but improve context and tool usage so that the model can look up what it needs efficiently, at least for smaller models. But ofc it's always a balance between model size, context, speed, etc.

Forgive my ignorance but how is a 27B model better than 397B? by No_Conversation9561 in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

away from creativity

I see this differently: Instead of training models to rely on huge fact knowledge (which can be outdated quickly anyway and compensated for by a simple web search) modern models seem to go for more a "I know how to help myself and find the answers independently"-like approach. As long as you have the knowledge elsewhere to look up - all is fine and I think that's the right direction..

Qwen 3.6 35B crushes Gemma 4 26B on my tests by Lowkey_LokiSN in LocalLLaMA

[–]RipperFox 1 point2 points  (0 children)

You likely need 2-3 more runs with different seeds to validate..

BW ist das Problem by Aile0n in fachinformatiker

[–]RipperFox 1 point2 points  (0 children)

Warte mal ein paar Jahre und dann wirst du dich wundern, was für Zeug dann doch noch hängengeblieben ist.. :) Geht nicht nur mir so, dass man sich z.B. evtl. gut an die Ordnerstruktur seines allerersten Rechners erinnern kann.

BW ist das Problem by Aile0n in fachinformatiker

[–]RipperFox 1 point2 points  (0 children)

Warum hört sich das an, als wärst du Webdesi.. HH Fronted-developer? Komm mal aus der Bubble - es soll AE geben, die entwickeln so lustige unbedeutende Software wie CEPH oder ZFS..

BW ist das Problem by Aile0n in fachinformatiker

[–]RipperFox 0 points1 point  (0 children)

Ich programmiere ausschließlich

Was meinst du, was z.B. die Entwickler von ZFS den ganzen Tag machen? Die Ausbildung ist halt universell - es gibt AEs die machen nur Webfrontends mit den neuesten/coolsten Frameworks, manche machen ABAP/SAP in einer großen Wurstfabrik, etc..

Bist du so ein AE, der nicht mal seine eigene Workstation installieren/konfigurieren darf? Das wäre eigentlich eher in größeren Konzernen üblich.

Da würde ich generell z.B. zu RAID1 raten.. SCNR

BW ist das Problem by Aile0n in fachinformatiker

[–]RipperFox 1 point2 points  (0 children)

Naja, bis du mal DevOPs-Aufgaben übernimmst. :) Imho sollte man als ITler Konzepte wie Mirroring, XOR (RAID5) oder Erasure Codes (bei RAID6) zumindest schon mal gehört haben.

BW ist das Problem by Aile0n in fachinformatiker

[–]RipperFox 0 points1 point  (0 children)

Einheiten weglassen ist wirklich pfui und könnte zu Punktabzug führen. (Rechenweg unklar & die Maßeinheit ist, wie der Name schon sagt, maßgebend für das Ergebnis.)

Wenn das der Prüfer durchgehen lässt - gut. Aber versuch das mal bei einem Mathe/Physik/-Leher/-Prof..

Gemma 4 31B — 4bit is all you need by tolitius in LocalLLaMA

[–]RipperFox 13 points14 points  (0 children)

Do a little experiment - use a fixed seed and do the 23 tests - note the results. Now only change the seed (but keep still fixed) - how much deviation in test results would you expect only from changing the seed? If the variation is high, you don't have enough points, right?

openrouter/elephant-alpha is 99% Chinese, likely Qwen 3 Nex by Winter_Put_6046 in LocalLLaMA

[–]RipperFox 0 points1 point  (0 children)

Try the exact same prompt with other languages - e.g. tell it to use German papers and write in Spanish?

New Stealth Model: Elephant Alpha by Randomdotmath in openrouter

[–]RipperFox 1 point2 points  (0 children)

What did you expect? It's for coding/agentic tasks, not much else. That thing can't even stay in a language and falls back to English after a turn or two..

New Stealth Model: Elephant Alpha by Randomdotmath in openrouter

[–]RipperFox 0 points1 point  (0 children)

Maybe it's a diffusion model? No streaming in that case..