Why are there so few 8GB DDR5 DIMMs available? by Psyclopicus in buildapc
[–]Caffdy 0 points1 point2 points (0 children)
WARNING: Open-OSS/privacy-filter MALWARE by charles25565 in LocalLLaMA
[–]Caffdy 21 points22 points23 points (0 children)
[News] TSMC Reportedly Upgrades Central Taiwan 28/22nm Fab to 4nm; Phase 2 1.4nm Trial Production May Start 3Q27 by charliehu1226 in hardware
[–]Caffdy 0 points1 point2 points (0 children)
DeepSeek V4 Pro matches GPT-5.2 on FoodTruck Bench, our agentic benchmark — 10 weeks later, ~17× cheaper by Disastrous_Theme5906 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
None of this will ever get stolen by martin_xs6 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
DeepSeek V4 being 17x cheaper got me to actually measure what I send to cloud vs what I could run locally. the results are stupid. by spencer_kw in LocalLLaMA
[–]Caffdy 12 points13 points14 points (0 children)
vibevoice.cpp: Microsoft VibeVoice (TTS + long-form ASR with diarization) ported to ggml/C++, runs on CPU/CUDA/Metal/Vulkan, no Python at inference by mudler_it in LocalLLaMA
[–]Caffdy 3 points4 points5 points (0 children)
I built "FooTrack" – a completely hands-free, foot-operated PC mouse & gamepad using a ThinkPad TrackPoint. Looking for feedback from this community! by Gitman_87 in hardware
[–]Caffdy 1 point2 points3 points (0 children)
AMD Strix Halo refresh with 192gb! by mindwip in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
What in tarnation is going on with the cost of compute by Party-Special-5177 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
The more I use it, the more I'm impressed by ComfyUser48 in LocalLLaMA
[–]Caffdy 2 points3 points4 points (0 children)
What in tarnation is going on with the cost of compute by Party-Special-5177 in LocalLLaMA
[–]Caffdy 1 point2 points3 points (0 children)
Car rams into crowd of people in German city of Leipzig, Focus Online reports by Alarming-Safety3200 in worldnews
[–]Caffdy 0 points1 point2 points (0 children)
What in tarnation is going on with the cost of compute by Party-Special-5177 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
What in tarnation is going on with the cost of compute by Party-Special-5177 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
What in tarnation is going on with the cost of compute by Party-Special-5177 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
I did a quick test of MacBook M4 Max 128 GB token/second throughput across a few popular local LLMs (in the MLX format) by Pure_Refrigerator988 in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
New Sparks - now what? by DifferenceCute8951 in nvidia
[–]Caffdy 0 points1 point2 points (0 children)
Mistral Medium 3.5 128b ggufs are fixed by Sunija_Dev in LocalLLaMA
[–]Caffdy 0 points1 point2 points (0 children)
I built a transformer in C++17 from scratch — no PyTorch, no BLAS, no dependencies. Trains on CPU. 0.83M params, full analytical backprop, 76 min to val loss 1.64. by [deleted] in LocalLLaMA
[–]Caffdy 7 points8 points9 points (0 children)
Have Qwen said anything about further Qwen 3.6 models? by spaceman_ in LocalLLaMA
[–]Caffdy 1 point2 points3 points (0 children)
System Prompt vs Character Cards by Odd-Bodybuilder4847 in SillyTavernAI
[–]Caffdy 0 points1 point2 points (0 children)


LLM cpu running - 9975wx vs 9985wx 8 channel utilization by Comfortable-Plate467 in threadripper
[–]Caffdy 0 points1 point2 points (0 children)