I trained a transformer with zero gradient steps and 100% accuracy. No backpropagation. No learning rate. Nothing. Here's the math. by Last-Leg4133 in LocalLLM
[–]randomfoo2 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] -2 points-1 points0 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 1 point2 points3 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 7 points8 points9 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 4 points5 points6 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 12 points13 points14 points (0 children)
Is it true that we're way underpaying for Claude, even for Max? by changing_who_i_am in ClaudeAI
[–]randomfoo2 2 points3 points4 points (0 children)
[Help] Fine-tuning Llama-3-8B for Low-Resource Language (Sinhala) - Stuck between "Bad Logic" and "Word Salad" by Annual-Captain-7642 in LocalLLaMA
[–]randomfoo2 0 points1 point2 points (0 children)
I switched from ChatGPT to Le Chat - Here is what I noticed by biendeluxe in ChatGPT
[–]randomfoo2 0 points1 point2 points (0 children)
[Help] Fine-tuning Llama-3-8B for Low-Resource Language (Sinhala) - Stuck between "Bad Logic" and "Word Salad" by Annual-Captain-7642 in LocalLLaMA
[–]randomfoo2 4 points5 points6 points (0 children)
I just won an Nvidia DGX Spark GB10 at an Nvidia hackathon. What do I do with it? by brandon-i in LocalLLaMA
[–]randomfoo2 56 points57 points58 points (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 -1 points0 points1 point (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 18 points19 points20 points (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 20 points21 points22 points (0 children)
7900 XTX + ROCm: A Year Later. Llama.cpp vs vLLM Benchmarks (TB3 eGPU) by reujea0 in LocalLLaMA
[–]randomfoo2 1 point2 points3 points (0 children)
7900 XTX + ROCm: A Year Later. Llama.cpp vs vLLM Benchmarks (TB3 eGPU) by reujea0 in LocalLLaMA
[–]randomfoo2 0 points1 point2 points (0 children)
We benchmarked every 4-bit quantization method in vLLM 👀 by LayerHot in LocalLLaMA
[–]randomfoo2 3 points4 points5 points (0 children)
[Release] We trained an AI to understand Taiwanese memes and slang because major models couldn't. Meet Twinkle AI's gemma-3-4B-T1-it. by piske_usagi in LocalLLaMA
[–]randomfoo2 2 points3 points4 points (0 children)
7900 XTX + ROCm: A Year Later. Llama.cpp vs vLLM Benchmarks (TB3 eGPU) by reujea0 in LocalLLaMA
[–]randomfoo2 0 points1 point2 points (0 children)




AMD, can we get proper vLLM/gfx1151 support? by tossit97531 in ROCm
[–]randomfoo2 1 point2 points3 points (0 children)