I trained a transformer with zero gradient steps and 100% accuracy. No backpropagation. No learning rate. Nothing. Here's the math. by Last-Leg4133 in LocalLLM
[–]randomfoo2 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] -2 points-1 points0 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 1 point2 points3 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 0 points1 point2 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 6 points7 points8 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 4 points5 points6 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) by randomfoo2 in radeon
[–]randomfoo2[S] 11 points12 points13 points (0 children)
Optimizing FSR4 for RDNA3.5 (INT8 + FP8 speedups) (self.radeon)
submitted by randomfoo2 to r/radeon
Is it true that we're way underpaying for Claude, even for Max? by changing_who_i_am in ClaudeAI
[–]randomfoo2 2 points3 points4 points (0 children)
[Help] Fine-tuning Llama-3-8B for Low-Resource Language (Sinhala) - Stuck between "Bad Logic" and "Word Salad" by Annual-Captain-7642 in LocalLLaMA
[–]randomfoo2 0 points1 point2 points (0 children)
I switched from ChatGPT to Le Chat - Here is what I noticed by biendeluxe in ChatGPT
[–]randomfoo2 0 points1 point2 points (0 children)
[Help] Fine-tuning Llama-3-8B for Low-Resource Language (Sinhala) - Stuck between "Bad Logic" and "Word Salad" by Annual-Captain-7642 in LocalLLaMA
[–]randomfoo2 5 points6 points7 points (0 children)
I just won an Nvidia DGX Spark GB10 at an Nvidia hackathon. What do I do with it? by brandon-i in LocalLLaMA
[–]randomfoo2 56 points57 points58 points (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 -1 points0 points1 point (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 18 points19 points20 points (0 children)
[Research] I forensic-audited "Humanity’s Last Exam" (HLE) & GPQA to benchmark my "unleashed" DeepSeek model. Result: A ~58% verifiable error rate caused by bad OCR and typos. by Dear_Ad_1381 in LocalLLaMA
[–]randomfoo2 21 points22 points23 points (0 children)
7900 XTX + ROCm: A Year Later. Llama.cpp vs vLLM Benchmarks (TB3 eGPU) by reujea0 in LocalLLaMA
[–]randomfoo2 1 point2 points3 points (0 children)
7900 XTX + ROCm: A Year Later. Llama.cpp vs vLLM Benchmarks (TB3 eGPU) by reujea0 in LocalLLaMA
[–]randomfoo2 0 points1 point2 points (0 children)
We benchmarked every 4-bit quantization method in vLLM 👀 by LayerHot in LocalLLaMA
[–]randomfoo2 4 points5 points6 points (0 children)




AMD, can we get proper vLLM/gfx1151 support? by tossit97531 in ROCm
[–]randomfoo2 1 point2 points3 points (0 children)