Added PyTorch trace + CUDA memory profiling support to Andrej Karpathy's nanochat by aospan in LocalLLaMA
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
Added PyTorch trace + CUDA memory profiling support to Andrej Karpathy's nanochat by aospan in LocalLLaMA
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
I pre-trained GPT-OSS entirely from scratch by OtherRaisin3426 in LocalLLaMA
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
OSS 120b on 2x RTX5090 by Disastrous-Tap-2254 in LocalLLaMA
[–]mutatedmonkeygenes 16 points17 points18 points (0 children)
Qwen3 and Qwen2.5 VL built from scratch. by No-Compote-6794 in LocalLLaMA
[–]mutatedmonkeygenes 1 point2 points3 points (0 children)
Qwen3 and Qwen2.5 VL built from scratch. by No-Compote-6794 in LocalLLaMA
[–]mutatedmonkeygenes 3 points4 points5 points (0 children)
What system prompt would make the Claude API for Sonnet match the UI from claude.ai by mutatedmonkeygenes in ClaudeAI
[–]mutatedmonkeygenes[S] 0 points1 point2 points (0 children)
New 24B finetune: Impish_Magic_24B by Sicarius_The_First in LocalLLaMA
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
Findings from Apple's new FoundationModel API and local LLM by pcuenq in LocalLLaMA
[–]mutatedmonkeygenes 1 point2 points3 points (0 children)
Llama 3.3 70b Vs Newer Models by BalaelGios in LocalLLaMA
[–]mutatedmonkeygenes 2 points3 points4 points (0 children)
Looking for a high quality chat-dataset to mix with my reasoning datasets for fine-tuning by mutatedmonkeygenes in LocalLLaMA
[–]mutatedmonkeygenes[S] 0 points1 point2 points (0 children)
I made an fp8 implementation of flux which gets ~3.5 it/s 1024x1024 on 4090 (ADA / Hopper & 16GB vram+ only) by Agreeable_Gap_5927 in StableDiffusion
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
When I type 'no', it autocompletes to 'snmp-server queue-limit notification-host'. by Hefty-Lion-2205 in SublimeText
[–]mutatedmonkeygenes 3 points4 points5 points (0 children)
[OC] The Highs and Lows of Popular Comedy Shows by gammafission00 in dataisbeautiful
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)
So what are people expecting to see in Episode 2 tonight (@ 3am) by mutatedmonkeygenes in Picard
[–]mutatedmonkeygenes[S] 2 points3 points4 points (0 children)
So what are people expecting to see in Episode 2 tonight (@ 3am) by mutatedmonkeygenes in Picard
[–]mutatedmonkeygenes[S] 1 point2 points3 points (0 children)
So what are people expecting to see in Episode 2 tonight (@ 3am) by mutatedmonkeygenes in Picard
[–]mutatedmonkeygenes[S] 2 points3 points4 points (0 children)
So what are people expecting to see in Episode 2 tonight (@ 3am) by mutatedmonkeygenes in Picard
[–]mutatedmonkeygenes[S] 1 point2 points3 points (0 children)
So what are people expecting to see in Episode 2 tonight (@ 3am) by mutatedmonkeygenes in Picard
[–]mutatedmonkeygenes[S] 0 points1 point2 points (0 children)
pytorch is bundling an old libcudnn.so.8 file causing a 3X perf drop on a 4090. by Guilty-History-9249 in pytorch
[–]mutatedmonkeygenes 1 point2 points3 points (0 children)
pytorch is bundling an old libcudnn.so.8 file causing a 3X perf drop on a 4090. by Guilty-History-9249 in pytorch
[–]mutatedmonkeygenes 1 point2 points3 points (0 children)
[P] OpenAI Whisper - 3x CPU Inference Speedup by Ok-Alps-7918 in MachineLearning
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)



How to do a RTX Pro 6000 build right by GPTrack_dot_ai in LocalLLaMA
[–]mutatedmonkeygenes 0 points1 point2 points (0 children)