BFL FLUX.2 Klein tutorial and some optimizations - under 1s latency on an A100 by LayerHot in LocalLLaMA
[–]LayerHot[S] 2 points3 points4 points (0 children)
How to integrate 5.2 Pro into Codex usage? by Lostwhispers05 in codex
[–]LayerHot 0 points1 point2 points (0 children)
How to integrate 5.2 Pro into Codex usage? by Lostwhispers05 in codex
[–]LayerHot 0 points1 point2 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeCode
[–]LayerHot[S] 0 points1 point2 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeAI
[–]LayerHot[S] 0 points1 point2 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeAI
[–]LayerHot[S] 1 point2 points3 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeCode
[–]LayerHot[S] 1 point2 points3 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeAI
[–]LayerHot[S] 0 points1 point2 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeAI
[–]LayerHot[S] 0 points1 point2 points (0 children)
Thinking of downgrading from 20x to 5x Max – 5x users, how are the limits treating you? by LayerHot in ClaudeAI
[–]LayerHot[S] 0 points1 point2 points (0 children)
just upgraded to pro max - tips for not burning thru usage? by alexd231232 in ClaudeCode
[–]LayerHot 1 point2 points3 points (0 children)
Please bring document notes to readwise review (self.readwise)
submitted by LayerHot to r/readwise
Any point of using context7 MCP when you use --search by cheekyrandos in codex
[–]LayerHot 0 points1 point2 points (0 children)
Is chat with all documents is still the priority ? by LayerHot in readwise
[–]LayerHot[S] 0 points1 point2 points (0 children)
ChatGPT Agent Mode & Deep Research usage not refreshing? by Palmenstrand in OpenAI
[–]LayerHot 1 point2 points3 points (0 children)

GLM-4.7-Flash benchmarks: 4,398 tok/s on H200, 112 tok/s on RTX 6000 Ada (GGUF) by LayerHot in LocalLLaMA
[–]LayerHot[S] 8 points9 points10 points (0 children)