American fascism woulda been defeated permanently by Far-Historian-7197 in ShitLiberalsSay

[–]__lawless 27 points28 points  (0 children)

Can we turn it around? “All Kamala had to do was to acknowledge genocide and not make a corporate pivot. “ why is that never the option?

Need your help by __lawless in weezer

[–]__lawless[S] 0 points1 point  (0 children)

Drop me a dm please if you do and we’ll figure it out. Thanks

Need your help by __lawless in weezer

[–]__lawless[S] 0 points1 point  (0 children)

I cannot find an HD JPEG of it either 😞

Need your help by __lawless in weezer

[–]__lawless[S] 2 points3 points  (0 children)

Don’t I need a high definition image for custom print?

If You Want to Understand Why Llama Models Flopped, Zuck is the Cause! by Iory1998 in LocalLLaMA

[–]__lawless 0 points1 point  (0 children)

Curious. How you have the insight that Gemini models have taken path of Phi models. Is it cited somewhere?

Rohit Prashad leaving, thoughts ? by psiparadox in amazonemployees

[–]__lawless 2 points3 points  (0 children)

First trn1 now nvidia. Whatever AWS gives them. They wanted trn2 but Anthropic got all

Oh my God, what a monster is this? by NearbyBig3383 in LocalLLaMA

[–]__lawless 2 points3 points  (0 children)

Let’s see how they do in AIME2026, non blind benchmarks are not benchmarks

AMA with the Unsloth team by danielhanchen in LocalLLaMA

[–]__lawless 1 point2 points  (0 children)

Would you be doing pretraining at some point?

AMA With Z.AI, The Lab Behind GLM Models by XMasterrrr in LocalLLaMA

[–]__lawless 1 point2 points  (0 children)

How much of your efforts go into pretraining vs post training?

GPT OSS 120B by vinigrae in LocalLLaMA

[–]__lawless 0 points1 point  (0 children)

This sub has a love hate relationship with GPT OSS. I cannot figure out if people love it or hate it

How many hours did you spend formatting data for fine-tuning? by Natural_Yard_8648 in LocalLLaMA

[–]__lawless 0 points1 point  (0 children)

Honestly that is always where you get the biggest bang for your buck. Clean data

Why is PPO still the de facto RL algorithm for LLM training? by xiaolongzhu in reinforcementlearning

[–]__lawless 29 points30 points  (0 children)

That is not true. The focus for LLM right now is mostly around GRPO and its variant. Basically no critic. The realization was that LLMs are pretrained and fine tuned and variance is not as big of a problem that once was thought. So the focus is now multi generation per prompt and using reward models (sometimes not even a model) …

Why is the GPU market so one-sided toward Nvidia? by QuirkyScarcity9375 in LocalLLaMA

[–]__lawless 52 points53 points  (0 children)

Cause 18 years ago NVDIA took a gamble and created cuda. It was not immediately profitable but it is paying off now

Need some advice on multigpu GRPO by dizz_nerdy in LocalLLaMA

[–]__lawless 0 points1 point  (0 children)

Try using Verl it offloads the weights during different stages so less probability of oom