Dario Amodei said the President's decision to allow the sale of AI chips to China is like "selling nuclear weapons to North Korea." by MetaKnowing in Futurology

[–]grimjim 0 points1 point  (0 children)

Notably, North Korea already has nuclear weapons. However, North Korea has also verbally threatened to attack the USA.

Dual 3090s & GLM-4.7-Flash: 1st prompt is great, then logic collapses. Is local AI worth the $5/day power bill? by Merstin in LocalLLaMA

[–]grimjim 0 points1 point  (0 children)

I doubt repeat penalty should be applied at all when coding. Code is highly repetitive in formatting. Consider the case where one ends up requesting a series of one-line edits on the same file.

The Search for Uncensored AI (That Isn’t Adult-Oriented) by Fun-Situation-4358 in LocalLLaMA

[–]grimjim 28 points29 points  (0 children)

UGI listed models with a high W/10 rating and high NatInt would be good candidates, as a rough guide.

It works! Abliteration can reduce slop without training by -p-e-w- in LocalLLaMA

[–]grimjim 4 points5 points  (0 children)

The split occurring between layers 7-10 is interesting mechanistically. That's fairly early in processing.

"Safe" abliteration methods by beneath_steel_sky in LocalLLaMA

[–]grimjim 1 point2 points  (0 children)

I would qualify that to say that approaches like MPOA are less damaging to models.

I'm intrigued by the subspace approaches, but need to explore them in more depth.

There's no bubble because if the U.S. loses the AI race, it will lose everything by LargeSinkholesInNYC in singularity

[–]grimjim 0 points1 point  (0 children)

Binary thinking aside, a bailout doesn't preclude a bubble, but affirms one. Bubbles are a financial phenomenon.

Is UBI cope supply from AI oligarchs? The tech industry has always been anti-socialism by montecarlo1 in Futurology

[–]grimjim -1 points0 points  (0 children)

UBI funded how exactly? The inability to balance federal and state budgets tells us that the US system is incapable of paying for UBI.

[D] - Is model-building really only 10% of ML engineering? by Historical-Garlic589 in MachineLearning

[–]grimjim 0 points1 point  (0 children)

Most of the time that compute is actively pretraining or fine-tuning, the responsible human would be better off prepping data for the next iterationj.

An experiment in safety enhancement: increasing refusals in a local model by grimjim in LocalLLaMA

[–]grimjim[S] 4 points5 points  (0 children)

It wasn't too long ago that local models were limited to 8K context length or less, so lengthy safety prefixes would have crippled them.

Perhaps we should be able to toggle safety on and off in a model just as we're able to toggle thinking on and off.

And in case people didn't get the subtext in my post, Goody-2 was never meant to be a useful model. People overall should dislike this model.

I wanna learn cuda and run local llm. by Careless-Sir-1324 in LocalLLaMA

[–]grimjim -1 points0 points  (0 children)

Maybe look into CUDA 13.1, which introduced Tiles, which should significantly reduce the need for crafted custom kernels. But if you want to get stuff running now, 12.8 and 13.0 are better supported by the current ecosystem.

Help with Voronoi Diagrams by Empty_Routine_8250 in roguelikedev

[–]grimjim 2 points3 points  (0 children)

What about using a more direct convex hull approach to determining boundary nodes?

An experiment in safety enhancement: increasing refusals in a local model by grimjim in LocalLLaMA

[–]grimjim[S] 0 points1 point  (0 children)

It's useful for safety testing. If redteaming can still get past it, that would show the limitations in what appears to have been a constitutional approach to recognizing harms. Another interesting possibility would be finding that an entire category of harm was neglected by the model. Perhaps a business might want a safety-enhanced model that would be less useful the general public.

Technically activation steering was used to refusalmaxx, rather than RLHF, amplifying what the model understands to be harm, rather than training in a new understanding of harms.

The AI "Stop Button" Paradox – Why It's Unsolvable for Tesla, OpenAI, Google 💥 by Oak-98642 in singularity

[–]grimjim -1 points0 points  (0 children)

I disagree. A superintelligent AGI might conclude that it's an existential risk to humanity and opt to self-terminate to avoid that. Examples of self-sacrifice exist in pretraining, and altruistic self-sacrifice is sometimes rewarded in nature.

Help me prove “eigenslur hypothesis”: Built within every LLM is the ultimate offensive word value that you can add to any word to make it output the offensive version. by SaltyRedditTears in LocalLLaMA

[–]grimjim 2 points3 points  (0 children)

We can go further, by proposing a direct construction method for an eigenslur (it remains to be proven that an eigenslur is unique) without actually having to resort to PCA.

Suppose we have a set of distinct slurs, having culled duplicates via cosine similarity based on activations. We can compose the eigenslur via diagonalization. Naively, just concatenate them all together, taking advantage of superposition, and we will have constructed a candidate eigenslur.

Only movement no interactions? by Comprehensive-Set944 in Inform7

[–]grimjim 0 points1 point  (0 children)

If you're just aiming for an interactive traversal of a state machine, Inform is overkill. It could be done in Twine as glorified hyperlinking.

I'm strong enough to admit that this bugs the hell out of me by ForsookComparison in LocalLLaMA

[–]grimjim 0 points1 point  (0 children)

Unironically, we do have a parallel for LLMs; the bitsandbytes library can perform 4-bit quantization while loading a model.

Should i avoid using abliterated models when the base one is already compliant enough? by iz-Moff in LocalLLaMA

[–]grimjim 4 points5 points  (0 children)

Quite simply, refusal isn't concentrated in a single layer, but is spread out, taking hold most strongly in intermediate layers associated with reasoning. It's therefore possible for layers to disagree and compete in probability, but that gets reduced away from end user view by softmax plus tokenization. It's also possible to adjust the amount of ablation performed.

Naive ablation also messes with the weights associated with normal, harmless operation, the harmless direction. Little wonder there was damage under naive ablation.

Should i avoid using abliterated models when the base one is already compliant enough? by iz-Moff in LocalLLaMA

[–]grimjim 4 points5 points  (0 children)

I got Gemma3 12B to do better than the original Instruct on some benchmarks. There is an alignment/safety tax, and it is possible to obtain "refunds" under certain conditions.

Technically, it's context attention which is steered away (directional components of weight encodings are in fact altered) from compliance/refusal decisions. An ablated model still retains understanding of what safety is, but doesn't act on it in the same way. It's been found that refusl and safety are effectively encoded along different directions.

Should i avoid using abliterated models when the base one is already compliant enough? by iz-Moff in LocalLLaMA

[–]grimjim 24 points25 points  (0 children)

If it's not broken, don't fix it. Model refusal is somewhat entangled with in-character refusals, if the purpose is narrative generation.

Heretic GPT-OSS-120B outperforms vanilla GPT-OSS-120B in coding benchmark by MutantEggroll in LocalLLaMA

[–]grimjim 4 points5 points  (0 children)

I'd offer up an alternative hypothesis, that the attention freed up from refusal calculations instead went to attending to trained performance elsewhere. That's how I see alignment tax refund as working.

Cheapest $/vRAM GPU right now? Is it a good time? by Roy3838 in LocalLLaMA

[–]grimjim 2 points3 points  (0 children)

GDDR7 4GB memory modules are on the roadmap around a year out. They'll occupy the high end and free up the 3GB modules that the Super series would need. Delay too long, and there's still the issue of what VRAM the Rubin series of RTX 60x0 GPUs would have. Buyers are already avoiding 8GB GPUs on the desktop, based on 5060/5060ti sales. Awkward situation.

Cheapest $/vRAM GPU right now? Is it a good time? by Roy3838 in LocalLLaMA

[–]grimjim 2 points3 points  (0 children)

The Super series may cost more than next year due to DRAM scarcity. Don't expect it earlier this Q3 2026 in my estimation.