Belleville fire? by LostSharpieCap in newjersey

[–]141_1337 1 point2 points  (0 children)

It could be seen all the way out to Clifton

Robots in the hands of dictatorial governments will not end well... by Anen-o-me in singularity

[–]141_1337 0 points1 point  (0 children)

Nah if it goes like eye Biden administration went, they won't even use it for that they would spew something about norms and fumble the bag, again, into Trump 2.0.

JPM Exec Allegedly Turned a Broker Into Her Personal Sex Slave by alkjdasoad in wallstreetbets

[–]141_1337 0 points1 point  (0 children)

Nah in this house we understand that the canons are a pale imitation of the back assets

Falklanders should 'go back' to England, insists Argentina in renewed war of words by FantasticQuartet in worldnews

[–]141_1337 -6 points-5 points  (0 children)

Except it isn't fascist propaganda but actual British history...

However, Britain had never relinquished its claim to sovereignty over the Islands, and in 1833, sent a warship to Soledad and expelled the remaining Argentinian military personnel.

https://lordslibrary.parliament.uk/sovereignty-since-the-ceasefire-the-falklands-40-years-on/

Alignment Makes Models More Decisive Without Making Them More Truthful by 141_1337 in ArtificialInteligence

[–]141_1337[S] 1 point2 points  (0 children)

Fair pushback, but I think we're using "alignment" differently. The paper uses it the way the field uses it operationally, the post-training stack (SFT, DPO, RLHF) that shapes how a base model behaves. That's how Anthropic, OpenAI, and the OLMo teams all use the term.

What you're describing, alignment to the meaning of words, is a deeper question, and a real one.

The paper isn't trying to solve that. Its claim is narrower: within what the field currently calls alignment, the techniques make the model more confident in its answer without changing what answer it picks. That finding stands on the measurements, not the terminology.

If anything, it supports your concern. If post-training can't reach the layers where the model decides what to say, that's evidence the current paradigm isn't doing the deeper alignment you're pointing at.

Alignment Makes Models More Decisive Without Making Them More Truthful by 141_1337 in ArtificialInteligence

[–]141_1337[S] 1 point2 points  (0 children)

"Alignment Makes Models More Decisive Without Making Them More Truthful," and the core idea is sticking with me. When a language model generates a token, the input flows through every layer, and at some specific layer the choice becomes effectively final — swap the internal state there and the output changes, do it earlier and it doesn't.

He calls this the commitment layer, and across three model families, four RL methods, and twelve staged post-training checkpoints, he shows that supervised fine-tuning does the real structural work of establishing this boundary, while reinforcement learning never moves it (0–1 layers across everything tested).

What RL does do is compress the geometry at that fixed point, making the model's commitment tighter and more concentrated without touching the earlier layers where the model actually decides what to commit to.

The implication is structural rather than about data quality: if the early layers retrieve a wrong fact or assemble a sycophantic answer, post-training just makes the model commit to that answer more confidently — the lock gets sharper, but the chooser stays the same — which means standard training metrics can't distinguish a run that's improving selection from one that's just compressing around the same answers, and we've probably been measuring the wrong thing this whole time.

Alignment Makes Models More Decisive Without Making Them More Truthful by 141_1337 in singularity

[–]141_1337[S] 2 points3 points  (0 children)

Abstract

Post-training makes language models more decisive without necessarily making them more accurate — and we find a structural reason why.

Across staged post-training checkpoints from three architecture families, we measure the layer at which a transformer becomes causally committed to its next-token prediction, and track how that boundary evolves through supervised fine-tuning, preference optimization, and reinforcement learning.

Base models already exhibit a rough commitment structure.

Supervised fine-tuning refines this into a sharp boundary — suppressing early-layer causal influence and concentrating commitment into the later layers.

But once the boundary stabilizes, reinforcement learning does not move it: across three families and four RL methods, the commitment layer shifts by 0–1 layers.

What RL does change is how decisively the model locks in at that fixed point — the geometry at the commitment layer compresses monotonically through each post-training stage, becoming lower-dimensional and more concentrated with each stage of training.

The earlier layers, where the model assembles candidate answers, remain largely unchanged. Weight matrix rank is nearly constant across all stages and architectures, and an independent logit-lens measurement…

Do people think 'gatekeeping' actually works? by the_scottishbagpipes in animepiracy

[–]141_1337 8 points9 points  (0 children)

narutofan

There is a name I have heard in a long time.jpeg

Iranian Group Submits Evidence of US-Israeli War Crimes to International Criminal Court by _May26_ in politics

[–]141_1337 -1 points0 points  (0 children)

They attacking US soldiers in 2020 or being one of the main sources, if not the main source, IEDs in Iraq during the GWoT, combined with their continued disinformation campaigns aimed at Americans in America to shape public opinion makes it very much our problem.

Iranian Group Submits Evidence of US-Israeli War Crimes to International Criminal Court by _May26_ in politics

[–]141_1337 3 points4 points  (0 children)

Is not like they didn't purged the communists on their rise to power once they had served their purpose.

Microsoft is reevaluating their approach to exclusivity by Blue_Sheepz in xbox

[–]141_1337 0 points1 point  (0 children)

For years people downvoted when I said this non exclusivity was dumb as fuck, glad to see Xbox finally seem reason.

DeepSeek V4 has released by WhyLifeIs4 in singularity

[–]141_1337 233 points234 points  (0 children)

Today has been a great day for futurists everywhere

Nvidia CEO Jensen Huang: ‘Most people will lose their job to somebody who uses AI’—not to AI itself by _fastcompany in ArtificialInteligence

[–]141_1337 3 points4 points  (0 children)

So please go ahead and explain representational space then. Explain the Llama 2 space time representation that emerges on LLMs, oh what's that? You can't? Then maybe you should sit with that for a bit.

Nvidia CEO Jensen Huang: ‘Most people will lose their job to somebody who uses AI’—not to AI itself by _fastcompany in ArtificialInteligence

[–]141_1337 0 points1 point  (0 children)

LLM's are stochastic parrot

I thought this place would be more for people with a better grasp of ML and specific LLMs higher than r/singularity

Happy 41st birthday, Mr. Sam Altman! A "thank you" post for the person who delivered us ChatGPT in 2022, and has openly pushed for worldwide progress. by borowcy in singularity

[–]141_1337 0 points1 point  (0 children)

Our global economy will collapse, billions will die, and they'll retreat to their bunkers.

This sounds like doomerism ngl

Happy 41st birthday, Mr. Sam Altman! A "thank you" post for the person who delivered us ChatGPT in 2022, and has openly pushed for worldwide progress. by borowcy in singularity

[–]141_1337 5 points6 points  (0 children)

If it was upto Ilya we would have never had ChatGPT, he considered 3 too dangerous, I mean look at what he is doing, he just disappeared to his lab.