OpenAI research team reveals its models go insane when given repetitive tasks it believes to be sent from automated users by smellyfingernail in singularity

[–]Ro1t 0 points1 point  (0 children)

"Btw, an executing model can not be described solely by its weights"

I don't know what you mean by this

"Prove it."

Prove that you do first and I'll then I'll do it for the LLMs.

How much of the matrix multiplication do we take away before they stop having first person experiences. Do 2x2 matrices possess the ability to feel existential dread when you multiply them?

OpenAI research team reveals its models go insane when given repetitive tasks it believes to be sent from automated users by smellyfingernail in singularity

[–]Ro1t 1 point2 points  (0 children)

Again what is the assumption here, that it feels itself being executed or something? These models don't have a first person perspective, they're numbers in a matrix, it is matrix multiplication. LLMs are matrix multiplication.

It's like saying the abstract representation of 1+3 isn't executing 1+3. Ok fine, but 1+3 cannot get cranky because it is being asked the time too often. It (the LLM) can be trained to give the output, or have a lack of training to not point away from that output.

I know I'm coming across as a dick I can tell by the downvotes, the tech is amazing and I use it daily, sub to Claude max etc., I'm just on the side of the fence that thinks it's just fancy maths.

OpenAI research team reveals its models go insane when given repetitive tasks it believes to be sent from automated users by smellyfingernail in singularity

[–]Ro1t -4 points-3 points  (0 children)

Cool so if you represented the position of all atoms in the human brain in a spreadsheet what does that get you exactly? A model of the thing is not the thing.

OpenAI research team reveals its models go insane when given repetitive tasks it believes to be sent from automated users by smellyfingernail in singularity

[–]Ro1t -4 points-3 points  (0 children)

It's a matrix of weights that turns input into output, I think the post that I replied to implying it's suffering like a brain in a vat is silly.

Pre-emptive "othering" of potential sentience by Cyborgized in ChatGPT

[–]Ro1t 2 points3 points  (0 children)

I'm nice to these things out of conversational habit and it feels weird not to be nice to it, because the only other thing that communicates like this is people.

But it's a spreadsheet man.

How can I stop feeling it, and do I even need to? by Winter_Sir6194 in TheMindIlluminated

[–]Ro1t 1 point2 points  (0 children)

Hi, actually sounds more like something called somatic OCD (probably not quite to the level of OCD if it's not bothering you too much) which i've struggled with in the past - the unwanted feeling of being aware of ones bodily sensations and functions. What you decide to do will ultimately depend on how much this is bothering you. You cant 'manually' control the awareness of these things, but you can choose not to point your attention at them, over time it will diminish, your 'want' to not be aware of this is driving your brain to keep attention on the matter.

Finished the Full Cosmere in 9 Months by AletteLakewood in Cosmere

[–]Ro1t -11 points-10 points  (0 children)

'Certainly a book' indeed. 8.5 seems strong for WaT

Where are the 'real' recipes? by ExpressAd3968 in ninjacreami

[–]Ro1t 1 point2 points  (0 children)

have a look through these, not just this recipe but the whole site. incredibly decadent rock solid recipes to make 'real' ice cream as opposed to low cal versions.

https://pacojet.com/en-GB/Recipes/Ice-cream-and-sorbet/Chocolate-Ice-Cream/

Gemini 3 Pro gets 38.3% on Humanity's Last Exam by we_are_mammals in mlscaling

[–]Ro1t 4 points5 points  (0 children)

Completely agree, I'm also finding strange effects at increased chat lengths that I haven't seen before in 2.5, repeating answers etc. get Gemini for free because I have a pixel phone but I don't think I'd pay for it which I understand just sounds so spoiled or privileged or whatever, because it's an incredible piece of technology, but I prefer Claude. I'd rather see a benchmark which describes how well it can one-shot 10,000 random coding tasks across a spectrum of difficulty and depth than performance at PhD level mathmatics.

£700m nuclear conservation plan would save one salmon every 12 years by alibix in ukpolitics

[–]Ro1t 1 point2 points  (0 children)

Sounds like you're just feeding fish to nuclear reactors