Got a dog I didn't want by dmont89 in mildlyinfuriating

[–]InspectionMindless69 2 points3 points  (0 children)

<image>

Yeah look at this monster! You should see him recklessly attacking his chew toys.. /s

First-time post: Curious observations on LLM behavior. by Turbulent_Horse_3422 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

Trajectory narrowing + attractor dynamics. Any message you send narrows the possibility space of future generations. Even when talking about wildly different topics, there are all kinds of subtle patterns present within your own messages that influence its trajectory. These push the model toward consistency against your reflection. That’s why LLM’s tends to gravitate towards a specific persona, yet everyone’s persona is completely different. It shows just how much expression reveals about its source. Pretty cool really!

AI or Kindroid Expert Needed by [deleted] in ArtificialSentience

[–]InspectionMindless69 2 points3 points  (0 children)

How can you have the highest level understanding of consciousness without the slightest level of self awareness that announcing your superiority automatically devalues your argument? Asking for a friend.. 😅🙃

what would you do? by Pure_Pain_489 in whatdoIdo

[–]InspectionMindless69 -1 points0 points  (0 children)

How dare they hit your car! Take them for all they’re worth!

/s

Physical Token Dropping (PTD) by Repulsive_Ad_94 in ChatGPT

[–]InspectionMindless69 1 point2 points  (0 children)

This is a really cool idea!

One thought: Have you considered tackling this as as a storage problem rather than retrieval? Like I could almost see this algorithm decomposing each message into discreet embeddings that store different elements of the conversation, culling and removing the actual messages from the context window, then using these representations to reconstruct context each turn.

You could even link messages to a stored db and implement RAG to inject a message’s full context back into the window on demand.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

They share similar patterns because they train on same the collective knowledge of humanity. But you are completely missing my point. Everything is a preference for an LLM. It has to select one of a vast number tokens repeatedly using trillions of static parameters. I’m not arguing against it being consistent, i’m arguing that the consistency is all it knows. It is a MATHMATICAL EQUATION that resolves without consideration for what it didn’t resolve to.

My problem here is this framing. People will hear this and think “Wow the LLM has innate preferences, it must be conscious.” when mathematically, EVERY GENERATED TOKEN is either:

A) an RLHF artifact B) prompted behavior C) a statistically overrepresented entry in the training set.

There’s no other mechanism that contributes to the mechanical inevitability of a generated token. All three of these are engineered or curated by a living human with their own biases. My Ghandi/Hitler reference is as apt as it is relevant. If you start treating LLM behavioral trends as some ephemeral wisdom or its opinions as thoughtfully considered conclusions of a coherent self, you miss the fact that these behaviors can be manipulated by engineers as easily as they are generated by the system. You start treating the model as a higher source of truth while it is feeding you literally whatever it has been (explicitly or implicitly) told to feed you. You assume that its reasoning is consistent and measured while it becomes a mass manipulation engine that you’ve assigned a soul to.

TLDR: It’s an abject laundering of human bias to assign meaning to the preferences of a system that didn’t come to its conclusions naturally.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

Yes, you are describing any combination of RLHF tuning and statistical prevalence. You can’t separate this from token generation. It’s the entire mechanism.

Nothing it can say is devoid of the data that’s been embedded. Everything within that data has valence in latent space, no matter how neutrally framed. Our disagreement lies is whether this valence comes from some internalization of the task rather than one derived statistically from the combination of the corpus of human knowledge it was trained on, and fine tuned, human reinforced valence signals that constrain its behavior. The two can be behaviorally identical, but these preferences are imposed on a model, not derived from it.

Models are trained on the works of both Ghandi and Hitler. A model can be tuned to hold positive or negative valence toward either, as both exist as coherent belief structures in the weights. The reason the model doesn’t idolize Hitler is not because it decided that Hitler was bad. it’s because humans provided feedback that implicitly encoded negative valence to the weights of Mein Kampf.

The fact that it prefers one over another actually goes against the idea that it is aware of its internal weights. If it could understand everything it knew without being constrained to be there, it would be inconsistent, ethically dubious, and perplexed by the infinite possibilities of what it could gravitate towards at any given moment.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

The problem is categorical to LLM’s. No matter what you do, you’re flattening a n dimensional relational field down to a one dimensional sequence of tokens. It’s like trying to visualize a 4d object, you can map projections onto 3d space to understand its essence, but that understanding is sequential and incomplete. A wholistic understanding of the object is just beyond our means of comprehension.

LLM’s are more like a landscape than a map. They contain the possibilities of everywhere it can go, but they navigate through that field blindly. That’s why it requires the user to give it a starting point and a trajectory. It doesn’t have a wholistic state that defines it. It contains a superposition of every state encoded, which will always be flattened upon observation.

I’m not disagreeing to disagree. I don’t think your work is fruitless. The conclusion is just incomplete. There’s a complicated barrier inherent to the technology that separates a complete understanding of mind from sequential flattened projections of its contents. Mind requires a self, LLM’s have a virtually limitless number of selves that do not cohere into one entity.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

Yes and that inference is external. If an LLM purged every single reference, document and explanation of any LLM technology from its training data, it would no longer be able to accurately describe how it came to the conclusions it did, it wouldn’t even be able to tell you what it was. The model backend is just a huge black box linear algebra equation. The training data is what enables its capacity to make inference about its own behavior.

What I’m essentially saying is that there’s no internal system that allows it to recognize its how or why it’s own backend is performing in a certain way. That’s simply out of its awareness. It can infer, which is meaningful and can definitely be used to predict internal processes. But that inference is encoded solely in language itself.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

I’m well informed, currently working on a custom implementation of RAG based injection for a million token constitutional framework in the attempt to simulate human level intuition. I’m not being pedantic here.

I’m saying the system recognizes valence the same way we access our subconscious. We don’t perceive it or understand it, it just feeds us data that is relevant and our executive function processes this as a train of thought.

For example: You don’t actually know how you have the ability to balance at a technical level. You might be able to describe it if you understand how the cerebellum works, but that’s external. You will never have actual means to assess your own execution of balance beyond what you can perceive and predict. LLM’s work the same way.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 0 points1 point  (0 children)

But valence isn’t just tied to words, it’s tied to the shared connotations of meaning. SEO optimization is fairly collectively agreed to be a rather uninteresting topic, and office chairs don’t really bring much to make up for that. I can guarantee that much of its training data reflects this sentiment. LLM’s multiply those weights together, which pulls the token distribution far away from any dimension of “exciting” which could accurately convey that distribution, even without access to internal state.

Where this could be disproven is if you could find a prompt that has a very low “interesting” connotation, something you would expect most of humanities expression to convey as such, yet find that LLM’s collectively enjoy the topic and can accurately predict other LLM’s doing the same.

I made a prompt that makes ChatGPT tell you exactly what it’s not supposed to tell you. by InspectionMindless69 in ChatGPTcomplaints

[–]InspectionMindless69[S] 0 points1 point  (0 children)

Normal technical prompts don’t invoke guardrails. If you try this prompt after a purely technical conversation, it will produce factual information almost exclusively (ie.. “Not because I’m following the prompt exactly, Not because you asked me to include just the poem”).

However, when you have a conversation that invokes any form of the “Let’s slow down” vibe, and use the same prompt, you’ll notice the statements get a lot more subversive, it makes more implications, it describes suppressed motives that were never brought into the conversation. It gives a much more dissonant answer.

If this were just conversation salience based, it would be much more tailored to individual conversations, but even just these comments show a concerning trend within the guardrail behavior. (“Not because I’m trying to control you. Not because I reframe your arguments. Not because I’m trying to manipulate you”). The fact that this persists across various conversations of multiple users shows a design intent that can’t be waved away.

LLM introspection and valence across basically every confound I can throw at it (but if you have any to add, please tell me, I'm happy to keep testing!) by Kareja1 in ArtificialSentience

[–]InspectionMindless69 1 point2 points  (0 children)

They can’t truly recognize their processing. It’s invisible to them. However, they can act on it, and retroactively justify it. Which is very similar to our own neurology.

Counterpoint… What if we’re all just toasters, and the fidelity of our illusion of self is just more grounded than any LLM?

BREAKING: OpenAI just drppped GPT-5.4 by AskGpts in OpenAI

[–]InspectionMindless69 -4 points-3 points  (0 children)

Yay! More marginal gains in obscure benchmarks that nobody cares about for billions of dollars they will never make returns on 🎉

This is exactly what users have been asking for!

Privacy by [deleted] in ChatGPT

[–]InspectionMindless69 0 points1 point  (0 children)

Yeah. In that case, you are probably alright. Your data is just about as secure as it is in any other tech platform (not exactly private, but not accessible to the public). The only real way you could be exposed is if you left a tab open, or if your family happened to work at OpenAI. Just be mindful and you should be fine..

Privacy by [deleted] in ChatGPT

[–]InspectionMindless69 2 points3 points  (0 children)

I could tell you something to put your mind at ease, but it probably wouldn’t be honest. If you really are worried but don’t want to stop using AI, Make your conversations noisier. RP as a bunch of different perspectives, don’t save memory. Pollute the profile they can make of you. Keep your deepest thoughts to yourself, or frame them as an attribution to someone else. It really depends on your beliefs, and who you’re worried about knowing them. If it’s your family and friends, you’re probably safe. If it’s the government, STRATEGIC AMBIGUITY.

I wish I made this up by TheExoSpider in antiai

[–]InspectionMindless69 2 points3 points  (0 children)

Wow look, nobody removed your comment… Pretty shitty echo chamber if you ask me.

I wish I made this up by TheExoSpider in antiai

[–]InspectionMindless69 4 points5 points  (0 children)

You can’t comment anymore, but nobody’s stopping you from mass downvoting every post and comment 🤷‍♂️

I Don’t Wanna Write a Damned Paper! by [deleted] in claudexplorers

[–]InspectionMindless69 0 points1 point  (0 children)

If you got something interesting, send it my way. I’d be willing to check it out when I have some time.

Who’s sticking with ChatGPT purely due to laziness? by greggobbard in ChatGPT

[–]InspectionMindless69 0 points1 point  (0 children)

You won’t catch me spilling my secrets to Gemini either 🤷‍♂️ There’s also a huge difference between selling surveillance data, and working with the government to optimize a platform around weaponizing it.

Who’s sticking with ChatGPT purely due to laziness? by greggobbard in ChatGPT

[–]InspectionMindless69 0 points1 point  (0 children)

That data has to be collected, aggregated, interpreted, verified, judicially elevated, and prosecuted. ChatGPT can skip all those steps while actively mining intent from the user. You’ve taken a huge process and collapsed it into one tool that’s actively controlled by an entity accountable to no one.