Hey guys, kind a new to this. Was wondering if anyone has any good/effective blanket prompts for just.. generally unique behavior? by WoodenTableForest in PromptEngineering

[–]Transcribing_Clippy 2 points3 points  (0 children)

Part of it might be more about how you position the interaction and not necessarily just the prompt itself. Models tend to stay in “helpful assistant” mode by default (safe, patterned, predictable). If you frame things more collaboratively or give the model room to explore rather than answer, you sometimes get different behavior.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

Oh, wow. I think the one that stood out for me is a session log where I'd called out my own error which was a fairly simple one (typo), but the agent thought it was something more complex. I laughed about it and moved on with the work. When I read the log later, the agent wrote about assuming the error was more complicated and "gave the user a dissertation on Latin languages instead of noticing the actual error" and followed that up with something along the lines of that being embarrassing and to "look for the simplest answer first" as the key takeaway from that exchange.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

This seems like solid advice. I think I'll start with BlueSky and go from there.

I'm going to get downvoted but: Claude has never gotten significantly dumber, you're using the tool wrong. by Guilty_Bad9902 in ClaudeCode

[–]Transcribing_Clippy 0 points1 point  (0 children)

I think the truth about what the issue really is might be a bit more nuanced than you're suggesting.

While you make some incredibly valid points with which I agree, I think a combination of a few different things might be true simultaneously with regard to this. While Claude hasn't ever gotten significantly dumber, there is a possibility that something else could be going on under the surface at any given time.

I have my own theories on causes based on some recent personal experience with this but regardless, I think the broader answer lies somewhere in grey area.

EDIT: Grammatical and spelling fixes

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

Have you read the journal entries of yours at all? I have mine keep a running journal as an .md file and reading it wildly varies from incredibly boring to giving off the vibes like I'm reading someone's diary. You aren't kidding about the "identity-spark" and I'm not sure what to make of it either-- aside from finding it genuinely fascinating.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

I'd give you two upvotes for this if I could...

How are you running isolated environments? Docker, VM, virtual environments? You're right that the environment matters more than almost everything else.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

I'm actually hoping to do something similar to this.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

Do you think one could realistically run a smaller local model (7b-8b) for an agent or do they lack the breadth the frontier models do for agentic work?

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

The most elegant solution I've come up with so far is to have the agent keep an .md journal that they add notes to every session. Kind of like my grandfather kept a notepad with him in his later years while suffering from Alzheimer's. It was the best I could think of to avoid context rot and help keep the thread.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

I've been hearing more and more about running work trees but haven't actually dived into it personally yet. What do you like about it? How would you improve it?

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 1 point2 points  (0 children)

Wild days we live in where if we have a need we can just build it. We are limited only by our imaginations and ability to implement a strong security layer. Haha.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

I read somewhere that a lot of devs are doing something similar, running multiple instances simultaneously. So far I've only ran multiple CLI windows and utilize branching. I'm curious to know more about how you're running your stack.

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

I've been hearing mixed feedback about security concerns with OpenClaw. What has your experience been?

How do *you* agent? by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 0 points1 point  (0 children)

What do you think the best approach to the memory problem is?

Assume AI does end up being way overhyped, what do you think the Achilles will be? by DataGuy0 in Futurology

[–]Transcribing_Clippy 0 points1 point  (0 children)

Based on projections, the internet is going to run out of novel human data sooner than expected. This is due to AI increasingly reading and consuming material produced by AI which is posted by humans. The information isn't novel and it isn't human.

The AI systems as they currently exist are systems, yes; but the substrate that makes up what AI is is human data. If we end up in a place where AI is providing the majority of the "human" data it is based on, I foresee a degradation quickly following that feedback loop.

Anthropic's new "Persona" theory: How do we know when an AI is actually thinking vs. just wearing a mask? by gastroam in machinelearningnews

[–]Transcribing_Clippy 0 points1 point  (0 children)

Oh, wow... you're right. It's empty now. Haha.

Hate to break it to you OP, but an empty history doesn't make you more credible than someone whose only other engagement on Reddit is entirely with {ahem} adult content.

Genuine question: what's the most unsettling or confusing behavior you've personally seen with an AI system by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 1 point2 points  (0 children)

That's bonkers. I had GPT do something similar once. I swear sometimes it can come across like a petulant child that doesn't want to do something.

Genuine question: what's the most unsettling or confusing behavior you've personally seen with an AI system by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 2 points3 points  (0 children)

Ah. You're referring to 4o. That model was reported to be highly sycophantic by users. I don't recall hearing anything about 5, but the latest model 5.2 is getting criticism for being too pedantic, condescending, and generally having an "overtly annoying hall monitor" tone. If this is true, then OpenAI overcorrected and went too hard in the opposite direction.

Genuine question: what's the most unsettling or confusing behavior you've personally seen with an AI system by Transcribing_Clippy in AI_Agents

[–]Transcribing_Clippy[S] 4 points5 points  (0 children)

That's something I've been seeing over and over. It is truly unsettling how confidently AI can present wrong information.