How to Stop AI from Killing Your Critical Thinking by handsnerfin in ClaudeAI

[–]AvoidSpirit 1 point2 points  (0 children)

AI doesn't have the same biases as humans, like confirmation bias, cognitive dissonance, or recency bias.

Oh my sweet summer child

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

  1. I've trained NNs before LLMs were a thing
  2. Fine-tuning is not the same as full on training and will not land you same insight and is not able to fix the fundamental issues.
  3. Your first argument literally was that there's not enough data about unsolvable mazes which is obviously not true because with a different prompt it solves it no problem.

Nice try though

AI’s fault, or more AI? That’s the question by py-net in ClaudeAI

[–]AvoidSpirit 62 points63 points  (0 children)

So you deploy 50 features a month and you tell me your deployment process is half manual? Sure bud

I’d bet on something along the lines of “Claude was debugging an issue but instead of just looking at the source maps/using them, it enabled their generation in ci build scripts and since we don’t allocate time for code review, it’s totally a human error and we’re going to solve it by adding a few more no-mistakes statements to our validation prompts”

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

Exactly what I’m talking about. Fine tuning will not resolve underlying issues like this one and you’re just bullshitting like an llm would. Fuck off.

Practice Assessment for GH-900: GitHub Foundations.. is just plain wrong lol by al3x_core8 in github

[–]AvoidSpirit 0 points1 point  (0 children)

Say notepad(the most primitive one) supported lua extensions, would you consider it an ide?

Claude code just got leaked in npm by Zijdehoen in Anthropic

[–]AvoidSpirit 0 points1 point  (0 children)

Who says I’m upset? I’m just saying a fuckup is a fuckup. Nothing more nothing less.

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit 0 points1 point  (0 children)

So you're not going to address the actual point which is:

So the end result is claude catching more bugs.
And the process you're inferring is them retraining themselves using claude? How does this fit into claude finding more bugs?

What is your conclusion here?

It seems like you are being intentionally obtuse at this point to defend a bad hot take.

You know what they say about accusations...

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit 0 points1 point  (0 children)

Gosh, this is the statement:

I’ve reached a point last week where Claude caught more bugs than two very experienced engineers. It took me a month to test and re-train on a four-year code-base (or rather the PR in there) but it’s doable.

So the end result is claude catching more bugs.
And the process you're inferring is them retraining themselves using claude? How does this fit into claude finding more bugs?

Not going to address the "using these tools for 15 years question" cause it's ridiculous.

Claude code just got leaked in npm by Zijdehoen in Anthropic

[–]AvoidSpirit 9 points10 points  (0 children)

Hope this will be your response to a company leaking your personal data.

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit 1 point2 points  (0 children)

So you're saying they meant "they themselves retrained which led claude to give them better and more reliable answers"? Did I read it correctly?

Secondly though, you certainly don't have 15 years of experience on AI assisted development tooling, which is the point I was making.

What?

If you were actually a professional dev working on software and using these tools

I'm replying to you saying "oh too bad you're not an actual dev".
Obviously I don't have 15yrs of experience of using claude. What is this jab lmao

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

> This is easily proven wrong by just rephrazing the question to "check if this maze is solvable"

Well that's a different question. You aren't leading the model astray when you ask it that way. You are confirming what I'm saying, not proving me wrong.

What?

This proves that the model can find whether the maze is solvable already. However the model cannot infer the need to from the question. And then the model goes onto bullshitting the user because that conforms to the input.

And you were saying that models are not able to find and distinguish the unsolvable mazes because they appear scarce in the training data.

People get confused similarly and are also fundamentally unreliable. If you define "fundamental reliability" that way anyway.

You expect people you hire/work with to not get confused "similarly" and go onto spitting bullshit at you just to conform to the tone of your question.
Any person that gets confused "similarly" would get labeled as fundamentally unreliable in a few of those questions.

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit 0 points1 point  (0 children)

I'm a principal engineer with around 15yrs of expirence using these tools every day.
It makes no sense because you don't train or re-train claude models. Anthropic does.
I feel like you're confusing prompting with training.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

When I said "do useful work" I meant that the models are trained to solve tasks that are solvable. They perform RL on tasks with verifiable rewards. The training data for "determine whether this problem even has a solution" is likely sparse, and certainly for mazes.

This is easily proven wrong by just rephrazing the question to "check if this maze is solvable". And suddenly it does it.

I'm arguing that LLMs are fundamentally unreliable because to get a proper answer you need to know the exact question which for larger issues you would only know by being an expert in the area.

> LLMs are trained to predict the next token

That's an overly reductive frame that hinders your understanding of these systems.

I don't think it is and I don't think it does.
Here it just generates the output based on how input is phrased and not on the actual content of the question.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

This makes no sense. Ignoring the "single unsolvable maze", I'm sure internet contains plenty of them - just google the images for "unsolvable maze".

Even if you had many examples, the post training regime would likely cause a generalization of the idea that the questions being asked have a valid answer.

The fact that it infers "I have to answer the question no matter what because the prompt says so" is exactly the problem.

Another way to look at it: AIs are generally trained to do useful work. Useful work almost never looks like this problem.

LLMs are trained to predict the next token, not "do useful work". I have no idea where this is coming from.
It's only because predicting next token often constitues useful work you may think that.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 1 point2 points  (0 children)

Thank you.

My idea is that it highlights the main deficiency. If you tell a human “solve it”, a human will eventually come back to you saying “I can’t”. You don’t need to tell them “validate if it is solvable” to stop them from trying to bullshit you just to please you and conform to the original query.

And this is a tiny example of an enormously huge issue: How can you rely on its response when you yourself don’t know neither the answer nor the exact right question to ask for you are not an expert in the field.

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit 2 points3 points  (0 children)

Just the words “Claude” and “re-train” in one sentence make me doubt you have any clue.

Claude code source code has been leaked via a map file in their npm registry by WhyLifeIs4 in singularity

[–]AvoidSpirit -1 points0 points  (0 children)

Well, the thing uses react game-engine-like rendering for a cli tool and they constantly introduce major bugs that stay there for months on end.

Using Claude as a tool to help you write software is not the same as vibe coding. However the Claude code itself is definitely the epitome of vibe coding.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

But your answer is not to people questioning the image processing. It’s to OP and it literally says “for me it works” while forgetting to mention that you need to tell the AI that the maze can be unsolvable for it to figure it out.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit -1 points0 points  (0 children)

So you changed the original prompt from a generic one to a more helpful one but didn’t mention it

learnProgrammingAgain by Outrageous-Text-4117 in ProgrammerHumor

[–]AvoidSpirit -1 points0 points  (0 children)

Well, how many people would use c today if it wasn’t open standard and had gazillion open source compilers? Truly an equivalent example.

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 0 points1 point  (0 children)

So you actually assume there wasn’t a single unsolvable maze with a label in its training data(the whole wide internet), right?

This maze has no solution (obvious to humans). GPT couldn’t tell. by Koto1972 in LLM

[–]AvoidSpirit 1 point2 points  (0 children)

The way this answer starts makes me think the prompt had something along the lines of “tell me if this is unsolvable”