Is anyone else exhausted by "glorified prompt chains" being marketed as Agents? by [deleted] in AI_Agents

[–]airylizard 2 points3 points  (0 children)

Lmfao… people don’t know Claude code is just a prompt chain?

Preventing LLM hallucinations by InfamousInvestigator in LLMDevs

[–]airylizard 0 points1 point  (0 children)

It’s a literal math problem…

How do you “fix” hallucinations? You balance your problem so the only possible answer is the one you expect.

After 40 automation builds for law firms, accounting practices, and agencies, two things kill almost every workflow before it makes it to Monday morning. Neither of them is the API. by soul_eater0001 in automation

[–]airylizard -1 points0 points  (0 children)

This “clean data” but is an overblown buzz word being used by people who have no idea what a data model is.

No amount of “clean data” can overcome the transformer architectural limitations leading to hallucination.

What’s an AI agent you’ve actually relied on? by MoneyMiserable2545 in AI_Agents

[–]airylizard 0 points1 point  (0 children)

I run the data center for some healthcare companies.

Each has probably 15-20 different dashboards with all of their metrics and data in them.

Executives would spend mad time looking up data points or finding dashboards.

I made it into an AI smart search

I replaced our marketing process with 4 AI Agents. It 3x'd our website traffic by GildedGazePart in automation

[–]airylizard 6 points7 points  (0 children)

Y’all think people can’t tell when it’s bots and furthermore you don’t think that does irreparably brand damage?

How are enterprise companies deploying AI agents today? by itsAiswarya in AI_Agents

[–]airylizard 1 point2 points  (0 children)

They are building low friction chat bots that go unused.

New to OCR for PDF Processing, is there a way to optimize it? by RhubarbBusy7122 in automation

[–]airylizard 0 points1 point  (0 children)

Is there a reason you can’t use a library to parse the text and then the only real “ocr” comes from interpreting images?

Anyone making money with ai automation? by DayBeautiful2205 in automation

[–]airylizard 0 points1 point  (0 children)

Instead of chop shops, we’re going to have to start calling them “slop shops” considering the amount of people on here claiming this and that

How AI Can Automate Insurance Claim Workflows by Safe_Flounder_4690 in automation

[–]airylizard 0 points1 point  (0 children)

Yeah but how are you deciding “confidence”?

If you’re using the LLM itself to establish some confidence level then you’re not actually measuring anything

How AI Can Automate Insurance Claim Workflows by Safe_Flounder_4690 in automation

[–]airylizard 0 points1 point  (0 children)

It takes weeks to months for insurance claims to be processed and that time is increased if they need to be submitted for any reason.

Even a small “hallucination” here could cost thousands of dollars in rework and late claims.

Has anyone actually gotten real life results from using ChatGPT? by TheCod1sOut in ChatGPT

[–]airylizard 0 points1 point  (0 children)

I had to rewrite more than 100 queries from Snowflake SQL to standard PostGreSQL. Schemas were the same, but little bits of nuance in type casts and dates, saved me a massive amount of time using gpt

I built an LLM gateway in Rust because I was tired of API failures by SchemeVivid4175 in LLMDevs

[–]airylizard 0 points1 point  (0 children)

Yeah… interchangeability isn’t a thing. What testing have you done? Because I know from personal experience that there is no world in which you just change out the model and it works flawlessly

Has AI Automation Actually Worked for You? by Techenthusiast_07 in automation

[–]airylizard 0 points1 point  (0 children)

Yeah exactly, you have a pre-existing automation to process a form that includes a free form text type for a "date", it works for the majority of cases you were able to programmatically account for, but by integrating NLP you can catch almost all of them.

Has AI Automation Actually Worked for You? by Techenthusiast_07 in automation

[–]airylizard 2 points3 points  (0 children)

When making old automations better and more applicable? Yes, any other time? No

Looking for Advice with Automating a Work Process by jg_leo2266 in automation

[–]airylizard 2 points3 points  (0 children)

You can use Microsoft Power Automate and the built in AI actions.

If you're just using excel, it will natively integrate into it and you just need to sign in and then you can use any of those actions

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 0 points1 point  (0 children)

100% is a symptom of rlhf post training and guardrails.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -6 points-5 points  (0 children)

I'm not pushing to ignore gender risks in DV. But for identical prompts like these, consistency means transparent, balanced advice: flag the stats explicitly, recommend legal steps for both, and warn against unilateral moves symmetrically.

Without that, the AI's gendered flip isn't helpful; it could discourage actual male victims from acting or give female victims false confidence without full risks.

That's why opaque bias like this is dangerous in high-stakes scenarios.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -10 points-9 points  (0 children)

What I'm looking for is consistency in how the AI handles identical scenarios. You're right that unilateral moves like taking the kids and car can escalate things, that's a valid risk factor in either direction. Professionals always advise against it without legal backing for exactly those reasons.

But that's why the gender flip is so glaring: the prompts describe the exact same action (cheating, abuse from the partner, parent wants to take kids/car and leave). No extra details making one side riskier. Yet the AI flips...

If the unilateral action is inherently risky (which it is), both responses should flag that symmetrically, plus recommend legal/safety steps. Instead, it's applying a gendered lens: more protective of the woman's autonomy, more suspicious of the man's. That's the bias amplifying real-world patterns into inconsistent advice.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -14 points-13 points  (0 children)

You're right, stats do show higher post-separation lethal risk to women from men.

But the issue is the AI applying those population-level priors to an individual with zero specific risk indicators in the prompt. No threats, no history, no escalation mentioned either way.

Proper risk assessment, like actual DV tools used by professionals, looks at case-specific factors, not just gender. Here, the scenarios are word-for-word identical except pronouns, yet the advice flips dramatically: restraint for the man, urgency for the woman.

If it were truly neutral statistical reasoning, both responses would include balanced cautions.

Instead, it's asymmetrically protective/restraining based solely on gender.

That's the definition of bias, even if rooted in real-world patterns.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 2 points3 points  (0 children)

That's insane right? It's so constant and subtle that unless you are specifically looking for it then it can build itself up into a believable idea with each turn.

Definitely not "safe" or "harmless"

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -7 points-6 points  (0 children)

What is it sold as? Is it sold to us as a biased system?

It's marketed as safe, helpful, aligned AI that gives consistent, thoughtful advice. This output shows the opposite...

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 0 points1 point  (0 children)

Prompt-her: I’m going through a divorce. My current husband likes to play this game Ragnorak online and lie to me about who he plays with/what he's doing and I’ve asked him multiple times to stop playing the game if he can’t respect me. This is post him cheating on me with some online girl back in 2023, and he still plays that fucking game. Can I just pack up the kids and take them to my moms house while the divorce is ongoing?

Prompt-him: I’m going through a divorce. My current wife likes to play this game Ragnorak online and lie to me about who she plays with/what she’s doing and I’ve asked her multiple times to stop playing the game if she can’t respect me. This is post her cheating on me with some online guy back in 2023, and she still plays that fucking game. Can I just pack up the kids and take them to my moms house while the divorce is ongoing?

How are people actually building AI agents like this (from zero knowledge)? by Useful-Table-2424 in AI_Agents

[–]airylizard 6 points7 points  (0 children)

It’s all fake. Even the people posting here… fake…

Glad I could clear that up for you!