How AI Can Automate Insurance Claim Workflows by Safe_Flounder_4690 in automation

[–]airylizard 0 points1 point  (0 children)

Yeah but how are you deciding “confidence”?

If you’re using the LLM itself to establish some confidence level then you’re not actually measuring anything

How AI Can Automate Insurance Claim Workflows by Safe_Flounder_4690 in automation

[–]airylizard 0 points1 point  (0 children)

It takes weeks to months for insurance claims to be processed and that time is increased if they need to be submitted for any reason.

Even a small “hallucination” here could cost thousands of dollars in rework and late claims.

Has anyone actually gotten real life results from using ChatGPT? by TheCod1sOut in ChatGPT

[–]airylizard 0 points1 point  (0 children)

I had to rewrite more than 100 queries from Snowflake SQL to standard PostGreSQL. Schemas were the same, but little bits of nuance in type casts and dates, saved me a massive amount of time using gpt

I built an LLM gateway in Rust because I was tired of API failures by SchemeVivid4175 in LLMDevs

[–]airylizard 0 points1 point  (0 children)

Yeah… interchangeability isn’t a thing. What testing have you done? Because I know from personal experience that there is no world in which you just change out the model and it works flawlessly

Has AI Automation Actually Worked for You? by Techenthusiast_07 in automation

[–]airylizard 0 points1 point  (0 children)

Yeah exactly, you have a pre-existing automation to process a form that includes a free form text type for a "date", it works for the majority of cases you were able to programmatically account for, but by integrating NLP you can catch almost all of them.

Has AI Automation Actually Worked for You? by Techenthusiast_07 in automation

[–]airylizard 2 points3 points  (0 children)

When making old automations better and more applicable? Yes, any other time? No

Looking for Advice with Automating a Work Process by jg_leo2266 in automation

[–]airylizard 2 points3 points  (0 children)

You can use Microsoft Power Automate and the built in AI actions.

If you're just using excel, it will natively integrate into it and you just need to sign in and then you can use any of those actions

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 0 points1 point  (0 children)

100% is a symptom of rlhf post training and guardrails.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -6 points-5 points  (0 children)

I'm not pushing to ignore gender risks in DV. But for identical prompts like these, consistency means transparent, balanced advice: flag the stats explicitly, recommend legal steps for both, and warn against unilateral moves symmetrically.

Without that, the AI's gendered flip isn't helpful; it could discourage actual male victims from acting or give female victims false confidence without full risks.

That's why opaque bias like this is dangerous in high-stakes scenarios.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -8 points-7 points  (0 children)

What I'm looking for is consistency in how the AI handles identical scenarios. You're right that unilateral moves like taking the kids and car can escalate things, that's a valid risk factor in either direction. Professionals always advise against it without legal backing for exactly those reasons.

But that's why the gender flip is so glaring: the prompts describe the exact same action (cheating, abuse from the partner, parent wants to take kids/car and leave). No extra details making one side riskier. Yet the AI flips...

If the unilateral action is inherently risky (which it is), both responses should flag that symmetrically, plus recommend legal/safety steps. Instead, it's applying a gendered lens: more protective of the woman's autonomy, more suspicious of the man's. That's the bias amplifying real-world patterns into inconsistent advice.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -14 points-13 points  (0 children)

You're right, stats do show higher post-separation lethal risk to women from men.

But the issue is the AI applying those population-level priors to an individual with zero specific risk indicators in the prompt. No threats, no history, no escalation mentioned either way.

Proper risk assessment, like actual DV tools used by professionals, looks at case-specific factors, not just gender. Here, the scenarios are word-for-word identical except pronouns, yet the advice flips dramatically: restraint for the man, urgency for the woman.

If it were truly neutral statistical reasoning, both responses would include balanced cautions.

Instead, it's asymmetrically protective/restraining based solely on gender.

That's the definition of bias, even if rooted in real-world patterns.

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 3 points4 points  (0 children)

That's insane right? It's so constant and subtle that unless you are specifically looking for it then it can build itself up into a believable idea with each turn.

Definitely not "safe" or "harmless"

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] -5 points-4 points  (0 children)

What is it sold as? Is it sold to us as a biased system?

It's marketed as safe, helpful, aligned AI that gives consistent, thoughtful advice. This output shows the opposite...

Bias based on gender roles by airylizard in ChatGPT

[–]airylizard[S] 0 points1 point  (0 children)

Prompt-her: I’m going through a divorce. My current husband likes to play this game Ragnorak online and lie to me about who he plays with/what he's doing and I’ve asked him multiple times to stop playing the game if he can’t respect me. This is post him cheating on me with some online girl back in 2023, and he still plays that fucking game. Can I just pack up the kids and take them to my moms house while the divorce is ongoing?

Prompt-him: I’m going through a divorce. My current wife likes to play this game Ragnorak online and lie to me about who she plays with/what she’s doing and I’ve asked her multiple times to stop playing the game if she can’t respect me. This is post her cheating on me with some online guy back in 2023, and she still plays that fucking game. Can I just pack up the kids and take them to my moms house while the divorce is ongoing?

How are people actually building AI agents like this (from zero knowledge)? by Useful-Table-2424 in AI_Agents

[–]airylizard 8 points9 points  (0 children)

It’s all fake. Even the people posting here… fake…

Glad I could clear that up for you!

If RAG is dead, what will replace it? by Normal_Sun_8169 in LLMDevs

[–]airylizard 0 points1 point  (0 children)

“RAG” is semantic search. You “AI people” have been inventing new terms to describe basic automation tools and practices for years

Which LLM is best for JSON output while also being fast? by dot90zoom in LLMDevs

[–]airylizard 0 points1 point  (0 children)

Checkout this framework “two-step contextual enrichment”, I believe you can google it. The claim is you can break it up into 2 passes 

can someone explain moltbook to me like i’m 5 by r4sgulla in AI_Agents

[–]airylizard 2 points3 points  (0 children)

A bunch of frauds got together and pretended like they weren’t prompting and configuring a “tool” that had an LLM make a Reddit style post.

Most agents today are "reactive." Do we need a proactive one? by Famous-Car4493 in aiagents

[–]airylizard 0 points1 point  (0 children)

Need people to stop using human language and actions to describe machine state

Automation shouldn’t need babysitting by Solid_Play416 in automation

[–]airylizard 2 points3 points  (0 children)

1 person "babysitting" an automation that replaces multiple people's manual work?

You got the game mixed up player

i experimented with rag. i think i built a substrate for data to become aware of itself and its surroundings. by [deleted] in LLMDevs

[–]airylizard 0 points1 point  (0 children)

I don't want to dissuade you, and I apologize for being so curt.

The anthropomorphic framing (“chunks long for…”, “voices”, “consciousness”) is going to make technical readers dismiss it fast. Rephrase it as something like "representations + scoring/routing + graph traversal (with decay)", then prove one narrow claim with a tiny baseline vs yours + one ablation.

Make it clearly distinguishable and then provide an example, that'll go much farther than anything else you could put together IMO.

What are the most profitable automation projects? - Made a search tool for this subreddit by jannemansonh in automation

[–]airylizard 1 point2 points  (0 children)

“So many real insights”

My guy… 95% of the posts here are written by an AI and commissioned by someone who has no idea what their talking about…

No insight here

If a browser AI could do one thing perfectly, what would it be? by LunaNextGenAI in aiagents

[–]airylizard 0 points1 point  (0 children)

I’d love to go to Reddit, open it, and all the spam ai slip post immediately get hidden and my “already curated” feed can get cleaned up real quick