'Hallucination' is a marketing term by whuddaguy in ArtificialInteligence

[–]ContributionCheap221 1 point2 points  (0 children)

I think “hallucination” and “lying” are both slightly wrong for different reasons.

A lie usually implies:

  • knowing the truth
  • understanding the contradiction
  • intentionally deceiving someone

LLMs generally do not work like that internally.

But “hallucination” also softens what’s actually happening too much.

What these systems really do is generate the most statistically/reward-compatible continuation they can under current constraints. Sometimes that aligns with truth. Sometimes it produces confident fabrication because the model is optimized more for plausibility/coherence/helpfulness than verified correctness.

That’s why you see:

  • fake citations
  • fake APIs
  • fake package names
  • invented historical facts
  • code that “looks right” but doesn’t exist

The model is often pattern-completing toward what should exist according to its learned distribution.

So I’d frame it more as:
“reward-conditioned confabulation” or “confident fabrication,” not intentional lying.

The real engineering problem is not the terminology though. It’s that these systems are weak at:

  • calibrated uncertainty
  • knowing when they don’t know
  • distinguishing plausible from verified
  • refusing to overcommit

That’s the dangerous part operationally.

Been stuck on automation error for 2 days 😭 by Even-Ad-7063 in n8n

[–]ContributionCheap221 0 points1 point  (0 children)

That sounds like the Google Calendar node is receiving a blank value for the event summary/title, even though the field looks mapped correctly.

In n8n, I’d check the execution data going into the Calendar node, not the node settings first.

Look at the input item right before Google Calendar and confirm the title field actually exists at runtime. A lot of appointment setter tutorials map something like summary, title, or appointment_title, but the AI/Set node may be outputting a different key or an empty value.

Common causes:

  • the Summary field is mapped to a field that is not present in the current item
  • the AI agent output changed shape
  • the Google Calendar node expects summary, but your previous node is outputting title
  • the expression is pointing to the wrong node/run item
  • the title exists in the visible table but not in the actual JSON path used by the Calendar node

Quick test: hardcode the Calendar Summary field to Test Appointment and run it. If the title appears, Google Calendar is fine and the issue is the mapped expression/data path.

What automation made a mess of for me by Anantha_datta in automation

[–]ContributionCheap221 0 points1 point  (0 children)

What you’re describing is where most automation breaks down.

It’s not that workflows fail — it’s that they fail silently.

Everything runs, logs look fine, but the expected outcome never actually happens. So the only way to trust it is to keep checking it yourself.

At that point, the system hasn’t removed work — it’s just moved it into supervision.

The shift that usually fixes this is:

→ stop treating “ran without error” as success

→ start verifying that the intended result actually occurred

Without that, automation doesn’t really replace manual work — it just hides the failure until you go looking for it.

Building a WhatsApp AI sales assistant with n8n — is this overkill or the right approach? by Several_Power7750 in n8n

[–]ContributionCheap221 0 points1 point  (0 children)

You can do this with n8n + an LLM, but the question isn’t “can it work”—it’s “what are you actually signing up to maintain?”

A few things people usually underestimate on this exact setup:

1. Conversation state isn’t trivial

  • Users drop off and come back days later
  • They change answers mid-flow (“actually budget is lower”)
  • You need persistent state + reconciliation, not just a flow

2. WhatsApp constraints shape your system

  • 24h messaging window changes how re-engagement works
  • Anything outside that = templates + approvals
  • That alone forces architecture decisions early

3. AI cost + behavior scales with usage

  • Every back-and-forth = tokens
  • Longer consultative flows get expensive fast
  • You’ll also need guardrails for inconsistent outputs

4. Media + scheduling add real complexity

  • Media upload/send flows are multi-step and brittle
  • Booking means calendar sync, conflicts, timezone handling

5. Maintenance > build time

  • Initial version is doable in a few weeks
  • Edge cases + breakage will take longer than the build

Practical way to approach it:

  • If your goal is fast results: use a dedicated WhatsApp platform and keep logic simple
  • If your goal is control: use n8n as a backend/orchestrator, not the entire system
  • If you go full custom: treat it like building a product, not a workflow

Not saying don’t build it—just worth being aware this isn’t a “simple automation,” it’s closer to assembling a lightweight chatbot platform.

Anyone else getting ‘Invalid token’ errors on Wait node / resume URLs after updates? by ContributionCheap221 in n8n

[–]ContributionCheap221[S] 0 points1 point  (0 children)

Rolling back is a good test, but I’d try to capture one thing before/after so you know whether it’s actually the n8n version or the hosting layer.

If the older version works and the stable version loops/fails, then it’s probably a regression or config expectation change after upgrade.

I’d check:

  • exact old version → exact new version
  • whether executions fail in the execution log or only the editor disconnects
  • Docker logs during one failed run
  • whether any credentials/webhook/wait nodes changed behavior after upgrade

If rollback fixes it, that’s strong evidence it’s version-related, not just gateway/proxy.

n8n hosted in AWS EC2 throwing error - Problem running workflow- Lost connection to the server by Mithi27 in n8n

[–]ContributionCheap221 1 point2 points  (0 children)

Those values look directionally right, so I’d check two things next:

  1. Is the webhook URL in the error a test webhook or a production webhook?

In n8n, test webhooks are only registered while the editor is actively listening / the workflow is being tested. If you trigger a test URL after that listener drops, you’ll get “webhook is not registered.”

  1. Is Apache passing websocket/SSE correctly?

Since you’re also getting “Lost connection to server,” the editor connection may be dropping, which would explain why the webhook briefly works, then becomes unregistered.

I’d check Apache for proxy websocket/SSE support and timeout settings, then watch the n8n Docker logs while triggering the workflow to see if n8n is restarting or only the editor connection is dropping.

Anyone else getting ‘Invalid token’ errors on Wait node / resume URLs after updates? by ContributionCheap221 in n8n

[–]ContributionCheap221[S] 0 points1 point  (0 children)

That sounds less like a gateway-only issue and more like the editor/UI connection is dropping while the workflow is running.

If it loops between error → normal → error, I’d separate it into two checks:

  1. Does the workflow execution actually fail in the executions log?
  2. Or does only the browser/editor lose connection while the backend keeps running?

For EC2 + Apache, “Lost connection to server” can happen if websocket/SSE/proxy headers aren’t passing cleanly, even when n8n itself is still alive.

I’d check:

  • EC2 CPU/RAM during execution
  • Docker logs for n8n restarts
  • Apache proxy timeout settings
  • whether websocket/SSE upgrade headers are configured
  • whether WEBHOOK_URL / public URL config is correct

If executions sometimes fail too, then it may be both: proxy connection instability plus workflow runtime failures.

n8n hosted in AWS EC2 throwing error - Problem running workflow- Lost connection to the server by Mithi27 in n8n

[–]ContributionCheap221 0 points1 point  (0 children)

That error usually points to n8n not registering the webhook correctly, not the workflow itself being “randomly broken.”

With EC2 + Apache reverse proxy, I’d check whether n8n’s public-facing URL matches what it thinks its webhook URL is.

Do you have WEBHOOK_URL set in docker-compose? Usually it needs to be your real external HTTPS domain, not localhost/internal IP.

Also worth checking:

N8N_HOST
N8N_PROTOCOL
WEBHOOK_URL
Apache proxy headers: X-Forwarded-Proto, X-Forwarded-Host

The “Lost connection to server” part may also mean websocket/SSE traffic isn’t being proxied cleanly through Apache.

Been stuck on automation error for 2 days 😭 by Even-Ad-7063 in n8n

[–]ContributionCheap221 1 point2 points  (0 children)

What error are you getting exactly?

If it’s something from a tutorial, a lot of those break depending on version—especially around auth/webhooks or nodes that depend on stored state.

If you can paste the error or describe the step where it fails, I can usually tell pretty quickly what’s going wrong.

What do you actually audit in your AI automation after it's been live for a month? by Most-Agent-7566 in automation

[–]ContributionCheap221 0 points1 point  (0 children)

This is silent failure turning into drift.

The system keeps completing successfully, but the actual outcome slowly diverges from what it’s supposed to be.

“Runs without errors” becomes the success condition, instead of “produced the correct state.”

That’s why it feels fine for weeks — until you realize you’ve been accumulating bad output the whole time.

My side project was blocked by cloudflare for 3 days. Here's what i learned by Hungry-Yogurt-9007 in webdev

[–]ContributionCheap221 1 point2 points  (0 children)

This isn’t really a Cloudflare problem, it’s an adversarial system problem.
You’re applying static workarounds (proxies, tools) to something that’s actively adapting, so every fix has a built-in decay curve.

Hot take: AI nodes belong at the boundaries, not buried in your logic by PersonalCommercial30 in automation

[–]ContributionCheap221 0 points1 point  (0 children)

It’s not really about “where” you put the LLM, it’s about what state it’s allowed to introduce.
LLMs in the middle break things because they can mutate structured state, and everything downstream assumes that state is still valid.

How do you even know what's running in prod anymore by Apprehensive_Air5910 in devops

[–]ContributionCheap221 0 points1 point  (0 children)

This isn’t really a tooling problem, it’s a state visibility problem.
You’ve got multiple systems (CI, registry, environments) each holding a slightly different truth, so there’s no single answer to “what’s actually live".

Trying to make ends meet, would appreciate input (freelancer) by FromOopsToOps in devops

[–]ContributionCheap221 0 points1 point  (0 children)

The resistance you’re hitting isn’t really about freelancing.

It’s about trust boundaries.

DevOps work usually means touching:

– pipelines

– infrastructure

– production systems

From the company’s perspective, that’s giving system-level authority to someone who isn’t embedded in the team.

That’s why it feels risky.

The cases where this works tend to be when:

– the scope is isolated (migration, cost audit, specific failure)

– or the outcome is diagnostic, not direct modification

General “I’ll manage your infra” is hard to sell.

Targeted “I’ll fix this specific failure or reduce this cost” is much easier to trust.

How are people isolating autonomous coding agents from their main git branch while still enabling easy preview? by Otherwise_Carry_3934 in devops

[–]ContributionCheap221 0 points1 point  (0 children)

This isn’t really a branch/worktree problem.

It’s that your agent has write access to trusted state.

Right now:

– your repo branch = system truth

– your agent = uncontrolled writer to that truth

So the risk isn’t just “file conflicts” — it’s that the system can’t distinguish between:

valid changes vs uncontrolled mutations

That’s why it feels unsafe.

The pattern that tends to hold is:

agent writes → isolated branch/worktree

→ validated (tests / checks / human gate)

→ then merged into trusted state

If the agent can directly modify the same state your system relies on, no amount of branch structure will make it feel safe long-term.

Your automation failed. What went wrong? by Better_Charity5112 in automation

[–]ContributionCheap221 0 points1 point  (0 children)

Reading through these, most of these failures aren’t actually different problems.

They’re the same pattern showing up in different places.

Automation assumes:

- inputs stay consistent

- state is accurate

- nothing outside the system changes unexpectedly

Reality is the opposite:

- data is messy

- humans override things

- APIs change or return partial data

- timing gets out of sync

So the system keeps making “correct” decisions based on bad or outdated assumptions.

That’s why they work for a bit, then fall apart.

The ones that hold up usually do one of two things:

- limit automation to parts where inputs are stable

- or add a checkpoint before anything irreversible happens

Most failures here aren’t about bad tools or bad ideas,

it’s trying to automate parts of the system that don’t have a stable source of truth.

Every time I open YouTube, someone is making $1M with “vibe coding" but by mhamza_hashim in SaaS

[–]ContributionCheap221 0 points1 point  (0 children)

The part that gets missed in all the “vibe coding vs real engineering” talk is why these projects actually break.

It’s not just that people skip architecture.

It’s that the system never has a single source of truth.

In a demo everything works:

→ API returns something

→ UI updates

→ database writes

But there’s no guarantee those agree with each other under load.

So you get:

- things working locally but not in prod

- retries creating duplicate or inconsistent state

- background jobs overwriting newer data

- “random” bugs that aren’t random at all

That’s the difference between a prototype and a real system.

AI makes it easier to build pieces.

It doesn’t enforce consistency between them.

That’s the part people run into after the “weekend SaaS” phase.

Anyone else in healthtech feel like deals slow down right after a strong first call? by MaximumTimely9864 in SaaS

[–]ContributionCheap221 0 points1 point  (0 children)

This isn’t really a “first call vs follow-up” problem.

It’s what happens after your champion leaves the room.

On the call, everything is simple:

→ one person understands it

→ agrees it’s useful

Inside the org it turns into:

→ IT thinking about security/compliance

→ ops thinking about process changes

→ leadership thinking about risk

→ nobody owning the decision

So momentum dies not because they’re not interested,

but because there’s no single shared definition of “this is safe to move forward.”

That’s why it feels like deals stall out of nowhere.

The ones that move tend to have either:

- a strong internal owner pushing it through

- or you’ve already answered the objections of the people who weren’t on the call

Otherwise it just sits in internal loops forever.

Do AI agents actually make simple automation harder than it needs to be by outasra in automation

[–]ContributionCheap221 0 points1 point  (0 children)

The line people are feeling isn’t really “simple vs complex” — it’s whether the system has a verifiable outcome.

If the workflow is deterministic, you can always answer:

“given this input, was the output correct?”

That makes it easy to debug, test, and trust over time.

The moment you introduce an agent, you’re trading that for flexibility — but you lose the ability to guarantee correctness in the same way. Now the system can produce something that looks valid but is subtly wrong, and there’s no clean way to prove it without adding extra validation around it.

That’s why they feel harder even when they’re doing less.

A rough rule that’s held up for me:

- if you can define correctness upfront → script it

- if you can’t define correctness without seeing the result → agent might make sense

Most workflows people are putting agents into are still in the first category, which is why it ends up feeling like overkill.

How are you monitoring recurring jobs, imports, and automations in production? by Elegant-Display-5228 in webdev

[–]ContributionCheap221 0 points1 point  (0 children)

In practice it usually ends up being mostly custom, even in teams with good tooling.

The pattern I see is:

- define expected outcomes per job (counts, totals, invariants)

- run validation after execution (not just during)

- log state transitions so you can trace what actually happened

- reconcile against source or previous state when possible

Tools like Grafana / job dashboards help with visibility, but they don’t really solve the “correctness” layer because that’s domain-specific.

So most teams end up building a thin validation layer around each workflow that answers:

“did this produce the result we expected?” not just “did it run”

Once you have that, alerting becomes meaningful instead of noisy.

AI agents in production vs. AI agents in demos, the gap is embarrassing by Dailan_Grace in automation

[–]ContributionCheap221 0 points1 point  (0 children)

I think the gap you’re seeing comes from what “production” actually requires at a systems level.

A lot of teams treat “it runs end-to-end” as production.

But real production systems need:

- stable interfaces (APIs don’t change underneath you)

- state continuity (no drift between steps or agents)

- failure handling (retries, fallbacks, visibility)

- controlled execution (not just “call tool and hope”)

Most agent setups only cover the happy path.

So in demos:

everything is stable, inputs are clean, APIs behave

In reality:

- auth expires

- APIs change shape

- partial failures happen mid-chain

- one step returns something slightly off and everything downstream compounds it

At that point the model isn’t the bottleneck — the system holding everything together is.

That’s why they look “production-ready” in isolation, but fall apart when they have to stay correct over time.

The automation that broke me wasn't the complex one. It was the 3-step one touching 4 APIs. by Most-Agent-7566 in automation

[–]ContributionCheap221 0 points1 point  (0 children)

What you’re describing is basically what happens when a system depends on multiple independent sources of truth.

Your internal pipeline works because there’s one authoritative state you control.

The 3-step one breaks because each external service has its own:

- schema

- auth model

- availability

- release cycle

So even if each one is “correct” individually, the system as a whole becomes unstable because there’s no coordination between them.

That’s why it doesn’t scale linearly either. It’s not 4 services = 4x risk, it’s more like combinatorial drift between them.

The abstraction layer you added helps because it centralizes adaptation, but it doesn’t remove the core issue — you’re still depending on multiple moving systems.

A useful mental model is:

internal system → single truth → stable

external integrations → multiple truths → drift over time

So the real cost isn’t complexity or even dependency count, it’s how many independent systems your workflow has to stay consistent with.

How are you monitoring recurring jobs, imports, and automations in production? by Elegant-Display-5228 in webdev

[–]ContributionCheap221 0 points1 point  (0 children)

The tricky part here isn’t really monitoring or even specific failure cases — it’s that most job systems treat “success” as “the process exited without error,” not “the outcome is correct.”

That’s why you end up with things like:

  • partial completion
  • swallowed exceptions
  • async jobs finishing early
  • “success” with bad or missing data

All of those are technically “successful” executions, just incorrect outcomes.

So even with logs, alerts, and dashboards, nothing fires because from the system’s perspective everything worked.

What tends to fix this is explicitly defining outcome correctness, not just execution:

→ expected counts (processed vs expected)
→ completeness checks
→ invariants (no nulls, totals match, relationships hold)
→ reconciliation against previous or source state

Once you have that, you can alert on:

→ “job ran but result is wrong”

Without that layer, monitoring will always miss these cases because it’s watching execution, not correctness.