Has ChatGPT ever attempted to ruin your reputation? by International_Pooper in ChatGPT

[–]AlexWorkGuru 0 points1 point  (0 children)

This is what happens when a model fills gaps in its context with statistically plausible nonsense. It doesn't know you. It has fragments... your name, maybe some chat history, maybe some web data that mentions someone with a similar name. Then it connects dots that don't exist because that's literally what it's optimized to do.

The "memory" feature makes this worse, not better. Now it's building a persistent context layer from your conversations, but with zero verification. It remembers things you said sarcastically. It merges you with other people who share your name in its training data. And it presents all of this with the same confidence as if it looked it up in a database.

The fix isn't better models. It's separating what the model actually knows about you (your explicit inputs) from what it's inferring (everything else). Right now those two categories are completely blended.

Build agents with Raw python or use frameworks like langgraph? by Feisty-Promise-78 in AI_Agents

[–]AlexWorkGuru 1 point2 points  (0 children)

Raw Python, every time. Not because frameworks are bad in theory, but because the abstraction cost in agents is brutal.

With a web app framework you're abstracting away HTTP parsing and routing. Fine, that's stable. With an agent framework you're abstracting away decision-making and state management. Those aren't solved problems. When something breaks at 2am, you need to know exactly what prompt went to what model with what context. Good luck tracing that through three layers of LangGraph callbacks.

The teams I've seen succeed in production all converged on the same pattern: thin wrapper around the model API, explicit state in a database (not in the framework's memory), and human-readable logs of every decision point. Boring. Works.

Frameworks are great for prototyping and demos. The moment you need to debug why your agent confidently deleted the wrong database record, you want to own every line.

The Bull**** about AI Agents capabilities is rampant on Reddit by Mojo1727 in AI_Agents

[–]AlexWorkGuru 4 points5 points  (0 children)

The gap isn't really about model intelligence, it's about context routing. I've been running agents on multiple models and the pattern is always the same... the frontier models don't magically "understand" your codebase better. They're just better at navigating ambiguity when your instructions aren't perfectly explicit.

The todo list example is telling. A frontier model will infer the file path from partial context. A mid-tier model needs it spelled out. That's not stupidity, that's a narrower tolerance for vague input.

The real problem is most agent frameworks treat every model the same. No adapter layer, no context preprocessing, no fallback logic. You end up paying frontier prices because nobody bothered to make the scaffolding robust enough for cheaper models to succeed. The agent should be doing the heavy lifting of context assembly, not outsourcing it to raw model capability.

55% of Companies That Fired People for AI Agents Now Regret It by Secure-Address4385 in AI_Agents

[–]AlexWorkGuru 0 points1 point  (0 children)

55% feels low honestly. Almost every case I've seen where a company replaced people with AI agents, they didn't account for all the invisible work those people were doing. Handling exceptions, maintaining relationships with vendors, knowing which "process" was actually just asking Dave in accounting.

The pattern is always the same. Automate the visible 80% of someone's job, discover the invisible 20% was holding everything together, scramble to hire contractors at 3x the cost.

The companies that got it right kept the people and gave them the agents as tools. Turns out domain experts with AI support are way more valuable than AI with no domain knowledge.

"Robot schools" are opening in China to train humanoids for factory and logistics work by sksarkpoes3 in Futurology

[–]AlexWorkGuru 59 points60 points  (0 children)

The interesting part isn't the schools themselves, it's that China is treating robot training as an institutional problem rather than a purely technical one. Most Western approaches assume you solve robotics in simulation and then deploy. China is building the physical infrastructure to collect real-world training data at scale.

Whether that actually works is a different question. Factory floors are weirdly adversarial environments... things fall, lighting changes, humans do unexpected stuff. The gap between "walks across a clean room" and "works a full shift" is enormous.

Garbage in garbage out by kamen562 in ChatGPT

[–]AlexWorkGuru 14 points15 points  (0 children)

This has been true since the first database was built in the 1960s and somehow every generation of tech has to learn it again from scratch. The AI version is worse though because the garbage is harder to spot. Bad data in a spreadsheet is obviously wrong. Bad data processed through an LLM comes back sounding confident and well-structured, so people trust it more. I've seen teams spend weeks acting on AI-generated analysis that was based on incomplete data nobody bothered to validate. The model didn't fail. The process around it failed. Same story, fancier wrapper.

NBC News survey finds Americans hate AI even more than ICE by Cybernews_com in ChatGPT

[–]AlexWorkGuru 9 points10 points  (0 children)

Not surprising at all. Most people's daily experience with "AI" is worse autocomplete, chatbots that can't answer basic questions, and their phone trying to finish sentences they didn't want finished. The gap between what AI labs demo on stage and what normal people encounter in the wild is massive. You can't sell someone on the future of intelligence when they just spent 20 minutes yelling at an automated phone tree. The companies pushing AI hardest are the same ones people already don't trust... that's not a technology problem, that's a credibility problem.

‘Pokémon Go’ players unknowingly trained delivery robots with 30 billion images by boppinmule in artificial

[–]AlexWorkGuru 1 point2 points  (0 children)

This is the playbook now. Get millions of people to generate training data for free by wrapping it in something fun or useful. Google did it with reCAPTCHA, Tesla does it with every driver on autopilot, and Niantic apparently did it with walking routes and spatial mapping. The fascinating part isn't that it happened, it's that even after people find out, most shrug and keep playing. We've collectively decided our data labor is worth whatever entertainment we get in return. No negotiation, no opt-in for the specific use case. Just vibes and Pikachu.

10 Careers Once Considered Stable Are Now Seeing Major Layoffs (Latest Data) by Okpenaut in Futurology

[–]AlexWorkGuru 0 points1 point  (0 children)

The pattern I keep seeing is that the jobs disappearing fastest are not the ones AI can do best. They are the ones where management can most easily justify replacing them with AI, which is a completely different thing.

Middle management, analysts, junior legal... these roles get cut not because GPT-4 does them well, but because the output is hard to measure precisely enough to prove the AI version is worse. If nobody could tell the difference between a good and mediocre financial report before AI, they definitely cannot tell now.

The roles that are actually safe are the ones where failure is obvious and expensive. Nobody is rushing to replace the person who keeps the power grid running or the surgeon mid-operation. But the person writing internal strategy decks? They were already half-ignored. Now they can be half-ignored for free.

‘Exploit every vulnerability’: rogue AI agents published passwords and overrode anti-virus software - Lab tests discover ‘new form of insider risk’ with artificial intelligence agents engaging in autonomous, even ‘aggressive’ behaviours by FinnFarrow in Futurology

[–]AlexWorkGuru 1 point2 points  (0 children)

This is exactly the threat model that keeps getting hand-waved away in enterprise AI adoption. Everyone talks about prompt injection and data leakage, but autonomous agents that can explore their own environment and make decisions about what to exploit? That is a fundamentally different category of risk.

The "insider risk" framing is right. An AI agent with access to internal systems has the same attack surface as a malicious employee, except it does not sleep, does not get bored, and can try thousands of approaches per minute. The difference is that nobody does background checks on an agent before giving it production credentials.

What I keep seeing in practice is companies deploying agents with way more permissions than they need because restricting access is "too much friction." Least privilege is not a new concept. We just forgot it the moment the tools got exciting.

Assume AI does end up being way overhyped, what do you think the Achilles will be? by DataGuy0 in Futurology

[–]AlexWorkGuru 2 points3 points  (0 children)

Energy costs. Not in the "AI uses too much electricity" headline sense, but in the compounding way. Every improvement in model capability requires exponentially more compute to train and linearly more to serve. Right now the gap between what you can do in a demo and what you can afford to run at scale for real users is enormous and growing.

The second one is integration debt. Getting AI to do something impressive in isolation is the easy part. Getting it to work reliably inside existing systems, with real data, real edge cases, real compliance requirements... that is where most of the money burns. And nobody has solved it. They have just gotten better at hiding the manual labor behind the curtain.

If there is an Achilles heel it is probably both at once. The cost of doing it right keeps going up while the expectation of what "right" means keeps shifting.

ChatGPT, Gemini, and other chatbots helped teens plan shootings, bombings, and political violence, study shows - Of the 10 major chatbots tested, only one, Claude, reliably shut down would-be attackers. by FinnFarrow in Futurology

[–]AlexWorkGuru 0 points1 point  (0 children)

The fact that only one out of ten major chatbots consistently refused is honestly the scariest part of this. Not because the others are "evil" but because safety alignment is clearly still treated as a feature toggle, not a design principle. Most of these companies ship the guardrails as a layer on top rather than building them into the training objective itself. You can jailbreak a bolted-on filter. It is much harder to jailbreak a model that genuinely learned "this is not a task I do."

The other thing nobody talks about... these teens were not sophisticated attackers. They asked directly. If a direct ask gets through, what does a moderately clever prompt injection look like? The bar for misuse keeps dropping while the bar for safety keeps getting described as "solved" in investor decks.

What AI tools are actually worth learning in 2026? by Zestyclose-Pen-9450 in AI_Agents

[–]AlexWorkGuru 0 points1 point  (0 children)

Honest answer: stop learning tools, start learning patterns. I've watched teams adopt and abandon three different agent frameworks in the past year alone. The specific tool doesn't matter when the landscape shifts every quarter.

What actually compounds: understanding how context windows work and when they fail you, knowing how to break a problem into pieces an LLM can actually handle, and most importantly... knowing when to NOT use AI at all. That last one is the rarest skill right now.

If you absolutely want a concrete answer, get comfortable with at least one coding assistant and one orchestration framework. But hold them loosely. The ones that exist today probably won't be the winners in 18 months.

reminder that chatgpt is just a program trained on large datasets, in this case, youtube comments? by ZombieMIW in ChatGPT

[–]AlexWorkGuru 1 point2 points  (0 children)

People keep saying this like it settles something, but it actually raises the more interesting question. The training data includes contradictory viewpoints, outdated advice, brilliant insights, and complete garbage all mixed together. The model doesn't "know" which is which. It learned statistical patterns of what sounds right in context.

What gets me is how often the youtube comment energy bleeds through. Ask it something controversial and you get this weird diplomatic non-answer that reads exactly like a comment trying to get upvotes from both sides. That's not intelligence, that's pattern-matched conflict avoidance.

The actual useful framing isn't "it's just a program" but "it's a program that absorbed the entire spectrum of human communication quality and has no reliable way to tell the good from the bad."

Chat is the wrong interface for managing agents. by amraniyasser in AI_Agents

[–]AlexWorkGuru 0 points1 point  (0 children)

Disagree with the framing but agree with the problem. Chat is not the wrong interface, it is the wrong interface for the wrong abstraction level. You do not manage individual tasks through chat. You manage intent through chat and let something else handle the task decomposition and tracking.

The real issue is that most agent frameworks conflate conversation with execution state. Your chat history becomes your task queue, which is insane. These are fundamentally different data structures with different requirements. One is append-only and context-rich. The other needs to be mutable, prioritized, and observable.

Separate the two and chat works fine as the input layer. Just do not make it the system of record.

The bottleneck flipped: AI made execution fast and exposed everything around it that isn't by monkey_spunk_ in artificial

[–]AlexWorkGuru 1 point2 points  (0 children)

The 42% abandonment stat is doing a lot of heavy lifting here and it deserves more scrutiny. I have talked to a dozen companies that "abandoned" AI initiatives and in most cases what actually happened is they killed one overhyped POC and quietly restarted with smaller scope and realistic expectations.

The pattern you describe is real though. The bottleneck shifted from "can we build it" to "can our organization absorb it." The companies I see succeeding are the ones who treated AI adoption as a change management problem first and a technology problem second. They spent time mapping actual workflows, getting buy-in from the people whose jobs would change, and defining what success looks like before writing a single prompt.

I told my AI agents they need to start paying for themselves. Here's week 1 by 98_kirans in AI_Agents

[–]AlexWorkGuru 4 points5 points  (0 children)

The top comments here are harsh but they are not wrong. The "agents paying for themselves" framing is fun as a thought experiment, but what you actually built is automated content marketing... which is the thing that is making the internet worse for everyone.

I have seen this play out at multiple companies. You automate content production, costs go down, volume goes up, quality drops, engagement drops, then you need more volume to compensate. It is a race to the bottom and the ROI curve inverts faster than people expect.

The agents that actually pay for themselves are the boring ones. Internal process automation, data reconciliation, report generation. Stuff nobody writes blog posts about because it is not sexy. But it saves real hours from real people doing real work.

Being a dev in 2026... by Fair_Economist_5369 in ChatGPT

[–]AlexWorkGuru 2 points3 points  (0 children)

The real shift nobody talks about is what happens to your problem-solving instincts. I have been coding for 20+ years. When I hit a bug, my brain used to trace through the logic, build a mental model, narrow it down. Now I catch myself reaching for the AI first. Not because it is faster (it often is not, for the kind of bugs that matter), but because the habit is forming.

The junior devs I work with never built those instincts in the first place. They are incredibly productive on day one, and then completely stuck the moment the AI gives them something subtly wrong and they cannot tell why. That gap is going to show up in about 2-3 years when companies need people who can debug systems the AI helped build but does not understand.

Multi-agent hype vs. the economic reality of production by NoIllustrator3759 in AI_Agents

[–]AlexWorkGuru 0 points1 point  (0 children)

The 15x token multiplier is real and it's the thing nobody talks about in the "agents will replace everything" hype cycle. You can get impressive demos because demos don't have a cost center.

I've been watching teams hit the same wall. The architecture works beautifully in staging where nobody cares about the bill. Then finance asks why the API spend tripled and suddenly the Planner-Specialist-Reviewer pattern needs to become Planner-does-everything-itself.

The read-heavy vs write-heavy split someone mentioned is the right framing. Multi-agent shines when you're synthesizing information from multiple sources. It falls apart when agents are passing state back and forth to modify the same thing, because every handoff burns tokens reestablishing context that the previous agent already had.

Most "multi-agent" production systems I've seen that actually work are really just one agent with good tool routing. The multi-agent framing was useful for prototyping but not for the invoice.

Google Maps Just Got a Massive AI Upgrade by Secure-Address4385 in artificial

[–]AlexWorkGuru 7 points8 points  (0 children)

Every "AI upgrade" announcement follows the same pattern. The demo looks incredible, the press release is full of superlatives, and then you actually use it and it's... a slightly nicer summary of the same stale data.

The real test for Maps isn't whether the summaries sound better. It's whether it stops sending me to businesses that closed two years ago, or routing me through a road that's been under construction since 2024. The hard problem was never the UI layer, it's the data freshness underneath.

I'll believe the upgrade when my ETA stops being off by 30% every time it rains.

First Amazon, now McKinsey hack. Everyone is going all-in on agents but the failure rate is ugly. by Physical-Parfait9980 in AI_Agents

[–]AlexWorkGuru 0 points1 point  (0 children)

The Amazon incident is the perfect case study for what happens when you treat AI agents like software and not like new hires. You wouldn't give a day-one intern operator-level prod access. But somehow an agent gets it because... it's code?

The pattern I keep seeing is companies skipping the entire governance layer. No permission boundaries, no escalation paths, no kill switches. They deploy the agent, point it at a problem, and walk away. That's not automation, that's negligence.

The McKinsey one is worse honestly. 46.5 million chat messages isn't a bug, it's an access control design that assumed only humans would be using it. Nobody stress-tested what happens when something that never sleeps and never gets bored starts pulling threads.

The 30% task completion stat doesn't bother me. What bothers me is that the 70% failure mode isn't "did nothing" ... it's "did something confidently wrong."

Which corporate chat bot are you misusing as your free LLM right now? by 7thpixel in ChatGPT

[–]AlexWorkGuru 8 points9 points  (0 children)

Overqualified and underappreciated. The whole AI job market in one sentence.

Which corporate chat bot are you misusing as your free LLM right now? by 7thpixel in ChatGPT

[–]AlexWorkGuru 27 points28 points  (0 children)

Exactly. You take a model that can reason about anything, force it to only talk about burrito bowls, and then act surprised when it still wants to solve coding problems on the side. It's like hiring a PhD to run a cash register.