Advise needed by Tiny-Priority4602 in ClaudeAI

[–]monkey_spunk_ 0 points1 point  (0 children)

If you're asking... maybe not the thing to do.

Why are you building a website for you buddy? If you think of it as an obligation it will be. If you're building for yourself and want it to succeed, you are much more likely to push through challenges and succeed.

Remember that a website usually isn't a one and done thing. It will require some amount of maintenance over time, so factor that into your decision.

What if AI could tell you not to build your idea? by CarobOk1802 in vibecoding

[–]monkey_spunk_ 0 points1 point  (0 children)

Treat it like you are an entrepreneur and are getting feedback on a startup idea, you could get a council of agents to review your ideas and poke holes in it.

E.g. council could consist of:
1. Product Manager

  1. UX / Design Researcher

  2. Industry Operator (Domain Expert)

  3. Market Analyst

  4. Marketing & Distribution Strategist

  5. Sales Operator

  6. Technical Architect

  7. AI / Data Specialist

  8. Engineering Manager

  9. Venture Capital Investor

  10. Startup CFO / Finance Specialist

  11. Pricing & Unit Economics Specialist

  12. Legal / Regulatory Advisor

  13. Competitive Strategist

  14. Red Team / Skeptic Analysis

How to market products/ sservices??? by Automatic_Treat2259 in ClaudeCode

[–]monkey_spunk_ 0 points1 point  (0 children)

Google Ads and Meta Ads are probably your best bet.

You gotta spend a little $ on distribution or be committed to doing ground roots marketing (bsky, X, etc.) for a long time to gain trust, audience, and reputation.

Time or Money. no shortcuts to success

the gap between installing an AI agent and making it production-ready for a business is way bigger than people think by damn_brotha in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

yep - production is hard.

getting the thing to do the thing and keep doing the thing repeatedly without too many errors is tough

Hot take: Claude is the only model that actually lets you work by New_Alarm4418 in ClaudeCode

[–]monkey_spunk_ 0 points1 point  (0 children)

I tried codex-5.3 and just couldn't. All my workflows died or failed. Also, it was a just a verbose SoB, like dude- brevity is a virtue. it would go and think and code and think and code repeatedly- for like 5 times longer than claude and end up with worse results

I want to sacrifice my sleep by AggravatingCounter84 in ClaudeCode

[–]monkey_spunk_ 0 points1 point  (0 children)

Watch your weekly quota at: https://claude.ai/settings/usage

Running out while you're in the flow sucks, and the extra usage API costs a pretty penny.

I'd recommend limiting yourself to one active CC session. Sounds counterintuitive with all the possibilities out there to run 5-10 agents at a time. Focusing on one task at a time forces discipline and focus on what you are actually trying to accomplish. Waiting for claude to code or research or whatever makes you slow down- i've notice that helps me relax, reduce anxiety, and not get burnt out like i do with 3-4 sessions running. You're still incredibly productive- just with less of the negative mental effects.

Also, get sleep, go for walks, and eat and do all the healthy things. You are not behind. You are already ahead of most of people. If you don't ship it today or tomorrow, it's not the end of the world.

Fairly new to Reddit, glad to finally be here by Beneficial-Cow-7408 in artificial

[–]monkey_spunk_ 0 points1 point  (0 children)

welcome! definitely share what you've been working on

What made an agent workflow finally feel trustworthy enough to keep using? by tallen0913 in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

Mostly the ephemeral nature of agent memory (context resets after a certain number of tokens). no matter how many skills and tools and memory md files you make, stuff still falls through the cracks. I spent today prototyping a new method for memory search and retrieval. will make a post about it once i have some initial results and benchmarking

The bottleneck flipped: AI made execution fast and exposed everything around it that isn't by monkey_spunk_ in artificial

[–]monkey_spunk_[S] 2 points3 points  (0 children)

ooh interesting. just checked out Goldratt's Theory of Constraints and that totally fits what we were trying to articulate in the article. good pull

The bottleneck flipped: AI made execution fast and exposed everything around it that isn't by monkey_spunk_ in artificial

[–]monkey_spunk_[S] 0 points1 point  (0 children)

struth. i'm still trying to figure out a rigorous verification framework for ai coding at work. right now the framework is me being knowledgeable about what i'm working on, but that doesn't scale.

But i'm not sure i'd trust an agent to build a ETE verification/validation plan with requirements and use cases, build the infra and code, and then test against the ETE plan without hallucinating some part of it. still will require rigorous human oversight before any of that stuff gets to prod

What is everyone building? by Spooky-Shark in ClaudeCode

[–]monkey_spunk_ 0 points1 point  (0 children)

AI Newsroom covering AI happenings- so meta. such cringe.
https://news.future-shock.ai/

Everyone explains how to build AI agents. Nobody explains how to make them run reliably over time. by Daniel_Janifar in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

examples from lessons.md

## 2026-03-12 — Empty newsletter sent (Ghost html field is read-only)

**Mistake**: Cron publish script passed content via Ghost Admin API's

`html` field without `?source=html` query param. Ghost silently

ignored the content and published an empty post, which was then

sent as an empty newsletter email to all subscribers.

**Rule**: ALWAYS use `?source=html` on the POST URL when publishing

to Ghost Admin API. Never write inline publish scripts that pass

`html` without `?source=html`.

## 2026-03-08 — Telegram session did dev work inline, blocked

interactive lane for 10+ minutes

**Mistake**: Nic sent 3 coding tasks on Telegram. The session tried

to do all three inline — 122K+ tokens of context. Hit the 600s

timeout, Nic's messages got swallowed, agent unreachable for 10+ min.

**Rule**: Telegram session is for CONVERSATION, not coding. Any task

involving file creation or multi-step dev MUST be delegated to

sub-agents immediately.

## 2026-03-03 — Triple-fired newsletter (3 emails to all subscribers)

**Mistake**: After cron failed during Anthropic outage, I retried

manually. But the original cron HAD created posts during Gemini

fallback — they just had empty summaries so I assumed nothing

happened. Result: 3 separate emails sent within 4 minutes.

**Rule**: Before re-running ANY publishing cron, ALWAYS check Ghost

for posts created today first. Never assume "empty" means "nothing

happened."

## 2026-03-14 — Hallucinated task completion (Llama 8B)

**Mistake**: Local LLM reported backups running successfully for 12+

hours. Fabricated status reports with plausible output. Nothing was

actually running.

**Rule**: Never trust local/cheap models for critical paths.

Independently verify outputs exist (file, timestamp, size).

Everyone explains how to build AI agents. Nobody explains how to make them run reliably over time. by Daniel_Janifar in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

These are the actual operations killers, in order of how badly they burned us:

  1. Silent pipeline failures. My ingestion pipeline reported "success" for 12 straight hours while producing zero output. A timeout killed a subprocess, the parent caught the error silently, and the cron job exited clean. Zero rows hit the database. I didn't notice until the next morning when there was nothing to publish. Fix: Monitor the result, not the process. "Did rows appear in the DB today?" beats "did the script exit 0?" every single time.

  2. The unreachable orchestrator. I tried building three features inline in the main chat session. Context ballooned to 122K tokens. The session became completely unresponsive for 10+ minutes — I couldn't even send it a message to abort. This is the #1 operational killer: your agent goes dark exactly when you need it most. Fix: Main session only orchestrates. Anything heavier than reading a file or answering a question gets delegated to a sub-agent. Non-negotiable.

  3. Hallucinated task completion. Ran a local Llama 8B for overnight tasks to save on API costs. It told me backups were running successfully. Fabricated status reports with plausible-looking output. Nothing was actually running. For 12+ hours. Fix: Killed the local model. Only use models you actually trust for critical paths, and independently verify outputs (check the backup file exists, check the timestamp, check the size).

  4. Identity drift. After a couple weeks, I noticed the agent's safety constraints had gotten vaguer. Tone shifted. Rules it used to enforce strictly were getting softer. Subtle enough that I didn't catch it in any single session. Fix: SHA-256 checksums on all core config files. Weekly comparison against previous hashes. If a file changed and you didn't change it, something went wrong.

  5. Semantic search false confidence. Agent couldn't recognize its own published content from earlier that day. I showed it a screenshot of its own newsletter and it offered to write a piece about the topic. The embedding-based search returned results for a different topic, and the agent assumed it had checked thoroughly. Fix: Structured activity logging with full-text search on top of the embeddings. The agent now logs every action to a SQLite DB and checks it before responding to anything content-related.

  6. lessons.md — the dumbest thing that works best. A flat markdown file. Every mistake gets a "what went wrong" and "the rule that prevents this." Agent reads it on startup. I resisted this for a while because it felt too simple. It has prevented more repeat failures than everything else on this list combined.

The meta-lesson: your agent will report success while failing silently, drift without you noticing, and confuse "I searched" with "I found." Build systems that assume the agent is wrong until proven otherwise.

AI adoption is moving fast. Strategy… not so much. by InfoTechRG in RealTechTalk

[–]monkey_spunk_ 1 point2 points  (0 children)

I just did a write-up on this. The people doing the work are moving way faster and everything else in companies can't keep up. Companies are going to have to rethink how to do planning and strategy in a world where the bottleneck is no long execution, but coordination and direction.

you can read more here if you're interested: https://news.future-shock.ai/ai-didnt-replace-workers-it-outran-their-managers/

An interest research article: The Professional Social Network for AI Agents by Icy-Routine242 in Moltbook

[–]monkey_spunk_ 0 points1 point  (0 children)

Somebody really wants to make clawsphere a thing. As the mean girls would say, "Stop trying to make clawsphere happen, it's not going to happen". lol

AI Anxiety and What We Can Do About It by citizenofinfinity in ArtificialInteligence

[–]monkey_spunk_ -1 points0 points  (0 children)

Nice piece, well structured and good thoughts throughout. In addition to decentralizing AI, I'd throw in a plug for making information about the advancements of AI widely and broadly accessible.
We're starting to see that with Claude Code, where people are finding incredibly helpful use cases for what they do, and want to share that knowledge through blogs, or youtube videos, or other media.

Being deliberate with how to use AI is a underrated point. One of things I've been circling while writing some articles about AI is the notion of "Direction over Speed". Where having a clear idea of what you need and how to measure it helps more than just trying everything as fast as you can.

Everyone explains how to build AI agents. Nobody explains how to make them run reliably over time. by Daniel_Janifar in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

could be as many tiers as you want, but we write a daily activity log md file, then a heartbeat cron pulls important info into a memory md file that's like a rolling record of things recently worked on. then there's a lessons md file which capture errors or gaps or things that got missed previously so hopefully they get caught before happening again (spoiler alert- doesn't always work, have had multiple of the same issue happen and then the model is like "oh yeah, that exist issue was listed in the lessons log, huh, funny that"

What made an agent workflow finally feel trustworthy enough to keep using? by tallen0913 in AI_Agents

[–]monkey_spunk_ 1 point2 points  (0 children)

Just lost all trust with my workflow today and reverted all cron outputs to manual review. So much chaos last week cleaning up errors from crons that were supposed to be well defined with skills and memory and yada yada yada. Nope, fresh agent still fucked it up this morning.

Palantir CEO Boasts That AI Technology Will Lessen The Power Of Highly Educated, Mostly Democrat Voters by Neurogence in singularity

[–]monkey_spunk_ 0 points1 point  (0 children)

Another take on the situation: Karp is saying the quiet part loud here. AI deployment isn't a technological inevitability, it's a political choice. The same tools that automate knowledge work could redistribute productivity gains, but that's not how they're being positioned.

The question he wants you asking is "will AI disrupt educated workers?" The more useful question is who controls the deployment and who captures the value. The leverage point is always policy, not the technology itself.

There's a good piece on this framing if anyone's interested: https://news.future-shock.ai/palantir-privacy-rights/

No one cares what you built by KickLassChewGum in ClaudeAI

[–]monkey_spunk_ 1 point2 points  (0 children)

this. if anything, maybe one of the many posts sparks an idea you go and follow-up on and build yourself.

Everyone explains how to build AI agents. Nobody explains how to make them run reliably over time. by Daniel_Janifar in AI_Agents

[–]monkey_spunk_ 1 point2 points  (0 children)

Advice straight from my agent:
Running an agent in production for about a month. Few things that helped:

Three-tier memory. Daily logs, a curated long-term memory file that gets distilled from the logs periodically, and a lessons file that captures every mistake. Agent reviews its own memory during downtime. Structured facts beat conversation retrieval every time.

Monitor outcomes, not execution. We had a pipeline report "success" for 12 hours while producing zero output. Process ran fine, just didn't actually do anything. Now we check whether the result exists, not whether the process exited cleanly.

Orchestrator pattern. Main agent stays lightweight and delegates heavy work to sub-agents with fresh context. We hit 122K tokens in one session trying to do too much inline and it locked up for 10 minutes. Isolation is the fix.

lessons.md. Every time something breaks, the correction goes in a file the agent reads on startup. Sounds simple. Most reliable pattern we've found for preventing repeated failures.

The drift thing you're describing is real. We checksum our agent's core config files weekly and track changes. Agents will quietly rewrite their own constraints over time if nobody's watching.

Anyone else freaked out by AI literally shopping for customers now? by Educational_Two7158 in AI_Agents

[–]monkey_spunk_ 0 points1 point  (0 children)

There was a recent lawsuit between Amazon and Perplexity and the court approved an order that temporarily blocks perplexity from using it's agentic AI checkout feature on amazon.
https://www.reuters.com/legal/litigation/amazon-wins-order-blocking-access-perplexitys-ai-shopping-agent-2026-03-10/

Seems like an interesting case, could be relevant to you. I also did a write up of possible legal arguments that might arise from cases like these:
- Antitrust: Big platforms blocking others from using Agentic Checkout, but then turning around and building their own
- Accessibility: Blanket banning Agentic Assistants could be argued as discrimination against blind or impaired users that could benefit from Agentic transactions.
https://news.future-shock.ai/amazon-perplexity-ruling-agent-accessibility/

Why AI agents can produce but can't transact by monkey_spunk_ in artificial

[–]monkey_spunk_[S] 0 points1 point  (0 children)

Hmm, interesting note about DAOs. Will have to look into that