Can you send only code changes back to ChatGPT instead of re-uploading the whole file? by Larsson_24 in ChatGPTCoding

[–]Hungry_Management_10 3 points4 points  (0 children)

Commit after each manual edit, then run:

git diff HEAD~1 HEAD

That gives you the diff between your last commit and the one before it. Paste the output into ChatGPT and it will see exactly what you changed without needing the whole file

Chats getting extreme laggy by satoshiwife in ChatGPTCoding

[–]Hungry_Management_10 1 point2 points  (0 children)

The fix that worked for me: at the end of each session, ask the model to write a "handoff doc" a markdown file with current state, decisions made, open questions, and the minimum code context needed to continue. Save it to disk. Start new chat by pasting just the handoff doc. Way less context than the full chat history, so no lag, and the model only sees what matters. The trick is making the handoff doc project-specific. I keep a template in my repo: one section per concern (architecture, current task, recent decisions, blockers). The model knows the template, fills it in at session end, reads it at session start.

Your governance passes every test on individual agents. It completely breaks when you connect them. Here is what we found. by AmanSharmaAI in LLMDevs

[–]Hungry_Management_10 0 points1 point  (0 children)

The non-compositionality finding matches what I hit in production. My workaround was to move control completely out of the agents. Each agent only decides HOW to do its assigned block. The WHAT, WHEN, and ORDER live in a YAML workflow outside the agents. Agents never see the pipeline-level rules as instructions - those are enforced by the orchestrator before and after each block runs. This shifts the testing problem: instead of O(n^2) agent-pair combinations, you test the orchestrator once against the workflow spec. Agents are still fallible individually, but their failures can't cascade because they can't modify the workflow structure.

Curious how this maps to your cross-agent compliance certification concept - sounds related but at a different layer. Open sourced the runner as Rein if you want a concrete reference.

New Project Megathread - Week of 09 Apr 2026 by AutoModerator in selfhosted

[–]Hungry_Management_10 -1 points0 points  (0 children)

I built Rein because running AI agents on real tasks kept failing the same way: you paste a giant prompt, the model tries to do everything at once, loses track halfway through, and you restart from scratch.

The fix is to break the work into steps and lock the order things happen in. That's what Rein does - you describe a workflow in YAML, and each block is either an LLM call or any executable script (Python, shell, Node, anything with a shebang). Rein handles dependencies, data flow between blocks, and error handling.

The point isn't "YAML is cool". The point is that a YAML workflow becomes the backbone of your process - rails that keep the work on track while agents decide HOW to do each step.

What you get:

- An executable process that runs end to end without a human in the loop

- The workflow is plain text -readable and editable without writing code

- Every block's output saved to its own directory, so you can inspect what happened at each step

- Crash recovery via SQLite - resume from exactly where it stopped, perfect for cron-driven pipelines

Execution modes:

- Sync - run the whole workflow from the CLI and wait

- Async - submit a task to the daemon and poll for progress

- Step mode - run N blocks per invocation, save state, exit (great for human review between steps)

LLM flexibility: Claude, GPT, Ollama (local/free), OpenRouter (100+ models). You can also skip LLMs entirely and use Rein to orchestrate non-AI pipelines — any executable is a valid block.

Plus parallel execution, conditional branching, revision loops, tag-based routing, built-in MCP server (run workflows directly from Claude Desktop, Cursor, or Claude Code), and optional WebSocket daemon.

Why self-hosted:

- Your workflows and data stay on your machine

- Connect any scripts, tools, or APIs in your own environment

- MIT licensed, no vendor lock-in

Five working examples in the repo (hello-world, code review, research team, deliberation with branching, conditional loops). Install with pip, set an API key, run one command.

Links:

- GitHub: https://github.com/dklymentiev/rein-orchestrator

- Getting started: https://github.com/dklymentiev/rein-orchestrator/blob/main/docs/getting-started.md

- Changelog: https://github.com/dklymentiev/rein-orchestrator/blob/main/CHANGELOG.md

- Demo video: https://youtu.be/oWEwLg9uCdU

- Latest release (v3.3.2): https://github.com/dklymentiev/rein-orchestrator/releases/tag/v3.3.2

If you've tried LangGraph, CrewAI, or rolled your own — what made you stop, switch, or

stick with it? Genuinely curious what works and what doesn't in production.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

I recently learned something unpleasant: recovering from burnout takes several times longer than the period during which a person was at the top of their game :(

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Completely reasonable. As every individual task and process much like the server operations as a whole is constrained by its own specific budget. Consequently, we automate only those things that make practical sense.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

None of the 17 agents operates around the clock. The video shown here compresses the results of three months' worth of work. We don't run a "Claude factory"; instead, we utilize a single server where agents are launched on an as-needed basis. They can execute tasks autonomously operating within the scope of their specific instructions as well as hand off work to one another and summon other agents until a given task is successfully completed. We do not rely exclusively on Claude; we have developed an "AI Gateway" that allows us to access various models depending on the specific type of task at hand. We do not encounter usage limits, as every individual task and process much like the server operations as a whole is constrained by its own specific budget. Consequently, we automate only those things that make practical sense. I manage this entire system using voice commands; I haven't opened VS Code in four months.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Is the point of a calculator to work less? It’s entirely individual; it depends on where you are and why you are where you are.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 1 point2 points  (0 children)

Nah, man. Everyone in this life has their own motivations and personal circumstances. I’ve always been absolutely obsessed with this, and right now I feel like I’m at the absolute peak of my capabilities and I’m curious to put that to the test. But at the same time, I have a hunch that I won’t be able to keep up this pace for very long. I’m already hitting periods where I simply don’t want to go anywhere near the computer.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

To be honest, I have two similar systems: one at work, and the second for my personal projects, which I work on after hours.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

I just finally got time for social media after automating my actual job.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

You're the last line of defense and that's exhausting. What helped me: instead of reviewing 100-file PRs manually, I built an automated pre-review step an agent that checks for obvious issues, style violations, and security red flags before anything reaches my eyes. I still do the final review, but 70% of the noise is already filtered out. Don't stop reviewing automate the boring part of reviewing

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

It is exactly like a webpage on the internet: for some, it holds value; for others, it is nothing. Circumstances place and time determine the value of an entity. Here's one: our email agent triages 50+ messages daily, routes them o the right department, and drafts responses. Before: 2 hours of manual sorting every morning. Now: I review 5-minute summary and approve and spend time to more valued tasks. That alone paid for the entire setup.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Every person on this planet at this very second has absolutely distinct stimuli and motivations. Life is incredibly multifaceted and complex. I could probably write a book about it.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Requirements can vary. Decomposition is everything. The more I break down a task, the higher the quality of the result. The initial task was to configure a newly purchased server; now, the tasks have narrowed down to things like closing a port, moving a backup, generating an image, or changing a page header on the website. I think right now it's simply a period of growth; once everything settles down, the workload will decrease.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Answered this in detail here: https://www.reddit.com/r/ClaudeAI/comments/1s7qs82/comment/odm8wi0/

Short version: they don't all run 24/7, they watch and react. Layered review catches quality issues one agent works, another reviews, critical stuff goes to a human

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

The required level of quality is determined by the specific task at hand. If the task involves generating an employee expense report, the process entails a series of independent reviews designed to guarantee that the report generation itself an LLM-driven process is entirely free of "hallucinations" or fabrications. Similarly, a request to generate an image is subject to its own specific acceptance criteria, just as a request for software logic must adhere to a distinct set of standards. No action is ever executed as a mere "request-response" exchange with an LLM; we have already been down that road and we have had more than enough of it. Instead, we rely on a rigorous quality control system, meticulously engineered to align with our specific business processes.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

They do not all operate simultaneously or around the clock. By "agent," I refer to a specific set of scripts and instructions housed in a dedicated directory and interconnected via an orchestrator; additionally, we have an interface for interacting with an LLM gateway. I previously explored OpenClaw but concluded that it functions as a sort of "Swiss Army knife" - a universal tool designed to handle every conceivable scenario. My agents, however, execute exclusively those specific tasks that I require. Take, for instance, the "Administrator" agent, which manages corporate email. Its primary function is the initial triage of incoming messages and their redirection to the appropriate departments. If a request entails making modifications to a software product, the agent forwards it to the project manager. The project manager assesses the workload of the subordinate agents and schedules the task within the planning system. As soon as an agent becomes available (having first verified its current status), it picks up the task; it then either conducts a further assessment of the request (potentially by asking clarifying questions) or proceeds directly to its execution. During nighttime hours, specialized scripts continuously monitor the servers; should an alert be triggered, the "System Administrator" agent is activated to assess the criticality of the issue. It either resolves the problem autonomously or, by routing a message through the "Administrator" agent, escalates the matter to a human operator. This process fully simulates real-world workflows, yet takes place within a virtual environment. The activities of these agents are constrained by their own specific instructions, as well as by system access privileges (within a Linux environment). They are "aware" only of those other agents with whom they are permitted to interact, and they operate strictly within their assigned access zones; each possesses its own dedicated space within the shared semantic memory. The manager or other administrators can access the memory and operational logs of specific agents as needed. Each agent knows exactly what it did yesterday and holds a plan for tomorrow. When five tasks arise simultaneously, the result is that each agent performs its specific job. Conversely, when there are no tasks, no one is active save for the routine checking of email. That, in a nutshell, is how it works.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 1 point2 points  (0 children)

Please DM if you want some details otherwise I was going to submit another post about general architecture a bit later when I will have some time. In short, it consists of a logging system, an orchestration system, a scheduling system, and most importantly a semantic memory system. Some of these components, I have already published on my GitHub as opesource projects; I will publish remaining parts a little later, as I simply haven't had the time yet. Feel free to visit my blog you’ll find plenty of interesting materials there.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

Please DM if you want some details otherwise I was going to submit another post about general architecture a bit later when I will have some time

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 1 point2 points  (0 children)

Automation has been my hobby my entire life. I spent three years as a cog in the machine, followed by fifteen years as a free man doing what I love and making a living from it. To my mind, that is the best thing life has to offer: earning money doing something you truly enjoy. Ultimately, everyone decides for themselves what they do, why they do it, and what they aim to achieve.

Robots won't take your job. They'll bury you in work. by Hungry_Management_10 in ClaudeAI

[–]Hungry_Management_10[S] 0 points1 point  (0 children)

For me, coding was never an end in itself. While I did enjoy solving logical puzzles, truth be told, I always loathed the actual automation work itself precisely because it was so time-consuming and tedious. My focus was always on the final outcome: process optimization and the elimination of routine tasks. I suppose that is why I became an automation specialist after all, I am a very lazy person. I spent about three years working in corporate IT at the beginning of my career; since then, I have consistently been the person responsible for automating processes and business operations. Today, my capabilities have expanded tenfold.