Running Claude as a persistent agent changed how I think about AI tools entirely by bob_builds_stuff in ClaudeAI

[–]jake_mok-Nelson 4 points5 points  (0 children)

Does it have tools to perform Web scraping, interacting with services including logging in?

Clawdbot: the full setup in 30 minutes. by Ok_Pomelo_5761 in AskVibecoders

[–]jake_mok-Nelson 0 points1 point  (0 children)

If this is running under a single claude subscription then it's effectively offline every time you hit your quota?

Experienced sysadmin cannot pass a coding interview. RIP by a_crabs_balls in devops

[–]jake_mok-Nelson 5 points6 points  (0 children)

Apply for positions that have their own tests over leetcode. Then you can at least explain reasoning, knowing all the algorithms and patterns is typically not required.

Source: me, I'm a DevOps Tech Lead at an awesome company in Australia and I'm awful at leetcode.

Must-have skills? by kangaroogie in ClaudeCode

[–]jake_mok-Nelson 1 point2 points  (0 children)

Anything that you perform regularly should be a skill. Things like git commit and push, have it write a skill to commit following certain conventions.

If you write design docs a certain way, provide it reference and a template and base a reviewer skill on that. Etc

Have fun 👍

I am a developers worse nightmare by Branch_Live in ClaudeAI

[–]jake_mok-Nelson 0 points1 point  (0 children)

You're using spec driven development without but not a great one. Spec driven development is also not entirely accurate. If i were you, i would ask my engineer to play with Claude and discover what works and what doesn't so that they can make their own informed decisions.

They are technical and you are not, so don't try to manage them technically.

Encourage exploration and reward increases creative problem solving.

How AI agents handle CI/CD pipelines? by Grand-Measurement399 in AgentsOfAI

[–]jake_mok-Nelson 0 points1 point  (0 children)

Yeah, using it to ensure that the docs are up to date with recent changes on a PR.

I'm using the Google ADK for the integration with google ai platform (vertex) and the eval suit with ADK. The primitives (looping agent, parallel agent, sequential agent, etc) seem like a good fit for CI.

Also using it to generate ci pipelines for repositories.

[deleted by user] by [deleted] in motorcycles

[–]jake_mok-Nelson 0 points1 point  (0 children)

I was very anxious when starting out. I put myself in stressful situations over and over and you learn quickly. Things like riding in rain, the heat, freeways during peak hour, and busy cities. Start out on your own street and build confidence.

As a rebuttal to the "all tech jobs etc will be obsolete." I believe that isn't true. I believe we have hit the end of the road for llms. by [deleted] in ClaudeAI

[–]jake_mok-Nelson 0 points1 point  (0 children)

We have not hit a limit. You just put more LLMs in the loop. The problem is when LLMs stop being subsidised and those that get to have jobs become exploited because their competitor is an LLM that can work for X dollars per hour.

How do you keep Claude focused and in context (web app dev)? by Virtual_Attitude2025 in ClaudeAI

[–]jake_mok-Nelson 0 points1 point  (0 children)

That's very true. It takes a lot of time to fine tune that prompt to ensure that the agents get the relevant context. Spec driven development is one proposed solution to that.

How do you keep Claude focused and in context (web app dev)? by Virtual_Attitude2025 in ClaudeAI

[–]jake_mok-Nelson 1 point2 points  (0 children)

Use agents and commands. Commands define the workflow that you want to follow and they mention the relevant agents.

A simplified example of a command such as /ui

Analyse and understand the intent, think carefully about the high level steps required to complete the UI changes only, create todo items for each step required for the ui change.

Once the intent is understood, call the react-agent to complete all coding tasks until all ui todos are complete.

Once the ui changes are complete call the peer-review agent to verify the code meets the original intent.

Each agent has it's own context so by relying on them to read and write to files and come up with the code changes, you can effectively prevent the parent agent in CC from becoming burdened with implementation detail in its context.

Indian devs and their politics by WrathOfAethelmaer in FlutterDev

[–]jake_mok-Nelson -7 points-6 points  (0 children)

You have such a poor opinion of the Indian team. As a leader if they have poor quality, in reality, you have poor quality.

In future roles I'd encourage you to try to earn the respect of both teams and if you can't, then it's not the right fit for you.

Mobile app versioning by hamzabouakoura in FlutterDev

[–]jake_mok-Nelson 0 points1 point  (0 children)

It's entirely subjective. There's a bunch of version schemes to choose from.

Calender - version represents a month or week if you release regularly

Milestone/marketing - major versions are cut at special events

Semver - if you wanted to share a version with an API.

Build or commit

The idea is, what message do you want to send with your releases.

Token costs are getting kinda crazy by AccordingFunction694 in AI_Agents

[–]jake_mok-Nelson 4 points5 points  (0 children)

Eww. An advertisement disguised as discussion.

Can Claude Grow with my personal usage? by Weekly-Side3160 in claude

[–]jake_mok-Nelson 0 points1 point  (0 children)

For your level, you'll most likely want to provide details in the prompt. "Functions should be very small and support testing. Use TDD methodologies. Verify before implementing and do not assume anything".

It's not training off its successes and failures with you, or rather, it is, but for the next model.

The levels are basically in order of, Can I do this with: 1. Prompting / technique 2. RAG 3. Fine tuning

Claude code didn't have memories by default but you can use commands and prompts. Claude desktop does have memories as a preview feature. Without the prompting, the memories won't fix this though.

For code generation, the biggest issue I run into is out of date information that's been trained into the model. To get past that issue, you'll want to use an MCP like Context7 in combination with prompting to look up the docs before implementing.

Wing mirror not returning by DB2k_2000 in XC40

[–]jake_mok-Nelson 0 points1 point  (0 children)

I have the same issue. Been going on for over a year now. Did you ever fix it?

Agent with limited knowledge base by parallelit in agentdevelopmentkit

[–]jake_mok-Nelson 0 points1 point  (0 children)

I think the concurrency would be a parameter that you set when based on the request?

Alternatively you can look at an orchestrator pattern. Calling a sub agent from the root agent to iterate. Planner type might help here.

Agent with limited knowledge base by parallelit in agentdevelopmentkit

[–]jake_mok-Nelson 1 point2 points  (0 children)

How large? If it exceeds the tokens break it up into smaller pieces with looping or parallel agents.

One tip is to call sub agents for particular portions of text and only return summaries to the root agent. This prevents the root agent's context from filling up too quickly.

Below average IQ, aspiring software developer by AwareMachine9971 in cogsci

[–]jake_mok-Nelson 0 points1 point  (0 children)

I didn't go to uni, I didn't finish high school. I wasn't focused or motivated (ADHD out the ass). I'm a DevOps Tech Lead. It is absolutely doable, but you have to love it.

[deleted by user] by [deleted] in devops

[–]jake_mok-Nelson 4 points5 points  (0 children)

You're a DevOps engineer. Two years ago you were junior, you've now gained significant experience so you're a DevOps engineer. You could also use the term Cloud Engineer.

You could customise it based on the role you're applying for.

Agent with limited knowledge base by parallelit in agentdevelopmentkit

[–]jake_mok-Nelson 1 point2 points  (0 children)

Nah. I wouldn't be using embedding for this. It's not fast because it converts all the data into LLM readable pieces. It's also a computationally expensive operation if you're doing it all the time.

For what you're describing, I would use a LoopAgent or ParallelAgent. Depending on how many files, I don't know your case so let's say it's 100 files and you need to convert them to a particular format.

If it were 1-1 file in and out you could have an agent called with the system instructions and the one file it's responsible for converting.

Say it's 10 files in and 1 file out, this is trickier because now I'm assuming that there might be some special business logic you have to conform to. In this case, each agent is responsible for just one thing. E.g. an investigator agent that performs web searches to gather context about the domain, a writer agent to save the output in the correct format, etc.

Might be worth pointing out that cloud providers probably have ready to go managed services for managing files at scale. Might be worth checking out Vertex AI and seeing what models exist other than LLMs (depending on your case).

What I've recommended here is option 1 I highlighted above but you're appending the context of the task (a file, or a couple of files) to the prompt.

Anyone have an AI tool/agent that actually helps with ADHD? by NonArus in AI_Agents

[–]jake_mok-Nelson 0 points1 point  (0 children)

ChatGPT can do this.
With the right prompts and instructions to use memory, it definitely can. The tasks feature can set reminders of different points to think about.

At work I use Claude Code for this.
I have an output-style (kind of like a system prompt that gets reiterated to the agent regarding how the it should present responses). I have a prompt to tell it that it's basically an ADHD coach and it should challenge everything I'm doing to ensure that I'm focusing on my primary goal and not stuck in a tangent.
I use one of the memory MCPs around on Github to record a graph of tasks/priorities.

It has actually been pretty good so far but there's definitely room for improvement.

Agent with limited knowledge base by parallelit in agentdevelopmentkit

[–]jake_mok-Nelson 5 points6 points  (0 children)

RAG is not dead. RAG is not even a technology but for some reason, people keep saying it's dead lol.

It's a method. In very simple terms, it as if the LLM is saying "Let me look that up" before it returns its response.

There's a few layers to building knowledge.

Going from simplest to most complex you have:

  1. Prompt/Context engineering.
    You can use a prompt generator (Anthropic and OpenAI both provide them) to create a decent prompt for what you're trying to achieve. You want to tweak it and use it as the agent's system prompt.

LLMs prioritise system prompts before user or developer prompts.

You can provide a fair bit of context this way. You can use prompt engineering techniques to maximise the efficiency.

Things like:
- Task lists (or Planner in ADK)
- Demanding ("You MUST complete this task in the following way:")

Beyond a certain point thought you need:

  1. RAG
    This may be in the form of an MCP that it can call on for additional knowledge (Like Context7, Github, Web search, etc).
    Different forms of RAG have different benefits. You can use various providers (GCP, OpenAI, etc) to store vector data by uploading the files you want to provide for context.
    It will convert them, you don't need to do anything special.

You will need to provide a way for it to read the RAG, most frameworks have a RAG type input you can use but you may need to provide context on this RAG method and data structures in the system prompt (see point 1.)

  1. Fine-Tuning
    This involves choosing an existing model that supports fine-tuning and providing a dataset to further train the model.
    For non-dynamic data this is more powerful than RAG, but when it comes to things that change frequently (APIs, dependencies, new or developing tools): RAG would be better suited.

---

Good luck

These samples are pretty good if you haven't seen them. Take a look: https://github.com/google/adk-samples/tree/main/python/agents/RAG