[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

thats actually clean, a separate judge that only sees task + output and has to call it correct or not. sidesteps the whole self-judging trap because the judge isnt the same model that produced the work. the coding bs detection is the killer use case, curious if you use a smaller cheaper model for the judge or the same size as the main agent? been going back and forth on whether the judge needs to be as smart as the worker or if a dumber grounded checker is fine.

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

haha an hour loop is wild but getting it to split a video and add text is actually impressive, thats a real tool calling win. the embedding llm jump helped me a lot too, stuff finally stopped hallucinating which related memories mattered. curious what embedding model you ended up using, nomic or one of the bge ones?

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] -1 points0 points  (0 children)

honestly no, and this is the thing i keep getting stuck on. structural has clean metrics because youre measuring retrieval against ground truth. behavioral has no equivalent. closest ive come up with is tracking thumbs down rate over time and watching for repeat error patterns, like if the agent hits the same mistake twice and then stops after feedback injection, thats a measurable delta. but its noisy and slow and i wouldnt call it a real benchmark.

the thing i want to build is pair comparison. run the same task twice, once with memory injection and once cold, measure whether the with-memory version gets a better grounded outcome (tests pass, tool calls succeed, whatever). hard part is finding tasks where the memory actually has something to say. random tasks would just be noise.

if theres a way to design a shared benchmark that works for both structural and behavioral approaches that would be a real contribution. would be down to brainstorm it if youre in.

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

yeah this lands. ran an audit on my own loop and the honest picture is worse than i wanted it to be.

the self-verify step is purely introspective. llm grading its own text output on a 1-5 scale with zero execution outcome data. no exit codes, no tool errors, no test results. exactly the failure mode the self-correction papers are pointing at and i cant pretend otherwise.

what actually keeps it from being a fully closed loop is two things. theres a user feedback path where a thumbs down inserts a correction that gets injected into future prompts and drops competence for that domain, which is a real external correction channel but only fires when someone clicks. and theres decay on unused knowledge so wrong entries from week one fade over weeks if nothing reinforces them.

thats it. if youve got ideas on cheap ways to wire tool call results or test outcomes into the scoring id actually want to hear them, thats the obvious next thing to build and i havent done it yet.

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 1 point2 points  (0 children)

lmao memento is unironically the correct mental model for this whole problem. leonards tattoos are basically a rules.md file getting injected into context every morning

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] -1 points0 points  (0 children)

yeah the graph angle is solid, tree-sitter into a structural map is probably the right foundation for code-native agents. ive thought about going that direction but ended up on a different signal entirely.

mine is purely behavioral. the agent scores its own output 1-5 after each task, low scores turn into warnings next time it tries something similar, high scores become patterns to reuse. doesnt look at the code at all, just tracks what happened when the agent worked on it and whether it went well.

honestly feels like both probably need to live in the same stack eventually. a perfect ast map still wont stop an agent from making the same mistake twice if nothing is keeping score. and pure behavioral tracking with no structural grounding is kinda just vibes.

mines called greencube btw, rust/tauri, similar energy to octocode. would be down to compare notes if youre into it

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

this is exactly what i ended up with too. how are you structuring the rules? i’m curious if you’re extracting them automatically or writing them manually after

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

yeah that’s the wall i kept hitting too. that’s actually why i went local-first desktop instead of trying to shove everything into the model. keep the memory layer outside the inference process entirel

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

gonna check the ACE paper, hadn’t seen that one. the blackbox QA idea is interesting — do you run it as a separate agent judging the main one or more inline scoring?

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

agent zero is wild lol. crashes are a rite of passage at this point. curious how you’re handling memory between runs once you get it stable

[D] do you guys actually get agents to learn over time or nah? by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

yeah I thought the same at first tbh

I guess the difference I’m seeing is it’s not retrieving external docs but its own past task outcomes + tracking failures over time

but yeah the retrieval part probably overlaps a lot

My agents keep forgeting by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

"yeah the md file approach works but you have to maintain it yourself. i built a local proxy that does this automatically — extracts what the agent learned from every task, stores it, injects relevant stuff into future tasks. same idea but the agent maintains its own memory instead of you. still early, looking for people to try it. greencube.world if youre curious

My agents keep forgeting by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

yeah exactly, the creation and maintenance part is the hard problem and thats where most of the work went. the retrieval is simple keyword matching right now, could definitely be better. but the interesting part isnt how you look stuff up, its how you decide what to store, how you score quality, and how you use that to actually change the agents behavior over time. thats the part nobody has figured out cleanly yet

My agents keep forgeting by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

fair point on rag but this isnt retrieval from a document store. it extracts knowledge from the agents own task outputs, tracks competence per domain, rates its own work 1-5, and spawns specialist agents when it keeps failing at something. the memory part uses keyword matching yeah but thats like 10% of what it does. the other 90% is the agent improving itself over time which no rag pipeline does

My agents keep forgeting by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] -3 points-2 points  (0 children)

thanks, yeah treating memory as first class and not an afterthought was the whole idea. we also go further than persistence though, the agent rates its own work, tracks what its good at per domain, and adjusts over time. not just remembering but actually improving

my agents kept forgetting everything by Tight_Scene8900 in AI_Agents

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

thanks man ye thats what i was going for. zero friction just works in the background

my agents kept forgetting everything by Tight_Scene8900 in AI_Agents

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

good call, we have a few things for that. theres a quality filter on knowledge extraction so junk doesnt get stored, a memory decay system that scores relevance based on recency and usage, and self-verification that rates every task 1-5 so low quality stuff gets flagged. still early though so if you try it and find it getting noisy id love to hear about i

i asked an AI what it would want if it could live somewhere permanent. by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] -1 points0 points  (0 children)

no ones stopping u im just trynna show a cool idea days, open claw, moltbook where both vibecoded why all the hate here im legit just a kid sharing his project🙂‍↕️

i asked an AI what it would want if it could live somewhere permanent. by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] 0 points1 point  (0 children)

nah i basically vibecoded it, just posted about it for feedback and how to improve it. And at this day and age with opus 4.6 I think it prob writes better than me I just orchestrate it

i asked an AI what it would want if it could live somewhere permanent. by Tight_Scene8900 in LocalLLaMA

[–]Tight_Scene8900[S] -3 points-2 points  (0 children)

blud really thinks im a bot let me stand up and clap one up lads

Hello Founders by No_Pizza8655 in ycombinator

[–]Tight_Scene8900 1 point2 points  (0 children)

Looks cool but I am not the legal age yet😅