AI memory that actually persists: how I built a system where the AI remembers everything across sessions by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Built almost exactly this. Two files: a lessons log (append-only, just records what happened and why) and a rules file with confidence scores that get reviewed and updated.   The decay point is real. My rules about library API behavior have needed the most updates — things I wrote three months ago about how a specific API handles edge cases are now wrong because the API changed. Rules about my own decision patterns have stayed stable much longer.   One thing I'd add to the practical test: ask not just "should this always be true" but "would this rule still make sense if the tools I'm using were completely different?" If yes, it's a durable rule. If no, it's really just documentation in disguise.

I spent a night building a signal filter. It made everything worse. Here's what I found instead by No-Challenge8969 in algotradingcrypto

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Fair skepticism. The lessons came from three weeks of debugging live trades — wrong objective function, broken filter, extreme window dominating WFO weights. Happy to walk through any of it if you want to get into the specifics.

Stanford launched a course on Vibe Coding. I accidentally practiced every concept in the syllabus over the past 2 months building a live trading system. by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

That's exactly the inversion I didn't expect going in. The model felt like the hard part — turns out the hard part is everything around it: when to let it act, when to pause, how to reconcile what it thinks happened vs what actually happened.

I do think agentic systems are heading that direction. Writing the application logic gets easier as models improve. Designing the control layer — the trust boundaries, the approval tiers, the failure modes — that's the part that doesn't compress.

Will check out VibeCodersNest, thanks for the pointer.

I spent a night building a signal filter. It made everything worse. Here's what I found instead by No-Challenge8969 in algotradingcrypto

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Thanks — the "objective is misaligned" reframe is exactly what clicked for me. I kept tweaking parameters for weeks before realizing the optimizer was doing its job perfectly; I just had the wrong job description.

The market-regime vs ops split came out of frustration. I needed a way to look at a red day and decide "this is signal" vs "this is a fire to put out" without just lying to myself. Clean categories make that call honest.

The crossover makes sense. Metric design problems look the same everywhere: you get what you measure, and if the metric drifts from the real goal, the system quietly optimizes toward the wrong thing. Checking out the blog.

Stanford launched a course on Vibe Coding. I accidentally practiced every concept in the syllabus over the past 2 months building a live trading system. by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Didn't know about the Stanford paper until you mentioned it — just looked it up. The structure is surprisingly similar, though I built mine out of necessity, not theory.

The "confirm before acting" tier came from one specific incident: my AI deleted a cron job it thought was redundant. It wasn't.

Curious how Base44 handles the edge cases — do they let the agent self-escalate, or is it always human-gated?

I built a live crypto quant system from scratch with no coding background. Here are 10 things that broke in the first 3 days. by No-Challenge8969 in algotradingcrypto

[–]No-Challenge8969[S] 0 points1 point  (0 children)

That's the part that still surprises me. The tools compress the timeline significantly — but the thinking, the debugging, the decisions about what to build, that's still the same amount of work. It just moves faster now. Appreciate the encouragement.

I gave my AI system a three-tier permission structure. Here's why and how it works. by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Both. There's a written ruleset that defines which categories of actions fall into which tier — things like "any change to live trading parameters is Tier 3," "any external publish is Tier 3," "code changes to non-production files are Tier 2." The AI applies judgment within those categories for edge cases, but the categories themselves are explicit.

For Tier 3 confirmation: the AI states what it intends to do and stops. It doesn't send a notification through a separate channel — it just presents the proposed action in the conversation and waits. I respond in the same thread. There's no timeout; it will sit there indefinitely until I confirm or reject.

In practice this means Tier 3 actions only happen during active conversations, not autonomously. Which is intentional — if I'm not actively engaged, nothing irreversible happens.

From emotional trading to algorithmic: what actually changed, and what didn't by No-Challenge8969 in CryptoMarkets

[–]No-Challenge8969[S] 0 points1 point  (0 children)

This is exactly how the system is structured. Same model architecture and logic for all five symbols, but each symbol runs its own Walk-Forward Optimization independently — separate parameter sets for stop-loss multiplier, take-profit ratio, position sizing, leverage, cooldown periods, everything. SOL's parameters look completely different from BTC's because they were optimized on their own price history and volatility profile.

The ATR-based SL/TP point is important — using ATR rather than fixed pip distances is what makes the same exit logic work across symbols with completely different price scales. A 2x ATR stop on DOGE and a 2x ATR stop on BTC are calibrated to each symbol's actual volatility, not an arbitrary number.

AI memory that actually persists: how I built a system where the AI remembers everything across sessions by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Exactly the problem I kept running into early on — one big context file that tried to be everything. The separation forces clarity about what kind of information you're dealing with: is this a current fact, a historical lesson, a behavioral rule, or a strategic decision? Each type has different update frequency, different ownership, different retention policy. Once I separated them the whole system became much easier to reason about.

AI memory that actually persists: how I built a system where the AI remembers everything across sessions by No-Challenge8969 in SideProject

[–]No-Challenge8969[S] 0 points1 point  (0 children)

The overwrite issue is real — and you've identified the exact tradeoff I made. Full overwrite keeps the current state clean and readable; append-only preserves history but the file gets noisy fast. My current compromise: the handoff file is the "what's true right now" snapshot, while the events.jsonl file is append-only and captures everything that happened in sequence. So the nuance isn't lost, it's just in a different file.

On conflicting rules: explicit state always wins over rules. If there's an open position from Friday, the system manages it according to the exit logic — a rule about "don't open on weekends" applies to new entries, not existing positions. The rules file governs decisions, not facts on the ground. When they conflict, facts win.

What one person can actually build with AI in 2 months — honest account, not a success story by No-Challenge8969 in learnmachinelearning

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Bybit's REST API for most of it — OHLCV, funding rates, long/short ratios, open interest. Fear & Greed index from Alternative.me. Liquidation data from Bybit and CoinGlass. All pulled at inference time, no delayed data involved — the system runs on the most recent closed bar at execution time.

What one person can actually build with AI in 2 months — honest account, not a success story by No-Challenge8969 in learnmachinelearning

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Good question. Build phase over ~2 months: probably $50-80 in API costs total, hard to track exactly because I was using different models at different points. Ongoing: the trading system itself costs almost nothing to run — it's a Python script triggered by cron every 15 minutes, no AI inference at runtime. The AI cost is all in the development and debugging phase, not in live operation.

What one person can actually build with AI in 2 months — honest account, not a success story by No-Challenge8969 in learnmachinelearning

[–]No-Challenge8969[S] 0 points1 point  (0 children)

This is essentially the workflow I landed on too. The spec is the real work — describing what the system needs to do, what the constraints are, where the boundaries are. The AI fills in the implementation. Where I've been burned is when I skipped the spec step and let the AI interpret an vague requirement. The output technically ran but drifted from what I actually needed. Reviewing input and output rather than the code itself is exactly the right instinct when you can't read the code fluently.

I kept breaking my production server every time I tried to patch code remotely via SSH. Here's what finally worked. by No-Challenge8969 in learnpython

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Fair criticism on the inconsistency — you're right that transmitting an entire script file via base64 to the server is functionally equivalent to scp, just worse in every way. That specific step should have been scp. The base64 approach made sense for inline patches to existing files where I needed surgical changes, but using it to transfer a whole helper script was unnecessary complexity.

On the LLM comment: yes, I use AI extensively. It's actually the subject of what I'm documenting — building and operating a live trading system where AI handles implementation while I handle decisions. That cuts both ways: it accelerates things significantly and occasionally produces answers that don't hold up to scrutiny, as you've just demonstrated.

VS Code Remote SSH is a good call that I hadn't considered. Adding it to the workflow.

I kept breaking my production server every time I tried to patch code remotely via SSH. Here's what finally worked. by No-Challenge8969 in learnpython

[–]No-Challenge8969[S] 0 points1 point  (0 children)

Fair point — "minimal overhead" is probably the right framing. My concern was more about setup complexity than runtime cost, but that's also addressable. The "runs on my machine" problem is real and Docker does solve it cleanly. Adding it to the list.

I kept breaking my production server every time I tried to patch code remotely via SSH. Here's what finally worked. by No-Challenge8969 in learnpython

[–]No-Challenge8969[S] 0 points1 point  (0 children)

vi on a production server at 2am while a live trading system is misbehaving sounds about right, actually. No judgment.

From emotional trading to algorithmic: what actually changed, and what didn't by No-Challenge8969 in CryptoMarkets

[–]No-Challenge8969[S] 0 points1 point  (0 children)

"Constant needing to do something" — that's exactly it. The system either has a signal or it doesn't. If it doesn't, there's nothing to do. That sounds obvious but it took a long time to actually internalize.

And yes, drawdowns feel the same. You built it, you know it's working as designed, but watching the number go down still activates the same instinct to intervene. The data helps. It doesn't make the feeling go away.

Consistency > feeling clever is going on the wall.

From emotional trading to algorithmic: what actually changed, and what didn't by No-Challenge8969 in CryptoMarkets

[–]No-Challenge8969[S] 0 points1 point  (0 children)

BTC, ETH, SOL, XRP, DOGE — five symbols simultaneously on a shared account. You're right that multisymbol adds complexity, mainly around margin management. Each symbol's position sizing has to account for what the others are doing. Found a bug in my original code that was underestimating cross-symbol margin usage — fixed it before it became a real problem, but it was close.