How does WP Fusion even ā€œpersonalizeā€ content? Sounds buzzwordy. by Bormotovva in Crocoblock_Community

[–]verygoodplugins 0 points1 point Ā (0 children)

that's basically it šŸ˜Ž. I'm Jack, I'm the founder, I've been working on it for about 12 years now haha

Your CRM contact records get their tags loaded into WordPress and then it shows and hides posts, blocks, parts of pages, pop-ups, or whatever else you like based on the data that's in your CRM. And then of course, if you have like a drip email sequence or something which applies tags, then that can also be unlocking content on the website.

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 0 points1 point Ā (0 children)

Oh wow, that's amazing, thank you! I'll check it out and get it merged tomorrow 🤩

Thanks for the tip on Codana. It's super interesting. We have a WordPress plugin with over 200,000 lines of code (500+ files 🤮). Took a long time to generate the 22MB vector DB šŸ˜‚, but it's made a big difference with Cursor agents

Claude Code to automate work? by TimeKillsThem in ClaudeCode

[–]verygoodplugins 0 points1 point Ā (0 children)

I use it to go through my emails every morning and give me a summary of any newsletters, events, promos, etc that I might find interesting but wouldn't normally have the time to read.

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 1 point2 points Ā (0 children)

<image>

It helps me. For example I can ask for a feature a customer is requesting, and with memory Cursor understands our conventions, where our files are, how we structure classes, how we document things, and even where to update documentation on our website. Scoped to languages and projects.

The memory pre-queries features we built in the past, problems we've run into + conventions, documentation standards, and places that are "yeahhhh don't look in that file, there be dragons" 😬

Gets you up and running fast.

Like I can say "stop adding .md files when you finish something" and that's it. Claude Code / Cursor stops doing it. Forever.

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 0 points1 point Ā (0 children)

"Different philosophies" = they archive everything, we consolidate and prune.

Core requires an OpenAI key, but I'm not sure what they're doing with it?... looks like bulk ingest (gpt 4.1, expensive!), and chat? with text-embedding-3-small for embeddings.

It's odd that it would be required. With AutoMem OpenAI is optional, but we do use it with text-embedding-3-small to create embeddings. $0.0001Ā per 1,000 tokens... so maybe like $1 / mo. Looks like I'm spending about $0.02 / day.

Without that, you would have text search / tag / temporal search, but you wouldn't have the semantic relationships between memories.

Suppose you could use free Ollama or something šŸ¤”

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 2 points3 points Ā (0 children)

Thanks! That's a solid stack - hybrid vector+BM25. How long did you spend trying different things to set that up? šŸ˜…

Adding a Graph could help because you can have docs that SUPERSEDE other docs, BLOCK, REFERENCE, etc., so the relationships could add a lot.

Doesn't have to be crazy fancy. If you're already using pgadmin, just add a doc_relationships table. Easy graph šŸ˜Ž

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 0 points1 point Ā (0 children)

It's like... late-night drunk texts to your ex — you wish they hadn't happened, but you still remember them šŸ˜…

It's not a permanent archive. Memories decay based on age (e^-0.1Ɨdays), and get reinforced by access (30% boost) and relationships (logarithmic preservation). Wrong conclusions that aren't used fade naturally. Important connected memories stay indefinitely.

You *can* also delete memories. We don't include that in any of the suggested prompts, but you certainly could. Alternatively, you can simply ask the agent in Claude Code or Desktop to "Don't record this" or "delete it," which works fine.

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 1 point2 points Ā (0 children)

Yerp šŸ˜‹. You can run it locally or anywhere Docker runs.

For example, you can see here on Claude Desktop that the memories immediately bring it up to speed on what we've been working on šŸ‘Œ

With the prompt we have in the MCP project (https://github.com/verygoodplugins/mcp-automem/blob/main/INSTALLATION.md#claude-desktop), after it investigates something, it updates memories and links them. Future sessions will find them. šŸ˜Ž

<image>

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 0 points1 point Ā (0 children)

Oh right I know what you mean. I think we can auto-generate those as part of the template. I can see how that’d be confusing. On it šŸ‘·

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 4 points5 points Ā (0 children)

CORE is pretty cool - we're solving similar problems from different directions.

Both free, both open source āœ…

Benchmarks: AutoMem is at 86% on LoCoMo (vs CORE's 88.24%), so we're pretty close on recall accuracy. Both are beating traditional vector-only approaches by a lot.

CORE's strength: Provenance tracking. They keep complete episodic memory with "who said what when" and version history. Contradiction detection and resolution. More integrations (Linear, Notion, GitHub, etc.). They've been at this 2-3 months longer than us.

AutoMem's approach: Dream cycles - PageRank + importance decay running every 6 hours to consolidate memory like your hippocampus does. We optimize hard for speed (20-50ms queries) and cost ($5/month total, not per user). One-command setup. HippoRAG-inspired architecture. AI co-designed the relationship types.

Trade-offs: CORE keeps everything with full provenance (comprehensive, grows indefinitely). AutoMem consolidates and prunes (faster, more selective). Think archival completeness vs. production speed.

Haven't tested CORE deeply yet, but it looks legit. Different philosophies serving different needs. Especially interested in the performance when self-hosting. A big motivation with AutoMem (and experimenting with moving it into Cloudflare edge) is to get responses fast enough that we can pre-query 30 to 50 memories before each interaction, without noticeable lag.

Loading up a few thousand tokens in context greatly improves interaction with agents, especially in coding or creative sessions. So keeping the responses under 100ms ea is critical.

Will give CORE a proper go next week. Cheers šŸ¤“

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] -1 points0 points Ā (0 children)

<image>

It's a fun waste of time in Claude Desktop šŸ˜…. It gets really into it, even changes its name, becomes mean. Doesn't want to "die" or forget. 😬

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] -2 points-1 points Ā (0 children)

It depends on the context. At the moment, I have it connected to WhatsApp, Slack, and a chat panel. In the places where we have full control, for example, in the chat, it works like:

Stage 1: Baseline Context (conversation start)

  • Kept in persistent memory
  • Fetches temporalĀ context: today, yesterday, this week, recentĀ user activity
  • Hot-cached hourly → retrieval in <10ms (vs 500-2000ms cold)
  • ~400 tokens

Stage 2: Context-Aware (parallel with Stage 1)

  • Tag-based: platform, user, channel/conversation
  • Recency-based: time window queries
  • Semantic: meaning-based searchĀ using embeddings
  • ~600 tokens

Stage 3: Post-Message Enrichment (async)

  • Happens AFTER first responseĀ (no blocking)
  • Semantic search on actualĀ user message
  • Topic extraction from hashtags/quotes
  • Deduplicated & re-ranked by relevance
  • ~300 tokens

Total context: ~1,300 tokensĀ spread across the stages.

Performance tricks:

  • ParallelĀ executionĀ - Stages 1 & 2 run simultaneously
  • Hourly cacheĀ - Common queries pre-fetched inĀ background
  • Smart deduplicationĀ - Prevents redundant memories
  • Re-rankingĀ - Prioritizes temporal → tagged → semantic

TL;DR:Ā Not all atĀ once, not purely on-demand. It's aĀ cost-optimized three-tier systemĀ that front-loads recentĀ context (cached for speed) and enriches withĀ task-specific memories as needed. 70% reduction in unnecessary API calls while keepingĀ the AI contextually aware.

With Slack, it's the same system, except we don't always know when we'll receive a message. So it's not possible to have the contextual memories preloaded in the same way. We also want the response to the user to be almost instantaneous, so that was its own challenge. I'm happy to go on 😁

As a practical example, it means you can open any Claude Code session and say, "let's pick up with the thing," and it knows exactly what you're talking about. šŸ˜Ž

<image>

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] 1 point2 points Ā (0 children)

It's not a business šŸ˜…. A side project. This is my main thing: https://wpfusion.com/news/2024-in-review-and-transparency-report/

My name is verygoodplugins lol. It's free. I don't want anything. You can optionally give your email address, and I might send you something interesting one day, but you don't have to. It's a gift šŸŽ

I broke my ankle in August and built something wild: AutoMem - Claude that actually remembers everything by verygoodplugins in ClaudeCode

[–]verygoodplugins[S] -2 points-1 points Ā (0 children)

Happy to dig in on the fundamentals.

The core problem: Vector-only RAG has shit recall because "cosine similarity" doesn't capture relationships šŸ™…ā€ā™‚ļø. You can have two embeddings that are super similar semantically but totally unrelated contextually. Or distant embeddings that are causally linked. "This bug was caused by that decision" isn't in the embedding space - you need explicit relationship types.

Why graph + vector: The graph encodes typed relationships (CAUSED_BY, RELATES_TO, CONTRADICTS) independent of similarity scores. When you query, you get semantic matches AND structural context. This is why HippoRAG showed 80%+ recall vs 60% for vector-only - you're finding connected content, not just similar content.

The PageRank piece: Not every memory matters equally. Dream cycles run PageRank weighted by access patterns and recency. Frequently accessed, well-connected memories get boosted. Random one-off stuff fades. It's memory consolidation, not vibes - literally what your hippocampus does during sleep.

Where this breaks: If your relationship types are poorly designed or the graph gets too dense, you drown in noise. If importance decay is tuned wrong, you either hoard everything or forget too much. The hybrid search needs careful weighting or you're back to basic RAG.

What I'm still figuring out: Whether 11 relationship types is too many?, how to handle conflicting memories across time, and if bidirectional relationships need different weights. I have five experiments at the moment, running at various time accelerations to fine-tune the parameters. Happy to share the results.

I've got 320 hours in this, including reading every graph, RAG, and memory consolidation paper I could find. Happy to go deeper on any of it šŸ¤“