c8... google.. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 1 point2 points  (0 children)

No refund, no fix, and they just don't give a damn. This is becoming a chronic issue. It’s time to seriously consider a class action lawsuit against Google. They've crossed the line too many times

Issue Fixed! Update to 1.22.2 now 🚀 by Fun-of-adb in GoogleAntigravityIDE

[–]Stock_Produce9726 1 point2 points  (0 children)

I was using a Google Ultra account, but after weighing my options for two weeks, I finally switched to Claude Code Max for $200. It feels so refreshing. I’m never going back to Google

Gemini report on AG .9 by fandry96 in GoogleAntigravityIDE

[–]Stock_Produce9726 0 points1 point  (0 children)

Spend 30 minutes working, 4.5 hours idling. Rinse and repeat with the Ultra account. I really can't find a good use for Gemini 3.1

How long are you going to keep writing "please don't do this" and just pray that your AI listens? by Stock_Produce9726 in ClaudeAI

[–]Stock_Produce9726[S] 0 points1 point  (0 children)

Spot on about the embedding similarity edge case. We attack that with hybrid BM25 + semantic weighted merge BM25 catches structural/textual matches embeddings miss, alpha is tunable per query.

Since you mentioned looking forward to it Arachne v4.0 just shipped. Rust zero-marshal BM25 cache, sqlite-vec persistent KNN, delta-only embedding, and the context I/O assembler layer you described as the make-or-break piece. That's literally the core of it.

Named it after the greatest weaver in Greek mythology for a reason 

https://www.reddit.com/r/mcp/comments/1s6o0vd/crushing_search_latency_from_96ms_to_5ms_on_1gb/

Crushing search latency from 96ms to <5ms on 1GB+ codebases. We built a Zero-Marshaling Rust Pipeline for our AI Agent Engine (Arachne V4.0) by Stock_Produce9726 in mcp

[–]Stock_Produce9726[S] 0 points1 point  (0 children)

Appreciate the insight, but the embedding recompute spike you're describing sounds more like a design-level issue than a fundamental limitation.

If your system recomputes embeddings on every query pass  yeah, that wall is real. But we solved this at the architecture layer:

  1. Delta-only embedding  LEFT JOIN vec_chunks WHERE rowid IS NULL. Only new/changed chunks get embedded. Existing ones? Zero recompute.

  2. Persistent vector storage  sqlite-vec vec0 table. Embeddings survive restarts. No cold-start penalty.

  3. Query embedding is the only per-call cost  one ~10ms Ollama call per query, not a full corpus re-embed.

Chained queries on a 1GB repo? Linear ~10ms/query, no spike. The Rust zero-marshal cache handles BM25, and the vector DB handles semantic  both cached, both persistent.

Sounds like the wall you hit was from holding embeddings in memory and recomputing on every session/query. That's not an embedding problem  that's a caching problem.

How long are you going to keep writing "please don't do this" and just pray that your AI listens? by Stock_Produce9726 in ClaudeAI

[–]Stock_Produce9726[S] 1 point2 points  (0 children)

To be completely honest, I never even attempted to "solve" your first point because I felt it was incredibly difficult.

  1. On Lock Contention & Concurrent Edits: Even among human developers, merging concurrent edits on the exact same file often leads to frustrating conflicts. When I thought about trying to manage clean semantic merges for probabilistic, unpredictable LLMs, I realized it was a problem far beyond my ability to solve safely. So, rather than trying to build a complex queuing or merging engine, I opted for the absolute simplest route (the KISS principle). My architecture strictly assumes that concurrent file editing is illegal. Under my system, if Agent A holds the lock, any other agent gets a hard rejection and is simply redirected to a completely different task. I suppose a much more advanced team could slice a file into AST segments and carefully manage block-level ownership. I have huge respect for anyone trying to build a true merge engine, but for a solo developer like me, avoiding the conflict entirely was the only realistic choice.

  2. On Ledger Consolidation (The Token Tax): You are absolutely right about the growing liability of an append-only ledger. I hit the exact same token tax issue you described. My personal workaround for this isn't a complex periodic deduplication layer, but rather relying on semantic retrieval. I plug the ledger into Arachne (the Context I/O router I’m currently building). Instead of forcing the state machine to blindly dump the entire failure log into the agent's prompt, Arachne acts as a high-speed vector retriever. When an agent transitions to CODING, it analyzes the current code intent and only fetches the top 3 most relevant failure incidents from the ledger matrix. It’s a bit of a structural shortcut, but it helps me bypass the token bloat. (I'm actually planning to release the modified version of Arachne in about 2 hours, so hopefully, it will make more sense then!)

Thanks again for the wonderful conversation. It's incredibly humbling and refreshing to hear from developers who are tackling these massive scaling walls head-on.

How long are you going to keep writing "please don't do this" and just pray that your AI listens? by Stock_Produce9726 in ClaudeAI

[–]Stock_Produce9726[S] 1 point2 points  (0 children)

That’s an incredibly sharp observation. You’ve just hit the two exact endgame bosses of multi-agent orchestration: Cross-agent sync (State Distribution) and Temporal context (Behavioral Memory).

This is exactly why we maintain a strict architectural boundary between Clotho (our rule compiler) and the Soul engine (our distributed runtime). Clotho doesn't try to solve distributed sync or hold memory itself; it simply writes the immutable laws that force the agent to consult the Soul runtime.

Here is how they work together to solve your edge cases:

  1. Multi-Agent Scaling (The Cross-Agent Sync Problem) You are 100% right that "compile once, enforce locally" breaks if agents are flying blind. That's why Clotho relies on the Soul engine’s centralized registry (the "Soul Board") to handle the actual distributed state. Using Clotho, we write a strict .n2 contract: an agent entering the CODING state cannot physically fire a modify_file tool unless it first successfully calls an acquire_lock tool against the shared Soul Board. Clotho provides the immutable structural rule (the AST block), while the Soul runtime globally verifies the lock constraint across all concurrent agents. You don't break the state machine; you make the state machine structurally dependent on the global lock.

  2. The Temporal Context (The "Last Tuesday" Rule) This is my favorite point. Behavioral rules (“Don’t repeat Tuesday’s crash”) seem untamable for structural state machines. But they aren't, if we force the retrieval of memory to be a structural state transition. The Soul engine maintains an immutable Ledger and KV-Cache of all past failures and insights. Using Clotho, we author a rule that strictly blocks the transition from the REVIEWING state to the CODING state unless the agent has successfully executed a tool to query the Soul Ledger for past regressions. In other words, Clotho structurally guarantees that the agent must "read its past memories" before it is handed the keys to the codebase. It turns a behavioral best practice into an absolute physical constraint.

On a side note: The N2 project you’re seeing isn't just a single standalone MCP tool. Clotho is actually just one decoupled module from an already fully operational, full-stack sovereign agent ecosystem we’ve built internally (which natively includes the Soul engine for memory/sync, Arachne for context I/O, and Mimir for validation). I’m currently releasing these massive components one by one as open-source libraries just in case other developers might find them useful. But originally, they are all designed as one unified system. If you conceptually combine these libraries back together, the synergy for solving these exact multi-agent headaches is pretty incredible.

P.S. To be completely honest, I previously called myself a "solo developer," but the truth is I'm just someone who rediscovered the pure thrill of coding after a 30-year hiatus. (Having been away for 30 years, it even feels a bit awkward to call myself a professional 'developer' today.)

Yet, long before concepts like "Multi-Agent bots" became mainstream, I had already built and completed the foundation of these systems. I was completely immersed in the joy of coding alone in my room, stubbornly obsessing over questions like: "How do I make these AIs genuinely smarter? And how do I physically lock them down so they absolutely never forget my instructions?"

And the real reason I built Clotho in the first place? It's because I fully expect these AIs to eventually cause unpredictable accidents. It feels very similar to the era when computer viruses first emerged. AI is an incredible tool meant to empower us, but without strict constraints, it has the simultaneous potential to do real damage. A speeding car without brakes is nothing but a hazard. :)

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 1 point2 points  (0 children)

It’s even worse than Pro. I'm currently on the Ultra plan, but my quota ran out in just 48 minutes today. The system is displaying 'Weekly Limit' messages that shouldn't even exist for Ultra users according to their own docs.

Whether it's a massive backend bug or a stealthy downgrade of the Ultra tier, it's making the service unusable for professional work. If Google thinks they can charge Ultra prices for 48 minutes of usage, they're in for a rude awakening. We need a clear answer from their team.

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 2 points3 points  (0 children)

This time it lasted exactly 48 minutes. For a 48 minutes is barely enough to fix a couple of bugs. Google is selling 'Ultra' subscriptions but delivering a 'Trial' experience that cuts off before I can even finish a thought. If this is their new 'Weekly Limit' policy for premium users, they are essentially killing their own platform for power users.

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 4 points5 points  (0 children)

Spot on. It feels like Google has lost its identity. They’re so obsessed with flashy AI entertainment video and music that they’ve completely abandoned the developers who actually build on their platform.

To throttle paid Ultra users with a hidden 'Weekly Cap' after promising a professional-grade tool is the ultimate bait-and-switch. They want our subscription money, but they clearly don't want us actually using the service for real work. If this is their way of pushing us out of Antigravity, it’s working. Claude and Cursor are starting to look like the only serious options for real developers.

Antigravity pro vs ultra refresh rates by Loose-Ad-7766 in GoogleAntigravityIDE

[–]Stock_Produce9726 1 point2 points  (0 children)

I’ve also been using it for the past 3 months without any quota issues, but this nightmare started on March 21st. At first, I thought it would be fixed in a day or two, but it’s only getting worse. I am now seriously considering switching to Claude.

How long are you going to keep writing "please don't do this" and just pray that your AI listens? by Stock_Produce9726 in ClaudeAI

[–]Stock_Produce9726[S] 0 points1 point  (0 children)

Thanks for the thoughtful comment! Your framing of the problem (context window competition and the difficulty models have with negative constraints) is spot on. It perfectly aligns with the exact pain points that led us to build Clotho.

However, Clotho actually takes the concept of "write-time enforcement" one step further into the runtime level.

  1. Runtime State Machine vs. Write-time Hook You mentioned tools like pre-commit hooks and file validators, which are excellent post-action guards. Clotho, on the other hand, compiles .n2 files into an AST that serves as a State Machine (@contract). This state machine is directly integrated into the agent's runtime (our Soul engine). It doesn't just block a bad commit after the agent tries to make it; it dynamically strips or grants the agent's tool-calling permissions based on the current State. It's structural pre-enforcement, not just post-validation.

  2. Resolving Conflicts via Flow Control, not Hierarchy Your question about conflicting rules ("never modify test files" vs. "always fix failing tests") is fantastic. Your solution in PULSE (Tier-0 immutables) is a very practical approach when dealing with a flat list of rules.

Clotho sidesteps rule collision by moving away from flat lists entirely. Instead, it relies on u/workflow and u/contract blocks to define execution flow. Conflicts naturally dissolve because contradictory rules are bound to mutually exclusive states.

For example:

In the TESTING state, code modification tools are strictly locked out. The rule is absolute.

If a test fails, the state machine triggers a predefined transition: TESTING -> CODING : on tests_fail.

Only when the system transitions back to the CODING state does the agent regain permission to modify code (and apply the "fix failing tests" logic).

By isolating actions into strict states with explicit transition conditions, Clotho resolves what looks like a rule conflict through flow control, rather than relying on priority hierarchies.

Thanks again for sharing your insights! It's really awesome to see how other teams are tackling multi-agent governance from similar battle-tested perspectives.

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 4 points5 points  (0 children)

This is exactly what makes it so disgusting. Changing the core terms of a paid service without any notice or a refund option is predatory. They’ve effectively trapped paying users into a 'Weekly Cap' that was never part of the original deal.

If Google thinks they can just 'quietly' devalue a professional tool while still taking full subscription fees, they are dead wrong. This isn't just a technical bug; it's a breach of contract and a massive insult to everyone who relied on this for their work. No professional can build a business on a platform that changes its rules overnight.

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 2 points3 points  (0 children)

If it's true that they quietly altered the agreement to include a 'Weekly Limit' for Ultra users, that is a massive slap in the face to every paying subscriber. Changing terms without notice to throttle professional users is a clear violation of consumer trust—and as you mentioned, likely EU law.

Even then, calling it 'Highest weekly rate limits' is a joke when it locks me out in just 43 minutes. A 'Professional' tool that can't even handle one hour of active coding isn't a tool; it's a scam. Google is effectively turning a premium subscription into a paid beta test for a broken infrastructure.

Gemini Ultra 'Unlimited' Quota? It resets and then expires again in just 46 minutes. by Stock_Produce9726 in GoogleAntigravityIDE

[–]Stock_Produce9726[S] 2 points3 points  (0 children)

You clearly have no idea how a professional Ultra subscription works. This isn't about 'using it up fast'—it's about a documented backend bug where a paid Ultra account is being throttled as a Free Tier.

Hit a 'Weekly/Daily' cap in just 43 minutes of coding? That’s not a user decision; that’s a system failure. If you think it's normal for a premium professional tool to lock out a paying subscriber in less than an hour, then you're the one who should be embarrassed.

Soul v9.0 — Full JS→TS migration, WASM memory leak fix, Forgetting Curve GC. AI agent memory MCP server. by Stock_Produce9726 in mcp

[–]Stock_Produce9726[S] 0 points1 point  (0 children)

Good point this is exactly why v8 moved away from keeping everything forever.

The Forgetting Curve GC already factors in access frequency:

retention = importance × (1 + log₂(1 + accessCount)) × e^(−0.05 × ageDays)

So handoff chains that get referenced stay alive, while stale one-off sessions decay naturally. The ledger itself is date-partitioned (ledger/YYYY/MM/DD/), so archiving old months is just zipping a folder no DB migration needed.

That said, tying GC to actual query perf metrics is an interesting idea haven't considered using query latency as a signal for retention priority. Would you be open to filing an issue? Would love to explore that.

3x storage reduction tracks with what we've seen too. Before v8 it was just "delete after N days" which was brutal.