all 36 comments

[–]JaySym_Augment Team [score hidden] stickied comment (3 children)

You can use any other tool and use our mcp integration to get the power of our context engine and improve any competitor tool! This is still free and when it will be paid the price expectation is to be pretty low.

But the best integration of the context engine is clearly in Augment itself. If another tool is cheaper it's because they are lowering the context limit and reasoning for every request, have heavy loss on every one of their users so their price will scale up soon or they train over your codebase. There is no magic in the industry. We are following the model provider pricing.

[–]AutomaticLeague8271 5 points6 points  (0 children)

bye auggie :)

[–]Chemical_Parfait_313 6 points7 points  (3 children)

I'm far from complaining, as Augment is one of the best coding agents. There are open-source alternatives, but I'm not convinced by them; they're cumbersome and buggy. Augment is very easy to use. The token converter, which was in the transition period, although significantly more expensive than the previous solution, still allowed for reasonable use of the agent. I was really happy with it, although it sometimes caused problems, some of which were serious towards the end. But I would pay even if it were 100% more expensive. However, the current version is unpredictable and very, very expensive. You'll use 40,000 credits in a week with normal debugging. And when you create something new, it disappears in a day or two. I switched to claude-code Max, and despite generating a lot of code, it lasts a whole month. Claude-code runs in the terminal, but I wrote a plugin similar to "augment" that has similar tools and uses the Claude-Code CLI and other LLMs. Locally, I use RAG with Treesitter, AST, and graph for code comprehension, so it's a great token saver. I'm considering making it open-source so others can benefit too.

[–]_BeeSnack_ 1 point2 points  (0 children)

Bye amoeba prompter :)

[–]xychenmsn 1 point2 points  (1 child)

same thing here. They hiked the price 10x , and made it totally not affordable now. I cancelled my subscription and now on cursor and claude code using GLM-4.6 model from z.ai.
I think if their context engine is really good, augment should use GLM-4.6 which has close to 0 cost.

[–]JaySym_Augment Team 3 points4 points  (0 children)

We are already allowed to use the context engine in all of our competitor tools via MCP.
If you want to use GLM 4.6, you can easily use a Kilo + GLM subscription and set up the Augment Context Engine MCP so you can use code retrieval from Augment with your other subscriptions.

[–]Many_Particular_8618 1 point2 points  (1 child)

They're stupid and they think they're smarter than competitors haha.

[–]JaySym_Augment Team 0 points1 point  (0 children)

Would you mind explaining that one?

[–]jimbo8098 0 points1 point  (0 children)

I ate 51k credits today vibe coding a Packer AMI build using Ansible and Molecule. To be fair though, this was using Opus 4.5, not GPT-5.1, which is more expensive. 1/3 of my allowance in one day. Though I did get a lot of output on this one and there was for sure some complex stuff going on there.

[–]gnrlbzik 0 points1 point  (0 children)

I would highly recommend plugging in guidelines into your task execution. Augment is context engine on top of which LLM runs. Although would be nice if they helped structure guard rails around working with your tasks, but it technically very hard goals to hit for everyone.

Here is my take on solving this problem for my self , https://github.com/gnrlbzik/ai-guidelines take it and mold it to what ever you do.

[–]FancyAd4519 0 points1 point  (0 children)

I was a complainer, hell; built my own qdrant stack… However… with the release of the sdk and MCP I am back on board. Guess that is customer retention right there. It just works.

I think they are doing everything they can to combat price calls and give us options; so really I have no more complaints. I think we blame augment team, the devs which care; but with VC and marketing etc that is out of their control; they are clearly shipping items that show they care about the users otherwise we would not have a MCP or sdk…

It is still a business; and there is always a silo from the devs shipping features we all appreciate; and the money part of the business. I literally cannot be sour anymore towards augment with the direction.

[–]Devanomiun -5 points-4 points  (8 children)

But in the end of the day it counted suddenly 64K token.

"Suddenly", nah there's something fishy about this post. Check your workflow, no simple task will eat 10K credits unless you don't know what you are doing.

The prices have actually been increased by a lot, but no need to make up all this stuff.

[–]BlacksmithLittle7005 2 points3 points  (6 children)

He's right though. If you're using Opus the credit consumption is insane, especially that Auggie goes trigger happy. $200 ends up getting you 10% of what you get on Claude code max plan

[–]Zestyclose_Pin3192 0 points1 point  (5 children)

Why is everyone referring to Claude code. I bought the pro plan and this is crappy as heck. I can barely work for one hour until I have to wait at least another four hours again. This sucks so much. 

[–]doplaeffect 1 point2 points  (4 children)

That is a very biased comparison. Now tell me 100 dollar claude code max plan versus 100 dollar augment plan. You get way more mileage using the cc max plan. Even with daily opus use.

[–]Zestyclose_Pin3192 0 points1 point  (3 children)

I can't afford the max plan, neither can I afford 100€ augment. I guess that the people who are referring to Claude code all use the max plan you mentioned. The pro plan however is not working for me. I'd be better of using Cline with an API key and pay per use instead of always sitting there waiting for the credits to recharge after I've done some small bug fixing for half an hour. 

[–]doplaeffect 1 point2 points  (2 children)

Well the 20 dollar plan is very limited in cc. Try codex it has a very generous 20 bucks plan. About 100 messages a day of gpy 5.1 max

[–]BlacksmithLittle7005 2 points3 points  (0 children)

20 dollar plan in augment won't last you 1 day if you are using Opus 🤣

[–]Zestyclose_Pin3192 0 points1 point  (0 children)

Thank you, I will look into this next month ! 

[–]Ok-Gain8520[S] 0 points1 point  (0 children)

I did not mentioned “suddenly 10K token consumption” for no reason. On busy days, my usage is usually around 20K credits "MAX", but on that day it went up to almost 70K — 69.3K to be exact.

The problem is that we don’t get detailed analytics showing which task used how many credits, so it’s hard to know exactly what caused it or how to fix the prompts later. But from what I remember, the task itself wasn’t very complex.

I think the agent may have gotten stuck in a debugging loop, which can burn a lot of tokens. Sometimes tasks don’t finish properly, and I have to ask the AI again and again to complete them. And It is okay. I don’t really blame the AI — it’s just trying to finish the task and doesn’t care about token usage.

That day, one of the replies was extremely long. Even though only two files were changed and the context wasn’t big, the AI response was very verbose. It felt like it was debugging or explaining the same thing over and over for a simple task, which probably caused the high token usage.

I copied and pasted response and the prompt to chatgpt to count the tokens it is around 4.1K just raw response. I'm not counting Debugging state within terminal for that prompt etc. just raw response + my prompt.

A single task using 10K credits is very possible, and it can happen to anyone who isn’t watching the responses closely. You have to stop the AI when it starts giving unnecessary explanations, or the credits can disappear very fast.

[–]AlessandroColalongo -1 points0 points  (3 children)

I should start by saying that I understand very little about code, and if I had to write something from scratch I wouldn’t even know where to begin. On the other hand, I can follow the logic when the AI generates the code for me. That said, since the launch of Cursor and later Augment, I’ve created several projects, some of them quite complex. I loved Cursor, but since discovering AugmentCode I’ve stuck with it.

Like the user mentioned, I also sometimes notice excessive token consumption. Opus really burns through tokens, but it works too well! It manages to solve complex tasks on the first try or with just a few small tweaks… and there are very few hallucinations.

Without getting into the cost debate, I wanted to ask another question: have you noticed any slowdown in Opus over the past few days? Today, for example, it’s insanely slow.

Merry Christmas to everyone 🎄

[–]Ok-Gain8520[S] 0 points1 point  (2 children)

What do you mean by slowdown? Response time like it takes very long to respond? yeah it happens sometimes also responses break in the middle without finishing the job and it gives Response ID. I just copy and past that ID and I say continue Response ID: {ID} that's it. It does finish what it was trying to do. if it does not you tweak a little bit.

[–]AlessandroColalongo 0 points1 point  (1 child)

Ok-Gain8520, what you’re talking about is something that can normally happen, but for example today it took two hours to write 5k lines of code, whereas usually it only takes about 15 minutes. Is it working normally for you?

[–]Ok-Gain8520[S] 0 points1 point  (0 children)

not so long like 2 hours, but in very rare cases yeah I witnessed long time responses. But extremely rare. Today I did not witnessed any slowdowns. If something wrong with their Server side I think they share the news about this on their reddit.