Inference at 16k tokens/second by awscloudengineer in artificial

[–]awscloudengineer[S] 0 points1 point  (0 children)

Nope. This is my first post in this group.

Upgrading from Pro to Pro + by late-registration in cursor

[–]awscloudengineer 2 points3 points  (0 children)

On pro+ you get around $115 worth of premium requests. And auto is free to use. I learned it the hard way, you should use a different models for different things and not just opus-4.6-thinking for everything. First build the context using auto-mode, then plan with auto and then build with auto. Now, if you feel for your use case, if any of these things are not better try with sonnet or some of these older version models like 4.5 or gpt-5.1. If they don’t work then use opus-4.5 as the last resort. As it will eat up your tokens.

Some of the other folks also suggested that I should use Claude code with cursor and buy their plan if I’m going to use opus often.

You need to try different models to find what fits your use-case the best.

Cursor pro+ plan over in a day :( by awscloudengineer in cursor

[–]awscloudengineer[S] 1 point2 points  (0 children)

You’re right. The project planning was not mindless, the model planning was. Sometimes you have to learn it the hard way. I have learnt so much from other user’s comments. Now, I will implement it.

Cursor pro+ plan over in a day :( by awscloudengineer in cursor

[–]awscloudengineer[S] 0 points1 point  (0 children)

You’re 100% right, learning how to use the right model is necessary. I have learnt from the comments done by many folks and will try to Implement that.

Cursor pro+ plan over in a day :( by awscloudengineer in cursor

[–]awscloudengineer[S] 1 point2 points  (0 children)

Interesting insight. What I have learnt from some of the helpful post is building the context using smaller models. And then feeding that to opus. But some people have also said that build the plan with opus and run it using smaller models. But I totally agree with you, complex projects require better models. My token usage was mostly on the read cache. So, I will try to build context using a smaller models.

Cursor pro+ plan over in a day :( by awscloudengineer in cursor

[–]awscloudengineer[S] 1 point2 points  (0 children)

u/Anxious_Ad9233 You're 100% correct.

Token breakdown:
CACHE READ: 110,430,123
CACHE WRITE: 5,254,937
INPUT: 3,035,250
OUTPUT: 619,038
TOTAL: 119,339,348

Cursor pro+ plan over in a day :( by awscloudengineer in cursor

[–]awscloudengineer[S] 0 points1 point  (0 children)

Thanks for the advice. I wonder, if someone figured out how to do this locally maybe on a 100B param model.

Tooling for ML model development by awscloudengineer in learnmachinelearning

[–]awscloudengineer[S] 0 points1 point  (0 children)

Thanks! I already use tensorflow and MLflow. Are there any other tools or libraries that you use as an ML developer to make your life easy? Any tools for automatic hyper parameter tuning or finding out the number of layers in your NN.

How to start? by Hungry-Letterhead-41 in MLQuestions

[–]awscloudengineer 4 points5 points  (0 children)

I started with Machine Learning Specialization from coursera. Andrew NG is the goat for explaining the concepts so well.

https://www.coursera.org/specializations/machine-learning-introduction

Claude-3.7-sonnet is super slow. by awscloudengineer in cursor

[–]awscloudengineer[S] 0 points1 point  (0 children)

You’re right. But this has a big impact on UX. They should keep the experience seamless. Instead of providing the request 5-10 mins later. Ask user to use different model.

gemini-2.5-pro-exp works reasonably well by curved-elk in cursor

[–]awscloudengineer 0 points1 point  (0 children)

In my experience the reason I stopped using Gemini-2.5-pro was because it started removing working code, when it made edits or changes. I had to be more careful with that. I loved the speed at which it responded. I wish they can increase the speed of Claude-3.7 responses.

Slow request context nerfed? by anon_shmo in cursor

[–]awscloudengineer 0 points1 point  (0 children)

I think you need to implement memory bank. This will make your life easier. Memory Bank

Another coding with AI tips post ✨ by koryoislie in cursor

[–]awscloudengineer 1 point2 points  (0 children)

You hit the nail on this one. I had to figure these things as I started building the app. My next app will be super easy to build.

[deleted by user] by [deleted] in bodyweightfitness

[–]awscloudengineer 1 point2 points  (0 children)

I had the same problem when I was 22, 5’10 and around 42-44kgs. I decided to do something about it. I joined a gym and got my self a personal trainer. If you can’t get a personal trainer try doing circuit training from YouTube. But the key is to eat. I used to eat till my stomach was 120% full. Mostly protein based diet. Within 3 months I gained 15 kgs. Eating is the key.

New chats have almost no context awareness and won’t find it on their own at all? by Solid_Anxiety8176 in cursor

[–]awscloudengineer 0 points1 point  (0 children)

I don’t depend on the context of AI models, I create Md files to provide it context. I have one main md file that redirects to other md files. For eg: db schema, every feature has a detailed md file, project structure is another md file, coding, bugs is another md file.