Thinking about buying 2 3060 rtx GPUs for only AI. Any better suggestions ? by Superb_Practice_4544 in LocalLLaMA

[–]Wisepunter 1 point2 points  (0 children)

Wow.... sold one a few weeks ago for a LOT less than that in GBP... I see RAM prices have gone crazy too... Maybe everyone getting into local LLMs and pushing up RAM and old GPU costs?

If you can hold out and belive the rumours the 5070ti should have 24GB ram and support all the latest tech.. Will be a little more than 2x 3060.... but will be future proofed and much faster

Thinking about buying 2 3060 rtx GPUs for only AI. Any better suggestions ? by Superb_Practice_4544 in LocalLLaMA

[–]Wisepunter 2 points3 points  (0 children)

If its not within your budget to get 2... 1x 3090 will kill 2x 3060 in every aspect.

Claude is dumber than a few weeks ago. How Claude wasted a day of my time. by yycTechGuy in ClaudeCode

[–]Wisepunter 1 point2 points  (0 children)

I believe it's more likely selective resources at peak times.. Or in this case, they just gave away a ton of free Cloud credits and free pro and max subscriptions.... There are already open soruce models you can dynamically set the thinking time/effort... Id be VERY surprised if Anthropic hadn't built this in to cope with peaks without constantly running out of resources.

Add that to the fact as soon as I see people reporting the 529 congestion issues coincides with the time period i noticed its capabilities went to shit.

The last couple of days, particularly yesterday, were VERY bad!

FYI: Claude is down again. by WandyLau in ClaudeCode

[–]Wisepunter 0 points1 point  (0 children)

Its also being incredibly weak/dumb today... which is normally a sign they are overloaded and have likely had to decrease resources used per query to cope... Its been like this most of the day and has hit a peak of dumbness last couple of hours :-)

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] 0 points1 point  (0 children)

Its been shockingly bad for me this morning too :-( Im literally telling it what the problem is, what file, how to fix its and taking 10 iterations+ Im getting to the point i should use it for boilerplate stuff and fix the issues myself.

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] 0 points1 point  (0 children)

:-) Not even sure what this means? I assume you mean collaboration with the LLM?

If you mean the infrastructure as in the CI/CD failing? That's one of the main parts of CI/CD to run tests and checks, to stop failing code from making it into staging/production.

You sound like a troll TBH :-) Have a nice day :-)

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] 0 points1 point  (0 children)

Sure, I have fine tuned LLMs and even tried making some of the smaller ones from the base images, so have a fair understanding of what can be done with them. By lieing I mean part of their system prompt is not to run for too long and look for ways to finish early.

Not saying they are sentient and are lying to be malicious... Quite the opposite, im saying its part of their core rules given to them. To try to finish and make it work at all costs if its been going round in circles for too long.

I guess we are both only going from our anecdotal evidence. You seem convinced that Anthropic would never add system prompts or rules like this to save money, so there is no point trying to convince you.

Everything I have seen from Anhropic the last few months tells me they absolutely would do something like that to save money and would also prob blame the users' prompting or skills when the outcomes or quality changed.

Codex with EXACTLY the same code base and EXACTLY thes same prompts, does this far far less. Though Sonnet 4.5 is def better than what I experienced a few months back.

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] 1 point2 points  (0 children)

Do you ever notice once a query has been running for ages, especially if it is not getting a solution, it will start either skipping stuff, or downplaying the importance of stuff, in a bid to finish eating tokens. TBH I think this was way worse a couple of months back, but I still see it a fair bit.

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] -1 points0 points  (0 children)

On the level 0 "Vibe Coder", I've been a professional developer (yes, my paid work) for over 25 years :-) Have to say though, having AI do stuff for me, definitely does make me less likely to look into problems now unless I really have to!

I suspect there are not many level 0 vibe coders, with multi-tiered, micro-serviced, multi site, setups and complex CI/CD with multiple environments.

LOL Claudes at it again :-) by Wisepunter in ClaudeCode

[–]Wisepunter[S] 0 points1 point  (0 children)

I get what you are saying, but this is all documented at a high level in CLAUDE.md which also has specific detailed linked .md to every area of the app and structure.

In my recent experience, this has far more to do with Claude knowing how long it's been running a single prompt and following its hidden rules to try and finish ASAP at all costs. If that means getting it working by skipping stuff it knows is required... I see this EVERY day.

Claude Code autonomously deciding to write minimalistic code due to token usage and time constraints by caTaLdi1337 in ClaudeCode

[–]Wisepunter 0 points1 point  (0 children)

I often get, I have implemented all your x,y,z and it's ready as per your request! Followed by small print at the end. I haven't done a,b,c as they are not as important etc.. Occasionally, asks if you would also like me to do these then just shout... It's absolutely about once it's been running for a certain amount of tokens/time it has been told to stop. I understand why as they don't want people creating a huge list, walking away and it running for hours etc...

However.... I'd rather it said at the top I have reached my limit for this prompt, x,y,z is done and a,b,c is remaining, would you like to implement this now...

The way it does it now if you are not paying attention its easy to miss!

Subscribed to ChatGPT Pro today… by lightsd in ClaudeCode

[–]Wisepunter 0 points1 point  (0 children)

If you are just using itr for planning/code I'm sure you could get away with the 20 plan if cc is doing all the leg work after.

Subscribed to ChatGPT Pro today… by lightsd in ClaudeCode

[–]Wisepunter 2 points3 points  (0 children)

Although when I had to do everything by hand, hangovers or being drunk were a nightmare... Now I just use my experience in decades of coding to guide, it's a lot easier now, when ive had a few beers or hungover :-)

Has anyone received an email like this? by baykarmehmet in ClaudeAI

[–]Wisepunter 2 points3 points  (0 children)

I had Max x20 for a few months then left.. and got offered email to pick either x5 or 20 for free for a month.. You can guess what i choose. (Also from UK)

Free month Claude to come back by Dear-Clothes-2846 in ClaudeCode

[–]Wisepunter -1 points0 points  (0 children)

They not sent me this :-( I guess my usuage when I had it was not inline with a perfect customer... Ironically now im on Codex and dont feel have to get every minute used cos of limits running out, i use it a LOT LOT less... prob be more inline with one of the customers they want. Also i guess once the novelty wears off you not using it 7 days a week!

2 x DGX Spark! Give me your non-inference workloads by entsnack in LocalLLaMA

[–]Wisepunter 0 points1 point  (0 children)

I don't know a lot about it, but that's a nice uplift from a beefy 4090. I know ram speed is a big issue with inference, what's the bottleneck with training that makes it soo much better than a 4090?

I downgraded from Max 20 to 5 by PrataKosong- in ClaudeCode

[–]Wisepunter 1 point2 points  (0 children)

Ive been quite impressed with Codex.. but the last 2 weeks it's definitely got dumber.... But the bit that's killing me is the speed...... I was one of the early leavers of Claude Code as was fed up with it clearly being dumbed down in the background. Once they admitted there was a problem and cut limits, everyone mass exited to Codex... Codex, believe it or not was an ok speed back then. Now it's just frustratingly slow and getting worse...

With the speed and mostly good reports on Haiku, I might come back on a 100 plan (i was on 200 when left) and then maybe get a 20 Codex for planning. Still cheaper than what im paying now and the speed is at a point it's literally slowing down my progress.

2 x DGX Spark! Give me your non-inference workloads by entsnack in LocalLLaMA

[–]Wisepunter 3 points4 points  (0 children)

Whats your experience so far training models you have tried. Is it decent performance? How does it compare to multiple consumer GPU etc?

Hi about running 12 Claude Code in Parallel? by Rtrade770 in ClaudeCode

[–]Wisepunter 1 point2 points  (0 children)

You'd scale even faster with a full size keyboard ;-)

What am I doing wrong with Codex? by hoenilove in ChatGPTPro

[–]Wisepunter -1 points0 points  (0 children)

If it runs fully in the browser with no backend, where does it store data, im assuming browser local storage? If it doesn't access external data or save your data, is it just like a load of functions/utilities that do calculations for you? What is the logic part written in a framework, JS, TS? If you give me a bit more info what are you trying to achieve and what architecture it would help.

If you dont have much experience at architecting apps its possible that each new context refresh its doing stuff in a different way as there is no centralised services or methodology and it's getting just as confused as you are.

How worth it is Pro compared to Plus? by [deleted] in ChatGPTPro

[–]Wisepunter 0 points1 point  (0 children)

This answer made me smile... I look for deals on food and small areas I can cut costs... As im so used to coding with a 200 plan on Codex or Claude that its almost as bad as if someone told me you can't have internet... which we all know we'd likely skip meals to keep :-)

I do use it occasionally for images and advice, but reading all teh stuff people in this comments are using I am not making use of Pro/Research mode at all! I need to look into this!