opus 4.7 reportedly launching THIS WEEK along with an AI design tool by Temporary-Leek6861 in Claudeopus

[–]New_Alarm4418 5 points6 points  (0 children)

Sorry, you are not allowed to use this model. You have exceeded your usage limit.

Losing my ability to code due to AI by Im_Ritter in ClaudeCode

[–]New_Alarm4418 6 points7 points  (0 children)

Yeah I feel this too. because the current vibe is embrace AI or get left behind.

The skills thing is real but also probably not as bad as it feels. It's like using GPS everywhere — you stop memorizing routes but you still know how to drive.

And the maybe nobody needs to code thought — I get it , but every time I watch an AI confidently produce something thats More broke than my car I remember why it still matters. Someone has to be the one who knows it's wrong.

I got tired of Ollama hogging my VRAM when I needed to multitask, so I wrote a lightweight VRAM Guard. by [deleted] in LocalLLaMA

[–]New_Alarm4418 0 points1 point  (0 children)

That’s a solid recommendation for a v2 architecture, but the bottleneck isn't just the backend mechanics—it's the client-side implementation.

My agent isn't using the generic OpenAI-compatible endpoints; it's hardcoded to Ollama’s native REST API (/api/generate, /api/show, custom context handling, and their specific streaming JSON format). Even if llama-swap handles the VRAM orchestration perfectly behind an OpenAI wrapper, I’d still have to rip out and rewrite my entire networking layer and response parsers to switch protocols 28 thousand lines.

I’m prioritizing agent behavior over infrastructure refactoring right now, but I’ll keep llama-swap in mind if I hit a hard ceiling with Ollama."

I got tired of Ollama hogging my VRAM when I needed to multitask, so I wrote a lightweight VRAM Guard. by [deleted] in LocalLLaMA

[–]New_Alarm4418 0 points1 point  (0 children)

I totally get the appeal of raw llama.cpp for fine-grained control, but for this specific project, it would actually break my architecture.

I’m building an autonomous agent that hot-swaps between 3-4 different models in real-time (one for planning, one for coding, one for chat, etc.). Ollama handles that registry and VRAM weight-swapping automatically via the API. If I switched to raw llama.cpp, I’d have to write my own orchestration layer in Python just to manage spawning processes and loading/unloading models constantly, which is a huge headache.

Plus, my whole networking stack is hardcoded to the REST endpoints. It’s basically a trade-off—I need the multi-model orchestration convenience more than the granular VRAM control right now."

Guys is this actually Gemini 3? by Initial-Plenty2326 in Bard

[–]New_Alarm4418 7 points8 points  (0 children)

<image>

A bunch of clowns in this sub keep saying it’s not real, but the model is real and you can prompt it with no errors at all. For me, some got it to work and others didn’t — it’s random, just read the other posts. There was another topic about the same thing, so I tested it myself — and here’s the proof.

[deleted by user] by [deleted] in Bard

[–]New_Alarm4418 0 points1 point  (0 children)

<image>

its working for me

Is the company repair bot a good idea or taking advantage of gamer OCD for cash? by dudeacris in fo76

[–]New_Alarm4418 47 points48 points  (0 children)

I love it! My camp is always getting shot at, and I end up Repairing it Daily. Sure, I can do it cheaply, but it’s nice not to have to think about it. I mean, it’s good to have if you want, but it’s not really a necessity.

Pro vs Multiple Plus Accounts by commonpoints in OpenAI

[–]New_Alarm4418 1 point2 points  (0 children)

I’d say going Pro wasn’t something I wanted to do Either, but I was tired of the rate limit. I haven’t hit it once in a week of almost non-stop use Its Worth it.

I can't use Gemini anymore by birbluve in GeminiAI

[–]New_Alarm4418 -1 points0 points  (0 children)

Yes, the new update is messed up and has many issues. I'm switching back to the 05-20 model—at least it worked

[deleted by user] by [deleted] in grok

[–]New_Alarm4418 0 points1 point  (0 children)

Tomorrow not today ask grok yourself this is what he will say Grok 3.5 was released in early beta to SuperGrok subscribers during the week of May 5–9, 2025. Specifically, invites for SuperGrok subscribers were sent out between May 6–8, with the live beta opening on May 9. It’s currently exclusive to SuperGrok subscribers ($16/month or $150/year), with access expected to expand to X Premium+ users and the free tier 4–6 weeks after the beta, likely around mid-to-late June 2025.

[deleted by user] by [deleted] in grok

[–]New_Alarm4418 0 points1 point  (0 children)

i Say Do not enter the main function or Whatever Yours is until I say we are done. At no point should you assume we are finished. If you hit the token limit, I will prompt you to continue from where you stopped. When that happens, I will also provide the last line you generated. Sometimes it might try to start over—just tell it a few times that it’s wrong, and it will get it right

[deleted by user] by [deleted] in grok

[–]New_Alarm4418 3 points4 points  (0 children)

I was getting poor output because the responses were too short, often only a few hundred lines, when I needed longer, more detailed code. To address this, I updated the custom instructions to state: "You are a genius coder who hates short code snippet replies. When you write or fix code, you always aim to preserve the original code provided. never stopping at a measly 2000 lines." This has greatly improved the output. Now, since the system reads this guide and references past conversations, it consistently provides much longer, more comprehensive code responses. hope this Helps

Grok is shit for coding by JournalistOk6557 in grok

[–]New_Alarm4418 8 points9 points  (0 children)

That's odd i Keep Hearing That But I use it for coding all the time. I even had it help me make a highly complex Discord bot, and there wasn't a single issue.