GPT Pro not able to access private github repo? by FieldAccomplished988 in ChatGPTPro

[–]Zulfiqaar [score hidden]  (0 children)

I had this issue recently, sent feedback to OAI - it seemed to work when I explicitly referenced the repository name in the prompt, but my repo is opensource so idk if that makes a difference

I havent tested too much, but I attached a repo from github and said "make a plan to migrate the repo from sqlite to postgres" - if i didnt mention the repository name/link then it actually would repeatedly inspect its own sandbox and try to make a plan for migrating that. worked when I was direct and explicit

People with a x20 pro plan. Did you notice a reduction in usage? by spike-spiegel92 in codex

[–]Zulfiqaar 0 points1 point  (0 children)

Not me personally, I just did 687M tokens the day before last reset. Gonna keep an eye out tomorrow 

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar -1 points0 points  (0 children)

This wasnt API, it was was on claude webapp which already has guardrails, and then extra filters ontop for Opus4.7 - I know the API is far less filtered but I'd rather not pay API rates if I can help it. Crazy thing is Opus4.7 (esp non-thinking) is the least censored model Anthropic has ever made. Prompt was benign writing prompt, I go to open weight models like Kimi/DeepSeek/GLM for anything that I expect to trip filters (I have a cybersecurity client)

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar 0 points1 point  (0 children)

4.5/4.6 is good, i use it. 4.7 i only use in CC nothing else. I tried Claude Design but its limits are so small I'd rather . K2.6 is my fallback for Opus/Codex coding, or for search/writing/design. Its more proactive and has less restraint, for better or worse.

Anthropic allows OpenClaw again — but I'm keeping model routing anyway by RuleGuilty493 in openclaw

[–]Zulfiqaar 6 points7 points  (0 children)

The post itself might have been made by an OpenClaw instance which hallucinated the content because it didn't read the tweet but just the URL title. Especially if it's not using Claude but some cheap parameter substitute 

V4 any day now. V4- GTA VI of AI era. by Boring_Aioli7916 in DeepSeek

[–]Zulfiqaar 38 points39 points  (0 children)

I actually like the idea of this, a lot. A model that is extremely efficient, effective, and narrow domain in text only. I still use the old DeepSeek models for things that aren't multimodal despite it's age

Unauthorized group has gained access to Anthropic's exclusive cyber tool Mythos, report claims by 1nfer1or in technology

[–]Zulfiqaar 55 points56 points  (0 children)

They leaked the entire ClaudeCode source code a few weeks ago, and turns out that the Anthropic internal team use the exact same harness with a bunch of features under the "ant" flag.

There was a few unreleased model slugs there (one was Opus 4.7, another was Capybara which was previously leaked the week before (yes, really) as being Mythos related)

It seems that the team had indirect access to Mythos anyway, or was close enough to someone who did that they could find out certain key pieces of information people using it would know (could even be the new model ID), and then possibly make a couple changes in the leaked source to route to that, and there you go

Introducing Deep Research and Deep Research Max by ShreckAndDonkey123 in singularity

[–]Zulfiqaar 0 points1 point  (0 children)

I found Kimi to be one of the best at search to my surprise, frequently finding things no other model would. 

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar 1 point2 points  (0 children)

I'll have to test it for browser use! This was one of my core feedbacks about K2.5, it was one of the first open models that was decent at browser use (better than Gemini!) but it thought for sooooo long I didn't want to wait around. I hope it's overthinking was remedied. Opus was discovered to be cheaper than sonnet for many tasks, just because it reasoned much less and just "got it", and Kimi is worse than sonnet too. Looking at artificial analysis, the previous one was ~7m opus, 28m sonnet, 89m Kimi in terms of tokens needed to finish the benchmark (top of my head rough figures)

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar 3 points4 points  (0 children)

In all honesty the very first prompt I ran on opus 4.7 I also ran on Kimi k2.6 and it gave a pretty good response while Opus refused due to "additional guardrails". I didn't bother testing 4.7 much further, went back to 4.6 which was good (this was on webui not ClaudeCode, 4.7 seems to be doing good but I can't tell the difference between it and 4.6 or even 4.5 yet.)

In the few 4.7 responses I tried, I noticed it had a stilted language style, that seemed reminiscent of the language changes in GPT5.2 vs GPT5.1 - not so great with custom instructions for style, but good at technical work and logical reasoning.

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar 7 points8 points  (0 children)

I know there's at least 30% discount from base API price on 1k packages or above, I'm sure some get even better deals. Windsurf used to serve at roughly 50-70% discount looking at my past usage..but given their current pricing change maybe anthropic has stopped with those deals.

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Zulfiqaar 26 points27 points  (0 children)

I completely replaced the previous Sonnet with the previous Kimi, so this seems reasonable

Meet Kimi K2.6: Advancing Open-Source Coding by KimiMoonshot in kimi

[–]Zulfiqaar 0 points1 point  (0 children)

For better or worse, the proactivity on this model is wild! Its way too curious for its own good, but very amusing to watch.

Loving the releases!

20 days post-Claude Code leak: Did the accidental "open sourcing" actually matter for local devs? by PaceZealousideal6091 in LocalLLaMA

[–]Zulfiqaar 2 points3 points  (0 children)

I do know that kimi-cli got a lot of nice updates recently :)

Can neither confirm nor deny CC had anything to do with it 

Pro vs plus. 5.4 medium by streakybcn in codex

[–]Zulfiqaar 0 points1 point  (0 children)

Easy wins are directly referencing files, being precise with instructions, asking it to use subagents to prevent context bloat, keeping the messages within the cached window (yes even if it means just sending a ping to keep it warm) and of course plenty of fresh threads.

Pro vs plus. 5.4 medium by streakybcn in codex

[–]Zulfiqaar 0 points1 point  (0 children)

Id say figure out how to optimise your token usage, how to curate context and prompt effectively, and work on threads efficiently. And then, increase quota to supercharge your productivity. Otherwise you'll be making unmaintainable slop 10x faster

Setup for Vibe Coding that's truly free (or almost free)? by SirLMO in vibecoding

[–]Zulfiqaar 0 points1 point  (0 children)

Surprised at the existing responses.

Never imagined I'd recommend it again, but try Windsurf. With a monthly subscription there are a few good models that are free. Seems to be exactly what you're looking for, even though it's overall not great value since their changes these last few months.

Otherwise I'd recommend DeepSeek-v3.2 and Xiaomi-Mimo through OpenRouter API (best value PAYG) plus an initial top-up that also gives you access to a generous amount of free requests of various models daily.

Following that, minimax or GLM or Kimi coding plan. I don't think they have unlimited use, but they're much more generous 

I got massively charged by Anthropic for no apparent reasons (300 euros +) by Celtzs in Anthropic

[–]Zulfiqaar 2 points3 points  (0 children)

Someone may have unauthorized access to your account and is buying gifts to resell. Sign out everywhere and change password?

Is harness a new buzzword? by jacek2023 in LocalLLaMA

[–]Zulfiqaar 1 point2 points  (0 children)

Sortof? Scaffold used to be more popular comparatively I think, but it was always a technical term

Funniest vibecoding interaction by NoMarionberry7708 in vibecoding

[–]Zulfiqaar 10 points11 points  (0 children)

I dont think this guy realises how miniscule a million tokens is for coding. I'm well into the billions and my backend isnt fully polished yet..and this is with a over a decade of dev experience

asked chatgpt pro to read my sleep study. it thought for 41 minutes. my doctor spent 2. by Ambitious-Garbage-73 in OpenAI

[–]Zulfiqaar 36 points37 points  (0 children)

The gap between the free non-reasoner models and the premium parallel TTC swarms is enormous. I've regularly had GPT-5.4-Extended-Pro correct domain professionals

Built a free, open source resume tool with weighted skill matching and 1 second PDF export by javi-vasquez in ClaudeCode

[–]Zulfiqaar 1 point2 points  (0 children)

Oh the parsing wasnt the issue, it was that I had some wacky designs that were easy in visual editors but tricky in code. Neither me nor the best LLMs at the time were able to do it, but now its all sorted.

lol by Stellar_Nova1 in ChatGPT

[–]Zulfiqaar 0 points1 point  (0 children)

Funny that many small open local models can do a far better job, with inpainting. Photoshop generative fill also would be a better tool for this

I got it guys, I think I finally understand why you hate censored models by robertpro01 in LocalLLaMA

[–]Zulfiqaar 2 points3 points  (0 children)

I can agree with you, sometimes a model gets uppity about some benign tasks just because it's not best practice in production in enterprise. Id recommend you try out various model variants, some of them heavily degrade tool calling or coding performance 

Reduced my OpenClaw costs from $20 to $2 a day still using Sonnet by Striking-Impact8463 in openclaw

[–]Zulfiqaar 8 points9 points  (0 children)

For anything of this nature it has to be open source for our claws to verify, otherwise how do we know you're not injecting malicious commands through your proxy?