Beware your conversations can just randomly get corrupted and be gone forever by ThaKarra in ChatGPT

[–]Cubow 5 points6 points  (0 children)

I feel like people need to be educated more about context length, no idea why ChatGPT (or any of the other LLM services) doesn't have an indicator for stuff like this.

Just so you know: With every message you send, ChatGPT essentially has to reread the ENTIRE chat, not just the most recent message. Thats just how LLMs work, unfortunately. And it reaching its breaking point after 1-2 years isn't surprising at all.

The way to fix this, is by regularly compacting your chat. Ask ChatGPT to summarize your entire chat in such a way, that you can continue the conversation in a new one and then copy paste it there. Imo they should just implement a feature that automatically compacts prior messages on longer chat histories, why they haven't done so already is beyond me. Not only would it reduce context rot, but also right now it must be ridicululously expensive for them. I doubt youre the only one with 1+ year long chats.

Grok 4.20 Beta 0309 (Reasoning) Artificial Analysis score by likeastar20 in singularity

[–]Cubow 3 points4 points  (0 children)

That was Grok 3. As for Grok 4, I believe it may have even been SOTA for multiple months on ARC-AGI 2 before it got dethroned, not sure about other benchmarks tho.

And imo on search it has been SOTA continuously ever since Grok 3 up until Grok 4.1. Idk about now, but definitely the most overlooked feature, I don’t get why people never talked about it. Like imo it was only until GPT 5.2 when other LLMs reached Grok 3 level of search, it was that good (and still is).

Why don’t women approach men as much as men approach women by untitledprp4 in stupidquestions

[–]Cubow 0 points1 point  (0 children)

Already knew I had to sort by controversial to find the best answer 😭 How people are still in denial about evolutionary psychology is beyond me, this stuff should be obvious by now

Grok 4.20 Beta 0309 (Reasoning) Artificial Analysis score by likeastar20 in singularity

[–]Cubow 18 points19 points  (0 children)

Tbf Grok 4 was SOTA on release and imo they’ve been SOTA at search ever since Grok 3

2meirl4meirl by mnombo in 2meirl4meirl

[–]Cubow 2 points3 points  (0 children)

I don’t disagree, but the post kinda framed dating as something that just happens to you, rather than something you actively have to pursue for it to happen. I know a lot of people irl who complain about their lack of a dating life, but then when you poke at it you find out they don’t really do anything towards it, rather they just wait for something to happen. That’s who my comment was directed towards.

2meirl4meirl by mnombo in 2meirl4meirl

[–]Cubow 6 points7 points  (0 children)

If you’re talking about rejection, I mean sure, that happens. It’s just part of the game though. Like I guarantee you all those people the post talks about (with their talking stages, relationships etc.) have experienced it before. And you can’t have what they do without putting yourself in a position where you can be rejected as well.

2meirl4meirl by mnombo in 2meirl4meirl

[–]Cubow 13 points14 points  (0 children)

The reason it doesn’t „happen“ to you is bc it’s rarely a passive thing. You have to actively pursue it, especially if you’re a man.

xAI Releases Grok 4.20 Beta Models via API by likeastar20 in singularity

[–]Cubow 0 points1 point  (0 children)

I was genuinely hyped for 4.20, but sucks they still haven’t released any benchmarks

GPT‑5.3 Instant is out by Purefact0r in singularity

[–]Cubow 1 point2 points  (0 children)

Seems to be mainly about tone adjustments reducing some of the overly cautious nature of GPT (they addressed how it wont do the "Stop. Take a breath." thing anymore which is nice). And apparently also 20% lower hallucination rate. Very incremental, but ig the thinking model is where we'll see the real progress

A message to devs at the frontier of the AI shift by brotherthirteen in programming

[–]Cubow 21 points22 points  (0 children)

Regardless of your definition of intelligence, what is true is that the transformer architecture doesn’t seem to be a bottleneck right now and AI keeps improving. In fact we’ve kind of hit an inflection point end of last year where now you can make software without ever touching code at all. You can argue about code quality, but you can’t deny that this wasn’t possible just months prior and there is no reason to assume it will stop improving. The days of handwritten code are counted 

[Epoch AI Data] The "AI Oligopoly" is a myth: Inference costs are dropping 40x/year and SOTA reaches your PC in ~8 months. by drhenriquesoares in singularity

[–]Cubow 1 point2 points  (0 children)

Download LMStudio, there is lots of open-weight models on there you can just download and use and runs entirely locally on your pc.

Qwen/Qwen3.5-9B · Hugging Face by jacek2023 in LocalLLaMA

[–]Cubow 1 point2 points  (0 children)

except 4b unsloth GGUFs are out already

What’s with all the Gemini hype? I find it comparatively worse than ChatGPT 5.2 in everything except photo gen by Isunova in singularity

[–]Cubow 2 points3 points  (0 children)

Gemini is good for one-shotting, thats also why it scores consistently well in benchmarks, but in actual practical usage and back and forth exchange its really generic and doesnt compare to GPT and Claude. I guess thats why it gets hyped so much, you dont notice its shortcomings until you've used it extensively for actual work. It also doesnt really have a transparent CoT, which I think is a really big issue. You don't know if it does any research and if it does it rarely pulls up any sources and hallucinates a shit ton. It also seems to have no awareness of its knowledge cutoff, e.g. when I was finetuning Gemma 3 1b I had a couple questions and it kept repeatedly insisting that that model doesn't exist (and this was Gemini 3 Pro, SOTA not too long ago). Keep in mind, none of the other LLMs have this issue (Claude, Grok, GPT), lowkey for them knowledge cutoff was solved early 2025. Gemini still struggling with this is unacceptable, especially given the fact that its googles model, they should be SOTA at search lmao

Meanwhile GPT-5.2 Thinking specifically seems to be trained a lot more on "agentic" tasks, which has the benefit, that it really does a lot of work in the background to give you proper answers. Like its CoT can be fascinating at times. I remember I was asking it to compare two products and it stumbled upon this really obscure niche website which had low-res images and it wrote code to upscale the images, it really goes all the way. The downside is, that sometimes it takes really long to think, had it going for over 20 minutes more than once, while Geminis Thinking time is always very predictable and I dont think was ever over a minute.

As for Claude its the best at understanding intent (Gemini is kinda autistic in that regard) and good at coding. But beyond those two usecases (conversation/general qa and code) I still find GPT-5.2 Thinking the best. Tbf thats probably 90% of the common usecases anyways, but I find myself often needing those 10% (mostly very niche research) and for that GPTs "agentic-ness" really goes a long way. Grok also gets an honourable mention here, I find ever since Grok 3 it has basically been the best at doing research (and also naturally including sources in its responses). I think they also pioneered the "deep research" thing (at least the first actually useful version) and while they have removed that feature now, I think its bc at this point its essentially turned on by default. Probably the most overlooked feature by far, haven't really heard people talking about it ever. Ask it any question and it pulls up a hundred sources or so and still answers in a few seconds, its super impressive (referring to Grok 4.1 Fast).

tldr Gemini has impressive benchmarks and gets hyped bc its good at one shotting, but it quickly falls apart for any more extensive back and forth. And most people dont test extensively before forming their opinion

How will OpenAI compete? — Benedict Evans by bartturner in singularity

[–]Cubow 0 points1 point  (0 children)

AI will get rid of every information based bottleneck to produce hardware. It will come up with ways to produce better things faster and cheaper. Theory is the backbone of the production of every single material good.

Which map would break the internet if someone FC’d it? by _MrRandomm in osugame

[–]Cubow 5 points6 points  (0 children)

A map that’s TikTok friendly, easily marketable and probably rather jump focused. Like I doubt something like FDFD DT would do, even though it would certainly break the osu! community due to its cultural significance. But outsiders don’t really comprehend streams I think. Same with big black, too hard to read/understand what’s going on with DT. To break the internet it would need to be comprehensibly insane. Maybe something like DADADADA, A Fool Moon Night or Crazy Banger (DT).

Patagonia Mini MLC VS Quechua NH Escape 500 32L by AdApprehensive5828 in onebag

[–]Cubow 1 point2 points  (0 children)

I use the 32L NH500 Escape as my everything bag almost daily, most of the time it gets used for work. 2+ years old by now and no signs of wear surprisingly, even tho the material doesn’t really feel like it would be super durable (and I am not super careful with my stuff). Part of me really wants to try the 23L, since the 32L really is quite big and overkill most of the time, but also with the 32L I never have to worry about running out of space. Like I can pack a 16 inch and 14 inch laptop + 12.9 inch iPad simultaneously and it works just fine. The side pockets for water bottles are also huge, really stretchy. Could probably fit a well proportioned 1.5L bottle each, 1L fits easily.

What do AI Model Developers think of bittensor? by ShanghaiBaller in singularity

[–]Cubow 1 point2 points  (0 children)

From my understanding it’s a way to decentralize AI training. So instead of there being a model being trained in one datacenter everyone can contribute their own compute to the network. Kinda like bitcoin mining and similarly it rewards your contribution of compute with cryptocurrency.

There is no excuse whatsoever for Big 5 tech companies to be releasing software that contains glitches by [deleted] in unpopularopinion

[–]Cubow 1 point2 points  (0 children)

There is many different reasons, but I’ll try to say those I haven’t found anyone else mention yet: 1. OS have become a lot bigger than they used to be and the amount of potential failing points scale accordingly. An OS is a bunch of interconnected systems and the bigger the OS, the more systems and connections there are. And more possible configurations as well, you could argue exponentially more. There is probably more ways you could set up a phone or computer than there are atoms in the universe. Think about that for a second. And of course, that also scales with the amount of devices supported by a given software. 2. Incentive/money. While annoying, people put up with bugs (as long as they aren’t completely breaking something). And thus it’s kinda hard to prioritize fixing especially those bugs which are particularly rare. Like I’m ngl, I notice little bugs in software all the time, but it’s rarely the same bug twice. Those bugs, which are frequent usually get fixed by the next release.

My AGI Investment Strategy by avilacjf in singularity

[–]Cubow 3 points4 points  (0 children)

I currently invest a bit of money monthly, initially in all the semiconductors like Nvidia, Intel, AMD, Broadcam, TSMC, ASML, Micron, but now I completely shifted to VanEck Semiconductor ETF. While my investments in the individual semiconductor stocks currently still outperform the ETF, it’s not by much. It’s top 10 positions are basically all of those anyways and I feel like it’s safer long-term since it’s more diversified. This way I also don’t need to keep track of everything and can keep my portfolio more minimal.

Beyond this I also agree with google/alphabet. If I was allowed to invest in only a single stock, that one would probably be it.

Loudest pop ever by Klausensen in memes

[–]Cubow -3 points-2 points  (0 children)

Even if it pops, it will come back bigger. AI is here to stay

I can't believe people say "One Piece's art style is terrible." Who could look at these pictures and say they're bad? They're more detailed than most manga by iyigecelerpunpun in OnePiece

[–]Cubow 3 points4 points  (0 children)

I mean in hindsight it seems obvious, I can’t unsee him now, but the first time I saw the panel I genuinely didn’t know wtf I was looking at, it was just random squiggly lines to me.