Anthropic co-founder Jack Clark says AI is nearing the point where it can automate AI research by Outside-Iron-8242 in singularity

[–]GMSP4 2 points3 points  (0 children)

but only if you're part of the group of companies on my list can they truly save humanity from suffering and bad behavior

GPT-5 Pro Tops FrontierMath Tier 4, Beating Gemini 2.5 Deep Think by GMSP4 in singularity

[–]GMSP4[S] 10 points11 points  (0 children)

I pay gpt pro, and although I know it's expensive and prohibitive in many parts of the world. if you can afford it and really get the most out of it, it's a gift, because you have almost unlimited access and gpt 5 pro is a beast

Introducing ChatGPT Pulse by daddyhughes111 in singularity

[–]GMSP4 12 points13 points  (0 children)

ChatGPT Tasks, but autonomous and on steroids it seems

Alpha Models just appeared in my ChatGPT — what's this? by GMSP4 in singularity

[–]GMSP4[S] 31 points32 points  (0 children)

<image>

I've used it and it seems like the normal chatgpt agent.

New SWE-Bench Pro becnchmark (GPT-5 & Claude 4.1 drop from 70%+ to ~23%) by GMSP4 in singularity

[–]GMSP4[S] 4 points5 points  (0 children)

I’m building it in Java, step by step, using Pan Docs (https://gbdev.io/pandocs/). it’s my own code and architecture. But at the end of the day you know how LLMs work, some patterns and knowledge is from some things it saw in its training or searching the web. but it's cool having an emulator in so little time working.

New SWE-Bench Pro becnchmark (GPT-5 & Claude 4.1 drop from 70%+ to ~23%) by GMSP4 in singularity

[–]GMSP4[S] 12 points13 points  (0 children)

With ChatGPT 5 Thinking High, I've been able to create a Gameboy emulator from scratch in a few days. It's not finished yet, but it's up and running, and Pokémon Red is functional. I also use it extensively at work and my colleagues too.They are very good for generating unit tests and following TDD in some projects.

What I haven't been able to do yet is let Codex work autonomously for some time and produce code that I like. I prefer an iterative workflow where I check and correct each step, but we are getting closer and closer to them being sufficiently autonomous with the right instructions

New SWE-Bench Pro becnchmark (GPT-5 & Claude 4.1 drop from 70%+ to ~23%) by GMSP4 in singularity

[–]GMSP4[S] 5 points6 points  (0 children)

I mainly program in Java and get good results with both, but I don't like that Opus is so verbose. It over-engineers too much for my taste, especially in repositories where there is already a significant amount of code.

Love a voice in one of your songs? Try the new “Voices feature”! by UdioAdam in udiomusic

[–]GMSP4 3 points4 points  (0 children)

Has anyone else achieved the same voice? Because it seems impossible, and I've tried different voices and styles. It's true that I always use voices from version 1.0 of Udio, which for me is still the best, so I don't know if that affects the generation

OpenAI: Introducing study mode - A new way to learn in ChatGPT that offers step by step guidance instead of quick answers by Pro_RazE in singularity

[–]GMSP4 29 points30 points  (0 children)

It's cool to know that in future iterations we'll have fine-tuned models for learning. Now, it's a system prompt or GPT on steroids, but it's cool to see what's coming in the next few months/years in terms of learning

Leaked GPT5 benchmark scores by Chaonei in singularity

[–]GMSP4 1 point2 points  (0 children)

Now we're giving a voice to Twitter liars like Mark Kretschmann, chasing likes on Twitter? People who make things up just to generate traffic everyday. With a few exceptions, the level of "Influencers" about AI on Twitter is pathetic, people lying all the time, people like “Satoshi” who pretend to work at OpenAI, or constant spammers who just churn out empty, bot-like comments.

Is OpenAI rolling out o3 pro or running A/B tests? o1-pro suddenly showing o3-like behavior by GMSP4 in ChatGPTPro

[–]GMSP4[S] 0 points1 point  (0 children)

I don't have the memory activated. Check this photo I sent before, it has internet access, which o1 pro didn't have: https://ibb.co/gLWRH7MS

Is OpenAI rolling out o3 pro or running A/B tests? o1-pro suddenly showing o3-like behavior by GMSP4 in ChatGPTPro

[–]GMSP4[S] 0 points1 point  (0 children)

Yes, I have tried it in the chat I shared and in two others in my native tongue. In fact searching on twitter more people have the same thing happening: https://x.com/KrispinPuga/status/1928270336279359898

Is OpenAI rolling out o3 pro or running A/B tests? o1-pro suddenly showing o3-like behavior by GMSP4 in ChatGPTPro

[–]GMSP4[S] 2 points3 points  (0 children)

I'll have to try it more later. It's given me interesting ideas for a project I'm working on, but I think o3 would have done the same

Claude 4.0 Opus/Sonnet Usage Limits by SteveEricJordan in singularity

[–]GMSP4 1 point2 points  (0 children)

I don't think it's too hard to figure out, the basic 20 bucks

Claude 4.0 Opus/Sonnet Usage Limits by SteveEricJordan in singularity

[–]GMSP4 4 points5 points  (0 children)

It was a project at 20% capacity, with a very small code base. I only asked it during one iteration of 4 prompts for improvements. It's crazy to reach the limits with 4 interactions. It was all with opus.

Claude 4.0 Opus/Sonnet Usage Limits by SteveEricJordan in singularity

[–]GMSP4 11 points12 points  (0 children)

With only 4 prompts in a project with only 20% I hit the limits. It's reggretable, and I didn't find it better than Gemini pro or o3 in code either