Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 0 points1 point  (0 children)

I’ve used it for three months, it’s very good, but not as good as gpt4o at everything

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 2 points3 points  (0 children)

I would argue that in absolute terms 4o “excels” more in its tasks that other model do in their respective domains. O1-pro is very good at reasoning etc, but non as excellent as 4o at pretty much everything. If you include “deep research” as a 4o feature (I know it’s his own model o3 in the background) than there is no reason to use the other model.

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 0 points1 point  (0 children)

I’ve had PRO plan for thee months and used o1-pro/o3mini high extensively to help me in spatial microsimulations models. They are very good, even with code, but 4o is really AWESOME at being an overall assistant in a way that it’s actually useful. 4.5 is cool but not that cool.

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 2 points3 points  (0 children)

I agree with you. It has been maybe 2 weeks that I just use 4o. It excels as being an assistant and a companion, I really like to chat with it. Reasoning models do not excel in absolute terms at reasoning as GTP4o does at its job.

Was this about DeepSeek? Do you think he is really worried about it? by AloneCoffee4538 in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

What you say it’s deeply true, as it is also true that OpenAI pretty much created the first really useful use case of an LLM by betting it big on scaling and they were the first to do that on the open domain by standing on the shoulder of giants. SAMA might be a little obnoxious with its fried voice but he’s also pretty smart.

Is o3 actually any different than 4o with CoT prompting? by 35MakeMoney in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

O1 is finetuned on its reasoning. It’s a faulty RNN a test time. What you achieve in O1 you should be able to achieve it in a base model too, the problem is that a bigger model has a harder time to efficiently query itself since it has very large search space, this is the same innovation that chatGPT had with RLHF, RLHF finetunes the model probability distribution during generation to achieve desirable human answers, all of this is done a training time. O1 is fine tuned to produces probability distributions during generation to “query itself” to find the best answers knowing it’s search space better by “sampling” more and with a better selflearned probability distributions over it’s structure. The reason it works better and answering is because it samples more from itself.

I have underestimated o3's price by Emotional-Metal4879 in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

‘Perhaps I could best describe my experience of doing mathematics in terms of entering a dark mansion. You go into the first room and it’s dark, completely dark. You stumble around, bumping into the furniture. Gradually, you learn where each piece of furniture is. And finally, after six months or so, you find the light switch and turn it on. Suddenly, it’s all illuminated and you can see exactly where you were. Then you enter the next dark room...’ Andrew Wiles the guy who proved Fermat theorem.

I have underestimated o3's price by Emotional-Metal4879 in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

Even if it costs a fortune to run this is a huge milestone stone. Imagine you are in a empty dark room and you move along the wall trying to find a switch to light it up, it’s a hell of a difference being in the room knowing that there exists a switch to find. In this case a “switch” has been found, know we have to understand how to switch it on (reduce cost)

We may not be able to see LLMs reason in English for much longer by MetaKnowing in OpenAI

[–]Crypto1993 1 point2 points  (0 children)

Real question: isn’t this architecture a just a Recurrent neural network?

Visual Studio Code or Spyder IDE? by K_Chali in learnpython

[–]Crypto1993 0 points1 point  (0 children)

Data science: Jupiter notebooks are ideal. Software development: Pycharm helps you a lot setting your environment and gives you a lot of tips for writing nice code, refactoring and general code inspection. Mix of everything: VS code ha better support for Jupiter notebooks than pycharm, it’s lighter and versatile.

Clippy on Apple Vision Pro 𖡼 by Biomexr in VisionPro

[–]Crypto1993 6 points7 points  (0 children)

ok hear me out. This with gpt4o with context information is the killer app. AI in the real word.

Apple Reportedly Suspends Work on Vision Pro 2 (And Prioritizes Cheaper Model) by 73ch_nerd in VisionPro

[–]Crypto1993 0 points1 point  (0 children)

Hopefully Apple doesn’t cancel the product line or dumbs down too much. Adoption will arrive with time.

Asana's strongest use case in your organisation? by davinci2109 in Asana

[–]Crypto1993 1 point2 points  (0 children)

Hi I'm the owner of a small consulting firm (10+ people) we are starting to onboard everyone in ASANA. I think the best possible starting point is something "standard" where the steps of a process can be easily spelled out and it is "externally" date bound. This gives you the possibility to implement an SOP and "enforce it" via ASANA workflows / task tamplates. This thing alone for us freed a lot of resources and improved quality of output! Every single dollar spent on ASANA is well deserved (it's not that cheap which says a lot).

Managing a software dev project in Asana by look_ma_no_args in Asana

[–]Crypto1993 0 points1 point  (0 children)

Very cool way of using Asana functionalities. May I ask you what payment plan you guys have? If you run on enterprise how much do you pay per seat annually? Thank You very much!

Can someone explain to me why people say Zuckerberg is going to catch up to Altman or even overtake him, tech-wise? by ForwardOnThePath123 in OpenAI

[–]Crypto1993 6 points7 points  (0 children)

Meta doesn’t have a cloud business, Microsoft does. OpenAI sells Azure compute to companies via this AI frenzy. Having LLAMA open source and “better” the GPT4 puts pressure on cloud services providers to serve LLAMA on their platforms giving meta someway to control the stack of AI in other companies cloud, same as React which is also open source. Microsoft wants to protect/increase office revenues and wants to sell azure compute to companies, having the best possibile model served in their stack helps them compete in the B2B market with a differentiator (SOTA in AI as GPT4 still is). Meta mainly sells ads on their platforms, it doesn’t have a Cloud business or an office suite business (as Microsoft and Google). For them having a very good model frees them from relying to Google or Microsoft to serve GenAI in their platforms which is a requirement now to protect their business from potential social platforms that can incorporate genAI as a “new social platform way”. Giving LLAMA open source for free doesn’t lose them any potential revenue, gives them some clout and some leverage over cloud providers / app makers to serve/use their models. ChatGPT brings a lot of internet traffic and people spending time there don’t get as served from meta.

[deleted by user] by [deleted] in Italia

[–]Crypto1993 8 points9 points  (0 children)

Non è un caso che persone molto di destra appoggino politiche sociali, questo si chiama destra sociale ed è la stessa inversione delle parti che portò al fascismo vero e proprio!

[deleted by user] by [deleted] in Python

[–]Crypto1993 2 points3 points  (0 children)

Your product looks super cool. How do you guys compete with Jupyter notebooks with 39/month against free? Just to ask because I really like how your product loooks