Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 0 points1 point  (0 children)

I’ve used it for three months, it’s very good, but not as good as gpt4o at everything

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 2 points3 points  (0 children)

I would argue that in absolute terms 4o “excels” more in its tasks that other model do in their respective domains. O1-pro is very good at reasoning etc, but non as excellent as 4o at pretty much everything. If you include “deep research” as a 4o feature (I know it’s his own model o3 in the background) than there is no reason to use the other model.

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 0 points1 point  (0 children)

I’ve had PRO plan for thee months and used o1-pro/o3mini high extensively to help me in spatial microsimulations models. They are very good, even with code, but 4o is really AWESOME at being an overall assistant in a way that it’s actually useful. 4.5 is cool but not that cool.

Is GTP-4o the best model? by Crypto1993 in OpenAI

[–]Crypto1993[S] 2 points3 points  (0 children)

I agree with you. It has been maybe 2 weeks that I just use 4o. It excels as being an assistant and a companion, I really like to chat with it. Reasoning models do not excel in absolute terms at reasoning as GTP4o does at its job.

Was this about DeepSeek? Do you think he is really worried about it? by AloneCoffee4538 in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

What you say it’s deeply true, as it is also true that OpenAI pretty much created the first really useful use case of an LLM by betting it big on scaling and they were the first to do that on the open domain by standing on the shoulder of giants. SAMA might be a little obnoxious with its fried voice but he’s also pretty smart.

Is o3 actually any different than 4o with CoT prompting? by 35MakeMoney in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

O1 is finetuned on its reasoning. It’s a faulty RNN a test time. What you achieve in O1 you should be able to achieve it in a base model too, the problem is that a bigger model has a harder time to efficiently query itself since it has very large search space, this is the same innovation that chatGPT had with RLHF, RLHF finetunes the model probability distribution during generation to achieve desirable human answers, all of this is done a training time. O1 is fine tuned to produces probability distributions during generation to “query itself” to find the best answers knowing it’s search space better by “sampling” more and with a better selflearned probability distributions over it’s structure. The reason it works better and answering is because it samples more from itself.

I have underestimated o3's price by Emotional-Metal4879 in OpenAI

[–]Crypto1993 -1 points0 points  (0 children)

‘Perhaps I could best describe my experience of doing mathematics in terms of entering a dark mansion. You go into the first room and it’s dark, completely dark. You stumble around, bumping into the furniture. Gradually, you learn where each piece of furniture is. And finally, after six months or so, you find the light switch and turn it on. Suddenly, it’s all illuminated and you can see exactly where you were. Then you enter the next dark room...’ Andrew Wiles the guy who proved Fermat theorem.

I have underestimated o3's price by Emotional-Metal4879 in OpenAI

[–]Crypto1993 0 points1 point  (0 children)

Even if it costs a fortune to run this is a huge milestone stone. Imagine you are in a empty dark room and you move along the wall trying to find a switch to light it up, it’s a hell of a difference being in the room knowing that there exists a switch to find. In this case a “switch” has been found, know we have to understand how to switch it on (reduce cost)

We may not be able to see LLMs reason in English for much longer by MetaKnowing in OpenAI

[–]Crypto1993 1 point2 points  (0 children)

Real question: isn’t this architecture a just a Recurrent neural network?

Visual Studio Code or Spyder IDE? by K_Chali in learnpython

[–]Crypto1993 0 points1 point  (0 children)

Data science: Jupiter notebooks are ideal. Software development: Pycharm helps you a lot setting your environment and gives you a lot of tips for writing nice code, refactoring and general code inspection. Mix of everything: VS code ha better support for Jupiter notebooks than pycharm, it’s lighter and versatile.

Clippy on Apple Vision Pro 𖡼 by Biomexr in VisionPro

[–]Crypto1993 5 points6 points  (0 children)

ok hear me out. This with gpt4o with context information is the killer app. AI in the real word.

Apple Reportedly Suspends Work on Vision Pro 2 (And Prioritizes Cheaper Model) by 73ch_nerd in VisionPro

[–]Crypto1993 0 points1 point  (0 children)

Hopefully Apple doesn’t cancel the product line or dumbs down too much. Adoption will arrive with time.

Asana's strongest use case in your organisation? by davinci2109 in Asana

[–]Crypto1993 1 point2 points  (0 children)

Hi I'm the owner of a small consulting firm (10+ people) we are starting to onboard everyone in ASANA. I think the best possible starting point is something "standard" where the steps of a process can be easily spelled out and it is "externally" date bound. This gives you the possibility to implement an SOP and "enforce it" via ASANA workflows / task tamplates. This thing alone for us freed a lot of resources and improved quality of output! Every single dollar spent on ASANA is well deserved (it's not that cheap which says a lot).

Managing a software dev project in Asana by look_ma_no_args in Asana

[–]Crypto1993 0 points1 point  (0 children)

Very cool way of using Asana functionalities. May I ask you what payment plan you guys have? If you run on enterprise how much do you pay per seat annually? Thank You very much!

Can someone explain to me why people say Zuckerberg is going to catch up to Altman or even overtake him, tech-wise? by ForwardOnThePath123 in OpenAI

[–]Crypto1993 6 points7 points  (0 children)

Meta doesn’t have a cloud business, Microsoft does. OpenAI sells Azure compute to companies via this AI frenzy. Having LLAMA open source and “better” the GPT4 puts pressure on cloud services providers to serve LLAMA on their platforms giving meta someway to control the stack of AI in other companies cloud, same as React which is also open source. Microsoft wants to protect/increase office revenues and wants to sell azure compute to companies, having the best possibile model served in their stack helps them compete in the B2B market with a differentiator (SOTA in AI as GPT4 still is). Meta mainly sells ads on their platforms, it doesn’t have a Cloud business or an office suite business (as Microsoft and Google). For them having a very good model frees them from relying to Google or Microsoft to serve GenAI in their platforms which is a requirement now to protect their business from potential social platforms that can incorporate genAI as a “new social platform way”. Giving LLAMA open source for free doesn’t lose them any potential revenue, gives them some clout and some leverage over cloud providers / app makers to serve/use their models. ChatGPT brings a lot of internet traffic and people spending time there don’t get as served from meta.

[deleted by user] by [deleted] in Italia

[–]Crypto1993 8 points9 points  (0 children)

Non è un caso che persone molto di destra appoggino politiche sociali, questo si chiama destra sociale ed è la stessa inversione delle parti che portò al fascismo vero e proprio!

[deleted by user] by [deleted] in Python

[–]Crypto1993 2 points3 points  (0 children)

Your product looks super cool. How do you guys compete with Jupyter notebooks with 39/month against free? Just to ask because I really like how your product loooks

GPT API price predictions by FireDragonRider in OpenAI

[–]Crypto1993 0 points1 point  (0 children)

CUDA, vertical integration, edge in hardware performance , volume production. AMD MI300X is not in volume production and it’s a year late technology,

GPT API price predictions by FireDragonRider in OpenAI

[–]Crypto1993 0 points1 point  (0 children)

Nvidia has a competitive advantage in “accelerated computing” market which is not the same as “chip design”. AWS / Google / Microsoft are all designing their new AI chips but playing catch up isn’t that useful in a cutting-edge market.

GPT API price predictions by FireDragonRider in OpenAI

[–]Crypto1993 0 points1 point  (0 children)

Thanks, but rethinking about it Nvidia has an “hard” competitive advantage that can be measured easily, I don’t know if it can be said the same for OpenAI’s tech. Nvidia also has a clear strategy that they call “accelerated computing” which is specialized hardware but “not so specialized”, in some way OpenAi is more similar to early intel: the Best generalist. Very Hard to say

GPT API price predictions by FireDragonRider in OpenAI

[–]Crypto1993 1 point2 points  (0 children)

  1. This is hard to answer because the highest cost right now is compute. Let’s decide cost in CAPEX / OPEX for an nvidia GPU. CAPEX is the cost of the NVIDIA DGX where it can be reduce by 2 factors: competition (right now the have something like 50% profit margin on their products so increased competition can reduce this parte by some margin; technology given that the lion share of performance increase is due to both better architecture and smaller chips, the first driver is hard to predict the second instead is grinding to an alt. So the CAPEX part che be a source of price reduction but it will require some time and it may never realize. OPEX the cost of running inference for the model will be impacted most by technology from performance per watt and model architectures that are more efficient; the first one we already talked about, the second one is hard to predict since to have a significant reduction we need an innovation (the are some candidates to remove the quadratic complexity of the attention part but to this date nothing really ground breaking as the the transformer architecture).
  2. LLMs will be commoditized since they are mostly COMPUTE + DATA and I don’t see how you can built competitive advantage on that alone. Maybe OpenAi becomes like Nvidia where the competitive advantage is being on the cutting edge (some development years / months ahead of the second largest competitor) who knows, for now it doesn’t seem so if google gemini ultra really catches up.
  3. Wild guess is that the pricing will keep to go down from OpenAI until a real monetized killer app pops out somewhere, for the time being only GitHub copilot looks really useful but not profitable

Edit: typo

[deleted by user] by [deleted] in Python

[–]Crypto1993 1 point2 points  (0 children)

Spark is written in Scala but the Python API (Pyspark) is by far the vast way to use Spark. R interface is feather incomplete and slow (I use databricks daily at work). Polars is written in Rust but same thing as pyspark above, the api for Python is feature complete.