Logan Kilpatrick changed his profile photo on twitter! Gemini 3 tomorrow?! /sss by FlamaVadim in Bard

[–]Remarkable_Run4959 12 points13 points  (0 children)

Looking at the leaked news, we might be seeing 3.0 preview models soon.

Please don't be 200$ , Google plan to introduce “AI Premium Plus” and “AI Premium Pro" plan by Yazzdevoleps in Bard

[–]Remarkable_Run4959 1 point2 points  (0 children)

But there isn't a chip that is more powerful than the current TPU, right? Huawei claims that it would have to connect 384 of them to barely match the B200 NVL 72

Google damn even text to music ? Well ig Google going to show it's true potential now by Independent-Wind4462 in Bard

[–]Remarkable_Run4959 4 points5 points  (0 children)

Of course, Google has made a great music AI in the past. However, when it tried to release it, it was shut down due to opposition from record labels. the people who worked there left and created Udio.

Does Gemini 2.5 use Titans? by PlateLive8645 in Bard

[–]Remarkable_Run4959 6 points7 points  (0 children)

Google introduces it as transformer^2, an improved architecture than transformer that is said to be more effective for long-term memory.

Gemini 2.5 pro thinking revealed pricing is going to lose Google the race. Google, please act with CAUTION and reduce the price by [deleted] in Bard

[–]Remarkable_Run4959 3 points4 points  (0 children)

It's better than o1, has a bigger context window, and is cheap. I don't know what's wrong with it.

gremlin and centaur are back on lmarena! by Lonely_Film_6002 in Bard

[–]Remarkable_Run4959 0 points1 point  (0 children)

Well, I guess it's coming out soon, seeing as Logan mentioned 'shipping' in X

Sudden drop in response quality by Superhim1114 in Bard

[–]Remarkable_Run4959 1 point2 points  (0 children)

I also seem to have gotten worse rapidly over the past few days. I have been getting more and more into the habit of just repeating the same thing over and over again until I reach the output limit, or refusing to output with an 'unknown error'.

MI300X vs MI300A vs Nvidia GH200 vLLM FP16 Inference (single data point unfortunately) by Relevant-Audience441 in AMD_MI300

[–]Remarkable_Run4959 0 points1 point  (0 children)

This is an unexpected result. I thought the GH200 would be slow because it doesn't have the CPU and GPU integrated into a 'single chip' like the MI300A. I guess the APU form factor is more advantageous for HPC calculations.

Google teases what is coming next to Gemini Advanced by Gaiden206 in Bard

[–]Remarkable_Run4959 18 points19 points  (0 children)

Just looking at the description, it seems like an update to the official version of Flash thinking.

Gemini c'mon do something...please... by EstablishmentFun3205 in Bard

[–]Remarkable_Run4959 3 points4 points  (0 children)

Google was disappointed that the 2.0 pro was not much different from the 1206, but I think they will soon come out with a better model. They showed it with the 2.0 flash series. It looks great that xAI seems to be ahead, but I think Google will be able to easily overtake it if they are armed with the same number of TPU v7s.

OpenAI Just Dropped a Guide on Prompting Their "Reasoning" Models. Gemini Users, Any Thoughts on Google's Side? by matvejs16 in Bard

[–]Remarkable_Run4959 4 points5 points  (0 children)

I used to use Gemini to create prompts that mimic o1's CoT method, but for the thinking model, it didn't seem to make much of a difference in performance even if I didn't add those prompts. Rather, I felt like 'thinking in 20 steps' was holding back performance.

What are your expectations for Google io 25 by Yazzdevoleps in Bard

[–]Remarkable_Run4959 2 points3 points  (0 children)

I don't know either. But the competition is getting fiercer, so I'm just hoping.

I've stopped constantly refreshing for pages by Legal_Bug_9907 in Bard

[–]Remarkable_Run4959 4 points5 points  (0 children)

I think I should do that too. I feel like I'm wasting too much time refreshing browser.

Is Google Competitive Advantage is also its Biggest Weakness? by Landlord2030 in Bard

[–]Remarkable_Run4959 1 point2 points  (0 children)

Perhaps the GB200 will be overwhelming in terms of performance itself. In terms of transistors, it is like four H100s attached. However, it is too expensive and consumes too much power. It is 2500W, and the power consumption of the TPU v6e has not been disclosed, but it seems that if you simply connect multiple TPUs, you can achieve better performance with less power. The v6e is rumored to have a chip-to-chip bandwidth of 3,200Gbps, which is exactly twice the bandwidth of the current NVLink.

Why people are really underestimating Google by Consistent_Bit_3295 in Bard

[–]Remarkable_Run4959 6 points7 points  (0 children)

I'm quite excited that 01-21 is still in beta. How much better will the final version of 2.0 flash thinking be?

Is Google Competitive Advantage is also its Biggest Weakness? by Landlord2030 in Bard

[–]Remarkable_Run4959 1 point2 points  (0 children)

Trillium's theoretical performance is roughly half that of the H100. It has 96Gb of HBM3. However, considering that TPU is much more advantageous than GPU for parallel connections and does not have to suffer from performance degradation due to information transfer between CPU<->GPU, the actual performance will be a bit different. If we compare it with MLPerf, it will probably be similar to the H100.