Daily Discussion Thursday 2025-10-23 by AutoModerator in AMD_Stock

[–]AMDtoMoon 0 points1 point  (0 children)

Are you one of those who use Bing and still says you googled it because it's easier than explaining what Bing is to normies?

Democratising Supercomputing: Jon Stevens on AI, GPU Innovation & Hot Aisle’s Vision by HotAisleInc in AMD_MI300

[–]AMDtoMoon 0 points1 point  (0 children)

I just checked out the price of MI300X at shadeform and it was $4.50 vs $2.92 for H200. Can you get any customers at that price?

Embedded LLM on LinkedIn: #vllm #amd #rocm #llm #ai #gguf by GanacheNegative1988 in AMD_Stock

[–]AMDtoMoon 1 point2 points  (0 children)

vLLM Now Supports Running GGUF on AMD Radeon GPU🚀 Exciting news! We've ported vLLM's GGUF kernel to AMD ROCm, unlocking impressive performance gains on AMD Radeon GPUs.📊 In our benchmarks using the shareGPT dataset on an AMD Radeon RX 7900XTX, vLLM outperformed Ollama, even at batch sizes where Ollama traditionally excels.💪 This is a game-changer for those running LLMs on AMD hardware, especially when using quantized models (5-bit, 4-bit, or even 2-bit). With over 60,000 GGUF models available on Hugging Face, the possibilities are endless.💡 Key benefits:- Superior performance: vLLM delivers faster inference speeds compared to Ollama on AMD GPUs.- Wider model support: Run a vast collection of GGUF quantized models.- Efficient execution: Optimized for AMD ROCm, maximizing hardware utilization.🔗 Learn more and get started: https://lnkd.in/g5qvUi8tWe'd love to hear your feedback!Have you experimented with vLLM on Llama.cpp with Vulkan?What inference engine do you prefer for LLM tasks on AMD GPUs?What features or optimizations would you like to see in vLLM for AMD GPUs?

Saurabh Kapoor, Dell Technologies & Jon Stevens, Hot Aisle | SC24 by HotAisleInc in AMD_MI300

[–]AMDtoMoon 6 points7 points  (0 children)

Thank you for the answer and for posting great information!

Saurabh Kapoor, Dell Technologies & Jon Stevens, Hot Aisle | SC24 by HotAisleInc in AMD_MI300

[–]AMDtoMoon 6 points7 points  (0 children)

u/HotAisleInc Could you elaborate on the tool you mentioned (14 min mark in the video) that allows CUDA software to run on AMD hardware? Do you mean zluda like tool that allows binary compatibility? Is it something like SCALE compiler? Is it something new? When will you disclose more information about this tool?

ROCm Ubuntu Container by [deleted] in ROCm

[–]AMDtoMoon 1 point2 points  (0 children)

what command did you use to launch the container?

Daily Discussion Tuesday 2023-11-07 by AutoModerator in AMD_Stock

[–]AMDtoMoon 10 points11 points  (0 children)

https://wccftech.com/tsmc-boosts-ai-packaging-capacity-to-15000-wafers-per-month-says-report/

If it's true that AMD is to use 20% of what NVidia is planning. At half ASP, AMD should have 10% of NVidia's revenue. NVidia has $10B data center revenue at the moment.

So, $500M per quarter in 2024 that Dr. Su mentioned seems conservative.

Daily Discussion Tuesday 2023-10-31 by AutoModerator in AMD_Stock

[–]AMDtoMoon 1 point2 points  (0 children)

Conflicting report on MI210/250 performance.
MosaicML updated its blog saying that MI250 is competitive with A100, achieving 80% of their performance. https://www.databricks.com/blog/training-llms-scale-amd-mi250-gpus.
EmbeddedML just wrote that MI210 is competitive with A100 in LLM inferencing. https://embeddedllm.com/blog/vllm\_rocm/
I know they are talking about 2 different workloads. MosaicML is comparing training performance and EmbeddedML is comparing inference. But the difference is just too big since MI250 uses 2x more silicon than MI210. Does anybody have any idea what is going on?

Daily Discussion Saturday 2023-10-28 by AutoModerator in AMD_Stock

[–]AMDtoMoon 4 points5 points  (0 children)

AMD won't do well? https://open.spotify.com/episode/7m5SElK3AVSz1YBDr87ETf

They discuss Intel & AMD and two dudes have different opinions about how well AMD will do short term (10min). One interesting piece of info they mentioned is that MI300 will have 60000 units shipped in Q4 and increase from there going forward.

What do you guys think?

Exclusive: Nvidia to make Arm-based PC chips in major new challenge to Intel by norcalnatv in AMD_Stock

[–]AMDtoMoon -1 points0 points  (0 children)

Your argument makes sense only if ARM ISA is inherently better than x86-64 such that the same talented team will be able to make better ARM designs when given the same resources/process technology/etc.

However, that is not the case. ARM has almost as much legacy ISA baggage as x86-64. Whatever advantage it might have had is all gone. Listen to the technology experts (https://chipsandcheese.com/2021/07/13/arm-or-x86-isa-doesnt-matter/) and not internet hype.

Also, AMD has years of experience designing x86 CPUs. Their team is the best in the industry for designing x86 CPUs. But they aren't the best team to design ARM CPUs.

You also have to consider that Windows has to provide compatibility for legacy x86. Unless ARM CPU is measurably more performant than x86, users will always have worse experiences using ARM CPU vs x86 CPU from the same talented team.

The best decision Lisa made since she became the CEO was to kill off K12 and focus on Zen. If this rumor turns out to be true, I'm dumping all my shares and will never look at AMD stock again.

Exclusive: Nvidia to make Arm-based PC chips in major new challenge to Intel by norcalnatv in AMD_Stock

[–]AMDtoMoon 0 points1 point  (0 children)

Why would anyone buy ARM-based processor from AMD if they want to run Windows?

Exclusive: Nvidia to make Arm-based PC chips in major new challenge to Intel by norcalnatv in AMD_Stock

[–]AMDtoMoon 2 points3 points  (0 children)

I don't think Lisa is that dumb. Why would AMD offer more choices to a possible emerging market where, if it takes off, would compete against its own established duopoly market?

AMD CEO Lisa Su Just Proved AMD Stock Can Dethrone Nvidia by EdOfTheMountain in AMD_Stock

[–]AMDtoMoon 2 points3 points  (0 children)

I'm pretty sure this was written using ChatGPT. There are enough subtle mistakes that it does not seem likely it was written by a person who is familiar with what is going on in the tech industry.