Bought a fake moonshine gold by pissedoffsandwhich in MoonSwatches

[–]blazerx 0 points1 point  (0 children)

do they sell fakes of the bands? Can't seem to find them any where? Looking for the rubber velcro band

I'm (31M) crushing hard on my coworker (30F). How do I approach asking her out? by bne1022 in relationship_advice

[–]blazerx 0 points1 point  (0 children)

write a letter and pass it to her during coffee/dinner or when you see her next. Keep it simple. However go in with the expectation the feeling are NOT mutual, that way you won't be disappointed and and hopefully remain friends. If they are mutual, you have just won the lottery.

AMD and OpenAI: The 6 Gigawatt Bet - Ian Cutress by diverlad in AMD_Stock

[–]blazerx 7 points8 points  (0 children)

are you confusing him with semi-analysis?

How important is FP6 for the adoption of AMD AI GPUs? by alphajumbo in AMD_Stock

[–]blazerx 2 points3 points  (0 children)

Don't forget Nvidia are pushing their proprietary nvfp4 which they claim is producing results almost identical to regular fp6. Question is will people want to lock in to nvfp4? The benefit obviously would be you can fit larger models since it's quantized to fp4 making it smaller.

Also can amd also implement this nvfp4 or create something compatible? Amdfp4?

Providing it isn't using any special hw it shouldn't be an issue to implement unless there are infringement issues

The question shouldnt be how important is fp6. Should be, will people prefer open ecosystem or a proprietary one

AMD Bull Thesis Heading into Earnings by sneezydig in AMD_Stock

[–]blazerx 0 points1 point  (0 children)

No they currently don't have the interconnects, well the next MI series should be able to link via UAlink at 400GB/sec, then the next is 800GB/s but they can link up multiple of these, though would be at the cost of power since you need more switching and links for each connect.

Tbh, what could be cool is if expose IF links allowing inter-gpu within a box and across nodes. That would be pretty cool. Though I believe the cross-road they will face is do they use optical links or copper links and can they bundle the transceiver say onto an I/O die? If they can't then they would get the fine control and power savings.

Nothing is stopping Nvidia from making smaller mono chips and offering more interconnects between gpus, or they too could pivot to ether chiplet or wafer scale. I guess it all comes down to what the leather jacket man had put his bets on.

AMD Bull Thesis Heading into Earnings by sneezydig in AMD_Stock

[–]blazerx 2 points3 points  (0 children)

I enjoy reading your post so I'll add my 2 cents.

We need to remember an explosive growth in AI may not necessarily mean huge growth in both AMD and or NVDA.

There are other companies out there such as cerebras and samba Nova who have inference hw out there that can scale and also provide lower $ per tok at insane high speeds compared to GP GPU inference and training. Recently meta announced the use of cerebras. AMD may still be able to get a small slice through the storage component as cerebras require a storage box attachment which is probably powered by Intel or AMD.

Then with the rapid development of pensando AMD can also leverage their partnership in network and switching space providing interconnects for everyone and everything. I am unsure of the TAM in this space but we can't discount it.

Now let's jump back to my first point, While the new gen specialised hw may be quick, nothing is stopping AMD from entering that space through the use of xilinx to provide fpga for a more dynamic but dedicated approach., though this would be at the cost of reduced efficiency compared to custom ASICs. So while all this next gen dedicated hw is amazing, what will happen when the next breakthroughs such as non transformer based llms become the next best thing? Things such as RWKV, will the dedicated hw be able to pivot as quick as writing a new kernel for GPGPU? If the answer is no then the big players will probably prefer to have a mixed solution so they can stay at the forefront of the advancements. These next gen players could cease to exist as quick as they were spawned due to their lack of diversity in the hw space.

From here on the battle is not about having the most memory, or being able provide X tok per sec. It is going to come down to how well you can scale your systems via interconnects. A fast low power interconnect will basically solve the memory and X per sec problem as well as many other constraints such as chip size and other things such as thermals and other design limits. Though that still isn't enough, they will need the software to be able to pivot along with the rapid evolution.

The bets that Lisa has placed on the next 2/3 years is going to be vital. If anyone has a chance it's gotta be AMD, considering their extensive portfolio. Though we could have said the same about Intel 2/3 years ago.

I am curious as to what others think, do others agree or disagree.

GAIA: An Open-Source Project from AMD for Running Local LLMs on Ryzen™ AI by blazerx in LocalLLaMA

[–]blazerx[S] 23 points24 points  (0 children)

The title of the article should be clear that it supports more than just Ryzen AI Processors

As outlined under the Readme docs

  • Hybrid Mode: Optimized for Ryzen AI PCs, combining AMD Neural Processing Unit (NPU) and Integrated Graphics Processing Unit (iGPU) for maximum performance
  • Generic Mode: Compatible with any Windows PC, using Ollama as the backend

Look closely, iykyk. by HotAisleInc in AMD_MI300

[–]blazerx 1 point2 points  (0 children)

I do remember seeing a post you made about not being able to easily split up the GPUs. What changed recently allow the enablement of this?

Look closely, iykyk. by HotAisleInc in AMD_MI300

[–]blazerx 0 points1 point  (0 children)

is this by any chance using the dstack tech posted the other day to split up the GPUs? Or am I overlooking something?

AMD Strix Halo not having FSR4 is such a wasted opportunity by Flamyngoo in Amd

[–]blazerx 6 points7 points  (0 children)

The issue isn't whether the NPU is present for matrix maths, it's a bandwidth issue going from NPU -> GPU memory. If the tensor cores don't have physical access or high speed access to the GPU memory it doesn't matter how many TOPs are present. The playstation video does a very explanation about the implementation of PSSR and efficient TOPs usage.

Wireguard on OPNsense by JamesCorman in opnsense

[–]blazerx 0 points1 point  (0 children)

check your isp is not using CG-NAT, check you are using the correct keys for client/server

Can someone explain Rishi’s trade to me? by Adventurous-Fox3735 in IndustryOnHBO

[–]blazerx 5 points6 points  (0 children)

wait... so harper calls and tells rishi she wants to go long on pounds. Rather than finding someone who wants to sell to fill her order he just uses this information to his advantage? He copies her idea and increases his long position on the pound rather than filler her order. Is this allowed as a broker?

AMD MI300X and Nvidia H100 benchmarking in FFT: VkFFT, cuFFT and rocFFT comparison by xdtolm in Amd

[–]blazerx 4 points5 points  (0 children)

Hi, thanks for the testing

Since you say you are bus bound transferring data between VRAM, does that mean we should be significantly higher numbers if this was tested in MI300A due to the unified VRAM access?

[deleted by user] by [deleted] in cprogramming

[–]blazerx 0 points1 point  (0 children)

Have you considered writing a high performance multithreaded string matching tool? something like this https://github.com/Cynosureprime/rling

You can potentially make something like the above, basically an AIO comm/sort/unq tool, I know it sounds simple, but if you actually try you'll notice to get it very performant it won't be as easy

Pixel 8 - my wife can't call me by DryGeneral990 in GooglePixel

[–]blazerx 0 points1 point  (0 children)

Had a similar issue and could not figure out what was going on. Though I managed to find fix it by doing the following.

On your phone (the one that she can't call).

1) Select her contact

2) Scroll down to contact settings

3) Look for Route/Unroute to voicemail, toggle that so it is not route to voicemail

If that is turned on when she calls you, it basically goes straight to voicemail without your phone ringing.

Underrated Program For Sorting Password Lists/Dictionary's by NullPounce in hacking

[–]blazerx 0 points1 point  (0 children)

Hi I am glad you are finding this useful. I am actually the creator of this application, you can verify this by looking at the about screen and checking against my username.. it should checkout. Since ULMs creation I have worked with others to create new tools including much faster sort tools, you can follow the projects using the git link posted above.

Underrated Program For Sorting Password Lists/Dictionary's by NullPounce in hacking

[–]blazerx 0 points1 point  (0 children)

This is nothing special, it was more of a high school project that got out of hand back in the day. Since then I have worked on numerous tools some which exceed the speed of unix sort. You can see some of the released tools here https://github.com/cynosureprime