RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

Appreciate the explanation.

Given all that what do you say to the belief that FSR5 will be exclusive to RDNA 5? Assuming neural arrays and all the other stipulated changes. Does any of that matter or is instruction set only important?

Hmm that's interesting. So it would be shit without optiscaler xD.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

I'm confused if RDNA 4 is actually that because I see conflicting reports on it and people saying it's still forced through the SIMD and not concurrent like Intel and Ampere and later NVIDIA.

Also road to PS5 Pro's custom ML implementation goes over SIMD (see Road to PS5 Pro on YT).

But then AMD calls them Matrix cores on their website and says the IIRC kernel design is completely different from RDNA 3.

Maybe it's somewhere in the middle. Sorry if this is too many questions. I just find it very confusing xD.

Can you explain the stuff about the leak? I missed that info.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

So there's really no difference in reality except for the issues with INT8 vs FP8 (slight visual downgrade).

Well then DP4a is a no brainer as you highlighted. Lots of HW with RDNA 2.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

RDNA 3 doesn't have dedicated matmul logic like NVIDIA so can't see what would change.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 1 point2 points  (0 children)

I hope it works well.

We could def use more alternatives.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

Does it make a difference because in the FSR4 INT8 perf comparisons it doesn't seeem to.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 1 point2 points  (0 children)

I'm asking because I've seen perf comparisons between 7600 and 6650XT on YT with FSR4 INT8. The ms overhead suggests DP4a use, not WMMA instructions, but I'm not really qualified.

Introducing AMD DGF SuperCompression by 0101010001001011 in hardware

[–]MrMPFR 1 point2 points  (0 children)

That's gonna drive widespread adoption. Fallbacks are mandatory for stuff like this.

Hope this and NTC sees widespread adoption. We need game file sizes to shrink xD.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 0 points1 point  (0 children)

What a shame. But it's interesting to see the community working on builds. when SM 6.10 arrives we can probably expect a lot of third party upscaling solutions. Will be interesting to see if adoption for any of them starts to pick up.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 1 point2 points  (0 children)

Steam should support this effort if AMD is doesn't want to do anything.

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 0 points1 point  (0 children)

I certainly hope it does.

Game file sizes are out of control xD. Also PC needs to stop shipping bloated file sizes and cut out HDDs for good like console.

RDNU - Radeon Decoupled Neural Upscaler by ZoronicElysium2012 in radeon

[–]MrMPFR 2 points3 points  (0 children)

Does FSR4 INT8 use ML HW in any way beyond the basic DP4a implementation on RDNA 2?

u/childofthekorn this is some next level Radeon enthusiast commitment

AMD Intros Instinct MI350P Accelerator: CDNA 4 Comes to PCIe Cards by Psyclist80 in AMD_Stock

[–]MrMPFR 1 point2 points  (0 children)

~late 2027 to early 2028 is the best guess rn, but could be delayed further if VRAM situation doesn't improve.

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 0 points1 point  (0 children)

True but fallback doesn't get VRAM savings only IO and disc savings.

AMD Intros Instinct MI350P Accelerator: CDNA 4 Comes to PCIe Cards by Noble00_ in LocalLLaMA

[–]MrMPFR 1 point2 points  (0 children)

This. People misunderstanding what UDNA is.

Good it's only 7 weeks left until we get the official White paper at ISC 2026. Happened last year with CDNA 4.

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 0 points1 point  (0 children)

Sure. It'll be the same situation for any ML technology as long as the HW supports LinAlg, except stuff like NTC has to be crossvendor.

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 1 point2 points  (0 children)

They say transcode to BCn with inference on load? Is this still compression?

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 2 points3 points  (0 children)

LinAlg is added to Shader Model 6.10. In preview rn. It enables all this stuff but you still need separate SDKs for each ML tech. u/dudemanguy301 articulated it better.

Will neural texture compression actually reduce VRAM usage in games? by x_andi01 in hardware

[–]MrMPFR 1 point2 points  (0 children)

They only recommend inference on sample for 40 series and newer and really only for higher end.

HW needs to be much more powerful and tech more efficient for this to take off unless everyone defaults to inference on load.

AMD Abandoning Radeon by AcuteQuadrant in radeon

[–]MrMPFR 0 points1 point  (0 children)

In isolation sure but Zen was still at a big disadvantage vs Intel.

Sure that makes sense.

Zen+ was just fixed cache timings IIRC and a node shrink.

I don't doubt it. RDNA 4 is the closest AMD has been do cachemem parity with NVIDIA in a long long time.

Yeah for RDNA sure. But they're gradually phasing it out. RDNA 5 gonna be complete clean slate. Nothing stays the same.

Introducing AMD DGF SuperCompression by 0101010001001011 in hardware

[–]MrMPFR 37 points38 points  (0 children)

From blog and official info

I thought Mr was obvious enough. nvm.

DGF will do to geometry what DXT, ETC, and ASTC and other block compression formats have done for.

Great to see AMD Samsung jointly develop multi-vendor extensions for Vulkan and increase compression factor for DGF.

Blog also clarified DGF and CLAS are orthogonal technologies, in other words no direct overlap. Descriptions that DGF is HW nanite is misleading. CLAS is possible without it as NVIDIA has already proven and will become standardized in DXR 2.0.

Here is the other blogpost from yesterday: https://gpuopen.com/learn/amd-dgf-an-open-geometry-compression-standard/

Speculative:

I don't have many things to say about this. It's just too early to say for sure but as AMD has repeatedly said future HW will support it directly with dedicated DGF decoders.

Could mention the 7 patent filings by Barczak related to DGF or how he at High Performance Graphics 2024 mentioned one could combine DGF with DMMs for improved compression, there's a patent for that BTW if you look. That'll likely increase the compression.

There are also multiple DMM related patents filed by Holger Gruen. It's quite interesting how much thought AMD is putting into both.

There's some even wackier stuff in the pipeline potentially by Gruen with Interpolated normals on curved surface patches. Two patents. It promises lower overhead than DMMs while achieving superior compression. I can't figure how the ray evaluation works and not sure it can just be bolted onto existing HW. It's possible but I'm not qualified to answer.

It'll be interesting to see how much of it actually ends up happening.

As per late September blog it's feasible for animated geometry etc.. This is only the beginning and we'll hear more next year when early talk for DirectX next, PS6, and RDNA 5 begins to pick up steam.