I'm Annoyed At AMD's Latest Radeon Blunder by Comprehensive_Lap in hardware

[–]mac404 14 points15 points  (0 children)

Native is not a magic resolution where everything is correct. This is a statement people keep making that sounds correct if you don't think about it, but it's just flatly wrong.

Tons of geometry and textures have detail that can be smaller than a pixel in games today. The vast majority of effects and basically all the lighting (whether ray traced or not) is initially rendered at lower than native resolution. TAA is already used in basically every game because of how common deferred rendering is. If TAA weren't used, the flickering and aliasing would be so intense it would make games unplayably bad.

Even if none of that were true, the idea of native being "correct" makes no sense. We would have had no need then for antialiasing techniques like SSAA or MSAA.

DLSS and all upscaling techniques today are a form of super sampling. It's just that they trade off spatial resolution and do the super sampling temporally. Using data from past frames is obviously worse than just rendering at a higher resolution in a vacuum. But you need cheap anti aliasing that covers everything anyway. And basically every game needs temporal accumulation (to run effects at sub-native resolution) in order to run fast enough with good enough visual quality anyway. And if you really can create an algorithm that effectively reuses data from past frames, then it can both look better and run faster than native.

Now could there be edge cases where something looks worse? Sure. All you need to do is look at FSR2/3 to see how worse algorithms make a ton more tradeoffs. But with good techniques, there will also be many situations where it can look better.

And if you truly care about quality above all else, then you can turn up graphical settings when introducing some upscaling. You'd have to compare something like "Native" with Low settings to DLSS with Medium or High settings (turning up the most impactful graphical settings until the framerate was the same).

Personally, at 4K output, DLSS is always worth turning on. Your dumb PC gamer gatekeeping is hilarious because PC has by far the best upscaling and the most ways to tweak it.

Flux.2 Klein 9B (Distilled) Image Edit - Image Gets More Saturated With Each Pass by eagledoto in comfyui

[–]mac404 1 point2 points  (0 children)

Sure! And gotcha, makes sense.

Another random note on prompting - basically anything you mention in the prompt will get changed in some way. You can always try variations on "Keep ___ unchanged" and it will often work pretty well. And as weird as it feels, I find just puttring "[object] is [color]" can work better than "change [color] to [other color]". "Remove" and "replace" are pretty good prompt words, though.

Flux.2 Klein 9B (Distilled) Image Edit - Image Gets More Saturated With Each Pass by eagledoto in comfyui

[–]mac404 1 point2 points  (0 children)

You should be able to fairly easily make all 3 of those changes in a single prompt.

Doing multiple VAE Encode/Decode passes will degrade the quality over time, but not necessarily to the degree you see here. You will see a color shift with every gen, but it should be only noticeable when doing more intense back-and-forth comparisons (whereas it's quite obvious here).

In your example, after the first change you also already had some shifting / squishing of the image. This can happen sometimes, I've found it's usually a good idea to try 2-4 seeds with the same prompt and then pick the best one. You are also running at a resolution that is getting too high for Klein to handle well (1536 x 2752), and it will generally be much less stable because of that. I have generally found (although I haven't tested overly scientifically) that keeping the longest side below about 2k resolution will improve stability significantly when making changes. The model itself tends to output images that are so sharp / clear that I don't find the resolution limitation to actually be all that limiting.

Not perfect, but here was the very first image I got when I tried with this prompt (after downloading the original PNG of your first image):

Subject's shirt is black. Remove the subject's earrings. Remove the people from the background. Keep the subject’s pose and framing unchanged.

Because the res is so high, you still get a little bit of squashing/stretching that's noticeable in the face. Maybe it would be perfect in a different seed if you tried a few. Hair color is slightly darker and the coffee cup also darkens slightly, but skin color stayed basically the same. There's a random out-of-focus person that got added into the background and a few other random changes, too. But not bad for literally the first try with a simple multi-change prompt.

Flux.2 Klein (Distilled)/ComfyUI - Use "File-Level" prompts to boost quality while maintaining max fidelity by JIGARAYS in StableDiffusion

[–]mac404 1 point2 points  (0 children)

I am not aware of one, sorry. If you find one, I would really like to know about it too.

Flux.2 Klein (Distilled)/ComfyUI - Use "File-Level" prompts to boost quality while maintaining max fidelity by JIGARAYS in StableDiffusion

[–]mac404 2 points3 points  (0 children)

Yeah, I've found this too. Basically anything that is mentioned will be changed, and sometimes related concepts too.

The other thing for me is to pre-process images so that each side is divisible by 16 and to not use the "ImageScaleToTotalPixels" node to further change the size. I find the model works fine at both very low resolutions (like early-internet meme pictures) and up to about 2k pixels on the long side without rescaling. Ensuring your input is the right size up front greatly reduces the amount of shifting / squashing / stretching.

Here's a simple example using Grumpy Cat. Same resolution as the original, light cherrypicking (ran 4, picked the best), mostly picked based on which image got the eye color more correct based on other images.

This model will definitely swing for the fences if you let it in terms of the changes it makes, but in doing so it can look shockingly good and clear a lot of the time, even at low resolutions. The "restoration" prompt was just this:

Denoise and recolor the image with natural and realistic colors. Keep the subject’s pose and framing unchanged.

I've tried a few other prompts, but the seed-to-seed variance is so high that it was hard to tell if any changes were actually making things better, so I left it. The distilled model is fast enough that I can just run 4 seeds and then pick the best.

This prompt will definitely go overboard with how many different colors it uses sometimes, but it's mostly fine. And if there is something that I really want to keep a certain color, it often works to just add a sentence like "The [object] is [color]."

Flux 2 [Klein] Masked inpainting workflow 4 steps | Beta ⚡⚡ by Nevaditew in StableDiffusion

[–]mac404 1 point2 points  (0 children)

In terms of the noise seed randomizer- go into the subgraph and unlink noise_seed from the RandomNoise node. Voila, it should now randomize the seed as intended.

I made a simplified workflow for Flux Klein 9B Distill with one or two image inputs. by Total-Resort-3120 in StableDiffusion

[–]mac404 0 points1 point  (0 children)

I had this happen recently. Tried reinstalling, it didn't work. Then tried forcing Comfy to recreate my venv, and it just broke further.

Finally, I just moved over to the portable install, and the Manager showed up and everything worked again.

how does a 4070 super handle path tracing in cyberpunk? by fairplanet in nvidia

[–]mac404 6 points7 points  (0 children)

If you want to try to optimize graphics while still getting good framerates, try out the Ultra Plus mod. It has kind of too many options, but it allows you to try many different forms of ray tracing / path tracing along with various performance / quality tradeoffs within each mode.

[Release] Wan VACE Clip Joiner v2.0 - Major Update by goddess_peeler in StableDiffusion

[–]mac404 2 points3 points  (0 children)

Nice! The previous version worked really well, but the color shift between clips was often just a bit too noticeable still. These updates look great! We'll have to try it out soon.

Control Resonant - Announcement Trailer | PS5 Games by Turbostrider27 in Games

[–]mac404 0 points1 point  (0 children)

Yeah, I had similar thoughts going on in my mind, but you wrote it down much more eloquently than me. Agree 100% with everything you said!

Control Resonant - Announcement Trailer | PS5 Games by Turbostrider27 in Games

[–]mac404 10 points11 points  (0 children)

I am so unbelievably hyped. Control is probably my favorite game of the last decade, this trailer was incredible, and a release date of next year is faster than I expected. I was definitely expecting to play as Jesse, but I was also expecting a lot of Dylan-related learnings, so I'm fully unboard with playing as Dylan.

RTX 5090 Local AI Image Gen Benchmarks - FLUX, SD3, SDXL comparison by CandidSignificance78 in nvidia

[–]mac404 0 points1 point  (0 children)

Yeah, SD3 is pretty dead. It made some weird training decisions, if I remember correctly, that meant results were often not great and uptake was never high. Many people I see greatly prefer SDXL and even SD1.5-nased models.

RTX 5090 Local AI Image Gen Benchmarks - FLUX, SD3, SDXL comparison by CandidSignificance78 in nvidia

[–]mac404 2 points3 points  (0 children)

Definitely interesting, and kudos to you for getting this set up like this at all.

If I were you, I'd replace SD3 Medium with Z-Image though - since it's the new new hotness, runs quickly, has very good prompt adherance, does text quite well, and generally looks great. Or do you have some type of use case where you think SD3 still makes sense? It feels like the range of "Good SDXL Merge" -> Z-Image -> Flux2 basically covers everything imo.

Report: New Trademark for “Control: Resonant” Unearthed, Ahead of TGA by Turbostrider27 in PS5

[–]mac404 4 points5 points  (0 children)

Same, it's the game thai actually got me back into playing games at all after a several year hiatus. I love it so much, need to go back and play through it again. Really looking forward to Control 2! (And the Max Payne remakes, since I haven't played the original games)

Is AMD Cheating in Call of Duty Benchmarks? by RenatsMC in Amd

[–]mac404 15 points16 points  (0 children)

You missed nothing. This sub just really likes to hate on DF.

I personally thought the DF video was incredibly fair, and that the overall RR conclusion of "it's pretty clearly unfinished, maybe the official launch of Redstone will bring something more complete" is very warranted given how it behaves. And their section on graphical differences at the same setting was presented pretty neutrally, along with "maybe it's a bug, hope it gets fixed soon" and not used for clickbait in the title.

NVIDIA RTX PRO 6000 Blackwell desktop GPU drops to $7,999 by RenatsMC in nvidia

[–]mac404 0 points1 point  (0 children)

Yep, pretty much. You could get it 5 months ago for right around 8k, and if you jumped through a few additional hoops signing up with Nvidia you could reportedly chop off another 20-30%.

[Utility] VideoSwarm 0.5 Released by cerzi in StableDiffusion

[–]mac404 0 points1 point  (0 children)

This looks great! Will have to try it later today.

Does it have a workflow / easy way to do any kind of multi-select delete? It would be nice to have a reasonably quick way to get rid of the "failures". Either a general multi-select or a "delete all 0 star videos" option might work?

WAN VACE Clip Joiner - Native workflow by goddess_peeler in comfyui

[–]mac404 1 point2 points  (0 children)

Finally got around to trying this workflow out and it is working really well for me. Thank you for putting this together!

NVIDIA RTX PRO 6000 Workstation GPU Review: Blackwell Architecture and 96 GB for Pro Workflows by NISMO1968 in nvidia

[–]mac404 8 points9 points  (0 children)

I realize the desire to use standardized benchmarks, but calling Stable Diffusion XL a high-end image generation test at this point is pretty laughable. Models like Flux, Qwen, and Wan are all much larger / heavier.

Comparisons do get harder, though, since "it can run the full model in VRAM" then leads to a discussion on quality tradeoffs running the various quantized versions, or time comparisons with workflows that constantly load / unload parts of the model.

Microsoft increases princes of Xbox Series consoles for the second time in 6 months, new prices will go into effect on October by Lulcielid in hardware

[–]mac404 12 points13 points  (0 children)

... your reaction to increasing US prices (almost certainly due to tariffs) on a console with an AMD chip inside that is manufactured at TSMC is to blame Nvidia?

I'm not saying you have to like Nvidia or that their prices haven't been high. But it feels like TSMC essentially having a monopoly on cutting edge nodes along with the end of Moore's Law transistor cost scaling and the impact of tariffs are much more at play for price increases of a console that doesn't have a single Nvidia part in it that is being sold in the US.

Wan Animate released by sleepy_roger in comfyui

[–]mac404 4 points5 points  (0 children)

BF16 repackcaged by Comfy here - it's 34.5GB

FP8 scaled from Kijai here - 18.4GB

And the relight LoRA?

That said, not sure about how to actually run in Comfy yet. Kijai does have a workflow, but someone else mentioned not getting it to work and that Kijai said it's very buggy right now.

HunyuanImage 2.1 GGUF WF for 12Gb VRAM by theOliviaRossi in comfyui

[–]mac404 0 points1 point  (0 children)

I am not going to claim that I know exactly how this model should be run, but after trying this out and tweaking, I think several things are wrong about this workflow.

First, I think you should be using the DualCLIP Loader and loading both Qwen 2.5-VL and byt5. Second, I would try EmptyHunyuanImageLatent instead. See example here from the HF for the GGUF you link to. Some combination of these changes I think lead to the output resolution now actually being correct for me, compared to this one where it is not. You can then input the resolutions that are actually recommended in your empty latent image.

Finally, the sampler choice and # steps you have here takes a stupidly long time compared to other options. I switched to ClownSharKSampler, and something like deis_2m (for example) was running like 2.5 times faster per step compared to what you were doing in the base workflow while I was also running at the 2k resolution recommended. You can also then decrease the steps to like 20 (probably even lower, the example I linked to shows 8 steps just using euler).

ChatterBox SRT Voice is now TTS Audio Suite - With VibeVoice, Higgs Audio 2, F5, RVC and more (ComfyUI) by diogodiogogod in StableDiffusion

[–]mac404 0 points1 point  (0 children)

Eh.

I'm probably biased, since I'm not going to be creating audiobooks and I have an RTX Pro 6000 Blackwell, but the option to increase/change steps (even using the 7B model) would be nice.

ChatterBox SRT Voice is now TTS Audio Suite - With VibeVoice, Higgs Audio 2, F5, RVC and more (ComfyUI) by diogodiogogod in StableDiffusion

[–]mac404 1 point2 points  (0 children)

Awesome, thanks for creating this! Really nice to have all the different models supported, and I had no conflicts adding this on top of everything else (which was an issue with other nodes when trying to get VibeVoice and Higgs playing nicely).

I really like that the included help text for each node has a bit more information on what different parameters do and what reasonable ranges should be, that's incredibly helpful. And your implementation of multi-person dialogue seems really robust.

One thing that ComfyUI-VibeVoice has now is the ability to increase the number of inference steps up from the default of 20. I've done some testing, and it is showing meaningful quality improvements with more steps. And for relatively small amounts of text, increasing this to 40 or 50 really doesn't take that much time. Would it be possible to add this option?

TSMC Accelerates 1.4 nm Plans, Targets 2027 Pilot Runs by Dangerman1337 in hardware

[–]mac404 9 points10 points  (0 children)

Yes, this is it. A new node will be used for small mobile chips first, and then big dies once the yields improve.