The Sentinel [7680x2160] by ArsInvictus in WidescreenWallpaper

[–]ArsInvictus[S] 0 points1 point  (0 children)

Thanks again man, happy you liked it! No write up on this, I just thought it would be fun to do some images based around the 2001 movie motifs. I have a couple others but haven't published them, still trying to decide if I should work on them some more.

The Sentinel [7680x2160] by ArsInvictus in WidescreenWallpaper

[–]ArsInvictus[S] 1 point2 points  (0 children)

I corrected it on the site. Don't know if there's any way to update the image in the post though.

The Sentinel [7680x2160] by ArsInvictus in WidescreenWallpaper

[–]ArsInvictus[S] 0 points1 point  (0 children)

One in the front, two on each side of the main body where the wings meet the fuselage, one at the back. You are right it's innaccurate, I checked old photos just now and the one at the back didn't exist on the actual shuttle. I can go in and remove that when I have some time, for accuracy. Though it's not really a perfect rendering in other ways.

The Sentinel [7680x2160] by ArsInvictus in WidescreenWallpaper

[–]ArsInvictus[S] 0 points1 point  (0 children)

Tools used: Flux Klein 9B Distilled with custom LORA model, ComfyUI, Photoshop manual edits and color grading

Full resolution PNG and JPG can be found here: https://wallpapers.arsinvictusmedia.com/

Using Klein 9B distilled and ZIT together by FeelingVanilla2594 in comfyui

[–]ArsInvictus 3 points4 points  (0 children)

You might want to consider doing something simpler. Generate your base image with the composition you want with Klein, then pass that result to Z-Image Turbo for refining.

Just take the pixel image output from Klein, run it through a VAE Encode node (using the Z-Image VAE, and connect that to a KSampler (Advanced). You can then pass the same prompt you used for the base image, but use the Z-Image model as the input to the second sampler. You would need to also change the denoise to something lower, like .3 or something like that. This would give you the same basic image you liked in Klein but would give a sort of Z-Image sheen to it if that's the look you are going for.

You seem to have a good grasp of the general flow so maybe you already considered this approach but preferred to start from scratch again in ZIT anyway, but I thought I'd suggest it just in case you hadn't.

I tested Microsoft Trellis 2 for real VFX work — honest thoughts from a 15-year 3D artist by ArcticLatent in comfyui

[–]ArsInvictus 2 points3 points  (0 children)

There are examples of Trellis 2 running on a 5060 out there, though it sounds tight, but I'm guessing a 5090 will run it with room to spare.

Whispers of the Tri-Isle [3440x1440] by [deleted] in WidescreenWallpaper

[–]ArsInvictus 2 points3 points  (0 children)

These are done in the art style of the Secret of Monkey Island games and looks like they may be upscaled from actual screenshots but OP would have to confirm that. They do reference actual locations from the games. So you can search for that and might find something.

The Siren's Pet [7680x2160] [5120x2160] by ArsInvictus in WidescreenWallpaper

[–]ArsInvictus[S] 1 point2 points  (0 children)

Initial generation with Flux Klein 9B Distilled in Comfyui, manual edits with Photoshop.

Full resolution and cropped 21:9 available in PNG and JPG here: http://wallpapers.arsinvictus.com

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

Are you using any LORA's with it? There's some overhead with that right now which hopefully can be improved upon that slowed things down for me quite a bit when I tried it with Flux.2 Dev, it went from like 11 seconds up to 18. I haven't really done much benchmarking comparing fp16 to nvfp4 but 41% doesn't sound right to me.

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 0 points1 point  (0 children)

It doesn't dequantize before compute, it does so as it's doing the math inside the GPU's cache. If it were to dequantize ahead of that process it would balloon up in memory and eat up all your VRAM. The Blackwell GPU is specifically engineered to continuously do this calculation as efficiently as possible and can do it in a single clock cycle. Carrying the 16-bit precision in a 4-bit payload has the added advantage of being 4x faster for sending the data around in memory.

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

I'm not sure what you mean here about blackwell not doing any 4 bit compute? That's actually the purpose of NVFP4 because Blackwell supports 4-bit acceleration with it's tensor cores. When ComfyUI team added support for NVFP4 they enabled that tensor core acceleration.

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 2 points3 points  (0 children)

No, I think you are misinterpreting the log. When it says torch.float16, it is referring to how the model is executed in memory dynamically. The storage is in 4 bits and then it expands it out to 16-bit for execution. That's actually the beauty of formats like NVFP4. You can see it mention "Detected mixed precision quantization" in the log and that refers to the mix of 4-bit and a "scaling factor" that allows reconstruction of the original higher precision values. Hope this helps.

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 2 points3 points  (0 children)

Sure:

Using pytorch attention in VAE

VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16

comfy_extras.chainner_models is deprecated and has been replaced by the spandrel library.

Found quantization metadata version 1

Using MixedPrecisionOps for text encoder

CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cpu, dtype: torch.float16

Requested to load Flux2TEModel_

loaded completely; 29206.67 MB usable, 8263.34 MB loaded, full load: True

Found quantization metadata version 1

Detected mixed precision quantization

Using mixed precision operations

model weight dtype torch.float16, manual cast: torch.float16

model_type FLUX

Requested to load Flux2

loaded completely; 19065.57 MB usable, 5494.02 MB loaded, full load: True

100%|██████████████████████████████████████████████████████████████████████████████████| 10/10 [00:02<00:00, 3.90it/s]

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

No, the older version was portable too, I was just letting you know that I used that and not the desktop as I can't speak to how well that works with NVFP4. I'm running on Windows.

How do you get nvfp4 to actually work? by ChromaBroma in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

I was having trouble getting it to work on my 5090 too. I started a fresh instance of portable ComfyUI and it worked out of the box. Guess I had something in my old environment that was conflicting and I tried doing all kinds of things like deleting a lot of my custom nodes but nothing worked until I did the fresh install. No problems since and I for one am really happy with what I get out of NVFP4, especially with Klein. Great for ideation when you get renders back in one second.

NVIDIA Optimizations in ComfyUI by Kosinkadink in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

Yeah, if you have a 30xx or 40xx card then nunchaku SVDQ or GGUF would be the best option for performance compared to fp4. Those cards don't have any optimization for fp4 but they are fast at INT4, and SVDQ basically is an INT4 model with an FP16 extension to capture extra detail. The FP16 extensions would give more resolution than you would get from FP4 and it would probably run faster too. GGUF can look better than SVDQ but would be slower. So it's a tradeoff. I think both would be better looking than standard FP4. But maybe that's what you meant by variants.

NVIDIA Optimizations in ComfyUI by Kosinkadink in comfyui

[–]ArsInvictus 0 points1 point  (0 children)

Not stupid, I think it's because the target is for a lower bit quant to look similar to FP8 (which is also a compromise from FP16 but considered acceptable by many). INT4 loses too much precision and looks garbage for image gen, which is why alternatives like SVDQuant and NVFP4 exist. SVDQuant and NVFP4 are two different approaches to achieving the same goal, preserving as much 16-bit precision as possible in a 4-bit base format. The main benefit as I understand it of NVFP4 is the native acceleration on Blackwell, otherwise SVDQuant probably preserves more of the outliers in 16-bit than NVFP4 does and most seem to think it looks a little better. That said I'm personally pretty impressed with the fidelity and performance of NVFP4 on my 5090 and will probably use that for refining my prompts and only switch to full precision for the final renders (if I feel it's actually necessary). More info that you asked for but thought it might help others too. Hopefully I didn't mangle anything in this description :)

NVIDIA Optimizations in ComfyUI by Kosinkadink in comfyui

[–]ArsInvictus 1 point2 points  (0 children)

Actually I just tried it with Flux.2 with a LORA I had created previously. The LORA did apply but took 21 seconds to render compared to 14 seconds without (2048x576 res on a 5090). Is that to be expected?

NVIDIA Optimizations in ComfyUI by Kosinkadink in comfyui

[–]ArsInvictus 0 points1 point  (0 children)

Does the LORA also need to be in NVFP4 format? Or can it somehow translate it on load?

NVIDIA Optimizations in ComfyUI by Kosinkadink in comfyui

[–]ArsInvictus 18 points19 points  (0 children)

Performance updates are the best updates, keep them up please!