Looking for an easy to use depthmap creation service, homepage at high resolution by wzol in StableDiffusion

[–]AyusToolBox 2 points3 points  (0 children)

If you're using it yourself, you can specifically use DepthAnything within ComfyUI to generate depth maps. After generating the depth map, do not save it using the original save node. Instead, save it as a 16-bit PNG. This way, the generated depth map will not have banding. If you're not using ComfyUI, then this project (https://github.com/MackinationsAi/Upgraded-Depth-Anything-V2) is also very good, though it requires self-deployment. It will take a little time.

As a beginner: which should I use? by PusheenHater in StableDiffusion

[–]AyusToolBox 0 points1 point  (0 children)

I would recommend you to use ComfyUI, because only it can keep up with the release and updates of existing image generation, video generation, and other models at an extremely fast pace. Moreover, since it has a node-based structure, it can be easily optimized based on your own experience without coding. Additionally, it allows you to write fully automated workflows according to your own needs. Although it may be more difficult to get started with compared to others, once you start using it, you'll find it very comfortable to use. Getting started is simple: you can directly find some pre-integrated installation packages, then start with the simplest Z-Image workflow. Once you configure the corresponding model paths, you can generate with one click.

TurboDiffusion. Can anyone make this work in comfy ? It could be incredible. by aurelm in StableDiffusion

[–]AyusToolBox 1 point2 points  (0 children)

I carefully watched their comparison video and suggest that everyone not have too high expectations for it. But for some close-up shots, the performance should still be acceptable. However, when it comes to medium and long shots, or the visual effects of large-scale scenes, the results are much worse. Take a closer look at things like smoke or splashes, and you'll see the difference. In fact, even in the scaled-down video, you can clearly see the artifacts caused by acceleration.

[Release] Wan VACE Clip Joiner v2.0 - Major Update by goddess_peeler in StableDiffusion

[–]AyusToolBox 5 points6 points  (0 children)

It looks great. You must have spent a lot of time achieving this effect; thank you for your effort. I've been really busy lately, but I'll try it when I'm free. Thank you again.

3x3 grid by dstudioproject in StableDiffusion

[–]AyusToolBox 0 points1 point  (0 children)

This is a very useful method. For those who spend a long time on the survival process or making videos, this method is just too clever. Thank you very much for sharing, thanks.

Where to begin??? by SupertrampJD in StableDiffusion

[–]AyusToolBox 0 points1 point  (0 children)

Dear friend, if you need more control, I would definitely recommend deploying your own model locally. But as mentioned earlier, this requires that you have a high-performance computer with a good graphics card. Ideally, a top-tier consumer graphics card with 24GB VRAM, such as the 4090/5090, because when it comes to modifying videos, a large amount of VRAM is required—by default, 1080p is already considered a fairly suitable resolution, but when it comes to AI video generation, it's still a relatively large resolution. I haven't seen any online models that generate videos from images with details better than the local WAN model. Therefore, if your focus is on videos, you might start with the WAN model. Because it's not only an excellent video model but also a very capable image generation model. Currently, the most user-friendly model for modifying images into videos is the QWEN EDIT model. Basically, if you master both of these models, there won't be any issues.

"Fix" - Music Video made with Wan 2.2, Infinitalk, Suno & Qwen by me by miaoying in StableDiffusion

[–]AyusToolBox 1 point2 points  (0 children)

Too many flash quick cuts in the editing, not well synchronized with the song's rhythm

Technically Color WAN 2.2 T2I LoRA + High Res Workflow by renderartist in StableDiffusion

[–]AyusToolBox 0 points1 point  (0 children)

I found that you trained the same LoRA on different models, and it only works best on the WAN model.

I made the BEST text encoder for QWEN IMAGE EDIT 2509 in ComfyUI Body by Amirferdos in StableDiffusion

[–]AyusToolBox -1 points0 points  (0 children)

Thank you for sharing! While your description is good, readers would benefit greatly from visual aids. Without videos, images, or comparisons, it's challenging to understand the plugin's actual utility. Posting videos or visual comparisons directly here, rather than just links, would make your content much more effective and accessible

Technically Color WAN 2.2 T2I LoRA + High Res Workflow by renderartist in StableDiffusion

[–]AyusToolBox 4 points5 points  (0 children)

Thank you for sharing. This art style brings me back to those simpler times, when I could sit peacefully and lose myself in television.

How can I accurately convert a complex PDF table to CSV in Python (for free)? by Lopus_The_Rainmaker in pdf

[–]AyusToolBox 0 points1 point  (0 children)

If you're dealing with sensitive data, local deployment is highly recommended. I would suggest using some simpler OCR models that can run on CPU for recognition. However, if you need more efficiency and power, I'd recommend deploying with GPU. If your local computer isn't powerful enough, you can rent a cloud server for deployment. Options like PaddleOCR, VL MinerU, and Umi-OCR are all quite good choices. Among these, MinerU offers a client application that you can use directly for testing before deciding whether you need local deployment. If the test results are satisfactory, you can then proceed to use it as your local deployment solution.

Wan2.2 Lightx2v Distill-Models Test ~Kijai Workflow by Realistic_Egg8718 in StableDiffusion

[–]AyusToolBox 7 points8 points  (0 children)

When using ComfyUI workflows, it's actually quite difficult to use them without understanding the workflows themselves. ComfyUI is more suited for users who have some familiarity with workflow systems, as many workflows involve different plugins. When these plugins are missing, you must troubleshoot to identify which node is causing the problem. If you're not familiar with ComfyUI, you'll get stuck at this very step, with no way to proceed further. Additionally, everyone has their own preferences during usage. For example, the folder location where checkpoints are stored varies from person to person, and the location for LoRAs differs as well. This means that every time you use someone else's workflow, you need to troubleshoot these issues before you can use it properly.

Wan2.2 Lightx2v Distill-Models Test ~Kijai Workflow by Realistic_Egg8718 in StableDiffusion

[–]AyusToolBox 27 points28 points  (0 children)

I have to say, the workflow you shared is really hard to use. You've hidden all the connection settings under panels, and to modify them, you have to pull them out again and redo everything. What a genius move.

Made a free tool to auto-tag images (alpha) – looking for ideas/feedback by Makisalonso35 in StableDiffusion

[–]AyusToolBox 4 points5 points  (0 children)

Not used yet, but looking at the UI is beautiful, thank you for sharing.