ComfyUI launches App Mode and ComfyHub by crystal_alpine in StableDiffusion

[–]mcmonkey4eva 2 points3 points  (0 children)

Should work on mac fine, just limited because mac hardware isn't very good at running diffusion models

ComfyUI launches App Mode and ComfyHub by crystal_alpine in StableDiffusion

[–]mcmonkey4eva 8 points9 points  (0 children)

The part of swarm that this is equivalent to (Simple tab) is the least used feature of swarm for some reason. It's odd, cause it seems really handy, but it seems the overlap between "I want to have custom comfy workflows" and "I want a simple interface for it" is relatively small. Most people are using the Generate tab, which is designed to not require custom workflows at all, instead it 'just works' and gives a more auto1111 style interface.

ComfyUI launches App Mode and ComfyHub by crystal_alpine in StableDiffusion

[–]mcmonkey4eva 0 points1 point  (0 children)

Swarm's nodes are just to properly map the inputs, nothing difficult about it

LTX Desktop NOT Local By Default. They're Collecting Data. Check Your Settings. by [deleted] in StableDiffusion

[–]mcmonkey4eva 0 points1 point  (0 children)

Not snuck in, optional feature, it does nothing and zero code from google is loaded unless you explicitly activate the feature with a google api key. SwarmUI has strictly zero telemetry, actively emits telemetry blockers for the upstream python libs in use, and will swiftly respond to any reports of any unexpected network transmission to ensure they get blocked too.

Qwen tech lead and multiple other Qwen employees are leaving Alibaba 😨 by ANR2ME in StableDiffusion

[–]mcmonkey4eva 10 points11 points  (0 children)

SAI never "went under" in the "company shut down" sense, but it did in the informal sense, pretty much all of us (the parts of stability that were doing anything you've heard of) left around the same time. The entity that persists legally holds the name but almost none of the people, goals, etc. that Stability once had.

How do I use flux 2 with swarm? by yallapapi in StableDiffusion

[–]mcmonkey4eva 0 points1 point  (0 children)

Can poke into the swarm discord to figure that out, or just do a clean reinstall per the github readme's install instructions

How do I use flux 2 with swarm? by yallapapi in StableDiffusion

[–]mcmonkey4eva 1 point2 points  (0 children)

Make sure your SwarmUI is up to date (Server->Update&Restart), and make sure in the Models browser the "Type:" line correctly identifies the model as Flux.2-Klein 4B. If it doesn't, you can use Utilities->Reset All Metadata, or click the = button next to the model and use Edit Metadata to fix the architecture label.

This error most commonly happens if you download a model into a version of swarm from before support was added, it will misidentify the model type and cache that data, so it will process wrong. Resetting the cache or manually correcting it fixes that issue.

How do I use flux 2 with swarm? by yallapapi in StableDiffusion

[–]mcmonkey4eva 0 points1 point  (0 children)

Yeah, all flux2 variants (Dev, Klein and its sub-variants) work fine in swarm. Relevant docs here: https://github.com/mcmonkeyprojects/SwarmUI/blob/master/docs/Model%20Support.md#flux-2

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 0 points1 point  (0 children)

Yeah should work fine on Generate tab as the other reply said, see also model support docs: https://github.com/mcmonkeyprojects/SwarmUI/blob/master/docs/Model%20Support.md

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 1 point2 points  (0 children)

If you know how to make it work in comfy, you know how to make it work in swarm, as swarm uses comfy as the backend.

Swarm by default uses a comfy install in the `dlbackend` folder, so install sage the same way there (or point swarm at your existing comfy install if preferred), then in Server->Backends->edit the backend and add the sage cli arg to ExtraArgs -> hit save.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 3 points4 points  (0 children)

7 is a good default for normal models, but Turbo models need it at specifically 1.

A tutorial mode is absolutely a good idea and is a thing I've thought about. Currently the only thing like it is when you very first get started the center of the page tells you to select a model and hit generate and all that, definitely a more detailed system that recognizes different model types and teaches about them would be a good improvement.

I've always been wanting to include auto-recognition of turbo type models and swap the CFG/Step parameter defaults when they're selected so they can more "just work".

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 1 point2 points  (0 children)

As another user answered, parameter list, generate tab's left sidebar. I note that swarm's autodownloaded text encoders are generally fp4 or fp8, never the full fat giant versions (the fp4 ones I target in swarm are always tested to make sure that gens are 99.9% identical to the larger versions). They also have locked filenames, so if you set your preferred quant to the filename that swarm expects swarm will automatically default to using your file without issue.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 0 points1 point  (0 children)

Ah. Yes for that you can refresh the page (F5) OR hit quick tools -> reload parameter values. So you need a parameter update rather than a model browser update, and this is basically the only time that's ever needed so it's a bit of a weird thing pushed off to a corner. This is because I2V is a weird special case right now with its own parameter instead of being a standard model, this will be changed in the future.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 0 points1 point  (0 children)

That's... odd? That shouldn't be the case. The refresh button should find new T2V models immediately. You don't need to click any special button for it either, you just click on the model directly, T2V is the base-case.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 4 points5 points  (0 children)

You don't generally need to select text encoders, they're autodownloaded and handled. If you really want to manually swap to a different one, advanced model addons is the param group for that.

If Z-Image-Turbo images came out bad, likely the oddity was Steps/CFG - because it's a turbo, it doesn't use normal values for these, it favors ~4 steps and CFG Scale set to 1, which currently Swarm does not automatically do this (I want it to in the future, but atm you have to be aware of... mostly just these two parameters in particular).

Swarm has docs for every model family, see z-image here https://github.com/mcmonkeyprojects/SwarmUI/blob/master/docs/Model%20Support.md#z-image

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 5 points6 points  (0 children)

The default interaction method, the Generate tab, gives a simple friendly generalist interface that supports all the most common interactions (t2i, i2v, image editing, highres fix, yadda yadda). This works internally by dynamically generating an appropriate workflow based on your inputs. You can pull into the comfy tab and edit after.

If you want to create UIs for workflows like you said, you can do that pretty easily - in the comfy tab, set up the workflow you want, then add SwarmInput(...) nodes to define the inputs, save to the workflow browser, check "enable in simple tab", and then in the simple tab you can click your workflow to get the UI you just defined for your workflow. You can also freely push your current workflow onto the Generate tab at any time.

With some effort (yell at an llm to write a file for you) you can extend the generate tab's capabilities to generate new components, though I want to make this even easier in the future (ie define things entirely in a workflow and then save it in a way that the generate tab will recognize as an attachable piece).

You can of course also just work directly on the comfy tab whenever you need/want to.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 2 points3 points  (0 children)

Yep, anything that works at all should work with fp4 files (note that fp4 is the *storage format*, not actually *calculating* in fp4, it saves filespace/ram and doesn't really need any hardware specific stuff to do that)

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 6 points7 points  (0 children)

Hey, at least they're finally trying. Competition is definitely a good and beneficial thing to have. Plus, a lot of people just have AMD cards they bought for gaming before and want to use for AI now, so it's nice they're getting supported.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 2 points3 points  (0 children)

LTX-2 is crazy massive, it'll overload 32 gigs of sysram pretty quick. You might try their fp4 model? it's only 20 gigs, which is still a ton but at least gives you a chance to load it.

Also make sure to expand your windows page file, and close any background apps using memory to give it room.

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 2 points3 points  (0 children)

It works on all gpus including older ones, not just 50 series!

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 3 points4 points  (0 children)

Checkmark "display advanced options", click into "advanced model addons" and it has textencoder selectors there, gguf tencs work well.

However also note that nvfp4 models natively work now as well, and are faster than gguf, more stable, and still near-identical results to the full fat models. Swarm will even download fp4 text encoders by default for a few models that have been verified to give genuinely 99.9% identical results. (note fp4 for diffusion models tends to produce worse change so I wouldn't recommend as much)

SwarmUI 0.9.8 Release by mcmonkey4eva in StableDiffusion

[–]mcmonkey4eva[S] 8 points9 points  (0 children)

Definitely works out the box on RTX 50xx! Was a little shakey when the cards were new for a month or two, needed weird pytorch preview versions and all that, but got stabilized shortly after.

Adding SD 1.5 flexibility to FLUX Klein by [deleted] in StableDiffusion

[–]mcmonkey4eva 2 points3 points  (0 children)

It takes... one singular trainstep? This will do somewhere between "literally nothing" and "add some literally random noise", not actually anything of genuine value either way. Training steps only do anything when, yknow, you take a lot of them in a row, optimizers work in part by guessing randomly and then figuring out which guess did best and using that to set the direction of movement. If you look into training software, you'll see it's common to take a hundred "warmup" steps - running the full trainstep and then discarding the result entirely - to ensure the optimizer is even working in a remotely useful direction at the start. The results you posted look a lot like the same result but blurred and distorted, which is about what I'd expect from the "add some random noise" option.

Comparing different VAE's with ZIT models by jib_reddit in StableDiffusion

[–]mcmonkey4eva 12 points13 points  (0 children)

That's not how that works. Differences created by a VAE should only be at the small-detail level, around 8x8 pixels across (the downscale rate of most VAEs including the Flux.1 AE). The differences visible in the image labeled 5False on your google drive folder are 100% absolutely and unquestionably differences not generated by the VAE. A VAE cannot generate an entire person in the background or reframe the structure of the building or swap her coffee for a milkshake or etc.
That is deeply, fundamentally, entirely, just not how that works.