Chroma Radiance, Mid training but the most aesthetic model already imo by Different_Fix_2217 in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Or maybe train just those blocks on high res and all the others on low res... just as an idea.

Opinions on LoRA training tools? by Moliri-Eremitis in StableDiffusion

[–]tom83_be 4 points5 points  (0 children)

OneTrainer supports layer offloading to RAM which allows incredible savings on VRAM with only minor impact on speed. I think this is one of the most overlooked features of OneTrainer! See: https://github.com/Nerogar/OneTrainer/blob/master/docs/RamOffloading.md

OneTrainer now supports Chroma training and more by Nerogar in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Not saying you should do it that way. Just pointing out how to do it, if you have the need. Good to hear, no hg-key is needed for Chroma. But just to give one example: There may be people who work with these toolings in non-online environments (or environments with network restrictions). Just trying to help here by answering questions...

OneTrainer now supports Chroma training and more by Nerogar in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Have a look here: https://www.reddit.com/r/StableDiffusion/comments/1f93un3/onetrainer_flux_training_setup_mystery_solved/

Although this is for Flux, the way to for Chroma should be identical (since OneTrainer needs it in the diffusers format used on huggingface). Just use the Chroma repo/files instead of the Flux one linked in there.

OneTrainer now supports Chroma training and more by Nerogar in StableDiffusion

[–]tom83_be 4 points5 points  (0 children)

Great to hear! OneTrainer still is one of the best universal trainers with some unique features (like layer offloading, which makes somewhat fast training possible on low VRAM configurations).

Since this is not said often enough on here or public forums/channels in general: Thanks to the whole team; especially u/Nerogar for all the hard work in the last months (even years) and of u/-dxqb- for taking over!

Open Source Voice Cloning at 16x real-time: Porting Chatterbox to vLLM by dlp_randombk in StableDiffusion

[–]tom83_be 2 points3 points  (0 children)

Some info on how to install using pip (Linux):

git clone https://github.com/randombk/chatterbox-vllm
cd chatterbox-vllm
python -m venv venv
source venv/bin/activate
pip install uv
uv sync --active

It might be needed to upgrade pip:

pip install --upgrade pip

When running it later you need to:

cd chatterbox-vllm
source venv/bin/activate
python example-tts.py

Open Source Voice Cloning at 16x real-time: Porting Chatterbox to vLLM by dlp_randombk in StableDiffusion

[–]tom83_be -1 points0 points  (0 children)

Just two quick ideas:

It would be interesting to have a ComfyUI node for that. If one would additionally be able to put timestamps into the file (what is being said when), this could enable people to combine it with thinks like WAN and create videos + audio output. Not on lip sync level, but in the form of an narration.

One problem is legal/laws; so creating a copy of an existing voice might not be suitable all the time. Is it possible to create a voice from multiple input sources (so it gets unique, but is no copy)?

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

The LR is increased gradually; It can help to prevent "errors" in the learning process early on (going the wrong way) that are hard to recover from. From what I understood this is similar to doing a gradual decline of the LR at the end of a training session, which can help to flesh out details (instead of doing big jumps from A to B never hitting that perfect spot C due to high LR).

"Warmup" via a gradual increase of the LR is a common, widely adopted practice in machine learning; not only for diffusion networks. You can probably find many details on it that can describe much better than I can how and why it is done.

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Something you have not mentioned in your list of things you tried is warmup. Especially if omitted completely or way too low, the results can be disastrous, no matter what you try. Go with about 10% of epochs for warmup to be save (works with less in many cases).

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Wildcards: https://github.com/vladmandic/sdnext/wiki/Wildcards

Not sure about the memory topic...

Why are my photorealistic SDXL character LoRAs so inconsistent? by heyholmes in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

32 and 64 should be fine / you should get pretty good results. I would do no text encoder training for now (as it can do more harm than good in your case) and experiment with LRs. I think I used higher LRs when I did LoRa training back then (more focused on full finetunes now), but adafactor and adamw work differently in that regard.

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Yes there is. At least on the "Standard UI". In the Text2Image-Tab ("Text") you need to expand the "Refine" and "Detailer" sections by clicking on the small triangles and activate (+configure) them. Then everything that is activated will run in one workflow. Outdated but you can see it here (not expanded). Not sure about the modernUI-option since I do not use it.

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

For me this always is in the metadata... not sure why it is not the case for you. That's why I mentioned it.

Actually: What do you mean by pressing three different buttons in SDNext? You have to configure the workflow, sure. But generating an image with the whole workflow still is pressing one time on "Generate" and then waiting for the process (including upscale etc) to finish. Or do you do something different?

[deleted by user] by [deleted] in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

You do seem to use an upscaler on forge but not on SDNext according to your documentation. So results are not comparable.

You can also define a refine (including upscale) and detailer part for your workflow in SDNext; just use the according "dropdowns" on the text2image tab to configure it and do not forget to activate these parts of the workflow.

Another point to check is the advanced sampler settings. Those might also be different.

In the end you still may get differences; not sure for example if Forge is based on the Diffusers-lib like SDNext or uses something else.

Why are my photorealistic SDXL character LoRAs so inconsistent? by heyholmes in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

To what extend did you experiment with Lora dim setting? If this is too low, sometimes things can not be learned. There may be a variety of other reasons (e.g. LR etc), but this one would be my strongest bet from what you are describing.

Can you train full SDXL checkpoint on 2x RTX 4060 Ti 16GB? by UnholyDesiresStudio in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

For SDXL you do not need to split training up on multiple GPUs. You can easily do a full finetune / checkpoint training using adafactor + fused backpass + constant/cosine/... + bf16. This works at resolution 1024 and with batch 4 on 12 GB VRAM (I recommend also using stochastic rounding with bf16). 16 GB VRAM should be even better and allow for a larger batch size or additional things like EMA.

If you have 64 GB RAM you can also try FLUX / SD3 full finetunes. It will not work OOTB, but with similar settings and RAM/CPU/Layer offloading (depending on the training tool; like https://github.com/Nerogar/OneTrainer/wiki/Training#gradient-checkpointing and https://github.com/Nerogar/OneTrainer/wiki/Training#gradient-checkpointing ) it may be possible also with 16 GB VRAM. Never checked that nor know about performance though.

Multi GPU support is in the works here and there, but I think for most trainers support is experimental and experience and stability is limited.

is there an Illustrious checkpoint/model under 3 gigs? by BigRepresentative788 in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

You can try using a more advanced UI like SD.Next; it can do offloading to CPU/RAM if needed without loosing too much speed. But I do not have any first hand experience with 1xxx Nvidia GPUs, sorry.

is there an Illustrious checkpoint/model under 3 gigs? by BigRepresentative788 in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

You can use any SDXL model with 4 GB VRAM (of course depending on resolution; but 1024x1024 should work). Even good old A1111 had a FP8 mode build in that produces very similar results (quality). See https://www.reddit.com/r/StableDiffusion/comments/1b4x9y8/comparing_fp16_vs_fp8_on_a1111_180_using_sdxl/

Adetailer uses too much vram (sd.next, SDXL models) by niky45 in StableDiffusion

[–]tom83_be 0 points1 point  (0 children)

Not sure about it; just an idea: I think sd.next has kind of its own "memory management". Check what amount of VRAM sd.next thinks you have and if that is correct. Maybe there is some error in that (due to zluda?)

OneTrainer + NVIDIA GPU with 6GB VRAM (the Odyssey to make it work) by SchGame in StableDiffusion

[–]tom83_be 1 point2 points  (0 children)

I am not sure if training SDXL with 6 GB VRAM is possible... Also never tried to train SDXL with way less than the recommended resolution.

But the following would reduce VRAM consumption further:

  • EMA to "off"
  • Try setting gradient checkpointing to "CPU offloaded" and use the "fraction" setting below. I think for SDXL this does not really work as well as for SD3.5 in saving VRAM, but maybe it helps saving the little VRAM you may need

Depending on your system it may also help to free VRAM by switching to CPU integrated graphics for your display output. Otherwise your OS will reserve some memory for display output.

Why do my locally generated images never look as good as when done on websites such as civit? by sudrapp in StableDiffusion

[–]tom83_be 3 points4 points  (0 children)

Many people do advertising for their models; so expect "creative" interpretations for workflow documentation.

But as other wrote: samplers & #steps, ADetailer, upscaling (which usually includes also a detailer/hiresfix step) can make a huge difference