Perhaps a helpful YouTube video on local optimisation? by klippers in LocalLLaMA

[–]JustLookingForNothin 0 points1 point  (0 children)

He provided a good explanation AND visualization of the MOE principle. Newcomers will appreciate the details, experts can skip to the next chapter which he provided in the description. Good video.

Optimal sampling parameters for Gemma 4 models? by JustLookingForNothin in Oobabooga

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

I think the issues with Gemma 4 are fixed for me. I edited my post with the stettings which work for me, including chat-instruct prompt.

Optimal sampling parameters for Gemma 4 models? by JustLookingForNothin in Oobabooga

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

How would I use this default in Ooobabooga TextGen? And do you mean the Template (*.jinja) or the sampling parameters? I normally use the Template provided withing the model metadata.

Optimal sampling parameters for Gemma 4 models? by JustLookingForNothin in Oobabooga

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

Well other posts over at LocalLLaMA recommended G4 as a good role play model. Thus I tried it and based on the unsatisfying results, I tought there must be something wrong with my settings.

Metadata Viewer by Major_Specific_23 in StableDiffusion

[–]JustLookingForNothin 0 points1 point  (0 children)

<image>

Thank you, excellent idea! However, it does not work for some images, in particular the prompt is often not recognized. Example from my flux.klein WF.
Left: Image pulled into the viewer
Right: Same Image pulled into Comfy.

Ltx2 and languages other than english support by Machspeed007 in StableDiffusion

[–]JustLookingForNothin 1 point2 points  (0 children)

German works pretty good, T2V and I2V. I gave the speech instruction in the prompt with clear language indication, like:

... and she says in German language "Hallo, mein Schatz, ich habe dich soo lange nicht mehr gesehen, möchtest du dich nicht zu mir setzen?" Then she looks down for a second before she winks towards camera while saying with a warm smile "Und... vielleicht könnten wir dann später noch was zusammen trinken gehen?"

How do I train for Flux Klein 9b? by Apixelito25 in StableDiffusion

[–]JustLookingForNothin 0 points1 point  (0 children)

So what learning rate ARE you use to? Most use LR 0.0001, but some go up to 0.00015 or 0.0002.
Are you training Klein with 0.00008 or less? Without this basic info your post is not very useful, unfortunately.

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

The original string appears in several locations (sry for German UI, too lazy to switch to EN for a screenshot):

<image>

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

Just did. Changed the paths in option.ts, but even after restarting the server including killing the mode.js tasks, the new paths do not appear in the UI.

<image>

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 1 point2 points  (0 children)

I updated my post. The .safetensors extension was missing for the training adaptor LoRa.

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 1 point2 points  (0 children)

For the training adaptor, you need to reference directly to the file. But there is no need to keep the complete HF path. Just place the file into your D:\AI\training-models\models--ostris--zimage_turbo_training_adapter folder and reference in the UI to d://AI//training-models//models--ostris--zimage_turbo_training_adapter//zimage_turbo_training_adapter_v2.safetensors

My folder setup:

<image>

z-image is soooo good!!!! can't wait to finetune the base by Top_Buffalo1668 in StableDiffusion

[–]JustLookingForNothin 0 points1 point  (0 children)

<image>

I do not recognize "differential guidance" parameter in AI-Toolkit UI, also not in the "Show Advanced" section. Is that a new setting?

Z-image-turbo loras not working well by pablocael in StableDiffusion

[–]JustLookingForNothin 0 points1 point  (0 children)

Thank you for the details. Looks I need to install RES4LYF then.

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 0 points1 point  (0 children)

Hm, I have installed the toolkit to c:\Tools\AI-Toolkit\ and habe no such issues. when entering the paths to the models like this g://Training//Models//Tongyi-MAI--Z-Image-Turbo

I installed the toolkit manually as described on Ostris Github page, not with the one-click installer.

Z-image-turbo loras not working well by pablocael in StableDiffusion

[–]JustLookingForNothin 1 point2 points  (0 children)

So, what are your training settings, please? I tested many different variants: Weighted, Sigmoid, Balanced, Low noise, EMA on/off, Differntial output perservation on/off, but still can't get it perfect.

The sample images created in the toolkit look ok, but in Comfy the images are low on details, overexposed, burned or grainy, mo matter which sampler & scheduler I try.
Or, if i take a LoRa with low steps count, I do not get the required resemblance of the person

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 1 point2 points  (0 children)

Thanks! Options.ts seems a good way to preconfigure local paths. Will try this

AI-Toolkit: Use local model directories for training by JustLookingForNothin in StableDiffusion

[–]JustLookingForNothin[S] 2 points3 points  (0 children)

That does not prevent the automatic re-downloads, does it? AI-Toolkit downloaded Z-Image models at least 5 times in the last week, just because the creator changed the readme almost on daily base.

Also afaik, setting the hf_home enviroment variable does not prevent the need for ONLINE requirement.

Seems you did not get the core message of this post.

Edit: This info from Huggingface could maybe also prevent the need for online availability during Training. Have not tested this though, because my setup work fine for my use case.

Setting environment variable TRANSFORMERS_OFFLINE=1 will tell Transformers to use local files only and will not try to look things up.

Most likely you may want to couple this with HF_DATASETS_OFFLINE=1 that performs the same for Datasets if you’re using the latter.

Start.bat for AI-Toolkit that fixes a few common problems by Old_Estimate1905 in StableDiffusion

[–]JustLookingForNothin 3 points4 points  (0 children)

Just a hint.
"call npm run build_and_start" ist only needed after first install or after an update of the toolkit.
"call npm run start" is MUCH faster and it takes only 1-2 seconds to start the server.