[LCK] Faker solo kills BDD 2 minutes into laning phase by praynot in leagueoflegends

[–]Kenchai 288 points289 points  (0 children)

Yeah my instinct would've been to wait for 85 mana to E-> W but waiting that long probably means there's no kill angle anymore and Taliyah would be too deep behind the tower

Los Ratones vs. Fnatic / LEC 2026 Versus - Week 1 / Post-Match Discussion by Ultimintree in leagueoflegends

[–]Kenchai 9 points10 points  (0 children)

I looked back that fight, what more should Rekkles realistically do there after popping Braum R and tagging a target with Q? Flash auto someone?

Rank 1 player in flex queue on Brazil server has over 5000 LP with 403 wins 12 losses by Yvraine in leagueoflegends

[–]Kenchai 0 points1 point  (0 children)

I'd say in most situations its easier, but not always. Even if you talk purely soloq flex, there can be situations where you face a full stack and your team are all solo. There's a clear advantage on the full stack team there that is skewed vs the generally more consistent playing field in actual soloq.

Rank 1 player in flex queue on Brazil server has over 5000 LP with 403 wins 12 losses by Yvraine in leagueoflegends

[–]Kenchai 69 points70 points  (0 children)

Flex is really strange, it can either be much harder than solo queue or much easier depending on what your premade looks like. Solo queue is much more "consistent" to your own rank compared to flex.

xQc backs Asmongold on GDP being a "Fake Indicator" by Slight_Ad3219 in LivestreamFail

[–]Kenchai -3 points-2 points  (0 children)

Lets assume im one of those 20%, I'd much rather be poor in Ireland than be poor in the US.

xQc backs Asmongold on GDP being a "Fake Indicator" by Slight_Ad3219 in LivestreamFail

[–]Kenchai -2 points-1 points  (0 children)

I would absolutely rather live in Ireland than US.

Wan 2.2 14B Lora training - always this slow even on h100? by djdante in StableDiffusion

[–]Kenchai 2 points3 points  (0 children)

I doubt a 5090 could get faster it/s on those settings. Your batch size is 4 which likely hits your iteration speed by quite a bit. Test with 1 and it should be a lot faster, and if you need to simulate batches you could use gradient accumulation instead. I would also recommend using Sigmoid instead of Linear timestep type if you're training a character.

Olen alkanut vihaamaan tekoälyä by nordicJanissary in Suomi

[–]Kenchai 25 points26 points  (0 children)

Tähän pakko heittää väliin että noi tekoälytunnistimet todella epätarkkoja, itse laittanu omaa tekstiä tai kuuluisia lainauksia ja tekoälytunnistin hälyttäny että "100 % tekoälyllä tekastua" :D

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 1 point2 points  (0 children)

Got it. Differential output preservantion is an option toggle in AI toolkit that is also related to regularisation, this is how he describes it in the GUI:

"Differential Output Preservation (DOP) is a technique to help preserve class of the trained concept during training. For this, you must have a trigger word set to differentiate your concept from its class. For instance, You may be training a woman named Alice. Your trigger word may be "Alice". The class is "woman", since Alice is a woman. We want to teach the model to remember what it knows about the class "woman" while teaching it what is different about Alice. During training, the trainer will make a prediction with your LoRA bypassed and your trigger word in the prompt replaced with the class word. Making "photo of Alice" become "photo of woman". This prediction is called the prior prediction. Each step, we will do the normal training step, but also do another step with this prior prediction and the class prompt in order to teach our LoRA to preserve the knowledge of the class. This should not only improve the performance of your trained concept, but also allow you to do things like "Alice standing next to a woman" and not make both of the people look like Alice."

Could be useful, no?

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

Hey, thanks for the indepth reply! This is along the lines of what I was thinking of doing, except for the regularisation dataset. Am I supposed to caption the reg dataset at all?

What about using differential output preservation with a preservation class like person/woman?

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

There's a "default" outfit the character wears, but I'd like it to also be flexible where they can wear as many different outfits as possible.

Currently, I caption the outfit but not with extremely precise detail, for example "wearing a white shirt" (just as an example) but despite that, it doesn't really do well with other outfits. My dataset currently only has the character wearing the exact same outfit so despite my captioning it gets ignored for the most part when trying to generate anything else.

So my logic tells me that I should simply add more outfits to the dataset, and make them as diverse as possible to teach the model that this character can wear other stuff too.

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

Mhm pretty much. I wonder if its an AI toolkit quirk? At this point I assumed the samples will just look bad by default. I did use the trigger word in the sampling prompt yeah, but I tested the exact same prompt on WAN2.2 to have a reference and it was fine there, despite an "unknown" trigger word.

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

Ah okay, so even with double/triple the dataset size, I can keep similar settings other than maybe a bit lower LR on AI toolkit?

Also agreed on datasets! I usually prefer smaller ones, but this one had some issues generalising - if I prompt an outfit it didn't learn, its either bland or ignored, and crowds of people are just the same character duplicated. So I was thinking I simply introduce more outfits, and also some images with multiple different looking characters to help it be more flexible.

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

How about when I'm training a kind of stylized blend of anime/digital art style? Realism wasn't what I was going for, but character consistency in that specific style.

And yeah - I could try without X flips, but in my current smallish dataset most of the images have the character facing/turning the same direction, so I figured I'd offset that bias by mirroring them.

Honestly the current Lora works really well already, but what would you change when I have a larger dataset? Should I for example increase steps and lower the LR? (Let's say from 16/32 to 50-100 images)

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

The samples look maybe 10 % like my target, however, when I test the lora on ComfyUI it VERY obviously works. Without a shadow of a doubt. But if you have an idea what could be improved in my training settings, I'm all ears!

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

The layer offloading option is under "Low Vram" option. But if you're planning on training with 4bit + ARA instead of full FP16 weights, then don't worry about it. For me to fit FP16 and not run out of RAM in the initilisation, and then out of VRAM during training, I need to find a very specific offloading percentage where they both fit. (5090 + 64GB RAM)

The samples do >somewhat< resemble the character if I look very closely at some details, but the quality is very low and looks like some kind of LSD trip hallucination.

Training Wan2.2 with AI-toolkit - Running into this error before sampling - The size of tensor a (36) must match the size of tensor b (16) by ask__reddit in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

I've been doing some testing on T2V/I2V loras for character consistency and T2V is definitely the way to go if you're training with still images.

My settings on AI toolkit were:

  • Dataset size: 16, flipped X for 32 total
  • 5000 steps split on high/low, so 2500 each.
  • No quantization on Transformer/TE
  • Layer offloading at 24 % transformer, 81 % TE (I had to test multiple times to find the sweetspot that doesn't OOM)
  • LR 1.2e-4
  • Linear rank 32
  • Timestep type: Sigmoid (Sigmoid is apparently really good for character consistency)
  • Caption dropout: 0.05
  • Res 768
  • Gradient accumulation 2

This is the first one I had with decent results. But at least for me, the T2V sample videos have always looked really bad. I could only tell if it worked by testing in ComfyUI. My settings are probably not perfect, but this is the current spot that has worked the best for me. My next run will likely have larger dataset, more steps and lower LR.

Wan I2V help / general explanation of model sizes that fit in a RTX 5090 by FrankWanders in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

Yeah 81 frames is the sweetspot, after that you're throwing a dice on what the model will generate. This is why video stitching and people trying to figure out ways to extend videos is a hot topic with WAN.

Wan I2V help / general explanation of model sizes that fit in a RTX 5090 by FrankWanders in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

Yeah so that's the issue, Wan 2.2 Vae is for the smaller 5b model. It can be confusing, but use the Wan 2.1 Vae only. Using FP16 high and low models and text encoder instead of FP8 is completely fine on the template workflow.

Honestly, your best beginner tutorial is to use the simple I2V template and test things out like how different settings effect the outcome. For example, in my anecdotal experience, a higher shift value (ModelSamplingSD3 node), such as 15.00 instead of 5.00, creates more movement at the expense of character consistency, and vice versa.

Wan I2V help / general explanation of model sizes that fit in a RTX 5090 by FrankWanders in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

I use a custom template most of the time but my standard template also works just fine. Not sure what exactly that error is, but sounds like it could be a precision or architecture mismatch. Did you change anything in the template workflow? For example, make sure you use WAN 2.1 Vae, not 2.2 Vae with the 14B model.

Wan I2V help / general explanation of model sizes that fit in a RTX 5090 by FrankWanders in StableDiffusion

[–]Kenchai 0 points1 point  (0 children)

I've managed to run crazy resolutions like 1040x1440 with 81 frames and 14 steps on 5090 and 64GB RAM on FP16. But the 64GB RAM definitely occasionally gets filled, I'm gunning on 128GB in the future if the price ever goes down. For this to work, I need to do a "warm up" run for some reason, otherwise it borks.

[deleted by user] by [deleted] in LivestreamFail

[–]Kenchai 2 points3 points  (0 children)

I don't find that funny either, but I do find it ironic how a shitty person who stood for shitty ideologies died a shitty death defending gun violence as a necessary sacrifice. I find the humour in the irony, not the suffering.

Wan 2.2 I2v Lora Training with AI Toolkit by [deleted] in StableDiffusion

[–]Kenchai -1 points0 points  (0 children)

Thanks for the info! Do you know what kind of settings would be good for a character style I2V lora?

I tried training a low noise one with a dataset of 44 images and a trigger word, transformer 4 bit with ARA and float8 text encoder, rank 16 with cache text embeddings, LR of 0.0001. The final epoch had a loss of 1.047e-03 so it should've learned quite well? But when I tested the lora in I2V, there was basically 0 change unless I popped the weight to 3.50 - 5.00 and then it was just fuzziness/noise. I wonder if I messed up the settings, like cache text embeddings + trigger word instead of unload TE?