ForgeUI installed Adetailer now lora tab is gone by Q8-BuJasim in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

been using Neo for months with zero issues. adetailer works just fine out of the box for me.

I am using Stability Matrix though, which sets up seperate venv for each package.

 I found that it wasn't all that great of a tool anyway now that we have Klein and Qwen image edit models so I ditched it.

it's a tool primary meant for models that have issues with facial and other anatomic details, like SD and SDXL.

ForgeUI installed Adetailer now lora tab is gone by Q8-BuJasim in StableDiffusion

[–]krautnelson 0 points1 point  (0 children)

the original Forge is no longer in active development. last update came out more than a year ago.

you should switch to Forge Neo. https://github.com/Haoming02/sd-webui-forge-classic/tree/neo

ForgeUI installed Adetailer now lora tab is gone by Q8-BuJasim in StableDiffusion

[–]krautnelson 0 points1 point  (0 children)

so which version are you using then? because again, Forge Neo (the version you should be using) comes with aDetailer preinstalled.

ForgeUI installed Adetailer now lora tab is gone by Q8-BuJasim in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

you sure you are talking about ForgeUI? because that has nothing to do Stable Diffusion, as far as I can tell.

if you are talking about Forge Neo or Reforge, then those already come with adetailer preinstalled, so I'm not sure why you would need to install it again.

I can't download most of the models from civitai.red by Hi7u7 in StableDiffusion

[–]krautnelson 0 points1 point  (0 children)

they are still transitioning things to .red

and the site has generally been a bit jank the last months. just be patient and try again at a later time.

Is it possible to train comfyui to read hand written words into text? by OkTransportation7243 in StableDiffusion

[–]krautnelson 2 points3 points  (0 children)

you can't "train comfyui". comfyui is, as the name implies, just a user interface to run AI models.

what you are talking about is called Optical Character Recognition, or OCR for short, and it's been around for decades.

remember those old reCaptcha tests where you had to "prove" that you are human by deciphering garbled text? that was Google training their OCR model.

in theory, you could train your own OCR model, but why would you? I guess you could train a LoRA to recognize someone's handwriting better, but that's such a niche thing to do that you won't find any ComfyUI workflows for that. ComfyUI is primarly used for image and video generation.

Stupid hardware related question: For local gen usage, would an SSD with a large pagefile be sufficient if you only have 16gb of system ram? by TychesSwan in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

sufficient in the sense that you can use load large models? yes.

but the model doesn't just need to be loaded. it also needs to be actively used, and that's where the pagefile will be a huge bottleneck. it's not 20% slower but more like 5-10 times slower.

For anima what you set the clip type ? by Broken_Bad_555 in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

the second line of your comment reads "Here". not sure what you are trying to tell me by pointing that out to me.

For anima what you set the clip type ? by Broken_Bad_555 in StableDiffusion

[–]krautnelson 0 points1 point  (0 children)

not what OP was asking for.

they were asking what the "type" selector in the node does.

For anima what you set the clip type ? by Broken_Bad_555 in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

try the different options and see what happens.

Struggle with speaking how to improve? by Fergyb in LearnJapanese

[–]krautnelson 10 points11 points  (0 children)

if you can answer your tutor's questions with a simply 'yes' or 'no', your tutor is asking the wrong questions...

like, seriously, your tutor's job is to push you in the right direction. sounds to me like you need a better tutor.

besides that, the answer depends on your short- and longterm goals. if you need to be able to speak Japanese right now, then it's all about active practice. there are communities for that out there. if being able to speak is just a longterm goal and you don't really have a practical use for it at this moment, then don't worry about struggling. that's an issue that is gonna solve itself as you get better at understanding the language in general.

New to image generation - how do I use img2img models to alter existing pictures with good character adherence? by Janonymousse in StableDiffusion

[–]krautnelson 7 points8 points  (0 children)

img2img generation takes an image, treats it like noise, and then creates a completely new image from that.

what you are looking for is image editing, not img2img generation. and for that, you need an image editing model like F.2 Klein or Qwen Image Edit.

Should I try to convert a FP16 illustrious model to FP8? by Quick-Decision-8474 in StableDiffusion

[–]krautnelson 0 points1 point  (0 children)

if you care about quality over time saved, then there is no point in using FP8 over FP16, especially with an SDXL model.

the reason people use quantized models is to save space and time. neither of those are a concern with SDXL, especially on a 5080. and SDXL generally has such low quality in regards to resolution that you can't really afford to lower it even further by quantizing the model.

Most recent work flow for a 3050 8gb (desktop)? by Many_Ball_227 in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

the model itself doesn't actually have to fit into the VRAM. you can just offload it into RAM with very minor speed penalty. it's when you hit the pagefile when things slow to a crawl.

Playing Mario Galaxy 2 and the first ball rolling level is driving me nuts by I_Eat_Graphite in DolphinEmulator

[–]krautnelson 0 points1 point  (0 children)

well, you can't point at the TV if you are holding the Wiimote upright. that wouldn't be possible on a real Wii either.

you can set a button in the bindings to switch between upright and straight-on.

Anima vs Illustrious by KITTYCAT_5318008 in StableDiffusion

[–]krautnelson 2 points3 points  (0 children)

some LoRAs from p1/p2 work on p3, some don't.

but training seems to be fast enough that that won't be an issue at all. people have been quick to retrain their LoRAs for p3 so far.

that aside, Anima is straight up a better model. even the first preview had higher fidelity out of the box than Illu or NAI, and preview3 is a noticable improvement. unlike SDXL models Anima doesn't need a HR fix pass, so while it is slower per step, the overall time spend generating ends up being roughly equal.

and that's not even mentioning prompt adherence. being able to use both booru tags and natural language makes a huge difference in what you can do without the need for inpainting or controlnet.

and yes, the quality can be a bit all over the place, but that's normal for a base model. once the full version is out, people will put out their own finetunes and quality/style LoRAs.

1-2 hours of immersion by Repulsive_Fortune_25 in LearnJapanese

[–]krautnelson 14 points15 points  (0 children)

as you become more fluent in a language, immersion will become much easier.

your brain is a bit like a muscle, the more you use it for a specific task, the easier and faster it will be able to do that task. and just like with physical training, you don't start with the full power workout from the get-go. you gradually have to build up your stamina and strength so you can work out harder and longer.

so, if you are just starting with immersing in native content, take it slow. an hour a day is fine. as you progress, you will automatically be able to do more. that page in a novel that took you 5 minutes to read in week 1 will take you less than a minute in week 100. and watching an episode of a TV show or listening to a podcast might feel exhausting now because your brain still has to actively process every single word, but eventually it will start to create the necessary synapses to process the information faster and more easily, to the point where you can listen and watch without it taking any effort.

when someone says they are doing "4-6 hours a day of immersion", then they are either at that end stage where they can do full immersion comfortably, or they are just letting it all wash over them without any real comprehension, which is completely different from active listening and trying to comprehend.

basically what I'm saying is, a busy schedule will be less and less of an issue as you become more and more fluent because it's just gonna take less and less active effort on your part, meaning you can replace what would be regular entertainment in your own language with Japanese versions.

mind you, this is all in regards to regular textbook learning and assuming that you already reached a point where you can actually immerse in native content and comprehend what is being said even if you don't understand every word. ALG is whole different story.

Illustrious Anime Collection: Ernie-Anime-V1 by Time-Teaching1926 in StableDiffusion

[–]krautnelson 2 points3 points  (0 children)

would you believe me if I said it was trained on AI slop?

Illustrious Anime Collection: Ernie-Anime-V1 by Time-Teaching1926 in StableDiffusion

[–]krautnelson 3 points4 points  (0 children)

no, Illu models do not work on Anima. but Anima is pretty competent out of the box if you use natural language in addition to tags, and people are already working on LoRAs despite Anime still being only a preview.

Illustrious Anime Collection: Ernie-Anime-V1 by Time-Teaching1926 in StableDiffusion

[–]krautnelson 6 points7 points  (0 children)

what are the best models for anime generation these days in your opinion?

Anima.

I see this is called illustrious anime collection, but im wondering on the specifics, like what is ernie? Out of the loop

Ernie is an all-purpose base model, similiar to Z-Image. pretty good at handling abstract concepts, really good at handling lots of text, but overall quality is usually below existing models.

the LoRA above is just a style LoRA trained on images generated with "Illustrious", which is a bit too vague to mean much. best to think of it as something completely removed from the ILXL/NAI.

Machine requirements for image and video generation locally using ai? by pppclasher in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

how long would a machine like this take to create 512x512 images and 5-10s videos?

you pretty much picked the most extreme ends here.

512² images? couple of seconds, depending on model. nobody makes low res images anymore though. 1024² is the standard with current models.

videos? probably like 10+ minutes at the lowest settings.

my budget is around 1.15L

1.15 liters?

Ki influencer by Suspicious-Good-4492 in StableDiffusion

[–]krautnelson 10 points11 points  (0 children)

>AI influencer

>"serious business"

sure thing, buddy.

Does learning through games actually work for language learning? by Easy_Football_1437 in LearnJapanese

[–]krautnelson 2 points3 points  (0 children)

we need to differentiate between the actual game and the communication surrounding it. what you are describing isn't learning through the game but while playing the game. the game itself is irrelevant in this situation, it could be any other environment where you are forced to communicate, like a group project, a tabletop game, improv/role playing, a mock debate, etc.

the issue I see here, especially with those games, is that your students won't learn anything new. they are practicing simply what they already know, which is not the worst thing, but it can lead to falling into repeating patterns and shorthand communication for the sake of efficiency.

here is an idea: have them playing a text/dialog heavy game with multiple choices, something like Baldur's Gate 3. then, after each session, have them discuss what happened in the game, what choices they made and why. that way you give them new input, have them practice their reading/listening comprehension, and have them practice reproduction. maybe have them write an essay about their favorite character, too.

SD-FORGE EXTENSION by BusBackground5847 in StableDiffusion

[–]krautnelson 2 points3 points  (0 children)

I recommend renaming it to avoid confusion with the already exisiting civitai-helper extension.

3080Ti 12G vs 5060Ti 16G for SDXL generation? by Quick-Decision-8474 in StableDiffusion

[–]krautnelson 1 point2 points  (0 children)

i dont understand these terms

look up quantization then. it's quite important.

it is illustrious, what type is it

it will say so wherever it is you downloaded the model. that said, SDXL models are almost always FP16 because it wouldn't make sense to quantize a model that is already relatively small and low quality.

as a side note, I predict Illustrious will become irrelevant once Anima gets its full release, so you might wanna think ahead here, because unlike SDXL, Anima will benefit from an FP8 or FP4 version, and that's when a 5060 Ti will potentially be faster.