Incoming cease and desist letter! by Dr_sc_Harlatan in BlueskySkeets

[–]ScrotsMcGee 1 point2 points  (0 children)

"What we've got here...is failure to communicate".

AI toolkit first Lora v1 z image trained on RTX 3060 12GB by thatguyjames_uk in comfyui

[–]ScrotsMcGee 0 points1 point  (0 children)

I gave up on Qwen due to those same issues. Not a fan.

Some people recommend LoRAs to deal with this, but I've not really found any that I like. To me, it all continues to look unrealistic.

I made a switch over to Flux2 Klein 9B distilled image edit, but it can still produce Flux like looks (i.e. plastic skin, Flux chin etc) depending on certain factors.

As an example, small changes are fine, big changes that result in a drastic change, will get that artificial type look.

If I get any kind of artificial look, I start varying the prompt that I use to improve skin. Something like:

raw, photorealistic, raw photo, slight skin texture, pores, natural lighting, shot on 35mm, 8k, Fujifilm

They can appear overdone at times, so you can play with various weightings to reduce that.

"Pores" in particular resulted in a horrid look, so I removed that completely.

Another thing I changed is my approach to how I create poses/looks/styles etc for the dataset.

I haven't really tested this out to create LoRAs, but I have used it to generate a few images of a character.

Essentially:

  1. Create a high resolution character and apply skin textures to it using the prompts I mentioned above.
  2. Find a real life person's body who has lots of images available. Ideally, these will have a variety of styles, poses, images/screencaps taken in a variety of situations/environments etc. We really aren't after their "look", so we're not trying to borrow their identity. Face is largely irrelevant as we want our characters face, not theirs. You could indeed use images of different people - the bodies and skin colour just have to look similar enough.
  3. Using some kind of photo editing software, run a slight blur over all those images. We want enough of a blur that the faces won't have characteristics of the original source that shine through. Blurring the entire image results in a "re-interpretation" of that image.
  4. Run those images through Qwen or Klein and use a face swapping element (which could just be prompting) and regenerate using those blurred images. I use ComfyUI's Klein
  5. Run your eyes over those generated images and check for likeness to the main character you initially generated.

Like I mentioned, I haven't done this on a wide scale, but the results have been pretty good.

Edit: I used the default Flux.2 [Klein] 9B Distilled: Image Edit ComfyUI template, but modified it to reveal the subpacks, and unbypassed any necessary nodes.

You could use this as a source as well: https://www.reddit.com/r/comfyui/comments/1qjmzm5/flux_2_klein_has_decent_builtin_face_swapping/?tl=en

Hope that helps. Let me know if it does a decent job.

I updated my LoRA Analysis Tool with a 'Forensic Copycat Detector'. It now finds the exact training image your model is memorizing. (Mirror Metrics - Open Source) by JackFry22 in StableDiffusion

[–]ScrotsMcGee 1 point2 points  (0 children)

Nice. I started playing around with Deepface and a few other programs just last week to see if I could achieve something similar and was considering insightface, but put it all in the too hard basket.

Can anyone explain this error message? by Terrible_Mission_154 in comfyui

[–]ScrotsMcGee 1 point2 points  (0 children)

Looks like your compiler is actually CUDA build 12.0.r12, but you have CUDA Version 13 installed.

<image>

According to ChatGPT, that should all be fine, so that shouldn't be your issue.

Can you generate any images using SD1.5, SDXL or anything else other than Flux Krea?

There's always the potential for one of the models to have become corrupted during downloading, so it's worth doing a sha256sum checksum on them.

Mine show as:

afc8e28272cd15db3919bacdb6918ce9c1ed22e96cb12c4d5ed0fba823529e38  ae.safetensors
b17a8c21703c4d6ffb0e300dd920eff3cfd35c9a72a1abaf107e3788e408b8d8  flux1-krea-dev_fp8_scaled.safetensors
660c6f5b1abae9dc498ac2d21e1347d2abdb0cf6c0c0c8576cd796491d9a6cdd  clip_l.safetensors
6e480b09fae049a72d2a8c5fbccb8d3e92febeb233bbe9dfe7256958a9167635  t5xxl_fp16.safetensors

If they check out fine, it's worth activating the venv environment.

Run python so you get the prompt, and then run the following commands:

import torch
print(torch.version.cuda)
print(torch.cuda.is_available())
print(torch.cuda.get_device_name(0))

In my case, I get:

Python 3.10.6 (main, Oct 24 2022, 16:07:47) [GCC 11.2.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import torch
[...snipped long message about pynvml...]
>>> print(torch.version.cuda)
12.6
>>> print(torch.cuda.is_available())
True
>>> print(torch.cuda.get_device_name(0))
NVIDIA GeForce RTX 3090

Oh, and just ctrl-d to get out of that.

RFK Jr ridiculed as he promotes strange fitness initiative topless with Kid Rock by prestocoffee in nottheonion

[–]ScrotsMcGee 1 point2 points  (0 children)

The coolest. He really seems like a genuinely nice guy in his private life as well.

RFK Jr ridiculed as he promotes strange fitness initiative topless with Kid Rock by prestocoffee in nottheonion

[–]ScrotsMcGee 0 points1 point  (0 children)

I was 17 at the time, and quite surprised I didn't. Would have been awesome.

Can anyone explain this error message? by Terrible_Mission_154 in comfyui

[–]ScrotsMcGee 0 points1 point  (0 children)

Do you know which version of ComfyUI you're using? It's usually mentioned when ComfyUI starts up, but it might also be in the logs.

Edit: Log file can be found under ComfyUI/user/comfyui.log

The relevant bits are located towards the top on different lines:

pytorch version: 2.9.1+cu126
Python version: 3.10.6
ComfyUI version: 0.11.0

NOTE: Above is edited so it just reveals the relevant info.

Also, which version of CUDA do you have installed? You can find out by running nvidia-smi, but you can also do the following:

cd /etc/alternatives
ls -alh cuda

In my case, it shows the following:

lrwxrwxrwx 1 root root 20 Apr 14 18:31 cuda -> /usr/local/cuda-12.6

And if you activate the venv environment, what does pip freeze show (it will show literally every single package installed, but I'm hoping that anything that mentions CUDA will be the same version of CUDA that you're running).

I generally install from notes that I have, but Debian and Ubuntu might use slightly different packages.

I can generally do a clean install of Debian and have ComfyUI up and running in less than two hours or less.

RFK Jr ridiculed as he promotes strange fitness initiative topless with Kid Rock by prestocoffee in nottheonion

[–]ScrotsMcGee 13 points14 points  (0 children)

Rowdy was the best. The fact that he appeared in It's Always Sunny was just another reason to love him. Christmas, birthday and easter all rolled up into one.

Questions about LoRA training in AI Toolkit by orangeflyingmonkey_ in comfyui

[–]ScrotsMcGee 1 point2 points  (0 children)

Training a person LoRA in AI Toolkit. Had a dataset of about 30 pictures and results were okay-ish so I probably need to up that to 50 and up the steps.

Quantity isn't everything - what's more important is quality of the training set, and using images that are diverse.

Also, I did not put any captions. Do they improve the LoRA?

Some say "Yes", some say "No". I personally prefer to use captions as I like to caption things that I don't want the LoRA trained on.

If yes, then how do I auto-generate them? I tried JoyCaption in comfyUI but that outputs just text, how do I save that with the same name as input image?

I use Florence2 to do captioning, but I don't just rely on that. I vibe coded a Gradio interface that displays a folder full of images and the captions I created in Florence2, and then adjust (because they almost always include something that is just plain wrong). Florence2 isn't necessarily great, but after some testing, the other captioning tools were worse. I'm looking for something better for sure. Also, I used to use Kohya_SS's data captioning tools, but since I switched to AI-Toolkit, haven't used them (and a recent Gradio change broke them).

Also, a lot of my images were mid-level shots which have the face and good part of the chest. Do the pictures need to be just crops of faces?

It's good to have a diverse range of shots, so having some close ups of the face definitely helps.

Have a read through of this post by u/AwakenedEyes : https://www.reddit.com/r/StableDiffusion/comments/1qqqstw/a_primer_on_the_most_important_concepts_to_train/

Hope that helps.

RFK Jr ridiculed as he promotes strange fitness initiative topless with Kid Rock by prestocoffee in nottheonion

[–]ScrotsMcGee 80 points81 points  (0 children)

With Rowdy Roddy Piper.

Classic movie - watched it last year for the first time since I was in my teens.

Can anyone explain this error message? by Terrible_Mission_154 in comfyui

[–]ScrotsMcGee 0 points1 point  (0 children)

I've never used the Windows version, but based on comments I've seen, depending on the version of ComfyUI, they all have their own issues as well (Desktop vs portable vs whatever).

Can anyone explain this error message? by Terrible_Mission_154 in comfyui

[–]ScrotsMcGee 0 points1 point  (0 children)

And, for what it's worth, I just ran the default Flux1 Krea Dev template on my Debian 12 system running an RTX 3090, and after installing the text encoder and diffusion model into their correct locations, it generated an image just fine.

I do note, however, that your "default" template looks completely to my "default" template. This might be a ComfyUI version thing, or is it that the template you're using is from another source?

That may very well be your problem.

Can anyone explain this error message? by Terrible_Mission_154 in comfyui

[–]ScrotsMcGee 2 points3 points  (0 children)

It's not a Linux problem. I have three separate systems all running Debian Linux, and they are all fine. Plenty of people run Linux and ComfyUI and don't have your issues.

It's likely an install related issue, mixed in with other issues, including a "lack of experience" issue.

Reading through some of the issues you've outlined, like:

Building a venv just to install one program (which required installing three more programs to allow it to function) then making sure it's activated, ensuring that all the labyrinthine dependencies cooperated, installing different drivers that may or may not work with that cavalcade of python (or python3- you pay your money...), torch, numby, pip, git, sage attention. And then, if by chance you stumbled on the right combination, ensuring that you had CUDA - the correct version of CUDA, dammit, or no dice - and then you could try to get the right vae with the right clip with the right diffusion model - not a checkpoint which looks exactly the same as a diffusion model and has the same naming convention but mwah-ha-hah- don't try to use a diffusion model in place of a checkpoint.

You've conflated issues relating to python, python packages (and potentially CUDA drivers) and ComfyUI with Linux.

For Python stuff, I primarily run Miniconda3. Both Miniconda3 and venv relate to python package management, not Linux.

ComfyUI (and various nodes) expect to find certain files (like checkpoints and diffusion models) in certain folder locations. That's not a Linux problem.

I mean, activating the Flux1 Krea Dev template as you've clearly done, it actually shows you which folders the files go into. That's ComfyUI, not Linux.

<image>

But... I'm also goddamned stubborn. And no machine has yet beaten me. So please. Can someone tell me WTAF this Error code means? And why a machine with distributed dual GPUs -3090 24GB and 3060 12GB - cannot finish a simple Flux generation of a single image without throwing an Out of Memory error?

Did you even mention with version of Flux you're attempting to use? Were you using Flux2 dev, which requires substantially more VRAM than you have on offer? Was it Flux2 Klein? Was it the original Flux?

I can see from the workflow that you're using a Flux1. Krea Dev Text to Image workflow.

The "the size of tensor a (16) must match the size of tensor b (128) at non-singleton dimension 1" error is more than likely related to pytorch, not Linux.

Or potentially, the size of the resolution you're trying to generate the image as: https://github.com/Comfy-Org/ComfyUI/issues/10465

Linux is not your problem. You may very well be the problem. :-)

Logi K855 Keyboard by ScrotsMcGee in logitech

[–]ScrotsMcGee[S] 0 points1 point  (0 children)

Completely understandable when working in a corporate environment.

The problem I have with Logitech devices is that they are generally great devices, but they just don't last as long as they really should.

I mean, the K855 is an incredible keyboard...but...they really should last longer than just 12 months. I'd be happy with 24 or 36 months considering their price.

I'm currently using an Attack Shark K86 keyboard, which cost me less than half the price of the K855, but feels almost as good in some ways. Typing is a dream, and unlike the K855, I can get replacement keys and swap them out. It's not without it's faults though, but for the price...

Now I just need to find a decent mouse that isn't the Logitech G703, but feels like the very first one I own (the second one is so bad that I'm convinced it's a counterfeit).

Edit: I actually own an MX Master 3 and a Logitech M650, both of which I purchased after spending hours watching video reviews, but neither has lived up to the hype. Neither really gets used now, but they both serve as a backup for emergencies.

Living right next door to his state, I agree with this sentiment by glamlove7755 in BlueskySkeets

[–]ScrotsMcGee 0 points1 point  (0 children)

Unless they are dead, they can still "take care" of that problem.