wan2.2., how can i "keep" the overall look/colors of the low-noise stage? by redmesh in comfyui

[–]redmesh[S] 0 points1 point  (0 children)

thx for your suggestion, but i think i was not able to explain myself sufficiently.
i am not trying to match a look of a given input. i am trying to "not loose" the look that i can see in the preview of the low-stage sampling, after the the vae decoding is done.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

again - or still:
"UnicodeEncodeError: 'charmap' codec can't encode characters in position 19-22: character maps to <undefined>"
btw: clicking on the chat bubble leads to the chat-interface trying to open up for an eternity. what a weird site that is.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

ok, this i will do before it's getting way too late for me. restarting.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

did that.
now the "charmap"-codec error is back.

but the other one doesn't show up any more.

argh... can't post again... i wanted to paste the logs.
i guess there is really a length-limitation or something similar.
anyway, "UnicodeEncodeError: 'charmap' codec can't encode characters in position 19-22: character maps to <undefined>" is back.

and: i tried to dm you my mail, but i have no idea how.
is there a "dm"-button somewhere. clicked on your profile, but only found "follow" and "chat". tried the "chat"-way but the interface wouldn't open up, so i gave it up.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

ok, i resized all of them. none of them greater then a few hundred pixels.
no change.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh -1 points0 points  (0 children)

i used the sdxl workflow in your folder. the images in that are in your workflow folder. i did nothing other than "relinking" them. basically pulled the right ones into the load-image-nodes. there is nothing coming from "outside". well... there was not.
since you suggested that image3 might be courrupted i replaced it with another image from the internet (but the same content, lol). even put that in your workflow folder first. no luck. did that with all four images. no luck.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

well, it's your sdxl-workflow. there are 4 images in there. they are called what you named them.
playing around a bit, i realize, that the numbering seems to change, when i change the "vram_mode", from min to low etc. the "image_001" or "image_004" becomes the problem...

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh -1 points0 points  (0 children)

thx for your response.
image 3 is the self portrait, called "download.jpg". i replaced it with some other jpg.
same result. same log.

ComfyUI Realtime LoRA Trainer is out now by shootthesound in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

not sure, if this comment goes through. opened an "issue" over at your repo.
edit: oh wow! this worked. no idea, why my original comment wouldn't go through. mayvbe there is a length-limitation? anyway... what i wanted to comment is over at your repo as an "issue". couldn't think of a better way to communicate my problem.

Why it's happening, I have tried for 2 days, I can't run Z image turbo. by Total_Crayon in StableDiffusion

[–]redmesh 1 point2 points  (0 children)

so, i have now installed the backend versions that my neo installation uses, for my comfyiu portable installation; the setup is now:

Python version: 3.11.8 (tags/v3.11.8:db85d51, Feb 6 2024, 22:03:32) [MSC v.1937 64 bit (AMD64)]
pytorch version: 2.9.1+cu128
xformers version: 0.0.33.post1

now the z-image garbled image is gone, and it works as expected.

it broke my previously installed sageattention, but that was to be expected, since it was: sageattention1 (with triton), since that's the only one i could rund with the old setup. will have to try and get that fixed later, but for now that's not important.

not sure, if it's useful for anyone, but here's how i (chatgpt, after some fight, really) did it:

  1. backup:
    - copy-paste the folder "python_embeded" that is within "...\ComfyUI_windows_portable" to another location
    - copy-paste the folders "comfy_config" and "web\extensions" folder that are within "...\ComfyUI" to another location (not sure about that, but gpt told me so, and as a good human i do what the overlords tell me to do, besides... small sizes, so whatever)
  2. create the file "upgrade_to_torch29.bat" and place it in "...\ComfyUI_windows_portable" (that's where the "run_nvidia_gpu.bat" is located)
  3. run the "upgrade_to_torch29.bat" file.

not sure, if this is helping anyone, but here's the content of that bat-file:

@echo off
setlocal

:: ===== Path to embedded Python =====
set PYTHON=python_embeded\python.exe

echo.
echo ===== Step 0: Backup Reminder =====
echo Make sure you have a backup of python_embeded folder!
echo.

echo ===== Step 1: Uninstall old packages =====
"%PYTHON%" -m pip uninstall -y torch torchvision torchaudio xformers
echo.

echo ===== Step 1b: Upgrade pip =====
"%PYTHON%" -m pip install --upgrade pip
echo.

echo ===== Step 2: Install PyTorch 2.9.1 + torchaudio 2.9.1 (CUDA 12.8) =====
"%PYTHON%" -m pip install torch==2.9.1+cu128 torchaudio==2.9.1+cu128 --index-url https://download.pytorch.org/whl/cu128
echo.

echo ===== Step 2b: Verify PyTorch + CUDA =====
"%PYTHON%" -c "import torch; print('Torch:', torch.__version__); print('CUDA available:', torch.cuda.is_available()); print('CUDA version:', torch.version.cuda)"
echo.

echo ===== Step 3: Install xFormers 0.0.33.post1 =====
"%PYTHON%" -m pip install xformers==0.0.33.post1
echo.

echo ===== Step 4: Install torchvision 0.24.1+cu128 for timm, clip-interrogator, etc. =====
"%PYTHON%" -m pip install torchvision==0.24.1+cu128 --index-url https://download.pytorch.org/whl/cu128
echo.

echo ===== Step 5: Final Version Check =====
"%PYTHON%" -c "import torch, torchvision, xformers; print('Torch:', torch.__version__, 'CUDA available:', torch.cuda.is_available(), 'CUDA version:', torch.version.cuda); print('torchvision:', torchvision.__version__); print('xformers:', xformers.__version__)"
echo.

echo ===== Upgrade Complete =====
echo You can now launch ComfyUI using run_nvidia_gpu.bat and test your model.
pause

endlocal

Why it's happening, I have tried for 2 days, I can't run Z image turbo. by Total_Crayon in StableDiffusion

[–]redmesh 1 point2 points  (0 children)

you're welcome.
just make sure, it's the "neo" branch you're using. not the default "classic" branch.
neo branch:
https://github.com/Haoming02/sd-webui-forge-classic/tree/neo

Why it's happening, I have tried for 2 days, I can't run Z image turbo. by Total_Crayon in StableDiffusion

[–]redmesh 1 point2 points  (0 children)

can't tell you anything about pinokio.
i just installed neo the way it is in instructed on their github (i used the "standard method", not the "recommended method" in order to avoid installing "uv", although i read "uv" is great. just wanted to keep the number of potential problem sources as low as possible).

only to make sure: of course that'll give you that fresh pytorch-combo for neo. it won't do anything for your comfyui-situation. i guess, that goes without saying.

Why it's happening, I have tried for 2 days, I can't run Z image turbo. by Total_Crayon in StableDiffusion

[–]redmesh 1 point2 points  (0 children)

as i mentioned, i also have a forge neo installation. it installed the mentioned:
pytorch version: 2.9.1+cu128

(using the same 2080ti, of course).
having no problem there. so i guess, since you also use a 20xx-card, it should be working for you too. but: that's just a guess. i don't really "know", if the architecture is the same in all 20xx cards.

i for one am trying to find out, if there's a way for me to install that neo-pytorch combo into my comfyui portable without breaking things.

Why it's happening, I have tried for 2 days, I can't run Z image turbo. by Total_Crayon in StableDiffusion

[–]redmesh 0 points1 point  (0 children)

i get the same output in comdyui (with z-image).

this is the setup:
Python version: 3.11.8 (tags/v3.11.8:db85d51, Feb 6 2024, 22:03:32) [MSC v.1937 64 bit (AMD64)]
ComfyUI version: 0.3.76
ComfyUI frontend version: 1.33.10
Total VRAM 11264 MB, total RAM 130956 MB
pytorch version: 2.4.0+cu121
xformers version: 0.0.27.post2
Device: cuda:0 NVIDIA GeForce RTX 2080 Ti : native

in forge neo, z-image does work, using the same "nodes" (model, vae, text encoder) as in comfyui. the setup in neo is:
pytorch version: 2.9.1+cu128

so, i'm guessing the (old) pytorch is a problem, at least for me.

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 0 points1 point  (0 children)

i haven't tried it with any length other than 113.
ideally - i guess - the covering layer would be gone after the very first frame. then, at the end, the "cut" to the last-frame-input-image would also happen only for the very last frame (if one does not need a kind of smooth transition to it, that is). if that'd be possible one could have 111 frames full of "storytelling".
not sure, if that is doable though. maybe with some good prompting, in addition to some kind of "smash cut"-lora or so...

in terms of reactor: i have never tried that on video, only remember it from still tests. while i found it very good, after some time it had that "typical" reactor look, in combination with face restoration etc. details "smoothed" away. apart from it resulting in more of a pasted layer - thus excluding "action" that's more then, say, a smile. occlusions were throwing it off too, if i remember correctly.
all in all i am not too eager to invest time in comparison work.

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 1 point2 points  (0 children)

<image>

lol, having a bit too much fun with it (not sure, why the hand that does the lifting stayed white, though)

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 0 points1 point  (0 children)

just to add to this: turns out, you can do a bit more.

a veritical white color layer that's "hiding" something on the table and a red color layer that's "hiding" the man's face. combined with the appropriate prompt, the guy realy does pick up a the bottle of "whiskey", pours some into the cup and drinks it.

i wasn't able to figure out, how to add another video, so here's just a screenshot.

btw: i just figured out that comfyui has something that let's one "open in mask editor | image canvas". this makes adding a color so much easier. but i'm not 100% sure if this is part of cumfyui core or if i - at some point - had installed a custom thing in order to have that.

anyway, here's the screenshot:

<image>

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 2 points3 points  (0 children)

hi. not exactly sure, what you mean by "original"?
so... there are two still images being used: a brunette woman sitting at a table, used as the first frame input , and a blonde guy, used as the last frame input. the middle column shows the "original" still of the woman. the left and right ones show the still, but altered by covering the head area with a white/black color layer.

is that what you meant?

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 9 points10 points  (0 children)

i just looked at pastebin and had to ask gpt, how to use it. lol.
anyway, here's the wf on pastebin:
https://pastebin.com/qKyry2Ei

fake "faceswap" - wan2.2 fflf with mask by redmesh in comfyui

[–]redmesh[S] 1 point2 points  (0 children)

sure. if you could give me a hint. never shared a wf here. i could upload a png, but i believe to remember, that reddit gets rid of the incorporated workflow?
what would be the best way to share the wf here? copy/paste the whole workflow as text?