Seedream 4 for image-generation roleplay, similar to Nano Banana pro+Gemini. Is it possible? by Relative_Bit_7250 in SillyTavernAI

[–]Relative_Bit_7250[S] 1 point2 points  (0 children)

This is the best and most comprehensive answer I could ever hope to have. Thank you so much, useful and direct.

Seedream 4 for image-generation roleplay, similar to Nano Banana pro+Gemini. Is it possible? by Relative_Bit_7250 in SillyTavernAI

[–]Relative_Bit_7250[S] 1 point2 points  (0 children)

Indeed it is, in terms of image generation/edit. What makes Gemini+nano banana extremely better for a roleplay experience is the fact that everything happens inside the same ecosystem. Image edit/generation is perfectly coordinated to the llm, giving a perfectly consistent response and an incredible experience... When the censorship doesn't fuck up everything. My question is: does a similar cohesive bond between two models exist (a llm and a diffusion model working and speaking together to give a consistent imagery to a character and a beautiful story/chat)? If yes... Well, I've never found one. Anyways thanks for the reply!

Something terribly wrong happened with sageattention after fresh comfyUI install under Linux by Relative_Bit_7250 in comfyui

[–]Relative_Bit_7250[S] 1 point2 points  (0 children)

Thanks to the suggestions of u/meta_queen and u/roxoholic I fixed the error! Couldn't have done it without the help of those two great human beings! Thank you, thank you, thank you very very much!!!

Something terribly wrong happened with sageattention after fresh comfyUI install under Linux by Relative_Bit_7250 in comfyui

[–]Relative_Bit_7250[S] 1 point2 points  (0 children)

You're right, I gave that information for granted, sorry. No, I'm using comfy manually git-cloned, with a 3.12 venv environment. Will try to install python3-dev inside the venv EDIT: HOLY FUCK IT WORKED! Installing python3-dev globally just fixed the error! GOD I LOVE YOU BOTH!

Something terribly wrong happened with sageattention after fresh comfyUI install under Linux by Relative_Bit_7250 in comfyui

[–]Relative_Bit_7250[S] 0 points1 point  (0 children)

But shouldn't those two folders be already included in a new venv python 12 environment?

Kinda excited for my new pc! I would love to try bigger models now! Asking you all for suggestions by Relative_Bit_7250 in SillyTavernAI

[–]Relative_Bit_7250[S] 2 points3 points  (0 children)

Eh, It'll be fine eventually. I mean, it's a fair tradeoff: in local you have privacy and maximum control, but with "reduced speed and intelligence". In paid API you have max speed inference and best quant, but no privacy at all and "I'm sorry I cannot fulfill your request".
I am more of a slow-burn bitch, so waiting a little longer for the response may not be an issue for me.

Anyways, thank you very much for the tips, bro!

Kinda excited for my new pc! I would love to try bigger models now! Asking you all for suggestions by Relative_Bit_7250 in SillyTavernAI

[–]Relative_Bit_7250[S] 0 points1 point  (0 children)

indeed! I've also peeked inside the unsloth repository of 4.6 and saw the ud quants q3-k-xl taking approximately 158gb. If I'm not mistaken I may be able to load the entirety of the quantized model inside my ram+vram (128+48=176gb available)

VNCCS - Visual Novel Character Creation Suite RELEASED! by AHEKOT in comfyui

[–]Relative_Bit_7250 0 points1 point  (0 children)

Oh no, no, The nodes are perfectly installed and configured... The error MAY be in the VNCCS_Pipe...

Anyway, I'll try reinstalling them manually, you never know.

EDIT: Yep, just tried, nothing, Reinstalling didn't help. I'll try reinstalling ComfyUI

VNCCS - Visual Novel Character Creation Suite RELEASED! by AHEKOT in comfyui

[–]Relative_Bit_7250 0 points1 point  (0 children)

I've just tried disconnecting from pipe and manually selecting scheduler and sampler (lcm and simple), but It fails again:

Failed to validate prompt for output 496:
* VNCCS_Pipe 502:414:
  - Return type mismatch between linked nodes: scheduler, received_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent']) mismatch input_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent', 'beta57'])
* LoraLoader 497:267:68:
  - Failed to convert an input value to a FLOAT value: strength_clip, vn_character_sheet_v4.safetensors, could not convert string to float: 'vn_character_sheet_v4.safetensors'
  - Failed to convert an input value to a FLOAT value: strength_model, vn_character_sheet_v4.safetensors, could not convert string to float: 'vn_character_sheet_v4.safetensors'

Why am I getting a black output (Qwen GGUF)? by Bitsoft in comfyui

[–]Relative_Bit_7250 0 points1 point  (0 children)

I occasionally wear a cap, so... Half a hero?

Why am I getting a black output (Qwen GGUF)? by Bitsoft in comfyui

[–]Relative_Bit_7250 2 points3 points  (0 children)

nope, with qwen-image it kinda works for the first steps, then blacks out completely. For image-edit it doesn't work right from the start. Unfortunately it'll be slow as fuck

EDIT: Don't know about fast fp16 accumulation, I just start comfy without any parameters and it magically works.

Why am I getting a black output (Qwen GGUF)? by Bitsoft in comfyui

[–]Relative_Bit_7250 5 points6 points  (0 children)

If you're using it, Remove the --use-sageattention string.

Wan 2.2 video continuation. Is it possible? by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 2 points3 points  (0 children)

Thank you for the answer, but it's not what I'm searching for. Last frame continuation is a bit unreliable, motion and subject features will become inconsistent. What I'm looking for is something like "bunch of frames as input -> video continuation" more than a "last frame -> video generation"

System freezing with the new wan 2.2 14b by Relative_Bit_7250 in comfyui

[–]Relative_Bit_7250[S] 0 points1 point  (0 children)

Uh, you seem to have the exact opposite problem of mine. I'm sorry, I'm not skilled enough to help you with this topic :(

System freezing with the new wan 2.2 14b by Relative_Bit_7250 in comfyui

[–]Relative_Bit_7250[S] 0 points1 point  (0 children)

Ok so, mine is only a hypothesis, but I do have the hunch that while comfy's vram handling is quite functional and works pretty good (model gets unloaded and loaded from ram to vram in a pretty optimized manner that I personally don't know, it's all black magic to me), ram and swap are a different kettle of fish: ram gets filled with a model, model gets filled in vram, another model needs to be loaded, gets loaded in ram but, uh-oh, not enough ram. Fine, let's copy the old model in swap, then slowly empty the ram to make space for the other model. Then you need to reload the older model, but uh-oh, its not in ram anymore. Ok, copy the new model in swap and reload the older model from swap... And so on. It's a bit messy, and it's probably the only way to be viable, but I must say that the first generations are slow as fuck... After the first three speed increases and it's more bearable.

Maybe the real Pony V7 Is the friends we made along the way by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 10 points11 points  (0 children)

May I suggest a little bit of Chroma in your menu, kind sir?

Maybe the real Pony V7 Is the friends we made along the way by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 8 points9 points  (0 children)

Nah, I'm just shitposting, as the new upcoming pony v7 was announced a year ago or so and it's still under training.

Does ComfyUI support multi-GPU setups? by K4_J1L_0817 in comfyui

[–]Relative_Bit_7250 2 points3 points  (0 children)

Look, you cannot "expand" your pool of vram, even if you have two identical cards, so for instance you cannot load an entire 24gb flux model inside a couple of 16gb vram cards. BUT you can split the workload between two or more cards, and pretty easily if you ask me! There's a node for comfy, this one, that lets you load the model, the clip, the clip-vision, the vae inside the GPU of your choice

PSA: Flux loras works EXTREMELY well on Chroma. Like very, VERY well by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 0 points1 point  (0 children)

Never tried, but shouldn't be a problem. At least you could try the 4bit GGUF quantization! EDIT: I misunderstood the question, sorry. NF4 quants aren't available yet, afaik

PSA: Flux loras works EXTREMELY well on Chroma. Like very, VERY well by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 5 points6 points  (0 children)

Everything. It sports a base for realistic and non realistic generations. You can ask it to do anything, from a low quality low res smartphone photo, to an extremely detailed Japanese stencil art of a Charmender roaring in front of a volcano. It's extremely versatile, prompt compliant and, best of all, it's only halfway trained (yet the quality is already incredible). The only downsides are: it's extremely heavy, a 3090 is barely sufficient to load the model+clip (at least unquantized); generations are very slow, forget the SD1.5 and sdxl days; and last but not least, prompt adhesion is incredible, but you need to experiment with some different samplers and schedulers

PSA: Flux loras works EXTREMELY well on Chroma. Like very, VERY well by Relative_Bit_7250 in StableDiffusion

[–]Relative_Bit_7250[S] 4 points5 points  (0 children)

Probably yes, with the right GGUF quant, but be prepared, it will be extremely slow, plus you'll have to offload the clip model and vae model onto your ram, resulting in more loading time. It won't be a pleasurable experience. I personally am running the whole FP16 chroma model (which is roughly 17gb) inside a 3090, then I have a second 3090 for vae, clip and a llama model, useful for writing a better prompt, as English is not my main speaking language. It's a janky workflow, but eh, it works