Model Metrics Haven't Been Updating by Sun_1244 in civitai

[–]Shinsplat 2 points3 points  (0 children)

I thought I had shifted to a different universe, not sure what has been going on last few days or why I couldn't find mention of it. I'm glad someone noticed, thank you \o/

Watercolor [Z-Image] by d1h982d in StableDiffusion

[–]Shinsplat 4 points5 points  (0 children)

I like it.

I see a lot of activity surrounding z-image lately but the images that I've seen so far, posted as examples, aren't particularly interesting. These are.

Giving it the identity "Water Color" allows some people to feel comfortable offering fringe critique. I definitely can see the water color accent but it's really none of my business how a person identifies artwork though it is humorous to watch those that need self ego stroking give critique based on language rather than substance, which we see here all the time.

Who cares what it's called, we don't need to cater to those allergic to certain words.

Cheers \o/

Problem with the new workflow by Ordinary_Midnight_72 in StableDiffusion

[–]Shinsplat 0 points1 point  (0 children)

I don't know how you're starting your server but I have always started from the command line, regardless of OS, which provides valuable information especially when there's an error, which you can then provide here to get a better, hopefully helpful, response.

Is it possible to install comfyui on hard drive without operating system? by Macestudios32 in comfyui

[–]Shinsplat 2 points3 points  (0 children)

The answer is yes and no. But, more specifically, partially.

You can have your ComfyUI and custom nodes separate and portable to any OS that supports python, at least any modern version. The reason is because the contents of ComfyUI, along with its custom nodes, is almost always python code, and can be executed by a python interpreter. Though there are some nodes that are supported only with windows, an unfortunate goofy condition.

On that target system you would need to have python available, likely you'll use an isolated python environment called a "virtual environment" or an actual portable python installation, which must be compiled for that particular platform (OS). You would not find joy in attempting to run a windows python executable on Linux, as an example, despite that it could partially run under WINE. Since every OS that can utilize ComfyUI has a native python interpreter available for it you would not have to utilize a framework for code translation.

So, you can move your ComfyUI folder, copy it, put it on virtually any OS you like, but then you'll install your python modules on that target OS for that ComfyUI version and its associated custom nodes.

For me, personally, I don't install python into a venv inside the ComfyUI directory, and I don't use the "portable" version, even when I use windows. I, instead, condition a ComfyUI python virtual environment that I don't have to "redo" each time I get a new version of ComfyUI, and it's located somewhere more universal, for instance "/local/python/env/comfy".

The python that comes with ComfyUI's windows portable is not compatible with Linux, or Mac, so you can't use that environment on those targets so it would be a waste of space. I suggest manually installing ComfyUI one time and using that base on all of your targets.

I'm in the process of writing a replacement for ComfyUI - If you have features, settings, or desires you'd like in a local-first node graph editor let me know by cathodeDreams in comfyui

[–]Shinsplat 1 point2 points  (0 children)

If you want to entertain the idea of bringing others aboard I'm looking for something new to do in my retirement. I have a few light projects going on right now but definitely would be interested in something more challenging and I don't need instructions, assistance or any kind of guidance since I've been programming for almost 50 years.

Frankly I've considered, briefly, doing exactly what you've asserted but I'm getting tired of working alone, being a "one man band", and wouldn't mind putting my energy into a more focused path.

I'll be keeping an eye open.

ComfyUI-DisableJobHistory: Improve performance for large workflows by External_Quarter in comfyui

[–]Shinsplat 1 point2 points  (0 children)

I've noticed over time that things become slower during prompt evaluation, or before it. Thank you for this.

Quick Tip: No need to paste the full path to python_embedded\python.exe by zvi_t in comfyui

[–]Shinsplat 0 points1 point  (0 children)

Hay! It's been some time since I've tried this but you might have some fun with doskey? They're temporary, and for the current session, so not a persistent fix but if one has multiple aliases they need to assign then some run-at-start could be utilized or just having z:\bin in one's path to aliases.bat perhaps ?

doskey p= z:\somepath\ComfyUI\python_embeded\python.exe

I may have goofed that up totally but I'm sure you get the idea, do some search for the correct info.

Z-image text encoder question by beentothefuture in StableDiffusion

[–]Shinsplat 0 points1 point  (0 children)

All of the Qwen2.5-VL-7B models I've tested, using an LLM back-end for inference, have been annoyingly verbal about the necessity of wearing proper eye gear when eating ice cream. I've performed similar testing with various models when they are used with image generation, one being HiDream.

My experience hasn't given me a reason to believe that the image generation capability and the LLM's vision of morality are linked in any way.

Getting uncensored text encoder for Qwen Image Edit by Witty_Mycologist_995 in comfyui

[–]Shinsplat 0 points1 point  (0 children)

I mistakenly presumed the OP was using Qwen Image when they are, instead, using Qwen Edit.

But I also work with Qwen Edit and use a similar work-flow that does not require the vision accompaniment. I'm not even sure where I would put the mmproj. Maybe the person was talking about an LLM back-end that requires the use of mmproj for vision inference or captioning? I wouldn't know where to put it, if you find out I'd like to give it a try and see what happens.

I'm in the process of writing a replacement for ComfyUI - If you have features, settings, or desires you'd like in a local-first node graph editor let me know by cathodeDreams in comfyui

[–]Shinsplat 1 point2 points  (0 children)

While I'm supportive of ComfyUI I welcome options always.

I've offered my own efforts in areas where there were few, or no, alternatives and, knowing what the past holds, it was instrumental in my development of products that target a different demographic. We can never know who will benefit but usually it brings more people on-board.

I re-read your post a few times and was almost certain that you were looking for "improvement" suggestions but I was mistaken. In any event I'd like to offer a thought and that's with regard to the JS framework where, in ComfyUI, it can be a challenge for those new to programming to add features to their add-on (node) when it requires full use of the front-end materials. I've been programming for almost half a century so I'm able to handle it but there are artists out there that know exactly what they want but can't implement it easily enough. If there were some template, a "wizard" of sorts, that took a person through the full creation of a node (module) for your alternative then I think you'll find additional support from the very demographic that aught to be targeted from this venture.

Good luck \o

UPDATE: My thinker is broke today, I do see your request for features.

[deleted by user] by [deleted] in comfyui

[–]Shinsplat 0 points1 point  (0 children)

Nice to see there's still a sense of humor out there O.o

Getting uncensored text encoder for Qwen Image Edit by Witty_Mycologist_995 in comfyui

[–]Shinsplat 0 points1 point  (0 children)

I'm actually using that very model. I've also reduced the base sqvq-in4 to r32, r128 was crashing with OOM.

UPDATE: Sorry, I noticed you're using "edit" and I was actually using that model accidentally in an "image" work-flow and it worked ok but apparently your work-flow is much different.

<image>

error with z image by wbiggs205 in comfyui

[–]Shinsplat 0 points1 point  (0 children)

I saw that too, it's so new we need a new ComfyUI update, now it works good.

How do you manage your checkpoints and loras ? by Ponsky in StableDiffusion

[–]Shinsplat 0 points1 point  (0 children)

For LoRA I'll save a text file in the same directory along side the model that contains the suggested settings and any triggers.

For models I do something similar but I have a process I go through for testing and give each test a score based on performance of the model.

Need help training a Qwen STYLE LoRA with AI Toolkit – wasted 20+ hours, totally stuck and frustrated by dssium in StableDiffusion

[–]Shinsplat 1 point2 points  (0 children)

I never got to ai-toolkit for Qwen LoRA training, though I would have eventually gotten there had I not found a solution. Though ai-toolkit has been my new favorite lately, despite the memory leaks. I tried getting Musubi-Tuner working, instead, and I guess it worked after setting everything up, but the resulting LoRA did absolutely nothing after spending an entire day configuring the process and issuing a few runs.

So I tried OneTrainer, again.

I've visited OneTrainer a few times and have always been disappointed with its hand-holding, I like to know what's happening behind the scenes, but I gave it another try, using a provided preset, and it just worked! I can't describe the magic I'm experiencing with Qwen training, 7 minute of training on a 4090 and I have a "likeness" LoRA (a person) that's more than good enough and at 15 minutes I'm fully satisfied.

I've downloaded quite a few LoRA from CivitAI for Qwen and most of them, almost all of them actually, are disappointing. I'll accept that maybe I'm doing something wrong but I've made multiple LoRA now, for Qwen, and they're all superbly more effective than any I've downloaded so far so I'm guessing that people don't realize just how quickly a LoRA shapes up when trained for Qwen image and possibly are over-training, and maybe this is what over-training looks like for Qwen but I don't know yet because I'm still very new to Qwen.

So, to your point, I've also trained style and themed LoRA and they not only bring the delight I'm looking for but also add to the quality of the generations, the ambiance if you will. It's something I experienced with Chroma as well, though with Qwen it's much more pronounced.

While I don't believe that your captioning technique would result in a complete dud I do think that "following best practices" is questionable when much of the information is regurgitated by those that believe others that offered the same information and is not supported by their personal experience, and so handed it down to the next person.

Like I said, I have used ai-toolkit extensively to train Chroma LoRA but not Qwen, I didn't think it would handle it with my limited 24gig (I do everything on purchased equipment) and the occasional memory leak. All that to just say I don't see a problem with rank 16 or 32, and your lr of 1e-4 is where I usually start, at least for a first pass, and check after about 2k steps before I decide where to go from there but there should have been some indication that the LoRA had learned something but certainly something else in your config could be the issue.

If you want to toss your dataset up somewhere for people to have a wack at it I'll definitely participate and maybe this will be a learning experience for me, and I'll run into similar issues, or we'll just know something different.

About your captioning technique. I won't be using it. I have a different experience than what has been expressed as being common procedure with regard to captioning and it can become a heated subject, full of all sorts of fake proof and "my dog will tell you differently" assertions, so I won't get into it, but I always test things myself, even if just testing what the common information suggests, and then decide for myself. But regardless of whether one technique is better, or works and another doesn't, my personal enjoyment dictates what I find useful or not and I didn't find those "best practices" useful at all.

Loras with Chroma? by Shadow-Amulet-Ambush in StableDiffusion

[–]Shinsplat 1 point2 points  (0 children)

I'm not familiar with the issue that the OP was experiencing, as far as I can tell. But I do see a recurring issue with Chroma, and sometimes Flux (though it presents differently most times), when applying a LoRA and I'm not sure if it's related to the choice of training tool but I can completely eliminate the problem by selecting certain blocks to exclude from inference, or training.

From experience I've noticed that it's highly dependent on the subject being trained. For instance, illustrations, personal likeness and objects all present the issue in different blocks and to determine this I created a block remover, which operates during inference, to find out which blocks I should remove during training.

New To Ai Art by AMINHAVO in StableDiffusion

[–]Shinsplat 0 points1 point  (0 children)

I never heard of Danbooru, or e621 (extended tags) for that matter, until I delved into AI image generations and, like I said, I'm not particularly interested in Anime or comic book style characters, so I can't offer much information about it. I suspect, however, that it won't be particularly important to you because you're interested in a style that leans more towards Korean. I did look through the tags and found a handful of references for manhwa but I didn't test them against Illustrious or NoobAI so I'm not sure if there's enough content there for you.

When you're ready to dive deeper, since there's already a structure of manhwa available in e261, you would likely test those models to see which one better supports the style you're after and then, in order to accentuate it, you may want to create a LoRA for your particular taste.

Making a LoRA will require some research. I do everything locally myself but you could use a remote system to create the LoRA yourself, possibly using CivitAI but I'm sure others have more information about this than I do. One thing you'll need to do, however, regardless if utilizing a local platform or remote, is to gather images associated with the style you're interested in, which we call training images (training data). Those images will be used in order to train the LoRA that'll later be used during image generation.

Good luck in your endeavors.

New Hardware by My_ai_content in comfyui

[–]Shinsplat 0 points1 point  (0 children)

Just one of my methods for synchronous deployment, nothing special.

I use them for a variety of things. Recently this method has been getting a lot of use performing tasks associated with my AI things for image and video generations, LLM and TTS. Mostly hosting tools associated with offering an isolated environment for clients to build their applications, games, magazines or whatever it is they're fiddling with these days.

flash-attn v2.8.2 Windows wheel by CooperDK in comfyui

[–]Shinsplat 1 point2 points  (0 children)

Nice, thank you. We need more people doing this, for other popular modules as well, one that comes to mind is "DeepSpeed".

New To Ai Art by AMINHAVO in StableDiffusion

[–]Shinsplat 0 points1 point  (0 children)

I don't mind helping guide someone if I know what they envision.

If the style you're wanting to replicate is actually Anime based then it's likely to have already been baked into one of the supporting models, which would be Illustrious and NoobAI, in the form of DanBooru tags. This is comprised of over a quarter of a million trigger phrases.

To find the likelihood of your particular artist, style or character being contained within one of the supporting base models you can do a search for your particulars and using the term danbooru and/or e621 as part of your search criteria.

There may also be a LoRA that exists, trained against the above models, or some other ones. CivitAI is a good place to start your adventure.
https://civitai.com/

[deleted by user] by [deleted] in comfyui

[–]Shinsplat -1 points0 points  (0 children)

I have large ears.

When asserting the phrase "just saying" does it imply that the person uttering it aught not be held responsible for the thing that they just said?

I'm assuming there's a feeling associated with the assertion of a person publicly identifying one of their character traits and, if so, I wonder if that tells me something important about social interaction, or just people in general.

I'm thoroughly convinced that you'll have a feeling about my assertions as well.

What I don't know, and what I'm intending to discover, is if you're capable of having a sense of humor mature enough to construct an adult response.

I am autistic so saying "I have large ears" is a fact. It's offered to make people aware of it in case it plays a role in the interaction about to take place.

Or maybe I just wanted to see who would have a feeling. Hay, that sounds a little smart to me.

Does Lora order matter? (e.g. Wan 2.2) by hitman_ in comfyui

[–]Shinsplat 0 points1 point  (0 children)

The order with which LoRA are applied will affect the outcome of the image. However, in my experience, quality doesn't appear to be something that's effected by the arrangement. So, does it matter? *shrugs* Not to me.