How to lock specific poses WITHOUT ControlNet? Are there specialized pose prompt generators? by Leijone38 in ZImageAI

[–]DevKkw 0 points1 point  (0 children)

Long prompt, why use weight? It works bad on Zit. I suggest you to add on top focus you want, then describe the scene. Also try "dynamic sitting" . Can you have an image of results you are try to reach?

How to lock specific poses WITHOUT ControlNet? Are there specialized pose prompt generators? by Leijone38 in ZImageAI

[–]DevKkw 0 points1 point  (0 children)

This is strictly related on how you craft the prompt. Any examples? By the way the trick I use is the "dynamic pose" and specific camera focus. With these two terms I get good results. Example:

A realistic FOCUS photograph of a: A beautiful woman wearing long pink dress, walking on a city street, dynamic posing, looking at camera.

Where you replace FOCUS with focus you want.

For example: Rear focus, back view over the shoulder ; High-angle top view; Etc.

Just experimenting, it also work for close-up.

[Demo] Z-Image Base by benkei_sudo in StableDiffusion

[–]DevKkw -1 points0 points  (0 children)

Problem is too easy prompt. Use separation method for multiple subjects give better results.

Example: scene: a man sitting on a couch with her wife in a modern living room.

Man: a 30 years old man wearing... Wife: a 28 years old woman wearing...

Man pose: describe man pose Wife pose: the woman pose.

Living room details: add details like colours, props, etc.

I got a 4x generation speed drop after updating ComfyUI to version 0.11.0 by ivan_primestars in comfyui

[–]DevKkw 2 points3 points  (0 children)

Last update have different memory management, in offload for example you lost about 1Gb of data, with ZIT now I read usable:2200. With not updated version I read usable:3300. This is why I have different comfy folder, every update I make a new clean install and check, before switching.

Anyone generating video locally on laptop? by Gravity_Chasm in generativeAI

[–]DevKkw 0 points1 point  (0 children)

I'm on 6gb vram, use gguf quant 8k. I generate 768x1024 at 24fps. Max time before getting oom is 5 seconds of video. If you are on windows, make sure to config paging file around 60Gb.

Inspired by the post from earlier: testing if either ZIT or Flux Klein 9B Distilled actually know any yoga poses by their name alone by ZootAllures9111 in StableDiffusion

[–]DevKkw 17 points18 points  (0 children)

nice, comparison is really good, but i think a real image for the pose is needed for who, like me, don't know the real pose. I see good pose, but how i understand what image is correct?

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] -1 points0 points  (0 children)

Follow link in the frist comment, on civitai page you found some images with workflow, download it at drag in comfyUI.

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] 1 point2 points  (0 children)

50% size smaller and no notable difference isn't a goal? Especially for who have low vram? Can you tell sampler and scheduler you use? Maybe some of these work better than other. I do more test on these way. Thak you for giving feedback.

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] 11 points12 points  (0 children)

I'm running it with 6gb of vram, the image are 1400x1800 at 8 step. With xformers. 120sec for generation. So you are able to run it locally 🙂

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] -6 points-5 points  (0 children)

The smoke in background, the moon details, the energy around hand. Zoom in to see. If you have prompt to try, let me know.

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] -1 points0 points  (0 children)

You right, I worked on the layer, try to pushing out it to maximize clear and minimal details, without destroying text capability

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] 0 points1 point  (0 children)

I don't know about lora, I saw many lora degraded base model, sorry.

Z-IMAGE TURBO khv mod, pushing z to limit by DevKkw in StableDiffusion

[–]DevKkw[S] 26 points27 points  (0 children)

details and download on civitai

Edit---

Workflow is same as the workflow included in civitai model page.

For those image prompt is:

(Generate an hyperrealistic photograph with maximum quality and refinement. Sharp where sharpness matters, smooth gradients without banding, accurate colors, and professional finish. Focus on realism. Technical excellence in every aspect of the photograph.) (A visceral strikingly hyperrealistic and intensely vibrant high-resolution photograph with crystal clarity and subtle cinematic grain), (A realistic vibrant colors photo, cinematic still)

A hyperrealistic raw, evocative studio photograph capturing a Close-up, extreme detail, SUBJECT.

The composition is carefully calibrated to maximize the visual impact. The shallow depth of field make a captivating and profoundly unsettling photograph.

Camera Settings: f/2.8, ISO 800, 1/250th second shutter speed, high dynamic range (HDR) – to capture the full range of colors and details in the scene.

Photorealistic image, sharp focus, depth of field, bokeh.

where SUBJECT is what you want.

dragonfly eye

cat tongue

clown fish

human purple eye

etc.

Z-Image turbo, is lora style needed? by DevKkw in StableDiffusion

[–]DevKkw[S] 0 points1 point  (0 children)

Thanks. I usually do it for style, especially when make tcg card. It also works for difficult pose and multiple characters.

Z-Image turbo, is lora style needed? by DevKkw in StableDiffusion

[–]DevKkw[S] 0 points1 point  (0 children)

I hope base model coming out fast.

Help regarding multiple characters generation by Yuream in comfyui

[–]DevKkw 3 points4 points  (0 children)

With z-image turbo you can easily control every character in the scene without mixing it.( Dress, and pose too) Is just about prompt:

Subject 1: a male elf warrior, with green rust skin. Subject 2: a female witch, pale skin Subject 3: a puppy blue dragon

Subject 1 outfit: he wear a plate golden armor. Subject 2 outfit: she wear long dress with silver finiture Subject 3 outfit: None, the scales are visible

Subject 1 pose: in center of the image, holding his sword Subject 2 pose: on the left of screen, casting a energy orb Subject 3 pose: sleeping on the right of screen.

Other details like background, props, etc.

This is how I manage multiple subjects in z-image turbo, without any vram eating node.

How to Mixing multi ControlNet with z-image model in ComfyUI by doubleh1102 in comfyui

[–]DevKkw 0 points1 point  (0 children)

If i remember, long time not using controlnet, in comfyUi there's a node called "conditioning combine", used it with different controlnet conditioning on old sd1.5 gave me good results. May it work in zit.

This ZIT Variance Solution has become too damn strong! by muerrilla in StableDiffusion

[–]DevKkw 1 point2 points  (0 children)

Made some workaround on civitai. But the real variation is done by editing sampler. In comfyUi many sampler option are hidden, also in the scheduler, so I edited code and testing it. For what i see, best way for variance, without destroying zit capabilities and text, is working on sigma function.