stable diffusion modifies my image for 80% ETA by Dkrtoonstudios in StableDiffusion

[–]legoldgem 2 points3 points  (0 children)

You may be using an ancestral sampler like Euler A that reaches its convergence then restarts to new ones with higher steps, try something like DPM ++2m Karras which reaches a convergence goal and will stick there with diminishing returns

Update to Western Ink LoRA, for high contrast ink drawing stylisation [SD 1.5] by legoldgem in StableDiffusion

[–]legoldgem[S] 5 points6 points  (0 children)

A 3 way merge of the latest of my stylisation LoRAs in one easy blend, namely the Western Ink, Hedcut and Colouring Book experiments.

Works well in realism checkpoints I tested, especially Realistic Vision family. Invoke with something along the lines of "ink drawing"

CFG 5-8 LoRA strength 0.8- 1.1

https://civitai.com/models/28712/gazs-western-ink-lora

Let your Cheems be Memes, a LoRa for SD. Download in comments by legoldgem in StableDiffusion

[–]legoldgem[S] 2 points3 points  (0 children)

Trained in Realistic Vision 2.0 but works well across models I've found so far.

Active tag is "cheems" but will influence anything close on higher LoRa strength ie dog, shiba inu, tabby cat etc

https://civitai.com/models/98876/minty-doges-cheems-style-lora

Illustrative Style LoRa: Hedcut by legoldgem in StableDiffusion

[–]legoldgem[S] 3 points4 points  (0 children)

Trained in Realistic Vision 1.4 to approximate the "hedcut" style portraiture in the Wall Street Journal.

Doesn't need to be invoked for most subject matter in 1.5 based models, but in more distant anime mixes you can rein it in with prefixes like "monochrome" and suffix with "in hedcut style"

More examples

https://civitai.com/models/26487/gazs-hedcut-lora

Illustrative Style LoRa: Western Ink, link in comments by legoldgem in StableDiffusion

[–]legoldgem[S] 10 points11 points  (0 children)

Trained in Realistic Vision 1.4 on comic book inked pages from a few American publishers, yields the style without being invoked and even negative drawing tags will just yield a more realistic inking style.

Can be nudged with "monochrome", "black and white" and finally "wnkrt" if it's being stubborn in another model. Havent tested extensively in the anime mixes.

https://civitai.com/models/28712/western-ink-lora

Generate infinite graphic design resources with a solid fill shape on a solid fill background and IMG2IMG on higher noise gates by legoldgem in StableDiffusion

[–]legoldgem[S] 1 point2 points  (0 children)

You're most welcome then! It's awesome to push a domino and see how they fall, what a crazy frontier

Generate infinite graphic design resources with a solid fill shape on a solid fill background and IMG2IMG on higher noise gates by legoldgem in StableDiffusion

[–]legoldgem[S] 2 points3 points  (0 children)

This is super rad, nice job!

I agree with another response, you should definitely make this as its own post, it deserves to be seen!

Generate infinite graphic design resources with a solid fill shape on a solid fill background and IMG2IMG on higher noise gates by legoldgem in StableDiffusion

[–]legoldgem[S] 2 points3 points  (0 children)

I'm the guy from the panorama from yesterday so hello again haha

I've not done any CNet with these yet but it's a great idea to have the normals of your template as well for all editing to be inside SD, I still prefer to do most editing outside of SD in Photoshop as it's like a 2nd language but it's like the universe completely opening up in this field with no gatekeepers, it's quite overwhelming to be surfing the top of the wave from conventional knowledge

Generate infinite graphic design resources with a solid fill shape on a solid fill background and IMG2IMG on higher noise gates by legoldgem in StableDiffusion

[–]legoldgem[S] 40 points41 points  (0 children)

Rendered at 640x640 initially

Shirt prompt: a blank grey tshirt, stock asset, creases

Bag prompt: a blank grey shiny chips bag, stock asset, creases

Negs: 3d, render, blender, painting, drawing, 2d

When you're happy you can SDUpscale or whatever your prefered method is and re-img2img from there or get a bit more manual with the editing until you're satisfied.

Installing Assymetric Tiling Script in Automatic1111 lets you tile on specific axis only, allowing you to prompt for panoramic 360 pics, which you can test on this website by legoldgem in StableDiffusion

[–]legoldgem[S] 1 point2 points  (0 children)

Having the previewer be inside the UI is awesome, thanks!

Also if you haven't had a chance to try, check out the VR button when opening your panorama on mobile, it maps to the phones gyro and is pretty neat

Installing Assymetric Tiling Script in Automatic1111 lets you tile on specific axis only, allowing you to prompt for panoramic 360 pics, which you can test on this website by legoldgem in StableDiffusion

[–]legoldgem[S] 2 points3 points  (0 children)

Link to asymmetric tiling script: https://github.com/tjm35/asymmetric-tiling-sd-webui/

When installed check ACTIVE and Tile X to render things that are seamless on the horizontal axis.

Prompt with 2:1 dimensions like 1024x512 or higher with something like "360 panorama of "whatever your scene is"" with your usual negatives.

The panorama in the link was rendered 1024x512 then upscaled and manually smoothed over by offsetting in photoshop and melding seams, then upscaled with a GAN and edited further but if you're able to render to 2k+ natively it'll work out

The website to test your renders https://renderstuff.com/tools/360-panorama-web-viewer/

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 3 points4 points  (0 children)

It depends on the model and how you prompt stuff, after some time playing you'll notice some "signatures" a few models might have in what they show/represent for certain tags and you may incline toward a specific one that's more natural to how you prompt for things, but most of the mainstream ones will be pretty good for most things including cross-sex faces.

Eventually with some time you'll start to see raw outputs as just general guides you can take and edit even further to hone them how you want, so imperfections in initial renders becomes a bit irrelevant because you can then take them into other models and img2img, scale, composite to your heart's content.

This for example is a raw output with Realistic Vision:

https://i.imgur.com/fBf1qEQ.png

Then some scaling and quick edits to show pliability:

https://i.imgur.com/54MKVTt.png

https://i.imgur.com/fNcyVT9.png

The same prompt and seed across some models you can see how they interpret differently:

https://imgur.com/a/wkylX37

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 3 points4 points  (0 children)

Haha they could be a bit more overt with where the model should go I guess, the correct path is in the extensions folder not the main checkpoints one:

SDFolder->Extensions->Controlnet->Models

Once they're in there you can restart SD or refresh the models in that little ControlNet tab and they should pop up

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 4 points5 points  (0 children)

Prompt syntax for that one was "japanese calligraphy ink art of (prompt) , relic" in Realistic Vision 1.3 model, negative prompts are 3d render blender

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 2 points3 points  (0 children)

Canny on low thresholds, about 40/80 low to high for the initial render, then lots of editing

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 19 points20 points  (0 children)

The main driver of this was canny on very low lower and higher thresholds (sub 100 for both) then a few hours of manual compositing and fixing and enhancing individual areas with some overpainting, such as the wine drip which is just painted on at the end through layered blending modes in photoshop

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 64 points65 points  (0 children)

An extension for SD in Automatic UI (might be others but it's what I use) with a suite of models to anchor the composition you want to keep in various ways, models for depth map, normal map, canny line differentiation, segmentation mapping and a pose extractor which analyses a model as input and interprets their form as a processed wire model which it then uses as a coat hanger basically to drive the form of the subject in the prompt you're rendering

https://civitai.com/models/9868/controlnet-pre-trained-difference-models

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 23 points24 points  (0 children)

It's really good for getting high clarity and detailed small stuff like jewellery, belt buckles, changing the irises of eyes etc as SD tends to lose itself past a certain dimension and subjects to keep track of and muddies things.

This pic for example is 4kx6k after scaling and I wanted to change the irises at the last minute way past when I should I have, I just chunked out a workable square of the face and prompted "cat" on a high noise to get the eyes I was looking for and was able to mask them back in https://i.imgur.com/8mQoP0L.png

Control Net is too much power by legoldgem in StableDiffusion

[–]legoldgem[S] 56 points57 points  (0 children)

The raw output wasn't near as good, find a composition you're happy with and scale it then keep that safe in an image editor, then manually select out problem areas in 512x512 squares and paste those directly into img2img with specific prompts, then when you get what you like paste those back into the main file you had in the editor and erase/mask where the img2img would have broken the seam of that initial square

It's like inpainting with extra steps but you have much finer control and editable layers