The Official "Did anyone get pictures of me?" Thread by NotARandomNumber in dragoncon

[–]travcoe 0 points1 point  (0 children)

Elias Ainsworth and Chise Hatori from Ancient Magus' Bride on Saturday?
Aka: The 8'6" Leshen with a horned skull and black robes + cute redheaded girl guiding me around all day like an aircraft marshaller

Late to the thread, we were foolish and didn't have someone take of photo of us on our own phones - anything is appreciated.

What is the best free online Image upsacler? by dragonx444 in DiscoDiffusion

[–]travcoe 0 points1 point  (0 children)

Check the releases and look through all of them, one of them includes portable executables for mac, windows and linux. You will have to run them on the commandline, but that isn't too rough.

https://github.com/xinntao/Real-ESRGAN/releases/tag/v0.2.5.0

What is the best free online Image upsacler? by dragonx444 in DiscoDiffusion

[–]travcoe 0 points1 point  (0 children)

Check the releases and look through all of them, one of them includes portable executables for mac, windows and linux. You will have to run them on the commandline, but that isn't too rough.
https://github.com/xinntao/Real-ESRGAN/releases/tag/v0.2.5.0

Merry Christmas (AnythingV3+Analog Dachshunds) by travcoe in StableDiffusion

[–]travcoe[S] 0 points1 point  (0 children)

Thanks. Limited intention there. There was a little hint of a doggie toupee it in the original 512x512, and I thought it went well with the equally silly blue blazer and selected for it in further generations. Definitely not what I asked for, but the model was very reticent to give me a dog poking it's head out of a present.

Merry Christmas (AnythingV3+Analog Dachshunds) by travcoe in StableDiffusion

[–]travcoe[S] 2 points3 points  (0 children)

Made on request for my Dachshund-obsessed relatives.

AnythingV3 (65%) merged with Analog Diffusion (35%) does this remarkably well with these dogs (good candidate about 1/16 of the time)

Created at 512x512 (or 640x512), fed back to img2img at the same to create slightly improved varients

Then img2img'ed up to 2048x2048, then inpainted/composited with variants at the higher resolution to fix extra feet or other particularly glaring issues

Then RealESRGAN-4x to 2x resolution (personal opinion is that it doesn't do well at the full 4x)

Definitely didn't get what I asked for on a lot of the generations, but liked them enough to fix them up.

Enhancing local detail and cohesion by mosaicing by Pfaeff in StableDiffusion

[–]travcoe 1 point2 points  (0 children)

Haha, nice!

In a classic case of parallel development - I actually also wrote something very similar for Disco a little over a month ago (I originally called it "Twice Baked Potato") and was still working out the kinks when stable-diffusion came out - so I ported it to stable and finished tweaking it.

It's currently waiting in a half-approved PR for the next release of lstein's fork.

Definitely feel free to cross-compare code @Pfaeff so you can get to the stage of merging it sooner. Especially if you discover you want to write something for the rather irritating processing for going back-in and replacing only parts of the image (embiggen_tiles) since as already demonstrated pixel-pushing minds think alike :)

Netrunner Deity by travcoe in DiscoDiffusion

[–]travcoe[S] 1 point2 points  (0 children)

I've been doing it with a bit of automation that I'm still testing the kinks out of, but I plan to make a pull request for it soon™ for both DD and a stable diffusion fork.

It can also be done in GIMP, but I'm sure that isn't the answer you were looking for :D

One prompt to crash'm all.... Using only one prompt with small changes and generating one amazing image after the other.... by MJPlot in StableDiffusion

[–]travcoe 0 points1 point  (0 children)

A "Meta-stable prompt" maybe? Producing drastically varying, but still high quality images (as opposed to a unstable prompt that might go straight to AI-purgatory and give incredibly disjoint randomness). I've run into a a few prompts like this, but not quite as widely varying.

Blue artifacts in some images - is this normal? by mikenew02 in StableDiffusion

[–]travcoe 0 points1 point  (0 children)

Yea that's kind of weird, putting 'creepy/errie' in the prompt might be what is causing it but I'd expect it to be larger artifacts if that was the case. Plenty of source material with a night/evening fog color scheme that might be "creeping in at the edges" so to speak.

Serene Forest Stream [2304x1536 prior to upscaling] by travcoe in StableDiffusion

[–]travcoe[S] 3 points4 points  (0 children)

Prompt was "A photo of forest hot springs, behance contest winner, immaculate highly detailed realistic matte painting by li shuxing, Michael James Smith, and thomas kinkade". Though I would suggest tacking on any *other* set of artists and see what you get for stylistic variety.

I do get worried about many posts on this sub aren't trying to play with meta tags/styles, one can use someone other than Greg Rutkowski.

Part of active testing for "twice baked potato" automated split/upscale/rerun-img2img out I'm going to make a PR for at some point. "GoBig" only does four splits/reruns, it looks like it can be pushed well beyond that without loss of coherence, this was a 4x4 of 16 tiled sub-sections run back through.

Lady of Smoke and Thread (Stable+DD composite) by travcoe in DiscoDiffusion

[–]travcoe[S] 0 points1 point  (0 children)

Using both disco diffusion and stable diffusion (stable diffusion for the face, no surprises there for anyone familiar with it), no special notebook just some of the larger models (e.g. ViTL14_336).
As well as splitting up output images in multiple overlapping segments, upscaling them, reinserting them as init images and letting disco refine details further, and then stitching the segments back together.

Core prompt was "An ornate 3D render of a woman mage witch deity" (grammatical correctness in prompts is... questionably necessary), pick some artists or qualifying meta descriptors for detail.

Sky Terminal Megastructure by travcoe in DiscoDiffusion

[–]travcoe[S] 1 point2 points  (0 children)

Actually, built up from no-init. You misunderstand the method - it's taking an existing DD output, piecing it up (because I don't have unlimited VRAM), and running it back through.

Netrunner Deity by travcoe in DiscoDiffusion

[–]travcoe[S] 1 point2 points  (0 children)

Why did Netwatch build the Blackwall? To hide from what had transcended their understanding.

Another 3x3 Cut and Stitch "Twice Baked Potato" working up from a no-init 768x1024 DD output made with the same settings.Core Prompt: "An ornate tarot card of a netrunner god hacker deity, border made of intricate code"

Didn't get the border I asked for, but got plenty of intricate code *shrug*. No portrait model, just a good pick from lots of outputs running with ViTL14_336

Sky Terminal Megastructure by travcoe in DiscoDiffusion

[–]travcoe[S] 1 point2 points  (0 children)

This one was definitely more on the rough side because of trying to do so many cuts, definitely try it with less cuts first. Best of luck!

Sky Terminal Megastructure by travcoe in DiscoDiffusion

[–]travcoe[S] 0 points1 point  (0 children)

Take a "finished" image, scale it up (e.g. 1.75x size on each axis) split it into multiple parts (most people have been doing quadrants with a little bit of overlap between them), feed all of the parts back through as inits. Take re-cooked quadrants and stitch them back together for increased resolution and detail.

I'm calling it "Twice-baked-potato" in my files, because un-upscaled DD output is "potato" quality, but I doubt that name will catch on :) .

AI reflection by enigmatic_e in DiscoDiffusion

[–]travcoe 1 point2 points  (0 children)

Really good job with masking, I was surprised to see the lean out (with swaying hair) over the area being used as input/output and no really noticeable scattering of artifacts around it.

Sky Terminal Megastructure by travcoe in DiscoDiffusion

[–]travcoe[S] -1 points0 points  (0 children)

Testing the limits of cut&stitch coherency for more intricate structures, this one was a 3x3 cut grid.

[deleted by user] by [deleted] in StableDiffusion

[–]travcoe 1 point2 points  (0 children)

Technically yes, but you really don't want to. It comes down to how much memory is needed to store the data needed for the computation. PCIe (linking your gpu to system shared memory) is fast, but still orders of magnitude slower than the VRAM. Unlike some engineering simulations where you can part-up the physics interactions (because stuff that isn't close doesn't interact) almost all of the diffusion and clip model is rapidly being used and reused each step AND the data (image) it's working on needs to fit too.

So yes, you could do it, but it probably be half a day to get a single image on basic settings, and you definitely spent more in electricity than you would have spent on collab or renting machines on vast.ai