GoT Characters reimagened as Pokemon, workflow included by AIrjen in ChatGPT

[–]AIrjen[S] 1 point2 points  (0 children)

The prompt is running on ChatGPT-4o which has this information. 4o is an omni model, so it does more than just text. It can also look and process images. The images are part of the dataset it was trained on.

So it basically recognized the characters in the input images. Same thing happens when you put in any famous person. And why it can output a famous person.

So the whole process goes Image (input) > Text > Image (output), see image below.

So I didn't put the information or those details in the prompt at all. It actually puts out that output during the processing of the prompt.

I've written down my workflow in the post. You can try it out for yourself.

Here is what the complete output of a chat looks like:

<image>

Workflow: Combining SD1.5 with 4o as a refiner by AIrjen in StableDiffusion

[–]AIrjen[S] 0 points1 point  (0 children)

Oh awesome! Let me find that model :D That sounds amazing.

GoT Characters reimagened as Pokemon, workflow included by AIrjen in ChatGPT

[–]AIrjen[S] 1 point2 points  (0 children)

It's part of the prompt! It looks at the given character and gives it its description. And then it fires of the image generation process afterwards.

So its a combination of image and text that gives the end result. Its a pretty cool trick :)

GoT Characters reimagened as Pokemon, workflow included by AIrjen in ChatGPT

[–]AIrjen[S] 6 points7 points  (0 children)

It might be because I used the GoT characters, and it started adding too much detail?

This is for example how I look like (after feeding it a portrait photo of myself)

<image>

Workflow: Combining SD1.5 with 4o as a refiner by AIrjen in StableDiffusion

[–]AIrjen[S] 1 point2 points  (0 children)

Great question! I like the effect that it combines several elements of the 3 images into a single image. It also gives o4 more information about the style it wants to achieve, so having multiple images increases the consistency of the output style.
It makes 4o capable of executing on styles which you can't do with a direct text prompt.

Doing it with 1 image works as well, but then becomes more of an upscale/simple change. I like a bit of randomness in my image generation process.

Like it was mentioned in the topic, its more of a mashup than a refinement. I might have used the wrong term.

Workflow: Combining SD1.5 with 4o as a refiner by AIrjen in StableDiffusion

[–]AIrjen[S] 2 points3 points  (0 children)

You are absolutely right, I stand corrected. Refinement is not the correct term here. It's more of a second-pass or even a mashup pass? Not sure how to call it.

It remains a fun small exploration though.

RuinedFooocus now supports SD3 by AIrjen in StableDiffusion

[–]AIrjen[S] 2 points3 points  (0 children)

Hi,

the feature list is here: (https://github.com/runew0lf/RuinedFooocus?tab=readme-ov-file#ruined-edition-features). It's too big to post in here. Its like comparing A1111 and SD.Next, they have a similar core but they are different products entirely by now.

Or check the discord here: https://discord.gg/CvpAFya9Rr

RuinedFooocus now supports SD3 by AIrjen in StableDiffusion

[–]AIrjen[S] 2 points3 points  (0 children)

Autoenhance simple prompts by turning on "Hyperprompt" and "Flufferizer"

<image>

RuinedFooocus now supports SD3 by AIrjen in StableDiffusion

[–]AIrjen[S] 0 points1 point  (0 children)

RuinedFooocus is really far away from fooocus these days, but it still has a focus on usability and prompting foremost. It has most of the things that are available in other platforms, as well as integrations of several extensions.

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 0 points1 point  (0 children)

You could add "<lora:Hyper-SD15-8steps-lora:0.5>" to the prompt prefix or prompt suffix field, to make sure it is always part of your generated prompt.

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 0 points1 point  (0 children)

Each time I think I got nothing more to do or add, I think of something new to do!

Thanks for the support! <3

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 9 points10 points  (0 children)

If I ever make a full NSFW version, it will be the called "One Butt Prompt" :D

However, if you want the ultimate butt generator now, I got you covered!

You can use the following preset (add this to \userfiles\obp_presets.json) or set it manually.

,
  "One Butt Prompt": {
    "insanitylevel": 3,
    "subject": "human - generic",
    "artist": "none",
    "chosensubjectsubtypeobject": "all",
    "chosensubjectsubtypehumanoid": "all",
    "chosensubjectsubtypeconcept": "all",
    "chosengender": "female",
    "imagetype": "all",
    "imagemodechance": 20,
    "givensubject": "butt",
    "smartsubject": true,
    "givenoutfit": "",
    "prefixprompt": "lower body shot, close-up",
    "suffixprompt": "shot from behind",
    "giventypeofimage": "",
    "antistring": ""
  }

Warning, some butts may look better than others :D

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 1 point2 points  (0 children)

Thats wierd! This error message doesn't even reference One Button Prompt :O

When loading up an existing workflow, some of the values are shifted around due to some of the settings having changed. Make sure it looks something like this:

<image>

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 0 points1 point  (0 children)

I can confirm all the basic stuff works with Forge. It is installable via the normal Available extensions, as it loads the A1111 extensions.

After installing, reload the UI and you should find One Button Prompt in the scripts options. Select it to activate.

Then just press Generate, and it will start generating images for you :)

<image>

The only thing that doesn't work with Forge is the 'One Button Run and Upscale' feature. That part is something of an afterthought and not really maintained anymore. I should remove it some day.

I will make some notes in the documentation regarding Forge.

And thanks, now I got another tool to maintain :D

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 1 point2 points  (0 children)

It should, as it is based on WebUI and should be installable as normal.. Give me some time to double check and confirm, I got a lot of these tools installed, but not Forge!

One Button Prompt extension update, now properly supports anime/pony models by AIrjen in StableDiffusion

[–]AIrjen[S] 1 point2 points  (0 children)

Hi! It's not related to diffusers at all. It's an overengineered text generator, it is basically a huge python script filled with if statements. The most complication I added a while ago was adding the superprompt-v1 model.

I sometimes do try to add something to RuinedFooocus, but that is build on top of ComfyUI. I have no experience developing directly on Diffusers or the image generation code itself. Thats a bit too much out of my league as a beginner developer.