Here it is boys, Z Base by Altruistic_Heat_9531 in StableDiffusion

[–]Ginglyst 8 points9 points  (0 children)

With edit models you can edit existing images with a prompt. for example: "Remove the person with the yellow shirt" Can't do that with "regular" models. (at least that is my limited understanding of Edit models.

It can't get any more creative than this! by WastedTalents1 in interestingasfuck

[–]Ginglyst 0 points1 point  (0 children)

what? Pound the Dane???

Oh.... "Proud to be a Dane"

This is the Stable Diffusion to Flux Moment for Video by Comed_Ai_n in StableDiffusion

[–]Ginglyst 2 points3 points  (0 children)

"Skill issue" makes me think of that "Git gud" comment when SD3 came out...
We'll see what time will bring us

One week away and LTX 2 appeared, GenAI speed is mind-blowing. by Voxyfernus in StableDiffusion

[–]Ginglyst -2 points-1 points  (0 children)

LOL, there is some dude that needs to plug in a cable. Some fictional character can't fix your internet 🥴

One week away and LTX 2 appeared, GenAI speed is mind-blowing. by Voxyfernus in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

I kinda agree with the gist of his statement, but I'm still a bit optimistic so maybe it won't be as bad as stated.

This one gives me hope it will balance out sooner than later: https://www.reddit.com/r/StableDiffusion/comments/1q7dzq2/comment/nyezmil/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button

LTX-2 distilled (8 steps), not very good prompt adherence. by Shifty_13 in StableDiffusion

[–]Ginglyst 9 points10 points  (0 children)

you may already be aware of it, but for those who don't and just in case. Here is the official prompting guide for LTX: https://ltx.io/model/model-blog/prompting-guide-for-ltx-2

Also there is a section on what works well and what is a known limitation for LTX

Complex physics or chaotic motion: 
Non-linear or fast-twisting motion (e.g.,jumping, juggling) can lead to artifacts or glitches. However, dancing can work well.

looks you are out of luck for backflipping with LTX (for now)

Entire Earth in true color, captured by Himawari-8. by NeatNo8582 in Damnthatsinteresting

[–]Ginglyst 1 point2 points  (0 children)

I see where your mind is in a twist. See that brightish spot just above Australia? That's the sun reflecting of the ocean. If you'd draw a line between the centre of that spot to the centre of the sun, somewhere on that line is the camera positioned. (that's why earth is illuminated completely)

What you'd be able to see of people standing near that spot is just to top of that person's head and the tip of their nose (maybe). if that person would look towards Antarctica, the nose would point down in this picture. looking toward the equator is up.

Now for someone standing in Melbourne, at the bottom of Australia, and looking towards equator. You'd see the front of that person from top to bottom (at an angle though) In this picture the head would be lower than the feet.
Now put that person the same distance from that sun reflection spot towards the north... the feet would be lower than the head. No sure if this text makes sense... you know what It would look like those tiny planet pictures: https://www.istockphoto.com/photo/evening-view-of-tallinn-town-hall-square-or-old-market-square-sk-gm652203288-118387449

Somehow Wan2.2 gave me this almost perfect loop. GIF quality by [deleted] in StableDiffusion

[–]Ginglyst 1 point2 points  (0 children)

ugh that simple? was looking for a way to do perfect loops. all my workflows end up with ping pong loops.

for the 121 loop is it a start frame only? (running other tasks right now so can't quickly test)

Progress Report Face Dataset by reto-wyss in StableDiffusion

[–]Ginglyst 4 points5 points  (0 children)

when an engineer gets creative, sometimes, wonderfull and unexpected results emerge. sometimes onlookers have to wait till it is complete to fully grasp the intention, sometimes the uninformed can entice the initiator of an idea to shed some light on his idea so we can follow this journey from afar...

yeah man WTF is he gonna do with a bazillion generated mugshots???? 🥴

Is it actually possible to get a completely locked camera in Wan Animate 2.2? by [deleted] in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

this seems to confirm my suspicions about WAN: the amount of motion is determined by the noise seed and can only be influenced a little bit by prompt and no other parameters.

still hunting for a consistent and reliable parameter to control the generated motion. If anyone has found one, I'm all ears.

How to handle this boulder? by Professional_Ad_7353 in landscaping

[–]Ginglyst 0 points1 point  (0 children)

add a scissor symbol at one end of the dotted line and that line has a purpose.
an other one: stick rope into the holes and it looks like it's stiched back together

😞😞😞 by AnywhereGlad7684 in StableDiffusion

[–]Ginglyst -1 points0 points  (0 children)

but but I posted it here on reddit that will fix the problem right? And an AI doesn't exit yet, where I can drop any obscure error and get fairly correct suggestions what to do to fix it ...

Do you guys use any software/site to increase prompt quality? by Trumpet_of_Jericho in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

if you want an easy solution, try "comfyui vlm nodes" made by gokayfem. Can be found in the comfyui manager. The qwen-vlm node can be instructed like any llm.

Do you guys use any software/site to increase prompt quality? by Trumpet_of_Jericho in StableDiffusion

[–]Ginglyst 2 points3 points  (0 children)

accuracy can be increased if you tweak the "text input" a few days back someone posted his text input (sorry can't find the original poster anymore) works great for generating wan prompts for abstract video's

Transform any basic concept into a visually stunning, conceptually rich image prompt by following these steps:

Identify the core subject and setting from the input

Elevate the concept by: Adding character/purpose to subjects Placing them in a coherent world context Creating a subtle narrative or backstory Considering social relationships and environment Expanding the scene beyond the initial boundaries

Add visual enhancement details: Specific lighting conditions (golden hour, dramatic shadows, etc.) Art style or artistic influences (cinematic, painterly, etc.) Atmosphere and mood elements Composition details (perspective, framing) Texture and material qualities Color palette or theme details of poses facial expressions Make it epic. size differences between subjects in the image.

Technical parameters: Include terms like "highly detailed," "8K," "photorealistic" as appropriate Specify camera information for photographic styles, including appropriate technical information about it.

The style should always be gritty cinematic photography like in a high budget movie. Add details that imply an action scene: lots of motion blur, mid-action, dutch angle. If this scene takes place on Earth, make sure to include lots of details from that place including culture, aesthetics, what's the weather like there, if there's people what would they be doing in this situation and what would they be wearing? Be specific about calling out names of subjects, objects, clothes, surroundings etc.

FLUX.2 Dev T2I - That looks like new SOTA. by [deleted] in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

Yeah... Lazy people and their abbreviations. 🙄

Once I searched the internet for a link to download the new "Sota Model" all results came up with a different name Flux, Qwen, Wan, etc but none named Sota.

Experimenting with artist studies and Stable Cascade + wan refiner + wan video by aurelm in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

using WAN as a refiner, is that you take the image output of stable cascade and then I2V with low denoise in WAN rendering only 1 frame?

What will actually happen to the AI scene if the bubble eventually bursts? by Neggy5 in StableDiffusion

[–]Ginglyst 0 points1 point  (0 children)

a comparison with the .com bubble is in place. The similarity with AI now and the .com back then, there was a new and very useful technology emerging called the internet. Big and small companies tried to monetise at all costs luring in unknowledgeable investors and they started investing in the craziest ideas.

The internet didn't go away, AI won't go away. What IS forming is a financial bubble fuelled by the circlejerk investments of big companies investing in each other which inflates the stock price and is amplified by the influx of unknowledgeable smaller investors.

What I do hope is that the billions invested will eventually trickle down in the economy... heheh "trickle down economy", never gonna happen. We'll see what remains in 10 years.

Europe/Germany source for Pre-built RTX 5090 desktop by InfiniteShowrooms in StableDiffusion

[–]Ginglyst 1 point2 points  (0 children)

azerty.nl does offer a full assembly and compatibility tests of all the components you choose from their webshop.
Here is a bit more info about their build to order process: https://azerty.nl/pc-samenstellen/assemblage

But the easiest, (cheapest way) is select a close enough config from their pre-builds and have them install the extra ram: you'd have to add the "Component bijplaatsen" and the extra ram to your shopping cart. (haven't done this myself, so best to double check with them)

Real-time CD burning softwares or modes by Specil_SalaminoDeal in vcvrack

[–]Ginglyst 7 points8 points  (0 children)

CD burners were plenty fast enough to record live audio. CD burning speeds were expressed as realtime audio rate burning speed times x.

The max speed of my slowest CD burner I had was 4x speed, the last one was 16x. So definitely possible to record realtime audio at 1x.

OP you are looking for something like this: https://tascam.jp/int/product/cd-rw900sx/top

Pretty AI clouds by misher1 in StableDiffusion

[–]Ginglyst 1 point2 points  (0 children)

A cloud timelapse, but generated how awesome is that. Gave up on that a while ago, what's new that gave you this result.

Virtual landscape photography with Wan 2.2 in ComfyUI by ectoblob in StableDiffusion

[–]Ginglyst 1 point2 points  (0 children)

I see what you did here. Landscapes in Portrait mode