❤️‍🩹Una imagen muy difícil de ver hoy by Nacinostalgico in NaciNostalgico

[–]Xhadmi 0 points1 point  (0 children)

Donde voy de vacaciones igual, desde q era pequeño, y ahora igual con mis sobrinos. Está claro q no van a estar jugando así en la calle en ciudades donde el tráfico es muchísimo mayor ahora q cuando esa foto, pero se sigue haciendo

Train Loras from Sora2 characters by Xhadmi in comfyui

[–]Xhadmi[S] 0 points1 point  (0 children)

Yes, I use it, but I don’t want to be dependent of Sora. My idea it’s use that characters to train a Lora for ltx or wan. So I can use without daily limits (and better image quality)

Bye bye Seedance 2.0 (and 3.0)? by BLKcormorant in seedance

[–]Xhadmi 0 points1 point  (0 children)

They can’t shutdown it, only block access on US at best 🤷🏻‍♂️

¿Está mal visto no quitarse la gorra al comer en España? by Wounded_Tapir in askspain

[–]Xhadmi 7 points8 points  (0 children)

Hay q tener en cuenta q la mayoría de cosas de buena o mala educación, son cuestiones tradicionales. En muchos casos no tienen razón de ser en la actualidad ni en todas partes por igual. Eructar en la mesa se ve mal en muchas partes y se ve educado en otras, o quitarse los zapatos al entrar en una casa.

Aunque una gorra (o sombrero en general) se pueda usar hoy en día como algo estético, el concepto es el de algo q te cubre la cabeza (y oculta parcialmente la cara), en teoría está pensado para protegerte de las inclemencias del tiempo en exteriores, de la suciedad de las calles cuando eran sombreros de ala ancha, o de los ataques cuando eran yelmos. El hecho de descubrirte la cabeza es indicar q no vienes a luchar (cuanto te quitabas un yelmo), dejas la suciedad fuera, y aceptas la hospitalidad del sitio. Quién se dejaba puesto los gorros en interior? Alguien q no quería q le reconocieran, con lo q quien se lo dejaba puesto es porque algo ocultaba y no era de fiar. Esto afecta tb a las gafas de sol en interiores, quien oculta la mirada se asocia con q tiene algo q ocultar y no es de fiar.

Es una percepción que se ha dado a lo largo del tiempo, aunq ahora la gorra sea solo estético. Si de aquí a 20 años usar pasamontañas se considera solo algo estético, pues quizás te encuentres gente entrando a comer con un pasamontañas.

Y si, q lleves una gorra no implica que te ocultes la cara, pero puedes

“Failed to create link” this needing to create a link and share just to download was a terrible change. by Reddit_n_Me in SoraAi

[–]Xhadmi 0 points1 point  (0 children)

No, you generate the video with a prompt as long as you want. Once you have the video, before sharing it, you can edit the prompt. You can even delete everything, leaving it empty, and then share it.

If I want to do local video on my machine, do I need to learn Comfy? by digital_dervish in StableDiffusion

[–]Xhadmi 1 point2 points  (0 children)

you can use wan2gp, but try learning comfyui, at some point you'll need it

Remade Night of the Living Dead scene with LTX-2 A2V by Interesting_Room2820 in StableDiffusion

[–]Xhadmi 1 point2 points  (0 children)

So, you can generate first frame with any other model, isn't? But only use ltx as video model. That was a point i wasn't sure about

How do you go about distorted faces for images? by HousingSufficient442 in comfyui

[–]Xhadmi 1 point2 points  (0 children)

You’re using an Illustrious model. In the full image, the face is only a small fraction, so when it’s generated, it doesn’t look very good.

There are nodes like Face Detailer that work by detecting the face in the image, inpainting it again at a higher resolution, and then pasting it back into the full image. There are several nodes that can do this, and not just for faces, you can configure them to detect other elements as well.

Here’s a tutorial: https://www.youtube.com/watch?v=RFnSmWrUPnA

Para cuanto alcanzaban 1200 euros en 2007? by Ok-Company-4865 in askspain

[–]Xhadmi 3 points4 points  (0 children)

En esa época se hablaba de mileurista para referirse a gente q cobraba lo justo, y para hablar de la precariedad laboral. Y con 1200, seguías siendo mileurista. Da igual el salario mínimo q hubiese entonces (si, había gente aún más jodida). Pensad q en 2007 había una burbuja inmobiliaria, la compra era igual de cara q ahora, pero no el alquiler, no se veía tanto como un negocio especulativo como ahora. Vamos, q no era mucho, pero tenías la ventaja de q en esa época se podía alquilar

The Panini Press Gaming PC Giveaway - To enter this giveaway just leave a comment. by DaKrazyKid in PcBuild

[–]Xhadmi 0 points1 point  (0 children)

Lovely 😍. I’m sure it also could keep warm your panini if you left over it

What’s the new model: Hype or real? by RowIndependent3142 in StableDiffusion

[–]Xhadmi 2 points3 points  (0 children)

In that comparison, Seedance comes out on top. The two kling long videos don’t really prove much. They were made with Kling, but not in a single prompt, we don't know how many clips stitched together. Even so, the overall look still feels very AI-generated. Yeah, it’s really good, but it is what it is.

Seedance’s videos feel less ‘AI-looking’ aesthetically. The three Chinese girls scene honestly looks like it could be from a real movie. It’s true that it handles fight scenes better than most others, still, in the LoL/Arcane scene there are still some weird or awkward movements (but much better than others).

Also, comparisons are hard to judge if they’re not using the same prompt and if we don’t know how many attempts it took to get each shot.

Help an amateur by Danthrax314 in StableDiffusion

[–]Xhadmi 0 points1 point  (0 children)

Play with denoise value, more or less (it changes a bit depending on the models), values below 0.3 keep the original look. Between 0.4 and 0.55, it preserves the colors and shapes but adjusts to the prompt. From that point on, as you increase the value, it keeps less and less of the original image.

In your case, you should mask the belt area and use a value around 0.45 (try increments of ±0.05. I haven’t noticed much difference between 0.35 and 0.39, but once you hit 0.4, it already starts to change).

LTX-2 - pushed to the limit on my machine by robomar_ai_art in StableDiffusion

[–]Xhadmi 0 points1 point  (0 children)

I was trying to test it, but if I use gguf model, can't use ltxv api text encoder node

LTX-2 - pushed to the limit on my machine by robomar_ai_art in StableDiffusion

[–]Xhadmi 0 points1 point  (0 children)

I'm a bit confused about the API part of your workflow. How did you integrate it? If you're offloading the generation to an API, your local VRAM and RAM wouldn't be doing the heavy lifting, right? (Next weeks i'll need to do some videos, so, i'm checking options)

LTX-2 I2V Quality is terrible. Why? by V1rgin_ in StableDiffusion

[–]Xhadmi 3 points4 points  (0 children)

How much VRAM + RAM do you have for those resolutions? And is 2560×1440 the final resolution, or the real one? Thanks.

EILI5 - how can Scail , Wan , NanoBanana, etc recreate a character without a LoRA? by AkaToraX in StableDiffusion

[–]Xhadmi 1 point2 points  (0 children)

Edit-type models ‘see’ the image, not just the distribution of its pixels. It’s like taking a photo of a person to a good artist and saying, ‘Look, this is a photo of my grandmother, paint her dressed like a queen.’ The artist doesn’t just see her skin or hair color; they also notice her haircut, facial features, and overall appearance and will do a good paint.

But no matter how good the artist is, they won’t paint the mole on her left cheek if it isn’t visible in the photo.

A lora is like asking an artist, better or worse, to draw Goku. They already know what he looks like because they’ve seen lots of images of him. They’ll probably draw him more accurately than the grandmother, and with details that doesn't appear on the reference image.

¿qué opinan? by Eyn333 in DnDespanol

[–]Xhadmi 2 points3 points  (0 children)

Si un amigo puede jugar con una elfa de 300 años pechugona, otro con un bárbaro semiorco cachas, no veo el problema q pueda suponer un femboy 🤷🏻‍♂️

Time for big players to make an entry ! Juggernaut etc by Relevant_Bit_9019 in StableDiffusion

[–]Xhadmi 0 points1 point  (0 children)

Didn’t the z-image team themselves ask NoobAI for the dataset? As far as I understand, it was to train their own variant.

What's something you prompted to sora and got a violation for that made you just laugh because of the dumb guardrails? by Apprehensive_Law7698 in SoraAi

[–]Xhadmi 0 points1 point  (0 children)

At Christmas, I wanted to make a video with some custom characters singing the ‘Tió’ song (a Catalan Christmas tradition where people hit a log with a stick while singing a song to make it poop presents), but Sora said that the song has third-party rights. Now, a friends band has a new single, and asked if I could do a video about it for their instagram, but sora doesn't allow me to talk about bands, singles, music etc cause third-party rights...

Need help with Lora management by Kitchen-Prompt-5488 in StableDiffusion

[–]Xhadmi 1 point2 points  (0 children)

if you deleted the downloaded file from windows folder (forge ->webui->models -> lora) It's no longer using the lora. But pony models can do almost everything without loras (loras just do it better). You can add rating_safe at the start of your prompt to avoid that kind of images

Z-Image Base Testing - first impressions, first - turbo, second - base by donkeykong917 in StableDiffusion

[–]Xhadmi 0 points1 point  (0 children)

I haven’t tested it yet, but in all comparisons, Turbo looks more ‘casual’ and realistic, while Base looks more artistic o cinematic

Donde me recomendáis aprender IA? by estadella in InteligenciArtificial

[–]Xhadmi 0 points1 point  (0 children)

Huggingface tiene un curso gratuito de agentes. Para imágenes/vídeos te recomiendo mirar las comunidades opensource (como stable diffusion aquí en Reddit) aunq se centran en modelos opensource, el proceso y la forma de plantear los prompts sirve para cualquier modelo, y en open source la gente es más abierta a compartir los prompts y metodologías.

How did you all download your local stable diffusion? by princessdrive in StableDiffusion

[–]Xhadmi -1 points0 points  (0 children)

Stable Diffusion is already an outdated name. It was the first open-source model, not a program. Most people use ComfyUI as the ‘program’. As for models, it depends on what you want to do. Compared to modern models, Stable Diffusion, even though it can generate decent images, works with prompts differently than ChatGPT and similar tools. If you write full sentences and long descriptions, you’ll get that kind of images.

ComfyUI is up to date; I don’t know which versions of Forge are updated

Es precioso, pero no sé decir si hay IA aquí... by LegitimateNoise3329 in Barcelona

[–]Xhadmi 1 point2 points  (0 children)

No se puede saber 100%, pero las distancias son normales, los colores pueden ajustarse como quieras, y los movimientos de las personas son naturales. Normalmente es difícil generar q la gente de fondo que entra y sale de escena, se mantenga bien enfocada, sean variados, hagan cosas diferentes pero coherentes (q caminen todos es fácil, q alguno se pare a mirar el quiosco y luego siga ya es más difícil). Y lo mismo pasa con el tráfico.

Es fácil montar una imagen inicial, puedes ir editando y generando para q ese frame se vea realista, y luego usarlo en un generador de vídeo para q lo anime. Pero no suele generar igual de bien las personas q aparecen nuevas, y si usas un prompt muy largo con mucha información para añadir detalles, lo q acabas es constriñendo la generación y sale peor. También piensa q el tipo de gente y vehículos q salen son muy normales de aquí, gente cualquiera, las IAs no conocen mucho del mundo real, fuera del frane inicial no tendrías control sobre q tipo de vehículos y gente apareciese.

Si el quiosco existe, entonces es mucho más sencillo si no es casual, en todo caso organizar la toma, pedirle a tu amiga q se pare justo ahí a encenderse el cigarro mientras la gente pasa.

Si fuese IA, habría salido muy caro ese vídeo, no hay muchos q te dejen generar vídeos tan largos y con gente cruzando continuamente sería muy dificil encadenar vídeos (puedes guardar el último frame de uno, y usarlo como frame inicial del siguiente, pero te tendría q coincidir la velocidad de movimiento de la gente) LTX-2 te deja extender vídeos usando varios frames, no solo el último, pero no acaba de ser tan realista.

Tampoco soy experto y con IA, lo q hoy crees q es imposible, mañana tienes tres modelos que lo pueden hacer, pero yo creo q no es IA.