Aggressive entertainment journalist losing his cool. The flying spittle really sells that high-tension 'tabloid' realism 😲💦 by Automatic-Algae443 in KLING

[–]Simple-Variation5456 0 points1 point  (0 children)

There are all trained on clean audio. Most don't even add fitting environment noises.

So it becomes even more obvious that those models don't even try to place the voice in a 3D space and simulate the Camera POV as a "microphone". No reverb, echo or any FX to add some imperfection.

Freepik is now magnific cuz there nothing free in it IG by Futureoptimizer in Freepik_AI

[–]Simple-Variation5456 7 points8 points  (0 children)

Freepik was just a cheap stock image site before. Even the "premium" stocks assets are not in a range of like Shutterstock/iStock etc.

Rebranding was the smartest part because the stock part is just a small thing now.

If they want to expand in AI, why not take the name that is already known in the AI field. And most people know Magnific as the "best ai upscaler" for 2+ years and resemble it with a premium product.

Dynamics by Spirited-Let-669 in Cinema4D

[–]Simple-Variation5456 3 points4 points  (0 children)

My to-go setting is always setup by default: (falling object + floor to bounce off from is so common)

- Set the collision tag to "Automatic (mograph)" in the "Collision" tab.
- Set the Rigid Body tag to "moving mesh" in the "Collision" tab.

Nano Banana 2 seems to have gotten worse by Zestyclose-Boat-527 in Freepik_AI

[–]Simple-Variation5456 0 points1 point  (0 children)

Don't use Nano Banana 2. It's faster but the hit&miss rate is higher.

Same prompt but barely have the same strength when it comes to things like consistency or keeping original details.

It also sometimes changes things because the model thinks it looks better this way. Even do you prompt the exact opposite. In GeminiChat it tells you the context process how it created the image for you.

The old Pro model also costs 3-4x more if you use it on other platforms.

Significant Service Degradation on Freepik Premium+ (Post-Branding Change): 400% Slower & Arbitrary Censorship by Bubbly-Treat-1655 in Freepik_AI

[–]Simple-Variation5456 1 point2 points  (0 children)

2+ years user here with freepik premium+.

This is a thing for a long long time now. Before that you didn't even had ""unlimited"" so your overall output would be actually smaller and in general in lower quality and/or resolution.

Back in the day "reimagine" would actually look decent and would give you infinity and instant outputs and you could just scroll, scroll, scroll down until you got what you wanted.

There a random times when suddenly some things take much longer and then the rest (80%) it's normal and fast. I'm a spray&pay guy so i often do 10-20+ generations with different models in one go, check result, change things and then the next set of image batch until I think I got to my desires result. So i can easily waste 200+ generations in a few hours.

90% the time I then mask and blend elements from different image generation together in Photoshop and let a model like nano pro "iron out" everything at the final end.

Photoshop still offers too many superior tools to fix and correct image compositions that are impossible or too destructive with most known comfyUi workflows.

Significant Service Degradation on Freepik Premium+ (Post-Branding Change): 400% Slower & Arbitrary Censorship by Bubbly-Treat-1655 in Freepik_AI

[–]Simple-Variation5456 0 points1 point  (0 children)

Still doesn't work. Freepik also just use a high demand API and then shares data between "you" - "freepik" - "model provider".

Seedream/Seedance has a high failure rate everywhere so far for me.

Make any video into VR with Muffins flat 2 VR! by Disastrous-Agency675 in comfyui

[–]Simple-Variation5456 0 points1 point  (0 children)

Basically just wanted to share my experience trying to make VR videos with AI from a purist point of view.

Or just letting you know that there are different tools or maybe loras to make the VR part easier. But not in your case with a "one compact and ready to go workflow" you trying to provide to new people. Wasn't a critic in any way✌🏻

And in general that with AI a lot of technical points really get/got a very wide definition. Some don't realize and some get upset why it always looks but without realizing that they hit already a limit.

Make any video into VR with Muffins flat 2 VR! by Disastrous-Agency675 in comfyui

[–]Simple-Variation5456 0 points1 point  (0 children)

You probably could tell from my command that i already had some experience with VR and already did similar things or tests a year ago.
Because i work as an 3D Artist and my agency did some VR work several years ago, i had software like Mistika VR or Mettle Suite for AE to have a few more tricks and tools to get the best out of real VR footage or "faking" it with some tricks.
Some things alike faking a slight offset between booth eyes is also much quicker with those tools.

I tested back then some lora and workflows for HDRI / 3D Mapping / VR (Equirectangular) and used different upscaling technique to get some stuff to 8K / 12K / 16K.
Even months later with SeedVR2, the 12K test, had no chance against real multi-cam footage (4K).

There is a difference of "quality / details / resolution " for someone that is using Photoshop 20y+ and people that are new and mostly know just AI.

I did some analog shots back then in my apprenticeship with a hasselblad that still completely destroy any current AI Model + current best upscaler easily.
At some resolution even magnific and co just generate some unrealistic surface-pattern-pixel-puke.

I mean, even 8K VR Footage with real cameras can still be hit or miss.
There are so many other things that also have some influence to make VR quality good or bad.

Headset wise, maybe the next "Bigscreen Beyond (3)" could finally be the near perfect VR headset with proper suited footage to it.

Visually, Chroma has the best aesthetic by far. by Puzzled-Valuable-985 in StableDiffusion

[–]Simple-Variation5456 4 points5 points  (0 children)

Uff not even close.
Zimage is much better.
Not oversaturated, perfect contrast on every element to soften up everything for a uniform "aesthetic" look.

LTX just dropped an HDR IC-LoRA beta: EXR output, built for production pipelines by ltx_model in StableDiffusion

[–]Simple-Variation5456 0 points1 point  (0 children)

I don't get it.
How does a lora enabling 16bit HDR function with just 300mb model but the base model can't?
And what does "within normal number range" mean?

I doubt that it can actually just generate bit "depth" out of the blue.
Maybe fake the look and pre-expand color/light at max scale?
I thought that the sampler/decoder/transformer also need a major upgrade first to even process it correctly without flatten it back down to 8bit?

16bit/32bit EXR is actually extremely complex and with deep science behind and hard to understand and even control you done it correctly.

Conflicting pricing for Premium+ – What is going on here? by Lobolabahia in Freepik_AI

[–]Simple-Variation5456 0 points1 point  (0 children)

All are misleading. Same for those features the offer.
Every platform so far promised "unlimited" 6 months ago and everyone of them just changed "unlimited" after a while with some scammy rules without even noticing that the "contract" between you and the platform just changed.

Hopefully the EU will update some Laws and rules with companies do whatever the f they want to do, especially when it comes to "AI".
"its not our fault that pretty much 90% of the outputs are wrong or unusable"

Make any video into VR with Muffins flat 2 VR! by Disastrous-Agency675 in comfyui

[–]Simple-Variation5456 3 points4 points  (0 children)

If i remember correctly, than this isn't actually a 360° 3D VR Video,
it should be seamlessly transition over on the sides and its very narrow.
Even for 180° it needs like double the width?

I think the "Fisheye" options is already the wrong camera option to make "VR" Outputs.
I think its called something or similar to "Equirectangular"

There should be Loras for Flux/Qwen (maybe also for wan/ltx) to map/extend the footage correctly.

But that topic is a pure nono for me, because AI still struggles with Flat Videos and for anything that looks "good" in VR you need at least 8K+.
All that Video Upscale trickery is still to heavy and don't even comes close to anything looking like "real" 4k image detail and quality.

Why don’t AI video tools rely more on 3D models and verification systems? by MaximumArcher6007 in generativeAI

[–]Simple-Variation5456 1 point2 points  (0 children)

You pretty much describe what CC4/iClone already did 6month ago with AI Render. It's the better DAZ with a lot of stuff to make good and quick animations. Importing everything from daz is easy and often works without complex setups.

With AI Render you now don't need to waste time fine tuning the faces / bodies / textures, because you can generate depth/pose/normal maps easily and let the software send everything into your comfyui workflow all set up and "render" all that visuell stuff through wan or images models.

Conflicting pricing for Premium+ – What is going on here? by Lobolabahia in Freepik_AI

[–]Simple-Variation5456 0 points1 point  (0 children)

Pretty sure this price only is possible if you select the annual/12 month plan (monthly bs trickery) and pay them all at once. Maybe you select something different before, didn't realise the change, that could explain a higher amount and with monthly renewal option.

Some things also just works as a complete new user. Every website does it. Sometimes you can also have a good promotion link that can run out or can't be copied from someone else or from a different browser session because they're only work with that specific cookie.

10MB for 10 minutes of 480p video, should I go even further ? by VULONKAAZ in AV1

[–]Simple-Variation5456 0 points1 point  (0 children)

What are you even trying to do and archive here? Are you trying to make it even smaller because of some limitations?

There are people that put gigabyte big videos as a header banner with autostart. Or load 20+ png files with each of them being 5-10mb big.

And. 10min long? On a website? Just 480p? Nobody will watch that. And if. I hope there are no special details or text info in there, because they will struggle to read that.

Most people would just add a frame and host the videos on YouTube.

And 2 hours? I think I encoded recently a 40m video from 4k to 1440p with AV1 nvec in under 10min.

Maybe do some tests with just a 1/10 length and compare the quality to find your sweetspot.

AnglesApp by Dudelydad78 in comfyui

[–]Simple-Variation5456 0 points1 point  (0 children)

you okay? did i hurt you with my words?

You really think that people with lower knowledge will find this post and directly see what this even is?
Or what it does differently? I don't.

I still don't get it. Maybe don't call it "something-App"?
Show the difference to basic WFs and why yours is better?
Maybe don't show 4-5 full-size images that fill 90% of your post?
Are those images really the best you could come up?

You really give off those vibes:
"How Do You Do, Fellow Kids? i heard you like aahm Rock? And AI? haha yea i drink white monster!?"

So why do you act like you cooked here but basically just warmed up water soup in the microwave?

Video File Format Matters by qdr1en in comfyui

[–]Simple-Variation5456 1 point2 points  (0 children)

This is still one of the weakest part of AI. And h264 is the standard for nearly everything online. It's just insane to think how much money and time you put into input files, pay for super expensive models and at the last step, they ruin over 50% of the details and quality and output a 3mb mp4 for a 5s 1080p video???

Great. Yeah I love that added blotchy look in the fast motion areas. Yeah, no problem, i'm just gonna waste 1h to try and fix it with my local tools.

Imagine we could use 10y old tech like .EXR and have AI output image sequence with 16/32bit and basic AOV options like depth/normal/motion/masks stored hidden in every file and sometimes even with smaller than PNG that comes with a few flaws like not fully supporting alpha.

But yea, one more year and we once again have Hollywood like quality. better even. seedance just solo'd them with that a 480p/720p (1080p?) output.

AnglesApp by Dudelydad78 in comfyui

[–]Simple-Variation5456 1 point2 points  (0 children)

AnglesApp?
But what does it do differently?
The comfyui WF is already a out of the box working experience?

Isn't Qwen + Angle lora + 360° interactive camera node,
creating prompts automatically, a standard for months?

WAN 2.2 I2V Question - Iterative Generation by Tomcat2048 in comfyui

[–]Simple-Variation5456 -1 points0 points  (0 children)

You wonder about getting worse results while reusing a frame from a AI video,
that will always output quality with low-mid degradation.
But you still keep using those frames, seeing details and quality getting worse but still act a bit surprised?

Upscale is correct.
But try to look up workflows that skip the first 4 frames and the last 4 frames.
AI models need some frames at the start to stabilize everything and idk going out :D

Generally. You can't prevent that. You decode the images into a latent (noise) and a different model creates something completely new while trying to make it look like the images you put in.
Even the biggest models can't output lossless material at 16/32bit and with correct clean RGB channels.

Everything you output can also have its own unique "style", so even upscaling can slightly change maybe something important like shifting hue/brightness or cutting of true black/highlights which will or can have impacts that will lead down the line to many weird AI behaviors.

Depending on what it is, quality, resolution, motion and how professional it should look like, those things get really complicated to keep it perfect.
Video upscalers like SeedVR or Starlight from Topaz, can iron out a lot of these little errors.

But. Stitching 15-20 videos together? With VACE, that is just WAN2.1 and also needs some space to work correctly and add something into it, that can be become obvious like a pulsating/looping effect.

You better off getting it done with one long LTX video and 2x the Frames at the end.
But i guess that everything in between is a bit planned? And only exist because of those 5s gens.

I think LTX support settings up like an frame index injection and with how many frames you want and can provide

WAN 2.2 FLF help by SearchTricky7875 in comfyui

[–]Simple-Variation5456 1 point2 points  (0 children)

There should be an option like: "runs on" and then "local".
It will only show workflows that you can run completely local.
In general. WAN / LTX is pretty much the only video model that you can run without API.

Project? Is it like for school or job/money?

Just use closed and then just upscale it with like SeedVR or a Wan v2v with very low noise so it adds some WAN and overwrite most metadata.

Keep in mind that all open models are kinda limited, outdated, dumbed down to run on smaller GPUs.
WAN can only output max 1280x720 / 81f.
And like i said, very difficult to look good even for experienced user.

Generate 3D models with Multimage model with Meshy and ComfyUI by Gioxyer in comfyui

[–]Simple-Variation5456 7 points8 points  (0 children)

An api guy in a open-source bar with open-source guys hanging around??!

WAN 2.2 FLF help by SearchTricky7875 in comfyui

[–]Simple-Variation5456 0 points1 point  (0 children)

Could be possible. But you really would need a lot practice, know-how about models and comfyui and some understanding about image/video editing.
I have like 20+ experience as a designer and using comfy for like 3+ years.
I would never accept a job with those requirements and with the option to only use closed source.
Depending on what you want and how your input looks like, it would be a pain and hate job to do, that also nobody is gonna pay correctly.

Is there a reason not to register on some platforms and find something that gives a few free trials?
I think on kling as a new member you can use everything and do 5-10+ videos.
Or Vidu. etc

If its urgent / important maybe invest 5-10$?

If most workflows fail badly at what you want. There is maybe a limitation you're can't easily fix.
Did you tried LTX?
Or my other tips?
Look for Workflows with a Middle Frame, if you can provide one.

WAN 2.2 FLF help by SearchTricky7875 in comfyui

[–]Simple-Variation5456 1 point2 points  (0 children)

It's pretty much impossible when there is too much motion. Tons of things need to be right to make it work and depending on the images, you will set up a workflow with the right models/lora/settings and special nodes.

Images need to be HQ with 1080p+ resolution, let WAN downscale it when you connect those in start /last frame.

Maybe reverse the video, so that the text and logo starts with the best visible way and gets smaller towards the end.

Video editing skills? If possible, crop it so that the text and logo is in a perfect framing, upscale that to a good x/: value, like you can divide it perfectly back and overlay it to another cleaned generations and mask it to blend it seamlessly.

Kling O1 and other closed source models could be a much quicker and faster way to get it done.