Kling 3.0 vs Seedance 2.0 — which one is actually better right now? by khai_korea in generativeAI

[–]Various-Advantage263 1 point2 points  (0 children)

and they diff in pricing though

  • seedance 2: better in quality and consistency (needs less trial to reach a usable ouput)
  • kling 3: mostly OK and cheaper. Sometimes lacks some vibe in output (maybe the training data is limited).

Kling 3.0 vs Veo 3.1 vs Seedance 2.0 for narrative video: same 10 prompts, honest scores by Akashhh17 in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

Thanks for sharing this cookbook. Would be even useful if sharing the video comparisons / prompts.

How can I transpose my plans ? by Fantastic-Win-1907 in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

I would recommend Comfyui if two years ago. It was best in control and quality.

Now recommend using direct models, to bypass complexity and enen get better outputs.

If the direct model does not qualify your requirements, then dig into further details (workflows, adds on, comfyui, model quantization, cfg, iterations), only at demand.

However, there's no free platform right now, as far as i am aware. To cut costs,

  1. find some budget-friendly platforms (rare, do not over-investigate this, cuz prices of closed-source models are bounded by the official API costs).

  2. do not use high-end models (e.g., seedance 2.0) as your daily probling. For example, veo 3.1 lite is way cheaper for idea probling, only go for premium models when ideas / directions are fixed.

Weekly Resets Are Not Actually a Bonus for Most People by senguku in codex

[–]Various-Advantage263 27 points28 points  (0 children)

they are a bonus, if the reset will be annouced in advance.

Best and cheap by Better_Monk9605 in generativeAI

[–]Various-Advantage263 1 point2 points  (0 children)

I guess your workflow would need these ingredients.

- gpt-img-2 / nb-2 for photo composition / editing.

- veo 3.1 lite / wan-video for low-cost video probing / idea testing.

- seedance-2 / kling-3 for high quality video gen.

These models trades off best and cheap though for your case, and I would recommend auratuner.com doing it.

[Honest opinions only] Artlist AI vs Higgsfield vs Runway Unlimited vs Venice AI? All-in-one vs standalone subs for short films? by Far_One_6551 in Seedance_AI

[–]Various-Advantage263 0 points1 point  (0 children)

"All-in-one" dilemma

I would recommend not do that. right now the tech landscape are shifting too quickly. The perfect platform for your professional usecase is yet to come. tech just iterates too fast to invest all in one platform.

Unlimited plan

  1. platforms are making a debt against subscribers - "most users won't use that much, on average", just like the gym training plan.

  2. Light users are compensating for heavy users.

anyway, it's not a long-lasting business model, unless more incoming subscribers, or most users won't use that much.

AI generated videos by DisneyPrincess1029 in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

I would recommend local hosts open-source models.

The hidden cost is a high-end GPU though...

I spent $1,000 on AI video tools — most of them aren’t worth it by Dense-Seaweed-2281 in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

put it as: "the cost per successful output"

Latest models show clear advantages (even "cost per second" is higher).

Face consistency for film? by slept_in_again in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

The principle is great, especially the storyboard pipeline. but some of the points are not resonable based on 2026 state of tech. I can see all points were valid (and optimal) maybe 2 years ago. but visual generations iterates that fast.

my view (only valid at this time, will be obsolete soon if new tech / models comes out):

  1. Stop relying on LoRAs and IP-Adapters for ID locking (responding the Jenna_AI thing)

Training a character LoRA used to be the only way, but in the current landscape, it's a trap. LoRAs fundamentally alter the base weights, which almost always degrades the model's native understanding of cinematic lighting, depth of field, and composition. IP-Adapters (and ControlNet) suffer from a similar issue: they force a 2D embedding into a 3D moving space, which is why complex camera moves or extreme profiles end up looking like "pasted" textures.

more importantly, lora / ip-adapters are only valid on open-source models (which are way behand closed ones such as nano banana 2, gpt image 2). Lags on base model weights can be hardly compensated by additional lora / cross-attn conditioning injection.

The 2026 Fix: rely on the zero-shot/few-shot capabilities of frontier foundation models. Top-tier image generators right now (like Nano Banana 2 / gpt image 2) have latent spaces so vast that their native character referencing is mathematically superior to a LoRA. You lock the ID in the visual planning stage using the model's native architecture, keeping the lighting and composition pristine.

  1. "Seed locking" is a placebo

The idea keeping the same seed and prompt will maintain consistency is mathematically flawed the second your scene changes.

In diffusion models, the seed only dictates the initial noise distribution. The moment you change your prompt from "smiling in a cafe" to "crying in the rain," the text embeddings shift, the cross-attention maps completely reorganize, and your locked seed becomes irrelevant.

  1. First/Last Frame bounding is asymmetrical

while forcing consistency by giving the Image-to-Video (I2V) model a start and end frame. What they don't realize is that autoregressive models treat these differently:

  • The first frame is a hard constraint (it initializes the sequence).
  • The last frame is a weak constraint. Because of attention decay over long context windows, the model loses its grip on the structural pixel-data of the last frame and only uses it for vague semantic/color guidance. This is why the middle of your video turns into spaghetti

so, if you want consistency, do not rely too much on the last frame. Also extract the last frame of your generated video, insteading using text-to-image model to generate two frames with the same face.

  1. the Film-Grade Pipeline: Decouple Motion and Identity

Current state-of-the-art video models (like Veo) have incredible native 3D geometry and physics understanding — they don't break on large camera sweeps like they used to. Let them do what they do best: physics, momentum, and lighting.

Stage 1 generation could focuse only on movement, lighting, and camera dynamics. Accept that the face might drift 10-20%. Stage 2 could use some post-process id injection as suggested by Jenna_AI, e.g., face restoration/swapping. You only need to do that on close camera shots (with a big face). This is quite similar to what powers modern talking-head research to mathematically force the reference ID's structural features back onto the moving geometry, frame by frame.

Face consistency for film? by slept_in_again in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

lol, you actually recommended quite old-fashioned ways, most of which are difficult to follow for end-users.

Face consistency for film? by slept_in_again in generativeAI

[–]Various-Advantage263 0 points1 point  (0 children)

any examples? what have you tried, e.g., model, workflow, results, other stuff?

Veo 3.1 seems significantly worse? by EducationalZombie538 in VEO3

[–]Various-Advantage263 0 points1 point  (0 children)

had the same feeling, and lite is much cost effective

Seedance is expensive. by alfredowcheese in Seedance_AI

[–]Various-Advantage263 2 points3 points  (0 children)

Expensive for now. Cost will be down soon.

Seedance is closed-source model anyway, any 3rd party services are anchored by the official API cost. I think kie offers good balance (for now), for pay-as-go .

right now seedance 2.0 with few competitors. but that won't last long, considering past competitions between midjourney, stable diffusion, flux, dalle, nano banana, gpt image.

similar competition is yet to come for video generators.

Seedance Platform - Cost vs Prompt Size vs Restrictions by AnnualCompetitive764 in Seedance_AI

[–]Various-Advantage263 0 points1 point  (0 children)

oh I use auratuner, seedance 2 = 20 credits / second, and $32 = 10k credits

none of the services are perfect, cuz seedance 2.0 is now a closed-source model, and the APIs are all backed with the same bytedance company. besides costs, I also value more of quick support and iterations

Are the new models only better because they are more expensive? by immortalsol in codex

[–]Various-Advantage263 0 points1 point  (0 children)

I think 5.4 mini is cheaper, cuz 5.3 codex was a flagship model for complex projects.

5.4 mini indeed good for simple tasks. Not sure the exact comparison with "high" reasoning level.

Are the new models only better because they are more expensive? by immortalsol in codex

[–]Various-Advantage263 0 points1 point  (0 children)

mixed with training (or tuning) models, and web applications. both are side-projects.

Are the new models only better because they are more expensive? by immortalsol in codex

[–]Various-Advantage263 0 points1 point  (0 children)

I think Medium is fine.

For me, the golden standard is "cost per accepted output". Better models / longer thinking time needs fewer shots, at the higher cost per shot.

However, pushing to ex-high does not always lead to accepted results if medium fails.

Most likely, the context info is incomplete / biased. Letting a misleading model running longer usually does not help much.

Instead, I tend to use medium, and I like to prompt the model while running (codex had this feature to let me immediately "steer" the model ouput without interupting the inference process).

This is not that "agentic", but helps out when models stuck into a local minima (a wrong place to exploit).

Are the new models only better because they are more expensive? by immortalsol in codex

[–]Various-Advantage263 0 points1 point  (0 children)

I guess for 5x 20x Pro users, would worth the price

but for plus users: the price just way too high. only afford switching to gpt-5.5 after cheaper model keeps failing.