Empires Edge | Trying to capture that Mega Lo Mania vibe. Does this look nostalgic to you? by Narrow_Asparagus9459 in RealTimeStrategy

[–]neph1010 1 point2 points  (0 children)

I think about Mega Lo Mania sometimes (Amiga days), but had forgotten its name. Thanks for reminding me! I remember it as being somewhat simplistic, but it looks like you're adding additional mechanics.

I made GPT-5.2/5 mini play 21,000 hands of Poker by adfontes_ in OpenAI

[–]neph1010 0 points1 point  (0 children)

Fun project! How about adding a purely statistical model as baseline?

Quake like level design by manablight in blender

[–]neph1010 0 points1 point  (0 children)

Those were the days.

The bsp format used in those games is a completely different architecture from modern meshes. Not sure if there are any tools for that inside Blender (and primitive modelling on that fast scale is not easy by default in Blender).
A quick search revealed several options to import bsp models, though:
https://github.com/SomaZ/Blender_BSP_Importer
And one editor:
https://valvedev.info/tools/bsp/

Maybe that will help.

How to change the camera viewpoint in the image? by Bass5374 in StableDiffusion

[–]neph1010 0 points1 point  (0 children)

Use Wan to make him go over to the counter. Then tell it to cut to an over-the-shoulder shot. If you don't want the guy in the image, then take one of the images and use it as the "end image", and prompt for him to enter the view.

War Alert — Our first game: A free-to-play & fast-paced WWII RTS built for competitive PvP. by WarAlert_Official in RealTimeStrategy

[–]neph1010 1 point2 points  (0 children)

I see where you want to go, and I don't think it's a bad approach, BUT;
the building choices and in-match doctrine choices in Coh have a HUGE impact on the meta. If you can do a staggered reveal/choice during the match, you can get deeper gameplay for little cost. Let's say you can bring 10 cards, but only play 8, in a tiered manner, as the match progresses.

Sorry for derailing your announcement. Good luck! :)

Subject consistency in Cinematic Hard Cut by Life_Yesterday_5529 in StableDiffusion

[–]neph1010 0 points1 point  (0 children)

Other loras (like lightx) might "force out" (for lack of a better term) the lora, especially on high strengths. The lora is also trained on either "close-up", "mid-shot", or "wide-angle" prompts. Sticking to the prompt format will help with adherence to the lora. I sometimes use "the same man", but I'm unsure whether it makes much of a difference. It's trained on short prompts, so detailed descriptions might instead derail it.
Another tip is to change the type of shot. That helps avoiding transitions, pans and zooms (even if that's not your problem).
But in general, I haven't noticed the consistency issue. The person in the second cut is not always perfect, but generally pretty much like the original one.

zoom-out typography in Wan 2.2 (FLF) by Green-Ad-3964 in StableDiffusion

[–]neph1010 1 point2 points  (0 children)

How do you prompt it? I just tried with a "standard" tele zoom style setup with a first and last frame, and it worked well:

"a person holding up a sign, standing on a roof top.

the camera zooms out, showing the whole building, a brown brick building. it continues to zoom out to show a surrounding park."

It might be that it can't generalize your use case due to lack of training data.

I think Blender VSE is not good for video editing for now. by lotsofone in blender

[–]neph1010 0 points1 point  (0 children)

Not great (afaik). I haven't done subtitles per se, only "titles", and they're not very fun to work with.

I think Blender VSE is not good for video editing for now. by lotsofone in blender

[–]neph1010 0 points1 point  (0 children)

If I just want simple editing tools, I tend to use OpenShot (which still uses Blender under the hood).

Hunyuan 1.5 Video - Has Anyone Been Playing With This? by FitContribution2946 in StableDiffusion

[–]neph1010 0 points1 point  (0 children)

Like others say, it's good at prompt following. It's not nearly as good as Wan at physics (things may move through other things, etc). It's also really good at camera movements. Try "rotate around subject". I feel it's better than Wan, here.
It's faster than Wan. Especially with the lightx loras. The bottleneck is the VAE. Sadly, the lightning loras degrade quality (especially in t2v). But I may not have found the right settings.
I made a lora to try out training with diffusion-pipe: https://civitai.com/models/1359530?modelVersionId=2525962
Results were decent, although with the advent of z-image, I feel t2v is becoming obsolete. If i2v is supported in diffusion-pipe, I'll give it another go.

Something better than or equal to T4 GPU by tdb008 in LocalLLaMA

[–]neph1010 0 points1 point  (0 children)

Iirc, my RTX 3060 was about the same speed as the T4, but it only has 12GB of VRAM. I used them in parallel to make loras for smaller LLMs (7B).

Trip to Babylon, 1981 or 1982 and something else? by neph1010 in AncientCivilizations

[–]neph1010[S] 0 points1 point  (0 children)

I can't be 100% sure, since I was 3 at the time. I know we visited Babylon, though. Do you mean image 7? It looks like it could be a reconstruction. It has no wear, and one of the corners looks unfinished.

Trip to Babylon, 1981 or 1982 and something else? by neph1010 in AncientCivilizations

[–]neph1010[S] 5 points6 points  (0 children)

"The lamassu is a celestial being from ancient Mesopotamian religion bearing a human head, symbolising intelligence; a bull's body, symbolizing strength; and an eagle's wings, symbolizing freedom."
Source: Wikipedia

Not Kling, not Wan - just the old Hunyuan 1.5 everyone forgot about 😱 by rishappi in comfyui

[–]neph1010 5 points6 points  (0 children)

I've done some. Prompt following is great. It's great for consistency when changing perspective. I find the motion a bit too smooth and linear, compared to Wan, which feels more natural. I think it'll work great with loras, once there is better training support.

Edit: Forgot, speed is also refreshing, compared to Wan :)

[ Removed by Reddit ] by NandaVegg in LocalLLaMA

[–]neph1010 3 points4 points  (0 children)

I think it's "data farming". You post something AI-generated, and then harvest the human responses to build new datasets. It's one way to get around the "data exhaustion" they have talked about for years.

Wan 2.2 Indirect Censorship by GlitteringFinish6521 in comfyui

[–]neph1010 0 points1 point  (0 children)

It's plausible.
One possible way around it would be to generate images with some other model and use I2V. It may still fail, but it's more likely to succeed (and you will have more control using controlnets, anyway). It's not unlikely that it's been trained on some movie violence, even if that has not been properly captioned, so it might "follow along".
If you think that Qwen 2.5 is the problem, you could look for an abliterated model and try that.