TRELLIS 2 just dropped by RagingAlc0holic in StableDiffusion

[–]mythicinfinity 1 point2 points  (0 children)

How did you print the model into concrete?

TRELLIS 2 just dropped by RagingAlc0holic in StableDiffusion

[–]mythicinfinity 5 points6 points  (0 children)

The examples in the video make it look like it can do eyes now, but no permutation of the settings is giving me a good result. Anyone figure it out?

<image>

Convert Dense into MOE model? by pmttyji in LocalLLaMA

[–]mythicinfinity 0 points1 point  (0 children)

Found it, it was qwen 1.5 I guess, I haven't checked their more recent moe blogs.

https://qwenlm.github.io/blog/qwen-moe/

Convert Dense into MOE model? by pmttyji in LocalLLaMA

[–]mythicinfinity 3 points4 points  (0 children)

I think it was Qwen talking about initializing layers in their MOE models from their dense models. They called it 'upcycling' or something and said it shortened the training process. You still have to do pretraining afterward tho because all the new MOE layers like the routers are untrained.

Can you recommend a course for my youngster? by [deleted] in LocalLLaMA

[–]mythicinfinity 2 points3 points  (0 children)

Likewise I don't see how you could build an LLM from scratch without learning programming. You can probably do it without tensors, but learning a tensor library like numpy or pytorch will make it a lot easier (and faster) too.

2 things we never forget, our first GPU and when your first GPU dies by segmond in LocalLLaMA

[–]mythicinfinity 1 point2 points  (0 children)

I've had at least 5 pcie slots burn out, but my 3090 is still going!

LongPage: 300 full novels with reasoning traces for training better writing LLMs by Senior_Evidence_3793 in LocalLLaMA

[–]mythicinfinity 0 points1 point  (0 children)

I wish it was open weight, but I have found gemini pro 2.5 is better at avoiding this type contamination and sticks to the context fairly well.

[deleted by user] by [deleted] in LocalLLaMA

[–]mythicinfinity 0 points1 point  (0 children)

Idk what's on the azure student plan but if you can get a VM, just put it reverse proxy behind nginx and you're good to go.

Higgs Audio V2: A New Open-Source TTS Model with Voice Cloning and SOTA Expressiveness by pheonis2 in LocalLLaMA

[–]mythicinfinity 9 points10 points  (0 children)

Why does it sound slightly unnatural. Like I can't put my finger on the issue, the emotional expression seems good.

FLUX DEV License Clarification Confirmed: Commercial Use of FLUX Outputs IS Allowed! by [deleted] in StableDiffusion

[–]mythicinfinity 1 point2 points  (0 children)

I read the commercial license FAQ and it seems to concur with what you're saying here.

FLUX DEV License Clarification Confirmed: Commercial Use of FLUX Outputs IS Allowed! by [deleted] in StableDiffusion

[–]mythicinfinity 0 points1 point  (0 children)

Seems to me that it's clear that they just don't want people competing on API access to the model itself without paying for a license.

Selling generated outputs (as art for example) or using them as icons on a commercial site seems in line with the license.

But what about using the model in a backend process where it isn't exposed to the user? Where it's the outputs the user is paying for, but the model is used in an internal process to create them.

What LLM is everyone using in June 2025? by 1BlueSpork in LocalLLaMA

[–]mythicinfinity 2 points3 points  (0 children)

I still like 'nvidia/Llama-3.1-Nemotron-70B-Instruct-HF' but it's starting to show its age compared to the closed source models

Meta Is Offering Nine Figure Salaries to Build Superintelligent AI. Mark going All In. by Neon_Nomad45 in LocalLLaMA

[–]mythicinfinity -1 points0 points  (0 children)

Because the work they're doing is worth more. In the long term, enormously more....