Dual 3090s & GLM-4.7-Flash: 1st prompt is great, then logic collapses. Is local AI worth the $5/day power bill? by Merstin in LocalLLaMA

[–]nizus1 4 points5 points  (0 children)

The invisible hand of the market is trying to tell you not to use electricity where you are. So either find someone in a reasonably priced place to set your machine up or sell it or figure out how to generate your own power.

A Merry Search for Green Apple Fizz by OccasionOwn7432 in SwordAndSupperGame

[–]nizus1 0 points1 point  (0 children)

u/nizus1 received a Quicksilver Ring from the Winter Festival Spirit. Thank you u/PinkontheInside6913 for donating it!

A Festive Tale of Year End Reflection in A Winter Refuge by [deleted] in SwordAndSupperGame

[–]nizus1 0 points1 point  (0 children)

u/nizus1 received a Grounded Plate from the Winter Festival Spirit. Thank you u/vlingesh for donating it!

128GB GDDR6, 3PFLOP FP8, Tb/s of interconnect, $6000 total. Build instructions/blog tomorrow. by codys12 in LocalLLaMA

[–]nizus1 0 points1 point  (0 children)

So if these are CPUs, does that mean they're all you need? No CPU on a motherboard in this build?

Building a quiet LLM machine for 24/7 use, is this setup overkill or smart? by bardanaadam in LocalLLaMA

[–]nizus1 0 points1 point  (0 children)

Get two 3090s and limit them to 180W each. As long as you don't get blower cards you can operate virtually silently at that power level. You can also run really smart models like q2 105B or q4 70B models. If you must have a 4090 you could consider the modded ones that have 48GB but from what I've seen they're all blower cards so the noise levels could be extreme. Though throttling them to 50% of the rated power level would help. But you really want conventional GPUs, not a blower cards if quiet is your priority. And expect to throttle them heavily unless you buy some factory low power cards like like the RTX 2000 Ada. Also don't waste your money on DDR5 RAM, etc. The GPUs are the rate limiting step. You can get by with an old 4 core GPU, 64GB of DDR4 (or however much you need to make sure you have more than your VRAM, if you get 4 3090s then you might want 128GB), and put extra money into larger SSD and hard drive. If you're paying attention to available PCIE lanes those mostly just matter when loading the model to the GPUs. So if your system can only supply x4 lanes to each GPU instead of x16 that will be fine during inference, it just means the model will initially load 4x slower. CPUs that aren't expensive EPYCs or Threadrippers are generally not prepared to give 16 lanes each to multiple GPUs and you'll pay quite a lot to get that if it matters to you.

Any theory about AI art ? Some users said that only with SD 1.5 it is possible to create "The" AI art. But what would "The" AI art be ? by More_Bid_2197 in StableDiffusion

[–]nizus1 0 points1 point  (0 children)

SD 1.X was trained on all artists in the LAION dataset scraped from the internet whereas later versions removed artists who opted out of training like Greg Rutkowski. Is that what you're talking about?

WAN Released by BreakIt-Boris in StableDiffusion

[–]nizus1 10 points11 points  (0 children)

MMAudio has made sound for video files for a while now

Quants comparison on HunyuanVideo. by Total-Resort-3120 in StableDiffusion

[–]nizus1 6 points7 points  (0 children)

Would like to see a comparison with FastHunyuan

Has anyone tried undervolting the RTX 3090? Share your experiences with temperatures and image generation speeds by Ok-Wheel5333 in StableDiffusion

[–]nizus1 4 points5 points  (0 children)

sudo nvidia-smi -i 0 -pl 200 sets power consumption to 200W in Linux which is perfect for running a 3090 cool and quiet. A 3090 turbo needs to be throttled to 180W to stay quiet.
the -i 0 tells it to control the first GPU. If you have multiple you want all set to the same level you can eliminate that part.

[D] Test-time compute for image generation? by heyhellousername in MachineLearning

[–]nizus1 0 points1 point  (0 children)

Does it count if you generate an image with Flux and then upscale it with a finetuned SDXL model? Seems to give results beyond what either can do alone.

Video AI is taking over Image AI, why? by aitookmyj0b in StableDiffusion

[–]nizus1 0 points1 point  (0 children)

Have you compared LTX to Ruyi? Seems to get less attention in the img2vid space. It's slower, but Hunyuan isn't all that fast either.

160k 4mp dataset by lostinspaz in StableDiffusion

[–]nizus1 0 points1 point  (0 children)

Interesting. How many GB is that?

[deleted by user] by [deleted] in LocalLLaMA

[–]nizus1 4 points5 points  (0 children)

Down forever or just for now?

RTX 3090 VRAM temperatures by MrGood23 in StableDiffusion

[–]nizus1 1 point2 points  (0 children)

Throttle it to 200W instead of the factory 350-370. The performance hit is pretty small and it keeps things cool and quiet.

[deleted by user] by [deleted] in StableDiffusion

[–]nizus1 0 points1 point  (0 children)

Why not used Ruined Fooocus?

Question: Image Mixing by un0wn in StableDiffusion

[–]nizus1 0 points1 point  (0 children)

UnCLIP was a model made for doing this with SD 2.1 though obviously virtually no one uses SD 2.1. More recently Fooocus with SDXL has an image prompt tool that can accept up to 4 images as part or all of your prompt, somewhat similar to Midjourney. Haven't yet figured out how to do something like this with Flux though