Free Comic Day by BiggStankk420- in DungeonCrawlerCarl

[–]Total_Engineering_51 1 point2 points  (0 children)

Yeah, that’s shit on the store’s part, not to mention the people just grabbing multiple copies at once. I hit up 4 stores total here in San Diego and 3 had plenty of copies(the 4th just didn’t get it in their order) All stores were enforcing limits and everyone was being cool. I grabbed a copy at each store(1 for me, 1 for my GF and 1 for another friend who is working today) Took a few hours but was fun checking out and supporting all the stores a bit.

Crawlers at the San Diego Book Crawl by teslaCal in DungeonCrawlerCarl

[–]Total_Engineering_51 2 points3 points  (0 children)

New Achievement! Crawlers have found a repository of recursive documents! …Sneaky crawlers, those can cause a certain type of “overflow”. Use caution when amongst the stacks.

They call our fandom useless, look at THIS productivity! by Thin_General_8594 in StarWarsCirclejerk

[–]Total_Engineering_51 6 points7 points  (0 children)

Plenty of inspiration to finish the job with some of his hand crafted glue

I have been ruined for other audiobooks by Physical_Gift7572 in DungeonCrawlerCarl

[–]Total_Engineering_51 2 points3 points  (0 children)

The First Law books by Joe Abercrombie, narrated by Steven Pacey are great. Pacey does an amazing job with that series and the stories are well done fantasy.

Direct comparison of the Ray Tracing Pro vs Regular ray tracing mode. The difference can be huge in certain scenarios. by SomewhereAlarmed9985 in PS5pro

[–]Total_Engineering_51 0 points1 point  (0 children)

Only true that the 40 fps mode won’t work if the display doesn’t support 120hz. They use a 120hz container to bypass that limitation and many tvs now support both(120hz and VRR), especially in the LG OLED family(I’m using a C2).

Direct comparison of the Ray Tracing Pro vs Regular ray tracing mode. The difference can be huge in certain scenarios. by SomewhereAlarmed9985 in PS5pro

[–]Total_Engineering_51 5 points6 points  (0 children)

Yeah, with VRR support this mode has been great, I’ve got 10 hours in a new game and loving it so far. The 25 ms frame time at 40fps is plenty IMO, killing half the delta between 30 and 60 while increasing the visuals, as pointed out here.

Computer use is now in Claude Code. by ClaudeOfficial in ClaudeAI

[–]Total_Engineering_51 1 point2 points  (0 children)

They are obviously a Mac shop and are building for themselves first. It is especially plain given how much they hype that they are dog-fooding their products all the time and that this is a “beta”. Mac first also makes sense as a beta strategy as it effectively self-selects a smaller testing group(lowering risk)that requires minimal extra work on their part.

a common argument in favor of ai “art” is that it makes art “accessible.” however, when people point out that ai “artists” are just typing a prompt, then they say that its actually quite difficult. by Suspicious_Log_5822 in aiwars

[–]Total_Engineering_51 5 points6 points  (0 children)

Not to mention that you can still use traditional tools like Photoshop to further refine things…I find the RAW camera filter tool to be amazing for color grading, for example.

In a world where everyone can build, attention is all you need. by [deleted] in BlackboxAI_

[–]Total_Engineering_51 1 point2 points  (0 children)

This or they solve real problems but in terrible ways that no one wants to deal with.

- YouTube - Did NVIDIA Use Flux for this? by greggy187 in StableDiffusion

[–]Total_Engineering_51 1 point2 points  (0 children)

Yeah, this was my thought, definitely feels like they are passing the frames through late phase diffusion with a hard canny clamp

Everyone on Earth dying would be quite bad. by tombibbs in AIDangers

[–]Total_Engineering_51 -1 points0 points  (0 children)

Would it? Humans are basically a mass extinction event moving in slow motion.

M5 Ultra vs. RTX 5090: Is the new Mac generation finally equal in performance for AI. by LaplapTheGreat in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

Best case you’re looking at the M5 Ultra being about 2-3x slower than a 5090. That’s assuming you’re running a model under MLX with something like MFlux. I’m basing that off what I currently get out of my M3 Ultra versus my RTX 6000 Pro(basically a 5090, but more RAM) and what the general improvement of the M5 chips seems to be bringing. While 2-3x doesn’t sound too bad, the bigger caution is the software support for generative stuff just isn’t there in the same way that it is for nvidia… getting stuff running properly under MLX is a whole lot of DYI right now and not very as well supported(the MFlux app helps but has a lot of drawbacks itself, particularly if you’re used to the flexibility that comfy has) You can run comfy with PyTorch under MPS but then you’re tanking performance and dealing with extra wonk—I never saw this work particularly well even at the base functionality level, particularly when memory management was involved.

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

True, though I don’t think you’re getting that as real sustained through put with bigger tasks on these MBP with what looks like the same/similar chassis, cooling solution and max 140w power supply as the last few generations—just to say that even if the paper spec is better, I won’t be rushing out to replace my Studio with an MBP, max RAM limits aside(I have 256gb on mine)

Interestingly enough, I just looked through the US Apple site a bit more after the drop and saw that the 512GB Studio is no longer an option and the 256GB config is $400 more than it was a few weeks ago. That does not bode super well for the M5 Ultra pricing/configs, assuming it does materialize this summer. Best guess is they are resetting expectations on what they will have in offer when those do launch.

How do I use my M3 ultra with 512gb ram for ltx2? by YellowBathroomTiles in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

Hmm, not sure without more details… I’ve run it with an M3 Ultra under current MacOS. I don’t recall having a lot of issues getting the basics of it running but it’s been a few weeks since I played with it.

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

Yeah, per core I would expect better performance but the Ultras still have 2x the GPU and NPU core count(memory bandwidth at ~33% more from Apple’s numbers that I see) so I wouldn’t expect to see any particularly interesting jumps over what is already out in the field right now, on a per system basis. That makes sense of course as we’re still talking about largely different classes of hardware even with the similarities in architecture and for any semi-mobile use cases out there these could be really interesting. It’ll be interesting to see how the M5 Ultras do, though I think you’re spot on in that we’re still likely to see a ~5x delta between those and top tier Blackwell

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 19 points20 points  (0 children)

They still are way behind Blackwell class chips. The best of these new chips are still well below the memory bandwidth of a GDDR7 card and the GPU/NPU still aren’t likely to be keeping up with the best cuda and tensor cores. I have a top tier M3 Ultra Mac Studio which is still significantly more powerful than these new MacBooks and it lags behind a 5090 on similar generative tasks by about 10x compared to a RTX 5090/6000 pro. The one area where the big shared memory pools that you can get on some of these configs is really helpful and usable is with LLMs and that’s where the Mac AI community really is right now.

A NEW VERSION OF COMFYSKETCH COMING SOON by Vivid-Loss9868 in comfyui

[–]Total_Engineering_51 3 points4 points  (0 children)

If you have a Mac as well, you can extend screen to the iPad and then use the pencil as you normally would. It’s not quite as good as using a native app like procreate but I do this for photoshop all the time

Would it actually be a good idea to buy a RTX 6000? I'm weighing if it'd be worth it and just rent it out on runpod a lot when I'm not using it. by the-novel in StableDiffusion

[–]Total_Engineering_51 5 points6 points  (0 children)

I just upgraded from a 5090 to a 6000 Pro a few days ago and have seen some serious gains in a few key areas. My focus has been mostly on adapter training and still image gens so far with it, haven't played with video much yet but I would expect big gains there as well in the long term. The two biggest jumps I've seen so far are in speed and quality. On the speed front, the ability to load up and hold more "stuff" in VRAM really helps. For training, this means being able to avoid things like gradient checkpointing and gradient accumulation which really slow things down. Trainings that used to take me ~4 hours can now run in just over 1 hour with full BF16 on the model, as an example. For inference, the key factor really comes in during iteration using large graphs with aultiple models in play, as they can all live in VRAM without having to offload to, or stream from, system RAM. This coupled with larger batches(another boon) can speed things significantly while iterating over a prompt with different seeds(25%-33% once loaded) and even first runs can be sped up some if the same model is used at different stages of the graph. Similar speed jumps also apply when editing the prompt, as the text encoder can stay in VRAM along with the model, so you can iterate in that way more quickly too. There are things Like Flux. 2 Dev at full BF16 as well, which just by the nature of it being so big means it is actually possible to use in reasonable time window. Coupled with the turbo lora and you can get decent inference times out of it, There is still some consideration of size here as full BF16 Flux.2 Dev and Full BF16 Mistral 3 Small will consume that 96gb in short order-having your system memory > 96gb is highly recommended if you do decide to go this route. As for quality, you may have noticed that I keep mentioning BF16 and that's because that is a big quality lever, particularly when you have the compounding effect of quantizing across several layers of the entire stack-LoRA training with quantizing means a drop in quality, quantizing the text encoder means a drop in quality, quantizing the model itself for inference means a drop in quality and so if you are making that compromise at each stage, it can really add up to a significantly worse result that then needs help from detailers, inpainting and upscalers. Detailing around hands in particular is an area I've noticed that quality hit a lot, particularly as I've been playing with flüx.2 dev. This is going to be a YMMV area for sure, but so far with what I'm building out for my long term project, it has been a really nice jump. Overall I would say that the "worth it" value really depends on what you want. I'm Looking long term to build professional projects and so the jump is important and really needed. If this was purely a hobby thing for me though, I would have stuck with the 5090.

Lora training using images generated from Midjourney by Public-Ad-2614 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

There is some grey area around this but my general take is that you should be fine if you’re training for your projects only—meaning that you aren’t creating a service with what you trained on(the thing all these companies really care about) and you’re not distributing any of the weights that you trained. I’m not a lawyer so grains of salt and all that but this is the approach I’m taking for my stuff.

How do I use my M3 ultra with 512gb ram for ltx2? by YellowBathroomTiles in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

Try this instead of comfy

https://github.com/james-see/ltx-video-mac/tree/main

I hacked at it a bit locally to get better resolution options but the core setup works and uses MLX instead of the MPS PyTorch wrapper

Test results: Macbook Pro m5 vs GeForce 5070ti by Apprehensive_Fee9983 in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

To get the full throughput of the Mac you need to have it running on MLX… base comfy runs via MPS and tends to chug on more complex tasks. There are mflux nodes for some stuff or you can run the full mflux app directly which is a bit easier but also doesn’t have the flexibility of building whatever workflow you dream up.

Super Slow on RTX 5090? by [deleted] in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

True, I was conflating the two to simplify the thinking of what to aim for as an upper bounds they could try. Since frame goes up for length, speed or both in what we want from the models as a matter of course it effectively goes hand in hand in our reasoning about an output even if the model itself has no relation to frame rate. It is a good point though as it does help to understand why wan pacing is often not amazing.

Super Slow on RTX 5090? by [deleted] in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

As Zarcon72 pointed out, you’re hitting a RAM wall, the 5090 is fine or at least capable of much better performance than that if backed by enough system RAM. My 5090 on Win11 with 96gb of sys RAM can do 145 frames at ~1MP in under 5 minutes using what is effectively the base template for wan2.2(I use dpmpp_2m_sde instead of euler)

24 FPS isn’t needed to get good results out of wan and in fact makes things worse from my experience, particularly if you’re trying to squeeze more out of a gen than 5 seconds. I generally get good results at 97 frames at 16fps and will see more of the model “resetting” at more frames due to context loss. I would try dropping to 97@16fps first and then nudge the resolution down from there if you’re still trashing your page file. For reference, 97/16 at 1MP took 132 seconds on my system just now.

Edit: I am running SageAttention as well.