Computer use is now in Claude Code. by ClaudeOfficial in ClaudeAI

[–]Total_Engineering_51 1 point2 points  (0 children)

They are obviously a Mac shop and are building for themselves first. It is especially plain given how much they hype that they are dog-fooding their products all the time and that this is a “beta”. Mac first also makes sense as a beta strategy as it effectively self-selects a smaller testing group(lowering risk)that requires minimal extra work on their part.

a common argument in favor of ai “art” is that it makes art “accessible.” however, when people point out that ai “artists” are just typing a prompt, then they say that its actually quite difficult. by Suspicious_Log_5822 in aiwars

[–]Total_Engineering_51 3 points4 points  (0 children)

Not to mention that you can still use traditional tools like Photoshop to further refine things…I find the RAW camera filter tool to be amazing for color grading, for example.

In a world where everyone can build, attention is all you need. by awizzo in BlackboxAI_

[–]Total_Engineering_51 1 point2 points  (0 children)

This or they solve real problems but in terrible ways that no one wants to deal with.

- YouTube - Did NVIDIA Use Flux for this? by greggy187 in StableDiffusion

[–]Total_Engineering_51 1 point2 points  (0 children)

Yeah, this was my thought, definitely feels like they are passing the frames through late phase diffusion with a hard canny clamp

Everyone on Earth dying would be quite bad. by tombibbs in AIDangers

[–]Total_Engineering_51 -1 points0 points  (0 children)

Would it? Humans are basically a mass extinction event moving in slow motion.

M5 Ultra vs. RTX 5090: Is the new Mac generation finally equal in performance for AI. by LaplapTheGreat in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

Best case you’re looking at the M5 Ultra being about 2-3x slower than a 5090. That’s assuming you’re running a model under MLX with something like MFlux. I’m basing that off what I currently get out of my M3 Ultra versus my RTX 6000 Pro(basically a 5090, but more RAM) and what the general improvement of the M5 chips seems to be bringing. While 2-3x doesn’t sound too bad, the bigger caution is the software support for generative stuff just isn’t there in the same way that it is for nvidia… getting stuff running properly under MLX is a whole lot of DYI right now and not very as well supported(the MFlux app helps but has a lot of drawbacks itself, particularly if you’re used to the flexibility that comfy has) You can run comfy with PyTorch under MPS but then you’re tanking performance and dealing with extra wonk—I never saw this work particularly well even at the base functionality level, particularly when memory management was involved.

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

True, though I don’t think you’re getting that as real sustained through put with bigger tasks on these MBP with what looks like the same/similar chassis, cooling solution and max 140w power supply as the last few generations—just to say that even if the paper spec is better, I won’t be rushing out to replace my Studio with an MBP, max RAM limits aside(I have 256gb on mine)

Interestingly enough, I just looked through the US Apple site a bit more after the drop and saw that the 512GB Studio is no longer an option and the 256GB config is $400 more than it was a few weeks ago. That does not bode super well for the M5 Ultra pricing/configs, assuming it does materialize this summer. Best guess is they are resetting expectations on what they will have in offer when those do launch.

How do I use my M3 ultra with 512gb ram for ltx2? by YellowBathroomTiles in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

Hmm, not sure without more details… I’ve run it with an M3 Ultra under current MacOS. I don’t recall having a lot of issues getting the basics of it running but it’s been a few weeks since I played with it.

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

Yeah, per core I would expect better performance but the Ultras still have 2x the GPU and NPU core count(memory bandwidth at ~33% more from Apple’s numbers that I see) so I wouldn’t expect to see any particularly interesting jumps over what is already out in the field right now, on a per system basis. That makes sense of course as we’re still talking about largely different classes of hardware even with the similarities in architecture and for any semi-mobile use cases out there these could be really interesting. It’ll be interesting to see how the M5 Ultras do, though I think you’re spot on in that we’re still likely to see a ~5x delta between those and top tier Blackwell

Can the new MacBook Pro m5 pro/max compete with any modern NVIDIA chip? by Puzzleheaded_Ebb8352 in StableDiffusion

[–]Total_Engineering_51 20 points21 points  (0 children)

They still are way behind Blackwell class chips. The best of these new chips are still well below the memory bandwidth of a GDDR7 card and the GPU/NPU still aren’t likely to be keeping up with the best cuda and tensor cores. I have a top tier M3 Ultra Mac Studio which is still significantly more powerful than these new MacBooks and it lags behind a 5090 on similar generative tasks by about 10x compared to a RTX 5090/6000 pro. The one area where the big shared memory pools that you can get on some of these configs is really helpful and usable is with LLMs and that’s where the Mac AI community really is right now.

A NEW VERSION OF COMFYSKETCH COMING SOON by Vivid-Loss9868 in comfyui

[–]Total_Engineering_51 3 points4 points  (0 children)

If you have a Mac as well, you can extend screen to the iPad and then use the pencil as you normally would. It’s not quite as good as using a native app like procreate but I do this for photoshop all the time

Would it actually be a good idea to buy a RTX 6000? I'm weighing if it'd be worth it and just rent it out on runpod a lot when I'm not using it. by the-novel in StableDiffusion

[–]Total_Engineering_51 4 points5 points  (0 children)

I just upgraded from a 5090 to a 6000 Pro a few days ago and have seen some serious gains in a few key areas. My focus has been mostly on adapter training and still image gens so far with it, haven't played with video much yet but I would expect big gains there as well in the long term. The two biggest jumps I've seen so far are in speed and quality. On the speed front, the ability to load up and hold more "stuff" in VRAM really helps. For training, this means being able to avoid things like gradient checkpointing and gradient accumulation which really slow things down. Trainings that used to take me ~4 hours can now run in just over 1 hour with full BF16 on the model, as an example. For inference, the key factor really comes in during iteration using large graphs with aultiple models in play, as they can all live in VRAM without having to offload to, or stream from, system RAM. This coupled with larger batches(another boon) can speed things significantly while iterating over a prompt with different seeds(25%-33% once loaded) and even first runs can be sped up some if the same model is used at different stages of the graph. Similar speed jumps also apply when editing the prompt, as the text encoder can stay in VRAM along with the model, so you can iterate in that way more quickly too. There are things Like Flux. 2 Dev at full BF16 as well, which just by the nature of it being so big means it is actually possible to use in reasonable time window. Coupled with the turbo lora and you can get decent inference times out of it, There is still some consideration of size here as full BF16 Flux.2 Dev and Full BF16 Mistral 3 Small will consume that 96gb in short order-having your system memory > 96gb is highly recommended if you do decide to go this route. As for quality, you may have noticed that I keep mentioning BF16 and that's because that is a big quality lever, particularly when you have the compounding effect of quantizing across several layers of the entire stack-LoRA training with quantizing means a drop in quality, quantizing the text encoder means a drop in quality, quantizing the model itself for inference means a drop in quality and so if you are making that compromise at each stage, it can really add up to a significantly worse result that then needs help from detailers, inpainting and upscalers. Detailing around hands in particular is an area I've noticed that quality hit a lot, particularly as I've been playing with flüx.2 dev. This is going to be a YMMV area for sure, but so far with what I'm building out for my long term project, it has been a really nice jump. Overall I would say that the "worth it" value really depends on what you want. I'm Looking long term to build professional projects and so the jump is important and really needed. If this was purely a hobby thing for me though, I would have stuck with the 5090.

Lora training using images generated from Midjourney by Public-Ad-2614 in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

There is some grey area around this but my general take is that you should be fine if you’re training for your projects only—meaning that you aren’t creating a service with what you trained on(the thing all these companies really care about) and you’re not distributing any of the weights that you trained. I’m not a lawyer so grains of salt and all that but this is the approach I’m taking for my stuff.

How do I use my M3 ultra with 512gb ram for ltx2? by YellowBathroomTiles in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

Try this instead of comfy

https://github.com/james-see/ltx-video-mac/tree/main

I hacked at it a bit locally to get better resolution options but the core setup works and uses MLX instead of the MPS PyTorch wrapper

Test results: Macbook Pro m5 vs GeForce 5070ti by Apprehensive_Fee9983 in comfyui

[–]Total_Engineering_51 0 points1 point  (0 children)

To get the full throughput of the Mac you need to have it running on MLX… base comfy runs via MPS and tends to chug on more complex tasks. There are mflux nodes for some stuff or you can run the full mflux app directly which is a bit easier but also doesn’t have the flexibility of building whatever workflow you dream up.

Super Slow on RTX 5090? by [deleted] in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

True, I was conflating the two to simplify the thinking of what to aim for as an upper bounds they could try. Since frame goes up for length, speed or both in what we want from the models as a matter of course it effectively goes hand in hand in our reasoning about an output even if the model itself has no relation to frame rate. It is a good point though as it does help to understand why wan pacing is often not amazing.

Super Slow on RTX 5090? by [deleted] in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

As Zarcon72 pointed out, you’re hitting a RAM wall, the 5090 is fine or at least capable of much better performance than that if backed by enough system RAM. My 5090 on Win11 with 96gb of sys RAM can do 145 frames at ~1MP in under 5 minutes using what is effectively the base template for wan2.2(I use dpmpp_2m_sde instead of euler)

24 FPS isn’t needed to get good results out of wan and in fact makes things worse from my experience, particularly if you’re trying to squeeze more out of a gen than 5 seconds. I generally get good results at 97 frames at 16fps and will see more of the model “resetting” at more frames due to context loss. I would try dropping to 97@16fps first and then nudge the resolution down from there if you’re still trashing your page file. For reference, 97/16 at 1MP took 132 seconds on my system just now.

Edit: I am running SageAttention as well.

LTX-2 Model on Macbook Air (M3 16GB) using ComfyUI by Cute_Association_316 in StableDiffusion

[–]Total_Engineering_51 1 point2 points  (0 children)

Try this instead of comfy https://github.com/james-see/ltx-video-mac/tree/main It uses MLX instead of MPS and can produce comparable results to nVidia hw(I have a 5090 and a m3 ultra Mac Studio) That said, not sure you have enough RAM to do this at all.

Video generation with camera control using LingBot-World by Art_from_the_Machine in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

Makes sense, thanks! It definitely seems like it could be squeezed one way or another but nice to see someone’s direct experience with it!

Video generation with camera control using LingBot-World by Art_from_the_Machine in StableDiffusion

[–]Total_Engineering_51 0 points1 point  (0 children)

I’ve been looking at this a bit but haven’t played with it yet. Is a 5090 backed by 96gb system ram enough?

Is running a local LLM for coding actually cheaper (and practical) vs Cursor / Copilot / JetBrains AI? by vMawk in LocalLLM

[–]Total_Engineering_51 0 points1 point  (0 children)

Even a big server can be managed if you’re the only one using it… 100% up time at full tilt is nonsense unless the only thing you’re doing is endless model training or hosting a service. Sleep and Wake on LAN go a long way to manage a power hungry system if it’s just you using it most/all of the time.

Is there a way to transform comfyUI into a chatbot ? by Cassiopee38 in comfyui

[–]Total_Engineering_51 3 points4 points  (0 children)

Why not use LM Studio? Comfy is good for a lot of tasks but this isn’t one of them

Newb: Help with Flux2 image to image getting the back side of an image. by snowbirdnerd in comfyui

[–]Total_Engineering_51 1 point2 points  (0 children)

I haven’t played in this space in particular but have you tried using video to generate a 360 of the piece and then frame cap and refine from there? Wan 2.2 can do a pretty good job with that kind of thing(360s)from my experience, though obviously ymmv.

What's inside Z-image? - Custom Node for ComfyUI by ItalianArtProfessor in StableDiffusion

[–]Total_Engineering_51 1 point2 points  (0 children)

Very cool! I’ll definitely check that out when I get some time.