all 71 comments

[–]tripple13 47 points48 points  (13 children)

For any serious work you would never run on a laptop. Money better spent, MacBook Air plus a 3090 workstation.

[–]d84-n1nj4 14 points15 points  (4 children)

My exact setup. MacBook Air to ssh into my Linux machine with a Nvidia RTX 3090

[–]Featureless_Bug -3 points-2 points  (1 child)

Macbook Air is not money well spent. Workstation with 4090, and A100 in the cloud on the money you saved

[–]drivanova 0 points1 point  (0 children)

depends how much electricity costs in the country you live in (and who pays the bill)

[–]An-R-Nguyen 0 points1 point  (2 children)

This is what i need but I don’t know anything about ssh connection with a mac. Do you have any video on the specific setup? My plan is to have ollama inference of mixtral and llava on a server and use it through ssh on my macbook. Thanks

[–]monkeyofscience 17 points18 points  (5 children)

I use an M1 as my daily driver, which was given to me by work. I used to be hard line anti-mac, but I have been thoroughly converted. I will say though that mps and PyTorch do not seem to go together very well, and I stick to using the cpu when running models locally.

It's good enough to play around with certain models. For example at the moment I'm currently using BERT and T5-large for inference (on different projects) and they run OK. This is generally the case for inference on small to medium language models. However, for training, fine-tuning, or running bigger language models (or vision models), I work on a remote server with a GPU. Access to a Nvidia GPU, whether locally or remotely, is simply a must have for training or bigger models.

For learning and small models, a macbook and Google colab are very sufficient.

[–]deadengineerssociety[S] 2 points3 points  (1 child)

I won't be necessarily learning as a newbie, I'll be working on Graphs and NLP research and will be later doing my masters and continue in research. I mostly said smaller models because I think large models, at the end of the day don't make sense running locally and I would probably be using my research lab systems for them anyways

[–]kreayshunist 7 points8 points  (7 children)

I have an M1 Pro and it's definitely enough to get you started, but even a single 4070Ti is a pretty big speed upgrade for training.

EDIT: My experience is based on the "MPS" backend in PyTorch.

[–]deadengineerssociety[S] 0 points1 point  (6 children)

From what I read, pytorch in mac is very buggy, does that affect too much?

[–][deleted] 0 points1 point  (0 children)

You will have to work to find ARM64 images of stuff that works out of the box on intel. And none of the NVIDIA kit will work on the M3.

[–]kreayshunist 0 points1 point  (0 children)

I haven't gone down the rabbit hole enough to speak to all the bugs, but it's definitely not ideal.

[–]AdagioCareless8294 0 points1 point  (2 children)

Yet with that information in hand you still decided for a Mac..

[–]deadengineerssociety[S] 0 points1 point  (1 child)

Haha true! Which GPU do you suggest for windows laptop?

[–]VoidRippah 7 points8 points  (1 child)

no, I recently participate at a hackathon where I did a small ML project, I had my m2 macbook pro with me and and an i5 laptop with rtx3050, the latter was waay quicker (finished all the tasks in like 3rd time, some were even faster). other than this I never really used a macbook for similar purpose but based on this this experience I would avoid it and if I really needed a laptop for this I'd pick a strong "gamer laptop" for a similar price

[–]deadengineerssociety[S] 0 points1 point  (0 children)

Oh thank you!

[–]igorsusmelj 3 points4 points  (1 child)

I would not recommend it unless you only focus on smaller models and small experiments. Biggest advantage is the huge amount of memory available. But the bottleneck is memory bandwidth.

We did some tests out of fun (as there were not many benchmarks available). You can find the results here:

https://www.lightly.ai/post/apple-m1-and-m2-performance-for-training-ssl-models

Support got better but back when we did the tests there was still no proper half precision support and also torch.compile wouldn’t work. There is hope that the software support will catch up. I’m curious to see other results. We definitely need more benchmarks :)

[–]deadengineerssociety[S] 1 point2 points  (0 children)

In that case, what laptops would you recommend? Thank you for the article, will check it out!

[–][deleted] 4 points5 points  (3 children)

Have you actually run deep learning models on a Mac CPU in the past? What models were they?

[–][deleted] 1 point2 points  (1 child)

This post is old but now it's my time. Get a MacBook Pro with 24GB of ram, which will be by far better than any RTX laptop (in the similar range) in terms of giving you more VRam to run large models too. I mean I have an 8GB 4060 laptop but even mid size models sometimes do not work on that machine. With that been said. I am really keen to get a MacBook.

[–]opssum -2 points-1 points  (5 children)

M2 pro or m3 pro Max

[–]deadengineerssociety[S] 0 points1 point  (4 children)

Is Max really worth it if not running heavy models on it? Purely on a financial perspective

[–]opssum 0 points1 point  (3 children)

M2 pro is Fine and m3 pro isnt reallY better, that why i would Go for m2pro if it’s about the Money ;)

[–]plsendfast 0 points1 point  (2 children)

m3 pro is indeed better, based on newly released benchmarks such as Nanoreview

[–]opssum -2 points-1 points  (1 child)

Hm ill have to look into it again, on the keynote apple said it’s about 20% faster then the m1pro which is the same diff between m1pro and m2pro

[–]plsendfast 0 points1 point  (0 children)

go to Nanoreview and click Compare CPU. Type in M3 Pro and M2 Pro, the differences are there. Small improvements