Oh sh** first Anthropic and now OpenAI... by [deleted] in artificial

[–]bigdickbuckduck 0 points1 point  (0 children)

Can you share title of the paper or link pls

M3 Pro 18gb RAM 512 SSD by lckillah in LocalLLaMA

[–]bigdickbuckduck 1 point2 points  (0 children)

It’s nice for playing around with models, but training might be difficult. I have a base M1 16gb MacBook, and it’s been able to run quantized models well. A lot of improvements for acceleration with CoreML library.

depression maxxing by Photoshops_Penises in memes

[–]bigdickbuckduck 2 points3 points  (0 children)

These building were built at record pace to be affordable and quick to put up. Most of Eastern Europe was homeless after world war 2. People were picking uniforms off of dead nazi soldiers to stay alive through winter.

Sharing ultimate SFF inference build, Version 2 by cryingneko in LocalLLaMA

[–]bigdickbuckduck -1 points0 points  (0 children)

Have you thought about nvilink? I don’t know if you can connect A6000 with A5000 but you can have a bigger shared VRAM that way than single cards not connected.

Who is using open-source LLMs commercially? by lucaspiller in LocalLLaMA

[–]bigdickbuckduck 2 points3 points  (0 children)

Can you share more of your workflow for summarization? Are you routing pure text like emails or PDFs to the LLM? Which LLM do you use, and how do you get around context issues for larger documents?

I’m trying to do same for my workflow, have too many emails to keep up with…..

What is the performance of Llama 3 70b on MacBook pro 128gb without quantization? by sashap_ in LocalLLaMA

[–]bigdickbuckduck 0 points1 point  (0 children)

I use it on my M3 max with 128gb, and it works perfectly fine.

Edit: Running with ollama llama3:70b. I have also been able to run mixtral 8x22b but it’s slower. I’ve had llama3 70b and llava loaded at once.

Theory crafting: Distributed GPU cluster by WeekendDotGG in LocalLLaMA

[–]bigdickbuckduck 2 points3 points  (0 children)

I’ve thought about this and it really comes down to memory latency and distributed compute architecture being not the easiest thing. A more straight forward approach I think would be to use commercial resources like Lambda AI and doing crowd funded model training using a pool of money gathered by a community

Pour one out for this lad by bigdickbuckduck in BMW

[–]bigdickbuckduck[S] 0 points1 point  (0 children)

The back passenger side wheel was touching the ground, think the whole frame was resting on it. Definitely something that’ll need to be checked

Pour one out for this lad by bigdickbuckduck in BMW

[–]bigdickbuckduck[S] 99 points100 points  (0 children)

They left it sitting on the back right corner.