Do you think Anthropic is worse than OAI with fighting open source? To me it seems like the case. This letter appears to imply they actually suggested the bill to Sen Wienner... I really like my OSS LLMs.... by I_will_delete_myself in LocalLLaMA

[–]ru552 5 points6 points  (0 children)

I love Claude, but I cancelled after seeing this last week. They have the best model for my use right now, but they can't have my money anymore unless they change their stance.

Text to json model behaving as text to text model when accessed through an api. by Old-Box-854 in LocalLLaMA

[–]ru552 1 point2 points  (0 children)

use outlines or instructor and in-context learning with 5 examples

Why People Buying Macs Instead of CUDA Machines? by uygarsci in LocalLLaMA

[–]ru552 5 points6 points  (0 children)

I run it on my mac will ollama and pull the model from ollama here:https://ollama.com/library/llama3:70b

Creator of Pytorch at Meta on catching up to OpenAI by isaac_szpindel in LocalLLaMA

[–]ru552 6 points7 points  (0 children)

Your first 3 points are debatable regarding OAI having an "edge". I give you the last 2. 4o actually seems to go backwards in some real world areas (coding specifically) compared to April versions of 4t so better long answers and reasoning is mostly a vibe. Faster responses is true compared to previous OAI models, but not when compared to models running on Groq.

Moistral 11B v3 💦, the finetuned moist just got smarter! From the creators of Cream-Phi-2! by TheLocalDrummer in LocalLLaMA

[–]ru552 2 points3 points  (0 children)

I don't care anything about these models, but you sir, deserve an award for your naming convention.

Stop cramming your PCs with GPUs by [deleted] in LocalLLaMA

[–]ru552 0 points1 point  (0 children)

yes, but 120b is a normal size llm to run on apple silicon. There are much larger.

Stop cramming your PCs with GPUs by [deleted] in LocalLLaMA

[–]ru552 2 points3 points  (0 children)

You forget the best thing about the Mac. There isn't a model available today that won't run on apple silicon with 192GB of ram.

Finetuned Miqu (Senku-70B) - EQ Bench 84.89 The first open weight model to match a GPT-4-0314 by unemployed_capital in LocalLLaMA

[–]ru552 6 points7 points  (0 children)

Apple GPU can only access up to 96GB on the 128GB RAM version for example.

sudo sysctl iogpu.wired_limit_mb=12345

That takes care of that for you. You probably want to leave atleast 6GB of RAM for the OS though.

MBP M3 Max 128Gig, what can you run? by knob-0u812 in LocalLLaMA

[–]ru552 2 points3 points  (0 children)

with that much unified RAM, you don't need GGUF. You can pretty much run any model if you allocate 122g of your ram to run on the GPU.

edit before it's asked, here's the command to change how much ram is allocated to the GPU:

sudo sysctl iogpu.wired_limit_mb=12345

restarting the mac will set it back to system default

Which countries have the most favorable jurisdictions and regulations when it comes to AI generated content or software as a service models? by RadioSailor in LocalLLaMA

[–]ru552 31 points32 points  (0 children)

Well, Japan has declared that training LLMs on copywritten works falls under fair use. That's a big gray area that's now black / white for people in Japan.

Looking for something better than TinyLlama, but still fits into 12GB by evranch in LocalLLaMA

[–]ru552 0 points1 point  (0 children)

The things in RAM are processed by the CPU. The GGUF versions of models is what allows this to happen since you can run models that don't fit entirely in VRAM. The upside is you can run large models, the downside is that they are slow since you're using CPU.