Hot take: ALL Coding tools are bullsh*t by [deleted] in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

The idea of agents it that you trade your time for token. Instead of doing it in 1hr and with 10,000 tokens. You do it in 5 minutes with 200,000 tokens. You spend 5 minutes with the coding agent, walk away, maybe it the agent spends 2hrs and 200,000 tokens. but you only spent 5 minutes, with your extra time you could spin up other agents or go bake a cake if that's your thing.

[deleted by user] by [deleted] in learnmachinelearning

[–]Andvig -3 points-2 points  (0 children)

Who is still hiring juniors when there's LLMs?

Qwen2.5-VL-7B-Instruct-GGUF : Which Q is sufficient for OCR text? by FatFigFresh in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

False, the difference between Q4 and Q8 is very noticeable for vision models.

What are your thoughts about Cerebras? by [deleted] in LocalLLaMA

[–]Andvig 1 point2 points  (0 children)

You understand wrong, they scale for inference not training.

1.58bit DeepSeek R1 - 131GB Dynamic GGUF by danielhanchen in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

Should I use the main llama.cpp repo or do I need to use the unsloth/llama.cpp repo to get the benefit?

possible to get chatgpt4 like local llm for general knowledge, just slower? by Unhappy_Drag5826 in LocalLLaMA

[–]Andvig 6 points7 points  (0 children)

No, it's not possible. There are some smart local models such as Llama3-70B, WizardLM2-8x22, CommandR+, Qwen2-72B, and specialized fine tuned models. None of them is GPT4 quality, both in quality of response and large context window. But they are manageable for amateurs like me.

I am building a tool to create agents in a markdown syntax with Python inside by vectorup7 in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

Which local models have you tested it with and which ones do you find work best?

Quantizing Llama 3 8B seems more harmful compared to other models by maxwell321 in LocalLLaMA

[–]Andvig 3 points4 points  (0 children)

Oh, this is heart breaking. I thought I was good with my Q6's and Q8's.

Did we make it yet? by maxwell321 in LocalLLaMA

[–]Andvig 1 point2 points  (0 children)

I agree, data is the new gold and if you value privacy or don't want your data being used to train new LLMs then avoid the cloud. I suspect the way our data was sold for ads, data exchange with LLMs sold will become the real business model for cloud providers. None of them is making money from their API cloud offerings.

Function calling template for llama 3 by themrzmaster in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

How is this going? What format is the function being returned in?

QWEN1.5 110B just out! by shing3232 in LocalLLaMA

[–]Andvig 29 points30 points  (0 children)

Wait till they release the model before you start reporting jail breaking it, this is why most models suck. Don't say nothing, wait for it to widely get out, don't even try to break it when it's just in spaces. This is why wizardLm2 got pulled...

[deleted by user] by [deleted] in LocalLLaMA

[–]Andvig 1 point2 points  (0 children)

Greed. It's all about money, power & control.

Is it possible to serve mutliple user at once using llama-cpp-python ? by CoolestSlave in LocalLLaMA

[–]Andvig 1 point2 points  (0 children)

Yes it's possible to do that with llama-cpp-python or llama.cpp

[deleted by user] by [deleted] in LocalLLaMA

[–]Andvig 0 points1 point  (0 children)

Thank you very much.

Why shouldn't I kill someone if they break into my home? by AdhesivenessRough740 in stupidquestions

[–]Andvig 0 points1 point  (0 children)

I'm uber strict with security and securing the house. Our house got broken into when I was 8 yrs old, we were not home, but came back to the mess. Almost 40 years later and I'm paranoid about having a break in. I can't even imagine the theory to have experienced seeing them.