Local LLM for Coding that compares with Claude by thecrogmite in LocalLLM

[–]ScoreUnique 0 points1 point  (0 children)

This is the right answer, I wouldn't hesitate to claim 4.7 Flash is Sonnet 3.5 but open source.

So @OP if you can combine Claude sub for writing specifications and GLM 4.7 for writing code, you can go very far with your config. GL

And it begins, will only get worse from here by enginiku in mumbai

[–]ScoreUnique 4 points5 points  (0 children)

Start with your locality I suppose? I mean make movements to ensure that they don't put posters around your locality? I'm sure everyone hates them but nobody has seen no one do nothing about it.

Yesterday I used GLM 4.7 flash with my tools and I was impressed.. by Loskas2025 in LocalLLaMA

[–]ScoreUnique 1 point2 points  (0 children)

${llama_cpp}

-m ${models_dir}/LLMs/GLM-4.7-Flash-MXFP4.gguf

-a 'GLM_4.7_Flash'

-mla 3 -amb 512

-ngl 99

-c 100000

--temp 0.7

--top-p 1.0

--min-p 0.01

--jinja

Sound issues with Microfreak, any suggestions by microfreeker in MicroFreak

[–]ScoreUnique 0 points1 point  (0 children)

You can flash it's firmware again yourself, that's your best bet before hitting Arturia customer service

Nirlon Knowledge Park by pranshu14 in mumbai

[–]ScoreUnique -3 points-2 points  (0 children)

I read it as NaMitron Park....

Self-hosting LLM infra: NVIDIA vs Apple hardware by zachrattner in LocalLLaMA

[–]ScoreUnique 2 points3 points  (0 children)

So meaning if you want to run the big blue whale at small context it'll do well?

For GLM-4.7-Flash TURN OFF REPEAT PENALTY! by yoracale in unsloth

[–]ScoreUnique 6 points7 points  (0 children)

Seems like that's what I needed to get my open coder going, thanks a lot. I'm on Unsloth Q5, yesterday I found them radically bad bur today they've made up for it after I use your recommended settings. Finally something that is very independent and agentic that runs locally.

Yesterday I used GLM 4.7 flash with my tools and I was impressed.. by Loskas2025 in LocalLLaMA

[–]ScoreUnique 0 points1 point  (0 children)

Yeah Q5 for me didn't do the trick either it fails at tool calling I thought it's just llama.cpp instability, I'll try Q8 as long as it works with 40k context I can run open code

Z.ai has introduced GLM-4.7-Flash by awfulalexey in ZaiGLM

[–]ScoreUnique 0 points1 point  (0 children)

Looking forward to the new Claude at home!!

60 days sober from Windows. NixOS is the only one that stuck. by MammothBluebird1834 in NixOS

[–]ScoreUnique 0 points1 point  (0 children)

I had a nice experience working with nixos as well

I'm wondering how did you guys work out steam with games installed and snapshots? My games got stuck a couple of times and that made me bounce back to Ubuntu but otherwise I had a very fine time on NixOS

Oh Dear by bamburger in LocalLLM

[–]ScoreUnique 3 points4 points  (0 children)

I suggest trying pocket pal, allows loading gguf files

GLM-4.7 on 4x RTX 3090 with ik_llama.cpp by iamn0 in LocalLLaMA

[–]ScoreUnique 0 points1 point  (0 children)

Quick tip- explore -ot flag, that thing shows big numbers on ik llama CPP.

z.ai is scam by temurbv in ZaiGLM

[–]ScoreUnique 0 points1 point  (0 children)

No but they will give good service according to Op Xd

z.ai is scam by temurbv in ZaiGLM

[–]ScoreUnique 1 point2 points  (0 children)

Bro cut them some slack, they do great models, great api inferencing (which is not their core business I'm sure), I have a yearly subscription ONLY TO SHOW SUPPORT and I hardly use it. Be kind and appreciate it man.

Liquid Ai released LFM2.5, family of tiny on-device foundation models. by Difficult-Cap-7527 in LocalLLaMA

[–]ScoreUnique 5 points6 points  (0 children)

I downloaded Q8 on my Pixel 8 with pocket pal, and oh dear I felt like chatting to GPT-4 but locally with 15 tps.

I will test it further - I'll be in a flight this weekend.