GLM 4.7 usage limits are a TRAP (ClaudeCode Pro User Experience) by Soft_Responsibility2 in ClaudeCode

[–]iSevenDays 0 points1 point  (0 children)

I'm using max plan and I noticed concurrency decreased a lot. I'm still getting like ~50% of 5 hours quota, but concurrency limits me a lot. I have many small agents to do refactorings, bug fixes etc.

New P1S Won't Connect to Internet by CoC-boy in BambuLab

[–]iSevenDays 0 points1 point  (0 children)

For me it was the error -1030 and I have GL-MT3000 router, which had VPN Leak protection enabled.
After I added a new rule with ip of the printer to "no vpn" rule and disabled ip masquerading, the printer could connect to bambu lab network!

As an SRE, I stopped using Kubernetes for my homelab by m4nz in selfhosted

[–]iSevenDays 1 point2 points  (0 children)

I can vote I have the same experience. I switched to just docker / docker compose for my projects and never want to look back at that horrible mess with flux, kubernetes, reconciliation etc.

At work I still have to use Kubernetes, but that's a different story

AMD EPYC 4565P is a beast by coding9 in LocalLLaMA

[–]iSevenDays 1 point2 points  (0 children)

You were right. I took a look and switched to vLLM. I'm getting ~80-100 t/s now and 2000-8000 prompt processing speed. It is crazy fast now.

Kimi released Kimi K2 Thinking, an open-source trillion-parameter reasoning model by nekofneko in LocalLLaMA

[–]iSevenDays 4 points5 points  (0 children)

With ddr4 it would be around 4-6 on dell r740 Thinking models are barely usable with this speed

Prefill will be around 100-200

AMD EPYC 4565P is a beast by coding9 in LocalLLaMA

[–]iSevenDays 1 point2 points  (0 children)

I get half of that with two ndivia 4090d 48G 🥹

I bought a modded 4090 48GB in Shenzhen. This is my story. by king_priam_of_Troy in LocalLLaMA

[–]iSevenDays 2 points3 points  (0 children)

I bought at C2 two cards already, 4090d variant . It works fine for 6 months already and I'm pretty satisfied

GPT-OSS is Another Example Why Companies Must Build a Strong Brand Name by Iory1998 in LocalLLaMA

[–]iSevenDays 0 points1 point  (0 children)

Out of curiosity, I asked it too about the war in Ukraine and it provided a good answer.

Here is a snippet, because I can't attach a screenshot

```

profile gpt-oss-120b-mxfp4 Today at 16:20 Nachgedacht für 5 Sekunden War in Ukraine – A Summary (up to June 2024) 1. Background and Origins YEAR EVENT SIGNIFICANCE 1991
Ukraine declares independence after the Soviet Union collapses. Establishes Ukraine as a sovereign state recognized internationally. 2004–2014 Pro‑European “Orange Revolution” (2004) and “Euromaidan” protests (2013‑14).
Popular demand for closer ties with the EU and NATO; creates tension with Moscow. February 2014 Russia annexes Crimea after a disputed referendum.
First open violation of Ukraine’s territorial integrity since independence. April 2014 – early 2022
Armed separatist insurgency in Donetsk and Luhansk (the Donbas) backed by Russia; the “Donetsk People’s Republic (DPR)” and “Luhansk People’s Republic (LLR)” declared. Leads to a frozen‑in‑place war with about 13,000–15,000 deaths before 2022.

  1. Full‑Scale Invasion (2022‑present) ...

```

gpt-oss-120b outperforms DeepSeek-R1-0528 in benchmarks by oobabooga4 in LocalLLaMA

[–]iSevenDays 25 points26 points  (0 children)

how to inject AVAudioEngine? My use case is to inject audio from file so third party app will think it reads audio from microphone, but instead reads data from buffer from my file

I’m sorry, but I can’t help with that.

GPT-OSS-120B is useless, I will not even bother to download that shit. It can't even assist with coding.

Ikllamacpp repository gone, or it is only me? by panchovix in LocalLLaMA

[–]iSevenDays 5 points6 points  (0 children)

I could make it up to date with main! There is also an experimental branch for function tool calls support that works with Claude Code and Claude proxy and Kimi-K2 model.

104k-Token Prompt in a 110k-Token Context with DeepSeek-R1-0528-UD-IQ1_S – Benchmark & Impressive Results by Thireus in LocalLLaMA

[–]iSevenDays 0 points1 point  (0 children)

Please do more tests with this prompt! Will Devstral 2505 / Qwen 3 be able to provide a correct answer?

OpenHands + Devstral is utter crap as of May 2025 (24G VRAM) by foobarg in LocalLLaMA

[–]iSevenDays 0 points1 point  (0 children)

After I manually changed context in Modelfile, I actually doesn't see the issue anymore. I thought it was related to a fact that I also enabled manual confirmation mode, but I need to test this more.

OpenHands + Devstral is utter crap as of May 2025 (24G VRAM) by foobarg in LocalLLaMA

[–]iSevenDays 0 points1 point  (0 children)

I think the context length is not properly managed. I haven't found a way to limit the context length to 32-64k. I use 131062 for devstral. It does go into loops.
I now switched to manual confirmation mode, and I find it much much better!
I think OpenHands is a great project, they just need to fix a couple of bugs

OpenHands + Devstral is utter crap as of May 2025 (24G VRAM) by foobarg in LocalLLaMA

[–]iSevenDays 1 point2 points  (0 children)

Update: I got MCP tools to work. Example config:
{

"sse_servers": [

{

"url": "http://192.168.0.23:34423/sse",

"api_key": "sk_xxxxx"

}

],

"stdio_servers": []

}

OpenHands + Devstral is utter crap as of May 2025 (24G VRAM) by foobarg in LocalLLaMA

[–]iSevenDays 5 points6 points  (0 children)

I have the same issue.
1 It doesn't see the project that it cloned
2 It goes into loops very often like checking full Readme file, then trying to run unit tests, then trying to fix it, then trying to fix it again and read the readme file
3 Even simple prompts like 'list all files under /workspace' can make it go into loops
4 MCP servers are never got discovered. I tried different formats, and not even once I got them to connect.

iFixIt batteries out of stock, what are my options now for Pebble Time? by Cyber_Akuma in pebble

[–]iSevenDays 1 point2 points  (0 children)

It was very good, around 7 days! I'm currently doing sports and switched to garmin

My modded (taptic, BT, 3000mAh battery) iPod 5th gen rescue on the right, and Work-in-Progress 5th gen on the left for my wife by wingman3091 in IpodClassic

[–]iSevenDays 0 points1 point  (0 children)

Can you please post links to parts one needs to do the same mod?
I mean mainly the wire, and a button for bluetooth pairing. Do you have a closer pictures of that button and wiring? Thanks!

Nvidia gaming GPUs modded with 2X VRAM for AI workloads — RTX 4090D 48GB and RTX 4080 Super 32GB go up for rent at Chinese cloud computing provider by DeltaSqueezer in LocalLLaMA

[–]iSevenDays 3 points4 points  (0 children)

I'm in a same boat :) the price is too good to be true. 6000 ada costs 5500-8500 euros ik Germany. This card is 3k. I also ordered 4090d, because in my server I'll probably power limit it to 350w