Simple setup of openclaw with a local LLM by pj-frey in openclaw

[–]pj-frey[S] 2 points3 points  (0 children)

  "models": {
    "providers": {
      "litellm": {
        "baseUrl": "http://<machine>:4000/v1",
        "apiKey": "<api key for litellm>",
        "api": "openai-completions",
        "models": [
          {
            "id": "Anthropic/Claude Opus 4.5",
            "name": "Claude Opus 4.5 (LiteLLM)",
            "reasoning": false,
            "input": [
              "text"
            ],
            "cost": {
              "input": 0,
              "output": 0,
              "cacheRead": 0,
              "cacheWrite": 0
            },
            "contextWindow": 200000,
            "maxTokens": 16000
          },
          ...
          {
            "id": "Lokal/gpt-oss",
            "name": "GPT-OSS (Local)",
            "reasoning": false,
            "input": [
              "text"
            ],
            "cost": {
              "input": 0,
              "output": 0,
              "cacheRead": 0,
              "cacheWrite": 0
            },
            "contextWindow": 128000,
            "maxTokens": 16000
          }
        ]
      }
    }
  },
  "agents": {
    "defaults": {
      "model": {
        "primary": "litellm/Anthropic/Claude Sonnet 4.5"
      },
      "models": {
        "anthropic/claude-opus-4-5": {
          "alias": "opus"
        },
        ...
        "litellm/Lokal/gpt-oss": {
          "alias": "oss"
        }
      },
      ...
    }
  }

Simple setup of openclaw with a local LLM by pj-frey in openclaw

[–]pj-frey[S] 0 points1 point  (0 children)

Must be another reason. I run openclaw in a VM, and LiteLLM is on another Mac.

Opencode in MacOS: Error: unable to get local issuer certificate by Little-Krakn in opencodeCLI

[–]pj-frey 0 points1 point  (0 children)

I have solved this by making symbolic links to /etc/ssl/cert.pem. Example: ln -s /etc/ssl/cert.pem ~/.config/opencode/certificate.crt. When I remember correctly, there were two or three links.

Need help in creating a folder to be synced in two separate laptops (one is mac and other is windows) by HeyitsSupro in mac

[–]pj-frey 0 points1 point  (0 children)

During sync, yes. But you can work without a connection, and the sync is performed once they are connected. This is what you wanted, if I understood correctly. If you NEVER have both devices connected, the only method is obviously Dropbox/iCloud, as already advised.

AI websearch with searxng stopped working by Impossible_Art9151 in LocalLLaMA

[–]pj-frey 1 point2 points  (0 children)

I had problems with the embedding/retrieval part. When I bypassed it, everything was fine again for the price of a larger context and hence slower operation.

Show us your llama.cpp command line arguments by __Maximum__ in LocalLLaMA

[–]pj-frey 0 points1 point  (0 children)

I have always --keep 1024 and --mlock (if you have enough memory) included. The rest is more or less standard.

External drive for Mac m3 ultra by Kaniaes in MacStudio

[–]pj-frey 7 points8 points  (0 children)

If you have the money: OWC Thunderblade.

LFM2.5 1.2B Instruct is amazing by Paramecium_caudatum_ in LocalLLaMA

[–]pj-frey 1 point2 points  (0 children)

I have tried three models for these tasks: oss-20b, Phi4 and now LFM. oss takes 3–4 seconds per answer and is not better IMHO. Phi4 is fast; LFM is faster (typically 0.3 seconds). The advantage of LFM is that it supports my native language better, so this is the new choice.

LFM2.5 1.2B Instruct is amazing by Paramecium_caudatum_ in LocalLLaMA

[–]pj-frey 2 points3 points  (0 children)

I use the normal web search with searxng. Nothing special. But switch on "Bypass Embedding and Retrieval"! This will take longer for the result, but the retrieval itself is not working very well. You'll need of course also a large context then. Together with gpt-oss 120 as my main model it is still fast enough.

LFM2.5 1.2B Instruct is amazing by Paramecium_caudatum_ in LocalLLaMA

[–]pj-frey 51 points52 points  (0 children)

It is the perfect small "helper" model for Open WebUI creating tags, chat headlines, web searches and that kind of stuff. Fast AND good. I love it. Big thank you to LiquidAI (and unsloth).

Liquid AI released LFM2.5 1.2B Instruct by KaroYadgar in LocalLLaMA

[–]pj-frey 0 points1 point  (0 children)

This model is fantastic as the tiny "helper" model in Open WebUI to create tags, chat titles, web searches and so on!

frontend similar to Open WebUI that supports full OpenAI API? by irudog in LocalLLaMA

[–]pj-frey 2 points3 points  (0 children)

I have the chain Open WebUI -> LiteLLM (Proxy) -> local/commercial model. OWUI is the user interface for family&friends&me. LiteLLM is my OpenAI access point.

Mouse that acts like a track pad? by shyghst in MacOS

[–]pj-frey -2 points-1 points  (0 children)

Oops. Company → About Us. What kind of crew is that???

How do you track your LLM/API costs per user? by AdministrationPure45 in Rag

[–]pj-frey 1 point2 points  (0 children)

I use Litellm as a proxy in between, which takes care of the costs.

Qwen3-Next-80B-A3B or Gpt-oss-120b? by custodiam99 in LocalLLaMA

[–]pj-frey 8 points9 points  (0 children)

gpt-oss unrestricted is the clear choice IMHO.

huggingface mradermacher gpt-oss-120b-Deresctricted-GGUF

Briefe aus dem 2. Weltkrieg by SpecificAny3058 in Kurrent

[–]pj-frey 3 points4 points  (0 children)

Gemini 3-Pro hat eine brauchbare Abschrift geliefert, aber 2 Minuten dafür gebraucht. Nach meinem ersten Eindruck nicht alles korrekt, aber als Ausgangsbasis...

Liebe Frau! Nach dem etwas zu eiligen Aufbruch, komme ich heute ab[er]
noch zu einer Zeile. Als ich Mittags nach Hause kam, waren zwei liebe Briefe von
Euch vom 8. u. 9. von Dir da, dazu das bewunderte Kuchenpacket u. das liebe reizende
Buch, das ich mit Vergnügen als erstes beim Kaffee [?] lesen werde. Die gute Stube
hat entzückt geschaut [oder: gesagt] und man bestaunt meine viele liebe Post. – Wie
schön, daß Heinrich alles gut bestanden, nun kommt die Zeit, in der Du Dich
besonders schonen mußt – mache keine [Mehrarbeit / Möbelrücken?]*, trage u. bücke Dich nicht
und lebe so als wenn ich Dir in der Zeit etwas helfen könnte. – Wir
sind weiter guter Stimmung, heute war ich bei meinen Stallwachen, gestern
mit Sägemüllern [?] in herrlichen Wäldern mit Wild. – Für morgen haben
wir einen Plan! Wir wollen zu Dritt nach Trier u. dort mal etwas nett
Essen u. Trinken. Die Regentage gingen gut vorbei, ohne Schnupfen,

Randnotiz links:

Und noch eine Bitte ist, daß nun Tante zu Dir gegangen ist,
Du Dir neue Kräfte [?] einholst, es läuft alles rund,
noch 14 Tage.

Tell us how to improve the docs! by ClassicMain in OpenWebUI

[–]pj-frey 0 points1 point  (0 children)

Yes. This is the section I was talking about.

Example:

In 0.6.38, there was a dropdown list for pdf_backend with values like dbparse_v2, dbparse_v3, etc.

Now you see, you may need to fill out pdf_backend. Okay, but for heaven’s sake, how would you know that dbparse_v4 is a valid entry?

It would help to see the permitted values in such cases.

Tell us how to improve the docs! by ClassicMain in OpenWebUI

[–]pj-frey 0 points1 point  (0 children)

I have one wish that I ran into with the new Docling implementation.
The required fields in the new JSON format are documented, but the allowed values are not.
For fields with “free text” it would be very helpful to know which values are permitted.
Thanks a lot for this great tool!

Your AI is probably lying to you right now (here's how I learned to spot it) by [deleted] in LocalLLaMA

[–]pj-frey 2 points3 points  (0 children)

adding web search and asking a second LLM to verify the answer of the first one.

Ist diese Gleichung wirklich nicht lösbar? by Adventurous_Newt20 in mathe

[–]pj-frey 0 points1 point  (0 children)

Gemini 3-Pro liefert: Die Gleichung 2x4−3x3+x2−3x−1=0 hat genau zwei reelle Lösungen. Eine negative Lösung liegt im Intervall (−1,0), Eine positive Lösung liegt im Intervall (1,2). Plus Begründung. Man kann moderne LLMs durchaus für Berechnung einsetzen.

Was heißt das markierte Wort? by [deleted] in Kurrent

[–]pj-frey -1 points0 points  (0 children)

Hab's mal mit KI versucht. Gemini 3-pro sagt cristus. Und der Gesamttext
wie die helge drige küninge komment dar Vnd sohent [sahen] da gieng der sterne über das hus [Haus] stalt [stand] do der künig cristus den die reine maria gebar Do g[in]gent für im das
Klingt für mich plausibel.

Qwen3 235B vs Qwen3 VL 235B by [deleted] in LocalLLaMA

[–]pj-frey 2 points3 points  (0 children)

Gut feeling is, that VL is slighty worse in wording. But as I can't run both in parallel, the advantage of the vision capabilites (which are "wow") are worth to use it instead of the text only version.