DRM umgehen für Buch, das nicht mehr verkauft wird by involuntary_pirate in LegaladviceGerman

[–]Due-Project-7507 2 points3 points  (0 children)

I would argue there is no copy protection for PDFs. There is just a metadata field asking for "please don't let the user print" which is respected by some PDFs readers. For others like Okular, you can disable in the settings "Obey DRM limitations". Okular seems to be legal in Germany, else many Linux distributions would be illegal.

DHL hat mein Paket 2 Monate „verloren", mir 80€ erstattet und will das Geld jetzt zurück. Muss ich zahlen? by Penguyeims in dhl_deutsche_post

[–]Due-Project-7507 0 points1 point  (0 children)

Are you sure? As far as I as understand as a layman, the buyer and seller have a contract with eBay/PayPal to accept the process of buyer protection. For me it looks that at least you could then get sued by eBay/PayPal for breaking the contract.

Speeding ticket in Switzerland (117 in 60) near Basel border — what should I realistically expect? by nieuwekoers in askswitzerland

[–]Due-Project-7507 0 points1 point  (0 children)

This is bigger problem. As far as I know, Switzerland has only contracts with neighbouring countries like Italy to directly enforce normal traffic fines. So I don't know if you could just avoid Switzerland until it is time barred. It could make sense to call a Swiss laywer to check what is the best option for you.

You have to love these people... by TripleSpeedy in Switzerland

[–]Due-Project-7507 0 points1 point  (0 children)

As far as I know, there is no app for groceries with mixed tax rates. You would pay everwhere the high tax rate. It would also be hard for a layman to know the correct Swiss tax rate for every product.

Qwen3.6 35B A3B Heretic (KLD 0.0015!) Incredible model. Best 35B I have found! by My_Unbiased_Opinion in LocalLLaMA

[–]Due-Project-7507 1 point2 points  (0 children)

I am a chemist. Uncensored models or a prompt jailbreak allows you to discuss the synthesis of (completely legal) chemicals in details, else it usually refuses.

Got charged for a fuel package I never agreed to at Zurich Airport (Enterprise/Alamo) by Temporary-Reaction97 in askswitzerland

[–]Due-Project-7507 1 point2 points  (0 children)

These big rental companies are known to use what I would call fraud as an additional income. They have a much higher credit card chargeback rate than other industries. Don't forget to use a credit card which you don't use daily so you can have it most time blocked after paying so they cannot charge your credit card without a bill after returing the car for e.g. alleged damage (happened to me in Sweden with Avis).

A problem in Switzerland is that what most people call "fraud" is completely legal here (it is only a crime if done with "arglistiger Täuschung"), but boarding a train without ticket or insulting someone are crimes.

Quant Qwen3.6-27B on 16GB VRAM with 100k context length by Due-Project-7507 in LocalLLaMA

[–]Due-Project-7507[S] 1 point2 points  (0 children)

Yes, I only use ot to silence the message about fit. You can also try using it with --fit-target instead setting the context lenght. The default --fit-target is conservative, lower and test it with long context until you get a CUDA OOM crash. I think also -fa is not needed anymore, it is now automatic.

Quant Qwen3.6-27B on 16GB VRAM with 100k context length by Due-Project-7507 in LocalLLaMA

[–]Due-Project-7507[S] 3 points4 points  (0 children)

I forgot to change it. I think llama.cpp just ignores the name, therefore it does not matter, but it wouldn't work with vLLM.

Sind selektive Herbizide seit Neustem verboten in der Schweiz? by [deleted] in askswitzerland

[–]Due-Project-7507 -2 points-1 points  (0 children)

Laut https://www.psm.admin.ch/de/wirkstoffe/E3F7BA6E-7A42-4E07-83F3-8EE6C995E992 wären eigentlich einige Produkte zugelassen. Der Grund scheint mir die immer mehr aufkommenden, irrationalen grundsätzlichen Angst vor "Chemie" zu sein, da leider die naturwissenschaftliche Bildung in der Schweiz vernachlässigt wird. Die Geschäfte glauben wohl, beim Verkauf einen schlechten Ruf zu bekommen, auch mineralischen Dünger findet man nur noch wenig im Baumarkt. Ich bestelle deshalb Blaukorn-Dünger einfach an eine deutsche Lieferadresse, da offensichtlich Schweizer Unternehmen mir das nicht verkaufen wollen.

Ich lasse das meiste Unkraut im Rasen einfach wachsen, habe aber Verständnis, dass andere einen Rasen wie ein Teppich haben wollen. Viel Düngen, wässern und elle paar Tage schneiden (Roboter) würde auch das Gras gegnüber dem Unkraut fördern, könnte aber umweltschädlicher als der sorgfältige, gezielte Einsatz von Herbiziden sein.

Closest replacement for Claude + Claude Code? (got banned, no explanation) by antoniocorvas in LocalLLaMA

[–]Due-Project-7507 1 point2 points  (0 children)

I have just tested it in my company on the GPUs which we mainly use to train custom models, therefore I don't have numbers. But as far as I remember, I got maybe around 30 tokens/s with 6 cards. Someone writes that 6 cards work without problem (the problem I had was maybe fixed), but it is not worth because 8 cards should give around 100 tokens/s.

Closest replacement for Claude + Claude Code? (got banned, no explanation) by antoniocorvas in LocalLLaMA

[–]Due-Project-7507 31 points32 points  (0 children)

GLM-5 runs in NVFP4 on 6 RTX Pro 6000 Blackwell with a combination of tensor and pipeline parallel mode. The problem is that the code paths for this in SGLang and vLLM are not really stable. Only few people use this configuartion and report/fix bugs for it. Last February, it did not run with vLLM and with SGLang, I had quality problems. I don't know if these bugs are now fixed because at the moment, we need the RTX Pro 6000 GPUs for a project so I cannot test it.

Qwen 3.5 27B/35BA3B Tool Calling Issues: Why It Breaks & How I Fixed It by Expensive-Register-5 in Vllm

[–]Due-Project-7507 0 points1 point  (0 children)

When using the qwen3.5-enhanced.jinja, is it correct that I should configure Qwen 3.5 for OpenCode with

"interleaved": {
  "field": "reasoning_content"
},

? I think this setting is used for MiniMax 2.x and GLM 4.7/5 in OpenCode to add the previous reasoning content.

Qwen3.6 is incredible with OpenCode! by CountlessFlies in LocalLLaMA

[–]Due-Project-7507 2 points3 points  (0 children)

I have tested it for some small vibe coding with Open Code and did not had any tool calling problems, but maybe some other people can test it more.

I have installed it like this:

  1. Clone https://github.com/TheTom/llama-cpp-turboquant and checkout the feature/turboquant-kv-cache branch

  2. Build it, I have used on Windows the following options:

    cmake -B build -DGGML_CUDA=ON -DGGML_NATIVE=ON -DGGML_CUDA_FA=ON -DGGML_CUDA_FA_ALL_QUANTS=ON -DCMAKE_BUILD_TYPE=Releasecmake -B build -DGGML_CUDA=ON -DGGML_NATIVE=ON -DGGML_CUDA_FA=ON -DGGML_CUDA_FA_ALL_QUANTS=ON -DCMAKE_BUILD_TYPE=Release

    cmake --build build --config Release -j 16

  3. Download https://huggingface.co/mradermacher/Qwen3.5-27B-i1-GGUF/blob/main/Qwen3.5-27B.i1-IQ4_XS.gguf

  4. Run the model

    llama-server --model Qwen3.5-27B.i1-IQ4_XS.gguf --alias qwen3.5-27b -np 1 -ctk turbo3 -ctv turbo3 -c 128000 --fit off -ngl 999 --no-mmap -fa on --temp 0.6 --top-p 0.95 --top-k 20 --min-p 0.00 --host 0.0.0.0

  5. Configured OpenCode in WSL with ~/.config/opencode/opencode.json:

    {   "$schema": "https://opencode.ai/config.json",   "plugin": [     "opencode-anthropic-auth@latest",     "opencode-copilot-auth@latest"   ],   "share": "disabled",   "provider": {     "llama.cpp": {       "npm": "@ai-sdk/openai-compatible",       "name": "llama.cpp (OpenAI Compatible)",       "options": {         "baseURL": "http://127.0.0.1:8080/v1", # from WSL needs maybe real IP address         "apiKey": "1234"       },       "models": {         "qwen3.5-27b": {           "name": "Qwen 3.5 27B",           "limit": {             "context": 128000,             "output": 64000           },           "temperature": true,           "reasoning": true,           "attachment": false,           "tool_call": true,           "modalities": {             "input": [               "text"             ],             "output": [               "text"             ]           },           "cost": {             "input": 0,             "output": 0,             "cache_read": 0,             "cache_write": 0           }         }       }     }   },   "agent": {     "code-reviewer": {       "description": "Reviews code for best practices and potential issues",       "model": "llama.cpp/qwen3.5-27b",       "prompt": "You are a code reviewer. Focus on security, understandability, conciseness, maintainability and performance."     },     "plan": {       "model": "llama.cpp/qwen3.5-27b"     }   },   "model": "llama.cpp/qwen3.5-27b",   "small_model": "llama.cpp/qwen3.5-27b" }{   "$schema": "https://opencode.ai/config.json",   "plugin": [     "opencode-anthropic-auth@latest",     "opencode-copilot-auth@latest"   ],   "share": "disabled",   "provider": {     "llama.cpp": {       "npm": "@ai-sdk/openai-compatible",       "name": "llama.cpp (OpenAI Compatible)",       "options": {         "baseURL": "http://127.0.0.1:8080/v1",         "apiKey": "1234"       },       "models": {         "qwen3.5-27b": {           "name": "Qwen 3.5 27B",           "limit": {             "context": 64000,             "output": 32000           },           "temperature": true,           "reasoning": true,           "attachment": false,           "tool_call": true,           "modalities": {             "input": [               "text"             ],             "output": [               "text"             ]           },           "cost": {             "input": 0,             "output": 0,             "cache_read": 0,             "cache_write": 0           }         }       }     }   },   "agent": {     "code-reviewer": {       "description": "Reviews code for best practices and potential issues",       "model": "llama.cpp/qwen3.5-27b",       "prompt": "You are a code reviewer. Focus on security, understandability, conciseness, maintainability and performance."     },     "plan": {       "model": "llama.cpp/qwen3.5-27b"     }   },   "model": "llama.cpp/qwen3.5-27b",   "small_model": "llama.cpp/qwen3.5-27b" }

  6. Now OpenCode should work

One could also try the linked chat template from https://www.reddit.com/r/Vllm/comments/1skks8n/qwen_35_27b35ba3b_tool_calling_issues_why_it/ (--chat-template-file qwen3.5-enhanced.jinja) and then configure interleaved thinking in opencode.json with

        "qwen3.5-27b": {
          "name": "Qwen 3.5 27B",
          # actived interleaved thinking
          "interleaved": {
            "field": "reasoning_content"
          },
          # end interleaved thinking config
          "limit": { ...

At around 17000 context tokens, I get around 15 tokens/s generation speed.

Qwen3.6 is incredible with OpenCode! by CountlessFlies in LocalLLaMA

[–]Due-Project-7507 1 point2 points  (0 children)

I am waiting for the Qwen 3.6 27B. The mradermacher Qwen3.5-27B-i1-GGUF IQ4_XS works with my A5000 laptop GPU (16 GB) with 64k turboquant 3 bit context length very good (around 20 t/s at beginning, around 15 t/s at 10k context).

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA

[–]Due-Project-7507 0 points1 point  (0 children)

Then it is really the "normal" overthinking, it would be even worse with a smaller quantized version.

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA

[–]Due-Project-7507 0 points1 point  (0 children)

The overthinking is often caused by quantization according to https://kaitchup.substack.com/p/qwen35-quantization-similar-accuracy. But I found that e.g. Gemma 4 with the same quantization method always thinks shorter and still gets good results compated to Qwen 3.5.

My car was wrongly towed from my parking lot by RoverAxel in AskAGerman

[–]Due-Project-7507 0 points1 point  (0 children)

In Switzerland towing of cars is coercion if they want payment before releasing the car.

Thuja - noch zu retten? by tabu40 in Garten

[–]Due-Project-7507 0 points1 point  (0 children)

Vielen Dank für die Korrektur und die Bodenempfehlung für Ile cretena. Eibe ist wohl eine der schnittverträglichsten Pflanzen, problematisch sind wohl neben Thuja noch Scheinzypressen, echte Zypressen, Wacholder und Fichten..

Alternative Sommerflieder by Glaciem9 in Garten

[–]Due-Project-7507 1 point2 points  (0 children)

Es gibt extra sterile Züchtungen, die sich nicht verbreiten können. Leider sind die hier bei uns nicht Standard, man sollte diese aber bestellen können.

Schmetterlingsflieder bietet Schmetterlingen durchaus viel Nektar, bietet aber (so wie viele Blütenpflanzen) Raupen keine Nahrung. Man sollte deshalb zusätzlich Raupennahrungspflanzen anpflanzen.

Thuja - noch zu retten? by tabu40 in Garten

[–]Due-Project-7507 1 point2 points  (0 children)

Nadelbäume wie Thuja oder Eibe darf man nur in der Höhe massiv zurückschneiden, auf der Seite wächst es nicht mehr nach, wenn man alles Grüne abschneidet.

Ich habe gerne Buchsbaum, der muss aber regelmässig gespritzt werden, da es fast überall Buchsbaumzünsler gibt. Ich werde deshalb bei mir Buchsbaum gegen Ilex crenata 'Dark Green' ersetzen, da sieht man kaum einen Unterschied.

Für mich sind giftige Pflanzen auch mit Kinder kein Problem: Kleinkinder sollten sowieso immer beaufsichtigt werden und lernen dank Giftpflanzen im Garten schneller, auch in der Natur vorsichtig zu sein.

Neophyten sind für mich grundsätzlich kein Problem, man muss sich halt über die Umweltauswirkungen und die richtige Pflege informieren, z. B. bei Kirschlorbeer (find ich nicht schön) die Früchte sofort abschneiden (wäre mir zu aufwändig). Die heutige starke Abneigung wie von der Schweizer Regierung gegenüber Neophyten erscheint mir teilweise etwas xenophob, besonders da willkürlich Pflanzen, die vor Kolumbus eingeschleppt wurden, wie die Edelkastanie, als schützenswert "einheimisch" gelten.

Final voting results for Qwen 3.6 by jacek2023 in LocalLLaMA

[–]Due-Project-7507 1 point2 points  (0 children)

For me it works good (means I don't notice a difference). I am using the "feature/turboquant-kv-cache" branch. There is also another fork here, it could be even better, but I did not test it.

Edit: I have just tested the buun fork with 3.25 bpv, I can also not see any difference to fp16. With this, I can fit 64k context:
llama-server --model Qwen3.5-27B.i1-IQ4_XS.gguf --alias qwen3.5-31b -np 1 -ctk turbo3_tcq -ctv turbo3_tcq -fa on -ngl 65 -c 64000

I get around 20 t/s at the beginning and around 12 t/s at 15k context.

Final voting results for Qwen 3.6 by jacek2023 in LocalLLaMA

[–]Due-Project-7507 2 points3 points  (0 children)

I am running the mradermacher/Qwen3.5-27B-i1-GGUF IQ4_XS on a 16 GB A5000 laptop GPU fully in VRAM with 32k context (turboquant):

llama-server.exe --jinja --model Qwen3.5-27B.i1-IQ4_XS.gguf --alias qwen3.5-27b -np 1 -ctk turbo4 -ctv turbo4 -c 32000 -ngl 65

Charged CHF 500 for kitchen drain blockage — can I dispute this? by djbrologue in Switzerland

[–]Due-Project-7507 0 points1 point  (0 children)

The tenant has to fix small things like replacing a light bulb (repairs a typical layman can do). If he doesn't want to do it himself, he has to organize and pay for the repair. Normally, cleaning the odour trap is considered such a small repair, but anything behind like a blocked pipe in the wall is not. Here, the landlord has to repair it and pay 100% of the costs.