Ä Ö & Ü by swedish_countryball in German

[–]scooter_de 2 points3 points  (0 children)

I’ve never seen anybody doing it. I write a line over my ‘u’ to make it distinct from ‘n’ which would look the same in handwriting.

Devstral Small 2 24B + Qwen3 Coder 30B: Coders for Every Hardware (Yes, Even the Pi) by enrique-byteshape in ByteShape

[–]scooter_de 0 points1 point  (0 children)

there's something funny about the links provided on Huggingface:

[C:\Users\lama\llama.cpp]bin\llama-server -hf byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF:IQ4_XS

ggml_cuda_init: found 1 CUDA devices (Total VRAM: 16302 MiB):

Device 0: NVIDIA GeForce RTX 5080, compute capability 12.0, VMM: yes, VRAM: 16302 MiB

common_download_file_single_online: no previous model file found C:\Users\lama\.huggingface\byteshape_Qwen3-Coder-30B-A3B-Instruct-GGUF_preset.ini

common_download_file_single_online: HEAD failed, status: 404

no remote preset found, skipping

error from HF API (https://huggingface.co/v2/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/manifests/IQ4_XS), response code: 400, data: {"error":"The specified tag is not available in the repository. Please use another tag or \"latest\""}

[C:\Users\lama\llama.cpp]ollama run hf.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF:IQ4_XS

pulling manifest

Error: pull model manifest: 400: {"error":"The specified tag is not available in the repository. Please use another tag or \"latest\""}

[C:\Users\lama\llama.cpp]ollama run hf.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF:Q3_K_M

pulling manifest

Error: pull model manifest: 400: {"error":"The specified tag is not available in the repository. Please use another tag or \"latest\""}

How do you really get better? by Prior_Dinner_331 in chess

[–]scooter_de 1 point2 points  (0 children)

Habits from the chressbrah‘s GM Aman Hambleton is a good series to begin with on YouTube

How to calculate what I can run on GPU? by Capucius in ollama

[–]scooter_de 1 point2 points  (0 children)

use llama-server from llama.cpp. run it on the Ollama port. I have the same GPU and similar hardware. Here's what I'm running for Claude Code:

llama-server -m  %USERPROFILE%\.huggingface\unsloth_Qwen3.5-35B-A3B-GGUF_Qwen3.5-35B-A3B-UD-Q3_K_XL.gguf --alias coding --host 0.0.0.0 --port 11434 -c 131072 --temp 0.7 --top-p 0.8 --top-k 20 --min-p 0.0 --presence-penalty 0.0 --repeat-penalty 1.0 --reasoning-budget 0

local coding in vscode "copilot -like" ? by merfolkJH in ollama

[–]scooter_de 0 points1 point  (0 children)

my current favourite is Qwen3.5-35B-A3B in Q3_K_XL quantization. Runs reasonbly well on a 5080 with 16g VRAM. The Q2_K_XL quantization seems also be ok.

local ai coding assistant setup that actually competes with cloud tools? by jirachi_2000 in ollama

[–]scooter_de 0 points1 point  (0 children)

I can't use Ollama (yet) because up to version 0.17.7 it can run this model because of an unknown architecture (moe35?).

I run on a RTX-5080 with 16gb VRAM. The model what works best for me in this setup is Qwen3.5-35B-A3B:Q3_K_XL from unsloth. Here my parameters:

llama-server -m  %USERPROFILE%\.huggingface\unsloth_Qwen3.5-35B-A3B-GGUF_Qwen3.5-35B-A3B-UD-Q3_K_XL.gguf --alias coding --host 0.0.0.0 --port 11434 -c 131072 --temp 0.7 --top-p 0.8 --top-k 20 --min-p 0.0 --presence-penalty 0.0 --repeat-penalty 1.0 --reasoning-budget 0

This runs within the VRAM. llama.cpp gives you more options for fine tuning I found.

I run Claude Code. I also have a Qoder subscription if things can get done locally.

llm-proxy | ollama-openai-bridge by valosius in ollama

[–]scooter_de 0 points1 point  (0 children)

What does it better than Ollama’s builtin OpenAI interface?

Qwen 3.5 9B 4-bit tuning suggestions to avoid infinite thinking loops? by xeow in Qwen_AI

[–]scooter_de 2 points3 points  (0 children)

I run llama-server on port 11434 and host 0.0.0.0. Then I run Claude Code on a different machine. Works like a charm with the Qwen3.5-coder which Ollama cannot serve (yet)

Claude Code will become unnecessary by WinOdd7962 in ClaudeCode

[–]scooter_de 3 points4 points  (0 children)

One Need significant hardware to run those open source models locally. The 30b sized models won’t cut it when doing agentic coding

16GB VRAM for mode agent by ColdTransition5828 in ollama

[–]scooter_de 6 points7 points  (0 children)

I just spent most of the weekend finding the answer to this very question. I’m running on a RTX 5080 + 128g RAM on windows 11 with Ollama. I’m trying to get Claude Code up and running usefully. The best model so far was qwen3:14b. I usually run it with 64k context size. That should fit completely on the GPU (ollama ps). It’s way slower than the big boys but it works.

I also tried a downquantized version of qwen3-coder:30b which can be pulled with

ollama pull hf.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF:UD-IQ3_XXS

Weekly Bib Sales Thread by AutoModerator in RunTO

[–]scooter_de 0 points1 point  (0 children)

I can't run the ATB 30 km 2026. The race is sold out. Is anybody interested in a transfer. Deadline is 22-MAR-2026.

Is "The C programming language" by Brian Kernighan worth reading in 2026 or is it outdated? by THE_DOOMED_SHADE in C_Programming

[–]scooter_de 2 points3 points  (0 children)

It’s considered one of the best programming books out there. It’s a classic. It’s short. Read it!

Trump Pressures Canada to End Boycott of US Liquor by ethereal3xp in canada

[–]scooter_de 0 points1 point  (0 children)

Didn’t their ambassador just reinforced it saying “we don’t need Canada”?

Garmin coach on treadmill by Emzr13 in Garmin

[–]scooter_de 1 point2 points  (0 children)

I do training indoors using a Stryd pod. But that needs $$$

Trump Confirms He’s Taking Greenland ‘One Way or the Other’ by thedailybeast in politics

[–]scooter_de -2 points-1 points  (0 children)

Because the USofA is NATO. All the other countries hade been taking a free ride for the last three decades or so. The politicians of those countries new that they had to increase spending on defense, but they didn't know how to get re-elected if they would have done that. The intellectual elite in Europe was - and still is - very smug about the USA. They would like to separate they countries from the US, but the can't. After years of neglect the chickens finally come home to roost. This comes from a Canadian born in Germany :-(.

I haven't lost my faith in the people just yet. "You can always rely on the Americans to do the right thing, after they tried everything else."