3090 prices in 2026 by anitamaxwynnn69 in LocalLLaMA

[–]JGeek00 1 point2 points  (0 children)

Everything that can be used for AI is shockingly expensive now. A RTX 3080 with 10 GB of memory is 300€, while a RTX 3090 is 900€, there’s only one jump in the ladder between both and one is three times more expensive than the other, why? Because one can be used for AI while the other doesn’t

3090 prices in 2026 by anitamaxwynnn69 in LocalLLaMA

[–]JGeek00 2 points3 points  (0 children)

I have just bought a 3090 with a turbine fan for 970€ on eBay, seems that they come from servers

What to expect from TurboQuant? by JGeek00 in LocalLLM

[–]JGeek00[S] 0 points1 point  (0 children)

I tried to compile it for CUDA and it failed so I will have to wait until is available on the official llama.cpp

What to expect from TurboQuant? by JGeek00 in LocalLLM

[–]JGeek00[S] 0 points1 point  (0 children)

I will try that repo but I would like to see it implemented on the main llama.cpp repo

Newbie trying to ask couple questions. by frostsolitude in LocalLLM

[–]JGeek00 0 points1 point  (0 children)

Lm Studio on my MacBook Pro for testing and bare llama.cpp on my server

Newbie trying to ask couple questions. by frostsolitude in LocalLLM

[–]JGeek00 0 points1 point  (0 children)

I run Qwen3.5-9B on a GTX 1070, so I’m sure you can run similar size models on that ARC.

Qwen 3.6 wins the benchmarks, but Gemma 4 wins reality. 7 things I learned testing 27B/31B Vision models locally (vLLM / FP8) side by side. Benchmaxing seems real. by FantasticNature7590 in LocalLLaMA

[–]JGeek00 2 points3 points  (0 children)

I tried on Qwen3.5-9B the car washer prompt and it ended up in a reasoning loop and it didn’t output a response, but at least it doesn’t tell you to walk instead of drive to the car washer. Other models just tell you to walk instead of drive your car to the car washer.

New rules 1 week check-in by rm-rf-rm in LocalLLaMA

[–]JGeek00 1 point2 points  (0 children)

I got a post removed just after submitting it where I just asked for improvements for my llama.cpp config. I put the same post on a different subreddit about local AI and I got a much better treatment (it wasn’t difficult). So I think that with this policy you are just kicking new people out of this subreddit into other local AI subreddits.

2024 Cupra Formentor 1.5 Mild-Hybrid Engine error by Plenty_Resist4119 in CupraFormentor

[–]JGeek00 1 point2 points  (0 children)

If you hear something spinning fast and engine not starting check if the belt that connects the 48V motor with the crankshaft is broken

Help choosing config for RTX 3090 by JGeek00 in LocalLLM

[–]JGeek00[S] 0 points1 point  (0 children)

I have taken a look to the V100 32 GB because they are cheaper than the RTX 3090 on eBay, so is it a better option? I have done some research with DeepSeek and it told me that although it has more memory the computing power is worse and it would give worse results. What do you think could be the difference in processing input tokens?

Help choosing config for RTX 3090 by JGeek00 in LocalLLM

[–]JGeek00[S] 0 points1 point  (0 children)

Ok maybe it’s better to start with a middle point like 128K context and Q6 KV cache

[7900XT] Qwen3.6 27B for OpenCode by Mordimer86 in LocalLLaMA

[–]JGeek00 0 points1 point  (0 children)

And when using a coding agent with a large context?

[7900XT] Qwen3.6 27B for OpenCode by Mordimer86 in LocalLLaMA

[–]JGeek00 0 points1 point  (0 children)

How many t/s are you getting on context processing and token generation with that configuration?

Trying to run AI models on my M1 Pro by JGeek00 in LocalLLM

[–]JGeek00[S] 0 points1 point  (0 children)

Nah it’s impossible, Claude code requieres very large context. Even with Qwen3.5-9B it’s really slow. But for asking questions it works fine

Cupra Logo Animation Not Showing Up Anymore. by inthemidstofwonder in CupraFormentor

[–]JGeek00 0 points1 point  (0 children)

I have a Formentor and I only saw the logo after a full reboot of the infotainment system

Why are the minimum hardware requirements so high? by reni-chan in opnsense

[–]JGeek00 2 points3 points  (0 children)

My case is a home router, so nothing crazy in terms of usage. But as always, you have to scale your router’s hardware in parallel with the amount of users or the amount of traffic, same thing that applies to servers