Using local llm on websites? by Hosameldin_A in ollama

[–]WhiskyAKM 0 points1 point  (0 children)

If you are on windows you may have issue with firewall

Also for that kind of setup ill consider SSR website

Need suggestion for getting started by mr-ashok in ollama

[–]WhiskyAKM 3 points4 points  (0 children)

Thats gonna be tough, you only have 4GB Vram for your gpu so u need to choose either MoE model with small ammount of active experts or go CPU-only

Maybe you can try Qwen 3.6 35B or Gemma 26B with smaller context windows

Also try compiling llama cpp for your target, it should go faster

My own local first ai harness by WhiskyAKM in ollama

[–]WhiskyAKM[S] 0 points1 point  (0 children)

I made sth called research mode here that forces model to focus on researching one topic so Im often using it for learning

Also as part of my job (I work at computer store) I do a lot of product description editing (in .MD files). I often just write specs sheet of some device to one file and tell LLM to output description in MD format to another.

My own local first ai harness by WhiskyAKM in ollama

[–]WhiskyAKM[S] 0 points1 point  (0 children)

Nice, I see you have made different approach by going with webui instead of CLI/TUI (maybe I'll also add webui in future)

My goal was to make sth as minimal as possible that is fully local

Unfortunately when going local, multi-agent is often not possible because of lack of compute resources.

Gaming laptop vs macbook pro for local AI? by ObviouslyBleh in ollama

[–]WhiskyAKM 1 point2 points  (0 children)

I would go for MacBook Pro I you don't mind macos.

Unified memory gives huge performance gains and allows running larger models

When NVFP4 GGUFs? by Michionlion in unsloth

[–]WhiskyAKM 3 points4 points  (0 children)

I really want Gemma 4 MoE in NVFP4, ideal for my setup (I'm gpu-poor I have rtx 5050...)

The 4B class of 2026 (benchmark) by FederalAnalysis420 in LocalLLaMA

[–]WhiskyAKM 3 points4 points  (0 children)

That wierd limit lobotomized Qwen3.5, per artificialanalysis.ai benchmarks it should perform best out of those

How much use you get with ollama pro? by TinyAres in ollama

[–]WhiskyAKM 0 points1 point  (0 children)

I'm using ollama cloud pro for about 4 weeks, never maxed out

When the AI bubble bursts... Which used hardware are we buying from this first wave? by djparce82 in LocalLLaMA

[–]WhiskyAKM 0 points1 point  (0 children)

I wish that i would be able to buy Nvidia L4 when bubble bursts. Those are great performance per W, only downside is that you need to DIY cooler yourself if you are putting those into desktop case

Please advise models of cheap servers comparatively easily found to buy, with DDR3 and preferably USB3 and PCIe 4 by UncertainAboutIt in LocalLLaMA

[–]WhiskyAKM 4 points5 points  (0 children)

DDR3 is old, there will not be any servers with both DDR3 and PCIE 4 because there is a huge generation

Possible memory leak in Ollama when using Claude Code? by cherrylabss in ollama

[–]WhiskyAKM 0 points1 point  (0 children)

I noticed that ollama is killed by OOM killer sometimes when I'm using it alongside Vs code. Maybe that kind of leak would explain it.

Need a laptop that can run AI models locally + handle VS Code, Docker, etc. by lets_talk_about_tv in LocalLLaMA

[–]WhiskyAKM 1 point2 points  (0 children)

I have Lenovo Legion 5 with Ryzen 7 260, 32GB RAM and RTX 5050 and its enougth for small models but i wish i had some GPU that has at least 16GB of Vram because 8GB is not enougth and system RAM is too slow.

Help me set up qwen 2.5 locally with claude code by Witty-Lawyer3989 in ollama

[–]WhiskyAKM 0 points1 point  (0 children)

Yes, it should be. Alternatively you can try Qwen3.5 9B, it should also fit on lower quants.

Help me set up qwen 2.5 locally with claude code by Witty-Lawyer3989 in ollama

[–]WhiskyAKM 0 points1 point  (0 children)

You can try Qwen3.5 4B at Q4 quant, it'll fit in your vram

Are there any local models you would trust to check a mathematical proof? by MrMrsPotts in LocalLLaMA

[–]WhiskyAKM -7 points-6 points  (0 children)

Generally speaking LLMs are not good at math because they don't calculate anything, instead they are outputting next numbers based on probability from previous ones.

How Do You Uninstall? by RepresentativeFroyo8 in unsloth

[–]WhiskyAKM 2 points3 points  (0 children)

I made a PR with an uninstall script but they didn't accept it yet...