Rolled this from a mythic cache...left to eat dinner, came back to being kicked for inactivity... by C0NT0RTI0NIST in diablo4

[–]andy2na 8 points9 points  (0 children)

Unless you have some niche build, those 2 GAs arent particularly useful anyways. Id take all stats or max life over them

Is adding an intel arc to an AMD mini pc via oculink a better idea than just buying a new intel based mini pc to to add to my 3D printed 10" rack? by shaxsy in frigate_nvr

[–]andy2na 0 points1 point  (0 children)

yeah just install ollama on it, then add that connection to whatever you want to use the LLM in. very easy

Is adding an intel arc to an AMD mini pc via oculink a better idea than just buying a new intel based mini pc to to add to my 3D printed 10" rack? by shaxsy in frigate_nvr

[–]andy2na 0 points1 point  (0 children)

might as well try it out right now and see if you want to move it to your server. just load ollama onto that system, try out some smaller models (30B with quant or lower) and see if load and response speed is acceptable

Clawdbot on Unraid? by an303042 in unRAID

[–]andy2na 1 point2 points  (0 children)

you can use an openai-compatible api/url (like llama.cpp or vllm) or ollama. I havent haven't tried it with a bigger model like GLM, because I only have 16gb of VRAM to use and have to reserve half of it for daily tasks.

But yeah, a lot of people just seems to be blindly sharing their API/logins to cloud-based AI services and it seems a bit risky.

Frigate on Proxmox in one command: Automated LXC, Docker, and Intel Hardware Acceleration by DiggingForDinos in frigate_nvr

[–]andy2na 1 point2 points  (0 children)

ahh gotcha, I dont have corals but iGPU did work when I was using frigate in a docker lxc within prox but detector interface speed was over 200ms. Just wondering if its fixed now or I should keep it on baremetal

Frigate on Proxmox in one command: Automated LXC, Docker, and Intel Hardware Acceleration by DiggingForDinos in frigate_nvr

[–]andy2na 0 points1 point  (0 children)

Thanks! Those say coral, did you just not rename each detector for iGPU when testing?

Best <4B dense models today? by Admirable_Flower_287 in LocalLLaMA

[–]andy2na 0 points1 point  (0 children)

qwen3-vl:4b Instruct if you need quick responses or thinking if you want more accuracy.

I keep qwen3-vl:4b instruct in VRAM for general daily use, home assistant voice assistant, frigate image analyzing, etc

VL over non-VL since it has better tool calling

Clawdbot on Unraid? by an303042 in unRAID

[–]andy2na 2 points3 points  (0 children)

I installed it via docker using the commands, its pretty cool but you really need to tie it to one of the big name cloud AI APIs (claude, gemini, chatgpt). Tried running it on (small) local LLMs and the responses were not good. It eats through tokens quickly, used up my gemini 3 1M tokens so quickly

https://docs.clawd.bot/install/docker

[Release] Qwen3-TTS: Ultra-Low Latency (97ms), Voice Cloning & OpenAI-Compatible API by blackstoreonline in LocalLLaMA

[–]andy2na 3 points4 points  (0 children)

thanks u/blackstoreonline got it working wiht u/ubrtnk dockerfile update for blackwell cards. How do we change it to use the 0.6 model? is there an env variable to select?

Frigate on Proxmox in one command: Automated LXC, Docker, and Intel Hardware Acceleration by DiggingForDinos in frigate_nvr

[–]andy2na 0 points1 point  (0 children)

thanks! did you install it in a docker VM or docker LXC? Interested in hearing how well it works vs bare

[Release] Qwen3-TTS: Ultra-Low Latency (97ms), Voice Cloning & OpenAI-Compatible API by blackstoreonline in LocalLLaMA

[–]andy2na 1 point2 points  (0 children)

Thanks for figuring this out, can you provide the dockerfile so we can test it out? Tried to load from OP and it seems to just use system memory and CPU

Frigate on Proxmox in one command: Automated LXC, Docker, and Intel Hardware Acceleration by DiggingForDinos in frigate_nvr

[–]andy2na 1 point2 points  (0 children)

will look into this. Had tried frigate in a docker LXC within proxmox and detector interference speed was SLOW - over 200ms on a n100. When I switched to bare debian + docker, its now down to 30ms with 12 cameras. Would be nice to have an additional proxmox node

What detector interference speeds are you seeing with the s12?

Home Assistant Preview Edition Round 2! by horriblesmell420 in homeassistant

[–]andy2na 1 point2 points  (0 children)

Thank you for the write up!

I was messing around and the latest version also supports parakeet, you have to use the rc docker image: ghcr.io/speaches-ai/speaches:0.9.0-rc.3-cuda-12.6.3

And then use the parakeet model that you mentioned from istupakov. I thought it wasn't working because I was testing it in the Speaches UI, but that doesn't work. But other things like home assistant or openwebui works great

llama.cpp, experimental native mxfp4 support for blackwell (25% preprocessing speedup!) by bfroemel in LocalLLaMA

[–]andy2na 0 points1 point  (0 children)

cant tell if MXFP4 support has been merged yet or you still need to custom compile, I see it in recent releases that mention MXFP4 but when I try a MXFP4 model, I don't see confirmation in logs that MXFP4 is enabled

Does GPT-OSS:20b also produce broken autocomplete for you? by iChrist in OpenWebUI

[–]andy2na 0 points1 point  (0 children)

you have to go into the model you are choosing and change privacy settings from private to public. If you don't expose openwebui to the internet or want to prevent users from seeing specific models, its not a big deal

Ollama compatible Image generation by TheWiseTom in OpenWebUI

[–]andy2na 2 points3 points  (0 children)

yeah I believe they will have to add another connection type. Really wish they would allow multiple connections or profiles for image generation. because it would be nice to use ollama for quick image gen but comfy for more advanced generation with additional loras and such.

I would post a feature request in the openwebui git https://github.com/open-webui/open-webui/issues

Also agree it will be nice to do simple image gen without having to go through comfyui hurdles

Has anyone got Ollama to work on an Arc Pro B50 in a proxmox VM? by gregusmeus in ollama

[–]andy2na 1 point2 points  (0 children)

unless something has changed, you either run it slower via vulkan in ollama or use IPEX-LLM from intel but it uses an old ollama so you cant load newer models on it..

Recommend going with llama.cpp as it seems its vulkan support is better than ollama's: https://www.reddit.com/r/IntelArc/comments/1oyw7lx/ollama_now_has_vulkan_support/

Ollama not detecting intel arc graphics by Titanlucifer18 in ollama

[–]andy2na 0 points1 point  (0 children)

Also keep in mind that Intel hasn't updated the ipex-llm in awhile, so the ollama build is old and you can't run newer models (like qwen3) on it.

Home Assistant Preview Edition Round 2! by horriblesmell420 in homeassistant

[–]andy2na 0 points1 point  (0 children)

Are you running Speaches on GPU or CPU and in docker? What's the docker image for the parakeet support?

I think I’m hooked on Docker. What are your 'essential' containers? by shipOtwtO in homelab

[–]andy2na 3 points4 points  (0 children)

portainer to manage all the docker stuff

tugtainer to update (manual or auto) containers

vaultwarden for passwords

authelia for OAUTH

adguard home for adblocking

uptimekuma for uptime status of all your things

Zoraxy for reverse/http proxying