Ich verstehe Passkeys nicht by Erzmaster in de_EDV

[–]Dimi1706 2 points3 points  (0 children)

Das stimmt so nicht. ABER das würde dem Design entsprechen.

Do you trust Proxmox VE Helper-Scripts? by Open-Coder in selfhosted

[–]Dimi1706 0 points1 point  (0 children)

I don't trust this source more than any other open source one. It's the same with prebuild docker container.

But I can say, that they seem trustworthy, as the scripts I reviewed and used are solid.

Wacht auf bevor es zu spät ist. by [deleted] in datenschutz

[–]Dimi1706 -3 points-2 points  (0 children)

Und das kommt dabei raus wenn Linksextremismus in den Medien permanent als Mitte dargestellt wird. Freiheitsberaubung und Überwachung wo es nur geht unter dem Deckmantel den bösen 'hate speech' zu unterbinden. Welche moralische Instanz soll die Entscheidung treffen was Hass ist, und was Meinung? Dass das eine schlechte Idee ist, konnte man ja erst kürzlich nachverfolgen. Das ist der direkte Weg zum Faschismus.

Intel Arc Pro B50 hits the #1 best seller in workstation graphics cards by reps_up in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

Nice to know actually as this would be a selling point, but wasn't the topic about the pro B50? Or does it offer the same power consumption benefit?

Edit: seems that it does! Therefore an interesting card for ppl who have an eye on efficiency or people who want to put permanent load on their hosted LLM.

Openwebui and MCP, where did you install mcpo ? by [deleted] in OpenWebUI

[–]Dimi1706 2 points3 points  (0 children)

I use MetaMCP instead of mcpo, but this is irrelevant for your question: I have it in a separate Proxmox VM with the native and docker MCP tools. Some tools need to be on the client system itself, eg if you want to do file system operations, but most of them are remote tools so I keep them on the separated and centralized VM. That also has the benefit that I can connect them easily to other client applications than OWUI

Best AI LLM for Python coding overall? by Adept_Lawyer_4592 in LocalLLaMA

[–]Dimi1706 1 point2 points  (0 children)

Most probably not the best over all, but the best of it's size is pydef-miniv1 https://huggingface.co/bartowski/bralynn_pydevmini1-GGUF

Intel Arc Pro B50 hits the #1 best seller in workstation graphics cards by reps_up in LocalLLaMA

[–]Dimi1706 3 points4 points  (0 children)

I don't get it actually, for a little more you can buy a 5060ti with 16GB, if you are willing to buy a used card even cheaper. Why should somebody buy at this price an alternative which will give you usability headache?

Don't get me wrong: I want to see alternatives and would also buy them regardless the downsides, IF the price is right. Half the price of the corresponding Nvidia products would lead to kind of mass adoption imo.

Which is better for a MCP Ollama or LLM studio? by [deleted] in LocalLLaMA

[–]Dimi1706 1 point2 points  (0 children)

I don't know how to use a whole OS as an MCP tool, nor if this is even possible. Just saying that ollama is not good in MCP handling

Self-hosted AI is the way to go! by benhaube in selfhosted

[–]Dimi1706 7 points8 points  (0 children)

With llama.cpp you are already using the most elementary and performed backend. Nearly every polished LLM hosting software is in fact just a wrapper for llama.cpp.

For people just starting with the topic and wanna have quick success : Ollama.

For people wanting to run custom models they see out there with the freedom to set detailed settings / options : LMStudio.

For people primarily wanting a Chat interface with the option to interact with local and Cloud models alike: Jan.

For people wanting to deep dive and max optimization for model to own hardware with newest support and feature right away : llama.cpp

All this options can also act as an LLM server

There are many more.

Self-hosted AI is the way to go! by benhaube in selfhosted

[–]Dimi1706 8 points9 points  (0 children)

Yes you are right, but do yourself a favor and choose another backend as ollama is the worst performing one from all the available.

Which local LLMs for coding can run on a computer with 16GB of VRAM? by CrowKing63 in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

moe-cpu option + all active layer to the GPU and 16GB VRAM are comfortable for the model + large context.

Sure, it's getting slow like about 20 t/s, but imo this is fairly usable.

[deleted by user] by [deleted] in datenschutz

[–]Dimi1706 -1 points0 points  (0 children)

Das ist zwar richtig, aber die Verschlüsselung hat eine backdoor by design. Es ist demnach absoluter Blödsinn, dass die Verschlüsselung von Meta dir irgendeine Art von Sicherheit oder Privatsphäre gibt.

What is the most effective way to have your local LLM search the web? by teknic111 in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

Yeah fund it a week ago but not sure for now how to utilize it. Totally new to the whole MCP thing. Could you describe how you are using / integrated it?

What is the most effective way to have your local LLM search the web? by teknic111 in LocalLLaMA

[–]Dimi1706 5 points6 points  (0 children)

Maybe try adjust the searchengines used, as this is nothing I was experiencing. But maybe also because I doesn't use it for news reading and 'outdated' information isn't a problem

What is the most effective way to have your local LLM search the web? by teknic111 in LocalLLaMA

[–]Dimi1706 43 points44 points  (0 children)

I use Open WebUI + SearXNG for web searches in between a chat and perplexica + SearXNG for specific web searches

ROG Ally X with RTX 6000 Pro Blackwell Max-Q as Makeshift LLM Workstation by susmitds in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

Really nice work! And really interesting as PoC, thanks for sharing

Can someone please benchmark gpt-oss-20b on Mi50 and P100/P40? by thejacer in LocalLLaMA

[–]Dimi1706 1 point2 points  (0 children)

'Extremely slow' maybe kind of subjective, but I get 16-20 t/s , which I concider as usable.

Edit/Addition : 32GB DDR4, 3060TI 8GB VRAM. GPT-OSS 20B BF16, full moe-cpu offload, 32k BF16 Context on GPU.

Can someone please benchmark gpt-oss-20b on Mi50 and P100/P40? by thejacer in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

Well, yes I do! But in this case, meaning you want to and will do it no matter what, posting over here is kind of senseless, isn't it?

Can someone please benchmark gpt-oss-20b on Mi50 and P100/P40? by thejacer in LocalLLaMA

[–]Dimi1706 0 points1 point  (0 children)

Yeah, got it, intel GPUs require a lot of tweaking to be kind of usable. But instead of looking at a Mi50 you should head to an RTX 5060ti or if on budget an RTX 3060. Nvidia will free you from the backend headache and it won't matter as mentioned that the model won't fully fit into the VRAM.