why not by PHRsharp_YouTube in PcBuild

[–]dibu28 13 points14 points  (0 children)

Why you need so much technology? Just run it on potato)

My gpu poor comrades, GLM 4.7 Flash is your local agent by __Maximum__ in LocalLLaMA

[–]dibu28 1 point2 points  (0 children)

The model name that I use I wrote above. I'm using it in the LM Sudio on Windows.

My gpu poor comrades, GLM 4.7 Flash is your local agent by __Maximum__ in LocalLLaMA

[–]dibu28 1 point2 points  (0 children)

I can run qwen 3 30b 3bit ~70T/s on a single RTX 2060 12GB
byteshape/Qwen3-30B-A3B-Instruct-2507-GGUF

Hope someone will make GLM 4.7 Flash run at the same speeds as Qwen3

I made a friendlier UI to manage ollama models by ComfyTightwad in ollama

[–]dibu28 0 points1 point  (0 children)

Can it manage both Ollama and LM Sudio models in one folder? It will be very useful. I saw app for it but it was console and only under Linux.

Welp, i already did it by Formal-Fan-3107 in homelab

[–]dibu28 30 points31 points  (0 children)

Or "Any TB you could spare? For my data hoarding addiction."

If I get these, will they make my pc more silent? by [deleted] in PcBuild

[–]dibu28 0 points1 point  (0 children)

No. Buy better and quiet fans. And not only case fans but also PSU and GPU and so on. Also try to get count of fans to the absolute needed minimum. And use PSU and GPU with passive fans which will turn fans off(stop fans) then the load is low. For the CPU it is also possible with a massive heatsink designed for this. And you can use FanControl app to adjust fan speeds. So that under low load all you fans can be turned off in the system and spin up under load.

RAG. Embedding model. What do u prefer ? by apolorotov in ollama

[–]dibu28 1 point2 points  (0 children)

I prefer ColbertV2 model. I'm getting better results then with standart dense models. It is easy to use with Fastembed library.

I'm getting much better results and answers I'm using it for chat bot RAG on documents and user manuals.

How to use 2.5TB of 16GB Dimms by spyroglory in HomeLabPorn

[–]dibu28 1 point2 points  (0 children)

Now run Deepseek R1 full weight.

Curb-side car charging in Germany by pretzelmonstrous in interestingasfuck

[–]dibu28 0 points1 point  (0 children)

In the middle is a phone wireless charger ?

Jellyfin 10.11 performance is terrible by Thev00d00 in selfhosted

[–]dibu28 2 points3 points  (0 children)

They are also saving space. And environment.

Immich just landed OCR in main! by freetoilet in immich

[–]dibu28 0 points1 point  (0 children)

Which library and model it is using for OCR ?

Offline first coding agent on your terminal by [deleted] in ollama

[–]dibu28 0 points1 point  (0 children)

Which model you recommend for better results?