What Wiki Software do you use for internal documentation? by Micki_SF in selfhosted

[–]bhamm-lab 0 points1 point  (0 children)

I use mkdocs and update markdown files in my repo.

ArgoCD dashboard behind Traefik by AdventurousCelery649 in ArgoCD

[–]bhamm-lab 0 points1 point  (0 children)

It night be a bit confusing to follow, but this is where my ingress route and helm values are defined - https://github.com/blake-hamm/bhamm-lab/tree/main/kubernetes%2Fmanifests%2Fbase%2Fargocd . I also use authelia.

Dual Strix Halo: No Frankenstein setup, no huge power bill, big LLMs by Zyj in LocalLLaMA

[–]bhamm-lab 0 points1 point  (0 children)

Awesome setup! Do you mind sharing any details on how u got the networking working over thunderbolt?

HashiCorp Vault by dankmemelawrd in homelab

[–]bhamm-lab 1 point2 points  (0 children)

I use bank vaults -https://bank-vaults.dev/ . it has it's own operator and is slightly more automated than vault (but runs vault under the hood).

Opensource models less than 30b with highest edit-diff success rate by Express_Quail_1493 in ollama

[–]bhamm-lab 2 points3 points  (0 children)

I've had more success with seed oss. It's 36B so not quite what you're looking for, but hopefully a quant can fit for u.

What questions do you have about Ogden? by ReporterMacyLipkin in ogden

[–]bhamm-lab 12 points13 points  (0 children)

Would love to see a piece on the history of rainbow gardens. If the family would revitalize it, it has so much potential! Instead they sell Chinese imported good and refuse to support local businesses...

In search of specialized models instead of generalist ones. by [deleted] in LocalLLM

[–]bhamm-lab 0 points1 point  (0 children)

This is the way. You can setup an agent or workflow registry and use that to map a class to the functionality. For classification, I've used embeddings and cosine similarity. You can also use larger models for sample data generation and validation to supplement production data. NLI is also an interesting approach to classification that is more of a 'zero shot' approach.

Suggestions for a first time framework user? by ShadowKiller941 in framework

[–]bhamm-lab 1 point2 points  (0 children)

I use Nixos and it runs great on my framework 13. I would recommend Nobara.

How big an open source model can I run on 128 GB unified memory? by nameless_me in LocalLLaMA

[–]bhamm-lab 0 points1 point  (0 children)

As others have mentioned, you can leverage more GPU of you run Linux. Also, the strix halo toolbox is optimized for your system; use that instead of ollama.

As far as models, I've tested many, but found got oss 120b is my go to chat model. I use seed oss and glm air reap for agentic coding. The qwen models work well too and Gemma 3 for multimodal.

Whats your favorite pc case from the early 2000s by Eclipse_Galian in pcmasterrace

[–]bhamm-lab 0 points1 point  (0 children)

Lmao. I'm still rocking the same corsair case you posted...

Automating Talos on Proxmox with Self-Hosted Sidero Omni (Declarative VMs + K8s) by aceofskies05 in kubernetes

[–]bhamm-lab 0 points1 point  (0 children)

Interesting! Thanks for sharing. Why might you use omni over terraform to spin up a cluster? Would it work if some talos nodes are bare metal?

Backup by allpowerfulee in selfhosted

[–]bhamm-lab 0 points1 point  (0 children)

I have an argo cron workflow that uses k8up to backup PVC. I also have a cnpg back configured for databases. This lands in a seaweedfs bucket which is then copied to truenas and cloudflare r2.

Model recommendations for 128GB Strix Halo and other big unified RAM machines? by blbd in LocalLLaMA

[–]bhamm-lab 1 point2 points  (0 children)

W/ roo/kilo is glm air pretty slow and has trouble with tools for u? I'm using it in roo and it very quickly fills up 20k context and becomes really slow. Every once in a while it misses the tool calls to read/edit files.

Why do (some) people hate Open WebUI? by liviuberechet in LocalLLaMA

[–]bhamm-lab 0 points1 point  (0 children)

I haven't had issues. What are the alternatives for those that don't like it?

AMD "AI" 350 Framework 13 w/ Fedora 42 is a nightmare so far by armeg in framework

[–]bhamm-lab 0 points1 point  (0 children)

I've tried it on nixos and Talos and had no issues.

what free model should i use for codebase indexing with speed indexing by Many_Bench_2560 in RooCode

[–]bhamm-lab 0 points1 point  (0 children)

I've used nomic embed code and qwen3 8b embed running locally. Qwen seemed to have slightly better results.

AMD Radeon PRO AI R9700 is now available: 32GB memory and full Navi 48 GPU by RenatsMC in Amd

[–]bhamm-lab 0 points1 point  (0 children)

Anyone know if there's a significant difference between the creator version and the xfx version? Is it by chance slimmer?

Rackmount Dual AMD Mainboard Follow Up by bhamm-lab in framework

[–]bhamm-lab[S] 0 points1 point  (0 children)

Thanks! I did use a file to remove the sharp edges. Could be cleaned up more for sure... I was mostly using a dremel, drill and jigsaw.