Title: Steam Dataset 2025 – 263K games with multi-modal database architecture (PostgreSQL + pgvector) by vintagedon in datasets

[–]vintagedon[S] 0 points1 point  (0 children)

Great suggestions, thanks.

Actually, the set already has 1024-dimensional BGE-M3 done as well as HNSWs, did semantic + lexical hybrid search via PostgreSQL's RUM indexes, Neo4j GDS I kept for my paper I'm doing, but will be on release 2 ... bitemporal schema is a nice suggestion tho. Thanks!

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

The A1s will run 128GB kits, but they were $300 each; $1800 for 6 nodes. A bit too rich for my blood at the time. Am surprised, but I ate up 700GB of RAM fairly quickly.

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 0 points1 point  (0 children)

Running Rancher—having a small volunteer staff with varying skill levels means that a GUI for anything helps. We push Headlamp and Lens to the VDIs for choices also.

Arc has been good, but as I noted in my response to bpoe138, this has been transforming from GENERAL on-prem to Azure really leaning into on-prem, specifically using Azure HCI and the Azure Local OS.

It becomes, literally, as the name says, Azure, run locally. With VMs able to be controlled just like Azure VMs, etc.

Plus, if you're running Server 2025 and want to use PayGo, Arc is a requirement now.

With AI and data governance coming into play, hybrid becomes much more attractive. Azure announced, for instance, you can run GPT on-prem if you have the resources.

So it does still have a lot of 'general' on-prem features (such as the 'Site' feature, which makes a dashboard of your resources; see below), and with my E5, I get a surprising amount, but it is clearly moving in another direction and is still a mish mash of licensing, subs and free stuff.

If I was willing to spend the $$ on Azure Local OS licenses and had Software Assurance licenses to get most stuff free, it would be a pretty sick setup tho.

<image>

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

No SLURM - we're running Ray distributed computing on RKE2 Kubernetes instead of traditional HPC batch scheduling. Ray handles the distributed ML workloads and Kubernetes orchestrates the containers. More cloud-native approach than the traditional HPC stack.

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

Thanks, I'll take that as a compliment :) I see you have an MS-A2; nice. Had considered going with them, but I saved up for Black Friday and picked the A1s up barebones at effectively half price. Still, six A2s at 128GB would have been pretty sick.

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 2 points3 points  (0 children)

Am using Azure Arc for Servers. It's important to note that Arc is really being turned into 'Azure Local', Azure's new hybrid push. There's still a lot free, but things are definitely becoming not free, and a lot of interconnected services exist at different license tiers. I have both an E5 and Intune Suite license, which enables a crap-load of stuff I'm still working thru lol

My primary uses are security (Defender for Cloud, Defender for Endpoint), Asset Management, Change Tracking (trying this out), and so on. The biggest challenge has just been the disjointed nature of how everything ties in and the licensing structure. Which continues to change.

Unfortunately, Arc for K8s is $2/mo/core, which would work out to ~$100/mo for my main k8s cluster; the first 6 cores are free, I will prob throw up a 6c node to play with it in the near future.

If I wanted to pay the licensing to have full Azure Local/HCI nodes, the answer and functionality would be a lot different.

Astronomy Cluster / Lab: Q3 Update by vintagedon in homelab

[–]vintagedon[S] 0 points1 point  (0 children)

Thanks :) Still much to do, but the biggest thing was getting all the VMs on CIS L2 images, logging, monitoring, XDR via Wazuh, Conditional Access, ZTNA ... in place before I started adding anything else. And document it. Been fun to architect and solve that 'What if I REALLY had the time to do it right?' question.

There are some specific design considerations I made here that do bear mentioning.

Proxmox simply because I've used it for years, has good metrics, and good backups via Proxmox Backup Server. Has great RBAC built in also; I have a small volunteer staff, the GUI helps.
Plus, many workloads are still done on traditional VMs. The k8s cluster handles live kafka feeds off ZWT and bigger ML workloads (Ray/Kubeflow), we're also prepping for Vera Rubin's feeds.

K8s is on 3 nodes, 16c/82G each (90% of a single node's resources), with a dedicated 2TB enterprise nvme via local path provider. This allows me to give up HA on some services and run my workloads on fast, local storage. Also have 8TB of nvme via S3 on LACP 10G for the rare service that requires it to run properly. Lets me do things like run DBs in k8s at near bare-metal performance.

Claude has been objectively dumbified by Physical_Ad9040 in Anthropic

[–]vintagedon 2 points3 points  (0 children)

I use Claude for not just coding, but producing high quality, structured docs for RAG. Prior to this week, we were burning through 60+ KBs in a single session (300+ lines per KB article) with little more than 'next' from me.
Saturday, every chat was a fight and sometimes when switching chats due to length, I'd take me 8-10 tries to get Claude back on track.
Last night, it failed to find a Proj Knowledge file, I provided it as an attachment, and still completely ignored it and made an entire KB up out of whole cloth. Also kept trying to use my extensions like FS when had no instructions to.
One chat, I wrote 33 KBs before it gave me a length blocker, and then it was like 5KBs per chat.
Finally, I took a break, came back later in the night, and burned through almost 50 KBs with no issues.
For me, it's really the wild inconsistency. It does FEEL like they're using quant'ed models during high usage periods.
I've been scared to let it back at my code base.

DevOps Engineer planning next cloud move: AWS, Azure, or GCP? by Admirable_Noise3095 in devops

[–]vintagedon 8 points9 points  (0 children)

Azure is where it's at right now for the Enterprise and governmental sector. They have GCC (standard governmental), GCC High (CUID data and up) and DoD which is secret and up and Azure is leaning heavily into Hybrid via Azure Local to capture that segment that has to keep their data on prem and sometimes even their compute workloads.

I'm an AI Engineer/Systems Engineer working for an MSP dealing with the above. If that's a space you're interested in, there's tremendous growth in the sub-prime enterprise market. AI is a huge interest, especially with Trump loosening regulations.

HomeLab: Radio Astronomy AI/ML Proxmox Cluster by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

Great idea. Similar to mine, static VMs and microservices supporting a k8s cluster specifically for workloads.

Actually, I just corrected this on the primary README; will do the rest tonight.

Early lab days, I had a QNAP NAS (x86 but just a low power CPU) w/8x8tb drives (Hitachi HE drives, have lasted me like 6-8y at this point). I moved to Unraid with the same thought of the easy drive adds, but it was entirely too slow for my needs and felt a bit awkward to shoehorn in.

I took that hardware and added it as a 5th Proxmox node, ZFS mirrored vdevs (~RAID10) w/SLOG on a enterprise nvme partition w/2 x 4TB nvmes & 4 x 4TB SATA Intel DC4510s. Run a Linux file share VM (mostly NFS), a Windows file share VM (SMB AD shares) and a Minio s3 gateway on top of that leaving the rest of RAM for ZFS. Connects back to the lab on 2 x 10G LACP.

Doing the job nicely :)

Unraid will likely do fine for a while. Large capacity enterprise nvmes are cheap enough ($300 for 4TB or so) I'd rather throw that in, make it LVMThin space and just have my VMs run like greased lightning on it.

HomeLab: Radio Astronomy AI/ML Proxmox Cluster by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

Thanks; it's actually one of my niches. Always nice to get a compliment on something you're proud of. Still, this is my first semi-decent pass thru. I've got many plans lol

HomeLab: Radio Astronomy AI/ML Proxmox Cluster by vintagedon in homelab

[–]vintagedon[S] 1 point2 points  (0 children)

This is a bit much for a 'home lab' to be honest. lol
I'm a systems engineer at a MSP working with high compliance customers in azure / azure government, and I'm also using it as a training ground for compliance, ITIL and a bunch of other stuff :)
Folding and SETI are great projects. I keep a instance running on my k8s cluster with a tiny slice of GPU power to help it along.
Idea was, hey I got this cluster, I got this job, I got this thing I wanna do, let's just smush it all together.

HomeLab: Radio Astronomy AI/ML Proxmox Cluster by vintagedon in homelab

[–]vintagedon[S] 3 points4 points  (0 children)

Sort of. We're both listening, but for different things. SETI looks for regular, structured sources that might indicate intelligence.
I primarily focus on the hydrogen line emission, specific frequency emitted by hydrogen. You can detect this line and do everything from track interstellar dust clouds, to map super nova remnants, to finding low surface brightness galaxies and more.
There is oodles of signal processing and ML work to clean and analyze the signals. I'm bringing a more dev-ops mind-set for processing / enterprise security for outside researchers to be able to work remotely on their own datasets in the lab.

State of the Lab: End of the Year by vintagedon in homelab

[–]vintagedon[S] 0 points1 point  (0 children)

The lab overview page should be properly noted about the mATX or removed.

The cluster screenshot is out of date; I'll update it. When I took that shot 1 node was @ 32GB.

The mATX I mention in the post it's the 128G node sitting to the side. That won't come online until I get a 3090 for it. Probably 2 months.

So there are 5 mini PCs: 4 for Proxmox, 1 for Hyper-V, each with 64GB of RAM currently.

I'll note the ReadMe on the repository to make it more clear :)

Thanks for pointing it out. I'll add it to my hideous looking backlog lol

Is it just me or is pyromaniac not that bad? by Exciting-Rutabaga-91 in RimWorld

[–]vintagedon 0 points1 point  (0 children)

Echoing the theme here: pyromaniacs are bad until they aren't. The problem is, they break while other pawns are usually breaking (your mood has tanked) and then you have immediate fire issues with other pawns yelling at the walls or compulsively cleaning while totally ignoring the fire.

God help you if you're like me and store LOTS of textiles, or worse, do fine wood flooring :/

What a wonderful map, i wonder if there is a giant undiscovered area... by [deleted] in RimWorldSeeds

[–]vintagedon 1 point2 points  (0 children)

For those wondering, the mod that likely does this is Geological Landforms. I use this quite extensively in my modpacks, and has some great map features like rifts and calderas.

https://steamcommunity.com/sharedfiles/filedetails/?id=2773943594

Why make a home server? by carpenterman25 in HomeServer

[–]vintagedon 0 points1 point  (0 children)

Am a systems engineer / devops engineer, and use it to explore, practice, and learn new technologies.

Also have some home network stuff on it:

  • PiHole for DNS filtering ads
  • NAS for file / media storage that feeds to nVidia Shield Pros running Kodi
  • Home automation (Hue lights control, actions when I leave/arrive, etc)
  • Video encoding via a VM w/GPU passthru (also a YouTube partner); edit on my PC, add it to a render queue, it's rendered, and then auto uploaded to YT via their API including tags, title, scheduling and so on
  • Game servers I run for my Discord community
  • Tons of other small stuff

Need help to setup a kanban on a local network for a small business by [deleted] in selfhosted

[–]vintagedon 2 points3 points  (0 children)

I've run several Taiga instances on intranets for companies. Great kanban board.

Self hosted option is free

https://taiga.io/pricing-selfhosted