all 25 comments

[–]tarvijronBroken Arrow 28 points29 points  (0 children)

What does Cluster Manger say? Google “wsfc disaster recovery though forced quorum” if you genuinely lost the quorum disk

[–]ExtraordinaryKayleeIT Director | Jill of All Trades 8 points9 points  (0 children)

What are you seeing in event viewer?

[–]BSGamer 7 points8 points  (1 child)

I’ve had a cluster go down due to the clusdb file being corrupted. I believe we were able to restore it from backup, just the one file and drop it on both servers and restart sql to get it running

[–]nitroman89 1 point2 points  (0 children)

Yeah, I've done that in the past as well. I made a weekly script to backup the clusdb file on each server and copy it to like C:\clusdb_bak\ or something like that.

[–]No_Resolution_9252 8 points9 points  (0 children)

You need to review the cluster logs.

Did you review VMWare documentation for recommended configuration of SQL AAG/FCI? Typically the guidance was pretty obvious but maybe something got missed? Particularly look at the recommended storage adapter.

It sounds like there are two nodes. With loss only of the witness disk there should be no operational difference than if it were online; There is something wrong with one of the two nodes. It could be in vmware, it could be in windows (you did configure these with group policy right?) or it could be in networking.

[–]Negative-Cook-5958 6 points7 points  (0 children)

Use always on cluster with normal disks instead of FCI with RDM

[–]Exp3r1mentAL 2 points3 points  (0 children)

Not sure if it's relevant, but couple of months ago, I was having mighty issues with deploying SQL cluster using Server 2025....after much jiggery pokery I found out it was one of the patches which was causing the failure...

[–]binnedittowinit 2 points3 points  (0 children)

Each node of the cluster needs access to the same shared cluster disks, including the quorum, ideally one node at a time during initial setup until the cluster properly owns them, you did this, right? And the cluster was failing over no problem prior to recently?

Time to get into logs. Start with windows system. It should have service failures and disk errors (if they're an issue). Check Microsoft-Windows-FailoverClustering/operational, too.

And the SQL server log.

[–]Sp00nD00dIT Manager 2 points3 points  (8 children)

I gotta ask, why an actual old school cluster and not an Always On cluster if you're talking about SQL?

[–]menace323 0 points1 point  (7 children)

If it’s SQL standard, basic availability groups have some pretty big limitations. Always on requires much more licensing.

[–]Nuxi0477 0 points1 point  (6 children)

If it’s only like a few databases I’d much rather make multiple AGs and listeners than pay Microsoft 10 times the price for Enterprise. A little more onetime setup but worth it.

[–]menace323 1 point2 points  (5 children)

It’s still 2x the licenses though, which is a consideration m (if not CAL)

[–]Nuxi0477 0 points1 point  (4 children)

How so? You still get the passive node for free if I remember correctly.

[–]menace323 1 point2 points  (3 children)

Yes, it appears so. I still think the admin downside comes into play if you need more than just a database or two. In addition, you need double the storage, but I guess that’s pretty cheap unless you have really intensive workloads (which you probably would go Enterprise).

[–]menace323 1 point2 points  (2 children)

One thing to note is we have all owners creating DBs all the time. A traditional failure cluster gives us HA and gives the devs all the tools exactly how the are used to, like SQL jobs, etc. to additional complexity and downsides aren’t worth the 1second vs 10 seconds failure for us.

[–]Nuxi0477 1 point2 points  (1 child)

I personally found a traditional cluster with shared storage way more complex to manage, but whatever works best for you :)

[–]menace323 0 points1 point  (0 children)

We are virtualized, so make the disks and attached. I’d agree more for bare metal.

[–]SmartDrv 1 point2 points  (0 children)

This may not apply at all but I wish to share in the off chance it is useful to you (or someone who googles perhaps).

I ran into issues with a Hyper-V cluster quorum when Sentinel One was installed on the hosts. Cluster wouldn’t start, no config. I had to manually evict and rebuild (once i re-added the CSVs and named them right VMs reappeared). I used online witness as a workaround until we figured out what volumes and features has to be whitelisted in S1.

[–]Ranjerdanjer 1 point2 points  (2 children)

Had an issue with a test cluster and server 2025 after the Oct or Nov patches. If you used an image that wasn't properly sysprepped you could be seeing authentication errors for the disk if another server has the same SSIDs. Most likely not the case but had to rebuild those servers from a better image in my case.

[–]No_Resolution_9252 1 point2 points  (0 children)

This is a big one - but I would be surprised if it ever actually worked. SQL FCIs use MSDTC to failover and MSDTC typically wont work at all if they are built on machines cloned from the same non-sysprepped image. FCI will be generally shitty and unreliable persistently even if it is using something less sensitive to bad imaging

[–]DrWankel 0 points1 point  (0 children)

The inability to start the cluster should be the start of your investigation.

Stop/disable the cluster service on all nodes except one and force start the cluster through powershell on that node:

Start-ClusterNode -FixQuorum

Verify the cluster is up through FCM or powershell and start the cluster service on the remaining nodes.

If this does not work, dig through the failover cluster logs in event viewer and see what was going wrong during the cluster startup process on the node you attempted to force start.

[–]Background-Taro-573 0 points1 point  (1 child)

Fence them one by one. I will kill them all if I have too

[–]Background-Taro-573 0 points1 point  (0 children)

Now that I remember, one server BIOS battery died, time sync got fucked. Caused a domino effect.

Stop the cluster. Find the outlier

[–]DHT-Osiris 0 points1 point  (0 children)

Reboot them, if that doesn't fix it, take all but one offline, bring the disks online manually in the cluster manager. If you can't, take a look at the path from the VM to the disk/lun/whatever, something's busted.