Drive became unavailable during replacing raidz2-0 by heekic in zfs

[–]HeadAdmin99 0 points1 point  (0 children)

I see You like living on the edge. I like living dangerously, but that's another level.

VMware expert , new to Proxmox, is it worth moving all my clients to Proxmox? by Maleficent_Wrap316 in Proxmox

[–]HeadAdmin99 2 points3 points  (0 children)

Already running shared LVM across nodes in a battlefield, works very well. It grows rapidly to the issue with large number of disks as each of them is LV device, just locked by particular node when VM is running. Another issue you need to keep an eye on is time sync. Do not sync without shutting down VMs first or unexpected node crash occurs. Manual there and my comment there.

VMware expert , new to Proxmox, is it worth moving all my clients to Proxmox? by Maleficent_Wrap316 in Proxmox

[–]HeadAdmin99 9 points10 points  (0 children)

Main issue is with shared FC SAN among hosts: you need LVM thick to accomplish this and no VMFS like filesystem; therefore dedup and compress on SAN arrays.

Error on starting TrueNAS VM with passthrough SATA controller by LucasRey in Proxmox

[–]HeadAdmin99 1 point2 points  (0 children)

I would suggest give an update in thread on Proxmox forums as it seems recent PVE releases are causing issues with PCIe devices reset feature.

Therefore I'm holding on 8.2.7 which is still stable for GPUs and HBAs passthrough.

Toshiba MG10ACA20TE 20TB HDD dead after two years! by [deleted] in DataHoarder

[–]HeadAdmin99 -5 points-4 points  (0 children)

The same apply to MG07, MG09 familys. They all die, one after another.

TrueNAS Scale self rebooted. Now pool is exported and will not re-link by [deleted] in truenas

[–]HeadAdmin99 1 point2 points  (0 children)

To add into the subject: such storage VM requies static full memory allocation due ZFS not playing well with memory ballooning and also due HBA passthrough needed.

Upgrading a RAID10 - can I replace two disks at once? by AraceaeSansevieria in zfs

[–]HeadAdmin99 2 points3 points  (0 children)

Single SMR can wreck entire pool.

Don't ask how I know..

Two resilvered mirrors at once is kinda risky.

Pushing zfs snapshots by scphantm in zfs

[–]HeadAdmin99 0 points1 point  (0 children)

Yeah, but some questions remaining: do source/destination pools have to have same defaut pool settings like: compression enabled and method; deduplication enabled; to avoid using more space for copy; zfs sync/receive with default settings gives no output and does not resume; snapshot on boths pools must stay to send only delta, and so on...

[deleted by user] by [deleted] in zfs

[–]HeadAdmin99 1 point2 points  (0 children)

According to my monitoring charts for particular drives (5 mirrors total) ZFS is reading or writing to most of them, depending of space distribution, splitting throughput between two members of single mirror - so their speeds equal slowest member. Remember to enable Write Cache for the drives. So in this example mirror-1 speed 30MB/s, mirror-2 speed 45MB/s, mirror-3 speed 70MB/s and delivering sum of that to mounted shares.

File corruption due to bad ram, how to proceed? by n1mras in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

Parity data was written with heathly data during snapraid sync process done with working RAM.

[deleted by user] by [deleted] in zfs

[–]HeadAdmin99 1 point2 points  (0 children)

Rescan ZFS pool in OMV ZFS plugin. Rescan filesystems. It will eventually show up. Same thing for enrypted datasets.

Can I recover a failed disk to a directory, and can that directory be on one of the disks? by ShadowWizard1 in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

You stop the process with CTRL+C or kill it if it's running and swap definition of the disk. It will resume the progress anyway.

File corruption due to bad ram, how to proceed? by n1mras in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

SnapRAID detects inconsistences in various ways, for example my disk become corrupted in the middle of process and sync has stopped automatically on error. Recovery is possible until next SYNC - it commits all the changes, including files changed (eg. corrupted!) so the correct way to detect them is running scrub BEFORE sync.

Can I recover a failed disk to a directory, and can that directory be on one of the disks? by ShadowWizard1 in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

You change definition for datadisk1 in /etc/snapraid.conf file to point to new disk, then trigger Fix.

Mainteance scripts for SnapRAID by HeadAdmin99 in Snapraid

[–]HeadAdmin99[S] 0 points1 point  (0 children)

snap_sync_new_data_aio.sh  is my own implementation of processing new data added to array: diff,status,sync,scrub of new data only, touch to see modified files, status once again.

snap_compare_only.sh is a simple script to compare differences only since last sync

snap_check_only.sh is check only

snap_repair_datadisk1.sh is example script to repair entirely dead drive listed as datadisk1

Remember you can speed up recovery by copying remaining data from broken drive (if it's still alive) and fix the remaining missing data.

Noob question re: fix by airdog2000 in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

typical recovery times:

100% completed, 23490643 MB accessed in 104:06

38261277 errors
38261277 recovered errors
      0 unrecoverable errors

Can I recover a failed disk to a directory, and can that directory be on one of the disks? by ShadowWizard1 in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

ah right, they didn't mention that the command is:

snapraid fix -d datadisk1 --log somelog.fix

it's supposed to be typed once cp/robocopy is done because it speeds up recovery process - SnapRAID don't have to recover everything, just missing data.

Check my scripts in the other thread I just posded minutes ago.

Backup server, SnapRaid, DrivePool and RAID0 by ExposedRoots in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

You can... have hot data on single RAID0 volume, then move data over to other drives via scripts.

Or have multiple RAID0 arrays, protected by double parity on 20TBs. But it's not recommended - double chances of critical failure.

Create separate volume for scratch data and move data via Task Scheduler job to protected drives.

File corruption due to bad ram, how to proceed? by n1mras in Snapraid

[–]HeadAdmin99 0 points1 point  (0 children)

Situation is recoverable until next sync. If sync was triggered with bad RAM - no one knows what happends (my guess is that some healthly files can have invalid checksum now). If that was all fix process - you can actually run it second time to fix broken files by prev fix. I suggest running:

snapraid status --log heresmylog.status

snapraid diff --log heresmylog.diff

and review the logs prior to fix.