Hey everyone,
I recently built my first NAS. It was bough used with SAS hardware. I’ve finally got past all the roadblocks and problems that were in my way (I basically bricked a whole SAS drive, a hero of a lemmy user helped me fix it).
Now after filling the 15 TB of RAIDZ2 with around 100gb of data. One of the drives started waiving its white flag and wants to die on me.
I am a complete beginner with no experience with these things.
Is my drive dying and should be replaced? or can it be fixed?
This is the output of the 507 errors that TrueNAS received form it and labelled the vDev as degraded and the drive as faulted:
Output of zpool status and sudo smartctl -a /dev/sdd
As a beginner it looks like this drive is cooked, please let me know if it needs replacing so I can order a new one and replace it right away.
Thank you sooo much!
Edit: SAS not SATA drives
I would replace it. Sometimes I push my luck and for minor or unexpected errors I just clear the error and re-add the drive, but this many errors is likely a solid sign.
Not necessarily. I would shut the system down completely and check the drive connectors. If it’s on a backplane, try swapping slots, or if it’s breakout connector, swap it with another drive (and clear the zpool errors). If the errors start happening on the other drive, it’s a cable problem. If they continue on the same drive, it’s a drive problem. If they stop happening, it was a bad connection and it ought to be fine now.
That’s kind of a short output from smartctl -a, though. Shouldn’t it include the attribute data? I’d run a smart test (after doing the swap above) and see what it says.
On a raidz2, I wouldn’t be too concerned about losing a drive, but you should always be prepared to order a replacement if you value your data.
I second this. SATA cables are cheaply made and can present issues that seem to indicate drive failure.
Had this issue once, 2 drives kept not initializing during boot, rebooting a few times got them to register but showed drive errors. I thought either the drives or my SAS card was dying. Fully reseating the connectors fixed it and haven’t had an issue since.
OP is using SAS, but it’s not too far from SATA.
Good catch. I don’t usually see SAS as /dev/sd* so I assumed. Almost the same cables, though *usually better made.
zpool has very reasonable thresholds for disk failure being enough to kick it from the pool. I’ve seen pool members have a batch of bad blocks and ZFS still chugged along for a few years just avoiding those blocks before the disk finally failed.
Heed truenas here, replace the disk if you can.
Does this drive not have SMART values? Could be a loose cable for all we know (UDMA CRC errors).
My rule for older hardware, before trusting the ZFS fault reporting, I would follow the following steps.
(Note these are homelabber steps and not what I would do in the enterprise, where risk and time is a lot more expensive than replacing hardware)
-
Check the Smart data of the drive. If it reports the drive as faulty, replace it.
-
Zpool clear the error and see if it comes back. Sometimes drive errors are not cause by the drive itself
-
Reseat the drive and the cables between the motherboard and the drive. Clear errors after this step. Especially with older hardware and it having travelled from its previous owner to you, something might not be seated properly.
-
Move the drive to another drive bay, or swap it with another drive. If the errors move with the drive, the drive is faulty. If the errors move to the bay, you probably have a good drive, but a faulty drive bay/cable.
-
In addition to other advice, with only hardware, have 1 cold spare drive for every 2 years of remaining life of the hardware. It gets difficult to find similar spec drives the older they get. So if you want to use the drives in the NAS for another 4 years, get 2 spares. After that you start getting into the territory of replacing the drives anyway.
Just RMA it now. If it has SMART failures, you can provide the codes and they’ll replace it no problem.
Buy two replacements