I've got a server with four Samsung hard drives. All drives are the same model and have been bought together. The drives are SAMSUNG HE753LJ with firmware 1AA01113.
I'm getting SMART errors but I have the feeling that smartctl does not understand the value he gets from the hard drive.
Here's the result of a SMART test:
asgard:~# smartctl -H /dev/sdb smartctl version 5.38 [i686-pc-linux-gnu] Copyright (C) 2002-8 Bruce Allen Home page is http://smartmontools.sourceforge.net/ === START OF READ SMART DATA SECTION === SMART overall-health self-assessment test result: FAILED! Drive failure expected in less than 24 hours. SAVE ALL DATA. Failed Attributes: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 3 Spin_Up_Time 0x0007 001 001 011 Pre-fail Always FAILING_NOW 60340
I don't trust SMART because:
- It's been over one year that all disks are about to fail within less than 24 hours. Nothing blew up yet.
- Wikipedia says that "Spin-Up Time is the average time of spindle spin up (from zero RPM to fully operational [millisecs])." That would mean that the drives need about one minute to wake up?!
I would like to follow smartctl's advice and change these disks but I just don't trust the results I read.
What do you think about this? What would you do?
Thanks for your help.
This is a ticking bomb.
Based on both the message from SMART and the quote above, you should change disks right away.
Since the drives have been bought together and are the same model, they will probably have the same weaknesses, and probably all fail simultaneously under the same condition...
The main concept of RAID is that disks fail at different times, giving you the opportunity to swap one disk at a time, and avoid data loss.
Others have reported simultaneous failure of an entire array of identical disks in a RAID configuration, coming from the same production batch, and thus being subject to the same weakness.
I can't stress this enough: You need to start swapping your drives!
I had a spare drive that I can still boot from that fails SMART checks every boot and requires a soft reset, has for years, but it's just a dump, not a system disk! So although SMART errors can persist for a long time they should ALWAYS be heeded in production, as the risks heavily outweigh the cost, time and data integrity benefits. Google studied 100,00 disks and found:
So it's not always a robust indicator. However SMART error significantly increase the likelihood of a disk crash in the time immediately after initial detection:
So statistically your disk is probably OK, as it's well exceeded the 60 day limit.
But are you willing to continue taking the risk? I'd change the disk ASAP to avoid having to get up in the early hours.
That part is not interpreted by smartctl (assuming I understand correctly, of course) - that drive has told smartctl that is isn't happy with its current state (for whatever reason) and smartctl is just echoing that warning to you. Even if it is misinterpreting the spin-up time reading, I don't think it is doing any interpretation on the "self assessment test" reading.
I would suggest moving your data off that drive ASAP, preferably before it next power cycles in case the spin-up problem is real and might get worse.
I would change the disks right away without thinking much about it. You'd be on the safe side, disks are dirt cheap and you'll sleep better. Your time spent diagnosing the disks is probably worth more than the disks themselves.
Make sure you have the latest copy of the smart utils not just the ones included in your OS. smart utils are updated frequently and some of the errors reporting from specific drives to get resolved.
Google's study was very informative. 30% of disk with SMART errors eventually fail. That's not odds I would be will to deal with. That's a 9% chance that two disks will fail and your RAID at that point will be destroyed.
I had a similar issues with some Seagate drives a few years ago. We bought about 8 disks at the same time and they all were from the same lot. About at about 3 years, one drive went. 18 hours later another drive went, 24 hours later a 3rd drive went.
Run a DST on the disks, and replace them accordingly.