* RAID5 up, but one drive removed, one says spare building, what now?
@ 2017-11-10 3:09 Jun-Kai Teoh
2017-11-10 11:31 ` Wols Lists
0 siblings, 1 reply; 12+ messages in thread
From: Jun-Kai Teoh @ 2017-11-10 3:09 UTC (permalink / raw)
To: Linux-RAID
Hi all,
I managed to get my RAID drive back up, content looks like it's still
there, but it's not resyncing or reshaping and my parity drive was
removed (I did it when I tried to get it back up).
So what should I do now? I'm afraid of doing anything else at this point.
/dev/md126:
Version : 1.2
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Array Size : 23441323008 (22355.39 GiB 24003.91 GB)
Used Dev Size : 3906887168 (3725.90 GiB 4000.65 GB)
Raid Devices : 8
Total Devices : 7
Persistence : Superblock is persistent
Intent Bitmap : Internal
Update Time : Thu Nov 9 18:57:18 2017
State : clean, FAILED
Active Devices : 6
Working Devices : 7
Failed Devices : 0
Spare Devices : 1
Layout : left-symmetric
Chunk Size : 512K
Delta Devices : 1, (7->8)
Name : livingrm-server:2 (local to host livingrm-server)
UUID : f7333d4f:8300969d:55148d64:93c8afc8
Events : 650582
Number Major Minor RaidDevice State
0 8 112 0 active sync /dev/sdh
1 8 48 1 active sync /dev/sdd
7 8 64 2 spare rebuilding /dev/sde
3 8 96 3 active sync /dev/sdg
4 8 32 4 active sync /dev/sdc
5 8 80 5 active sync /dev/sdf
6 8 16 6 active sync /dev/sdb
14 0 0 14 removed
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 3:09 RAID5 up, but one drive removed, one says spare building, what now? Jun-Kai Teoh
@ 2017-11-10 11:31 ` Wols Lists
2017-11-10 13:01 ` Phil Turmel
0 siblings, 1 reply; 12+ messages in thread
From: Wols Lists @ 2017-11-10 11:31 UTC (permalink / raw)
To: Jun-Kai Teoh, Linux-RAID, Phil Turmel
On 10/11/17 03:09, Jun-Kai Teoh wrote:
> Hi all,
>
> I managed to get my RAID drive back up, content looks like it's still
> there, but it's not resyncing or reshaping and my parity drive was
> removed (I did it when I tried to get it back up).
>
> So what should I do now? I'm afraid of doing anything else at this point.
>
> /dev/md126:
> Version : 1.2
> Creation Time : Thu Jun 30 07:57:36 2016
> Raid Level : raid5
> Array Size : 23441323008 (22355.39 GiB 24003.91 GB)
> Used Dev Size : 3906887168 (3725.90 GiB 4000.65 GB)
> Raid Devices : 8
> Total Devices : 7
> Persistence : Superblock is persistent
>
> Intent Bitmap : Internal
>
> Update Time : Thu Nov 9 18:57:18 2017
> State : clean, FAILED
> Active Devices : 6
> Working Devices : 7
> Failed Devices : 0
> Spare Devices : 1
>
> Layout : left-symmetric
> Chunk Size : 512K
>
> Delta Devices : 1, (7->8)
>
> Name : livingrm-server:2 (local to host livingrm-server)
> UUID : f7333d4f:8300969d:55148d64:93c8afc8
> Events : 650582
>
> Number Major Minor RaidDevice State
> 0 8 112 0 active sync /dev/sdh
> 1 8 48 1 active sync /dev/sdd
> 7 8 64 2 spare rebuilding /dev/sde
> 3 8 96 3 active sync /dev/sdg
> 4 8 32 4 active sync /dev/sdc
> 5 8 80 5 active sync /dev/sdf
> 6 8 16 6 active sync /dev/sdb
> 14 0 0 14 removed
Okay. I was hoping someone else would chime in, but I'd say this looks
well promising. You have seven drives of eight so you have no redundancy :-(
You say your data is still there - does that mean you've mounted it, and
it looks okay?
sde is rebuilding, which means the array is sorting itself out.
You need that eighth drive. If a fsck says you have no (or almost no)
filesystem corruption, and you have a known-good drive, add it in. The
array will then sort itself out.
I would NOT recommend mounting it read-write until it comes back and
says "eight drives of eight working".
Cheers,
Wol
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 11:31 ` Wols Lists
@ 2017-11-10 13:01 ` Phil Turmel
2017-11-10 13:13 ` Mark Knecht
0 siblings, 1 reply; 12+ messages in thread
From: Phil Turmel @ 2017-11-10 13:01 UTC (permalink / raw)
To: Wols Lists, Jun-Kai Teoh, Linux-RAID
Hi Jun-Kai,
On 11/10/2017 06:31 AM, Wols Lists wrote:
> On 10/11/17 03:09, Jun-Kai Teoh wrote:
>> Hi all,
>>
>> I managed to get my RAID drive back up, content looks like it's still
>> there, but it's not resyncing or reshaping and my parity drive was
>> removed (I did it when I tried to get it back up).
If you can see your content (mounted read-only, I hope), backup up
everything, in order from most critical to least critical.
>> So what should I do now? I'm afraid of doing anything else at this point.
Well, you need to provide more information. At least the mdadm -E
reports for all of the member devices. Including the "parity" device
you removed. (Parity is spread among all devices in a normal raid5
layout, so you may be suffering from a misunderstanding.)
>> /dev/md126:
>> Version : 1.2
>> Creation Time : Thu Jun 30 07:57:36 2016
>> Raid Level : raid5
>> Array Size : 23441323008 (22355.39 GiB 24003.91 GB)
>> Used Dev Size : 3906887168 (3725.90 GiB 4000.65 GB)
>> Raid Devices : 8
>> Total Devices : 7
>> Persistence : Superblock is persistent
>>
>> Intent Bitmap : Internal
>>
>> Update Time : Thu Nov 9 18:57:18 2017
>> State : clean, FAILED
>> Active Devices : 6
>> Working Devices : 7
>> Failed Devices : 0
>> Spare Devices : 1
>>
>> Layout : left-symmetric
>> Chunk Size : 512K
>>
>> Delta Devices : 1, (7->8)
^^^^^^
Somewhere in your efforts, you must have used mdadm --grow. That was
bad, and the reason for my suggestion above.
Please review your bash history and reconstruct what steps you took in
your efforts to revive your array. An extract of relevant dmesg text
from that time period may be helpful, too.
It would help if you summarized how you got into this jam in the first
place.
[trim /]
> Okay. I was hoping someone else would chime in, but I'd say this looks
> well promising. You have seven drives of eight so you have no redundancy :-(
>
> You say your data is still there - does that mean you've mounted it, and
> it looks okay?
>
> sde is rebuilding, which means the array is sorting itself out.
It's not progressing, because of the reshape.
> You need that eighth drive. If a fsck says you have no (or almost no)
> filesystem corruption, and you have a known-good drive, add it in. The
> array will then sort itself out.
No, backups are first.
> I would NOT recommend mounting it read-write until it comes back and
> says "eight drives of eight working".
Concur.
Phil
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 13:01 ` Phil Turmel
@ 2017-11-10 13:13 ` Mark Knecht
2017-11-10 17:25 ` Jun-Kai Teoh
0 siblings, 1 reply; 12+ messages in thread
From: Mark Knecht @ 2017-11-10 13:13 UTC (permalink / raw)
To: Phil Turmel; +Cc: Wols Lists, Jun-Kai Teoh, Linux-RAID
On Fri, Nov 10, 2017 at 6:01 AM, Phil Turmel <philip@turmel.org> wrote:
> Hi Jun-Kai,
>
> On 11/10/2017 06:31 AM, Wols Lists wrote:
>> On 10/11/17 03:09, Jun-Kai Teoh wrote:
>>> Hi all,
>>>
>>> I managed to get my RAID drive back up, content looks like it's still
>>> there, but it's not resyncing or reshaping and my parity drive was
>>> removed (I did it when I tried to get it back up).
>
> If you can see your content (mounted read-only, I hope), backup up
> everything, in order from most critical to least critical.
>
>>> So what should I do now? I'm afraid of doing anything else at this point.
>
> Well, you need to provide more information. At least the mdadm -E
> reports for all of the member devices. Including the "parity" device
> you removed. (Parity is spread among all devices in a normal raid5
> layout, so you may be suffering from a misunderstanding.)
>
Phil,
Please reference the thread entitled
"Raid 5 array down/missing - went through wiki steps"
All the mdadm -E info and much more was reported there a week
ago. Wols and I helped get him that far but neither of us was confident
in giving instructions to get him further.
Cheers,
Mark
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 13:13 ` Mark Knecht
@ 2017-11-10 17:25 ` Jun-Kai Teoh
2017-11-10 18:01 ` Phil Turmel
0 siblings, 1 reply; 12+ messages in thread
From: Jun-Kai Teoh @ 2017-11-10 17:25 UTC (permalink / raw)
To: Mark Knecht; +Cc: Phil Turmel, Wols Lists, Linux-RAID
I was reshaping/growing my array when it died. Computer came back up
but array wouldn't assemble.
I looked around and saw some commands on how to assemble without the
one drive that looked like it was kicked off sometime ago that I did
not realize (I used --assemble --force --verbose)
I kept getting the error that there were 6 drives, 1 rebuilding, and
not enough to get the array up. I unplugged the one drive that had the
incorrect superblock and ran "mdadm -A -R".
It ran and looked like it was resuming reshaping but it got really
slow - like 7 years slow.
I killed the process, rebooted the machine, and it just auto
reassembled the array and automounted it on boot.
Am in the process of backup right now.
And yes, the drive/array is mounted.
mdadm version 3.3, september 2013
Below is the updated mdadm examine info followed by smartctl (note
that the sd* have changed from the earlier post, likely because I
unplugged that drive):
/dev/sdb:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : 2df0b319:bdb18eee:27b318ec:da55d53d
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : d2b7f433 - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 6
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdc:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : f1a790a9:98e01257:d9ab257d:95c8f1fc
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : 1b42463e - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 4
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdd:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : 6f44eba1:29d246a4:c5e8312e:bac00a7b
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : 8ff6095a - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 1
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sde:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x47
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
Recovery Offset : 11264816 sectors
State : clean
Device UUID : 109b7a2f:0529794c:2cf95cc1:d6c0bd6b
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : 4a41aa92 - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 2
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdf:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : deb04cea:d6530966:6c70ca90:bebb143e
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : bfbc7a25 - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 5
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdg:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : 15d85573:6e78f040:8c028ef3:d1301f9d
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : e57acd25 - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 3
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdh:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x45
Array UUID : f7333d4f:8300969d:55148d64:93c8afc8
Name : livingrm-server:2 (local to host livingrm-server)
Creation Time : Thu Jun 30 07:57:36 2016
Raid Level : raid5
Raid Devices : 8
Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
Array Size : 27348210176 (26081.29 GiB 28004.57 GB)
Used Dev Size : 7813774336 (3725.90 GiB 4000.65 GB)
Data Offset : 262144 sectors
New Offset : 254976 sectors
Super Offset : 8 sectors
State : clean
Device UUID : 64d5961e:230e558c:3748b561:a7c6ab8c
Internal Bitmap : 8 sectors from superblock
Reshape pos'n : 39348736 (37.53 GiB 40.29 GB)
Delta Devices : 1 (7->8)
Update Time : Fri Nov 10 09:17:35 2017
Bad Block Log : 512 entries available at offset 72 sectors
Checksum : 39d43ba7 - correct
Events : 651498
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 0
Array State : AAAAAAA. ('A' == active, '.' == missing, 'R' == replacing)
------
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E2VKP3SV
LU WWN Device Id: 5 0014ee 2620754c8
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is: Fri Nov 10 09:24:03 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (51120) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 512) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 191 183 021 Pre-fail
Always - 7441
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 182
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 082 082 000 Old_age
Always - 13243
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 182
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 113
193 Load_Cycle_Count 0x0032 192 192 000 Old_age
Always - 26737
194 Temperature_Celsius 0x0022 118 101 000 Old_age
Always - 34
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E4JNHEC6
LU WWN Device Id: 5 0014ee 2614842c3
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 1.5 Gb/s)
Local Time is: Fri Nov 10 09:24:14 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (53940) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 539) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 189 180 021 Pre-fail
Always - 7516
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 214
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 076 076 000 Old_age
Always - 18057
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 214
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 128
193 Load_Cycle_Count 0x0032 189 189 000 Old_age
Always - 33162
194 Temperature_Celsius 0x0022 117 092 000 Old_age
Always - 35
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E2ZYE8AN
LU WWN Device Id: 5 0014ee 20bf26ce9
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 1.5 Gb/s)
Local Time is: Fri Nov 10 09:24:17 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (54960) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 550) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 186 178 021 Pre-fail
Always - 7658
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 214
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 076 076 000 Old_age
Always - 18060
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 214
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 121
193 Load_Cycle_Count 0x0032 189 189 000 Old_age
Always - 33798
194 Temperature_Celsius 0x0022 117 098 000 Old_age
Always - 35
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68N32N0
Serial Number: WD-WCC7K7PZ7R6Z
LU WWN Device Id: 5 0014ee 26410129e
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Form Factor: 3.5 inches
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-3 T13/2161-D revision 5
SATA Version is: SATA 3.1, 6.0 Gb/s (current: 1.5 Gb/s)
Local Time is: Fri Nov 10 09:24:20 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (46440) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 492) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x303d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 100 253 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 100 253 021 Pre-fail
Always - 0
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 5
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 100 253 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 100 100 000 Old_age
Always - 320
10 Spin_Retry_Count 0x0032 100 253 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 253 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 5
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 2
193 Load_Cycle_Count 0x0032 200 200 000 Old_age
Always - 61
194 Temperature_Celsius 0x0022 118 114 000 Old_age
Always - 32
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E4JNHKL9
LU WWN Device Id: 5 0014ee 2b69e186b
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 1.5 Gb/s)
Local Time is: Fri Nov 10 09:24:25 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (55260) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 552) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 1
3 Spin_Up_Time 0x0027 187 180 021 Pre-fail
Always - 7616
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 675
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 074 074 000 Old_age
Always - 19242
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 263
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 163
193 Load_Cycle_Count 0x0032 188 188 000 Old_age
Always - 36231
194 Temperature_Celsius 0x0022 116 091 000 Old_age
Always - 36
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E4K2JPZT
LU WWN Device Id: 5 0014ee 2b510b64f
Firmware Version: 80.00A80
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is: Fri Nov 10 09:24:28 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (52320) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 523) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 189 179 021 Pre-fail
Always - 7541
4 Start_Stop_Count 0x0032 098 098 000 Old_age
Always - 2626
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 064 064 000 Old_age
Always - 26956
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 312
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 185
193 Load_Cycle_Count 0x0032 188 188 000 Old_age
Always - 38926
194 Temperature_Celsius 0x0022 117 092 000 Old_age
Always - 35
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 192 000 Old_age
Always - 14
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-98-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Western Digital Red
Device Model: WDC WD40EFRX-68WT0N0
Serial Number: WD-WCC4E4VTX9TP
LU WWN Device Id: 5 0014ee 261485dc4
Firmware Version: 82.00A82
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 5400 rpm
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-2 (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is: Fri Nov 10 09:24:31 2017 PST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (54780) seconds.
Offline data collection
capabilities: (0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 548) minutes.
Conveyance self-test routine
recommended polling time: ( 5) minutes.
SCT capabilities: (0x703d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE
UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail
Always - 0
3 Spin_Up_Time 0x0027 189 181 021 Pre-fail
Always - 7516
4 Start_Stop_Count 0x0032 100 100 000 Old_age
Always - 212
5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail
Always - 0
7 Seek_Error_Rate 0x002e 200 200 000 Old_age
Always - 0
9 Power_On_Hours 0x0032 076 076 000 Old_age
Always - 18057
10 Spin_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
11 Calibration_Retry_Count 0x0032 100 100 000 Old_age
Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age
Always - 212
192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age
Always - 122
193 Load_Cycle_Count 0x0032 189 189 000 Old_age
Always - 33611
194 Temperature_Celsius 0x0022 118 092 000 Old_age
Always - 34
196 Reallocated_Event_Count 0x0032 200 200 000 Old_age
Always - 0
197 Current_Pending_Sector 0x0032 200 200 000 Old_age
Always - 0
198 Offline_Uncorrectable 0x0030 100 253 000 Old_age
Offline - 0
199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age
Always - 0
200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age
Offline - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
No self-tests have been logged. [To run self-tests, use: smartctl -t]
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
On Fri, Nov 10, 2017 at 5:13 AM, Mark Knecht <markknecht@gmail.com> wrote:
> On Fri, Nov 10, 2017 at 6:01 AM, Phil Turmel <philip@turmel.org> wrote:
>> Hi Jun-Kai,
>>
>> On 11/10/2017 06:31 AM, Wols Lists wrote:
>>> On 10/11/17 03:09, Jun-Kai Teoh wrote:
>>>> Hi all,
>>>>
>>>> I managed to get my RAID drive back up, content looks like it's still
>>>> there, but it's not resyncing or reshaping and my parity drive was
>>>> removed (I did it when I tried to get it back up).
>>
>> If you can see your content (mounted read-only, I hope), backup up
>> everything, in order from most critical to least critical.
>>
>>>> So what should I do now? I'm afraid of doing anything else at this point.
>>
>> Well, you need to provide more information. At least the mdadm -E
>> reports for all of the member devices. Including the "parity" device
>> you removed. (Parity is spread among all devices in a normal raid5
>> layout, so you may be suffering from a misunderstanding.)
>>
>
> Phil,
> Please reference the thread entitled
>
> "Raid 5 array down/missing - went through wiki steps"
>
> All the mdadm -E info and much more was reported there a week
> ago. Wols and I helped get him that far but neither of us was confident
> in giving instructions to get him further.
>
> Cheers,
> Mark
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 17:25 ` Jun-Kai Teoh
@ 2017-11-10 18:01 ` Phil Turmel
2017-11-10 18:15 ` Jun-Kai Teoh
0 siblings, 1 reply; 12+ messages in thread
From: Phil Turmel @ 2017-11-10 18:01 UTC (permalink / raw)
To: Jun-Kai Teoh, Mark Knecht; +Cc: Wols Lists, Linux-RAID
Hi Jun-Kai,
{Convention on kernel.org is to trim replies and avoid top-posting.}
On 11/10/2017 12:25 PM, Jun-Kai Teoh wrote:
> I was reshaping/growing my array when it died. Computer came back up
> but array wouldn't assemble.
>
> I looked around and saw some commands on how to assemble without the
> one drive that looked like it was kicked off sometime ago that I did
> not realize (I used --assemble --force --verbose)
>
> I kept getting the error that there were 6 drives, 1 rebuilding, and
> not enough to get the array up. I unplugged the one drive that had the
> incorrect superblock and ran "mdadm -A -R".
>
> It ran and looked like it was resuming reshaping but it got really
> slow - like 7 years slow.
>
> I killed the process, rebooted the machine, and it just auto
> reassembled the array and automounted it on boot.
>
> Am in the process of backup right now.
>
> And yes, the drive/array is mounted.
Ok. I reviewed the other thread. Consider creating a new array from
scratch after you complete your backups, using consistent partitioning
across all devices. I would put LVM on top and leave part of it
unallocated (for emergencies), but that's just my preference.
Phil
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 18:01 ` Phil Turmel
@ 2017-11-10 18:15 ` Jun-Kai Teoh
2017-11-10 19:04 ` Phil Turmel
0 siblings, 1 reply; 12+ messages in thread
From: Jun-Kai Teoh @ 2017-11-10 18:15 UTC (permalink / raw)
To: Phil Turmel; +Cc: Mark Knecht, Wols Lists, Linux-RAID
> On Nov 10, 2017, at 10:01 AM, Phil Turmel <philip@turmel.org> wrote:
>
> Hi Jun-Kai,
>
> {Convention on kernel.org is to trim replies and avoid top-posting.}
>
> On 11/10/2017 12:25 PM, Jun-Kai Teoh wrote:
>> I was reshaping/growing my array when it died. Computer came back up
>> but array wouldn't assemble.
>>
>> I looked around and saw some commands on how to assemble without the
>> one drive that looked like it was kicked off sometime ago that I did
>> not realize (I used --assemble --force --verbose)
>>
>> I kept getting the error that there were 6 drives, 1 rebuilding, and
>> not enough to get the array up. I unplugged the one drive that had the
>> incorrect superblock and ran "mdadm -A -R".
>>
>> It ran and looked like it was resuming reshaping but it got really
>> slow - like 7 years slow.
>>
>> I killed the process, rebooted the machine, and it just auto
>> reassembled the array and automounted it on boot.
>>
>> Am in the process of backup right now.
>>
>> And yes, the drive/array is mounted.
>
> Ok. I reviewed the other thread. Consider creating a new array from
> scratch after you complete your backups, using consistent partitioning
> across all devices. I would put LVM on top and leave part of it
> unallocated (for emergencies), but that's just my preference.
>
> Phil
My apologies all! I did not know this.
I'm backing up as much as I can, but I don't think I'll be able to back up everything. I can back up very little of the data, so I'm trying to prioritize carefully right now.
All devices are just 4TB drives, I didn't partition them separately for the array - or am I misunderstanding what you mean? <-- This is really common, all of you are far more knowledgeable than I am and I only understand some of the terms sometimes.
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 18:15 ` Jun-Kai Teoh
@ 2017-11-10 19:04 ` Phil Turmel
2017-11-10 19:22 ` Jun-Kai Teoh
0 siblings, 1 reply; 12+ messages in thread
From: Phil Turmel @ 2017-11-10 19:04 UTC (permalink / raw)
To: Jun-Kai Teoh; +Cc: Mark Knecht, Wols Lists, Linux-RAID
On 11/10/2017 01:15 PM, Jun-Kai Teoh wrote:
>>> Am in the process of backup right now.
>>>
>>> And yes, the drive/array is mounted.
>>
>> Ok. I reviewed the other thread. Consider creating a new array
>> from scratch after you complete your backups, using consistent
>> partitioning across all devices. I would put LVM on top and leave
>> part of it unallocated (for emergencies), but that's just my
>> preference.
> I'm backing up as much as I can, but I don't think I'll be able to
> back up everything. I can back up very little of the data, so I'm
> trying to prioritize carefully right now.
Ah, ok. Well, after you get the important stuff backed up, you can
add more devices as spares and try to let MD continue its grow and
rebuild operations. It might end up completing.
But don't add any more complete devices to the array -- use a partition
that starts at 1MB and covers the rest of the device (default for most
partition tools nowadays).
I recommend that when/if your array is stable again, you add one more
spare (with partition) and then use mdadm's --replace operation to move
complete-device members to the new member. When each is done, take the
newly freed device and partition it and do the next. When you have no
more complete-device members and have the last freed device partitioned,
consider converting to raid6 with that spare.
All of that will take many days, but can be done safely while running
the system.
> All devices are just 4TB drives, I didn't partition them separately
> for the array - or am I misunderstanding what you mean? <-- This is
> really common, all of you are far more knowledgeable than I am and I
> only understand some of the terms sometimes.
MD raid doesn't care if you use whole devices of if you use partitions.
But other system utilities can get confused, and can even write a
new, unwanted partition table. (I recall seeing that with a NAS
product a while back.) So, the 1MB lost is good insurance against
stupid tools.
Phil
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 19:04 ` Phil Turmel
@ 2017-11-10 19:22 ` Jun-Kai Teoh
2017-11-10 21:58 ` Phil Turmel
0 siblings, 1 reply; 12+ messages in thread
From: Jun-Kai Teoh @ 2017-11-10 19:22 UTC (permalink / raw)
To: Phil Turmel; +Cc: Mark Knecht, Wols Lists, Linux-RAID
> On Nov 10, 2017, at 11:04 AM, Phil Turmel <philip@turmel.org> wrote:
>
> On 11/10/2017 01:15 PM, Jun-Kai Teoh wrote:
>
>> I'm backing up as much as I can, but I don't think I'll be able to
>> back up everything. I can back up very little of the data, so I'm
>> trying to prioritize carefully right now.
>
> Ah, ok. Well, after you get the important stuff backed up, you can
> add more devices as spares and try to let MD continue its grow and
> rebuild operations. It might end up completing.
Again, pardon me if I sound super ignorant, I just want to do all of this right.
Do I just plug in another blank 4TB, and run "mdadm -A /dev/md126 /dev/sd[newdrive]" into it?
Or leave it as it is, and just ask it to continue growing with "mdadm -A -R", and *then* add a blank drive into the raid but don't ask it to grow?
> But don't add any more complete devices to the array -- use a partition
> that starts at 1MB and covers the rest of the device (default for most
> partition tools nowadays).
Where can I read about how to do this properly? I tried googling it but I'm pretty lost. Or what tool I should use etc.
> I recommend that when/if your array is stable again, you add one more
> spare (with partition) and then use mdadm's --replace operation to move
> complete-device members to the new member. When each is done, take the
> newly freed device and partition it and do the next. When you have no
> more complete-device members and have the last freed device partitioned,
> consider converting to raid6 with that spare.
I think I understand what you mean here, but it sounds like I need to figure out the top two parts first and that'll take me awhile.
I'll probably have questions again, later, haha.
I really appreciate all the support and patience from this list.
Y'all lifesavers.
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 19:22 ` Jun-Kai Teoh
@ 2017-11-10 21:58 ` Phil Turmel
2017-11-10 23:17 ` Wols Lists
0 siblings, 1 reply; 12+ messages in thread
From: Phil Turmel @ 2017-11-10 21:58 UTC (permalink / raw)
To: Jun-Kai Teoh; +Cc: Mark Knecht, Wols Lists, Linux-RAID
On 11/10/2017 02:22 PM, Jun-Kai Teoh wrote:
>> On Nov 10, 2017, at 11:04 AM, Phil Turmel <philip@turmel.org>
>> wrote:
>>
>> On 11/10/2017 01:15 PM, Jun-Kai Teoh wrote:
>>
>>> I'm backing up as much as I can, but I don't think I'll be able
>>> to back up everything. I can back up very little of the data, so
>>> I'm trying to prioritize carefully right now.
>>
>> Ah, ok. Well, after you get the important stuff backed up, you
>> can add more devices as spares and try to let MD continue its grow
>> and rebuild operations. It might end up completing.
>
> Again, pardon me if I sound super ignorant, I just want to do all of
> this right.
>
> Do I just plug in another blank 4TB, and run "mdadm -A /dev/md126
> /dev/sd[newdrive]" into it?
>
> Or leave it as it is, and just ask it to continue growing with "mdadm
> -A -R", and *then* add a blank drive into the raid but don't ask it
> to grow?
Neither.... With your array mounted and running as it is right now, but
after all backups are done, use:
mdadm --add /dev/mdXXX /dev/sdX1
with the new device's *partition*.
MD should recognize the opportunity to proceed and will resume
rebuilding and growing.
>> But don't add any more complete devices to the array -- use a
>> partition that starts at 1MB and covers the rest of the device
>> (default for most partition tools nowadays).
>
> Where can I read about how to do this properly? I tried googling it
> but I'm pretty lost. Or what tool I should use etc.
parted or gdisk or a very modern version of fdisk will let you partition
your new drives. They will default to a "GPT" disk label and will
generally place the first partition you create in the "right" place -- a
multiple of 4K, typically 1MB. When you save the new partition table
(aka disk label), the #1 partition will show up in your list of devices.
>> I recommend that when/if your array is stable again, you add one
>> more spare (with partition) and then use mdadm's --replace
>> operation to move complete-device members to the new member. When
>> each is done, take the newly freed device and partition it and do
>> the next. When you have no more complete-device members and have
>> the last freed device partitioned, consider converting to raid6
>> with that spare.
>
> I think I understand what you mean here, but it sounds like I need to
> figure out the top two parts first and that'll take me awhile.
Same partitioning procedure for each free disk.
> I'll probably have questions again, later, haha.
Sure, no problem.
> I really appreciate all the support and patience from this list.
>
> Y'all lifesavers.
You're welcome.
Phil
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 21:58 ` Phil Turmel
@ 2017-11-10 23:17 ` Wols Lists
2017-11-10 23:22 ` Wols Lists
0 siblings, 1 reply; 12+ messages in thread
From: Wols Lists @ 2017-11-10 23:17 UTC (permalink / raw)
To: Jun-Kai Teoh; +Cc: Linux-RAID
On 10/11/17 21:58, Phil Turmel wrote:
>>> But don't add any more complete devices to the array -- use a
>>> >> partition that starts at 1MB and covers the rest of the device
>>> >> (default for most partition tools nowadays).
>> >
>> > Where can I read about how to do this properly? I tried googling it
>> > but I'm pretty lost. Or what tool I should use etc.
> parted or gdisk or a very modern version of fdisk will let you partition
> your new drives. They will default to a "GPT" disk label and will
> generally place the first partition you create in the "right" place -- a
> multiple of 4K, typically 1MB. When you save the new partition table
> (aka disk label), the #1 partition will show up in your list of devices.
>
>>> >> I recommend that when/if your array is stable again, you add one
>>> >> more spare (with partition) and then use mdadm's --replace
>>> >> operation to move complete-device members to the new member. When
>>> >> each is done, take the newly freed device and partition it and do
>>> >> the next. When you have no more complete-device members and have
>>> >> the last freed device partitioned, consider converting to raid6
>>> >> with that spare.
>> >
>> > I think I understand what you mean here, but it sounds like I need to
>> > figure out the top two parts first and that'll take me awhile.
If you take the wiki slowly, you'll hopefully grasp what's going on.
https://raid.wiki.kernel.org/index.php/Linux_Raid#Overview
Especially the bit about setting up a new system or converting an
existing system. Okay, they're not at all about what you're doing, but
if you work your way through and get what's happening there straight in
your head, you'll then be able to apply it to your setup.
Note that they're written from my perspective as a gentoo user, so I
have to do everything from scratch - boot into a live/rescue CD, and
then configure/setup the system from there.
I know it doesn't cover LVM - that is very patchy at the moment - but I
just need to troubleshoot my new machine to get a bios screen, and I'll
be setting that up with LVM and all that should be appearing on the wiki.
(My new system at the moment isn't even POSTing :-( I'm not used to
troubleshooting builds, I've usually been lucky, so I hope everything is
working soon :-)
Cheers,
Wol
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: RAID5 up, but one drive removed, one says spare building, what now?
2017-11-10 23:17 ` Wols Lists
@ 2017-11-10 23:22 ` Wols Lists
0 siblings, 0 replies; 12+ messages in thread
From: Wols Lists @ 2017-11-10 23:22 UTC (permalink / raw)
To: Jun-Kai Teoh; +Cc: Linux-RAID
On 10/11/17 23:17, Wols Lists wrote:
> If you take the wiki slowly, you'll hopefully grasp what's going on.
>
> https://raid.wiki.kernel.org/index.php/Linux_Raid#Overview
>
> Especially the bit about setting up a new system or converting an
> existing system. Okay, they're not at all about what you're doing, but
> if you work your way through and get what's happening there straight in
> your head, you'll then be able to apply it to your setup.
>
> Note that they're written from my perspective as a gentoo user, so I
> have to do everything from scratch - boot into a live/rescue CD, and
> then configure/setup the system from there.
Just checked those two pages (should have done that before I hit send
:-) and they're not quite as appropriate as I thought. Still worth
reading, though.
But that's given me an idea for more detail in the wiki, so this should
lead to further improvement :-)
Cheers,
Wol
^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2017-11-10 23:22 UTC | newest]
Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-11-10 3:09 RAID5 up, but one drive removed, one says spare building, what now? Jun-Kai Teoh
2017-11-10 11:31 ` Wols Lists
2017-11-10 13:01 ` Phil Turmel
2017-11-10 13:13 ` Mark Knecht
2017-11-10 17:25 ` Jun-Kai Teoh
2017-11-10 18:01 ` Phil Turmel
2017-11-10 18:15 ` Jun-Kai Teoh
2017-11-10 19:04 ` Phil Turmel
2017-11-10 19:22 ` Jun-Kai Teoh
2017-11-10 21:58 ` Phil Turmel
2017-11-10 23:17 ` Wols Lists
2017-11-10 23:22 ` Wols Lists
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.