All of lore.kernel.org
 help / color / mirror / Atom feed
* --examine contradicts --create and --detail, again
@ 2009-02-11 10:48 whollygoat
  2009-02-11 13:59 ` Robin Hill
  0 siblings, 1 reply; 3+ messages in thread
From: whollygoat @ 2009-02-11 10:48 UTC (permalink / raw)
  To: linux-raid

I am on my 4th or 5th try at getting a raid5 array 
set up.  I did have one working quite nicely with 6 
(5 active, 1 spare) 40GB drives but have had a 
succession of failures since trying to grow the array
after putting bigger drives in the box.

Either I am misunderstanding something, or I have
a hardware issue I don't know how to diagnose.  Any 
suggestions on things to read or to try are welcome.

I've been told that zeroing the superblock is sufficient
but having had so many failures I wanted to be thourough


# dd if=/dev/zero of=/dev/hd{egikmo} (for brevity, each was
run separately)

# rm -R /dev/md0 /dev/md

# fdisk /dev/hd{egikmo}
 - creating a new, primary, partition starting cylinder 1, 80G
 - type da (non-fs data)
 
#mdadm -C /dev/md/0 -e 1.0 -v -l 5 -b internal -a yes\
  -n 5 /dev/hde1 /dev/hdg1 /dev/hdi1 /dev/hdk1 /dev/hdm1\
  -x 1 /dev/hdo1  --name=FlyFileServ_md

After waiting for the initial syncing of the array to complete

# mdadm -D /dev/md/0
/dev/md/0:
        Version : 01.00.03
  Creation Time : Tue Feb 10 14:45:39 2009
     Raid Level : raid5
     Array Size : 312501760 (298.02 GiB 320.00 GB)
    Device Size : 156250880 (74.51 GiB 80.00 GB)
   Raid Devices : 5
  Total Devices : 6
Preferred Minor : 0
    Persistence : Superblock is persistent

  Intent Bitmap : Internal
   
    Update Time : Tue Feb 10 16:41:48 2009
          State : active
 Active Devices : 5
Working Devices : 6
 Failed Devices : 0
  Spare Devices : 1
           
         Layout : left-symmetric
     Chunk Size : 64K
           
           Name : fly:FlyFileServ_md  (local to host fly)
           UUID : 684bf5f1:de2c0d2a:5a5ac88f:de7cf2d3
         Events : 2
    
    Number   Major   Minor   RaidDevice State
       0      33        1        0      active sync   /dev/hde1
       1      34        1        1      active sync   /dev/hdg1
       2      56        1        2      active sync   /dev/hdi1
       3      57        1        3      active sync   /dev/hdk1
       6      88        1        4      active sync   /dev/hdm1
         
       5      89        1        -      spare   /dev/hdo1

So, then, why oh why oh why does --examine, on any of the 
component devices, show slots for  7 devices, one failed, 
one empty?  I have recently changed RAM, mobo, system disk
and some IDE cables thinking each time that I had finally
come to the end of this.



# mdadm -E /dev/hde1
/dev/hde1:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x1
     Array UUID : 684bf5f1:de2c0d2a:5a5ac88f:de7cf2d3
           Name : fly:FlyFileServ_md  (local to host fly)
  Creation Time : Tue Feb 10 14:45:39 2009
     Raid Level : raid5
   Raid Devices : 5
     
    Device Size : 156250880 (74.51 GiB 80.00 GB)
     Array Size : 625003520 (298.02 GiB 320.00 GB)
   Super Offset : 156251008 sectors
          State : clean
    Device UUID : 88b0d67e:3e2cf8ee:83f58286:4040c5da

Internal Bitmap : 2 sectors from superblock
    Update Time : Tue Feb 10 16:41:48 2009
       Checksum : dae7896e - correct
         Events : 2

         Layout : left-symmetric
     Chunk Size : 64K

    Array Slot : 0 (0, 1, 2, 3, failed, empty, 4)
   Array State : Uuuuu 1 failed

Sorry for bringing up this new instance of the same old 
problem yet one more time.
-- 
  
  whollygoat@letterboxes.org

-- 
http://www.fastmail.fm - A no graphics, no pop-ups email service


^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: --examine contradicts --create and --detail, again
  2009-02-11 10:48 --examine contradicts --create and --detail, again whollygoat
@ 2009-02-11 13:59 ` Robin Hill
  2009-02-16  6:22   ` whollygoat
  0 siblings, 1 reply; 3+ messages in thread
From: Robin Hill @ 2009-02-11 13:59 UTC (permalink / raw)
  To: linux-raid

[-- Attachment #1: Type: text/plain, Size: 2162 bytes --]

On Wed Feb 11, 2009 at 02:48:09AM -0800, whollygoat@letterboxes.org wrote:

> Either I am misunderstanding something, or I have
> a hardware issue I don't know how to diagnose.  Any 
> suggestions on things to read or to try are welcome.
> 
> # mdadm -D /dev/md/0
>     
>     Number   Major   Minor   RaidDevice State
>        0      33        1        0      active sync   /dev/hde1
>        1      34        1        1      active sync   /dev/hdg1
>        2      56        1        2      active sync   /dev/hdi1
>        3      57        1        3      active sync   /dev/hdk1
>        6      88        1        4      active sync   /dev/hdm1
>          
>        5      89        1        -      spare   /dev/hdo1
> 
> So, then, why oh why oh why does --examine, on any of the 
> component devices, show slots for  7 devices, one failed, 
> one empty?  I have recently changed RAM, mobo, system disk
> and some IDE cables thinking each time that I had finally
> come to the end of this.
> 
> # mdadm -E /dev/hde1
> 
>     Array Slot : 0 (0, 1, 2, 3, failed, empty, 4)
>    Array State : Uuuuu 1 failed
> 
> Sorry for bringing up this new instance of the same old 
> problem yet one more time.
>
I'd say this is just an artifact of how the array creation works,
Because it starts as a degraded array, initially one of the RAID devices
is flagged as missing/faulty (so you have member numbers 0, 1, 2 and 3
in the array, 4 flagged as missing/faulty, and 5 & 6 as spares).  When
it does the initial build, it uses member number 6, leaving 5 as the
spare.  This leaves you with 1 "failed" member and one "empty" (the
spare - it's not part of the array so doesn't have a Device Number).

Unfortunately I don't have any RAID5 arrays around now to check whether
I see the same here.  However, if the system's working fine otherwise
then I certainly wouldn't worry about it.

Cheers,
    Robin
-- 
     ___        
    ( ' }     |       Robin Hill        <robin@robinhill.me.uk> |
   / / )      | Little Jim says ....                            |
  // !!       |      "He fallen in de water !!"                 |

[-- Attachment #2: Type: application/pgp-signature, Size: 197 bytes --]

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: --examine contradicts --create and --detail, again
  2009-02-11 13:59 ` Robin Hill
@ 2009-02-16  6:22   ` whollygoat
  0 siblings, 0 replies; 3+ messages in thread
From: whollygoat @ 2009-02-16  6:22 UTC (permalink / raw)
  To: linux-raid; +Cc: Robin Hill

On Wed, 11 Feb 2009 13:59:17 +0000, "Robin Hill" <robin@robinhill.me.uk>
said:
> On Wed Feb 11, 2009 at 02:48:09AM -0800, whollygoat@letterboxes.org
> wrote:
> 
[snip]

> > # mdadm -E /dev/hde1
> > 
> >     Array Slot : 0 (0, 1, 2, 3, failed, empty, 4)
> >    Array State : Uuuuu 1 failed
> > 
> > Sorry for bringing up this new instance of the same old 
> > problem yet one more time.
> >
> I'd say this is just an artifact of how the array creation works,
> Because it starts as a degraded array, initially one of the RAID devices
> is flagged as missing/faulty (so you have member numbers 0, 1, 2 and 3
> in the array, 4 flagged as missing/faulty, and 5 & 6 as spares).  When
> it does the initial build, it uses member number 6, leaving 5 as the
> spare.  This leaves you with 1 "failed" member and one "empty" (the
> spare - it's not part of the array so doesn't have a Device Number).
> 
> Unfortunately I don't have any RAID5 arrays around now to check whether
> I see the same here.  However, if the system's working fine otherwise
> then I certainly wouldn't worry about it.

Thanks for the reply.  Unfortunately, I can't claim to be enjoying a
finely working system.  Since your reply, I've tried unplugging the ide
expansion cards from the pci but, and playing around with the ide
cables.

I've managed to pull off one create run where the end result showed no
"failed" or "empty" slots for --examine.  Unfortunately, I have not been
able to reproduce that "clean" run, even with the same card removed.  So
I 
am guessing that it is either an ide cable or the sole ide card I have
not
been able to remove (kernel oopses when this card is not plugged into
the 
bus).

Ta,

wg
some bad h/w somewhere.
-- 
  
  whollygoat@letterboxes.org

-- 
http://www.fastmail.fm - The way an email service should be


^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2009-02-16  6:22 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2009-02-11 10:48 --examine contradicts --create and --detail, again whollygoat
2009-02-11 13:59 ` Robin Hill
2009-02-16  6:22   ` whollygoat

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.