* mdadm not assembling RAID?
@ 2022-08-13 19:58 David F.
2022-08-13 23:02 ` David F.
0 siblings, 1 reply; 3+ messages in thread
From: David F. @ 2022-08-13 19:58 UTC (permalink / raw)
To: linux-raid
I have an older system with Intel RAID (striped) I use to test updates
to my linux boot disk to ensure the RAID is still working. I just
went through a big update of all the utilities and shared libraries
along with moving from 5.10 to 5.15 kernel. There was a stage where
the kernel was updated to 5.15 but all the utilities and shared
libraries were not (and md raid was working). The problem is now
there are no md devices created, but there are a bunch of dm devices
created that never were before.
This group has been very helpful in the past. I hope you can help
with what is going on here? Does it have something to do with the
updated devmapper ? Should I revert?
I put both the old working information and the new information below -
separated by the 3 lines of equals.
============================================================
============================================================
============================================================
Here's the old boot information where it works (5.10.x kernel):
Contents of /sys/block:
fd0 loop1 loop3 loop5 loop7 md127 sdb sr0
loop0 loop2 loop4 loop6 md126 sda sdc
Contents of /dev:
block md126p4 tty19 tty57
bsg md126p5 tty2 tty58
btrfs-control md127 tty20 tty59
bus mem tty21 tty6
cdrom null tty22 tty60
cdrw port tty23 tty61
char psaux tty24 tty62
console ptmx tty25 tty63
core pts tty26 tty7
cpu_dma_latency random tty27 tty8
disk rfkill tty28 tty9
dvd rtc tty29 ttyS0
fb0 rtc0 tty3 ttyS1
fd sda tty30 ttyS2
fd0 sda1 tty31 ttyS3
full sda2 tty32 urandom
fuse sda3 tty33 vcs
fw0 sda5 tty34 vcs1
hidraw0 sdb tty35 vcs2
hidraw1 sdc tty36 vcs3
hidraw2 sdc1 tty37 vcs4
hidraw3 sg0 tty38 vcs5
hpet sg1 tty39 vcs6
input sg2 tty4 vcsa
kmsg sg3 tty40 vcsa1
lightnvm sr0 tty41 vcsa2
log stderr tty42 vcsa3
loop-control stdin tty43 vcsa4
loop0 stdout tty44 vcsa5
loop1 synth tty45 vcsa6
loop2 tty tty46 vcsu
loop3 tty0 tty47 vcsu1
loop4 tty1 tty48 vcsu2
loop5 tty10 tty49 vcsu3
loop6 tty11 tty5 vcsu4
loop7 tty12 tty50 vcsu5
mapper tty13 tty51 vcsu6
md tty14 tty52 vga_arbiter
md126 tty15 tty53 zero
md126p1 tty16 tty54
md126p2 tty17 tty55
md126p3 tty18 tty56
Contents of /proc/partitions:
major minor #blocks name
2 0 4 fd0
11 0 1048575 sr0
8 0 244198584 sda
8 1 32684242 sda1
8 2 1 sda2
8 3 514048 sda3
8 5 2088418 sda5
8 16 732574584 sdb
8 32 7897088 sdc
8 33 7895916 sdc1
9 126 488391680 md126
259 0 32684242 md126p1
259 1 1 md126p2
259 2 514048 md126p3
259 3 8032 md126p4
259 4 10040593 md126p5
Contents of /proc/mdstat (Linux software RAID status):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5]
[raid4] [multipath]
md126 : active raid0 sda[1] sdb[0]
488391680 blocks super external:/md127/0 128k chunks
md127 : inactive sda[1](S) sdb[0](S)
7945 blocks super external:imsm
unused devices: <none>
Contents of /run/mdadm/map (Linux software RAID arrays):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
md126 /md127/0 835de710:3d35bfb1:d159af46:6570f120 /dev/md/TestRAID
md127 imsm 788c3635:2e37de4b:87d08323:987f57e5 /dev/md/imsm0
Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# mdadm.conf
#
# Please refer to mdadm.conf(5) for information about this file.
#
# by default (built-in), scan all partitions (/proc/partitions) and all
# containers for MD superblocks. alternatively, specify devices to scan, using
# wildcards if desired.
DEVICE partitions containers
# automatically tag new arrays as belonging to the local system
HOMEHOST <system>
ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
Long listing of /dev/md directory and /dev/md* device files (Linux
software RAID devices):
brw-rw---- 1 root disk 9, 126 Aug 13 04:17 /dev/md126
brw-rw---- 1 root disk 259, 0 Aug 13 04:17 /dev/md126p1
brw-rw---- 1 root disk 259, 1 Aug 13 04:17 /dev/md126p2
brw-rw---- 1 root disk 259, 2 Aug 13 04:17 /dev/md126p3
brw-rw---- 1 root disk 259, 3 Aug 13 04:17 /dev/md126p4
brw-rw---- 1 root disk 259, 4 Aug 13 04:17 /dev/md126p5
brw-rw---- 1 root disk 9, 127 Aug 13 04:17 /dev/md127
/dev/md:
lrwxrwxrwx 1 root root 8 Aug 13 04:17 TestRAID -> ../md126
lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID1 -> ../md126p1
lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID2 -> ../md126p2
lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID3 -> ../md126p3
lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID4 -> ../md126p4
lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID5 -> ../md126p5
lrwxrwxrwx 1 root root 8 Aug 13 04:17 imsm0 -> ../md127
Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data
captured when 'start-md' is executed):
mdadm - v4.1 - 2018-10-01
Output of 'mdadm --examine --scan'
ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
Output of 'mdadm --assemble --scan --no-degraded -v'
mdadm: looking for devices for further assembly
mdadm: no RAID superblock on /dev/sdc1
mdadm: /dev/sdc is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/sdc
mdadm: no RAID superblock on /dev/sdc
mdadm: no RAID superblock on /dev/sda5
mdadm: no RAID superblock on /dev/sda3
mdadm: no RAID superblock on /dev/sda2
mdadm: no RAID superblock on /dev/sda1
mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot 1.
mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot 0.
mdadm: added /dev/sdb to /dev/md/imsm0 as 1
mdadm: added /dev/sda to /dev/md/imsm0 as 0
mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
mdadm: looking for devices for /dev/md/TestRAID
mdadm: looking in container /dev/md/imsm0
mdadm: found match on member /md127/0 in /dev/md/imsm0
mdadm: Started /dev/md/TestRAID with 2 devices
Output of 'dmesg | grep md:'
md: md127 stopped.
md: md126 stopped.
Output of 'mdadm --examine /dev/sda'
/dev/sda:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.0.00
Orig Family : a209f7f5
Family : a209f7f5
Generation : 0000016d
Attributes : All supported
UUID : 788c3635:2e37de4b:87d08323:987f57e5
Checksum : 7cb36503 correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk00 Serial : WD-WCANK2656465
State : active
Id : 00000000
Usable Size : 488391680 (232.88 GiB 250.06 GB)
[TestRAID]:
UUID : 835de710:3d35bfb1:d159af46:6570f120
RAID Level : 0
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 0
Sector Size : 512
Array Size : 976783360 (465.77 GiB 500.11 GB)
Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
Sector Offset : 0
Num Stripes : 1907780
Chunk Size : 128 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean
RWH Policy : off
Disk01 Serial : 3QK01SE5
State : active
Id : 00010000
Usable Size : 1465143680 (698.63 GiB 750.15 GB)
Output of 'mdadm --examine /dev/sdb'
/dev/sdb:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.0.00
Orig Family : a209f7f5
Family : a209f7f5
Generation : 0000016d
Attributes : All supported
UUID : 788c3635:2e37de4b:87d08323:987f57e5
Checksum : 7cb36503 correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk01 Serial : 3QK01SE5
State : active
Id : 00010000
Usable Size : 1465138766 (698.63 GiB 750.15 GB)
[TestRAID]:
UUID : 835de710:3d35bfb1:d159af46:6570f120
RAID Level : 0
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 1
Sector Size : 512
Array Size : 976783360 (465.77 GiB 500.11 GB)
Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
Sector Offset : 0
Num Stripes : 1907780
Chunk Size : 128 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean
RWH Policy : off
Disk00 Serial : WD-WCANK2656465
State : active
Id : 00000000
Usable Size : 488386766 (232.88 GiB 250.05 GB)
Output of 'mdadm --examine /dev/sdc'
/dev/sdc:
MBR Magic : aa55
Partition[0] : 15791832 sectors at 63 (type 0b)
Output of 'mdadm --detail /dev/md*', if any:
mdadm: /dev/md does not appear to be an md device
/dev/md126:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 488391680 (465.77 GiB 500.11 GB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md126p1:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 32684242 (31.17 GiB 33.47 GB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md126p2:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 1
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md126p3:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 514048 (502.00 MiB 526.39 MB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md126p4:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 8032 (7.84 MiB 8.23 MB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md126p5:
Container : /dev/md/imsm0, member 0
Raid Level : raid0
Array Size : 10040593 (9.58 GiB 10.28 GB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Chunk Size : 128K
Consistency Policy : none
UUID : 835de710:3d35bfb1:d159af46:6570f120
Number Major Minor RaidDevice State
1 8 0 0 active sync /dev/sda
0 8 16 1 active sync /dev/sdb
/dev/md127:
Version : imsm
Raid Level : container
Total Devices : 2
Working Devices : 2
UUID : 788c3635:2e37de4b:87d08323:987f57e5
Member Arrays : /dev/md/TestRAID
Number Major Minor RaidDevice
- 8 0 - /dev/sda
- 8 16 - /dev/sdb
Contents of /dev/mapper directory:
crw------- 1 root root 10, 236 Aug 13 04:17 control
============================================================
============================================================
============================================================
============================================================
Here is the new version that doesn't work (5.15.x kernel):
Contents of /sys/block:
dm-0 dm-2 dm-5 loop0 loop2 loop4 loop6 sda sdc
dm-1 dm-4 fd0 loop1 loop3 loop5 loop7 sdb sr0
Contents of /dev:
block mem tty20 tty58
bsg null tty21 tty59
btrfs-control port tty22 tty6
bus psaux tty23 tty60
cdrom ptmx tty24 tty61
cdrw pts tty25 tty62
char random tty26 tty63
console rfkill tty27 tty7
core rtc tty28 tty8
cpu_dma_latency rtc0 tty29 tty9
disk sda tty3 ttyS0
dm-0 sda1 tty30 ttyS1
dm-1 sda2 tty31 ttyS2
dm-2 sda3 tty32 ttyS3
dm-4 sda5 tty33 urandom
dm-5 sdb tty34 vcs
dvd sdc tty35 vcs1
fb0 sdc1 tty36 vcs2
fd sg0 tty37 vcs3
fd0 sg1 tty38 vcs4
full sg2 tty39 vcs5
fuse sg3 tty4 vcs6
fw0 sr0 tty40 vcsa
hidraw0 stderr tty41 vcsa1
hidraw1 stdin tty42 vcsa2
hidraw2 stdout tty43 vcsa3
hidraw3 synth tty44 vcsa4
hpet tty tty45 vcsa5
input tty0 tty46 vcsa6
kmsg tty1 tty47 vcsu
log tty10 tty48 vcsu1
loop-control tty11 tty49 vcsu2
loop0 tty12 tty5 vcsu3
loop1 tty13 tty50 vcsu4
loop2 tty14 tty51 vcsu5
loop3 tty15 tty52 vcsu6
loop4 tty16 tty53 vga_arbiter
loop5 tty17 tty54 zero
loop6 tty18 tty55
loop7 tty19 tty56
mapper tty2 tty57
Contents of /proc/partitions:
major minor #blocks name
2 0 4 fd0
11 0 1048575 sr0
8 16 732574584 sdb
8 0 244198584 sda
8 1 32684242 sda1
8 2 1 sda2
8 3 514048 sda3
8 5 2088418 sda5
252 0 488391936 dm-0
252 1 32684242 dm-1
252 2 514048 dm-2
252 4 8032 dm-4
252 5 10040593 dm-5
8 32 15687680 sdc
8 33 15687441 sdc1
Listing of /dev/dm-x devices (LVM, dmraid, etc.):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
brw-rw---- 1 root disk 252, 0 Aug 13 04:27 /dev/dm-0
brw-rw---- 1 root disk 252, 1 Aug 13 04:27 /dev/dm-1
brw-rw---- 1 root disk 252, 2 Aug 13 04:27 /dev/dm-2
brw-rw---- 1 root disk 252, 4 Aug 13 04:27 /dev/dm-4
brw-rw---- 1 root disk 252, 5 Aug 13 04:27 /dev/dm-5
Listing of /sys/block/dm-x/slaves:
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
/sys/block/dm-0/slaves:
lrwxrwxrwx 1 root root 0 Aug 13 11:28 sda ->
../../../../pci0000:00/0000:00:1f.2/ata3/host2/target2:0:0/2:0:0:0/block/sda
lrwxrwxrwx 1 root root 0 Aug 13 04:28 sdb ->
../../../../pci0000:00/0000:00:1f.2/ata4/host3/target3:0:0/3:0:0:0/block/sdb
/sys/block/dm-1/slaves:
lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
/sys/block/dm-2/slaves:
lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
/sys/block/dm-4/slaves:
lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
/sys/block/dm-5/slaves:
lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
Contents of /proc/mdstat (Linux software RAID status):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5]
[raid4] [multipath]
unused devices: <none>
Contents of /run/mdadm/map (Linux software RAID arrays):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file):
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# mdadm.conf
#
# Please refer to mdadm.conf(5) for information about this file.
#
# by default (built-in), scan all partitions (/proc/partitions) and all
# containers for MD superblocks. alternatively, specify devices to scan, using
# wildcards if desired.
DEVICE partitions containers
# automatically tag new arrays as belonging to the local system
HOMEHOST <system>
ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data
captured when 'start-md' is executed):
mdadm - v4.1 - 2018-10-01
Output of 'mdadm --examine --scan'
ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
Output of 'mdadm --assemble --scan --no-degraded -v'
mdadm: looking for devices for further assembly
mdadm: no RAID superblock on /dev/sdc1
mdadm: /dev/sdc is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/sdc
mdadm: no RAID superblock on /dev/sdc
mdadm: /dev/dm-5 is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/dm-5
mdadm: no RAID superblock on /dev/dm-5
mdadm: /dev/dm-4 is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/dm-4
mdadm: no RAID superblock on /dev/dm-4
mdadm: /dev/dm-2 is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/dm-2
mdadm: no RAID superblock on /dev/dm-2
mdadm: /dev/dm-1 is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/dm-1
mdadm: no RAID superblock on /dev/dm-1
mdadm: /dev/dm-0 is not attached to Intel(R) RAID controller.
mdadm: No OROM/EFI properties for /dev/dm-0
mdadm: no RAID superblock on /dev/dm-0
mdadm: no RAID superblock on /dev/sda5
mdadm: no RAID superblock on /dev/sda3
mdadm: no RAID superblock on /dev/sda2
mdadm: no RAID superblock on /dev/sda1
mdadm: /dev/sda is busy - skipping
mdadm: /dev/sdb is busy - skipping
mdadm: looking for devices for /dev/md/TestRAID
mdadm: Cannot assemble mbr metadata on /dev/sdc1
mdadm: Cannot assemble mbr metadata on /dev/sdc
mdadm: Cannot assemble mbr metadata on /dev/dm-5
mdadm: Cannot assemble mbr metadata on /dev/dm-4
mdadm: Cannot assemble mbr metadata on /dev/dm-2
mdadm: Cannot assemble mbr metadata on /dev/dm-1
mdadm: Cannot assemble mbr metadata on /dev/dm-0
mdadm: Cannot assemble mbr metadata on /dev/sda5
mdadm: Cannot assemble mbr metadata on /dev/sda3
mdadm: Cannot assemble mbr metadata on /dev/sda2
mdadm: no recogniseable superblock on /dev/sda1
mdadm: /dev/sda has wrong uuid.
mdadm: /dev/sdb has wrong uuid.
Output of 'dmesg | grep md:'
Output of 'mdadm --examine /dev/sda'
/dev/sda:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.0.00
Orig Family : a209f7f5
Family : a209f7f5
Generation : 0000016e
Attributes : All supported
UUID : 788c3635:2e37de4b:87d08323:987f57e5
Checksum : 7cb36504 correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk00 Serial : WD-WCANK2656465
State : active
Id : 00000000
Usable Size : 488391680 (232.88 GiB 250.06 GB)
[TestRAID]:
UUID : 835de710:3d35bfb1:d159af46:6570f120
RAID Level : 0
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 0
Sector Size : 512
Array Size : 976783360 (465.77 GiB 500.11 GB)
Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
Sector Offset : 0
Num Stripes : 1907780
Chunk Size : 128 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean
RWH Policy : off
Disk01 Serial : 3QK01SE5
State : active
Id : 00010000
Usable Size : 1465143680 (698.63 GiB 750.15 GB)
Output of 'mdadm --examine /dev/sdb'
/dev/sdb:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.0.00
Orig Family : a209f7f5
Family : a209f7f5
Generation : 0000016e
Attributes : All supported
UUID : 788c3635:2e37de4b:87d08323:987f57e5
Checksum : 7cb36504 correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk01 Serial : 3QK01SE5
State : active
Id : 00010000
Usable Size : 1465138766 (698.63 GiB 750.15 GB)
[TestRAID]:
UUID : 835de710:3d35bfb1:d159af46:6570f120
RAID Level : 0
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 1
Sector Size : 512
Array Size : 976783360 (465.77 GiB 500.11 GB)
Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
Sector Offset : 0
Num Stripes : 1907780
Chunk Size : 128 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean
RWH Policy : off
Disk00 Serial : WD-WCANK2656465
State : active
Id : 00000000
Usable Size : 488386766 (232.88 GiB 250.05 GB)
Output of 'mdadm --examine /dev/sdc'
/dev/sdc:
MBR Magic : aa55
Partition[0] : 31374882 sectors at 63 (type 0b)
Output of 'mdadm --detail /dev/md*', if any:
mdadm: cannot open /dev/md*: No such file or directory
Contents of /dev/mapper directory:
crw------- 1 root root 10, 236 Aug 13 04:27 control
brw-rw---- 1 root disk 252, 0 Aug 13 04:27 isw_chbifgccjd_TestRAID
brw-rw---- 1 root disk 252, 1 Aug 13 04:27
isw_chbifgccjd_TestRAID1
brw-rw---- 1 root disk 252, 2 Aug 13 04:27
isw_chbifgccjd_TestRAID3
brw-rw---- 1 root disk 252, 4 Aug 13 04:27
isw_chbifgccjd_TestRAID4
brw-rw---- 1 root disk 252, 5 Aug 13 04:27
isw_chbifgccjd_TestRAID5
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: mdadm not assembling RAID?
2022-08-13 19:58 mdadm not assembling RAID? David F.
@ 2022-08-13 23:02 ` David F.
2022-08-14 0:48 ` David F.
0 siblings, 1 reply; 3+ messages in thread
From: David F. @ 2022-08-13 23:02 UTC (permalink / raw)
To: linux-raid
So it does have something to with the mapper devices. If I add
"nodmraid" kernel parameter then mdadm raid works like it did before.
I want searching config parameters for the kernel to see if I enabled
DMRAID somehow but didn't find any. Why is the new disk/updated
programs automatically starting dmraid? could it be udev rules .. I
don't really understand them .. but dmraid is an option for non intel
type raid, so what happens is dmraid is tried first for all but the
intel items, then mdadm is used. But this is not coming from my call
of dmraid, it's coming from someplace automatic ??
On Sat, Aug 13, 2022 at 12:58 PM David F. <df7729@gmail.com> wrote:
>
> I have an older system with Intel RAID (striped) I use to test updates
> to my linux boot disk to ensure the RAID is still working. I just
> went through a big update of all the utilities and shared libraries
> along with moving from 5.10 to 5.15 kernel. There was a stage where
> the kernel was updated to 5.15 but all the utilities and shared
> libraries were not (and md raid was working). The problem is now
> there are no md devices created, but there are a bunch of dm devices
> created that never were before.
>
> This group has been very helpful in the past. I hope you can help
> with what is going on here? Does it have something to do with the
> updated devmapper ? Should I revert?
>
> I put both the old working information and the new information below -
> separated by the 3 lines of equals.
>
>
> ============================================================
> ============================================================
> ============================================================
>
> Here's the old boot information where it works (5.10.x kernel):
>
> Contents of /sys/block:
> fd0 loop1 loop3 loop5 loop7 md127 sdb sr0
> loop0 loop2 loop4 loop6 md126 sda sdc
>
> Contents of /dev:
> block md126p4 tty19 tty57
> bsg md126p5 tty2 tty58
> btrfs-control md127 tty20 tty59
> bus mem tty21 tty6
> cdrom null tty22 tty60
> cdrw port tty23 tty61
> char psaux tty24 tty62
> console ptmx tty25 tty63
> core pts tty26 tty7
> cpu_dma_latency random tty27 tty8
> disk rfkill tty28 tty9
> dvd rtc tty29 ttyS0
> fb0 rtc0 tty3 ttyS1
> fd sda tty30 ttyS2
> fd0 sda1 tty31 ttyS3
> full sda2 tty32 urandom
> fuse sda3 tty33 vcs
> fw0 sda5 tty34 vcs1
> hidraw0 sdb tty35 vcs2
> hidraw1 sdc tty36 vcs3
> hidraw2 sdc1 tty37 vcs4
> hidraw3 sg0 tty38 vcs5
> hpet sg1 tty39 vcs6
> input sg2 tty4 vcsa
> kmsg sg3 tty40 vcsa1
> lightnvm sr0 tty41 vcsa2
> log stderr tty42 vcsa3
> loop-control stdin tty43 vcsa4
> loop0 stdout tty44 vcsa5
> loop1 synth tty45 vcsa6
> loop2 tty tty46 vcsu
> loop3 tty0 tty47 vcsu1
> loop4 tty1 tty48 vcsu2
> loop5 tty10 tty49 vcsu3
> loop6 tty11 tty5 vcsu4
> loop7 tty12 tty50 vcsu5
> mapper tty13 tty51 vcsu6
> md tty14 tty52 vga_arbiter
> md126 tty15 tty53 zero
> md126p1 tty16 tty54
> md126p2 tty17 tty55
> md126p3 tty18 tty56
>
> Contents of /proc/partitions:
> major minor #blocks name
>
> 2 0 4 fd0
> 11 0 1048575 sr0
> 8 0 244198584 sda
> 8 1 32684242 sda1
> 8 2 1 sda2
> 8 3 514048 sda3
> 8 5 2088418 sda5
> 8 16 732574584 sdb
> 8 32 7897088 sdc
> 8 33 7895916 sdc1
> 9 126 488391680 md126
> 259 0 32684242 md126p1
> 259 1 1 md126p2
> 259 2 514048 md126p3
> 259 3 8032 md126p4
> 259 4 10040593 md126p5
>
> Contents of /proc/mdstat (Linux software RAID status):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5]
> [raid4] [multipath]
> md126 : active raid0 sda[1] sdb[0]
> 488391680 blocks super external:/md127/0 128k chunks
>
> md127 : inactive sda[1](S) sdb[0](S)
> 7945 blocks super external:imsm
>
> unused devices: <none>
>
> Contents of /run/mdadm/map (Linux software RAID arrays):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> md126 /md127/0 835de710:3d35bfb1:d159af46:6570f120 /dev/md/TestRAID
> md127 imsm 788c3635:2e37de4b:87d08323:987f57e5 /dev/md/imsm0
>
> Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> # mdadm.conf
> #
> # Please refer to mdadm.conf(5) for information about this file.
> #
>
> # by default (built-in), scan all partitions (/proc/partitions) and all
> # containers for MD superblocks. alternatively, specify devices to scan, using
> # wildcards if desired.
> DEVICE partitions containers
>
> # automatically tag new arrays as belonging to the local system
> HOMEHOST <system>
>
> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
> ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
> member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
>
> Long listing of /dev/md directory and /dev/md* device files (Linux
> software RAID devices):
> brw-rw---- 1 root disk 9, 126 Aug 13 04:17 /dev/md126
> brw-rw---- 1 root disk 259, 0 Aug 13 04:17 /dev/md126p1
> brw-rw---- 1 root disk 259, 1 Aug 13 04:17 /dev/md126p2
> brw-rw---- 1 root disk 259, 2 Aug 13 04:17 /dev/md126p3
> brw-rw---- 1 root disk 259, 3 Aug 13 04:17 /dev/md126p4
> brw-rw---- 1 root disk 259, 4 Aug 13 04:17 /dev/md126p5
> brw-rw---- 1 root disk 9, 127 Aug 13 04:17 /dev/md127
>
> /dev/md:
> lrwxrwxrwx 1 root root 8 Aug 13 04:17 TestRAID -> ../md126
> lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID1 -> ../md126p1
> lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID2 -> ../md126p2
> lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID3 -> ../md126p3
> lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID4 -> ../md126p4
> lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID5 -> ../md126p5
> lrwxrwxrwx 1 root root 8 Aug 13 04:17 imsm0 -> ../md127
>
> Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data
> captured when 'start-md' is executed):
> mdadm - v4.1 - 2018-10-01
> Output of 'mdadm --examine --scan'
> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
> ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
> member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
> Output of 'mdadm --assemble --scan --no-degraded -v'
> mdadm: looking for devices for further assembly
> mdadm: no RAID superblock on /dev/sdc1
> mdadm: /dev/sdc is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/sdc
> mdadm: no RAID superblock on /dev/sdc
> mdadm: no RAID superblock on /dev/sda5
> mdadm: no RAID superblock on /dev/sda3
> mdadm: no RAID superblock on /dev/sda2
> mdadm: no RAID superblock on /dev/sda1
> mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot 1.
> mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot 0.
> mdadm: added /dev/sdb to /dev/md/imsm0 as 1
> mdadm: added /dev/sda to /dev/md/imsm0 as 0
> mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
> mdadm: looking for devices for /dev/md/TestRAID
> mdadm: looking in container /dev/md/imsm0
> mdadm: found match on member /md127/0 in /dev/md/imsm0
> mdadm: Started /dev/md/TestRAID with 2 devices
> Output of 'dmesg | grep md:'
> md: md127 stopped.
> md: md126 stopped.
>
>
> Output of 'mdadm --examine /dev/sda'
> /dev/sda:
> Magic : Intel Raid ISM Cfg Sig.
> Version : 1.0.00
> Orig Family : a209f7f5
> Family : a209f7f5
> Generation : 0000016d
> Attributes : All supported
> UUID : 788c3635:2e37de4b:87d08323:987f57e5
> Checksum : 7cb36503 correct
> MPB Sectors : 1
> Disks : 2
> RAID Devices : 1
>
> Disk00 Serial : WD-WCANK2656465
> State : active
> Id : 00000000
> Usable Size : 488391680 (232.88 GiB 250.06 GB)
>
> [TestRAID]:
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> RAID Level : 0
> Members : 2
> Slots : [UU]
> Failed disk : none
> This Slot : 0
> Sector Size : 512
> Array Size : 976783360 (465.77 GiB 500.11 GB)
> Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
> Sector Offset : 0
> Num Stripes : 1907780
> Chunk Size : 128 KiB
> Reserved : 0
> Migrate State : idle
> Map State : normal
> Dirty State : clean
> RWH Policy : off
>
> Disk01 Serial : 3QK01SE5
> State : active
> Id : 00010000
> Usable Size : 1465143680 (698.63 GiB 750.15 GB)
>
>
> Output of 'mdadm --examine /dev/sdb'
> /dev/sdb:
> Magic : Intel Raid ISM Cfg Sig.
> Version : 1.0.00
> Orig Family : a209f7f5
> Family : a209f7f5
> Generation : 0000016d
> Attributes : All supported
> UUID : 788c3635:2e37de4b:87d08323:987f57e5
> Checksum : 7cb36503 correct
> MPB Sectors : 1
> Disks : 2
> RAID Devices : 1
>
> Disk01 Serial : 3QK01SE5
> State : active
> Id : 00010000
> Usable Size : 1465138766 (698.63 GiB 750.15 GB)
>
> [TestRAID]:
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> RAID Level : 0
> Members : 2
> Slots : [UU]
> Failed disk : none
> This Slot : 1
> Sector Size : 512
> Array Size : 976783360 (465.77 GiB 500.11 GB)
> Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
> Sector Offset : 0
> Num Stripes : 1907780
> Chunk Size : 128 KiB
> Reserved : 0
> Migrate State : idle
> Map State : normal
> Dirty State : clean
> RWH Policy : off
>
> Disk00 Serial : WD-WCANK2656465
> State : active
> Id : 00000000
> Usable Size : 488386766 (232.88 GiB 250.05 GB)
>
>
> Output of 'mdadm --examine /dev/sdc'
> /dev/sdc:
> MBR Magic : aa55
> Partition[0] : 15791832 sectors at 63 (type 0b)
>
>
> Output of 'mdadm --detail /dev/md*', if any:
> mdadm: /dev/md does not appear to be an md device
> /dev/md126:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 488391680 (465.77 GiB 500.11 GB)
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md126p1:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 32684242 (31.17 GiB 33.47 GB)
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md126p2:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 1
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md126p3:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 514048 (502.00 MiB 526.39 MB)
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md126p4:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 8032 (7.84 MiB 8.23 MB)
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md126p5:
> Container : /dev/md/imsm0, member 0
> Raid Level : raid0
> Array Size : 10040593 (9.58 GiB 10.28 GB)
> Raid Devices : 2
> Total Devices : 2
>
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 128K
>
> Consistency Policy : none
>
>
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> Number Major Minor RaidDevice State
> 1 8 0 0 active sync /dev/sda
> 0 8 16 1 active sync /dev/sdb
> /dev/md127:
> Version : imsm
> Raid Level : container
> Total Devices : 2
>
> Working Devices : 2
>
>
> UUID : 788c3635:2e37de4b:87d08323:987f57e5
> Member Arrays : /dev/md/TestRAID
>
> Number Major Minor RaidDevice
>
> - 8 0 - /dev/sda
> - 8 16 - /dev/sdb
>
> Contents of /dev/mapper directory:
> crw------- 1 root root 10, 236 Aug 13 04:17 control
>
> ============================================================
> ============================================================
> ============================================================
> ============================================================
>
> Here is the new version that doesn't work (5.15.x kernel):
>
> Contents of /sys/block:
> dm-0 dm-2 dm-5 loop0 loop2 loop4 loop6 sda sdc
> dm-1 dm-4 fd0 loop1 loop3 loop5 loop7 sdb sr0
>
> Contents of /dev:
> block mem tty20 tty58
> bsg null tty21 tty59
> btrfs-control port tty22 tty6
> bus psaux tty23 tty60
> cdrom ptmx tty24 tty61
> cdrw pts tty25 tty62
> char random tty26 tty63
> console rfkill tty27 tty7
> core rtc tty28 tty8
> cpu_dma_latency rtc0 tty29 tty9
> disk sda tty3 ttyS0
> dm-0 sda1 tty30 ttyS1
> dm-1 sda2 tty31 ttyS2
> dm-2 sda3 tty32 ttyS3
> dm-4 sda5 tty33 urandom
> dm-5 sdb tty34 vcs
> dvd sdc tty35 vcs1
> fb0 sdc1 tty36 vcs2
> fd sg0 tty37 vcs3
> fd0 sg1 tty38 vcs4
> full sg2 tty39 vcs5
> fuse sg3 tty4 vcs6
> fw0 sr0 tty40 vcsa
> hidraw0 stderr tty41 vcsa1
> hidraw1 stdin tty42 vcsa2
> hidraw2 stdout tty43 vcsa3
> hidraw3 synth tty44 vcsa4
> hpet tty tty45 vcsa5
> input tty0 tty46 vcsa6
> kmsg tty1 tty47 vcsu
> log tty10 tty48 vcsu1
> loop-control tty11 tty49 vcsu2
> loop0 tty12 tty5 vcsu3
> loop1 tty13 tty50 vcsu4
> loop2 tty14 tty51 vcsu5
> loop3 tty15 tty52 vcsu6
> loop4 tty16 tty53 vga_arbiter
> loop5 tty17 tty54 zero
> loop6 tty18 tty55
> loop7 tty19 tty56
> mapper tty2 tty57
>
> Contents of /proc/partitions:
> major minor #blocks name
>
> 2 0 4 fd0
> 11 0 1048575 sr0
> 8 16 732574584 sdb
> 8 0 244198584 sda
> 8 1 32684242 sda1
> 8 2 1 sda2
> 8 3 514048 sda3
> 8 5 2088418 sda5
> 252 0 488391936 dm-0
> 252 1 32684242 dm-1
> 252 2 514048 dm-2
> 252 4 8032 dm-4
> 252 5 10040593 dm-5
> 8 32 15687680 sdc
> 8 33 15687441 sdc1
>
> Listing of /dev/dm-x devices (LVM, dmraid, etc.):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> brw-rw---- 1 root disk 252, 0 Aug 13 04:27 /dev/dm-0
> brw-rw---- 1 root disk 252, 1 Aug 13 04:27 /dev/dm-1
> brw-rw---- 1 root disk 252, 2 Aug 13 04:27 /dev/dm-2
> brw-rw---- 1 root disk 252, 4 Aug 13 04:27 /dev/dm-4
> brw-rw---- 1 root disk 252, 5 Aug 13 04:27 /dev/dm-5
>
> Listing of /sys/block/dm-x/slaves:
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> /sys/block/dm-0/slaves:
> lrwxrwxrwx 1 root root 0 Aug 13 11:28 sda ->
> ../../../../pci0000:00/0000:00:1f.2/ata3/host2/target2:0:0/2:0:0:0/block/sda
> lrwxrwxrwx 1 root root 0 Aug 13 04:28 sdb ->
> ../../../../pci0000:00/0000:00:1f.2/ata4/host3/target3:0:0/3:0:0:0/block/sdb
>
> /sys/block/dm-1/slaves:
> lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
>
> /sys/block/dm-2/slaves:
> lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
>
> /sys/block/dm-4/slaves:
> lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
>
> /sys/block/dm-5/slaves:
> lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0
>
> Contents of /proc/mdstat (Linux software RAID status):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5]
> [raid4] [multipath]
> unused devices: <none>
>
> Contents of /run/mdadm/map (Linux software RAID arrays):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file):
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> # mdadm.conf
> #
> # Please refer to mdadm.conf(5) for information about this file.
> #
>
> # by default (built-in), scan all partitions (/proc/partitions) and all
> # containers for MD superblocks. alternatively, specify devices to scan, using
> # wildcards if desired.
> DEVICE partitions containers
>
> # automatically tag new arrays as belonging to the local system
> HOMEHOST <system>
>
> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
> ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
> member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
>
> Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data
> captured when 'start-md' is executed):
> mdadm - v4.1 - 2018-10-01
> Output of 'mdadm --examine --scan'
> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5
> ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5
> member=0 UUID=835de710:3d35bfb1:d159af46:6570f120
> Output of 'mdadm --assemble --scan --no-degraded -v'
> mdadm: looking for devices for further assembly
> mdadm: no RAID superblock on /dev/sdc1
> mdadm: /dev/sdc is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/sdc
> mdadm: no RAID superblock on /dev/sdc
> mdadm: /dev/dm-5 is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/dm-5
> mdadm: no RAID superblock on /dev/dm-5
> mdadm: /dev/dm-4 is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/dm-4
> mdadm: no RAID superblock on /dev/dm-4
> mdadm: /dev/dm-2 is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/dm-2
> mdadm: no RAID superblock on /dev/dm-2
> mdadm: /dev/dm-1 is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/dm-1
> mdadm: no RAID superblock on /dev/dm-1
> mdadm: /dev/dm-0 is not attached to Intel(R) RAID controller.
> mdadm: No OROM/EFI properties for /dev/dm-0
> mdadm: no RAID superblock on /dev/dm-0
> mdadm: no RAID superblock on /dev/sda5
> mdadm: no RAID superblock on /dev/sda3
> mdadm: no RAID superblock on /dev/sda2
> mdadm: no RAID superblock on /dev/sda1
> mdadm: /dev/sda is busy - skipping
> mdadm: /dev/sdb is busy - skipping
> mdadm: looking for devices for /dev/md/TestRAID
> mdadm: Cannot assemble mbr metadata on /dev/sdc1
> mdadm: Cannot assemble mbr metadata on /dev/sdc
> mdadm: Cannot assemble mbr metadata on /dev/dm-5
> mdadm: Cannot assemble mbr metadata on /dev/dm-4
> mdadm: Cannot assemble mbr metadata on /dev/dm-2
> mdadm: Cannot assemble mbr metadata on /dev/dm-1
> mdadm: Cannot assemble mbr metadata on /dev/dm-0
> mdadm: Cannot assemble mbr metadata on /dev/sda5
> mdadm: Cannot assemble mbr metadata on /dev/sda3
> mdadm: Cannot assemble mbr metadata on /dev/sda2
> mdadm: no recogniseable superblock on /dev/sda1
> mdadm: /dev/sda has wrong uuid.
> mdadm: /dev/sdb has wrong uuid.
> Output of 'dmesg | grep md:'
>
>
> Output of 'mdadm --examine /dev/sda'
> /dev/sda:
> Magic : Intel Raid ISM Cfg Sig.
> Version : 1.0.00
> Orig Family : a209f7f5
> Family : a209f7f5
> Generation : 0000016e
> Attributes : All supported
> UUID : 788c3635:2e37de4b:87d08323:987f57e5
> Checksum : 7cb36504 correct
> MPB Sectors : 1
> Disks : 2
> RAID Devices : 1
>
> Disk00 Serial : WD-WCANK2656465
> State : active
> Id : 00000000
> Usable Size : 488391680 (232.88 GiB 250.06 GB)
>
> [TestRAID]:
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> RAID Level : 0
> Members : 2
> Slots : [UU]
> Failed disk : none
> This Slot : 0
> Sector Size : 512
> Array Size : 976783360 (465.77 GiB 500.11 GB)
> Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
> Sector Offset : 0
> Num Stripes : 1907780
> Chunk Size : 128 KiB
> Reserved : 0
> Migrate State : idle
> Map State : normal
> Dirty State : clean
> RWH Policy : off
>
> Disk01 Serial : 3QK01SE5
> State : active
> Id : 00010000
> Usable Size : 1465143680 (698.63 GiB 750.15 GB)
>
>
> Output of 'mdadm --examine /dev/sdb'
> /dev/sdb:
> Magic : Intel Raid ISM Cfg Sig.
> Version : 1.0.00
> Orig Family : a209f7f5
> Family : a209f7f5
> Generation : 0000016e
> Attributes : All supported
> UUID : 788c3635:2e37de4b:87d08323:987f57e5
> Checksum : 7cb36504 correct
> MPB Sectors : 1
> Disks : 2
> RAID Devices : 1
>
> Disk01 Serial : 3QK01SE5
> State : active
> Id : 00010000
> Usable Size : 1465138766 (698.63 GiB 750.15 GB)
>
> [TestRAID]:
> UUID : 835de710:3d35bfb1:d159af46:6570f120
> RAID Level : 0
> Members : 2
> Slots : [UU]
> Failed disk : none
> This Slot : 1
> Sector Size : 512
> Array Size : 976783360 (465.77 GiB 500.11 GB)
> Per Dev Size : 488391944 (232.88 GiB 250.06 GB)
> Sector Offset : 0
> Num Stripes : 1907780
> Chunk Size : 128 KiB
> Reserved : 0
> Migrate State : idle
> Map State : normal
> Dirty State : clean
> RWH Policy : off
>
> Disk00 Serial : WD-WCANK2656465
> State : active
> Id : 00000000
> Usable Size : 488386766 (232.88 GiB 250.05 GB)
>
>
> Output of 'mdadm --examine /dev/sdc'
> /dev/sdc:
> MBR Magic : aa55
> Partition[0] : 31374882 sectors at 63 (type 0b)
>
>
> Output of 'mdadm --detail /dev/md*', if any:
> mdadm: cannot open /dev/md*: No such file or directory
>
> Contents of /dev/mapper directory:
> crw------- 1 root root 10, 236 Aug 13 04:27 control
> brw-rw---- 1 root disk 252, 0 Aug 13 04:27 isw_chbifgccjd_TestRAID
> brw-rw---- 1 root disk 252, 1 Aug 13 04:27
> isw_chbifgccjd_TestRAID1
> brw-rw---- 1 root disk 252, 2 Aug 13 04:27
> isw_chbifgccjd_TestRAID3
> brw-rw---- 1 root disk 252, 4 Aug 13 04:27
> isw_chbifgccjd_TestRAID4
> brw-rw---- 1 root disk 252, 5 Aug 13 04:27
> isw_chbifgccjd_TestRAID5
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: mdadm not assembling RAID?
2022-08-13 23:02 ` David F.
@ 2022-08-14 0:48 ` David F.
0 siblings, 0 replies; 3+ messages in thread
From: David F. @ 2022-08-14 0:48 UTC (permalink / raw)
To: linux-raid
What I did to make it work was add nodmraid to kernel parameter - but
since I need to use dmraid sometimes and I want to use the init script
to handle it, I removed 63-md-raid-array.rules and 85-dmraid.rules
from udev so raid doesn't try to automatically start at boot.
However, when trying to do: mdadm --assemble --scan --no-degraded -v
you end up with:
mdadm: looking for devices for further assembly
mdadm: Cannot assemble mbr metadata on /dev/sdc1
mdadm: Cannot assemble mbr metadata on /dev/sdc
mdadm: Cannot assemble mbr metadata on /dev/sda5
mdadm: Cannot assemble mbr metadata on /dev/sda3
mdadm: Cannot assemble mbr metadata on /dev/sda2
mdadm: no recongnizeable superblock on /dev/sda1
mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot 0.
mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot 1.
mdadm: added /dev/sdb to /dev/md/imsm0 as 1
mdadm: added /dev/sda to /dev/md/imsm0 as 0
mdadm: timeout waiting for /dev/md/imsm0
mdadm: looking for devices for further assembly
mdadm: looking for devices for further assembly
mdadm: Cannot assemble mbr metadata on /dev/sdc1
mdadm: Cannot assemble mbr metadata on /dev/sdc
mdadm: Cannot assemble mbr metadata on /dev/sda5
mdadm: Cannot assemble mbr metadata on /dev/sda3
mdadm: Cannot assemble mbr metadata on /dev/sda2
mdadm: no recongnizeable superblock on /dev/sda1
mdadm: /dev/sda is busy - skipping
mdadm: /dev/sdb is busy - skipping
mdadm: cannot open device /dev/md/imsm0: No such file or directory
If I left the rules in place, dmraid takes over, if I do this:
dmsetup remove_all
mdadm --assemble --scan --no-degraded -v
then it comes up fine as well.
Help!
Thanks.
On Sat, Aug 13, 2022 at 4:02 PM David F. <df7729@gmail.com> wrote:
>
> So it does have something to with the mapper devices. If I add
> "nodmraid" kernel parameter then mdadm raid works like it did before.
> I want searching config parameters for the kernel to see if I enabled
> DMRAID somehow but didn't find any. Why is the new disk/updated
> programs automatically starting dmraid? could it be udev rules .. I
> don't really understand them .. but dmraid is an option for non intel
> type raid, so what happens is dmraid is tried first for all but the
> intel items, then mdadm is used. But this is not coming from my call
> of dmraid, it's coming from someplace automatic ??
>
> On Sat, Aug 13, 2022 at 12:58 PM David F. <df7729@gmail.com> wrote:
> >
> > I have an older system with Intel RAID (striped) I use to test updates
> > to my linux boot disk to ensure the RAID is still working. I just
> > went through a big update of all the utilities and shared libraries
> > along with moving from 5.10 to 5.15 kernel. There was a stage where
> > the kernel was updated to 5.15 but all the utilities and shared
> > libraries were not (and md raid was working). The problem is now
> > there are no md devices created, but there are a bunch of dm devices
> > created that never were before.
> >
> > This group has been very helpful in the past. I hope you can help
> > with what is going on here? Does it have something to do with the
> > updated devmapper ? Should I revert?
> >
> > I put both the old working information and the new information below -
> > separated by the 3 lines of equals.
> >
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2022-08-14 0:49 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-08-13 19:58 mdadm not assembling RAID? David F.
2022-08-13 23:02 ` David F.
2022-08-14 0:48 ` David F.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.