* lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing [not found] <338941973.7699634.1473230038475.JavaMail.zimbra@redhat.com> @ 2016-09-07 6:43 ` Yi Zhang 2016-09-08 22:56 ` Shaohua Li 0 siblings, 1 reply; 10+ messages in thread From: Yi Zhang @ 2016-09-07 6:43 UTC (permalink / raw) To: linux-raid; +Cc: shli Hello I tried create one IMSM RAID10 with missing, found lots of "md: export_rdev(sde)" printed, anyone could help check it? Steps I used: mdadm -CR /dev/md0 /dev/sd[b-f] -n5 -e imsm mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing Version: 4.8.0-rc5 mdadm - v3.4-84-gbd1fd72 - 25th August 2016 Log: http://pastebin.com/FJJwvgg6 <6>[ 301.102007] md: bind<sdb> <6>[ 301.102095] md: bind<sdc> <6>[ 301.102159] md: bind<sdd> <6>[ 301.102215] md: bind<sde> <6>[ 301.102291] md: bind<sdf> <6>[ 301.103010] ata3.00: Enabling discard_zeroes_data <6>[ 311.714344] ata3.00: Enabling discard_zeroes_data <6>[ 311.721866] md: bind<sdb> <6>[ 311.721965] md: bind<sdc> <6>[ 311.722029] md: bind<sdd> <5>[ 311.733165] md/raid10:md127: not clean -- starting background reconstruction <6>[ 311.733167] md/raid10:md127: active with 3 out of 4 devices <6>[ 311.733186] md127: detected capacity change from 0 to 240060989440 <6>[ 311.774027] md: bind<sde> <6>[ 311.810664] md: md127 switched to read-write mode. <6>[ 311.819885] md: resync of RAID array md127 <6>[ 311.819886] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. <6>[ 311.819887] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for resync. <6>[ 311.819891] md: using 128k window, over a total of 234435328k. <6>[ 316.606073] ata3.00: Enabling discard_zeroes_data <6>[ 343.949845] capability: warning: `turbostat' uses 32-bit capabilities (legacy support in use) <6>[ 1482.314944] md: md127: resync done. <7>[ 1482.315086] RAID10 conf printout: <7>[ 1482.315087] --- wd:3 rd:4 <7>[ 1482.315089] disk 0, wo:0, o:1, dev:sdb <7>[ 1482.315089] disk 1, wo:0, o:1, dev:sdc <7>[ 1482.315090] disk 2, wo:0, o:1, dev:sdd <7>[ 1482.315099] RAID10 conf printout: <7>[ 1482.315099] --- wd:3 rd:4 <7>[ 1482.315100] disk 0, wo:0, o:1, dev:sdb <7>[ 1482.315100] disk 1, wo:0, o:1, dev:sdc <7>[ 1482.315101] disk 2, wo:0, o:1, dev:sdd <7>[ 1482.315101] disk 3, wo:1, o:1, dev:sde <6>[ 1482.315220] md: recovery of RAID array md127 <6>[ 1482.315221] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. <6>[ 1482.315222] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. <6>[ 1482.315227] md: using 128k window, over a total of 117217664k. <6>[ 2697.184217] md: md127: recovery done. <7>[ 2697.524143] RAID10 conf printout: <7>[ 2697.524144] --- wd:4 rd:4 <7>[ 2697.524146] disk 0, wo:0, o:1, dev:sdb <7>[ 2697.524146] disk 1, wo:0, o:1, dev:sdc <7>[ 2697.524147] disk 2, wo:0, o:1, dev:sdd <7>[ 2697.524148] disk 3, wo:0, o:1, dev:sde <6>[ 2697.524632] md: export_rdev(sde) <6>[ 2697.549452] md: export_rdev(sde) <6>[ 2697.568763] md: export_rdev(sde) <6>[ 2697.587938] md: export_rdev(sde) <6>[ 2697.607271] md: export_rdev(sde) <6>[ 2697.626321] md: export_rdev(sde) <6>[ 2697.645676] md: export_rdev(sde) <6>[ 2697.663211] md: export_rdev(sde) <6>[ 2697.681603] md: export_rdev(sde) <6>[ 2697.699117] md: export_rdev(sde) <6>[ 2697.716510] md: export_rdev(sde) Best Regards, Yi Zhang ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-07 6:43 ` lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing Yi Zhang @ 2016-09-08 22:56 ` Shaohua Li 2016-09-09 12:56 ` Artur Paszkiewicz 0 siblings, 1 reply; 10+ messages in thread From: Shaohua Li @ 2016-09-08 22:56 UTC (permalink / raw) To: Yi Zhang; +Cc: linux-raid, Jes.Sorensen On Wed, Sep 07, 2016 at 02:43:41AM -0400, Yi Zhang wrote: > Hello > > I tried create one IMSM RAID10 with missing, found lots of "md: export_rdev(sde)" printed, anyone could help check it? > > Steps I used: > mdadm -CR /dev/md0 /dev/sd[b-f] -n5 -e imsm > mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing > > Version: > 4.8.0-rc5 > mdadm - v3.4-84-gbd1fd72 - 25th August 2016 can't reproduce with old mdadm but can with upstream mdadm. Looks mdadm is keeping write the new_dev sysfs entry. Jes, any idea? Thanks, Shaohua > Log: > http://pastebin.com/FJJwvgg6 > > <6>[ 301.102007] md: bind<sdb> > <6>[ 301.102095] md: bind<sdc> > <6>[ 301.102159] md: bind<sdd> > <6>[ 301.102215] md: bind<sde> > <6>[ 301.102291] md: bind<sdf> > <6>[ 301.103010] ata3.00: Enabling discard_zeroes_data > <6>[ 311.714344] ata3.00: Enabling discard_zeroes_data > <6>[ 311.721866] md: bind<sdb> > <6>[ 311.721965] md: bind<sdc> > <6>[ 311.722029] md: bind<sdd> > <5>[ 311.733165] md/raid10:md127: not clean -- starting background reconstruction > <6>[ 311.733167] md/raid10:md127: active with 3 out of 4 devices > <6>[ 311.733186] md127: detected capacity change from 0 to 240060989440 > <6>[ 311.774027] md: bind<sde> > <6>[ 311.810664] md: md127 switched to read-write mode. > <6>[ 311.819885] md: resync of RAID array md127 > <6>[ 311.819886] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. > <6>[ 311.819887] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for resync. > <6>[ 311.819891] md: using 128k window, over a total of 234435328k. > <6>[ 316.606073] ata3.00: Enabling discard_zeroes_data > <6>[ 343.949845] capability: warning: `turbostat' uses 32-bit capabilities (legacy support in use) > <6>[ 1482.314944] md: md127: resync done. > <7>[ 1482.315086] RAID10 conf printout: > <7>[ 1482.315087] --- wd:3 rd:4 > <7>[ 1482.315089] disk 0, wo:0, o:1, dev:sdb > <7>[ 1482.315089] disk 1, wo:0, o:1, dev:sdc > <7>[ 1482.315090] disk 2, wo:0, o:1, dev:sdd > <7>[ 1482.315099] RAID10 conf printout: > <7>[ 1482.315099] --- wd:3 rd:4 > <7>[ 1482.315100] disk 0, wo:0, o:1, dev:sdb > <7>[ 1482.315100] disk 1, wo:0, o:1, dev:sdc > <7>[ 1482.315101] disk 2, wo:0, o:1, dev:sdd > <7>[ 1482.315101] disk 3, wo:1, o:1, dev:sde > <6>[ 1482.315220] md: recovery of RAID array md127 > <6>[ 1482.315221] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. > <6>[ 1482.315222] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. > <6>[ 1482.315227] md: using 128k window, over a total of 117217664k. > <6>[ 2697.184217] md: md127: recovery done. > <7>[ 2697.524143] RAID10 conf printout: > <7>[ 2697.524144] --- wd:4 rd:4 > <7>[ 2697.524146] disk 0, wo:0, o:1, dev:sdb > <7>[ 2697.524146] disk 1, wo:0, o:1, dev:sdc > <7>[ 2697.524147] disk 2, wo:0, o:1, dev:sdd > <7>[ 2697.524148] disk 3, wo:0, o:1, dev:sde > <6>[ 2697.524632] md: export_rdev(sde) > <6>[ 2697.549452] md: export_rdev(sde) > <6>[ 2697.568763] md: export_rdev(sde) > <6>[ 2697.587938] md: export_rdev(sde) > <6>[ 2697.607271] md: export_rdev(sde) > <6>[ 2697.626321] md: export_rdev(sde) > <6>[ 2697.645676] md: export_rdev(sde) > <6>[ 2697.663211] md: export_rdev(sde) > <6>[ 2697.681603] md: export_rdev(sde) > <6>[ 2697.699117] md: export_rdev(sde) > <6>[ 2697.716510] md: export_rdev(sde) > > Best Regards, > Yi Zhang > > > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-08 22:56 ` Shaohua Li @ 2016-09-09 12:56 ` Artur Paszkiewicz 2016-09-12 8:03 ` Yi Zhang 2016-09-14 21:05 ` Jes Sorensen 0 siblings, 2 replies; 10+ messages in thread From: Artur Paszkiewicz @ 2016-09-09 12:56 UTC (permalink / raw) To: Shaohua Li, Yi Zhang; +Cc: linux-raid, Jes.Sorensen On 09/09/2016 12:56 AM, Shaohua Li wrote: > On Wed, Sep 07, 2016 at 02:43:41AM -0400, Yi Zhang wrote: >> Hello >> >> I tried create one IMSM RAID10 with missing, found lots of "md: export_rdev(sde)" printed, anyone could help check it? >> >> Steps I used: >> mdadm -CR /dev/md0 /dev/sd[b-f] -n5 -e imsm >> mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing >> >> Version: >> 4.8.0-rc5 >> mdadm - v3.4-84-gbd1fd72 - 25th August 2016 > > can't reproduce with old mdadm but can with upstream mdadm. Looks mdadm is > keeping write the new_dev sysfs entry. > > Jes, any idea? > > Thanks, > Shaohua >> Log: >> http://pastebin.com/FJJwvgg6 >> >> <6>[ 301.102007] md: bind<sdb> >> <6>[ 301.102095] md: bind<sdc> >> <6>[ 301.102159] md: bind<sdd> >> <6>[ 301.102215] md: bind<sde> >> <6>[ 301.102291] md: bind<sdf> >> <6>[ 301.103010] ata3.00: Enabling discard_zeroes_data >> <6>[ 311.714344] ata3.00: Enabling discard_zeroes_data >> <6>[ 311.721866] md: bind<sdb> >> <6>[ 311.721965] md: bind<sdc> >> <6>[ 311.722029] md: bind<sdd> >> <5>[ 311.733165] md/raid10:md127: not clean -- starting background reconstruction >> <6>[ 311.733167] md/raid10:md127: active with 3 out of 4 devices >> <6>[ 311.733186] md127: detected capacity change from 0 to 240060989440 >> <6>[ 311.774027] md: bind<sde> >> <6>[ 311.810664] md: md127 switched to read-write mode. >> <6>[ 311.819885] md: resync of RAID array md127 >> <6>[ 311.819886] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >> <6>[ 311.819887] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for resync. >> <6>[ 311.819891] md: using 128k window, over a total of 234435328k. >> <6>[ 316.606073] ata3.00: Enabling discard_zeroes_data >> <6>[ 343.949845] capability: warning: `turbostat' uses 32-bit capabilities (legacy support in use) >> <6>[ 1482.314944] md: md127: resync done. >> <7>[ 1482.315086] RAID10 conf printout: >> <7>[ 1482.315087] --- wd:3 rd:4 >> <7>[ 1482.315089] disk 0, wo:0, o:1, dev:sdb >> <7>[ 1482.315089] disk 1, wo:0, o:1, dev:sdc >> <7>[ 1482.315090] disk 2, wo:0, o:1, dev:sdd >> <7>[ 1482.315099] RAID10 conf printout: >> <7>[ 1482.315099] --- wd:3 rd:4 >> <7>[ 1482.315100] disk 0, wo:0, o:1, dev:sdb >> <7>[ 1482.315100] disk 1, wo:0, o:1, dev:sdc >> <7>[ 1482.315101] disk 2, wo:0, o:1, dev:sdd >> <7>[ 1482.315101] disk 3, wo:1, o:1, dev:sde >> <6>[ 1482.315220] md: recovery of RAID array md127 >> <6>[ 1482.315221] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >> <6>[ 1482.315222] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. >> <6>[ 1482.315227] md: using 128k window, over a total of 117217664k. >> <6>[ 2697.184217] md: md127: recovery done. >> <7>[ 2697.524143] RAID10 conf printout: >> <7>[ 2697.524144] --- wd:4 rd:4 >> <7>[ 2697.524146] disk 0, wo:0, o:1, dev:sdb >> <7>[ 2697.524146] disk 1, wo:0, o:1, dev:sdc >> <7>[ 2697.524147] disk 2, wo:0, o:1, dev:sdd >> <7>[ 2697.524148] disk 3, wo:0, o:1, dev:sde >> <6>[ 2697.524632] md: export_rdev(sde) >> <6>[ 2697.549452] md: export_rdev(sde) >> <6>[ 2697.568763] md: export_rdev(sde) >> <6>[ 2697.587938] md: export_rdev(sde) >> <6>[ 2697.607271] md: export_rdev(sde) >> <6>[ 2697.626321] md: export_rdev(sde) >> <6>[ 2697.645676] md: export_rdev(sde) >> <6>[ 2697.663211] md: export_rdev(sde) >> <6>[ 2697.681603] md: export_rdev(sde) >> <6>[ 2697.699117] md: export_rdev(sde) >> <6>[ 2697.716510] md: export_rdev(sde) >> >> Best Regards, >> Yi Zhang Can you check if this fix works for you? If it does I'll send a proper patch for this. Thanks, Artur diff --git a/super-intel.c b/super-intel.c index 92817e9..ffa71f6 100644 --- a/super-intel.c +++ b/super-intel.c @@ -7789,6 +7789,9 @@ static struct mdinfo *imsm_activate_spare(struct active_array *a, IMSM_T_STATE_DEGRADED) return NULL; + if (get_imsm_map(dev, MAP_0)->map_state == IMSM_T_STATE_UNINITIALIZED) + return NULL; + /* * If there are any failed disks check state of the other volume. * Block rebuild if the another one is failed until failed disks ^ permalink raw reply related [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-09 12:56 ` Artur Paszkiewicz @ 2016-09-12 8:03 ` Yi Zhang 2016-09-12 10:58 ` Artur Paszkiewicz 2016-09-14 21:05 ` Jes Sorensen 1 sibling, 1 reply; 10+ messages in thread From: Yi Zhang @ 2016-09-12 8:03 UTC (permalink / raw) To: Artur Paszkiewicz, Shaohua Li; +Cc: linux-raid, Jes.Sorensen On 09/09/2016 08:56 PM, Artur Paszkiewicz wrote: > On 09/09/2016 12:56 AM, Shaohua Li wrote: >> On Wed, Sep 07, 2016 at 02:43:41AM -0400, Yi Zhang wrote: >>> Hello >>> >>> I tried create one IMSM RAID10 with missing, found lots of "md: export_rdev(sde)" printed, anyone could help check it? >>> >>> Steps I used: >>> mdadm -CR /dev/md0 /dev/sd[b-f] -n5 -e imsm >>> mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing >>> >>> Version: >>> 4.8.0-rc5 >>> mdadm - v3.4-84-gbd1fd72 - 25th August 2016 >> can't reproduce with old mdadm but can with upstream mdadm. Looks mdadm is >> keeping write the new_dev sysfs entry. >> >> Jes, any idea? >> >> Thanks, >> Shaohua >>> Log: >>> http://pastebin.com/FJJwvgg6 >>> >>> <6>[ 301.102007] md: bind<sdb> >>> <6>[ 301.102095] md: bind<sdc> >>> <6>[ 301.102159] md: bind<sdd> >>> <6>[ 301.102215] md: bind<sde> >>> <6>[ 301.102291] md: bind<sdf> >>> <6>[ 301.103010] ata3.00: Enabling discard_zeroes_data >>> <6>[ 311.714344] ata3.00: Enabling discard_zeroes_data >>> <6>[ 311.721866] md: bind<sdb> >>> <6>[ 311.721965] md: bind<sdc> >>> <6>[ 311.722029] md: bind<sdd> >>> <5>[ 311.733165] md/raid10:md127: not clean -- starting background reconstruction >>> <6>[ 311.733167] md/raid10:md127: active with 3 out of 4 devices >>> <6>[ 311.733186] md127: detected capacity change from 0 to 240060989440 >>> <6>[ 311.774027] md: bind<sde> >>> <6>[ 311.810664] md: md127 switched to read-write mode. >>> <6>[ 311.819885] md: resync of RAID array md127 >>> <6>[ 311.819886] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >>> <6>[ 311.819887] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for resync. >>> <6>[ 311.819891] md: using 128k window, over a total of 234435328k. >>> <6>[ 316.606073] ata3.00: Enabling discard_zeroes_data >>> <6>[ 343.949845] capability: warning: `turbostat' uses 32-bit capabilities (legacy support in use) >>> <6>[ 1482.314944] md: md127: resync done. >>> <7>[ 1482.315086] RAID10 conf printout: >>> <7>[ 1482.315087] --- wd:3 rd:4 >>> <7>[ 1482.315089] disk 0, wo:0, o:1, dev:sdb >>> <7>[ 1482.315089] disk 1, wo:0, o:1, dev:sdc >>> <7>[ 1482.315090] disk 2, wo:0, o:1, dev:sdd >>> <7>[ 1482.315099] RAID10 conf printout: >>> <7>[ 1482.315099] --- wd:3 rd:4 >>> <7>[ 1482.315100] disk 0, wo:0, o:1, dev:sdb >>> <7>[ 1482.315100] disk 1, wo:0, o:1, dev:sdc >>> <7>[ 1482.315101] disk 2, wo:0, o:1, dev:sdd >>> <7>[ 1482.315101] disk 3, wo:1, o:1, dev:sde >>> <6>[ 1482.315220] md: recovery of RAID array md127 >>> <6>[ 1482.315221] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >>> <6>[ 1482.315222] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. >>> <6>[ 1482.315227] md: using 128k window, over a total of 117217664k. >>> <6>[ 2697.184217] md: md127: recovery done. >>> <7>[ 2697.524143] RAID10 conf printout: >>> <7>[ 2697.524144] --- wd:4 rd:4 >>> <7>[ 2697.524146] disk 0, wo:0, o:1, dev:sdb >>> <7>[ 2697.524146] disk 1, wo:0, o:1, dev:sdc >>> <7>[ 2697.524147] disk 2, wo:0, o:1, dev:sdd >>> <7>[ 2697.524148] disk 3, wo:0, o:1, dev:sde >>> <6>[ 2697.524632] md: export_rdev(sde) >>> <6>[ 2697.549452] md: export_rdev(sde) >>> <6>[ 2697.568763] md: export_rdev(sde) >>> <6>[ 2697.587938] md: export_rdev(sde) >>> <6>[ 2697.607271] md: export_rdev(sdeautomate) >>> <6>[ 2697.626321] md: export_rdev(sdeautomateautomate) >>> <6>[ 2697.645676] md: export_rdev(sde) >>> <6>[ 2697.663211] md: export_rdev(sde) >>> <6>[ 2697.681603] md: export_rdev(sde) >>> <6>[ 2697.699117] md: export_rdev(sde) >>> <6>[ 2697.716510] md: export_rdev(sde) >>> >>> Best Regards, >>> Yi Zhang > Can you check if this fix works for you? If it does I'll send a proper > patch for this. Hello Artur With your patch, no "md: export_rdev(sde)" printed after create raid10. I found another problem, not sure whether it is reasonable, could you help confirm it, thanks. When I create one container with 4 disks[1], and create one raid10 with 3 disks(sd[b-d]) + 1 missing [2], but it finally bind the fourth disk: sde [3]. [1] mdadm -CR /dev/md0 /dev/sd[b-e] -n4 -e imsm [2] mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing --size=500M [3] # cat /proc/mdstat Personalities : [raid10] md127 : active raid10 sde[4] sdd[2] sdc[1] sdb[0] 1024000 blocks super external:/md0/0 128K chunks 2 near-copies [4/4] [UUUU] md0 : inactive sde[3](S) sdd[2](S) sdc[1](S) sdb[0](S) 4420 blocks super external:imsm unused devices: <none> > Thanks, > Artur > > diff --git a/super-intel.c b/super-intel.c > index 92817e9..ffa71f6 100644 > --- a/super-intel.c > +++ b/super-intel.c > @@ -7789,6 +7789,9 @@ static struct mdinfo *imsm_activate_spare(struct active_array *a, > IMSM_T_STATE_DEGRADED) > return NULL; > > + if (get_imsm_map(dev, MAP_0)->map_state == IMSM_T_STATE_UNINITIALIZED) > + return NULL; > + > /* > * If there are any failed disks check state of the other volume. > * Block rebuild if the another one is failed until failed disks ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-12 8:03 ` Yi Zhang @ 2016-09-12 10:58 ` Artur Paszkiewicz 2016-09-14 9:24 ` Yi Zhang 0 siblings, 1 reply; 10+ messages in thread From: Artur Paszkiewicz @ 2016-09-12 10:58 UTC (permalink / raw) To: Yi Zhang, Shaohua Li; +Cc: linux-raid, Jes.Sorensen On 09/12/2016 10:03 AM, Yi Zhang wrote: > Hello Artur > With your patch, no "md: export_rdev(sde)" printed after create raid10. > > I found another problem, not sure whether it is reasonable, could you help confirm it, thanks. > When I create one container with 4 disks[1], and create one raid10 with 3 disks(sd[b-d]) + 1 missing [2], but it finally bind the fourth disk: sde [3]. > > [1] mdadm -CR /dev/md0 /dev/sd[b-e] -n4 -e imsm > [2] mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing --size=500M > [3] # cat /proc/mdstat > Personalities : [raid10] > md127 : active raid10 sde[4] sdd[2] sdc[1] sdb[0] > 1024000 blocks super external:/md0/0 128K chunks 2 near-copies [4/4] [UUUU] > > md0 : inactive sde[3](S) sdd[2](S) sdc[1](S) sdb[0](S) > 4420 blocks super external:imsm > > unused devices: <none> I think that this is correct behavior. Because there is a spare disk available in the container, it is used for rebuilding the volume. This is equivalent to: mdadm -CR /dev/md0 /dev/sd[b-d] -n3 -e imsm mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing --size=500M mdadm -a /dev/md0 /dev/sde ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-12 10:58 ` Artur Paszkiewicz @ 2016-09-14 9:24 ` Yi Zhang 0 siblings, 0 replies; 10+ messages in thread From: Yi Zhang @ 2016-09-14 9:24 UTC (permalink / raw) To: Artur Paszkiewicz; +Cc: Shaohua Li, Jes.Sorensen, linux-raid On 09/12/2016 06:58 PM, Artur Paszkiewicz wrote: > On 09/12/2016 10:03 AM, Yi Zhang wrote: >> Hello Artur >> With your patch, no "md: export_rdev(sde)" printed after create raid10. >> >> I found another problem, not sure whether it is reasonable, could you help confirm it, thanks. >> When I create one container with 4 disks[1], and create one raid10 with 3 disks(sd[b-d]) + 1 missing [2], but it finally bind the fourth disk: sde [3]. >> >> [1] mdadm -CR /dev/md0 /dev/sd[b-e] -n4 -e imsm >> [2] mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing --size=500M >> [3] # cat /proc/mdstat >> Personalities : [raid10] >> md127 : active raid10 sde[4] sdd[2] sdc[1] sdb[0] >> 1024000 blocks super external:/md0/0 128K chunks 2 near-copies [4/4] [UUUU] >> >> md0 : inactive sde[3](S) sdd[2](S) sdc[1](S) sdb[0](S) >> 4420 blocks super external:imsm >> >> unused devices: <none> > I think that this is correct behavior. Because there is a spare disk > available in the container, it is used for rebuilding the volume. This > is equivalent to: > > mdadm -CR /dev/md0 /dev/sd[b-d] -n3 -e imsm > mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing --size=500M > mdadm -a /dev/md0 /dev/sde got, thanks Artur for the confirmation. Yi > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-09 12:56 ` Artur Paszkiewicz 2016-09-12 8:03 ` Yi Zhang @ 2016-09-14 21:05 ` Jes Sorensen 2016-09-15 8:01 ` Artur Paszkiewicz 1 sibling, 1 reply; 10+ messages in thread From: Jes Sorensen @ 2016-09-14 21:05 UTC (permalink / raw) To: Artur Paszkiewicz; +Cc: Shaohua Li, Yi Zhang, linux-raid Artur Paszkiewicz <artur.paszkiewicz@intel.com> writes: > On 09/09/2016 12:56 AM, Shaohua Li wrote: >> On Wed, Sep 07, 2016 at 02:43:41AM -0400, Yi Zhang wrote: >>> Hello >>> >>> I tried create one IMSM RAID10 with missing, found lots of "md: >>> export_rdev(sde)" printed, anyone could help check it? >>> >>> Steps I used: >>> mdadm -CR /dev/md0 /dev/sd[b-f] -n5 -e imsm >>> mdadm -CR /dev/md/Volume0 -l10 -n4 /dev/sd[b-d] missing >>> >>> Version: >>> 4.8.0-rc5 >>> mdadm - v3.4-84-gbd1fd72 - 25th August 2016 >> >> can't reproduce with old mdadm but can with upstream mdadm. Looks mdadm is >> keeping write the new_dev sysfs entry. >> >> Jes, any idea? >> >> Thanks, >> Shaohua [snip] > Can you check if this fix works for you? If it does I'll send a proper > patch for this. > > Thanks, > Artur Artur, You were too fast :) Did you intend to post a patch with a commit message? Cheers, Jes > > diff --git a/super-intel.c b/super-intel.c > index 92817e9..ffa71f6 100644 > --- a/super-intel.c > +++ b/super-intel.c > @@ -7789,6 +7789,9 @@ static struct mdinfo *imsm_activate_spare(struct active_array *a, > IMSM_T_STATE_DEGRADED) > return NULL; > > + if (get_imsm_map(dev, MAP_0)->map_state == IMSM_T_STATE_UNINITIALIZED) > + return NULL; > + > /* > * If there are any failed disks check state of the other volume. > * Block rebuild if the another one is failed until failed disks > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-14 21:05 ` Jes Sorensen @ 2016-09-15 8:01 ` Artur Paszkiewicz 2016-09-16 12:31 ` Jes Sorensen 0 siblings, 1 reply; 10+ messages in thread From: Artur Paszkiewicz @ 2016-09-15 8:01 UTC (permalink / raw) To: Jes Sorensen; +Cc: Shaohua Li, Yi Zhang, linux-raid On 09/14/2016 11:05 PM, Jes Sorensen wrote: > Artur, > > You were too fast :) Did you intend to post a patch with a commit > message? > > Cheers, > Jes > Hi Jes, I wanted to wait for feedback from Yi first. I just sent a proper patch with a commit message. Artur ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-15 8:01 ` Artur Paszkiewicz @ 2016-09-16 12:31 ` Jes Sorensen 2016-09-18 2:53 ` Yi Zhang 0 siblings, 1 reply; 10+ messages in thread From: Jes Sorensen @ 2016-09-16 12:31 UTC (permalink / raw) To: Artur Paszkiewicz; +Cc: Shaohua Li, Yi Zhang, linux-raid Artur Paszkiewicz <artur.paszkiewicz@intel.com> writes: > On 09/14/2016 11:05 PM, Jes Sorensen wrote: >> Artur, >> >> You were too fast :) Did you intend to post a patch with a commit >> message? >> >> Cheers, >> Jes >> > > Hi Jes, > > I wanted to wait for feedback from Yi first. I just sent a proper patch > with a commit message. Thats great, much appreciated! Jes ^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing 2016-09-16 12:31 ` Jes Sorensen @ 2016-09-18 2:53 ` Yi Zhang 0 siblings, 0 replies; 10+ messages in thread From: Yi Zhang @ 2016-09-18 2:53 UTC (permalink / raw) To: Jes Sorensen, Artur Paszkiewicz; +Cc: Shaohua Li, linux-raid Hello Artur I have test the patch you provided and that works, thanks. Best Regards, Yi Zhang ----- Original Message ----- From: "Jes Sorensen" <Jes.Sorensen@redhat.com> To: "Artur Paszkiewicz" <artur.paszkiewicz@intel.com> Cc: "Shaohua Li" <shli@kernel.org>, "Yi Zhang" <yizhan@redhat.com>, linux-raid@vger.kernel.org Sent: Friday, September 16, 2016 8:31:50 PM Subject: Re: lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing Artur Paszkiewicz <artur.paszkiewicz@intel.com> writes: > On 09/14/2016 11:05 PM, Jes Sorensen wrote: >> Artur, >> >> You were too fast :) Did you intend to post a patch with a commit >> message? >> >> Cheers, >> Jes >> > > Hi Jes, > > I wanted to wait for feedback from Yi first. I just sent a proper patch > with a commit message. Thats great, much appreciated! Jes -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 10+ messages in thread
end of thread, other threads:[~2016-09-18 2:53 UTC | newest] Thread overview: 10+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- [not found] <338941973.7699634.1473230038475.JavaMail.zimbra@redhat.com> 2016-09-07 6:43 ` lots of "md: export_rdev(sde)" printed after create IMSM RAID10 with missing Yi Zhang 2016-09-08 22:56 ` Shaohua Li 2016-09-09 12:56 ` Artur Paszkiewicz 2016-09-12 8:03 ` Yi Zhang 2016-09-12 10:58 ` Artur Paszkiewicz 2016-09-14 9:24 ` Yi Zhang 2016-09-14 21:05 ` Jes Sorensen 2016-09-15 8:01 ` Artur Paszkiewicz 2016-09-16 12:31 ` Jes Sorensen 2016-09-18 2:53 ` Yi Zhang
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.