All of lore.kernel.org
 help / color / mirror / Atom feed
* [linux-lvm] Problems with vgimport after software raid initialisation failed.
@ 2003-09-30 15:00 SystemError
  2003-10-02  4:23 ` Heinz J . Mauelshagen
  0 siblings, 1 reply; 5+ messages in thread
From: SystemError @ 2003-09-30 15:00 UTC (permalink / raw)
  To: linux-lvm

Hello out there,

after I migrating my precious volume group "datavg" from unmirrored
disks to linux software raid devices I ran into serios problems.
(Although I fear the biggest problem here was my own incompetence...)

First I moved the data from the old unmirrored disks away, using pvmove.
No Problems so far.

At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
So I did a vgreduce on them, then created a new raid1 
"/dev/md4" (containing both "hdf" and "hdh") and added it to my 
volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
No Problems so far.

Everything looked soooo perfect and so I decided to reboot the system...

At this point things started to go wrong, during the boot sequence
"/dev/md4" was not automatically activated and suddenly the PV
"/dev/hdf" showed up in "datavg", "/dev/md4" was gone.

Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
initialisation of "/dev/md4", and rebooted again. 
This was a probably a baaad idea. 
Shame upon me.

Now my pvscan looks like this:
"
[root@athens root]# pvscan 
pvscan -- reading all physical volumes (this may take a while...)
pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
0 free]
pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
(run vgscan)
pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
50.88 GB free]
pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
"

Or with the -u option:
"
[root@athens root]# pvscan -u
pvscan -- reading all physical volumes (this may take a while...)
pvscan -- ACTIVE   PV "/dev/md2" with UUID
"g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
10 GB free]
pvscan -- inactive PV "/dev/md3" with UUID
"R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
[132.25 GB / 0 free]
pvscan -- inactive PV "/dev/md4" with UUID
"szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
[57.12 GB / 50.88 GB free]
pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
pvscan -- inactive PV "/dev/hdf" with UUID
"szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
[57.12 GB / 50.88 GB free]
pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]

"

A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
"
[root@athens root]# vgimport datavg /dev/md3 /dev/md4
vgimport -- ERROR "pv_read(): multiple device" reading physical volume
"/dev/md4"
"

Using "md3" and "hdh" also fails:
"
[root@athens root]# vgimport datavg /dev/md3 /dev/hdh
vgimport -- ERROR "pv_read(): multiple device" reading physical volume
"/dev/hdh"
"

It also fails when I try to use "hdf", only the error message is
different:
"
[root@athens root]# vgimport datavg /dev/md3 /dev/hdf
vgimport -- ERROR: wrong number of physical volumes to import volume
group "datavg"
"

So here I am, with a huge VG an tons of data in it but no way to access
the VG. Has anybody out there an idea how I can still access
the data of datavg ?

By the way: 
I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
as provided by RedHat.

Bye
In desperation
Lutz Reinegger

PS:
Any comments and suggestions are highly appreciated, even if those
suggestions include the use of hex editors or sacrificing 
caffeine to dark and ancient deities.
;-)

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [linux-lvm] Problems with vgimport after software raid initialisation failed.
  2003-09-30 15:00 [linux-lvm] Problems with vgimport after software raid initialisation failed SystemError
@ 2003-10-02  4:23 ` Heinz J . Mauelshagen
  2003-10-02  5:58   ` SystemError
  0 siblings, 1 reply; 5+ messages in thread
From: Heinz J . Mauelshagen @ 2003-10-02  4:23 UTC (permalink / raw)
  To: linux-lvm

Lutz,

looks like you hit some strange LVM on top of MD bug :(

In order to get your VG active again (which of course is your highest
priority) and before we analyze the vgscan problem, you want to go for
the follwoing workaround (presumably /etc/lvmconf/datavg.conf is the last
correct archive of the metadata):

# cp /etc/lvmconf/datavg.conf /etc/lvmtab.d/datavg
# echo -ne "datavg\0" >> /etc/lvmtab
# vgchange -ay datavg

Warning: the next vgscan run will remove the above metadata again, so avoid
         running it for now by commenting it out in your boot script.

So far about firefighting ;)


For further analysis, please do the following and send the resulting
bzip2'ed tar archive containing your metadat to me in private mail
<mge@sistina.com>:

# for d in md2 md3 md4 hdf hdh
# do
#	dd bs=1k count=4k if=/dev/$d of=$d.vgda
# done
# tar cf Lutz_Reinegger.vgda.tar *.vgda
# rm *.vgda
# bzip2 Lutz_Reinegger.vgda.tar


Regards,
Heinz    -- The LVM Guy --



On Tue, Sep 30, 2003 at 09:59:19PM +0200, SystemError wrote:
> Hello out there,
> 
> after I migrating my precious volume group "datavg" from unmirrored
> disks to linux software raid devices I ran into serios problems.
> (Although I fear the biggest problem here was my own incompetence...)
> 
> First I moved the data from the old unmirrored disks away, using pvmove.
> No Problems so far.
> 
> At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
> So I did a vgreduce on them, then created a new raid1 
> "/dev/md4" (containing both "hdf" and "hdh") and added it to my 
> volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
> No Problems so far.
> 
> Everything looked soooo perfect and so I decided to reboot the system...
> 
> At this point things started to go wrong, during the boot sequence
> "/dev/md4" was not automatically activated and suddenly the PV
> "/dev/hdf" showed up in "datavg", "/dev/md4" was gone.
> 
> Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
> initialisation of "/dev/md4", and rebooted again. 
> This was a probably a baaad idea. 
> Shame upon me.
> 
> Now my pvscan looks like this:
> "
> [root@athens root]# pvscan 
> pvscan -- reading all physical volumes (this may take a while...)
> pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
> pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
> 0 free]
> pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
> (run vgscan)
> pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
> 50.88 GB free]
> pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> "
> 
> Or with the -u option:
> "
> [root@athens root]# pvscan -u
> pvscan -- reading all physical volumes (this may take a while...)
> pvscan -- ACTIVE   PV "/dev/md2" with UUID
> "g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
> 10 GB free]
> pvscan -- inactive PV "/dev/md3" with UUID
> "R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
> [132.25 GB / 0 free]
> pvscan -- inactive PV "/dev/md4" with UUID
> "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> [57.12 GB / 50.88 GB free]
> pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> pvscan -- inactive PV "/dev/hdf" with UUID
> "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> [57.12 GB / 50.88 GB free]
> pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> 
> "
> 
> A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
> "
> [root@athens root]# vgimport datavg /dev/md3 /dev/md4
> vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> "/dev/md4"
> "
> 
> Using "md3" and "hdh" also fails:
> "
> [root@athens root]# vgimport datavg /dev/md3 /dev/hdh
> vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> "/dev/hdh"
> "
> 
> It also fails when I try to use "hdf", only the error message is
> different:
> "
> [root@athens root]# vgimport datavg /dev/md3 /dev/hdf
> vgimport -- ERROR: wrong number of physical volumes to import volume
> group "datavg"
> "
> 
> So here I am, with a huge VG an tons of data in it but no way to access
> the VG. Has anybody out there an idea how I can still access
> the data of datavg ?
> 
> By the way: 
> I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
> as provided by RedHat.
> 
> Bye
> In desperation
> Lutz Reinegger
> 
> PS:
> Any comments and suggestions are highly appreciated, even if those
> suggestions include the use of hex editors or sacrificing 
> caffeine to dark and ancient deities.
> ;-)
> 
> 
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/

*** Software bugs are stupid.
    Nevertheless it needs not so stupid people to solve them ***

=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-

Heinz Mauelshagen                                 Sistina Software Inc.
Senior Consultant/Developer                       Am Sonnenhang 11
                                                  56242 Marienrachdorf
                                                  Germany
Mauelshagen@Sistina.com                           +49 2626 141200
                                                       FAX 924446
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [linux-lvm] Problems with vgimport after software raid initialisation failed.
  2003-10-02  4:23 ` Heinz J . Mauelshagen
@ 2003-10-02  5:58   ` SystemError
  2003-10-02  9:03     ` Heinz J . Mauelshagen
  0 siblings, 1 reply; 5+ messages in thread
From: SystemError @ 2003-10-02  5:58 UTC (permalink / raw)
  To: linux-lvm

Heinz,

I did as instructed and now I am able to access my 
precious "datavg" again.
Thanks a lot Heinz.
You really saved my day. :-)

I also edited /etc/rc.d/rc.sysinit and commented out all vgscans
(RedHat does 2 of them), and added a hard coded "vgchange -ay sysvg" in
order to bring my other VG with the /usr, /var, etc... 
filesystems online during boot time.

Immediately after another reboot I tried to activate "datavg" using
"vgchange -ay datavg"...
...which failed.
Vgchange complained:
"
The PVs /dev/md3 and /dev/md4 are not active.
Please run vgscan.
"
(Can't remember the excact error message)

So I gulped,shrugged and decided that it was time for another 
cup of coffee...
When I returned to my desk I ran the vgchange again in order to fetch
the error message for this email...
...et voila:
"
[root@athens root]# vgchange -ay datavg
vgchange -- volume group "datavg" successfully activated
"

So I was only able to activate datavg some minutes after my system was
back up and running ?
I really don't understand this...
OK, I am more than just happy to have my data back online, but WHY ?
Anyone any ideas on this one ?

Another question:
Is it save to extend lvs and filesystems right now in 
this strange situation ?

Bye
Lutz Reinegger

PS: The requested tar archvie is on its way.

PPS: Again: THANK YOU !!!

Am Don, 2003-10-02 um 11.11 schrieb Heinz J . Mauelshagen:
> 
> Lutz,
> 
> looks like you hit some strange LVM on top of MD bug :(
> 
> In order to get your VG active again (which of course is your highest
> priority) and before we analyze the vgscan problem, you want to go for
> the follwoing workaround (presumably /etc/lvmconf/datavg.conf is the last
> correct archive of the metadata):
> 
> # cp /etc/lvmconf/datavg.conf /etc/lvmtab.d/datavg
> # echo -ne "datavg\0" >> /etc/lvmtab
> # vgchange -ay datavg
> 
> Warning: the next vgscan run will remove the above metadata again, so avoid
>          running it for now by commenting it out in your boot script.
> 
> So far about firefighting ;)
> 
> 
> For further analysis, please do the following and send the resulting
> bzip2'ed tar archive containing your metadat to me in private mail
> <mge@sistina.com>:
> 
> # for d in md2 md3 md4 hdf hdh
> # do
> #	dd bs=1k count=4k if=/dev/$d of=$d.vgda
> # done
> # tar cf Lutz_Reinegger.vgda.tar *.vgda
> # rm *.vgda
> # bzip2 Lutz_Reinegger.vgda.tar
> 
> 
> Regards,
> Heinz    -- The LVM Guy --
> 
> 
> 
> On Tue, Sep 30, 2003 at 09:59:19PM +0200, SystemError wrote:
> > Hello out there,
> > 
> > after I migrating my precious volume group "datavg" from unmirrored
> > disks to linux software raid devices I ran into serios problems.
> > (Although I fear the biggest problem here was my own incompetence...)
> > 
> > First I moved the data from the old unmirrored disks away, using pvmove.
> > No Problems so far.
> > 
> > At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
> > So I did a vgreduce on them, then created a new raid1 
> > "/dev/md4" (containing both "hdf" and "hdh") and added it to my 
> > volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
> > No Problems so far.
> > 
> > Everything looked soooo perfect and so I decided to reboot the system...
> > 
> > At this point things started to go wrong, during the boot sequence
> > "/dev/md4" was not automatically activated and suddenly the PV
> > "/dev/hdf" showed up in "datavg", "/dev/md4" was gone.
> > 
> > Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
> > initialisation of "/dev/md4", and rebooted again. 
> > This was a probably a baaad idea. 
> > Shame upon me.
> > 
> > Now my pvscan looks like this:
> > "
> > [root@athens root]# pvscan 
> > pvscan -- reading all physical volumes (this may take a while...)
> > pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
> > pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
> > 0 free]
> > pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
> > (run vgscan)
> > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
> > 50.88 GB free]
> > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > "
> > 
> > Or with the -u option:
> > "
> > [root@athens root]# pvscan -u
> > pvscan -- reading all physical volumes (this may take a while...)
> > pvscan -- ACTIVE   PV "/dev/md2" with UUID
> > "g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
> > 10 GB free]
> > pvscan -- inactive PV "/dev/md3" with UUID
> > "R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
> > [132.25 GB / 0 free]
> > pvscan -- inactive PV "/dev/md4" with UUID
> > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > [57.12 GB / 50.88 GB free]
> > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > pvscan -- inactive PV "/dev/hdf" with UUID
> > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > [57.12 GB / 50.88 GB free]
> > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > 
> > "
> > 
> > A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
> > "
> > [root@athens root]# vgimport datavg /dev/md3 /dev/md4
> > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > "/dev/md4"
> > "
> > 
> > Using "md3" and "hdh" also fails:
> > "
> > [root@athens root]# vgimport datavg /dev/md3 /dev/hdh
> > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > "/dev/hdh"
> > "
> > 
> > It also fails when I try to use "hdf", only the error message is
> > different:
> > "
> > [root@athens root]# vgimport datavg /dev/md3 /dev/hdf
> > vgimport -- ERROR: wrong number of physical volumes to import volume
> > group "datavg"
> > "
> > 
> > So here I am, with a huge VG an tons of data in it but no way to access
> > the VG. Has anybody out there an idea how I can still access
> > the data of datavg ?
> > 
> > By the way: 
> > I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
> > as provided by RedHat.
> > 
> > Bye
> > In desperation
> > Lutz Reinegger
> > 
> > PS:
> > Any comments and suggestions are highly appreciated, even if those
> > suggestions include the use of hex editors or sacrificing 
> > caffeine to dark and ancient deities.
> > ;-)
> > 
> > 
> > 
> > _______________________________________________
> > linux-lvm mailing list
> > linux-lvm@sistina.com
> > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 
> *** Software bugs are stupid.
>     Nevertheless it needs not so stupid people to solve them ***
> 
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> Heinz Mauelshagen                                 Sistina Software Inc.
> Senior Consultant/Developer                       Am Sonnenhang 11
>                                                   56242 Marienrachdorf
>                                                   Germany
> Mauelshagen@Sistina.com                           +49 2626 141200
>                                                        FAX 924446
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [linux-lvm] Problems with vgimport after software raid initialisation failed.
  2003-10-02  5:58   ` SystemError
@ 2003-10-02  9:03     ` Heinz J . Mauelshagen
  2003-10-09 15:26       ` SystemError
  0 siblings, 1 reply; 5+ messages in thread
From: Heinz J . Mauelshagen @ 2003-10-02  9:03 UTC (permalink / raw)
  To: linux-lvm

On Thu, Oct 02, 2003 at 12:57:18PM +0200, SystemError wrote:
> Heinz,
> 
> I did as instructed and now I am able to access my 
> precious "datavg" again.
> Thanks a lot Heinz.
> You really saved my day. :-)

Great. Sounds like filesystem backup time now ;)

> 
> I also edited /etc/rc.d/rc.sysinit and commented out all vgscans
> (RedHat does 2 of them), and added a hard coded "vgchange -ay sysvg" in
> order to bring my other VG with the /usr, /var, etc... 
> filesystems online during boot time.
> 
> Immediately after another reboot I tried to activate "datavg" using
> "vgchange -ay datavg"...
> ...which failed.
> Vgchange complained:
> "
> The PVs /dev/md3 and /dev/md4 are not active.
> Please run vgscan.
> "
> (Can't remember the excact error message)
> 
> So I gulped,shrugged and decided that it was time for another 
> cup of coffee...
> When I returned to my desk I ran the vgchange again in order to fetch
> the error message for this email...
> ...et voila:
> "
> [root@athens root]# vgchange -ay datavg
> vgchange -- volume group "datavg" successfully activated
> "
> 
> So I was only able to activate datavg some minutes after my system was
> back up and running ?
> I really don't understand this...
> OK, I am more than just happy to have my data back online, but WHY ?
> Anyone any ideas on this one ?

Sounds like a race to me, where the mds aren't active when vgscan runs.

> 
> Another question:
> Is it save to extend lvs and filesystems right now in 
> this strange situation ?

No, the activation/metadata problems need to get sorted out first.

> 
> Bye
> Lutz Reinegger
> 
> PS: The requested tar archvie is on its way.
> 
> PPS: Again: THANK YOU !!!
> 
> Am Don, 2003-10-02 um 11.11 schrieb Heinz J . Mauelshagen:
> > 
> > Lutz,
> > 
> > looks like you hit some strange LVM on top of MD bug :(
> > 
> > In order to get your VG active again (which of course is your highest
> > priority) and before we analyze the vgscan problem, you want to go for
> > the follwoing workaround (presumably /etc/lvmconf/datavg.conf is the last
> > correct archive of the metadata):
> > 
> > # cp /etc/lvmconf/datavg.conf /etc/lvmtab.d/datavg
> > # echo -ne "datavg\0" >> /etc/lvmtab
> > # vgchange -ay datavg
> > 
> > Warning: the next vgscan run will remove the above metadata again, so avoid
> >          running it for now by commenting it out in your boot script.
> > 
> > So far about firefighting ;)
> > 
> > 
> > For further analysis, please do the following and send the resulting
> > bzip2'ed tar archive containing your metadat to me in private mail
> > <mge@sistina.com>:
> > 
> > # for d in md2 md3 md4 hdf hdh
> > # do
> > #	dd bs=1k count=4k if=/dev/$d of=$d.vgda
> > # done
> > # tar cf Lutz_Reinegger.vgda.tar *.vgda
> > # rm *.vgda
> > # bzip2 Lutz_Reinegger.vgda.tar
> > 
> > 
> > Regards,
> > Heinz    -- The LVM Guy --
> > 
> > 
> > 
> > On Tue, Sep 30, 2003 at 09:59:19PM +0200, SystemError wrote:
> > > Hello out there,
> > > 
> > > after I migrating my precious volume group "datavg" from unmirrored
> > > disks to linux software raid devices I ran into serios problems.
> > > (Although I fear the biggest problem here was my own incompetence...)
> > > 
> > > First I moved the data from the old unmirrored disks away, using pvmove.
> > > No Problems so far.
> > > 
> > > At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
> > > So I did a vgreduce on them, then created a new raid1 
> > > "/dev/md4" (containing both "hdf" and "hdh") and added it to my 
> > > volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
> > > No Problems so far.
> > > 
> > > Everything looked soooo perfect and so I decided to reboot the system...
> > > 
> > > At this point things started to go wrong, during the boot sequence
> > > "/dev/md4" was not automatically activated and suddenly the PV
> > > "/dev/hdf" showed up in "datavg", "/dev/md4" was gone.
> > > 
> > > Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
> > > initialisation of "/dev/md4", and rebooted again. 
> > > This was a probably a baaad idea. 
> > > Shame upon me.
> > > 
> > > Now my pvscan looks like this:
> > > "
> > > [root@athens root]# pvscan 
> > > pvscan -- reading all physical volumes (this may take a while...)
> > > pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
> > > pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
> > > 0 free]
> > > pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
> > > (run vgscan)
> > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
> > > 50.88 GB free]
> > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > "
> > > 
> > > Or with the -u option:
> > > "
> > > [root@athens root]# pvscan -u
> > > pvscan -- reading all physical volumes (this may take a while...)
> > > pvscan -- ACTIVE   PV "/dev/md2" with UUID
> > > "g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
> > > 10 GB free]
> > > pvscan -- inactive PV "/dev/md3" with UUID
> > > "R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
> > > [132.25 GB / 0 free]
> > > pvscan -- inactive PV "/dev/md4" with UUID
> > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > [57.12 GB / 50.88 GB free]
> > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > pvscan -- inactive PV "/dev/hdf" with UUID
> > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > [57.12 GB / 50.88 GB free]
> > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > 
> > > "
> > > 
> > > A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
> > > "
> > > [root@athens root]# vgimport datavg /dev/md3 /dev/md4
> > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > "/dev/md4"
> > > "
> > > 
> > > Using "md3" and "hdh" also fails:
> > > "
> > > [root@athens root]# vgimport datavg /dev/md3 /dev/hdh
> > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > "/dev/hdh"
> > > "
> > > 
> > > It also fails when I try to use "hdf", only the error message is
> > > different:
> > > "
> > > [root@athens root]# vgimport datavg /dev/md3 /dev/hdf
> > > vgimport -- ERROR: wrong number of physical volumes to import volume
> > > group "datavg"
> > > "
> > > 
> > > So here I am, with a huge VG an tons of data in it but no way to access
> > > the VG. Has anybody out there an idea how I can still access
> > > the data of datavg ?
> > > 
> > > By the way: 
> > > I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
> > > as provided by RedHat.
> > > 
> > > Bye
> > > In desperation
> > > Lutz Reinegger
> > > 
> > > PS:
> > > Any comments and suggestions are highly appreciated, even if those
> > > suggestions include the use of hex editors or sacrificing 
> > > caffeine to dark and ancient deities.
> > > ;-)
> > > 
> > > 
> > > 
> > > _______________________________________________
> > > linux-lvm mailing list
> > > linux-lvm@sistina.com
> > > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > 
> > *** Software bugs are stupid.
> >     Nevertheless it needs not so stupid people to solve them ***
> > 
> > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > 
> > Heinz Mauelshagen                                 Sistina Software Inc.
> > Senior Consultant/Developer                       Am Sonnenhang 11
> >                                                   56242 Marienrachdorf
> >                                                   Germany
> > Mauelshagen@Sistina.com                           +49 2626 141200
> >                                                        FAX 924446
> > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > 
> > _______________________________________________
> > linux-lvm mailing list
> > linux-lvm@sistina.com
> > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > 
> 
> 
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/

-- 

Regards,
Heinz    -- The LVM Guy --

*** Software bugs are stupid.
    Nevertheless it needs not so stupid people to solve them ***

=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-

Heinz Mauelshagen                                 Sistina Software Inc.
Senior Consultant/Developer                       Am Sonnenhang 11
                                                  56242 Marienrachdorf
                                                  Germany
Mauelshagen@Sistina.com                           +49 2626 141200
                                                       FAX 924446
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [linux-lvm] Problems with vgimport after software raid initialisation failed.
  2003-10-02  9:03     ` Heinz J . Mauelshagen
@ 2003-10-09 15:26       ` SystemError
  0 siblings, 0 replies; 5+ messages in thread
From: SystemError @ 2003-10-09 15:26 UTC (permalink / raw)
  To: linux-lvm

Hey there,

thanks to the help of Heinz I was able to get my VG online again...
Unfortunately we had a power outage here in Karlsruhe on 02.10.2003 and
so it took a little bit longer than expected.
(Of course the outage came right when it hurt the most)

Now I am able to varryon/varryoff my VG as often I as I like.
(during the chaos after the outage I renamed "datavg" -> "vaultvg")

My "pvscan" looks now like this:
"
[root@athens root]# pvscan 
pvscan -- reading all physical volumes (this may take a while...)
pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg"   [16 GB / 10 GB free]
pvscan -- ACTIVE   PV "/dev/md3" of VG "vaultvg" [132.25 GB / 0 free]
pvscan -- ACTIVE   PV "/dev/md4" of VG "vaultvg" [57.12 GB / 50.88 GB
free]
pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
pvscan -- WARNING: physical volume "/dev/hdf" belongs to a meta device
pvscan -- total: 5 [205.70 GB] / in use: 5 [205.70 GB] / in no VG: 0 [0]
"

And my here goes my "vgscan":
"
[root@athens root]# vgscan
vgscan -- reading all physical volumes (this may take a while...)
vgscan -- found active volume group "sysvg"
vgscan -- found active volume group "vaultvg"
vgscan -- ERROR "vg_read_with_pv_and_lv(): current PV" can't get data of
volume group "datavg" from physical volume(s)
vgscan -- "/etc/lvmtab" and "/etc/lvmtab.d" successfully created
vgscan -- WARNING: This program does not do a VGDA backup of your volume
groups
"

Here are my remaining questions:
a.) 
Why is "pvscan" complaining about the raid members "hdh" and "hdf" ?
It does not complain about my other raid devices.
Any ideas ?

b.)
And why is "vgscan" complaining about the VG "datavg",
even if this VG does not exist anymore ?

---> Would perhaps a vgexport/vgimport be a solution ?

>>> Great. Sounds like filesystem backup time now ;)
I would like very much to do so, Heinz.
:-)
But unfortunately that is not an option in 
this setup...
:-(
...better don' ask...
;-)

Still grateful for any hints
Lutz Reinegger

Am Don, 2003-10-02 um 15.51 schrieb Heinz J . Mauelshagen:
> On Thu, Oct 02, 2003 at 12:57:18PM +0200, SystemError wrote:
> > Heinz,
> > 
> > I did as instructed and now I am able to access my 
> > precious "datavg" again.
> > Thanks a lot Heinz.
> > You really saved my day. :-)
> 
> Great. Sounds like filesystem backup time now ;)
> 
> > 
> > I also edited /etc/rc.d/rc.sysinit and commented out all vgscans
> > (RedHat does 2 of them), and added a hard coded "vgchange -ay sysvg" in
> > order to bring my other VG with the /usr, /var, etc... 
> > filesystems online during boot time.
> > 
> > Immediately after another reboot I tried to activate "datavg" using
> > "vgchange -ay datavg"...
> > ...which failed.
> > Vgchange complained:
> > "
> > The PVs /dev/md3 and /dev/md4 are not active.
> > Please run vgscan.
> > "
> > (Can't remember the excact error message)
> > 
> > So I gulped,shrugged and decided that it was time for another 
> > cup of coffee...
> > When I returned to my desk I ran the vgchange again in order to fetch
> > the error message for this email...
> > ...et voila:
> > "
> > [root@athens root]# vgchange -ay datavg
> > vgchange -- volume group "datavg" successfully activated
> > "
> > 
> > So I was only able to activate datavg some minutes after my system was
> > back up and running ?
> > I really don't understand this...
> > OK, I am more than just happy to have my data back online, but WHY ?
> > Anyone any ideas on this one ?
> 
> Sounds like a race to me, where the mds aren't active when vgscan runs.
> 
> > 
> > Another question:
> > Is it save to extend lvs and filesystems right now in 
> > this strange situation ?
> 
> No, the activation/metadata problems need to get sorted out first.
> 
> > 
> > Bye
> > Lutz Reinegger
> > 
> > PS: The requested tar archvie is on its way.
> > 
> > PPS: Again: THANK YOU !!!
> > 
> > Am Don, 2003-10-02 um 11.11 schrieb Heinz J . Mauelshagen:
> > > 
> > > Lutz,
> > > 
> > > looks like you hit some strange LVM on top of MD bug :(
> > > 
> > > In order to get your VG active again (which of course is your highest
> > > priority) and before we analyze the vgscan problem, you want to go for
> > > the follwoing workaround (presumably /etc/lvmconf/datavg.conf is the last
> > > correct archive of the metadata):
> > > 
> > > # cp /etc/lvmconf/datavg.conf /etc/lvmtab.d/datavg
> > > # echo -ne "datavg\0" >> /etc/lvmtab
> > > # vgchange -ay datavg
> > > 
> > > Warning: the next vgscan run will remove the above metadata again, so avoid
> > >          running it for now by commenting it out in your boot script.
> > > 
> > > So far about firefighting ;)
> > > 
> > > 
> > > For further analysis, please do the following and send the resulting
> > > bzip2'ed tar archive containing your metadat to me in private mail
> > > <mge@sistina.com>:
> > > 
> > > # for d in md2 md3 md4 hdf hdh
> > > # do
> > > #	dd bs=1k count=4k if=/dev/$d of=$d.vgda
> > > # done
> > > # tar cf Lutz_Reinegger.vgda.tar *.vgda
> > > # rm *.vgda
> > > # bzip2 Lutz_Reinegger.vgda.tar
> > > 
> > > 
> > > Regards,
> > > Heinz    -- The LVM Guy --
> > > 
> > > 
> > > 
> > > On Tue, Sep 30, 2003 at 09:59:19PM +0200, SystemError wrote:
> > > > Hello out there,
> > > > 
> > > > after I migrating my precious volume group "datavg" from unmirrored
> > > > disks to linux software raid devices I ran into serios problems.
> > > > (Although I fear the biggest problem here was my own incompetence...)
> > > > 
> > > > First I moved the data from the old unmirrored disks away, using pvmove.
> > > > No Problems so far.
> > > > 
> > > > At a certain point I had emptied the 2 PVs "/dev/hdh" and "/dev/hdf".
> > > > So I did a vgreduce on them, then created a new raid1 
> > > > "/dev/md4" (containing both "hdf" and "hdh") and added it to my 
> > > > volume group "datavg" using pvcreate(->"/dev/md4") and vgextend.
> > > > No Problems so far.
> > > > 
> > > > Everything looked soooo perfect and so I decided to reboot the system...
> > > > 
> > > > At this point things started to go wrong, during the boot sequence
> > > > "/dev/md4" was not automatically activated and suddenly the PV
> > > > "/dev/hdf" showed up in "datavg", "/dev/md4" was gone.
> > > > 
> > > > Unfortunately I paniced and ran a vgexport on "datavg", fixed the broken
> > > > initialisation of "/dev/md4", and rebooted again. 
> > > > This was a probably a baaad idea. 
> > > > Shame upon me.
> > > > 
> > > > Now my pvscan looks like this:
> > > > "
> > > > [root@athens root]# pvscan 
> > > > pvscan -- reading all physical volumes (this may take a while...)
> > > > pvscan -- ACTIVE   PV "/dev/md2" of VG "sysvg" [16 GB / 10 GB free]
> > > > pvscan -- inactive PV "/dev/md3" is in EXPORTED VG "datavg" [132.25 GB /
> > > > 0 free]
> > > > pvscan -- inactive PV "/dev/md4"  is associated to unknown VG "datavg"
> > > > (run vgscan)
> > > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > > pvscan -- inactive PV "/dev/hdf"  is in EXPORTED VG "datavg" [57.12 GB /
> > > > 50.88 GB free]
> > > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > > "
> > > > 
> > > > Or with the -u option:
> > > > "
> > > > [root@athens root]# pvscan -u
> > > > pvscan -- reading all physical volumes (this may take a while...)
> > > > pvscan -- ACTIVE   PV "/dev/md2" with UUID
> > > > "g6Au3J-2C4H-Ifjo-iESu-4yp8-aRQv-ozChyW" of VG "sysvg"        [16 GB /
> > > > 10 GB free]
> > > > pvscan -- inactive PV "/dev/md3" with UUID
> > > > "R15mli-TFs2-214J-YTBh-Hatl-erbL-G7WS4b"  is in EXPORTED VG "datavg"
> > > > [132.25 GB / 0 free]
> > > > pvscan -- inactive PV "/dev/md4" with UUID
> > > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > > [57.12 GB / 50.88 GB free]
> > > > pvscan -- WARNING: physical volume "/dev/hdh" belongs to a meta device
> > > > pvscan -- inactive PV "/dev/hdf" with UUID
> > > > "szAa6A-rNM7-FmeU-6DHl-rKmZ-SePL-IURwtg"  is in EXPORTED VG "datavg"
> > > > [57.12 GB / 50.88 GB free]
> > > > pvscan -- total: 5 [262.97 GB] / in use: 5 [262.97 GB] / in no VG: 0 [0]
> > > > 
> > > > "
> > > > 
> > > > A vgimport using "md3"(no probs with this raid1) and "md4" fails: 
> > > > "
> > > > [root@athens root]# vgimport datavg /dev/md3 /dev/md4
> > > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > > "/dev/md4"
> > > > "
> > > > 
> > > > Using "md3" and "hdh" also fails:
> > > > "
> > > > [root@athens root]# vgimport datavg /dev/md3 /dev/hdh
> > > > vgimport -- ERROR "pv_read(): multiple device" reading physical volume
> > > > "/dev/hdh"
> > > > "
> > > > 
> > > > It also fails when I try to use "hdf", only the error message is
> > > > different:
> > > > "
> > > > [root@athens root]# vgimport datavg /dev/md3 /dev/hdf
> > > > vgimport -- ERROR: wrong number of physical volumes to import volume
> > > > group "datavg"
> > > > "
> > > > 
> > > > So here I am, with a huge VG an tons of data in it but no way to access
> > > > the VG. Has anybody out there an idea how I can still access
> > > > the data of datavg ?
> > > > 
> > > > By the way: 
> > > > I am using RedHatLinux 9.0 with the lvm-1.0.3-12 binary rpm package
> > > > as provided by RedHat.
> > > > 
> > > > Bye
> > > > In desperation
> > > > Lutz Reinegger
> > > > 
> > > > PS:
> > > > Any comments and suggestions are highly appreciated, even if those
> > > > suggestions include the use of hex editors or sacrificing 
> > > > caffeine to dark and ancient deities.
> > > > ;-)
> > > > 
> > > > 
> > > > 
> > > > _______________________________________________
> > > > linux-lvm mailing list
> > > > linux-lvm@sistina.com
> > > > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > > 
> > > *** Software bugs are stupid.
> > >     Nevertheless it needs not so stupid people to solve them ***
> > > 
> > > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > > 
> > > Heinz Mauelshagen                                 Sistina Software Inc.
> > > Senior Consultant/Developer                       Am Sonnenhang 11
> > >                                                   56242 Marienrachdorf
> > >                                                   Germany
> > > Mauelshagen@Sistina.com                           +49 2626 141200
> > >                                                        FAX 924446
> > > =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> > > 
> > > _______________________________________________
> > > linux-lvm mailing list
> > > linux-lvm@sistina.com
> > > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> > > 
> > 
> > 
> > 
> > _______________________________________________
> > linux-lvm mailing list
> > linux-lvm@sistina.com
> > http://lists.sistina.com/mailman/listinfo/linux-lvm
> > read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 
> -- 
> 
> Regards,
> Heinz    -- The LVM Guy --
> 
> *** Software bugs are stupid.
>     Nevertheless it needs not so stupid people to solve them ***
> 
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> Heinz Mauelshagen                                 Sistina Software Inc.
> Senior Consultant/Developer                       Am Sonnenhang 11
>                                                   56242 Marienrachdorf
>                                                   Germany
> Mauelshagen@Sistina.com                           +49 2626 141200
>                                                        FAX 924446
> =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> 
> _______________________________________________
> linux-lvm mailing list
> linux-lvm@sistina.com
> http://lists.sistina.com/mailman/listinfo/linux-lvm
> read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/
> 

^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2003-10-09 15:26 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2003-09-30 15:00 [linux-lvm] Problems with vgimport after software raid initialisation failed SystemError
2003-10-02  4:23 ` Heinz J . Mauelshagen
2003-10-02  5:58   ` SystemError
2003-10-02  9:03     ` Heinz J . Mauelshagen
2003-10-09 15:26       ` SystemError

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.