I used this procedure to archive what you need to do.

1: active cmirror on every cluster nodes
2: lvconvert -m 1 vg00/lvdata /dev/mapper/mapth1 --corelog #where mpath1 is the new lun

When the lvdata lv is in sync, now you can the dettach the old lun with

lvconvert -m 0 vg00/lvdata /dev/mapper/mapth0


2018-05-23 14:31 GMT+02:00 kAja Ziegler <ziegleka@gmail.com>:
Hi all,

 I want to ask if it is possible and safe to move online the clustered LV with GFS on the one PV (multipathed LUN from the old storage) to an other one (multipathed LUN on the new storage)?

I found these articles in Red Hat knowledgebase:

- Can I perform a pvmove on a clustered logical volume? - https://access.redhat.com/solutions/39894
- How to migrate SAN LUNs which has Clustered LVM configured on it? - https://access.redhat.com/solutions/466533

With regard to the mentioned articles it can be done, it is only needed to install and run the cmirror service. Should I expect any problems or other prerequisites?


My clustered environment:

- 8 nodes - CentOS 6.9
- LVM version:     2.02.143(2)-RHEL6 (2016-12-13)
  Library version: 1.02.117-RHEL6 (2016-12-13)

  Driver version:  4.33.1
- 7 clustered VGs overall
- 1 LV with GFS mounted on all nodes


- 1 clustered VG with 1 PV and 1 LV on which it is GFS:

[root@...]# pvdisplay /dev/mapper/35001b4d01b1da512
  --- Physical volume ---
  PV Name               /dev/mapper/35001b4d01b1da512
  VG Name               vg_1
  PV Size               4.55 TiB / not usable 2.00 MiB
  Allocatable           yes
  PE Size               4.00 MiB
  Total PE              1192092
  Free PE               1115292
  Allocated PE          76800
  PV UUID               jH1ubM-ElJv-632D-NG8x-jzgJ-mwtA-pxxL90

[root@...]# lvdisplay vg_1/lv_gfs
  --- Logical volume ---
  LV Path                /dev/vg_1/lv_gfs
  LV Name                lv_gfs
  VG Name                vg_1
  LV UUID                OsJ8hM-sH9k-KNs1-B1UD-3qe2-6vja-hLsrYY
  LV Write Access        read/write
  LV Creation host, time ,
  LV Status              available
  # open                 1
  LV Size                300.00 GiB
  Current LE             76800
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     256
  Block device           253:418

[root@...]# vgdisplay vg_1
  --- Volume group ---
  VG Name               vg_1
  System ID
  Format                lvm2
  Metadata Areas        1
  Metadata Sequence No  3898
  VG Access             read/write
  VG Status             resizable
  Clustered             yes
  Shared                no
  MAX LV                0
  Cur LV                1
  Open LV               1
  Max PV                0
  Cur PV                1
  Act PV                1
  VG Size               4.55 TiB
  PE Size               4.00 MiB
  Total PE              1192092
  Alloc PE / Size       76800 / 300.00 GiB
  Free  PE / Size       1115292 / 4.25 TiB
  VG UUID               PtMo7F-XIbC-YSA0-rCQQ-R1oE-g8B7-PiAeIR


- IO activity on the PV (LUN) is very low - from iostat and average per node: 2.5 tps , 20.03 Blk_read/s and 0 Blk_wrtn/s in 1 minute.


Thank you for your opinions and experience.

Have a great day and with best regards,
--
Karel Ziegler


_______________________________________________
linux-lvm mailing list
linux-lvm@redhat.com
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/



--
  .~.
  /V\
 //  \\
/(   )\
^`~'^