This post is more than 5 years old
1 Message
0
5546
zfs pool powerpath device
Hello, I have few questions regarding how zfs handles power path devices during reconfigure/DR situation on Solaris 11.3. Unfortunately, I could not get any information from the manual
I am in the process of creating a zfs pool using powerpath devices. Here are the questions.
1. What happens when powrepath is reconfigured and new name is allocated to the device that I used to create the pool. Would zfs recognize the new device and reconfigure the zfs pool?
2. LUNs on the server gets replicated to DR during DR test/situation. Would I be able to import ZFS pool on the DR server as and when replicated LUNs are presented to the target server.
Any input on above is much appreciated
unixian
31 Posts
0
August 15th, 2017 18:00
G'day. In answer to your 1st question:
Here is the audit of my session:
billy-the-kid# ls -l /dev/dsk/emcpower525a
lrwxrwxrwx 1 root root 35 Feb 2 2017 /dev/dsk/emcpower525a -> ../../devices/pseudo/emcp@525:a,blk
billy-the-kid# zpool create tank emcpower525a emcpower526a
billy-the-kid#
billy-the-kid# zpool list tank
NAME SIZE USED AVAIL CAP HEALTH ALTROOT
tank 1.12G 89K 1.12G 0% ONLINE -
billy-the-kid# zpool status tank
pool: tank
state: ONLINE
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
tank ONLINE 0 0 0
emcpower525a ONLINE 0 0 0
emcpower526a ONLINE 0 0 0
errors: No known data errors
From powermt ; policy=SymmOpt; queued-IOs=0
==============================================================================
--------------- Host --------------- - Stor - -- I/O Path -- -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3072 pci@1f,4000/lpfc@2/fp@0,0 c1t5000097208131D60d19s0 FA 9f:00 active alive 0 0
3074 pci@1f,4000/lpfc@4/fp@0,0 c2t5000097208131D68d19s0 FA 11f:00 active alive 0 0
billy-the-kid# emcpadm renamepseudo -s emcpower525a -t emcpower527a<<===
billy-the-kid# zpool export tank
billy-the-kid# emcpadm renamepseudo -s emcpower525a -t emcpower527a<<---
billy-the-kid# powermt display dev=emcpower527a
Pseudo name=emcpower527a
Symmetrix ID=000192601223
Logical device ID=2D0E
Device WWN=60000970000192601223533032443045
state=alive; policy=SymmOpt; queued-IOs=0
==============================================================================
--------------- Host --------------- - Stor - -- I/O Path -- -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3072 pci@1f,4000/lpfc@2/fp@0,0 c1t5000097208131D60d19s0 FA 9f:00 active alive 0 0
3074 pci@1f,4000/lpfc@4/fp@0,0 c2t5000097208131D68d19s0 FA 11f:00 active alive 0 0
billy-the-kid# zpool import tank
billy-the-kid# zpool status tank
pool: tank
state: ONLINE
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
tank ONLINE 0 0 0
emcpower527a ONLINE 0 0 0
emcpower526a ONLINE 0 0 0
errors: No known data errors
unixian
31 Posts
1
August 15th, 2017 18:00
To answer your 2nd question:
Please see the section on migrating storage pools at Migrating ZFS Storage Pools - Oracle Solaris Administration: ZFS File Systems