Overview of Red Hat Openshift Enterprise


Openshift is a container platform developed by Red Hat to deploy, develop and run applications.
It is designed using upstream community project Openshit Origin. Openshift Origin basically provides an open source application container platform. All source code for the Origin project is available under the Apache License (Version 2.0) on GitHub.

Openshift Origin used in OpenShift Online, OpenShift Dedicated, and OpenShift Container Platform which are different software products by Red Hat. It is Built around a core of Docker container packaging and Kubernetes container cluster management, Origin is augmented by application lifecycle management functionality and DevOps tooling.

OpenShift Online is Red Hat’s public cloud application development and hosting service.
OpenShift Dedicated is Red Hat’s managed private cluster offering, built around a core of application containers powered by Docker, with orchestration and management provided by Kubernetes, on a foundation of Red Hat Enterprise Linux.
OpenShift Container Platform (formerly known as OpenShift Enterprise) is Red Hat’s on-premise private platform as a service product, built around a core of application containers powered by Docker, with orchestration and management provided by Kubernetes, on a foundation of Red Hat Enterprise Linux.

If you do not have to manage your own data center then you can use OpenShift Online by Red Hat, a public cloud platform provided by Red Hat.


Openshift Enterprise builds on other different opensource project such as Atomic, Docker, and Kubernetes. Openshift services provide additional authentication, security, scheduling, networking, storage, and application life-cycle management over standard Kubernetes orchestration.

Applications are running as a container inside Openshift Enterprise with isolated from each other on single operating system. Containers has some benefits over Virtual Machines. As container are light weight virtual machines with minimal operating system packages and application dependencies installed. Each container has separate storage and network isolation. This leads to deploy applications rapidly inside the container.

Following diagram shows software stack included in Openshift Enterprise Product.


From above diagram starting from bottom,
1) Base Operating System (Red Hat Enterprise Linux).

2) Docker: A container platform service.

3) Kubernetes: It is Orchestration tool designed and developed by Google and written in Go programming language. It used to manage the deployment of containers using templates.

4) Containerized Service: fulfill many PaaS infrastructure functions such as networking and authorization. Some of them run all the time, while others are started on demand. Run times and xPaaS are base container images ready for use by developers, each preconfigured with a particular runtime language or database.

5) DevOps tools and user experience: Openshift provides Web and CLI management tools for developers and system administrators, allowing the configuration and monitoring of both applications and Openshift services and resources.

In upcoming articles, I will demonstrate you, how to install and configure Openshift Enterprise on Red Hat Enterprise Linux.







HP Integrity VM/vPar Administration

vPars/HPVM v6.1 software requirements:

HP VM/vPar can installed on following servers:

c-class i2/i4 blade server, Superdome Server 2, rx2800 server.

Host on which virtualization software is installed is called as VSP.

Virtualization server platform (VSP) requirements:

  • HP-UX11.31 with AR1203+ (1209+ for i4 blades) running v6.1 vPars/HPVM
  • BB068AA and VirtualBase (AR1203 VSE-OE,DC-OE or application media)
  • GUIDMGR (required separate ISU)
  • VMMGR (required separate ISU)

Install VSP software depot on base machine

# swinstall  –s -x autreboot=true BB068AA VirtualBase GUIDMGR VMMGR

# swlist BB068AA VirtualBase VMMGR

vPars/VM Configuration files:






vPar/VM log files:




/var/opt/hpvm/guests/name/console/consolelog *read with strings

VSP boot time configuration:


/sbin/rc2.d/S533hpvm   ->  /sbin/init.d/hpvm

/sbin/rc1.d/K467hpvm   ->  /sbin/init.d/hpvm

Verify VSP is running on base server:

# ps –ef | grep –e hpvm

# vparsratus

# hpvminfo

# more /var/opt/hpvm/common/hpvm_mon_log

Verify VSP Resources Displays

It displays available resources on vsp which will used to create vpar/vm

# hpvmstatus –s

Display Cores allocated/available to vsp and vpar/vm

# hpvmhwmgmt –lp cpu

Display Memory for vsp and vpar/vm

# hpvmhwmgmt –lp memory

Display Direct I/O

# hpvmhwmgmt –lp dio

Device Management Database file location


To view entries:

# hpvmdevmgmt –l rdev|gdev|env|server|all


# hpvmdevmgmt –l all

View the restricted devices listed in the HPVM device database. These are devices that are reserved for the VSP only to use:

# hpvmdevmgmt –l rdev | more

Example: Add /opt directory as restricted directory and change recursive attribute for /opt directory

# hpvmdevmgmt –a rdev:/opt

# hpvmdevmgmt –l rdev | grep /opt

# hpvmdevmgmt –m rdev:/opt:attr:RECURSIVE=DIR

# hpvmdevmgmt –l rdev:/opt

Managing Backing Store:

Resources on the VSP assigned to one or more vpar/vm as raw storage space.

# hpvmdevmgmt –S 20G /hpvmdisks/vmg0disk1

# hpvmdevmgmt –m gdev:/backing_store:attr:SHARE=YES

Creating Virtual Switches:

Vparnet and hpvmnet command are used to create virtual switches

Check network interfaces available on VSP first using

#  nwmgr


#  lanscan

Create virtual switch on selected network interface

#  vparnet –c –S vswlan1 –n 1


# hpvmnet –c –S vswlan1 –n 1

c – indicates creation of switch

S – Specify virtual switch name

n – select instance id of physical nic

#  vaprnet


#  hpvmnet

Cloning Virtual switch

# vparnet –C –S vswlan1 –N NewVsw –n 2

Show properties of virtual switch

# hpvmnet –p all –S vswlan1

# hpvmnet –p all –S NewVsw

List the devices in HPVM database. Find the entries related with new vswitch.

# hpvmdevmgmt –l all | grep –e SWITCH –e lan

Start Virtual Switch

# vparnet –b –S vswlan1

-b : boot virtual switch

-S : vswitch name

-r : restart vswitch

-h : halt vswitch

Check status

# vparnet


# nwmgr

Creation of vPar and VM:

Example1:  Create a vPar with following parameters

Name: vPar27

ID: 27

Cores: 4 total, min 2, max 6

Memory: 2GB

Graceful timeout: 90 seconds

# vparcreate   –P   vPar27   -p  27  -a   cpu::4:2:6  -a  mem:2048  -x  graceful_timeout=90

Example2: Create vm with following parameters

Name: VFinDB

vCPU: 2 total, min 1, max4

CPU entitlement: 40%

Memory: 4GB

Graceful timeout: 60 Second

# hpvmcreate –P vFinDB  -c  2:1:4  -r 4G  -e  40  -x  graceful_timeout=45

Additional Options available in vparcreate and hpvmcreate commands:

-P:          vpar/vm name

-a:           Add new resources/device to vpar/vm

-K :               ssh or telnet IP Address of the Virtual iLo/remote console

-L:           subnet mask

-l:            label to the vpar and vm

-u  login:admin | oper    Virtual console user access

-g  group:admin | oper  Virtual console grop access

-x  active_config=true | false

True= assign resources exclusively when created; this is default option

False= create in standby mode without assigning resources: can’t be started

-e :         cpu entitlement in %

-E : cpu entitlement in MHz or GHz

(cpu entitlement is a minimum percentage of the CPU processing power that is guaranteed to the guest)

Add Virtual Switch to the vPar/VM

# hpvmmodify –P vFinDB  –a  network:avio_lan::vswitch:vswlan0

Here,    avio_lan : Adapter Type

:: Adapter arguments (opt) bus, device, MAC

Add vNIC with a specific vMAC to the vPar, backed by a port on vswlan1

# vparmodify –P  vPar27  -a network:avio_lan:,,aa-aa-aa-05-20-56:vswitch:vswlan1

Add Direct Network to the vPar and VM

DIO pNIC must be defined as a DIO device in the database before assigning

# ioscan –fnClan

Add physical NIC to the HPVM device management database

# hpvmhwmgmt  -a  dio  0/0/0/4/0/0/1

Add a DIO vNIC with specific vMAC to the vPar

# vparmodify  -P  vPar27  -a  lan:dio:,,aa-aa-aa-05-47-56:hwpath:0/0/0/4/0/0/1

Adding Storage Device to the vPar/VM

aviostor (Emulated single ported parallel SCSI MPT storage adapter)

A virtual server is up to having 256 AVIO storage devices. All storage devices is addressed from virtual PCI buses. There are total seven PCI bus on each virtual server. Each PCI bus has eight slots into which virtual PCI adapters can be placed.

aviostor is used to connect 15 SCSI target devices to the virtual server.

-a  disk: aviostor: 1,1,2: disk:/dev/rdisk/disk34

Here,     disk: devicetype

Aviostor: Adapter type

1,1,2: PCI bus (0-6), PCI slot (0-7) , scsi target (0-14)

Example1: Add virtual disk backed by file to vPar

# hpvmdevmgmt –S 45G /BootDisks/vPar27boot

# vparmodify –p 27  -a  disk:aviostor::file:/BootDisks/vPar27boot

Example2: Add virtual disk backed by an LV to vPar

# lvcreate –L 400  –n   vPar27LV    /dev/vg4vPars

# vparmodify p -27 –a disk:aviostor::lv:/dev/vg4vPars/vPar27LV

Example3: Add a virtual disks backed by VSP physical disk or SAN LUN, with virtual hardware path 1/0/2/0.3.0 to VM

# hpvmmodify –p 2 –a disk:avio_stor:1,2,3:disk:/dev/rdisk/disk5


# vparmodify -p abc -a disk:avio_stor::disk:/dev/rdisk/disk5

Example4: Add virtual disk backed by a logical volume to the VM

# lvcreate –L 4000 –n  vFinDBLV  /dev/vg4vPars

# hpvmmodify –p 2  –a  disk:avio_stor::lv:/dev/vg4vPars/vFinDBLV

Guest Devices in the Database

# hpvmdevmgmt   –l    gdev

vPars/VM manage their device database in flat ASCII file  /var/opt/hpvm/common/hpvm_mgmtdb

only hpvmdevmgmt command will be used to display these file.

Viewing Status and Configuration

View Storage details of vm

# hpvmdevinfo  -Vp  27

# vparstatus   -vp   27

# hpvmstatus  -vp  27                       ( Use –V for more information)

Accessing Virtual Machine console

# vparconsole  -p  27

# hpvmconsole  -p  27

Powering Up a vPar or VM

# vparboot  -p  27

# vparconsole  -p  27  -c  “pc  -on”

# vparconsole  -p  27

[vpar27]  vMP>  pc  -on


# hpvmstart  -p  28

# hpvmconsole  -p  28  -c  “pc  -on”

# hpvmconsole  -p  28

[vFinDB]  vMP>  pc  -on

Power Down vPar

# vparstatus  -p  27

#vparreset   -p  27 –fh

Here, -g: waits graceful_stop_timeout # of seconds; reboot inhibited

-d: power off;  reboot inhibited

-h: reset and reboot

-t: ToC; default

-f: force; Required


# vparconsole  -p  27  -c  “pc  -off”

Powering Down VM

# hpvmstatus  -p   2

#  hpvmstop  -p  2


# hpvmconsole   -p  2  -c  “pc  -off”

Removing  vPar/VM

vPar/VM must be powerd off before removed

# vparremove  -fp  27

# vpvmremove  -P  vFinDB

Set Virtual Server Primary Boot Path

After creation of virtual machine set primary boot path for vm/vPar

# grep   –i  boot   /var/adm/syslog/syslog.log

Copy hardware path from this file

#  setboot  -p

Confirm with setboot command

# setboot

Monitoring VSP Resources

Display VSP Resources

# hpvmstatus –s     or   # vparstatus –A

VSP Resource Pools

# hpvmhwmgmt  -lp  memory

# hpvmhwmgmt  -lp cpu

# hpvmhwmgmt  -lp dio

Example: Move a core from the vPars/VM CPU pool to the VSP CPU Pools

# vparhwmgmt  -p cpu  -d  1

Now check available core to vPar/VM

# vparhwmgmt  -lp  cpu

Olarstarpolicy for CPU pools of vPar/VM and VSP

Set to host, changes VSP CPU Pool

Set to guest, changes vPar/VM Pool

# vparhwmgmt  -p  cpu   -x  olarstarpolicy=[host  | guest]

# hpvmhwmgmt  -p  cpu  -x  olarstarpolicy=[host  | guest]

Virtual iLo Remote Access Control

# vparmodify  -P vPar27  -K  -L

Here,  -K = iLO IP Address

-L = Subnet mask

Access to vPar/VM for Non-Super Users

# useradd –s  /opt/hpvm/bin/vparconsole  -r  no  -d  /var/opt/guests/vPar27   vPar27

# passwd vPar27

# hpvmmodify  -p  27 –u  vPar27:oper


# hpvmmodify  -p  27  -u  vPar27:admin

Managing Backing Store

Attached Devices: Tapes, DVD, USB

Example:  device_type : avio_stor :: attach_path : lun_path

Adding Tape drive to Vpar/VM

#  vparmodify  -P vPar27  -a  tape:avio_stor::attach_path:0/5/0/0/0/0.0×50001.0x0

# hpvmmodify –P vFinDB  -a  tape:avio_stor::attach_path: 0/5/0/0/0/0.0×50001.0x0

Devices are not shared by default, but can be made sharable

# hpvmdevmgmt –m  gdev:: 0/5/0/0/0/0.0×50001.0x0:attr:SHARE=YES


Virtual Storage devices are assigned to vPar/VM dynamically they are hot pluggable but Virtual adapters are not hot pluggable.

Adding DVD drive to vPar/VM

Map virtual DVD to Physical DVD

# vparmodify  -p 27  -a  dvd:aviostor::disk:/dev/rdisk/disk28

Map Virtual DVD to an .iso file

# vparmodify –p 27 –a dvd:aviostor::file:/ISOs/image1.iso

Map Virtual DVD to null Device

# vparmodify –p 27 –a dvd:aviostor::null:/ISOs

Control Media Selection on vMP of the vPar

vMP>  in

vMP>  ej

-Not recommended for OS installations, guest reboot will cause media to eject.

Installing Guest OS

Install from DVD

-hpvmstart and vparboot automatically boot from DVD before a R/W boot disk is installed

-Ignite adds the selected disk to the vEFI boot manager

-Map physical DVD drive to Virtual DVD drive

# vparmodify  -p  27 –a dvd:aviostor::disk:/dev/rdisk/disk28

-Map an ISO image of a bootable install DVD to an virtual DVD drive

# vparmodify –p 27 –a dvd:aviostor::file:/var/iso/image_1203_dvd1.iso

Network Installations

-Create an entry for the vPars vNIC into /etc/booptab file of Ignite server

-Start the vPar/VM

-from virtual console, interrupt the boot and go to EFI shell

Shell> lanboot select

Configuring Boot Parameters

Display vPar/VM boot manager menu

# hpvmnvram  -lp  vPar27

Configure vPar/VM vEFI boot options (vPar/VM must be down)

#  hpvmnvram –P vPar27  -a  /dev/rdisk/disk35: 3 :\efi\hpux\hpux.efi


-a  Add an entry for file-backed DVD to position 2

# hpvmnvram –p 27 –a /var/iso/Image_1203_dvd1.iso: 2 : \INSTALL.EFI

-m  Move Entry 3  to position 2

#  hpvmnvram –p 27  -m 3:2

-d  Delete entry  4

# hpvmnvram –p 47 –d 4


VLAN Configurations on Virtual Switches

Note: Ports are not assigned to vPar/VM by default

Vswitch ports may be assigned to VLAN before being assigned to vPars or VM’s

VLAN tagging is done at port level in vPar/VM

# vparnet –S vParsw0  -u  portid:1:vlanid:147

# vparnet –S vParsw0  -u  portid:2:vlanid:147

# vparnet  -S vParsw0  –p all

Assign particular port to vPar/VM

# hpvmmodify –p 27 –a network:aviolan::vswitch:vParsw0:portid:1


# vparmodify –p 27 –a network:aviolan::vswitch:vParsw0:portid:2

Configuring Dynamic Resources to Virtual Machine

VM must be down to set Dynamic memory parameters

Assigning Dynamic Memory to VM

# hpvmmodify –p 27 –x  ram_dyn_type=any  \

-x   ram_dyn_min=512M \

-x  ram_dyn_max=1G \

-x  ram_dyn_target_start=512M \

-x  dynamic_memory_control=1

Other Options:

-x ram_dyn_entitlement  (Guarenteed amount of memory to VM) in MB or GB

-x amr_enable (Enable automatic memory reallocation) 1  or 0

-x amr_chunk_size (Granule size for automatic memory reallocation in MB or GB)

Check Dynamic Memory Information of VM

# hpvmstatus   -VP vFinDB


On guest machine or vm

# hpvmmgmt  -V –l ram

To Enable Automatic Memory reallocation to VM edit parameter in /etc/rc.config.d/hpvmconf file of VSP

# vi  /etc/rc.config.d/hpvmconf

HPVMAMRENABLE=1        ; 1= start hpvmamrd deamon with host software

Guest Configuration

# hpvmmodify –P vPar27 –x  amr_enable=1 –x amr_chunk_resize=512

Dynamic vCPU Management in VM

On VSP check cpu status

# hpvmhwmgmt –lp  cpu


Check current cpu status of VM

# hpvmmgmt  -l  cpu

Enable only 2 vCPU in VM

# hpvmmgmt –c  2

Note: However disabled vCPU still appear in hpvmsar command, to changing actual number of vCPU in VM configuration requires VM to be down at the time of modification.

Cloning Virtual Server

Create Template to clone virtual server

# vparcreate –P vParTemplate –a cpu:::1:4

-a network:aviolan::vswitch:vswlan1

-a mem::2048

-x active_config=false

# hpvmclone –P vParTemplate –N New1

# hpvmstatus


While creating template active_config attribute set to false, because it just a template so resources are not reserved to template.

After creation of vPar/Vm it need to change active_config attribute to true

# vparmodify –P vPar27  -x  active_config=true

# hpvmmodify –P vPar27  -x active_config=true

When cloning a vPar, the clone fails when non sharable devices are configured. To clone vPar, all devices assigned to existing vPar must be sharable

When cloning a VM, warning message will displays when non sharable devices are configured, but the clone continues.

Converting a vPar to VM

A VSP can host a vPar or VM, but not both in v6.1

# vparstatus

# vparmodify –p 27 –x vm_type=shared

# vparstatus

Converting VM to vPar

# hpvmstatus

# hpvmmodify  -p 27 –x vm_type=vpar

# hpvmstatus

Monitoring Virtual Servers on VSP

Configurations, Consoles, Log Files


Vmm_config.current                    current configurations

Vmm_config.prev                          last known configurations

Log                                                        ASII Control files

Console/conslog                             vPar/VM data console; read with string command on VSP

VSP Log files:

Command Log file

/var/opt/hpvm/common/command.log   (on VSP # hpvmstatus –e to read this file)

Monitor Log File


vPars/VM Processes on the VSP

From a  ps –ef  following processes are displays

Powered-up processes

/opt/hpvm/lbin/hpvmapp –d vPar27

/opt/hpvm/lbin/hpvmapp –d vPar1

Vswitches processes

/opt/hpvm/lbin/hpvmnetd –c –I vparsw0 –n 0 –p

/opt/hpvm/lbin/hpvmnetd –c –I localnet

Glance on the VSP

On vsp xterminal

# xglance

Glance has the logical system report this report provides a list of virtual servers and displays information of them

Performance agent collects all metrics for virtual systems every 5 minutes. Edit the file /var/opt/perf/parm to enable logical system class to collect virtual machine performance data.

hpvmsar command to monitor cpu and lan performance

# hpvmsar –a

# hpvmsar –S vswlan1  -n 2

Other options

-s  seconds          sample rate

-n #                        # of additional samples (1+n); 0=continues; default=5

-h #                        # of lines of output between header redisplay; default=8

-M                          vCPU data

-A                           all guests

-a                            all running guests

Migrating Integrity Virtual Machines


  • Virtual machine backing stores likes SAN Lun’s are visible to source and target VSP
  • DIO Devices are labelled identically on source and target VSP
  • vswitch names used by the vPar or VM are identical on both VSP
  • Source and destination VSP network properly configured.

Online migration requires both VSP are having same processor family.

# machinfo –v  | grep –I family

Family  32, model 2, stepping 4

Online migration of virtual machine is limited to processor families 31 and 32.

VSP Network and Encryption Configuration

Two variables in /etc/rc.config.d/hpvmconf file set the encryption defaults


1= Encrypt unless overridden by hpvmmigrate –Y

0= clear text unless overridden by hpvmmigrate –y


Sets the encryption algorithm aes-128-cbc (default) or aes-256-cbc(stronger encryption)

SSH-Key based Configuration for both VSP


  • On both VSP, verify hostname/IP resolution includes files

# grep hosts  /etc/nsswitch.conf

Hosts:  files  dns

  • On both VSP verify that /etc/hosts resolves only host-hpvm-migr to the private network address.

# grep  migr  /etc/hosts         vsp1-hpvm-migr               vsp1-hpvm-migr.usv1.net         vsp2-hpvm-migr               vsp1-hpvm-migr.usv2.net

  • On both VSP secure communication between the VSP’s

On VSP1:

# /opt/hpvm/bin/secsetup  -r  VSP2-hpvm-migr


# /opt/hpvm/bin/secsetup  -r VSP1-hpvm-migr

  • Test Secure communication between the VSP’s

On VSP1:

#hpvmmigrate  -s  -h  VSP2-hpvm-migr

VSP ssh Configuration

Disable  /etc/issue banner for SSH

# vi /etc/opt/ssh/sshd_config:

#banner  /etc/issue

banner none

Restart SSH service

# /sbin/init.d/secsh stop

# /sbin/init.d/secsh start

Offline Virtual Server Migration

Perform Pre-Migration compatibility checking between the local VSP and target VSP

# hpvmmigrate –s  -h  target_IP_or_hostname

Here, -h = perform compatibility checking

Without –h option it actually performs migration.

# hpvmmodify –P  vm_name –x runnable_status= enabled

More Options for hpvmmigrate command-

-D = Delete vpar from source VSP; default is not to delete and disable runnable status

-b = automatically boot the vpar on target VSP

-n = start the migration in background

-N = Assign new name to vpar on the target VSP. If the vpar has same name on target VSP then its UUID must have same and marked not runnable on target VSP.

-d = power down the vpar after resources verification phase of the migration completed successfully.

-C = use with –m to copy local storage from source to target

For online migration of VM it is necessary to Enable Runnable status and Online Migration status of VM

Lab Activity: Migration of Virtual Servers

Part 1: Configuring the VSP’s for Virtual Server Migration

  • Shutdown the vpar/VM

# vparconsole –p 27                        OR          # hpvmconsole –p 27

# shutdown –h now

  • Verify vswitches on both the VSP


Verify vswitch name and port number if assigned.

  • Find shared LUN to be used as backing store on virtual server that will be migrated.
  • Verify the boot devices

# ioscan  -m dsf

Find World wide ID of boot device for your VM guest

# ioscan  – C disk –P wwid             <= on VSP1

#ioscan   -C disk –P wwid                              <= on VSP2

Note: A common WWN indicates that the same lun, whether viewed on first VSP or 2nd VSP, in the ioscan listing abouve that are 2 common WWN, also note there pDSF are different.

  • Determine if DNS is used on VSP by checking /etc/resolv.conf file. Edit /etc/hosts file to add 2 lines for private IP address for virtual machines.

Example:   vsp1-hpvm-migr.usv1.net                vsp1-hppvm-migr   vsp2-hpvm-migr.usv1.net                vsp2-hppvm-migr

# vi /etc/nsswitch.conf

Hosts: files, dns

# nsquery  hosts  vsp1_hostname-hpvm-migr

# nsquery  hosts  vsp2_hostname-hpvm-migr

  • Identify LAN instances

#ioscan  -fkC lan

Edit  /etc/rc.config.d/netconf    file to add private ip address that are mentioned in /etc/hosts files

Restart network service

# /sbin/init.d/net  start

Verify private network is configured

# ifconfig lan#

  • Disable /etc/issue banner for SSH

# vi /etc/opt/ssh/sshd_config:

#banner  /etc/issue

banner none

Restart SSH service

# /sbin/init.d/secsh stop

# /sbin/init.d/secsh start

  • On VSP1:

# /opt/hpvm/bin/secsetup  -r  VSP2-hpvm-migr


# /opt/hpvm/bin/secsetup  -r VSP1-hpvm-migr

  • Test the ssh trust between the VSP


# hpvmmigrate  -o  -s  -h  VSP2_hostname


# hpvmmigrate –o –s –h VSP1_Hostname

Part 2: Creating the VM Guest and Migrating it Online

  • On VSP1

# hpvmcreate  -P vsg3   -r 2G

# hpvmmodify –P vsg3  -a disk:aviostor::disk:/dev/disk/disk27

#hpvmmodify –P vsg3 –a network:aviolan::vswitch:vswlan0

# hpvmnvram  -lp  vsg3

# hpvmnvram  -P vsg3  -a /dev/rdisk/disk27:1:”\EFI\HPUX\HPUX.EFI”

# hpvmstart  -P vsg3

Now Migrate vsg3  on  VSP2

# hpvmmigrate   -o –P vsg3 –h VSP2_Hostname

Observe the migration phase.

When Migration is completed, verify the vsg3 is running on VSP2

  • On VSP2

# hpvmstatus

# hpvmstatus  -VP vsg3 | grep status


# hpvmstatus  -VP vsg3  | grep status

GUIDmgr and NPIV with vPar/VM v6.1


Node Port ID Virtualization is multiple unique number of WWN registered to the VSP fibre channel physical HBA port, so that virtual server get unique vPort on VSP’s physical HBA.

SAN knows the vPort directly, storage visibility to the virtual servers without VSP visibility.

GUIDmgr (Global Unique Identifier Manager):

It is a client server software that is used to assign vWWN to virtual server virtual fibre channel adapters.

OUID (Organizationally Unique Identifier):

It is having leading 7 hex digits of WWN. 50014c2 for GUIDmgr.

Example: Configuring GUIDmgr and Adding NPIV to Virtual Server

  • Verify GUIDmgr bundle is installed on VSP

# swlist –l bundle GUIDMGR

  • Check FC HBA of VSP supports NPIV or not

# ioscan –fnNC   fc

# /opt/fcms/bin/fcmsutil             /dev/fclp1

  • Initialize GUIDmgr Server

# cd /opt/guid/utils

# ./guid_server_prepare.sh

  • Define Hostname bl86-220 as GUIDmgr server

# /opt/guid/sbin/guidconfig  -h   bl86-220

  • Prepare the client side of the GUIDmgr server to install public key

# /opt/guid/bin/guid_client_prepare   -a

  • Define the WWN managed pool as 50014c2001007110 –  50014c200100711f

# /opt/guid/bin/guidmgmt  -S  wwn  001007110  00100711f

  • Verify the port WWN and Node WWN pools

# /opt/guid/bin/guidmgmt  -L  wwn

  • Add NPIV to Client machine

# vparconsole –p 27 –ic  “CO”

Login as root user

# ioscan  -C   fc   -P  wwid

# ioscan  -C  -fc   -P wwid  >  /tmp/vsgNameStorage

# shutdown  -h   now

vMP> X

# vparconsole   -p  27   -a  hba:aviostor::npiv:/dev/fcd0

#hpvmstatus –VP vpar27