Why should I migrate?
Before I discuss how to migrate to a newer version of the IBM AIX®
operating system, let's review why you should consider migrating at all.
One of the most important reasons to migrate is support. If you
are running an older version of AIX, such as 5.3 or earlier, you should
already be aware that these versions are no longer supported by IBM. If
problems arise on an older version of AIX you are most likely going to
be on your own. IBM support will no longer be able to help you.
AIX 5.3 officially went out of support on April 30th 2012.
IBM is offering extended support for a limited time for a fee. This
will help some customers in the interim while they migrate their systems
to AIX 7.1 or 6.1. If you have IBM POWER7® hardware, you might want to
consider AIX 5.2 and 5.3 Workload Partitions (WPARs). This will allow
you to continue running your legacy applications on AIX 5.2 or 5.3,
however they will run within an AIX 7.1 WPAR. These special systems are
known as versioned WPARs and are only supported on POWER7 hardware. You
can find more information on versioned WPARs.
Of course there are other pressing reasons to migrate as well. Newer
releases of AIX include a multitude of new enhancements, improvements,
features and performance boosts. I encourage you to read the >AIX 7.1 and 6.1 Differences Guides (IBM Redbooks® publications) to find out more – see the Resources section .
Migrating to AIX 7.1 with nimadm
I've discussed using nimadm to migrate to AIX 6.1 in the past. In
this article I'll briefly cover that same process but this time
migrating to AIX 7.1. Admittedly the steps are almost identical. So I'll
refer you to my 2010 article as a starting point for migrating to 7.1.
In this article I'll provide a brief guide to migrating both AIX 5.3 and
6.1 systems to 7.1. I'll also offer some general advice and tips.
The nimadm utility offers several advantages over a conventional
migration. For example, a system administrator can use nimadm to create a
copy of a NIM client's rootvg (on a spare disk on the client, similar
to a standard alternate disk installation alt_disk_install) and migrate
the disk to a newer version or release of AIX. All of this can be done
without disruption to the client (there is no outage required to perform
the migration). After the migration is finished, the only downtime
required will be a scheduled reboot of the system.
Another advantage is that the actual migration process occurs on the NIM
master, taking the load off the client client logical partition (LPAR).
This reduces the processing overhead on the LPAR and minimizes the
performance impact to the running applications.
For customers with a large number of AIX systems, it is also important
to note that the nimadm tool supports migrating several clients at once.
Just as I did in my previous article I'll assume you already have a NIM
master in your environment. And I'm going to assume that your NIM master
is already running AIX 7.1 with the latest latest technology level (TL)
and and service pack (SP) applied. If not, I recommend you refer to my
previous article and the associated Resources section.
My NIM master is running AIX 7.1 TL1 SP4.
# oslevel -s 7100-01-04-1216 # lslpp -l bos.sysmgt.nim.master Fileset Level State Description ---------------------------------------------------------------------------- Path: /usr/lib/objrepos bos.sysmgt.nim.master 7.1.1.15 APPLIED Network Install Manager - Master Tools
I created new lpp_source and SPOT NIM resources for AIX 7.1 TL1 SP4.
# lsnim -t lpp_source lpp_sourceaix710104 resources lpp_source # lsnim -t spot spotaix710104 resources spot # lsnim -l lpp_sourceaix710104 lpp_sourceaixaix710104: class = resources type = lpp_source arch = power Rstate = ready for use prev_state = unavailable for use location = /export/lpp_source/lpp_sourceaix710104 simages = yes alloc_count = 0 server = master # lsnim -l spotaix710104 spotaix7101014: class = resources type = spot plat_defined = chrp arch = power Rstate = ready for use prev_state = verification is being performed location = /export/spot/spotaix7101014/usr version = 7 release = 1 mod = 1 oslevel_r = 7100-01 alloc_count = 2 server = master if_supported = chrp.64 ent Rstate_result = success
To create these resources, I downloaded AIX 7.1 ISO images from the IBM Entitled Software support website. I placed these images into a temporary directory and then used the loopmount command to temporarily mount them.
# ls -ltr total 11301248 drwxr-xr-x 2 root system 256 May 14 15:47 lost+found -rw-r--r-- 1 root system 3361046528 May 14 16:22 NIMAIX71DVD1.iso -rw-r--r-- 1 root system 2425192448 May 14 16:23 NIMAIX71DVD2.iso # loopmount -i NIMAIX71DVD1.iso -o "-V cdrfs -o ro" -m /mnt/dvd1 # loopmount -i NIMAIX71DVD2.iso -o "-V cdrfs -o ro" -m /mnt/dvd2 # df | grep loop /dev/loop0 6563484 0 100% 1640871 100% /mnt/dvd1 /dev/loop1 4735648 0 100% 1183912 100% /mnt/dvd2 # ls -ltr /mnt/dvd1 /mnt/dvd2 /mnt/dvd1: total 84 drwxr-xr-x 3 4000 4000 2048 Sep 03 2010 ppc -rw-r--r-- 1 4000 4000 819 Sep 03 2010 README.aix drwxr-xr-x 2 4000 4000 2048 Sep 03 2010 7100-00 drwxr-xr-x 3 4000 4000 2048 Sep 03 2010 root -rw-r--r-- 1 4000 4000 15081 Sep 03 2010 image.data -rw-r--r-- 1 4000 4000 6252 Sep 03 2010 bosinst.data -rw-r--r-- 1 4000 4000 16 Sep 03 2010 OSLEVEL drwxrwxr-x 4 4000 4000 2048 Sep 03 2010 RPMS drwxr-xr-x 11 4000 4000 2048 Sep 03 2010 usr drwxr-xr-x 4 4000 4000 2048 Sep 03 2010 installp -rw-rw-r-- 1 4000 4000 42 Sep 03 2010 .Version /mnt/dvd2: total 16 drwxr-xr-x 3 4000 4000 2048 Sep 03 2010 ismp drwxrwxr-x 3 4000 4000 2048 Sep 03 2010 usr drwxrwxr-x 4 4000 4000 2048 Sep 03 2010 installp -rw-rw-r-- 1 4000 4000 42 Sep 03 2010 .Version
After they are mounted, I used smitty bffcreate to copy the contents of these images to my new AIX 7.1 lpp_source directory create (/export/lpp_source/lpp_sourceaix710104).
Copy Software to Hard Disk for Future Installation
Type or select values in entry fields. Press Enter AFTER making all desired changes. [Entry Fields] * INPUT device / directory for software /mnt/dvd1 * SOFTWARE package to copy [all] + * DIRECTORY for storing software package [/export/lpp_source/lpp_sourceaixaix710104] DIRECTORY for temporary storage during copying [/tmp] EXTEND file systems if space needed? yes + Process multiple volumes? yes +
After the base filesets had been copied over, I defined the new
lpp_source and SPOT within NIM. Then I downloaded the latest TL and SP
for AIX 7.1 and updated the lpp_source and SPOT.
My NIM clients are running AIX 5.3 and AIX 6.1. We will migrate both of these systems to AIX 7.1 using nimadm. We will review the migration process for each system and check that the systems are appropriately tuned after the migration.
root@lparlparaix53[/] > oslevel -s 5300-12-04-1119 root@lparlparaix61[/] > oslevel -s 6100-06-04-1112
I have NIM client definitions for both systems already in place on my NIM master.
# lsnim -t standalone lparlparaix53 machines standalone lparlparaix61 machines standalone # lsnim -l lparaix53 lparaix53: class = machines type = standalone connect = nimsh platform = chrp netboot_kernel = 64 if1 = 172_29_154 lparaix53 0 cable_type1 = N/A Cstate = ready for a NIM operation prev_state = not running Mstate = currently running cpuid = 00CDB5114C00 Cstate_result = reset # lsnim -l lparaix61 lparaix61: class = machines type = standalone connect = nimsh platform = chrp netboot_kernel = 64 if1 = 172_29_154 lparaix61 0 cable_type1 = N/A Cstate = ready for a NIM operation prev_state = not running Mstate = currently running cpuid = 00C8E4244C00 Cstate_result = reset
Each NIM client has a spare disk available that we will use for the alternate disk migration.
root@lparaix53[/] > lspv hdisk4 00f6050a2cd79ef8 rootvg active hdisk5 00cdb511757d999e None root@lparaix61[/] > lspv hdisk4 00f6050a2cd79ef8 rootvg active hdisk5 00c8e42485fabfd7 None
The NIM Master is configured with only three volume groups, rootvg, nimvg and nimadmvg. The nimadmvg volume group will be used as temporary client data cache location for the 7.1 migrations. The volume group is currently empty.
# lsvg rootvg nimvg nimadmvg # lspv | grep adm hdisk4 00c342c6395ff736 nimadmvg active # lsvg -l nimadmvg nimadmvg: LV NAME TYPE LPs PPs PVs LV STATE MOUNT POINT
Currently, nimadm requires rsh access to the NIM clients in order to function. Therefore we ensure that the NIM Master has rshaccess to each of the clients and that NIM can communicate with each client successfully.
# rsh lparaix53 date Fri May 18 14:06:07 EETDT 2011 # rsh lparaix61 date Sat May 18 14:06:40 EETDT 2012 # nim -o lslpp lparaix53 | grep -w "bos.rte " bos.rte 5.3.0.60 COMMITTED Base Operating System Runtime bos.rte 5.3.0.60 COMMITTED Base Operating System Runtime # nim -o lslpp lparaix61 | grep -w "bos.rte " bos.rte 6.1.0.0 COMMITTED Base Operating System Runtime bos.rte 6.1.0.0 COMMITTED Base Operating System Runtime bos.rte 6.1.1.0 COMMITTED Base Operating System Runtime bos.rte 6.1.2.0 COMMITTED Base Operating System Runtime bos.rte 6.1.3.0 COMMITTED Base Operating System Runtime bos.rte 6.1.4.0 COMMITTED Base Operating System Runtime bos.rte 6.1.6.0 COMMITTED Base Operating System Runtime bos.rte 6.1.6.15 COMMITTED Base Operating System Runtime
We will now initiate a migration for both systems. Starting with the 5.3 system, we run the following nimadm command on the NIM master to start the alternate disk-migration process.
NIMADM operation for AIX 5.3 system:
# nimadm -j nimvadmg -c lparaix53 -s spotaix710104 –l lpp_sourceaix710104 -d "hdisk5" –Y Initializing the NIM master. Initializing NIM client lparaix53. Verifying alt_disk_migration eligibility. Initializing log: /var/adm/ras/alt_mig/lparaix53_alt_mig.log Starting Alternate Disk Migration. +-----------------------------------------------------------------------------+ Executing nimadm phase 1. +-----------------------------------------------------------------------------+ Cloning altinst_rootvg on client, Phase 1. Client alt_disk_install command: alt_disk_copy -j -M 7.1 -P1 -d "hdisk5" Calling mkszfile to create new /image.data file. Checking disk sizes. Creating cloned rootvg volume group and associated logical volumes. Creating logical volume alt_hd5 Creating logical volume alt_hd6 Creating logical volume alt_hd8 ...ETC...
In a new secure shell (SSH) session on the NIM master we initiate a second nimadm operation to migrate our AIX 6.1 NIM client.
NIMADM operation for AIX 6.1 system:
# nimadm -j nimadmvg -c lparaix61 -s spotaix710104 –l lpp_sourceaix710104 -d "hdisk5" –Y Initializing the NIM master. Initializing NIM client lparaix61. Verifying alt_disk_migration eligibility. Initializing log: /var/adm/ras/alt_mig/lparaix61_alt_mig.log Starting Alternate Disk Migration. +-----------------------------------------------------------------------------+ Executing nimadm phase 1. +-----------------------------------------------------------------------------+ Cloning altinst_rootvg on client, Phase 1. Client alt_disk_install command: alt_disk_copy -j -M 7.1 -P1 -d "hdisk5" Calling mkszfile to create new /image.data file. Checking disk sizes. LOGICAL_VOLUME= hd11admin FS_LV= /dev/hd11admin Creating cloned rootvg volume group and associated logical volumes. Creating logical volume alt_hd5 Creating logical volume alt_hd6 Creating logical volume alt_hd8 ...ETC...
At this point the nimadm process is copying data from the
clients, to the cache file systems on the NIM master, and performing the
migration on the master itself. After the migration process for a
client is complete, the data is copied back to the client's alternate
rootvg disk. If you are interested in learning more about each phase of
the nimadm process, then again, I refer to my original nimadm article on the IBM developerWorks® website.
We observe that both clients now have a new volume group named altinst_rootvg. This volume group contains a copy of the original rootvg, now migrated to AIX 7.1.
root@lparaix53[/tmp] > lspv hdisk4 00f6050a2cd79ef8 rootvg active hdisk5 00cdb511757d999e altinst_rootvg active root@lparaix61[/] > lspv hdisk4 00f6050a2cd79ef8 rootvg active hdisk5 00c8e42485fabfd7 altinst_rootvg active
As the migrated data is being copied from the NIM master to the client,
we observe that the alternate rootvg file systems are temporarily
mounted on each client to receive the data.
root@lparaix53[/tmp] > df | grep alt /dev/alt_hd4 262144 261384 1% 10 1% /alt_inst /dev/alt_hd11admin 262144 261448 1% 4 1% /alt_inst/admin /dev/alt_hd1 262144 261448 1% 4 1% /alt_inst/home /dev/alt_hd10opt 1048576 1047760 1% 4 1% /alt_inst/opt /dev/alt_hd3 262144 247936 6% 5 1% /alt_inst/tmp /dev/alt_hd2 7077888 7076152 1% 4 1% /alt_inst/usr /dev/alt_hd9var 524288 523552 1% 4 1% /alt_inst/var root@lparaix61[/] > df | grep alt /dev/alt_hd4 1048576 1047696 1% 10 1% /alt_inst /dev/alt_hd11admin 524288 523552 1% 4 1% /alt_inst/admin /dev/alt_hd1 524288 523552 1% 4 1% /alt_inst/home /dev/alt_hd10opt 1572864 1571968 1% 4 1% /alt_inst/opt /dev/alt_hd3 524288 508256 4% 5 1% /alt_inst/tmp /dev/alt_hd2 8388608 8386672 1% 4 1% /alt_inst/usr /dev/alt_hd9var 524288 523552 1% 4 1% /alt_inst/var
On the NIM Master we discover temporary cache file system mounts for
each of the NIM clients. These cache file systems are housed in the nimadmvg volume group.
# df -g Filesystem GB blocks Free %Used Iused %Iused Mounted on /dev/hd4 0.38 0.16 58% 11761 22% / /dev/hd2 6.00 1.95 68% 93348 17% /usr /dev/hd9var 2.38 2.02 15% 7620 2% /var /dev/hd3 0.50 0.49 3% 187 1% /tmp /dev/hd1 0.12 0.10 24% 18 1% /home /dev/hd11admin 0.12 0.12 1% 7 1% /admin /proc - - - - - /proc /dev/hd10opt 2.62 2.01 24% 12726 3% /opt /dev/livedump 0.25 0.25 1% 4 1% /var/adm/ras/livedump /dev/cglv 25.00 19.61 22% 6 1% /cg /dev/lppsrclv 25.00 13.99 45% 7269 1% /lppsrc /dev/spotlv 25.25 23.96 6% 29122 1% /spot /dev/loop0 3.13 0.00 100% 1640871 100% /mnt/dvd1 /dev/loop1 2.26 0.00 100% 1183912 100% /mnt/dvd2 /dev/lv00 0.25 0.16 37% 7441 12% /lparaix53_alt/alt_inst /dev/lv01 0.25 0.24 4% 18 1% /lparaix53_alt/alt_inst/admin /dev/lv02 0.25 0.24 4% 25 1% /lparaix53_alt/alt_inst/home /dev/lv03 0.75 0.29 62% 6718 4% /lparaix53_alt/alt_inst/opt /dev/lv04 0.25 0.20 21% 225 1% /lparaix53_alt/alt_inst/tmp /dev/lv05 3.50 0.73 80% 80350 9% /lparaix53_alt/alt_inst/usr /dev/lv06 0.25 0.15 42% 7792 12% /lparaix53_alt/alt_inst/var /dev/lv07 0.25 0.03 89% 13533 21% /lparaix61_alt/alt_inst /dev/lv08 0.25 0.24 4% 17 1% /lparaix61_alt/alt_inst/admin /dev/lv09 0.25 0.24 4% 23 1% /lparaix61_alt/alt_inst/home /dev/lv10 0.75 0.61 19% 6409 4% /lparaix61_alt/alt_inst/opt /dev/lv11 0.25 0.24 6% 107 1% /lparaix61_alt/alt_inst/tmp /dev/lv12 4.25 0.26 94% 104878 10% /lparaix61_alt/alt_inst/usr /dev/lv13 0.25 0.12 54% 9917 16% /lparaix61_alt/alt_inst/var /usr/lib 6.00 1.95 68% 93348 17% /lparaix53_alt/alt_inst/usr/ lib/alt_mig/usr/lib /usr/ccs/lib 6.00 1.95 68% 93348 17% /lparaix53_alt/alt_inst/usr/ lib/alt_mig/usr/ccs/lib /lparaix53_alt/alt_inst 0.25 0.16 37% 7441 12% /lparaix53_alt/alt_inst/usr/ lpp/bos/inst_root /lparaix53_alt/alt_inst/var 0.25 0.15 42% 7792 12% /lparaix53_alt/alt_inst/usr/ lpp/bos/inst_root/var /lparaix53_alt/alt_inst/tmp 0.25 0.20 21% 225 1% /lparaix53_alt/alt_inst/usr/ lpp/bos/inst_root/tmp /lparaix53_alt/alt_inst/home 0.25 0.24 4% 25 1% /lparaix53_alt/alt_inst/usr/ lpp/bos/inst_root/home /lparaix53_alt/alt_inst/admin 0.25 0.24 4% 18 1% /lparaix53_alt/alt_inst/usr/ lpp/bos/inst_root/admin # lsvg -l nimadmvg nimadmvg: LV NAME TYPE LPs PPs PVs LV STATE MOUNT POINT loglv00 jfs2log 1 1 1 open/syncd N/A lv00 jfs 1 1 1 open/syncd /lparaix53_alt/alt_inst lv01 jfs 1 1 1 open/syncd /lparaix53_alt/alt_inst/admin lv02 jfs 1 1 1 open/syncd /lparaix53_alt/alt_inst/home lv03 jfs 3 3 1 open/syncd /lparaix53_alt/alt_inst/opt lv04 jfs 1 1 1 open/syncd /lparaix53_alt/alt_inst/tmp lv05 jfs 14 14 1 open/syncd /lparaix53_alt/alt_inst/usr lv06 jfs 1 1 1 open/syncd /lparaix53_alt/alt_inst/var lv07 jfs 1 1 1 open/syncd /lparaix61_alt/alt_inst lv08 jfs 1 1 1 open/syncd /lparaix61_alt/alt_inst/admin lv09 jfs 1 1 1 open/syncd /lparaix61_alt/alt_inst/home lv10 jfs 3 3 1 open/syncd /lparaix61_alt/alt_inst/opt lv11 jfs 1 1 1 open/syncd /lparaix61_alt/alt_inst/tmp lv12 jfs 17 17 1 open/syncd /lparaix61_alt/alt_inst/usr lv13 jfs 1 1 1 open/syncd /lparaix61_alt/alt_inst/var
Each of my migrations took around 60 minutes to complete. By using nimadm, that was an hour of downtime I was able to avoid.
We can review the associated nimadm log files for each client to verify the migration process was successful. You can monitor the migration by tailing each of the log files on the NIM master.
# cd /var/adm/ras/alt_mig # ls –ltr *.log total 7136 -rw-r--r-- 1 root system 420858 May 18 14:33 lparaix61_alt_mig.log -rw-r--r-- 1 root system 429109 May 18 14:33 lparaix53_alt_mig.log # tail –f lparaix53_alt_mig.log All rights reserved. US Government Users Restricted Rights - Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. . . . . . << End of copyright notice for bos.help.msg.en_US >>. . . . Filesets processed: 566 of 2038 (Total time: 14 mins 7 secs). installp: APPLYING software for: bos.help.msg.en_US.com 7.1.1.0 # tail –f lparaix61_alt_mig.log All rights reserved. US Government Users Restricted Rights - Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. . . . . . << End of copyright notice for bos.msg.Ja_JP >>. . . . Filesets processed: 418 of 2059 (Total time: 13 mins 17 secs). installp: APPLYING software for: bos.msg.JA_JP.rte 7.1.1.0
After the migration process is finished, the NIM master unmounts the
cache file systems on the master and clients. We observe that the
correct oslevel is returned at the end of the migration that is 7100-01.
lparaix53:
install_all_updates: Checking for recommended maintenance level 7100-01. install_all_updates: Executing /usr/bin/oslevel -rf, Result = 7100-01 install_all_updates: Verification completed. install_all_updates: Log file is /var/adm/ras/install_all_updates.log install_all_updates: Result = SUCCESS Known Recommended Maintenance Levels ... Expanding /alt_inst/usr client filesystem. Filesystem size changed to 8650752 Adjusting size for /var Syncing cache data to client ... +-----------------------------------------------------------------------------+ Executing nimadm phase 10. +-----------------------------------------------------------------------------+ Unmounting client mounts on the NIM master. forced unmount of /lparaix53_alt/alt_inst/var forced unmount of /lparaix53_alt/alt_inst/usr forced unmount of /lparaix53_alt/alt_inst/tmp forced unmount of /lparaix53_alt/alt_inst/opt forced unmount of /lparaix53_alt/alt_inst/home forced unmount of /lparaix53_alt/alt_inst/admin forced unmount of /lparaix53_alt/alt_inst Removing nimadm cache file systems. Removing cache file system /lparaix53_alt/alt_inst Removing cache file system /lparaix53_alt/alt_inst/admin Removing cache file system /lparaix53_alt/alt_inst/home Removing cache file system /lparaix53_alt/alt_inst/opt Removing cache file system /lparaix53_alt/alt_inst/tmp Removing cache file system /lparaix53_alt/alt_inst/usr Removing cache file system /lparaix53_alt/alt_inst/var +-----------------------------------------------------------------------------+ Executing nimadm phase 11. +-----------------------------------------------------------------------------+ Cloning altinst_rootvg on client, Phase 3. Client alt_disk_install command: alt_disk_copy -j -M 7.1 -P3 -d "hdisk5" ## Phase 3 ################### Verifying altinst_rootvg... Modifying ODM on cloned disk. forced unmount of /alt_inst/var forced unmount of /alt_inst/usr forced unmount of /alt_inst/tmp forced unmount of /alt_inst/opt forced unmount of /alt_inst/home forced unmount of /alt_inst/admin forced unmount of /alt_inst Changing logical volume names in volume group descriptor area. Fixing LV control blocks... Fixing file system superblocks... Bootlist is set to the boot disk: hdisk5 blv=hd5 +-----------------------------------------------------------------------------+ Executing nimadm phase 12. +-----------------------------------------------------------------------------+ Cleaning up alt_disk_migration on the NIM master. Cleaning up alt_disk_migration on client lparaix53.
lparaix61:
install_all_updates: Checking for recommended maintenance level 7100-01. install_all_updates: Executing /usr/bin/oslevel -rf, Result = 7100-01 install_all_updates: Verification completed. install_all_updates: Log file is /var/adm/ras/install_all_updates.log install_all_updates: Result = SUCCESS Known Recommended Maintenance Levels ... Expanding /alt_inst/usr client filesystem. Filesystem size changed to 8650752 Adjusting size for /var Syncing cache data to client ... +-----------------------------------------------------------------------------+ Executing nimadm phase 10. +-----------------------------------------------------------------------------+ Unmounting client mounts on the NIM master. forced unmount of /lparaix61_alt/alt_inst/var forced unmount of /lparaix61_alt/alt_inst/usr forced unmount of /lparaix61_alt/alt_inst/tmp forced unmount of /lparaix61_alt/alt_inst/opt forced unmount of /lparaix61_alt/alt_inst/home forced unmount of /lparaix61_alt/alt_inst/admin forced unmount of /lparaix61_alt/alt_inst Removing cache file system /lparaix61_alt/alt_inst/admin Removing cache file system /lparaix61_alt/alt_inst/home Removing cache file system /lparaix61_alt/alt_inst/opt Removing cache file system /lparaix61_alt/alt_inst/tmp Removing cache file system /lparaix61_alt/alt_inst/usr Removing cache file system /lparaix61_alt/alt_inst/var +-----------------------------------------------------------------------------+ Executing nimadm phase 11. +-----------------------------------------------------------------------------+ Cloning altinst_rootvg on client, Phase 3. Client alt_disk_install command: alt_disk_copy -j -M 7.1 -P3 -d "hdisk5" ## Phase 3 ################### Verifying altinst_rootvg... Modifying ODM on cloned disk. forced unmount of /alt_inst/var forced unmount of /alt_inst/var forced unmount of /alt_inst/usr forced unmount of /alt_inst/usr forced unmount of /alt_inst/tmp forced unmount of /alt_inst/tmp forced unmount of /alt_inst/opt forced unmount of /alt_inst/opt forced unmount of /alt_inst/home forced unmount of /alt_inst/home forced unmount of /alt_inst/admin forced unmount of /alt_inst/admin forced unmount of /alt_inst forced unmount of /alt_inst Changing logical volume names in volume group descriptor area. Fixing LV control blocks... Fixing file system superblocks... Bootlist is set to the boot disk: hdisk5 blv=hd5 +-----------------------------------------------------------------------------+ Executing nimadm phase 12. +-----------------------------------------------------------------------------+ Cleaning up alt_disk_migration on the NIM master. Cleaning up alt_disk_migration on client lparaix61.
At this point we restart each of the NIM clients into AIX 7.1. You'll observe that the bootlist for each client was modified bynimadm,
so that the alternate rootvg is now the only disk in the boot list.
With the clients successfully restarted on 7.1 the migration is now
complete.
AIX Tunables post migration
After an AIX migration, I usually like to run the tuncheck command
to verify the current tunable parameters are valid. One area that can
indicate a tuning problem is the AIX error report. If you see the
following messages in the errpt output, you might want to verify the current settings are valid:
IDENTIFIER TIMESTAMP T C RESOURCE_NAME DESCRIPTION D221BD55 0523115112 I O perftune RESTRICTED TUNABLES MODIFIED AT REBOOT --------------------------------------------------------------------------- LABEL: TUNE_RESTRICTED IDENTIFIER: D221BD55 Date/Time: Wed May 23 11:51:16 EET 2012 Sequence Number: 676 Machine Id: 00C342C64C00 Node Id: lparaix53 Class: O Type: INFO WPAR: Global Resource Name: perftune Description RESTRICTED TUNABLES MODIFIED AT REBOOT Probable Causes SYSTEM TUNING User Causes TUNABLE PARAMETER OF TYPE RESTRICTED HAS BEEN MODIFIED Recommended Actions REVIEW TUNABLE LISTS IN DETAILED DATA Detail Data LIST OF TUNABLE COMMANDS CONTROLLING MODIFIED RESTRICTED TUNABLES AT REBOOT, SEE FILE /etc/tunables/lastboot.log vmo
In the output above, you'll notice that we are advised to check the /etc/tunables/lastboot.log for a modified restricted vmo tuning parameter. At this point I usually like to run the tuncheck command against the current /etc/tunables/nextboot file
and review its output. As you can see, in the example below, we are
warned that several restricted tunables are not set to their default
values. These values might not be appropriate for your newly migrated
AIX 7.1 (or 6.1) system. Settings that worked well with 5.3 are most
likely no longer appropriate with 7.1.
# tuncheck -p -f /etc/tunables/nextboot Warning: restricted tunable lrubucket is not at default value Warning: restricted tunable strict_maxperm is not at default value Warning: unknown parameter lru_file_repage in stanza vmo Warning: restricted tunable maxperm% is not at default value Warning: restricted tunable maxclient% is not at default value Checking successful
Based on the output above, the tuning for this newly migrated 7.1 system
appears to be inappropriate. Unless we have a valid reason (which has
been verified by IBM AIX support) we should set these tunables to their
default AIX 7.1 settings.
You can reset individual tunables to their defaults using the –d flag and the corresponding tuning command. For example to set the maxperm% tunable to its default you would run the following vmo command:
# vmo -p -d maxperm% Modification to restricted tunable maxperm%, confirmation required yes/no yes Setting maxperm% to 90 in nextboot file Setting maxperm% to 90 Warning: a restricted tunable has been modified
If you want to set all the vmo tunables back to their defaults you would run the following vmo command with the –D option:
# vmo -r -D Setting maxfree to 1088 in nextboot file Setting minfree to 960 in nextboot file Setting minperm% to 3 in nextboot file Modification to restricted tunable maxperm%, confirmation required yes/no yes Setting maxperm% to 90 in nextboot file Modification to restricted tunable strict_maxperm, confirmation required yes/no yes Setting strict_maxperm to 0 in nextboot file Setting maxpin% to 80 in nextboot file ...etc... Warning: some changes will take effect only after a bosboot and a reboot Run bosboot now? yes/no yes bosboot: Boot image is 47016 512 byte blocks.
Note that setting all parameters to their defaults will require the bosboot command to be run and a reboot of the system for the changes to take effect.
The tundefault command can also reset tuning to default
parameters. The command launches all the tuning commands (ioo, vmo,
schedo, no, nfso, and raso) with the -D flag. This resets all the AIX tunable parameters to their default values. The –r flag
defers the reset to default value to the next reboot. This clears the
stanza(s) in the /etc/tunables/nextboot file and if necessary, runs
bosboot and warns that a reboot is needed.
# tundefault -r
After the tunables have been reset, re-run the tuncheck command and ensure it runs without errors:
# tuncheck -p -f /etc/tunables/nextboot Checking successful
You can verify when a systems tunables were lasted checked (with the tuncheck command) by reviewing the info stanza in the /etc/tunables/nextboot file. For example, the nextboot file (below) was last checked on 7th June 2012 and the system was running AIX 7.1.
# IBM_PROLOG_BEGIN_TAG # This is an automatically generated prolog. # # bos520 src/bos/usr/sbin/perf/tune/nextboot 1.1 # # Licensed Materials - Property of IBM # # (C) COPYRIGHT International Business Machines Corp. 2002 # All Rights Reserved # # US Government Users Restricted Rights - Use, duplication or # disclosure restricted by GSA ADP Schedule Contract with IBM Corp. # # IBM_PROLOG_END_TAG info: AIX_level = "7.1.1.1" Kernel_type = "MP64" Last_validation = "2012-06-07 12:29:54 EETDT (current, reboot)" vmo: nfso: portcheck = "1" nfs_use_reserved_ports = "1"
Unless you've permanently set restricted tunables in your
/etc/tunables/nexboot file, the migration will change the systems
default tuning to match the newer version of AIX. For example, we
observed the following tuning changes on our AIX 5.3 system after
migrating to 7.1.
- The maxperm default value changed from 80 to 90:
maxperm% 80 80 80 1 100 % memory D maxperm% 90 90 90 1 100 % memory D
- The minperm default value changed from 20 to 3:
minperm% 20 20 20 1 100 % memory D minperm% 3 3 3 1 100 % memory D
- Note that with AIX 7.1 lru_file_repage is hardcoded to 0 and removed from the list of vmo tunables. Please refer to the following document, for more information.
Oracle Architecture and Tuning on AIX v2.20
- As expected, we noted a number of new tunables with AIX 7.1. Some examples are shown below.
vmm_klock_mode 2 2 2 0 3 numeric B j2_inodeCacheSize 200 200 200 1 1000 D
To learn more about these new tunables we can run the corresponding tuning utility with the –h flag to obtain usage information.
# vmo -h vmm_klock_mode
Help for tunable vmm_klock_mode:
Help for tunable vmm_klock_mode:
Purpose:
Kernel locking prevents paging out kernel data. This improves system performance in many cases. If set to 0, kernel locking is disabled. If set to 1, kernel locking is enabled automatically if Active Memory Expansion (AME) feature is also enabled. In this mode, only a subset of kernel memory is locked. If set to 2, kernel locking is enabled regardless of AME and all of kernel data is eligible for locking. If set to 3, only the kernel stacks of processes are locked in memory. Enabling kernel locking has the most positive impact on performance of systems that do paging but not enough to page out kernel data or on systems that do not do paging activity at all. Note that 1, 2, and 3 are only advisory. If a system runs low on free memory and performs extensive paging activity, kernel locking is rendered ineffective by paging out kernel data. Kernel locking only has an impact on pageable page-sizes in the system.
Kernel locking prevents paging out kernel data. This improves system performance in many cases. If set to 0, kernel locking is disabled. If set to 1, kernel locking is enabled automatically if Active Memory Expansion (AME) feature is also enabled. In this mode, only a subset of kernel memory is locked. If set to 2, kernel locking is enabled regardless of AME and all of kernel data is eligible for locking. If set to 3, only the kernel stacks of processes are locked in memory. Enabling kernel locking has the most positive impact on performance of systems that do paging but not enough to page out kernel data or on systems that do not do paging activity at all. Note that 1, 2, and 3 are only advisory. If a system runs low on free memory and performs extensive paging activity, kernel locking is rendered ineffective by paging out kernel data. Kernel locking only has an impact on pageable page-sizes in the system.
Values:
Default: 2
Range: 0 - 3
Type: Bosboot
Unit: numeric
Default: 2
Range: 0 - 3
Type: Bosboot
Unit: numeric
Tuning:
If processes are being delayed waiting for compressed memory to become available, increase ame_minfree_mem to improve response time. Note, this must be at least 257kb less than ame_maxfree_mem.
If processes are being delayed waiting for compressed memory to become available, increase ame_minfree_mem to improve response time. Note, this must be at least 257kb less than ame_maxfree_mem.
- We also noticed a new subsystem, named aso. The Active System
Optimizer (ASO) is a new AIX 7.1 (on POWER7 only) feature that
autonomously tunes the allocation of system resources to improve
performance.
# lsitab -a | grep aso aso:23456789:once:/usr/bin/startsrc -s aso -e "NORMAL_RESPAWN_NOLOG"
- New network (no) tuning options also appeared after the migration.
The example below relates to TCP tuning for loopback network access.
# no -a | grep tcp_fast tcp_fastlo = 0 tcp_fastlo_crosswpar = 0
# no -h tcp_fastlo
Help for tunable tcp_fastlo:
Help for tunable tcp_fastlo:
Purpose:
Specifies whether TCP fastpath loopback is enabled (1) or disabled (0).
Specifies whether TCP fastpath loopback is enabled (1) or disabled (0).
Values:
Default: 0
Range: 0 - 1
Type: Connect
Unit: boolean
Default: 0
Range: 0 - 1
Type: Connect
Unit: boolean
Tuning:
This option allows the TCP loopback traffic to shortcut the entire TCP/IP stack (protocol and interface) in order to achieve better performances.
This option allows the TCP loopback traffic to shortcut the entire TCP/IP stack (protocol and interface) in order to achieve better performances.
# no -h tcp_fastlo_crosswpar
Help for tunable tcp_fastlo_crosswpar:
Purpose:
Specifies whether TCP fastpath loopback between WPARs of a system is allowed (1) or forbidden (0).
Specifies whether TCP fastpath loopback between WPARs of a system is allowed (1) or forbidden (0).
Values:
Default: 0
Range: 0 - 1
Type: Connect
Unit: boolean
Default: 0
Range: 0 - 1
Type: Connect
Unit: boolean
Tuning:
This option is valid only if TCP fastpath loopback is enabled (with tcp_fastlo option).
This option is valid only if TCP fastpath loopback is enabled (with tcp_fastlo option).
I highly recommend that you refer to the IBM AIX Version 7.1 Differences Guide Redbook publication for more information on the new features of the AIX 7.1 OS.
Considerations when migrating to AIX 7.1 (or 6.1)
- Do you have JFS file systems in rootvg? If you do, please be aware that nimadm does not convert rootvg file systems from JFS to JFS2. I have requested that IBM include this feature with nimadm in the future. Starting with AIX 6.1 TL4, thealt_disk_copy utility has a new –T flag to convert rootvg file systems to JFS2 during the cloning process. Unfortunately nimadm does not currently call this flag with alt_disk_copy.If you are already running AIX 6.1 TL4 or higher, then you can use alt_disk_copy –T to convert rootvg file systems to JFS2 first and then use nimadm to migrate to AIX 7.1.If you are on AIX 5.3, the alt_disk_copy command does not have the –T flag. In this case you might want to migrate to AIX 7.1 (or 6.1) with nimadm first, and then use the alt_disk_copy –T command to convert rootvg file systems to JFS2.
- You might need to upgrade your MPIO device driver software. For example, if you are using IBM SDDPCM, then you will need to uninstall the previous version of SDDPCM and then install the correct version of the software for your new version of AIX. You might be able to use pre and post-migration scripts to include this update as part of the overall AIX migration process. Please refer to the following link for an example.
Using a post migration script with nimadm
- As discussed in my previous nimadm article, multibos is not supported in nimadm environments. Before you start animadm migration make sure you have removed any old standby BOS instance and that your rootvg logical volumes are not using any bos_ LV names.
- During our tests we found that even though we removed the standy instance (multibos –R), the nimadm process failed with the following error:
+-----------------------------------------------------------------------------+ Executing nimadm phase 11. +-----------------------------------------------------------------------------+ Cloning altinst_rootvg on client, Phase 3. Client alt_disk_install command: alt_disk_copy -j -M 7.1 -P3 -d "hdisk1" ## Phase 3 ################### Verifying altinst_rootvg... alt_disk_copy: 0505-218 ATTENTION: init_multibos() returned an unexpected result. Cleaning up. forced unmount of /alt_inst/var/log forced unmount of /alt_inst/var/log forced unmount of /alt_inst/var forced unmount of /alt_inst/var forced unmount of /alt_inst/usr/local forced unmount of /alt_inst/usr/local forced unmount of /alt_inst/usr forced unmount of /alt_inst/usr forced unmount of /alt_inst/tmp forced unmount of /alt_inst/tmp forced unmount of /alt_inst/opt forced unmount of /alt_inst/opt forced unmount of /alt_inst/home forced unmount of /alt_inst/home forced unmount of /alt_inst/admin forced unmount of /alt_inst/admin forced unmount of /alt_inst forced unmount of /alt_inst 0505-187 nimadm: Error cloning altinst_rootvg on client. Cleaning up alt_disk_migration on the NIM master. Cleaning up alt_disk_migration on client lpar1. Client alt_disk_install command: alt_disk_install -M 7.1 -X Bootlist is set to the boot disk: hdisk0 blv=hd5
We also found the init_multibos error in the /var/adm/ras/alt_disk_inst.log file on the NIM client:Tue Nov 22 14:48:12 EETDT 2011 cmd: /ALT_MIG_SPOT/sbin/alt_disk_copy -j -M 7.1 -P3 -d hdisk1 Verifying altinst_rootvg... alt_disk_copy: 0505-218 ATTENTION: init_multibos() returned an unexpected result. Cleaning up.
Given that the error appeared to be related to init_multibos, we assumed the failure was due to some multibos checks being performed by alt_disk_copy on the client. The client system did not have an existing multibos standby instance. So, we tried two things: First we created a standby instance on the client (multibos –s –X) and re-tried the nimadmoperation. This failed. Next we removed the standby instance (multibos –R) and re-tried the nimadm operation. This worked and the client then migrated to AIX 7.1 successfully. We re-tried the same operations (that is, create standby instance, remove standby instance and nimadm) several times and each worked as expected.
Unfortunately, it appears that 'multibos –R' may not clean up the /bos_inst directory. If this directory exists the nimadm operation will most likely fail. The simple fix was (in our case) to remove the /bos_inst directory before attempting the AIX migration.# rm –r /bos_inst
- If you plan on using your AIX 7.1 NIM master to migrate your AIX 5.3
clients to AIX 6.1, then make sure that you install the AIX 7.1 bos.alt_disk_install.rte fileset into the AIX 6.1 SPOT resource first. Failure to do so will result in your nimadmoperation reporting the following error message:
# nimadm -j nimadmvg -c lparaix53 -s spotaix610605 -l lpp_sourceaix610605_NEW -d hdisk2 -Y Initializing the NIM master. Initializing NIM client lparaix53. 0042-001 nim: processing error encountered on "master": /usr/bin/lslpp: Fileset bos.alt_disk_install.rte not installed. 0505-204 nimadm: SPOT spotaix610605_NEW does not have bos.alt_disk_install.rte installed. 0505-205 nimadm: The level of bos.alt_disk_install.rte installed in SPOT spotaix610605_NEW (0.0.0.0) does not match the NIM master's level (7.1.1.0). Cleaning up alt_disk_migration on the NIM master.
You must install the AIX 7.1 bos.alt_disk_install.rte fileset into your AIX 6.1 SPOT resource.# smit nim_res_op ....etc... > spotaix610605 Customize a SPOT Type or select values in entry fields. Press Enter AFTER making all desired changes. [Entry Fields] * Resource Name spotaix610605 * Source of Install Images [lpp_sourceaix710104] + Fileset Names [bos.alt_disk_install.rte]
You can verify the correct fileset is installed in your 6.1 SPOT using the following nim command:# nim -o showres spotaix610605 | grep bos.alt_disk_install.rte bos.alt_disk_install.rte 7.1.1.15 A F Alternate Disk Installation
- Some administrators, migrating from AIX 5.3 to 6.1, reported issues with certain device filesets left behind after the migration. They identified there was an issue when the lppchk command returned errors after migration. To resolve the problem, IBM support advised that certain filesets be uninstalled and that several entries be deleted from the ODM. Please refer to the following link for further information.AIX 5.3 to 6.1 migration
- Another 6.1 related issue we discovered was that after migrating, the sys0 maxuproc attribute was returned to its default value. This resulted in performance issues and application crashes. Check your maxuproc value before and after the migration and ensure it has not changed. We did not experience this issue with AIX 7.1 migrations. Please refer to the following blog post for more information.AIX 6.1 migration: iostat and maxuproc change to their defaults?
- Read the latest AIX 7.1 installation tips document when planning your migrations. It can help you resolve and/or avoid issues either before, during or after the upgrade. For example:
7100-01 InstallationWhen applying the 7100-01 Technology Level with Service Pack 1 included, you will have to run smitty update_all a second time to update bos.aso and mcr.rte. Until this is done, the oslevel command will not indicate the correct level. - I recommend you migrate to the latest TL and SP for AIX 7.1 (or 6.1) to avoid any known issues. For example, SP3 for 7.1 TL1 is vulnerable to an issue where the netstat –f command can crash an LPAR. SP4 for 7.1 TL1 contains APARIV09942 which resolves this problem. We hit this problem during our migrations. One of our system monitoring tools was regularly running the netstat command, which resulted in a number of unwanted system outages on some systems.
- Customers using IBM DB2® Versions 8.2, 9.1, 9.5 or 9.7 should apply the following ifixes when applying SP4 for AIX 7.1 TL1 or AIX 6.1 TL7. Use APAR IV22132 for AIX 7.1 TL1 SP4 and APAR IV22062 for AIX 6.1 TL7 SP4. Please refer to the following website for further information:
- Before applying AIX 7.1 TL1 SP4 to your system, make sure you have removed any legacy tuning from your system. During our tests we discovered that if we left legacy AIX 5.2 or 5.3 tuning in place the LPAR would hang during restart at "Setting tunable parameters". The following entries in our nextboot file were removed and the system started OK. This issue only appeared with systems migrated to SP4. Systems on AIX 7.1 TL1 SP3 did not exhibit this problem however we still removed the legacy tuning as it was inappropriate for 7.1.
info: AIX_level = "5.2.0.31" Kernel_type = "MP64" Last_validation = "2004-12-08 13:48:10 EETDT (current, reboot)" vmo: lrubucket = "262144" strict_maxclient = "1" strict_maxperm = "1" lru_file_repage = "0" minperm% = "2" maxperm% = "5" maxclient% = "5" maxfree = "1200" ioo: aio_maxreqs = "12288" j2_maxPageReadAhead = "512" maxpgahead = "8" no: use_isno = "0" udp_sendspace = "65536" udp_recvspace = "65536" tcp_sendspace = "262144" tcp_recvspace = "262144" rfc132
You might want to simply move the existing nextboot file "out of the way" prior to migrating. You can then review the file post migration. For example:# mv /etc/tunables/nextboot /etc/tunables/nextboot.old
- For special considerations for SSH host keys and AIX migrations, refer to this blog.
Note: As of AIX Version 7.1, the 32-bit kernel has been
deprecated. Therefore, 64-bit hardware is required to run AIX Version
7.1 (IBM POWER4™, POWER5™, POWER6™ or POWER7 systems only).
further more information Source Link :http://www.ibm.com/developerworks/aix/library/au-aix-nimadm/index.html
No comments:
Post a Comment