Updating the system
The system update process involves the updating of your entire SAN Volume Controller environment.
Start here to update to version 7.5 or later from version 7.4 or later.
If you are updating from a release before version 7.4.0, follow the instructions in that previous release. However, you are required to confirm the update, which is not included in the instructions for your current release. After you follow the instructions for your release, return to the corresponding final instructions in this version. Follow the step that refers to receiving a status message and confirming the update.
For the most recent information about restrictions before you update, search for flashes, alerts, and bulletins at this support site:
| Sequence | Update task |
|---|---|
| 1 | Before you update, become familiar with the prerequisites and tasks involved. Decide whether you want to update automatically or update manually. During an automatic update procedure, the clustered system updates each of the nodes systematically. The automatic method is the preferred procedure for updating software on nodes. However, you can also update each node manually. |
| 2 | Ensure that CIM object manager (CIMOM) clients are working correctly. When necessary, update these clients so that they can support the new version of SAN Volume Controller code. |
| 3 | Ensure that multipathing drivers in the environment are fully redundant. |
| 4 | Update your system. The system update includes component firmware updates. The drive firmware update is a separate process. |
| 5 | Update other devices in the SAN Volume Controller environment. Examples might include updating hosts and switches to the correct levels. |
Note: The amount of time can vary depending on the amount of preparation work that is required and
the size of the environment. For automatic update, it takes about 20 minutes for
each node plus 30 minutes for each system. The 30-minute interval provides time for the multipathing
software to recover.
|
|
Firmware and software for the system and its attached adapters are tested and released as a single package. The package number increases each time that a new release is made.
Some code levels support updates only from specific previous levels, or the code can be installed only on certain hardware types. If you update to more than one level above your current level, you might be required to install an intermediate level. For example, if you are updating from level 1 to level 3, you might need to install level 2 before you can install level 3. For information about the prerequisites for each code level, see this website:
www.ibm.com/supportThe update process
During the automatic update process, each node in a system is updated one at a time, and the new code is staged on the nodes. While each node restarts, there might be some degradation in the maximum I/O rate that can be sustained by the system. After all the nodes in the system are successfully restarted with the new code level, the new level is automatically committed.
During an automatic code update, each node of a working pair is updated sequentially. The node that is being updated is temporarily unavailable and all I/O operations to that node fail. As a result, the I/O error counts increase and the failed I/O operations are directed to the partner node of the working pair. Applications do not see any I/O failures. When new nodes are added to the system, the update package is automatically downloaded to the new nodes from the SAN Volume Controller system.
The update can normally be done concurrently with normal user I/O operations. However, performance might be impacted. If any restrictions apply to the operations that can be done during the update, these restrictions are documented on the product website that you use to download the update packages. During the update procedure, most of the configuration commands are not available. Only the following commands are operational from the time the update process starts to the time that the new code level is committed, or until the process is backed out:
To determine when your update process completes, you are notified through the management GUI. If you are using the command-line interface, issue the lsupdate command to display the status of the update.
Because of the operational limitations that occur during the update process, the code update is a user task. However, if you have problems with an update, contact your support center. Do not try to troubleshoot update problems without technical assistance. For further directions, see the topic about how to get information, help, and technical assistance.
Multipathing driver
Before you update, ensure that the multipathing driver is fully redundant with every path available and online. You might see errors that are related to the paths that are going away (fail over) and the error count increasing during the update. When the paths to the nodes are back, the nodes fall back to become a fully redundant system. After the 30-minute delay, the paths to the other node go down.
If you are using IBM® Subsystem Device Driver (SDD) or IBM Subsystem Device Driver Device Specific Module (SDDDSM) as the multipathing software on the host, increased I/O error counts are displayed by the datapath query device or datapath query adapter commands to monitor the state of the multipathing software. For more information, see the IBM System Storage Multipath Subsystem Device Driver User's Guide for more information about the datapath query commands.
If you are using IBM Subsystem Device Driver Path Control Module (SDDPCM) as the multipathing software on the host, increased I/O error counts are displayed by the pcmpath query device or pcmpath query adapter commands to monitor the state of the multipathing software.
Updating SAN Volume Controller 2145-CG8 or 2145-CF8 systems with internal flash drives
The SAN Volume Controller update process reboots each node in the system in turn. Before the update commences and before each node is updated, the update process checks for dependent volumes.You can check for dependent volumes by using the lsdependentvdisks command-line interface (CLI) command with the node parameter.
- The update process takes each
node offline temporarily to process the update. While the node that contains an internal flash drive is offline, any data written to
volumes with a mirrored copy on the offline node are written only to the other online copy.
After the updated node rejoins
the system, data is resynchronized from the copy that remained online. The update process delays approximately 30
minutes before the update on the
partner node is started. The synchronization must complete within this time or the update stalls and requires manual
intervention. For any mirrored volume that uses disk extents on a flash drive that is on a SAN Volume Controller node for one or both of its volume
copies, set its synchronization rate set to 80 or above to ensure that the resynchronization
completes in time. Note: To increase the amount of time between the two nodes that contain volume copies and prevent them from going offline during the update process, consider manually updating the code.
- Table 2 defines the synchronization rates.
Table 2. Resynchronization rates of volume copies Synchronization rate Data copied/sec 1-10 128 KB 11-20 256 KB 21-30 512 KB 31-40 1 MB 41-50 2 MB 51-60 4 MB 61-70 8 MB 71-80 16 MB 81-90 32 MB 91-100 64 MB
- Updating systems with internal flash drives that use RAID 1 or 10
- The update process takes each node offline temporarily to
process the update. During this time, write operations to a mirrored
array on an offline node are written only to the drive that is in
the online node. When the node comes back online, the drive that was
offline is then resynchronized from the online mirrored array. However,
if this synchronization process does not complete before the partner
node needs to be updated,
the dependent volume process fails and the update stalls. Attention: To increase the amount of time between the two nodes going offline during the update process, consider manually updating the code.
Metro Mirror and Global Mirror relationships
When you update software on a system that has secondary volumes of running Metro Mirror or Global Mirror relationships, write performance might be degraded on the primary volumes, and Global Mirror relationships can be automatically stopped with one or more errors with error code 1920. You might want to proactively stop such relationships before you update the software to avoid the write performance degradation, and restart the relationships after the update completes.
chnodehw node id
chnodehw -legacy software_level node id
If the additional hardware is activated and a partnership is required to be established with a system that is running pre-6.4 software, then the additional hardware must be disabled first by using the chnodehw -legacy software version (pre 6.4) node id command.
When a node is added to a system, the system checks for (started) partnerships and determine the lowest software level of the partnered systems. This software level is passed to the node that is being added to the system. The node processes the equivalent of a chnodehw -legacy software level command as it joins the system.
After the system update
The audit log content that was on your system before the update is sent to a file in the/dumps/audit directory on the configuration node. The audit log will now contain content that occurs from commands that are run after a successful update of the system.