Escolar Documentos
Profissional Documentos
Cultura Documentos
Release Notes
AIX
Legal Notice
Copyright 2013 Symantec Corporation. All rights reserved.
Symantec, the Symantec Logo, the Checkmark Logo, Veritas, Veritas Storage Foundation,
CommandCentral, NetBackup, Enterprise Vault, and LiveUpdate are trademarks or registered
trademarks of Symantec Corporation or its affiliates in the U.S. and other countries. Other
names may be trademarks of their respective owners.
The product described in this document is distributed under licenses restricting its use,
copying, distribution, and decompilation/reverse engineering. No part of this document
may be reproduced in any form by any means without prior written authorization of
Symantec Corporation and its licensors, if any.
THE DOCUMENTATION IS PROVIDED "AS IS" AND ALL EXPRESS OR IMPLIED CONDITIONS,
REPRESENTATIONS AND WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT,
ARE DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD TO
BE LEGALLY INVALID. SYMANTEC CORPORATION SHALL NOT BE LIABLE FOR INCIDENTAL
OR CONSEQUENTIAL DAMAGES IN CONNECTION WITH THE FURNISHING,
PERFORMANCE, OR USE OF THIS DOCUMENTATION. THE INFORMATION CONTAINED
IN THIS DOCUMENTATION IS SUBJECT TO CHANGE WITHOUT NOTICE.
The Licensed Software and Documentation are deemed to be commercial computer software
as defined in FAR 12.212 and subject to restricted rights as defined in FAR Section 52.227-19
"Commercial Computer Software - Restricted Rights" and DFARS 227.7202, "Rights in
Commercial Computer Software or Commercial Computer Software Documentation", as
applicable, and any successor regulations. Any use, modification, reproduction release,
performance, display or disclosure of the Licensed Software and Documentation by the U.S.
Government shall be solely in accordance with the terms of this Agreement.
Symantec Corporation
350 Ellis Street
Mountain View, CA 94043
http://www.symantec.com
Technical Support
Symantec Technical Support maintains support centers globally. Technical
Supports primary role is to respond to specific queries about product features
and functionality. The Technical Support group also creates content for our online
Knowledge Base. The Technical Support group works collaboratively with the
other functional areas within Symantec to answer your questions in a timely
fashion. For example, the Technical Support group works with Product Engineering
and Symantec Security Response to provide alerting services and virus definition
updates.
Symantecs support offerings include the following:
A range of support options that give you the flexibility to select the right
amount of service for any size organization
For information about Symantecs support offerings, you can visit our website at
the following URL:
www.symantec.com/business/support/index.jsp
All support services will be delivered in accordance with your support agreement
and the then-current enterprise technical support policy.
Hardware information
Operating system
Network topology
Problem description:
Customer service
Customer service information is available at the following URL:
www.symantec.com/business/support/
Customer Service is available to assist with non-technical questions, such as the
following types of issues:
customercare_apac@symantec.com
semea@symantec.com
supportsolutions@symantec.com
Documentation
Product guides are available on the media in PDF format. Make sure that you are
using the current version of the documentation. The document version appears
on page 2 of each guide. The latest product documentation is available on the
Symantec website.
https://sort.symantec.com/documents
Your feedback on product documentation is important to us. Send suggestions
for improvements and reports on errors or omissions. Include the title and
document version (located on the second page), and chapter and section titles of
the text on which you are reporting. Send feedback to:
docs@symantec.com
For information regarding the latest HOWTO articles, documentation updates,
or to ask a question regarding product documentation, visit the Storage and
Clustering Documentation forum on Symantec Connect.
https://www-secure.symantec.com/connect/storage-management/
forums/storage-and-clustering-documentation
Release Notes
This document includes the following topics:
Changes in version
Fixed issues
Known issues
Software limitations
Documentation errata
Documentation
Release Notes
Component product release notes
This is Document version: 5.1SP1.3 of the Veritas Cluster Server Release Notes.
Before you start, ensure that you are using the latest version of this guide. The
latest product documentation is available on the Symantec Web site at:
http://www.symantec.com/business/support/overview.jsp?pid=15107
Release Notes
About Symantec Operations Readiness Tools
VCS provides a framework that allows for the creation of custom agents. Create
agents in situations where the Veritas High Availability Agent Pack, the bundled
agents, or the enterprise agents do not meet your needs.
For more information about the creation of custom agents, refer to the Veritas
Cluster Server Agent Developer's Guide. You can also request a custom agent
through Symantec consulting services.
Analyze your current deployment and identify the Symantec products and
licenses you are using
Upload configuration data to the SORT Web site, so you can share information
with coworkers, managers, and Symantec Technical Support
Patches
10
Release Notes
Important release information
Note: Certain features of SORT are not available for all products.
To access SORT, go to:
http://sort.symantec.com
The latest product documentation is available on the Symantec Web site at:
http://www.symantec.com/business/support/overview.jsp?pid=15107
For important updates regarding this release, review the Late-Breaking News
TechNote on the Symantec Technical Support website:
http://entsupport.symantec.com/docs/334998
Changes in version
This section lists the changes for Veritas Cluster Server.
Release Notes
Changes in version
You can perform a rolling upgrade using the script-based or Web-based installer.
See the Veritas Cluster Server Installation Guide.
The installer detects and lists the aggregated links that you can choose to
configure as private heartbeat links.
The installer provides an option to detect NICs on each system and network
links, and sets link priority to configure LLT over Ethernet.
The installer can copy CPI scripts to any given location using
-copyinstallscripts option
The installer can copy CPI scripts to given location using -copyinstallscripts
option. This option is used when customers install SFHA products manually and
11
12
Release Notes
Changes in version
You need the superuser passwords for the systems where you plan to install
VCS.
Upgrade changes
The following lists the upgrade changes in this release.
Release Notes
Changes in version
5.0MP3 to 5.1SP1
5.1RP1 to 5.1SP1
5.1RP2 to 5 .1SP1
For supported upgrade matrix, refer to the Veritas Cluster Server Installation
Guide.
Upgrades that follow any other upgrade paths require a reboot.
Packaging updates
The following lists package changes in this release.
13
14
Release Notes
Changes in version
2.
3.
4.
5.
6.
7.
8.
9.
Release Notes
Changes in version
See the Veritas Cluster Server Administrator's Guide for more information.
The following agents are IMF-aware in VCS 5.1 SP1:
Mount
Process
Application
Oracle
Netlsnr
CFSMount
CVMVxconfigd
CFSfsckd
New attributes
The following sections describes the attributes introduced in VCS 5.1SP1, VCS
5.1, and VCS 5.0MP3.
EnvFile: This attribute specifies the environment file that must be sourced
before running StartProgram, StopProgram, MonitorProgram or CleanProgram.
UseSUDash: This attribute specifies that the agent must run su - user -c
<program> or su user -c <program> while running StartProgram,
StopProgram, MonitorProgram or CleanProgram.
ReturnIntOffline: This attribute can take one of the following three values.
These values are not mutually exclusive and can be used in combination with
one another. You must set IntentionalOffline attribute to 1 for the
ReturnIntOffline attribute to work.
15
16
Release Notes
Changes in version
NONE: The disk group is imported without SCSI-3 reservation. The agent
does not care about the cluster-level UseFence attribute.
Note: This attribute must be set to NONE for all resources of type DiskGroup
in case of non-SCSI-3 fencing.
IPMultiNICB agent attribute
Options: Value of this attribute is used as the option for the ifconfig command.
For example, if value of this attribute is set to metric 2 then metric 2 option
is set on the interface while bringing IPMultiNICB resource online.
Lower: Defines the position of the NFSRestart resource in the service group.
The NFSRestart resource below the Share resource needs a value of 1. The
NFSRestart resource on the top of the resource dependency tree has a Lower
attribute value of 0.
Release Notes
Changes in version
NotifierSourceIP: : Lets you specify the interface that the notifier must use to
send packets. This attribute is string/scalar. You must specify an IP address
that is either DNS resolvable or appears in the/etc/hosts file.
PidFile: The absolute path to the Samba daemon (smbd) Pid file. This attribute
is mandatory if you are using Samba configuration file with non-default name
or path.
SocketAddress: The IPv4 address where the Samba daemon (smbd) listens for
connections. This attribute is mandatory if you are configuring multiple
SambaServer resources on a node.
PidFile: The absolute path to the Samba daemon (nmbd) PidFile. This attribute
is mandatory if you are using Samba configuration file with non-default name
or path.
Cluster-level attributes
17
18
Release Notes
Changes in version
LogOnly logs the message in Engine Log and SysLog. Trigger invokes a trigger
which has a default action of collecting the comms tar file. The Default value
of Trigger is LogOnly.
Release Notes
Changes in version
If set, the agent configures the WPAR to use only the resource set specified
by this attribute.
WorkLoad: Allows modification of resource control attributes of WPAR shares_CPU and shares_memory. This attribute has two keys, CPU and MEM.
The key CPU is used to specify the number of processor shares that are available
to the workload partition. The key MEM is used to specify the number of
memory shares that are available to the workload partition.
EPClass: Enables you to control the scheduling class for the agent functions
(entry points) except the online entry point.
EPPriority: Enables you to control the scheduling priority for the agent
functions (entry points) except the online entry point.
OnlineClass: Enables you to control the scheduling class for the online agent
function (entry point).
OnlinePriority: Enables you to control the scheduling priority for the online
agent function (entry point).
ContainerInfo: Specifies information about the container that the service group
manages.
19
20
Release Notes
Changes in version
AgentFile: Complete name and path of the binary for an agent. Use when the
agent binaries are not installed at their default locations.
AgentDirectory: Complete path of the directory in which the agent binary and
scripts are located. Use when the agent binaries are not installed at their default
locations.
Guests: List of users that have Guest privileges on the service group.
Release Notes
Changes in version
VolumeSet agent: The VolumeSet agent brings online, takes offline, and
monitors a Veritas Volume Manager (VxVM) volume set. Use this agent to
make a volume set highly available.
LVMVG agent
21
22
Release Notes
Changes in version
The RemoteGroup resource does not take the remote service group offline if
the resource is online anywhere in the cluster.
After an agent restarts, the RemoteGroup resource does not return offline if
the resource is online on another cluster node.
The RemoteGroup resource takes the remote service group offline if it is the
only instance of RemoteGroup resource online in the cluster.
VCS SambaServer and NetBios agents now support multiple resource instances
on a node.
These agents now support non default Samba configuration file and pid file
names.
For more information about bundled agents, refer to the Veritas Cluster Server
Bundled Agents Reference Guide.
Release Notes
Changes in version
The VCS agent binaries for Sybase are now a part of VRTSvcsea package. This
package also includes the VCS agent binaries for DB2 and Oracle.
Changes to LLT
This release includes the following new features and changes to LLT:
LLT startup time through the LLT init script is now optimized to use a constant
time. LLT takes less than 16 seconds to start irrespective of the number of
links specified in /etc/llttab file.
In the previous releases, LLT took around (5 * number_of_links_specified
_in_the_/etc/llttab_file) seconds to start.
23
24
Release Notes
Changes in version
See the Veritas Cluster Server Installation Guide and the Veritas Cluster Server
Administrator's Guide for more details.
Changes to GAB
This section lists the new features and changes related to GAB in this release.
Release Notes
Changes in version
Manually resynchronize the original Primary with the data from the new
Primary, once the original Primary comes back up. The applications are only
active on the new Primary site.
Automatically resynchronize the original Primary with the data from the new
Primary, once the original Primary comes back up. The applications are only
active on the new Primary site.
25
26
Release Notes
VCS system requirements
Beginning in Release 5.1SP1, you have a third option. Applications can be active
on both the original Primary and Secondary sites. After the original Primary
returns or the network disruption is corrected, you have the option of specifying
which site is the Primary going forward. This option is called the primary-elect
feature, and it is enabled through the VCS global clustering.
The key difference between the primary-elect feature and the other options is
that if a network disruption occurs, applications continue to run on the Primary
site and they are also failed over to the Secondary. This feature lets you maintain
application availability on both sites while the network is down.
Release Notes
VCS system requirements
For important updates regarding this release, review the Late-Breaking News
TechNote on the Symantec Technical Support website:
http://entsupport.symantec.com/docs/334998
Journaled File System (JFS) and Enhanced Journaled File System (JFS2) on
LVM
Veritas Storage Foundation (SF): Veritas Volume Manager (VxVM) with Veritas
File System (VxFS)
VCS 5.1 SP1 supports the following versions of SF:
SF 5.1 SP1
SF 5.1
Note: VCS supports the previous version of SF and the next version of SF to
facilitate product upgrades.
27
28
Release Notes
Features no longer supported
Table 1-1
Agent
Application
DB2
DB2 Enterprise
Server Edition
Oracle
Oracle
11gR1,
AIX 5.3, AIX 6.1
10gR2, 11gR2
Sybase
Sybase Adaptive
Server Enterprise
12.5.x, 15.x
See the Veritas Cluster Server Installation Guide for the agent for more details.
For a list of the VCS application agents and the software that the agents support,
see the Veritas Cluster Server Agents Support Matrix at Symantec website.
Configuration wizards
Disk agent
ServiceGroupHB agent.
This release does not support disk heartbeats. Symantec recommends using
I/O fencing.
VRTSWebApp
Oracle 8.0.x, Oracle 8.1.x, and Oracle 9i - not supported by the updated Oracle
agent.
Release Notes
Fixed issues
Fixed issues
This section covers the incidents that are fixed in this release.
This release includes fixed issues from the 5.1 Service Pack (SP) 1 Rolling Patch
(RP) 2 release. For the list of fixed issues in the 5.1 SP1 RP2 release, see the Veritas
Storage Foundation and High Availability Solutions 5.1 SP1 RP2 Release Notes.
See the corresponding Release Notes for a complete list of fixed incidents related
to that product.
Incident
Description
1908938
[GAB] In a large cluster, cascaded lowest node failures result in GAB panic
during sequence space recovery.
1840826
[GAB] Prevent 'gabconfig -c' while port 'a' is in the middle of iofence
processing.
2096584
[LLT] LLT unload may fail but should not cause hang/panic on AIX when
clients are registered.
1861439
[LLT] Removing the LLT links from a single node in a four-node cluster
causes other nodes to panic.
1849527
2066020
[LLT] The dlpiping utility exits with an error similar to "dlpiping: send
ECHO_REQ failed."
2005045
[LLT] The hastart command fails to start HAD on one of the nodes with
message GabHandle::open failed errno = 16 in syslog after HAD is stopped
on all the nodes in the cluster simultaneously.
1859023
29
30
Release Notes
Fixed issues
Table 1-2
Incident
Description
1846387
[Fencing] The vxfenswap and the vxfentsthdw utilities fail when rsh or
ssh communication is not set to the same node.
2084121
1922413
1847517
1992560
[Fencing] The vxfentsthdw utility uses scp to communicate with the local
host.
1512956
2143933
2151166
Incident
Description
251704
252354
2001039
Release Notes
Fixed issues
Table 1-3
Incident
Description
2019904
2045972
2111972
Incident
Description
254693
970971
1074707
1472734
1634494
1859387
31
32
Release Notes
Fixed issues
Table 1-4
Incident
Description
1861740
1866434
1874261
1874533
1915908
The hares command lets you have the "." special character
in a resource name.
1958245
1971256
2022123
2061932
2081725
2083232
2084961
Release Notes
Fixed issues
Table 1-4
Incident
Description
2111296
2128658
Table 1-5
Incident
Description
776230
2117378
Fixed
issues
Description
1967207
1946367
1941647
1922411
vxfentsthdw must detect storage arrays which interpret NULL keys as valid
for registrations/reservations.
1916004
33
34
Release Notes
Fixed issues
Table 1-6
Fixed
issues
Description
1915909
The hares command lets you create resources which has the "." special
character
1915016
1900450
1874267
1870424
1855196
System panics due to depleted memory during GAB broadcast stress and
reboot node 0.
1839091
Fixed
issues
Description
1986311
2011416
2019899
2031931
2059631
LLT links go off/on repeatedly after configuring LLT over udp in IPv6.
2069689
2070743
HAD gets stuck in a tight loop when HACLI forks and membership changes.
2077396
VCS dumps core under heavy load system which results in node panic.
2086251
Release Notes
Fixed issues
Table 1-7
Fixed
issues
Description
2086280
2102777
RemoteGroup resource must support remote groups that are a parallel SG.
2083268
2083181
2100421
2108816
2087190
2110465
vxfenclearpre cannot clear keys from coordinator disks and data disks when
there's a preexisting split brain.
2111294
Some VCS nodes in a group are online and some are not.
2116161
2119570
2120020
2125442
HAD must bring online the failover service group only if it is offline on all
nodes after the service group is unfrozen.
2126871
1937834
Partial parent group must auto-start after child comes online, and service
group must fail over after node panic.
2066967
2071549
Under heavy load, HAD may wait indefinitely in waitpid() when user requests
to run some command through "hacli". This may cause GAB to kill HAD.
2077363
Overriding static attribute for any resource with empty type definition
creates duplicate entries in main.cf
2077375
35
36
Release Notes
Known issues
Table 1-7
Fixed
issues
Description
2079617
2079622
nfs splitbrain test: Delay forced import of disk group agent to avoid errors
to client nodes.
2079664
2080703
2084977
2089182
The vxcpserv process memory keeps growing when you run cpsadm
commands.
2102764
RemoteGroup resource appears OFFLINE even after applying patch for issue
1834858.
2110465
vxfenclearpre cannot clear keys from coordinator disks and data disks when
there is a preexisting split brain.
Known issues
This section covers the known issues in this release.
See the corresponding Release Notes for a complete list of known issues related
to that product.
See Documentation on page 64.
APAR IZ65498
APAR IZ64768
For versions earlier than AIX 6.1 TL4, contact IBM for a suitable APAR.
Release Notes
Known issues
Workaround:
The warning is due to a software error and can be safely ignored.
37
38
Release Notes
Known issues
Before upgrading SFHA from 5.0 MP3 RP2 to 5.1SP1, you must apply the I/O
fencing hotfix 5.0MP3RP2HF2.
Note down the list of products configured on the node for keyless licensing.
# vxkeyless display
2.
3.
4.
Release Notes
Known issues
1.
Note down the list of products configured on the node for keyless licensing.
Run vxkeyless display to display the list.
2.
3.
Find and delete the keyless licenses left over in the system. To do this, perform
the following steps for every key stored in /etc/vx/licenses/lic:
Verify if the key has VXKEYLESS feature Enabled using the following
command:
# vxlicrep -k <license_key> | grep VXKEYLESS
4.
39
40
Release Notes
Known issues
Normally, the number of values in ArgListValues for a resource must not exceed
425. However, in case of a keylist, association or vector type of attribute appears
in the ArgList for a resource-type. Since these attributes can take multiple values,
there is a chance for the resource values in ArgListValues to exceed 425.
Release Notes
Known issues
localhost
Workaround: There is currently no workaround for this issue. This issue has no
impact on any functionality.
41
42
Release Notes
Known issues
Workaround
The installer will be modified in future to prevent automatic addition of nodes to
SystemList and AutoStartList.
As a workaround, use the following commands to remove the nodes from the
SystemList and AutoStartList:
# hagrp modify ClusterService SystemList delete newnode
# hagrp modify ClusterService AutoStartList delete newnode
Release Notes
Known issues
Edit the OracleTypes.cf file on one of the nodes. Add the following line to
Netlsnr and Oracle type:
static int ContainerOpts{}
= { RunInContainer=1, PassCInfo=0 }
After VCS has started on the node, start VCS on other cluster nodes:
# hastart
VCS agent for Oracle: Health check monitoring does not work
with Oracle 10.2.0.4
The health check monitoring in Oracle agent does not work with Oracle 10.2.0.4
due to incompatibility of the health check APIs provided by Oracle. [2101570]
Resolution: Disable health check monitoring by setting the MonitorOption attribute
to 0 (zero).
43
44
Release Notes
Known issues
VCS agent for Oracle: Make sure that the ohasd has an entry
in the init scripts
Make sure that the ohasd process has an entry in the init scripts so that when the
process is killed or the machine is rebooted, this automatically restarts the
process.[1985093]
Workaround: Respawn of ohasd process. Add the ohasd process in
the/etc/inittab file to ensure that this process is automatically restarted when
killed or the machine is rebooted.
Release Notes
Known issues
Where rD is the name of the dependent resource, and target_res_name is the name
of the reference attribute that contains the name of the target resource.
Set the value of the reference attribute (target_res_name) to the name of the target
resource (rT).
# hares -modify rD target_res_name rT
45
46
Release Notes
Known issues
FFDC_AGFWTimer_729_agent_type.log core
agent_typeAgent_stack_729.txt
Workaround: If you see that both of the above criteria are true, increase the value
of the AgentReplyTimeout attribute value. (Up to 300 seconds or as necessary.)
[1853285]
Release Notes
Known issues
The engine log file receives too many log messages on the
secure site in global cluster environments
When the WAC process runs in secure mode on one site, and the other site does
not use secure mode, the engine log file on the secure site gets logs every five
seconds. [1539646]
Workaround: The two WAC processes in global clusters must always be started
in either secure or non-secure mode. The secure and non-secure WAC connections
will flood the engine log file with the above messages.
Both these variables are integers. With constant traffic, recvbytes hits and rolls
over MAX_INT quickly. This can cause the value of recvbytes to be less than the
value of recvcnt. [1788315]
This does not impact the LLT functionality.
47
48
Release Notes
Known issues
occurs in some corner cases even though a port is not even registered on the peer
node. [1809827]
All nodes in a sub-cluster panic if the node that races for I/O
fencing panics
At the time of a network partition the lowest node in each sub-cluster races for
the coordination points on behalf of that sub-cluster. If the lowest node is unable
to contact a majority of the coordination points or the lowest node itself
unexpectedly panics during the race, then all the nodes in that sub-cluster will
panic. [1965954]
The fencing setup uses customized mode with one or more CP servers.
The node weight for a single node (say galaxy with node id 0) is more than the
sum total of node weights for the rest of the nodes.
A network fault occurs and the cluster partitions into two with the single node
(galaxy) on one part and the rest of the nodes on the other part.
Under such circumstances, for group-based preferred fencing, the single node
panics even though more high priority services are online on that node. For
system-based preferred fencing, the single node panics even though more weight
is assigned to the node. [2161816]
See the Veritas Cluster Server Administrator's Guide for more information on
preferred fencing.
Release Notes
Known issues
Reconfiguring VCS with I/O fencing fails if you use the same
CP servers
When you reconfigure an application cluster that uses server-based I/O fencing
(customized fencing mode), the installer does not remove the application cluster
information from the CP servers before the reconfiguration. As a result, if you
reconfigure the application cluster and choose to configure I/O fencing in
customized mode using the same CP servers, then reconfiguration of server-based
fencing for the application cluster fails. [2076240]
Workaround: Manually remove the application cluster information from the CP
servers after you reconfigure VCS but before you reconfigure server-based I/O
fencing for the application cluster.
See the Veritas Cluster Server Administrator's Guide for instructions to remove
the application cluster information from the CP servers.
49
50
Release Notes
Known issues
Workaround: After rebooting on one system, delay the reboot on the next system
by 180 seconds.
/opt/VRTScps/bin/cpsat
/opt/VRTSvcs/bin/vcsat
This issue occurs when the command requires some user interaction. [1841185]
Workaround:
Release Notes
Software limitations
This causes the RVGPrimary online at the failover site to fail and the following
message is displayed:
RVGPrimary:RVGPrimary:online:Diskgroup bunkerdgname
could not be imported on bunker host hostname. Operation
failed with error 256 and message VxVM
VVR vradmin ERROR V-5-52-901 NETWORK ERROR: Remote server
unreachable...
Timestamp VCS ERROR V-16-2-13066 (hostname) Agent is calling
clean for resource(RVGPrimary) because the resource
is not up even after online completed.
Software limitations
This section covers the software limitations of this release.
See Documentation on page 64.
51
52
Release Notes
Software limitations
Any group that you defined as VCShmg along with all its resources.
Any resource type that you defined as HostMonitor along with all the resources
of such resource type.
Release Notes
Software limitations
IMF registration on Linux for bind file system type is not supported.
IMF should not be enabled for the resources where the BlockDevice can get
mounted on multiple MountPoints.
If FSType attribute value is nfs, then IMF registration for nfs file system
type is not supported.
53
54
Release Notes
Software limitations
The default value for the DB2 agent RestartLimit is 3. This higher value spreads
out the re-start of the DB2 resources (after a resource online failure), which lowers
the chances of DB2 resources all starting simultaneously. [1231311]
Total number of clusters in a global cluster configuration can not exceed four.
VxVM site for the disk group remains detached after node reboot in
campus clusters with fire drill
When you bring the DiskGroupSnap resource online, the DiskGroupSnap agent
detaches the site from the target disk group defined. The DiskGroupSnap agent
Release Notes
Software limitations
invokes VCS action entry points to run VxVM commands to detach the site. These
commands must be run on the node where the disk group is imported, which is
at the primary site.
If you attempt to shut down the node where the fire drill service group or the disk
group is online, the node goes to a LEAVING state. The VCS engine attempts to
take all the service groups offline on that node and rejects all action entry point
requests. Therefore, the DiskGroupSnap agent cannot invoke the action to reattach
the fire drill site to the target disk group. The agent logs a message that the node
is in a leaving state and then removes the lock file. The agents monitor function
declares that the resource is offline. After the node restarts, the disk group site
still remains detached. [1272012]
Workaround:
You must take the fire drill service group offline using the hagrp -offline
command before you shut down the node or before you stop VCS locally.
If the node has restarted, you must manually reattach the fire drill site to the disk
group that is imported at the primary site.
If the secondary node has crashed or restarted, you must manually reattach the
fire drill site to the target disk group that is imported at the primary site using
the following command: /opt/VRTSvcs/bin/hares -action $targetres joindg
-actionargs $fdsitename $is_fenced -sys $targetsys.
If you use the Bronze configuration for the DiskGroupSnap resource, you could
end up with inconsistent data at the secondary site in the following cases:
[1391445]
After the fire drill service group is brought online, a disaster occurs at the
primary site during the fire drill.
After the fire drill service group is taken offline, a disaster occurs at the
primary while the disks at the secondary are resynchronizing.
Symantec recommends that you use the Gold configuration for the
DiskGroupSnap resource.
55
56
Release Notes
Software limitations
split the CPU, memory, and certain adapters (networking and storage), into smaller
virtual units that the partitions can then use.
In an Advanced POWER Virtualization (APV) environment, AIX uses the VIO
Server to monitor and manage the I/O paths for the virtualized client partitions.
At a very high level, the VIO server provides a partitions access to storage that
is external to the physical computer. The VIO server encapsulates the physical
hardware into virtual adapters called virtual SCSI adapters (server adapter). On
the client side, you can create virtual adapters (client adapters) that map to the
server adapter and enable a partition to connect to external storage.
The VIO server provides similar mechanisms to share limited networking resources
across partitions. Refer to the manual that came with your system to help set up
partitions, and to configure and use the various components such as VIO server
and HMC, which are integral parts of IBM's APV environment.
The minimum patch level for using VIO servers with VCS is: Fix Pack 7.1.2.0.0.
Supported storage
Refer to the IBM data sheet:
http://www14.software.ibm.com/webapp/set2/sas/f/vios/home.html
Disk Restrictions
When using VCS in combination with VIO servers and their client partitions, you
need to ensure that no reservations are placed on the shared storage. This enables
client partitions on different systems to access and use the same shared storage.
If the shared storage is under MPIO control, set the reserve_policy attribute
of the disk to no_reserve.
If the shared storage is not under MPIO control, look up the array
documentation to locate a similar attribute to set on the disk.
Internal testing on EMC disks shows that this field maps as the reserve_lock
attribute for EMC disks. In this case, setting it to no achieves the same result.
Release Notes
Software limitations
that you have mapped the physical adapters to the client partitions to provide
access to the external shared storage.
To create a shareable diskgroup, you need to ensure that the different partitions
use the same set of disks. A good way to make sure that the disks (that are seen
from multiple partitions) are the same is to use the disks serial numbers, which
are unique.
Run the following commands on the VIO server (in non-root mode), unless
otherwise noted.
Get the serial number of the disk of interest:
$ lsdev -dev hdisk20 -vpd
hdisk20
U787A.001.DNZ06TT-P1-C6-T1-W500507630308037CL401 0401A00000000 IBM FC 2107
Manufacturer................IBM
Machine Type and Model......2107900
Serial Number...............7548111101A
EC Level.....................131
Device Specific.(Z0)........10
Device Specific.(Z1)........0100
Make sure the other VIO server returns the same serial number. This ensures that
you are viewing the same actual physical disk.
List the virtual SCSI adapters.
$ lsdev -virtual | grep vhost
vhost0 Available Virtual SCSI Server Adapter
vhost1 Available Virtual SCSI Server Adapter
Note: Usually vhost0 is the adapter for the internal disks. vhost1 in the example
above maps the SCSI adapter to the external shared storage.
Prior to mapping hdisk20 (in the example) to a SCSI adapter, change the
reservation policy on the disk.
$ chdev -dev hdisk20 -attr reserve_policy=no_reserve
hdisk20 changed
For hdisk20 (in the example) to be available to client partitions, map it to a suitable
virtual SCSI adapter.
57
58
Release Notes
Software limitations
If you now print the reserve policy on hdisk20 the output resembles:
$ lsdev -dev hdisk20 attr reserve_policy
value
no_reserve
Finally on the client partition run the cfgmgr command to make this disk visible
via the client SCSI adapter.
You can use this disk (hdisk20 physical, and known as mp1_hdisk5 on the client
partitions) to create a diskgroup, a shared volume, and eventually a shared file
system.
Perform regular VCS operations on the clients vis-a-vis service groups, resources,
resource attributes, etc.
Cluster Manager does not work if the hosts file contains IPv6
entries
VCS Cluster Manager fails to connect to the VCS engine if the /etc/hosts file
contains IPv6 entries.
Release Notes
Software limitations
The operating system does not distinguish between IPv4 and IPv6
packet counts
In a dual-stack configuration, when you use packet counts and the IPv6 network
is disabled, the NIC agent might not detect a faulted NIC. It might not detect a
fault because while the IPv6 network is down its packet count still increases. The
packet count increases because the operating system does not distinguish between
the packet counts for IPv4 and IPv6 networks. The agent then concludes that the
NIC is up. If you are using the same NIC device for IPv4 as well as IPv6 resources,
set PingOptimize to 0 and specify a value for the NetworkHosts attribute for either
the IPv6 or the IPv4 NIC resource. [1061253]
A service group that runs inside of a WPAR may not fail over when its
network connection is lost
For a WPAR configuration when the WPAR root is on NFS, the WPAR service
group may not fail over if the NFS connection is lost. This issue is due to an AIX
operating system limitation. [1637430]
LLT over IPv6 UDP cannot detect other nodes while VCS tries
to form a cluster
LLT over IPv6 requires link-local scope multicast to discover other nodes when
VCS tries to form a cluster. If multicast networking is undesirable, or unavailable
in your environment, use the address of the peer nodes to eliminate the need for
the multicast traffic. [1533308]
Workaround: Add the set-addr entry for each local link into the /etc/llttab file.
You add the entry to specify the address of the peer nodes that are available on
the corresponding peer links. For example, you add the following lines into the
llttab file to specify the set-addr entry for a node. In this example, the node's IPv6
address is fe80::21a:64ff:fe92:1d70.
59
60
Release Notes
Software limitations
After you reboot the systems, open the system console using any available
method, for example, from HMC.
On the console, go to the terminal setting menu, and set the terminal of your
choice.
Release Notes
Documentation errata
Guide for a description of the problems a failed interconnect can create and the
protection I/O fencing provides.
I/O fencing uses SCSI-3 PR keys to implement data protection. Keys are placed
on I/O fencing coordinator points and on data disks. The VCS administrator must
be aware of several operational changes needed when working with clusters
protected by I/O fencing. Specific shutdown procedures ensure keys are removed
from coordinator points and data disks to prevent possible difficulties with
subsequent cluster startup.
Using the reboot command rather than the shutdown command bypasses shutdown
scripts and can leave keys on the coordinator points and data disks. Depending
on the order of reboot and subsequent startup events, the cluster may warn of a
possible split brain condition and fail to start up.
Workaround: Use the shutdown -r command on one node at a time and wait for
each node to complete shutdown.
Documentation errata
The following sections, if present, cover additions or corrections for Document
version: 5.1SP1.3 of the product documentation. These additions or corrections
may be included in later versions of the product documentation that can be
downloaded from the Symantec Support website and the Symantec Operations
Readiness Tools (SORT).
See the corresponding Release Notes for documentation errata related to that
component or product.
See Documentation on page 64.
See About Symantec Operations Readiness Tools on page 9.
61
62
Release Notes
Documentation errata
Make sure that the VCS cluster is online and check that the fencing mode is
customized.
# vxfenadm -d
On each node, set the value of the LLT sendhbcap timer parameter value as
follows:
Add the following line to the /etc/llttab file so that the changes remain
persistent after any reboot:
set-timer senhbcap:3000
Release Notes
Documentation errata
For each resource of the type DiskGroup, set the value of the
MonitorReservation attribute to 0 and the value of the Reservation
attribute to NONE.
# hares -modify <dg_resource> MonitorReservation 0
# hares -modify <dg_resource> Reservation "NONE"
Make sure that the UseFence attribute in the VCS configuration file main.cf
is set to SCSI3.
10 To make these VxFEN changes take effect, stop and restart VxFEN and the
dependent modules
After VCS takes all services offline, run the following command to stop
VxFEN:
# /etc/init.d/vxfen.rc stop
On each node, run the following commands to restart VxFEN and VCS:
# /etc/init.d/vxfen.rc start
# /etc/init.d/vcs.rc start
63
64
Release Notes
Documentation
Documentation
Product guides are available on the documentation disc in PDF formats. Symantec
recommends copying pertinent information, such as installation guides and release
notes, from the disc to your system's /opt/VRTS/docs directory for reference.
Documentation set
Table 1-8 lists the documents for Veritas Cluster Server.
Table 1-8
Title
File name
vcs_install_51SP1_aix.pdf
vcs_notes_51SP1_aix.pdf
vcs_admin_51SP1_aix.pdf
vcs_agent_dev_51sp1pr4.pdf
Release Notes
Documentation
Table 1-8
Title
File name
Document title
File name
vvr_planning_51SP1_aix.pdf
vvr_advisor_users_51SP1_aix.pdf
Table 1-10 lists the documentation for Symantec Product Authentication Service
(AT).
Table 1-10
Title
File name
vxat_admin.pdf
Manual pages
The manual pages for Veritas Storage Foundation and High Availability Solutions
products are installed in the /opt/VRTS/man directory.
Set the MANPATH environment variable so the man(1) command can point to the
Veritas Storage Foundation manual pages:
For the Bourne or Korn shell (sh or ksh), enter the following commands:
65
66
Release Notes
Documentation
MANPATH=$MANPATH:/opt/VRTS/man
export MANPATH