Note - Disk paths are automatically added to the monitoring list monitored when I/O devices
are added to a node by using the cldevice command. Disk paths are
also automatically unmonitored when devices are removed from a node by using Oracle
Solaris Cluster commands.

The procedures in the following section that issue the cldevice command include the
disk-path argument. The disk-path argument consists of a node name and a disk
name. The node name is not required and defaults to all if you
do not specify it.

How to Monitor a Disk Path

Perform this task to monitor disk paths in your cluster.

Caution - DPM is not supported on nodes that run versions that were released prior
to Sun Cluster 3.1 10/03 software. Do not use DPM commands while
a rolling upgrade is in progress. After all nodes are upgraded, the nodes
must be online to use DPM commands.

The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global
cluster.

This procedure provides the long forms of the Oracle Solaris Cluster commands.
Most commands also have short forms. Except for the long and short forms
of the command names, the commands are identical.

Become superuser or assume a role that provides solaris.cluster.modify RBAC authorization on
any node in the cluster.

Monitor a disk path.

# cldevice monitor -n nodedisk

Verify that the disk path is monitored.

# cldevice status device

Example 5-38 Monitoring a Disk Path on a Single Node

The following example monitors the schost-1:/dev/did/rdsk/d1 disk path from a single node.
Only the DPM daemon on the node schost-1 monitors the path to
the disk /dev/did/dsk/d1 .

How to Unmonitor a Disk Path

Use this procedure to unmonitor a disk path.

Caution - DPM is not supported on nodes that run versions that were released prior
to Sun Cluster 3.1 10/03 software. Do not use DPM commands while
a rolling upgrade is in progress. After all nodes are upgraded, the nodes
must be online to use DPM commands.

The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global
cluster.

This procedure provides the long forms of the Oracle Solaris Cluster commands.
Most commands also have short forms. Except for the long and short forms
of the command names, the commands are identical.

Become superuser or assume a role that provides solaris.cluster.modify RBAC authorization on
any node in the cluster.

Determine the state of the disk path to unmonitor.

# cldevice status device

On each node, unmonitor the appropriate disk paths.

# cldevice unmonitor -n nodedisk

Example 5-41 Unmonitoring a Disk Path

The following example unmonitors the schost-2:/dev/did/rdsk/d1 disk path and prints disk paths
with status for the entire cluster.

How to Print Failed Disk Paths

Use the following procedure to print the faulted disk paths for a
cluster.

Caution - DPM is not supported on nodes that run versions that were released prior
to Sun Cluster 3.1 10/03 software. Do not use DPM commands while
a rolling upgrade is in progress. After all nodes are upgraded, the nodes
must be online to use DPM commands.

Become superuser on any node in the cluster.

Print the faulted disk paths throughout the cluster.

# cldevice status -s fail

Example 5-42 Printing Faulted Disk Paths

The following example prints faulted disk paths for the entire cluster.

How to Resolve a Disk-Path Status Error

If the following events occur, DPM might not update the status of
a failed path when it comes back online:

A monitored-path failure causes a node reboot.

The device under the monitored DID path does not come back online until after the rebooted node is back online.

The incorrect disk-path status is reported because the monitored DID device is unavailable
at boot time, and therefore the DID instance is not uploaded to the
DID driver. When this situation occurs, manually update the DID information.

From one node, update the global-devices namespace.

# cldevice populate

On each node, verify that command processing has completed before you proceed to
the next step.

The command executes remotely on all nodes, even though the command is run
from just one node. To determine whether the command has completed processing, run
the following command on each node of the cluster.

# ps -ef | grep cldevice populate

Verify that, within the DPM polling time frame, the status of the faulted
disk path is now Ok.

How to Monitor Disk Paths From a File

Use the following procedure to monitor or unmonitor disk paths from a file.

To change your cluster configuration by using a file, you must first export
the current configuration. This export operation creates an XML file that you can
then modify to set the configuration items you are changing. The instructions in
this procedure describe this entire process.

Caution - DPM is not supported on nodes that run versions that were released prior
to Sun Cluster 3.1 10/03 software. Do not use DPM commands while
a rolling upgrade is in progress. After all nodes are upgraded, the nodes
must be online to use DPM commands.

The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global
cluster.

This procedure provides the long forms of the Oracle Solaris Cluster commands.
Most commands also have short forms. Except for the long and short forms
of the command names, the commands are identical.

Become superuser or assume a role that provides solaris.cluster.modify RBAC authorization on
any node in the cluster.

Export your device configuration to an XML file.

# cldevice export -o configurationfile

-oconfigurationfile

Specify the file name for your XML file.

Modify the configuration file so that device paths are monitored.

Find the device paths that you want to monitor, and set the
monitored attribute to true.

Monitor the device paths.

# cldevice monitor -i configurationfile

-iconfigurationfile

Specify the file name of the modified XML file.

Verify that device path is now monitored.

# cldevice status

Example 5-43 Monitor Disk Paths From a File

In the following example, the device path between the node phys-schost–2 and
device d3 is monitored by using an XML file.

The first step is to export the current cluster configuration.

# cldevice export -o deviceconfig

The deviceconfig XML file shows that the path between phys-schost–2 and d3
is not currently monitored.

For more detail about exporting cluster configuration and using the resulting XML file
to set cluster configuration, see the cluster(1CL) and the clconfiguration(5CL) man pages.

How to Enable the Automatic Rebooting of a Node When All Monitored Shared-Disk Paths Fail

When you enable this feature, a node automatically reboots, provided that the following
conditions are met:

All monitored shared-disk paths on the node fail.

At least one of the monitored disks is accessible from a different node in the cluster. The scdpm daemon uses the private interconnect to check if disks are accessible from a different node in the cluster. If the private interconnect is disabled, the scdpm daemon cannot obtain the status of the disks from another node.

Rebooting the node restarts all resource groups and device groups that are mastered
on that node on another node.

If all monitored shared-disk paths on a node remain inaccessible after the node
automatically reboots, the node does not automatically reboot again. However, if any disk
paths become available after the node reboots but then fail, the node automatically
reboots again.

When you enable the reboot_on_path_failure property, the states of local-disk paths are not
considered when determining if a node reboot is necessary. Only monitored shared disks
are affected.

On any node in the cluster, become superuser or assume a role that
provides solaris.cluster.modify RBAC authorization.

For all nodes in the cluster, enable the automatic rebooting of a node
when all monitored shared-disk paths to it fail.

# clnode set -p reboot_on_path_failure=enabled +

How to Disable the Automatic Rebooting of a Node When All Monitored Shared-Disk Paths Fail

When you disable this feature and all monitored shared-disk paths on a node
fail, the node does not automatically reboot.

On any node in the cluster, become superuser or assume a role that
provides solaris.cluster.modify RBAC authorization.

For all nodes in the cluster, disable the automatic rebooting of a node
when monitored all monitored shared-disk paths to it fail.