天天看点

ASM diskgroup dismount with "Waited 15 secs for write IO to PST" (文档 ID 1581684.1)

ASM diskgroup dismount with "Waited 15 secs for write IO to PST" (文档 ID

1581684.1)

ASM diskgroup dismount with "Waited 15 secs for write IO to PST" (文档 ID 1581684.1)
ASM diskgroup dismount with "Waited 15 secs for write IO to PST" (文档 ID 1581684.1)
ASM diskgroup dismount with "Waited 15 secs for write IO to PST" (文档 ID 1581684.1)

In this Document

<a target="_blank" href="https://support.oracle.com/epmos/faces/DocumentDisplay?_afrLoop=341592462704979&amp;id=1581684.1&amp;_adf.ctrl-state=ov8k99qxg_168#SYMPTOM">Symptoms</a>

<a target="_blank" href="https://support.oracle.com/epmos/faces/DocumentDisplay?_afrLoop=341592462704979&amp;id=1581684.1&amp;_adf.ctrl-state=ov8k99qxg_168#CAUSE">Cause</a>

<a target="_blank" href="https://support.oracle.com/epmos/faces/DocumentDisplay?_afrLoop=341592462704979&amp;id=1581684.1&amp;_adf.ctrl-state=ov8k99qxg_168#FIX">Solution</a>

<a target="_blank" href="https://support.oracle.com/epmos/faces/DocumentDisplay?_afrLoop=341592462704979&amp;id=1581684.1&amp;_adf.ctrl-state=ov8k99qxg_168#REF">References</a>

Oracle Database - Enterprise Edition - Version 11.2.0.3 to 12.1.0.1 [Release 11.2 to 12.1]

Information in this document applies to any platform.

Normal or high redundancy diskgroup is dismounted with these WARNING messages.

//ASM alert.log

Mon Jul 01 09:10:47 2013

WARNING: Waited 15 secs for write IO to PST disk 1 in group 6.

WARNING: Waited 15 secs for write IO to PST disk 4 in group 6.

....

GMON dismounting group 6 at 72 for pid 44, osid 8782162 

Generally this kind messages comes in ASM alertlog file on below situations,

Delayed ASM PST heart beats on ASM disks in normal or high redundancy diskgroup,

thus the ASM instance dismount the diskgroup.By default, it is 15 seconds.

By the way the heart beat delays are sort of ignored for external redundancy diskgroup.

ASM instance stop issuing more PST heart beat until it succeeds PST revalidation,

but the heart beat delays do not dismount external redundancy diskgroup directly.

The ASM disk could go into unresponsiveness, normally in the following scenarios:

+    Some of the paths of the physical paths of the multipath device are offline or lost

+    During path 'failover' in a multipath set up

+    Server load, or any sort of storage/multipath/OS maintenance

The Doc ID 10109915.8  briefs about Bug 10109915(this fix introduce this underscore parameter). And the issue is with no OS/Storage tunable timeout mechanism in a case of a Hung NFS Server/Filer. And then  _asm_hbeatiowait 

helps in setting the time out.

1]    Check with OS and Storage admin that there is disk unresponsiveness.

2]    Possibly keep the disk responsiveness to below 15 seconds. 

This will depend on various factors like

+    Operating System

+    Presence of Multipath ( and Multipath Type )

+    Any kernel parameter

So you need to find out, what is the 'maximum' possible disk unresponsiveness for your set up.

For example, on AIX  rw_timeout  setting affects this and defaults to 30 seconds.

Another example is Linux with native multipathing. In such set up, number of physical paths and  polling_interval value in multipath.conf file, will dictate this maximum disk unresponsiveness.

So for your set up ( combination of OS / multipath / storage ), you need to find out this.

3]    If you can not keep the disk unresponsiveness to below 15 seconds, then the below parameter can be set in the ASM instance ( on all the Nodes of RAC ):

    _asm_hbeatiowait

As per internal bug 17274537 , based on internal testing the value should be increased to 120 secs, the same will be fixed in 12.2

Run below in asm instance to set desired value for _asm_hbeatiowait

alter system set "_asm_hbeatiowait"=&lt;value&gt; scope=spfile sid='*';

And then restart asm instance / crs, to take new parameter value in effect.

继续阅读