1ea8dc4b6Seschrock /* 2ea8dc4b6Seschrock * CDDL HEADER START 3ea8dc4b6Seschrock * 4ea8dc4b6Seschrock * The contents of this file are subject to the terms of the 5ea8dc4b6Seschrock * Common Development and Distribution License (the "License"). 6ea8dc4b6Seschrock * You may not use this file except in compliance with the License. 7ea8dc4b6Seschrock * 8ea8dc4b6Seschrock * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9ea8dc4b6Seschrock * or http://www.opensolaris.org/os/licensing. 10ea8dc4b6Seschrock * See the License for the specific language governing permissions 11ea8dc4b6Seschrock * and limitations under the License. 12ea8dc4b6Seschrock * 13ea8dc4b6Seschrock * When distributing Covered Code, include this CDDL HEADER in each 14ea8dc4b6Seschrock * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15ea8dc4b6Seschrock * If applicable, add the following below this CDDL HEADER, with the 16ea8dc4b6Seschrock * fields enclosed by brackets "[]" replaced with your own identifying 17ea8dc4b6Seschrock * information: Portions Copyright [yyyy] [name of copyright owner] 18ea8dc4b6Seschrock * 19ea8dc4b6Seschrock * CDDL HEADER END 20ea8dc4b6Seschrock */ 21ea8dc4b6Seschrock /* 22e7cbe64fSgw * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23ea8dc4b6Seschrock * Use is subject to license terms. 24ea8dc4b6Seschrock */ 25ea8dc4b6Seschrock 26ea8dc4b6Seschrock #pragma ident "%Z%%M% %I% %E% SMI" 27ea8dc4b6Seschrock 28ea8dc4b6Seschrock #include <sys/spa.h> 29ea8dc4b6Seschrock #include <sys/spa_impl.h> 30ea8dc4b6Seschrock #include <sys/vdev.h> 31ea8dc4b6Seschrock #include <sys/vdev_impl.h> 32ea8dc4b6Seschrock #include <sys/zio.h> 33ea8dc4b6Seschrock 34ea8dc4b6Seschrock #include <sys/fm/fs/zfs.h> 35ea8dc4b6Seschrock #include <sys/fm/protocol.h> 36ea8dc4b6Seschrock #include <sys/fm/util.h> 37ea8dc4b6Seschrock #include <sys/sysevent.h> 38ea8dc4b6Seschrock 39ea8dc4b6Seschrock /* 40ea8dc4b6Seschrock * This general routine is responsible for generating all the different ZFS 41ea8dc4b6Seschrock * ereports. The payload is dependent on the class, and which arguments are 42ea8dc4b6Seschrock * supplied to the function: 43ea8dc4b6Seschrock * 44ea8dc4b6Seschrock * EREPORT POOL VDEV IO 45ea8dc4b6Seschrock * block X X X 46ea8dc4b6Seschrock * data X X 47ea8dc4b6Seschrock * device X X 48ea8dc4b6Seschrock * pool X 49ea8dc4b6Seschrock * 50ea8dc4b6Seschrock * If we are in a loading state, all errors are chained together by the same 5132b87932Sek * SPA-wide ENA (Error Numeric Association). 52ea8dc4b6Seschrock * 53ea8dc4b6Seschrock * For isolated I/O requests, we get the ENA from the zio_t. The propagation 54ea8dc4b6Seschrock * gets very complicated due to RAID-Z, gang blocks, and vdev caching. We want 55ea8dc4b6Seschrock * to chain together all ereports associated with a logical piece of data. For 56ea8dc4b6Seschrock * read I/Os, there are basically three 'types' of I/O, which form a roughly 57ea8dc4b6Seschrock * layered diagram: 58ea8dc4b6Seschrock * 59ea8dc4b6Seschrock * +---------------+ 60ea8dc4b6Seschrock * | Aggregate I/O | No associated logical data or device 61ea8dc4b6Seschrock * +---------------+ 62ea8dc4b6Seschrock * | 63ea8dc4b6Seschrock * V 64ea8dc4b6Seschrock * +---------------+ Reads associated with a piece of logical data. 65ea8dc4b6Seschrock * | Read I/O | This includes reads on behalf of RAID-Z, 66ea8dc4b6Seschrock * +---------------+ mirrors, gang blocks, retries, etc. 67ea8dc4b6Seschrock * | 68ea8dc4b6Seschrock * V 69ea8dc4b6Seschrock * +---------------+ Reads associated with a particular device, but 70ea8dc4b6Seschrock * | Physical I/O | no logical data. Issued as part of vdev caching 71ea8dc4b6Seschrock * +---------------+ and I/O aggregation. 72ea8dc4b6Seschrock * 73ea8dc4b6Seschrock * Note that 'physical I/O' here is not the same terminology as used in the rest 74ea8dc4b6Seschrock * of ZIO. Typically, 'physical I/O' simply means that there is no attached 75ea8dc4b6Seschrock * blockpointer. But I/O with no associated block pointer can still be related 76ea8dc4b6Seschrock * to a logical piece of data (i.e. RAID-Z requests). 77ea8dc4b6Seschrock * 78ea8dc4b6Seschrock * Purely physical I/O always have unique ENAs. They are not related to a 79ea8dc4b6Seschrock * particular piece of logical data, and therefore cannot be chained together. 80ea8dc4b6Seschrock * We still generate an ereport, but the DE doesn't correlate it with any 81ea8dc4b6Seschrock * logical piece of data. When such an I/O fails, the delegated I/O requests 82ea8dc4b6Seschrock * will issue a retry, which will trigger the 'real' ereport with the correct 83ea8dc4b6Seschrock * ENA. 84ea8dc4b6Seschrock * 85ea8dc4b6Seschrock * We keep track of the ENA for a ZIO chain through the 'io_logical' member. 86ea8dc4b6Seschrock * When a new logical I/O is issued, we set this to point to itself. Child I/Os 87ea8dc4b6Seschrock * then inherit this pointer, so that when it is first set subsequent failures 88ea8dc4b6Seschrock * will use the same ENA. If a physical I/O is issued (by passing the 89ea8dc4b6Seschrock * ZIO_FLAG_NOBOOKMARK flag), then this pointer is reset, guaranteeing that a 90ea8dc4b6Seschrock * unique ENA will be generated. For an aggregate I/O, this pointer is set to 91ea8dc4b6Seschrock * NULL, and no ereport will be generated (since it doesn't actually correspond 92ea8dc4b6Seschrock * to any particular device or piece of data). 93ea8dc4b6Seschrock */ 94ea8dc4b6Seschrock void 95ea8dc4b6Seschrock zfs_ereport_post(const char *subclass, spa_t *spa, vdev_t *vd, zio_t *zio, 96ea8dc4b6Seschrock uint64_t stateoroffset, uint64_t size) 97ea8dc4b6Seschrock { 98ea8dc4b6Seschrock #ifdef _KERNEL 99ea8dc4b6Seschrock nvlist_t *ereport, *detector; 100ea8dc4b6Seschrock uint64_t ena; 101ea8dc4b6Seschrock char class[64]; 102c5904d13Seschrock int state; 103ea8dc4b6Seschrock 104ea8dc4b6Seschrock /* 105ea8dc4b6Seschrock * If we are doing a spa_tryimport(), ignore errors. 106ea8dc4b6Seschrock */ 107ea8dc4b6Seschrock if (spa->spa_load_state == SPA_LOAD_TRYIMPORT) 108ea8dc4b6Seschrock return; 109ea8dc4b6Seschrock 110ea8dc4b6Seschrock /* 111ea8dc4b6Seschrock * If we are in the middle of opening a pool, and the previous attempt 112ea8dc4b6Seschrock * failed, don't bother logging any new ereports - we're just going to 113ea8dc4b6Seschrock * get the same diagnosis anyway. 114ea8dc4b6Seschrock */ 115ea8dc4b6Seschrock if (spa->spa_load_state != SPA_LOAD_NONE && 116ea8dc4b6Seschrock spa->spa_last_open_failed) 117ea8dc4b6Seschrock return; 118ea8dc4b6Seschrock 119bf82a41bSeschrock if (zio != NULL) { 120bf82a41bSeschrock /* 121bf82a41bSeschrock * Ignore any errors from I/Os that we are going to retry 122bf82a41bSeschrock * anyway - we only generate errors from the final failure. 123bf82a41bSeschrock * Checksum errors are generated after the pipeline stage 124bf82a41bSeschrock * responsible for retrying the I/O (VDEV_IO_ASSESS), so this 125bf82a41bSeschrock * only applies to standard I/O errors. 126bf82a41bSeschrock */ 127bf82a41bSeschrock if (zio_should_retry(zio) && zio->io_error != ECKSUM) 128bf82a41bSeschrock return; 129ea8dc4b6Seschrock 130bf82a41bSeschrock /* 131bf82a41bSeschrock * If this is not a read or write zio, ignore the error. This 132bf82a41bSeschrock * can occur if the DKIOCFLUSHWRITECACHE ioctl fails. 133bf82a41bSeschrock */ 134bf82a41bSeschrock if (zio->io_type != ZIO_TYPE_READ && 135bf82a41bSeschrock zio->io_type != ZIO_TYPE_WRITE) 136bf82a41bSeschrock return; 137bf82a41bSeschrock 138bf82a41bSeschrock /* 139bf82a41bSeschrock * Ignore any errors from speculative I/Os, as failure is an 140bf82a41bSeschrock * expected result. 141bf82a41bSeschrock */ 142bf82a41bSeschrock if (zio->io_flags & ZIO_FLAG_SPECULATIVE) 143bf82a41bSeschrock return; 144*51ece835Seschrock 145*51ece835Seschrock /* 146*51ece835Seschrock * If the vdev has already been marked as failing due to a 147*51ece835Seschrock * failed probe, then ignore any subsequent I/O errors, as the 148*51ece835Seschrock * DE will automatically fault the vdev on the first such 149*51ece835Seschrock * failure. 150*51ece835Seschrock */ 151*51ece835Seschrock if (vd != NULL && vd->vdev_is_failing && 152*51ece835Seschrock strcmp(subclass, FM_EREPORT_ZFS_PROBE_FAILURE) != 0) 153*51ece835Seschrock return; 154bf82a41bSeschrock } 155b468a217Seschrock 156ea8dc4b6Seschrock if ((ereport = fm_nvlist_create(NULL)) == NULL) 157ea8dc4b6Seschrock return; 158ea8dc4b6Seschrock 159ea8dc4b6Seschrock if ((detector = fm_nvlist_create(NULL)) == NULL) { 160ea8dc4b6Seschrock fm_nvlist_destroy(ereport, FM_NVA_FREE); 161ea8dc4b6Seschrock return; 162ea8dc4b6Seschrock } 163ea8dc4b6Seschrock 164ea8dc4b6Seschrock /* 165ea8dc4b6Seschrock * Serialize ereport generation 166ea8dc4b6Seschrock */ 167ea8dc4b6Seschrock mutex_enter(&spa->spa_errlist_lock); 168ea8dc4b6Seschrock 169ea8dc4b6Seschrock /* 170ea8dc4b6Seschrock * Determine the ENA to use for this event. If we are in a loading 171ea8dc4b6Seschrock * state, use a SPA-wide ENA. Otherwise, if we are in an I/O state, use 172ea8dc4b6Seschrock * a root zio-wide ENA. Otherwise, simply use a unique ENA. 173ea8dc4b6Seschrock */ 174ea8dc4b6Seschrock if (spa->spa_load_state != SPA_LOAD_NONE) { 175ea8dc4b6Seschrock if (spa->spa_ena == 0) 176ea8dc4b6Seschrock spa->spa_ena = fm_ena_generate(0, FM_ENA_FMT1); 177ea8dc4b6Seschrock ena = spa->spa_ena; 178ea8dc4b6Seschrock } else if (zio != NULL && zio->io_logical != NULL) { 179ea8dc4b6Seschrock if (zio->io_logical->io_ena == 0) 180ea8dc4b6Seschrock zio->io_logical->io_ena = 181ea8dc4b6Seschrock fm_ena_generate(0, FM_ENA_FMT1); 182ea8dc4b6Seschrock ena = zio->io_logical->io_ena; 183ea8dc4b6Seschrock } else { 184ea8dc4b6Seschrock ena = fm_ena_generate(0, FM_ENA_FMT1); 185ea8dc4b6Seschrock } 186ea8dc4b6Seschrock 187ea8dc4b6Seschrock /* 188ea8dc4b6Seschrock * Construct the full class, detector, and other standard FMA fields. 189ea8dc4b6Seschrock */ 190ea8dc4b6Seschrock (void) snprintf(class, sizeof (class), "%s.%s", 191ea8dc4b6Seschrock ZFS_ERROR_CLASS, subclass); 192ea8dc4b6Seschrock 193ea8dc4b6Seschrock fm_fmri_zfs_set(detector, FM_ZFS_SCHEME_VERSION, spa_guid(spa), 194ea8dc4b6Seschrock vd != NULL ? vd->vdev_guid : 0); 195ea8dc4b6Seschrock 196ea8dc4b6Seschrock fm_ereport_set(ereport, FM_EREPORT_VERSION, class, ena, detector, NULL); 197ea8dc4b6Seschrock 198ea8dc4b6Seschrock /* 199ea8dc4b6Seschrock * Construct the per-ereport payload, depending on which parameters are 200ea8dc4b6Seschrock * passed in. 201ea8dc4b6Seschrock */ 202ea8dc4b6Seschrock 203c5904d13Seschrock /* 204c5904d13Seschrock * If we are importing a faulted pool, then we treat it like an open, 205c5904d13Seschrock * not an import. Otherwise, the DE will ignore all faults during 206c5904d13Seschrock * import, since the default behavior is to mark the devices as 207c5904d13Seschrock * persistently unavailable, not leave them in the faulted state. 208c5904d13Seschrock */ 209c5904d13Seschrock state = spa->spa_import_faulted ? SPA_LOAD_OPEN : spa->spa_load_state; 210c5904d13Seschrock 211ea8dc4b6Seschrock /* 212ea8dc4b6Seschrock * Generic payload members common to all ereports. 213ea8dc4b6Seschrock * 214ea8dc4b6Seschrock * The direct reference to spa_name is used rather than spa_name() 215ea8dc4b6Seschrock * because of the asynchronous nature of the zio pipeline. spa_name() 216ea8dc4b6Seschrock * asserts that the config lock is held in some form. This is always 217ea8dc4b6Seschrock * the case in I/O context, but because the check for RW_WRITER compares 218ea8dc4b6Seschrock * against 'curthread', we may be in an asynchronous context and blow 219ea8dc4b6Seschrock * this assert. Rather than loosen this assert, we acknowledge that all 220ea8dc4b6Seschrock * contexts in which this function is called (pool open, I/O) are safe, 221ea8dc4b6Seschrock * and dereference the name directly. 222ea8dc4b6Seschrock */ 223ea8dc4b6Seschrock fm_payload_set(ereport, FM_EREPORT_PAYLOAD_ZFS_POOL, 224ea8dc4b6Seschrock DATA_TYPE_STRING, spa->spa_name, FM_EREPORT_PAYLOAD_ZFS_POOL_GUID, 225ea8dc4b6Seschrock DATA_TYPE_UINT64, spa_guid(spa), 226ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_POOL_CONTEXT, DATA_TYPE_INT32, 227c5904d13Seschrock state, NULL); 228ea8dc4b6Seschrock 22932b87932Sek if (spa != NULL) { 23032b87932Sek fm_payload_set(ereport, FM_EREPORT_PAYLOAD_ZFS_POOL_FAILMODE, 23132b87932Sek DATA_TYPE_STRING, 23232b87932Sek spa_get_failmode(spa) == ZIO_FAILURE_MODE_WAIT ? 23332b87932Sek FM_EREPORT_FAILMODE_WAIT : 23432b87932Sek spa_get_failmode(spa) == ZIO_FAILURE_MODE_CONTINUE ? 23532b87932Sek FM_EREPORT_FAILMODE_CONTINUE : FM_EREPORT_FAILMODE_PANIC, 23632b87932Sek NULL); 23732b87932Sek } 23832b87932Sek 239ea8dc4b6Seschrock if (vd != NULL) { 240ea8dc4b6Seschrock vdev_t *pvd = vd->vdev_parent; 241ea8dc4b6Seschrock 242ea8dc4b6Seschrock fm_payload_set(ereport, FM_EREPORT_PAYLOAD_ZFS_VDEV_GUID, 243ea8dc4b6Seschrock DATA_TYPE_UINT64, vd->vdev_guid, 244ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_VDEV_TYPE, 245ea8dc4b6Seschrock DATA_TYPE_STRING, vd->vdev_ops->vdev_op_type, NULL); 246ea8dc4b6Seschrock if (vd->vdev_path) 247ea8dc4b6Seschrock fm_payload_set(ereport, 248ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_VDEV_PATH, 249ea8dc4b6Seschrock DATA_TYPE_STRING, vd->vdev_path, NULL); 250ea8dc4b6Seschrock if (vd->vdev_devid) 251ea8dc4b6Seschrock fm_payload_set(ereport, 252ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_VDEV_DEVID, 253ea8dc4b6Seschrock DATA_TYPE_STRING, vd->vdev_devid, NULL); 254ea8dc4b6Seschrock 255ea8dc4b6Seschrock if (pvd != NULL) { 256ea8dc4b6Seschrock fm_payload_set(ereport, 257ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_PARENT_GUID, 258ea8dc4b6Seschrock DATA_TYPE_UINT64, pvd->vdev_guid, 259ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_PARENT_TYPE, 260ea8dc4b6Seschrock DATA_TYPE_STRING, pvd->vdev_ops->vdev_op_type, 261ea8dc4b6Seschrock NULL); 262ea8dc4b6Seschrock if (pvd->vdev_path) 263ea8dc4b6Seschrock fm_payload_set(ereport, 264ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_PARENT_PATH, 265c25056deSgw DATA_TYPE_STRING, pvd->vdev_path, NULL); 266ea8dc4b6Seschrock if (pvd->vdev_devid) 267ea8dc4b6Seschrock fm_payload_set(ereport, 268ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_PARENT_DEVID, 269ea8dc4b6Seschrock DATA_TYPE_STRING, pvd->vdev_devid, NULL); 270ea8dc4b6Seschrock } 271ea8dc4b6Seschrock } 272ea8dc4b6Seschrock 273ea8dc4b6Seschrock if (zio != NULL) { 274ea8dc4b6Seschrock /* 275ea8dc4b6Seschrock * Payload common to all I/Os. 276ea8dc4b6Seschrock */ 277ea8dc4b6Seschrock fm_payload_set(ereport, FM_EREPORT_PAYLOAD_ZFS_ZIO_ERR, 278ea8dc4b6Seschrock DATA_TYPE_INT32, zio->io_error, NULL); 279ea8dc4b6Seschrock 280ea8dc4b6Seschrock /* 281ea8dc4b6Seschrock * If the 'size' parameter is non-zero, it indicates this is a 282ea8dc4b6Seschrock * RAID-Z or other I/O where the physical offset and length are 283ea8dc4b6Seschrock * provided for us, instead of within the zio_t. 284ea8dc4b6Seschrock */ 285ea8dc4b6Seschrock if (vd != NULL) { 286ea8dc4b6Seschrock if (size) 287ea8dc4b6Seschrock fm_payload_set(ereport, 288ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_OFFSET, 289ea8dc4b6Seschrock DATA_TYPE_UINT64, stateoroffset, 290ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_SIZE, 291c3c6d682Seschrock DATA_TYPE_UINT64, size, NULL); 292ea8dc4b6Seschrock else 293ea8dc4b6Seschrock fm_payload_set(ereport, 294ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_OFFSET, 295ea8dc4b6Seschrock DATA_TYPE_UINT64, zio->io_offset, 296ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_SIZE, 297c3c6d682Seschrock DATA_TYPE_UINT64, zio->io_size, NULL); 298ea8dc4b6Seschrock } 299ea8dc4b6Seschrock 300ea8dc4b6Seschrock /* 301ea8dc4b6Seschrock * Payload for I/Os with corresponding logical information. 302ea8dc4b6Seschrock */ 303ea8dc4b6Seschrock if (zio->io_logical != NULL) 304ea8dc4b6Seschrock fm_payload_set(ereport, 305e7cbe64fSgw FM_EREPORT_PAYLOAD_ZFS_ZIO_OBJSET, 306e7cbe64fSgw DATA_TYPE_UINT64, 307e7cbe64fSgw zio->io_logical->io_bookmark.zb_objset, 308ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_OBJECT, 309ea8dc4b6Seschrock DATA_TYPE_UINT64, 310ea8dc4b6Seschrock zio->io_logical->io_bookmark.zb_object, 311ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_LEVEL, 312c25056deSgw DATA_TYPE_INT64, 313ea8dc4b6Seschrock zio->io_logical->io_bookmark.zb_level, 314ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_ZIO_BLKID, 315ea8dc4b6Seschrock DATA_TYPE_UINT64, 316c3c6d682Seschrock zio->io_logical->io_bookmark.zb_blkid, NULL); 317ea8dc4b6Seschrock } else if (vd != NULL) { 318ea8dc4b6Seschrock /* 319ea8dc4b6Seschrock * If we have a vdev but no zio, this is a device fault, and the 320ea8dc4b6Seschrock * 'stateoroffset' parameter indicates the previous state of the 321ea8dc4b6Seschrock * vdev. 322ea8dc4b6Seschrock */ 323ea8dc4b6Seschrock fm_payload_set(ereport, 324ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_PREV_STATE, 325ea8dc4b6Seschrock DATA_TYPE_UINT64, stateoroffset, NULL); 326ea8dc4b6Seschrock } 327ea8dc4b6Seschrock mutex_exit(&spa->spa_errlist_lock); 328ea8dc4b6Seschrock 329ea8dc4b6Seschrock fm_ereport_post(ereport, EVCH_SLEEP); 330ea8dc4b6Seschrock 331ea8dc4b6Seschrock fm_nvlist_destroy(ereport, FM_NVA_FREE); 332ea8dc4b6Seschrock fm_nvlist_destroy(detector, FM_NVA_FREE); 333ea8dc4b6Seschrock #endif 334ea8dc4b6Seschrock } 335ea8dc4b6Seschrock 3363d7072f8Seschrock static void 3373d7072f8Seschrock zfs_post_common(spa_t *spa, vdev_t *vd, const char *name) 338ea8dc4b6Seschrock { 339ea8dc4b6Seschrock #ifdef _KERNEL 340ea8dc4b6Seschrock nvlist_t *resource; 341ea8dc4b6Seschrock char class[64]; 342ea8dc4b6Seschrock 343ea8dc4b6Seschrock if ((resource = fm_nvlist_create(NULL)) == NULL) 344ea8dc4b6Seschrock return; 345ea8dc4b6Seschrock 346ea8dc4b6Seschrock (void) snprintf(class, sizeof (class), "%s.%s.%s", FM_RSRC_RESOURCE, 3473d7072f8Seschrock ZFS_ERROR_CLASS, name); 348ea8dc4b6Seschrock VERIFY(nvlist_add_uint8(resource, FM_VERSION, FM_RSRC_VERSION) == 0); 349ea8dc4b6Seschrock VERIFY(nvlist_add_string(resource, FM_CLASS, class) == 0); 350ea8dc4b6Seschrock VERIFY(nvlist_add_uint64(resource, 351ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_POOL_GUID, spa_guid(spa)) == 0); 352ea8dc4b6Seschrock if (vd) 353ea8dc4b6Seschrock VERIFY(nvlist_add_uint64(resource, 354ea8dc4b6Seschrock FM_EREPORT_PAYLOAD_ZFS_VDEV_GUID, vd->vdev_guid) == 0); 355ea8dc4b6Seschrock 356ea8dc4b6Seschrock fm_ereport_post(resource, EVCH_SLEEP); 357ea8dc4b6Seschrock 358ea8dc4b6Seschrock fm_nvlist_destroy(resource, FM_NVA_FREE); 359ea8dc4b6Seschrock #endif 360ea8dc4b6Seschrock } 3613d7072f8Seschrock 3623d7072f8Seschrock /* 3633d7072f8Seschrock * The 'resource.fs.zfs.removed' event is an internal signal that the given vdev 3643d7072f8Seschrock * has been removed from the system. This will cause the DE to ignore any 3653d7072f8Seschrock * recent I/O errors, inferring that they are due to the asynchronous device 3663d7072f8Seschrock * removal. 3673d7072f8Seschrock */ 3683d7072f8Seschrock void 3693d7072f8Seschrock zfs_post_remove(spa_t *spa, vdev_t *vd) 3703d7072f8Seschrock { 3713d7072f8Seschrock zfs_post_common(spa, vd, FM_RESOURCE_REMOVED); 3723d7072f8Seschrock } 3733d7072f8Seschrock 3743d7072f8Seschrock /* 3753d7072f8Seschrock * The 'resource.fs.zfs.autoreplace' event is an internal signal that the pool 3763d7072f8Seschrock * has the 'autoreplace' property set, and therefore any broken vdevs will be 3773d7072f8Seschrock * handled by higher level logic, and no vdev fault should be generated. 3783d7072f8Seschrock */ 3793d7072f8Seschrock void 3803d7072f8Seschrock zfs_post_autoreplace(spa_t *spa, vdev_t *vd) 3813d7072f8Seschrock { 3823d7072f8Seschrock zfs_post_common(spa, vd, FM_RESOURCE_AUTOREPLACE); 3833d7072f8Seschrock } 384