1fa9e4066Sahrens /* 2fa9e4066Sahrens * CDDL HEADER START 3fa9e4066Sahrens * 4fa9e4066Sahrens * The contents of this file are subject to the terms of the 5ea8dc4b6Seschrock * Common Development and Distribution License (the "License"). 6ea8dc4b6Seschrock * You may not use this file except in compliance with the License. 7fa9e4066Sahrens * 8fa9e4066Sahrens * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9fa9e4066Sahrens * or http://www.opensolaris.org/os/licensing. 10fa9e4066Sahrens * See the License for the specific language governing permissions 11fa9e4066Sahrens * and limitations under the License. 12fa9e4066Sahrens * 13fa9e4066Sahrens * When distributing Covered Code, include this CDDL HEADER in each 14fa9e4066Sahrens * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15fa9e4066Sahrens * If applicable, add the following below this CDDL HEADER, with the 16fa9e4066Sahrens * fields enclosed by brackets "[]" replaced with your own identifying 17fa9e4066Sahrens * information: Portions Copyright [yyyy] [name of copyright owner] 18fa9e4066Sahrens * 19fa9e4066Sahrens * CDDL HEADER END 20fa9e4066Sahrens */ 21fa9e4066Sahrens /* 22f80ce222SChris Kirby * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved. 23b77b9231SDan McDonald * 24b77b9231SDan McDonald * Portions Copyright 2010 Robert Milkowski 25b77b9231SDan McDonald * 26047c81d3SSaso Kiselkov * Copyright 2017 Nexenta Systems, Inc. All rights reserved. 27b7edcb94SMatthew Ahrens * Copyright (c) 2012, 2017 by Delphix. All rights reserved. 28810e43b2SBill Pijewski * Copyright (c) 2013, Joyent, Inc. All rights reserved. 29c3d26abcSMatthew Ahrens * Copyright (c) 2014 Integros [integros.com] 30fa9e4066Sahrens */ 31fa9e4066Sahrens 32fa9e4066Sahrens /* 33fa9e4066Sahrens * ZFS volume emulation driver. 34fa9e4066Sahrens * 35fa9e4066Sahrens * Makes a DMU object look like a volume of arbitrary size, up to 2^64 bytes. 36fa9e4066Sahrens * Volumes are accessed through the symbolic links named: 37fa9e4066Sahrens * 38fa9e4066Sahrens * /dev/zvol/dsk/<pool_name>/<dataset_name> 39fa9e4066Sahrens * /dev/zvol/rdsk/<pool_name>/<dataset_name> 40fa9e4066Sahrens * 41681d9761SEric Taylor * These links are created by the /dev filesystem (sdev_zvolops.c). 42fa9e4066Sahrens * Volumes are persistent through reboot. No user command needs to be 43fa9e4066Sahrens * run before opening and using a device. 44fa9e4066Sahrens */ 45fa9e4066Sahrens 46fa9e4066Sahrens #include <sys/types.h> 47fa9e4066Sahrens #include <sys/param.h> 48fa9e4066Sahrens #include <sys/errno.h> 49fa9e4066Sahrens #include <sys/uio.h> 50fa9e4066Sahrens #include <sys/buf.h> 51fa9e4066Sahrens #include <sys/modctl.h> 52fa9e4066Sahrens #include <sys/open.h> 53fa9e4066Sahrens #include <sys/kmem.h> 54fa9e4066Sahrens #include <sys/conf.h> 55fa9e4066Sahrens #include <sys/cmn_err.h> 56fa9e4066Sahrens #include <sys/stat.h> 57fa9e4066Sahrens #include <sys/zap.h> 58fa9e4066Sahrens #include <sys/spa.h> 59810e43b2SBill Pijewski #include <sys/spa_impl.h> 60fa9e4066Sahrens #include <sys/zio.h> 61e7cbe64fSgw #include <sys/dmu_traverse.h> 62e7cbe64fSgw #include <sys/dnode.h> 63e7cbe64fSgw #include <sys/dsl_dataset.h> 64fa9e4066Sahrens #include <sys/dsl_prop.h> 65fa9e4066Sahrens #include <sys/dkio.h> 66fa9e4066Sahrens #include <sys/efi_partition.h> 67fa9e4066Sahrens #include <sys/byteorder.h> 68fa9e4066Sahrens #include <sys/pathname.h> 69fa9e4066Sahrens #include <sys/ddi.h> 70fa9e4066Sahrens #include <sys/sunddi.h> 71fa9e4066Sahrens #include <sys/crc32.h> 72fa9e4066Sahrens #include <sys/dirent.h> 73fa9e4066Sahrens #include <sys/policy.h> 74fa9e4066Sahrens #include <sys/fs/zfs.h> 75fa9e4066Sahrens #include <sys/zfs_ioctl.h> 76fa9e4066Sahrens #include <sys/mkdev.h> 7722ac5be4Sperrin #include <sys/zil.h> 78c5c6ffa0Smaybee #include <sys/refcount.h> 79c2e6a7d6Sperrin #include <sys/zfs_znode.h> 80c2e6a7d6Sperrin #include <sys/zfs_rlock.h> 81e7cbe64fSgw #include <sys/vdev_disk.h> 82e7cbe64fSgw #include <sys/vdev_impl.h> 83810e43b2SBill Pijewski #include <sys/vdev_raidz.h> 84e7cbe64fSgw #include <sys/zvol.h> 85e7cbe64fSgw #include <sys/dumphdr.h> 861209a471SNeil Perrin #include <sys/zil_impl.h> 8780901aeaSGeorge Wilson #include <sys/dbuf.h> 88810e43b2SBill Pijewski #include <sys/dmu_tx.h> 89810e43b2SBill Pijewski #include <sys/zfeature.h> 90810e43b2SBill Pijewski #include <sys/zio_checksum.h> 911271e4b1SPrakash Surya #include <sys/zil_impl.h> 92047c81d3SSaso Kiselkov #include <sys/dkioc_free_util.h> 93*79315247SMatthew Ahrens #include <sys/zfs_rlock.h> 94fa9e4066Sahrens 95fa9e4066Sahrens #include "zfs_namecheck.h" 96fa9e4066Sahrens 97c99e4bdcSChris Kirby void *zfsdev_state; 98503ad85cSMatthew Ahrens static char *zvol_tag = "zvol_tag"; 99fa9e4066Sahrens 100e7cbe64fSgw #define ZVOL_DUMPSIZE "dumpsize" 101e7cbe64fSgw 102fa9e4066Sahrens /* 103c99e4bdcSChris Kirby * This lock protects the zfsdev_state structure from being modified 104fa9e4066Sahrens * while it's being used, e.g. an open that comes in before a create 105fa9e4066Sahrens * finishes. It also protects temporary opens of the dataset so that, 106fa9e4066Sahrens * e.g., an open doesn't get a spurious EBUSY. 107fa9e4066Sahrens */ 108c99e4bdcSChris Kirby kmutex_t zfsdev_state_lock; 109fa9e4066Sahrens static uint32_t zvol_minors; 110fa9e4066Sahrens 111e7cbe64fSgw typedef struct zvol_extent { 11288b7b0f2SMatthew Ahrens list_node_t ze_node; 113e7cbe64fSgw dva_t ze_dva; /* dva associated with this extent */ 11488b7b0f2SMatthew Ahrens uint64_t ze_nblks; /* number of blocks in extent */ 115e7cbe64fSgw } zvol_extent_t; 116e7cbe64fSgw 117fa9e4066Sahrens /* 118fa9e4066Sahrens * The in-core state of each volume. 119fa9e4066Sahrens */ 120fa9e4066Sahrens typedef struct zvol_state { 121fa9e4066Sahrens char zv_name[MAXPATHLEN]; /* pool/dd name */ 122fa9e4066Sahrens uint64_t zv_volsize; /* amount of space we advertise */ 12367bd71c6Sperrin uint64_t zv_volblocksize; /* volume block size */ 124fa9e4066Sahrens minor_t zv_minor; /* minor number */ 125fa9e4066Sahrens uint8_t zv_min_bs; /* minimum addressable block shift */ 126701f66c4SEric Taylor uint8_t zv_flags; /* readonly, dumpified, etc. */ 127fa9e4066Sahrens objset_t *zv_objset; /* objset handle */ 128fa9e4066Sahrens uint32_t zv_open_count[OTYPCNT]; /* open counts */ 129fa9e4066Sahrens uint32_t zv_total_opens; /* total open count */ 13022ac5be4Sperrin zilog_t *zv_zilog; /* ZIL handle */ 13188b7b0f2SMatthew Ahrens list_t zv_extents; /* List of extents for dump */ 132*79315247SMatthew Ahrens rangelock_t zv_rangelock; 1338dfe5547SRichard Yao dnode_t *zv_dn; /* dnode hold */ 134fa9e4066Sahrens } zvol_state_t; 135fa9e4066Sahrens 136e7cbe64fSgw /* 137e7cbe64fSgw * zvol specific flags 138e7cbe64fSgw */ 139e7cbe64fSgw #define ZVOL_RDONLY 0x1 140e7cbe64fSgw #define ZVOL_DUMPIFIED 0x2 141c7f714e2SEric Taylor #define ZVOL_EXCL 0x4 142701f66c4SEric Taylor #define ZVOL_WCE 0x8 143e7cbe64fSgw 14467bd71c6Sperrin /* 14567bd71c6Sperrin * zvol maximum transfer in one DMU tx. 14667bd71c6Sperrin */ 14767bd71c6Sperrin int zvol_maxphys = DMU_MAX_ACCESS/2; 14867bd71c6Sperrin 149893c83baSGeorge Wilson /* 150893c83baSGeorge Wilson * Toggle unmap functionality. 151893c83baSGeorge Wilson */ 152893c83baSGeorge Wilson boolean_t zvol_unmap_enabled = B_TRUE; 153893c83baSGeorge Wilson 1541c9272b8SStephen Blinick /* 1551c9272b8SStephen Blinick * If true, unmaps requested as synchronous are executed synchronously, 1561c9272b8SStephen Blinick * otherwise all unmaps are asynchronous. 1571c9272b8SStephen Blinick */ 1581c9272b8SStephen Blinick boolean_t zvol_unmap_sync_enabled = B_FALSE; 1591c9272b8SStephen Blinick 16092241e0bSTom Erickson extern int zfs_set_prop_nvlist(const char *, zprop_source_t, 1614445fffbSMatthew Ahrens nvlist_t *, nvlist_t *); 162681d9761SEric Taylor static int zvol_remove_zv(zvol_state_t *); 1631271e4b1SPrakash Surya static int zvol_get_data(void *arg, lr_write_t *lr, char *buf, 1641271e4b1SPrakash Surya struct lwb *lwb, zio_t *zio); 165e7cbe64fSgw static int zvol_dumpify(zvol_state_t *zv); 166e7cbe64fSgw static int zvol_dump_fini(zvol_state_t *zv); 167e7cbe64fSgw static int zvol_dump_init(zvol_state_t *zv, boolean_t resize); 16867bd71c6Sperrin 169fa9e4066Sahrens static void 170c61ea566SGeorge Wilson zvol_size_changed(zvol_state_t *zv, uint64_t volsize) 171fa9e4066Sahrens { 172c61ea566SGeorge Wilson dev_t dev = makedevice(ddi_driver_major(zfs_dip), zv->zv_minor); 173fa9e4066Sahrens 174c61ea566SGeorge Wilson zv->zv_volsize = volsize; 175fa9e4066Sahrens VERIFY(ddi_prop_update_int64(dev, zfs_dip, 176681d9761SEric Taylor "Size", volsize) == DDI_SUCCESS); 177fa9e4066Sahrens VERIFY(ddi_prop_update_int64(dev, zfs_dip, 178681d9761SEric Taylor "Nblocks", lbtodb(volsize)) == DDI_SUCCESS); 179e7cbe64fSgw 180e7cbe64fSgw /* Notify specfs to invalidate the cached size */ 181e7cbe64fSgw spec_size_invalidate(dev, VBLK); 182e7cbe64fSgw spec_size_invalidate(dev, VCHR); 183fa9e4066Sahrens } 184fa9e4066Sahrens 185fa9e4066Sahrens int 186e9dbad6fSeschrock zvol_check_volsize(uint64_t volsize, uint64_t blocksize) 187fa9e4066Sahrens { 188e9dbad6fSeschrock if (volsize == 0) 189be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 190fa9e4066Sahrens 191e9dbad6fSeschrock if (volsize % blocksize != 0) 192be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 1935c5460e9Seschrock 194fa9e4066Sahrens #ifdef _ILP32 195e9dbad6fSeschrock if (volsize - 1 > SPEC_MAXOFFSET_T) 196be6fd75aSMatthew Ahrens return (SET_ERROR(EOVERFLOW)); 197fa9e4066Sahrens #endif 198fa9e4066Sahrens return (0); 199fa9e4066Sahrens } 200fa9e4066Sahrens 201fa9e4066Sahrens int 202e9dbad6fSeschrock zvol_check_volblocksize(uint64_t volblocksize) 203fa9e4066Sahrens { 204e9dbad6fSeschrock if (volblocksize < SPA_MINBLOCKSIZE || 205b5152584SMatthew Ahrens volblocksize > SPA_OLD_MAXBLOCKSIZE || 206e9dbad6fSeschrock !ISP2(volblocksize)) 207be6fd75aSMatthew Ahrens return (SET_ERROR(EDOM)); 208fa9e4066Sahrens 209fa9e4066Sahrens return (0); 210fa9e4066Sahrens } 211fa9e4066Sahrens 212fa9e4066Sahrens int 213a2eea2e1Sahrens zvol_get_stats(objset_t *os, nvlist_t *nv) 214fa9e4066Sahrens { 215fa9e4066Sahrens int error; 216fa9e4066Sahrens dmu_object_info_t doi; 217a2eea2e1Sahrens uint64_t val; 218fa9e4066Sahrens 219a2eea2e1Sahrens error = zap_lookup(os, ZVOL_ZAP_OBJ, "size", 8, 1, &val); 220fa9e4066Sahrens if (error) 221fa9e4066Sahrens return (error); 222fa9e4066Sahrens 223a2eea2e1Sahrens dsl_prop_nvlist_add_uint64(nv, ZFS_PROP_VOLSIZE, val); 224a2eea2e1Sahrens 225fa9e4066Sahrens error = dmu_object_info(os, ZVOL_OBJ, &doi); 226fa9e4066Sahrens 227a2eea2e1Sahrens if (error == 0) { 228a2eea2e1Sahrens dsl_prop_nvlist_add_uint64(nv, ZFS_PROP_VOLBLOCKSIZE, 229a2eea2e1Sahrens doi.doi_data_block_size); 230a2eea2e1Sahrens } 231fa9e4066Sahrens 232fa9e4066Sahrens return (error); 233fa9e4066Sahrens } 234fa9e4066Sahrens 235fa9e4066Sahrens static zvol_state_t * 236e9dbad6fSeschrock zvol_minor_lookup(const char *name) 237fa9e4066Sahrens { 238fa9e4066Sahrens minor_t minor; 239fa9e4066Sahrens zvol_state_t *zv; 240fa9e4066Sahrens 241c99e4bdcSChris Kirby ASSERT(MUTEX_HELD(&zfsdev_state_lock)); 242fa9e4066Sahrens 243c99e4bdcSChris Kirby for (minor = 1; minor <= ZFSDEV_MAX_MINOR; minor++) { 244c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 245fa9e4066Sahrens if (zv == NULL) 246fa9e4066Sahrens continue; 247fa9e4066Sahrens if (strcmp(zv->zv_name, name) == 0) 248f80ce222SChris Kirby return (zv); 249fa9e4066Sahrens } 250fa9e4066Sahrens 251f80ce222SChris Kirby return (NULL); 252fa9e4066Sahrens } 253fa9e4066Sahrens 254e7cbe64fSgw /* extent mapping arg */ 255e7cbe64fSgw struct maparg { 25688b7b0f2SMatthew Ahrens zvol_state_t *ma_zv; 25788b7b0f2SMatthew Ahrens uint64_t ma_blks; 258e7cbe64fSgw }; 259e7cbe64fSgw 260e7cbe64fSgw /*ARGSUSED*/ 261e7cbe64fSgw static int 2621b912ec7SGeorge Wilson zvol_map_block(spa_t *spa, zilog_t *zilog, const blkptr_t *bp, 2637802d7bfSMatthew Ahrens const zbookmark_phys_t *zb, const dnode_phys_t *dnp, void *arg) 264e7cbe64fSgw { 26588b7b0f2SMatthew Ahrens struct maparg *ma = arg; 26688b7b0f2SMatthew Ahrens zvol_extent_t *ze; 26788b7b0f2SMatthew Ahrens int bs = ma->ma_zv->zv_volblocksize; 268e7cbe64fSgw 269a2cdcdd2SPaul Dagnelie if (bp == NULL || BP_IS_HOLE(bp) || 27043466aaeSMax Grossman zb->zb_object != ZVOL_OBJ || zb->zb_level != 0) 27188b7b0f2SMatthew Ahrens return (0); 272e7cbe64fSgw 2735d7b4d43SMatthew Ahrens VERIFY(!BP_IS_EMBEDDED(bp)); 2745d7b4d43SMatthew Ahrens 27588b7b0f2SMatthew Ahrens VERIFY3U(ma->ma_blks, ==, zb->zb_blkid); 27688b7b0f2SMatthew Ahrens ma->ma_blks++; 277e7cbe64fSgw 27888b7b0f2SMatthew Ahrens /* Abort immediately if we have encountered gang blocks */ 27988b7b0f2SMatthew Ahrens if (BP_IS_GANG(bp)) 280be6fd75aSMatthew Ahrens return (SET_ERROR(EFRAGS)); 281e7cbe64fSgw 28288b7b0f2SMatthew Ahrens /* 28388b7b0f2SMatthew Ahrens * See if the block is at the end of the previous extent. 28488b7b0f2SMatthew Ahrens */ 28588b7b0f2SMatthew Ahrens ze = list_tail(&ma->ma_zv->zv_extents); 28688b7b0f2SMatthew Ahrens if (ze && 28788b7b0f2SMatthew Ahrens DVA_GET_VDEV(BP_IDENTITY(bp)) == DVA_GET_VDEV(&ze->ze_dva) && 28888b7b0f2SMatthew Ahrens DVA_GET_OFFSET(BP_IDENTITY(bp)) == 28988b7b0f2SMatthew Ahrens DVA_GET_OFFSET(&ze->ze_dva) + ze->ze_nblks * bs) { 29088b7b0f2SMatthew Ahrens ze->ze_nblks++; 29188b7b0f2SMatthew Ahrens return (0); 292e7cbe64fSgw } 293e7cbe64fSgw 29488b7b0f2SMatthew Ahrens dprintf_bp(bp, "%s", "next blkptr:"); 295e7cbe64fSgw 29688b7b0f2SMatthew Ahrens /* start a new extent */ 29788b7b0f2SMatthew Ahrens ze = kmem_zalloc(sizeof (zvol_extent_t), KM_SLEEP); 29888b7b0f2SMatthew Ahrens ze->ze_dva = bp->blk_dva[0]; /* structure assignment */ 29988b7b0f2SMatthew Ahrens ze->ze_nblks = 1; 30088b7b0f2SMatthew Ahrens list_insert_tail(&ma->ma_zv->zv_extents, ze); 30188b7b0f2SMatthew Ahrens return (0); 30288b7b0f2SMatthew Ahrens } 303e7cbe64fSgw 30488b7b0f2SMatthew Ahrens static void 30588b7b0f2SMatthew Ahrens zvol_free_extents(zvol_state_t *zv) 30688b7b0f2SMatthew Ahrens { 30788b7b0f2SMatthew Ahrens zvol_extent_t *ze; 308e7cbe64fSgw 30988b7b0f2SMatthew Ahrens while (ze = list_head(&zv->zv_extents)) { 31088b7b0f2SMatthew Ahrens list_remove(&zv->zv_extents, ze); 31188b7b0f2SMatthew Ahrens kmem_free(ze, sizeof (zvol_extent_t)); 312e7cbe64fSgw } 31388b7b0f2SMatthew Ahrens } 314e7cbe64fSgw 31588b7b0f2SMatthew Ahrens static int 31688b7b0f2SMatthew Ahrens zvol_get_lbas(zvol_state_t *zv) 31788b7b0f2SMatthew Ahrens { 3183adc9019SEric Taylor objset_t *os = zv->zv_objset; 31988b7b0f2SMatthew Ahrens struct maparg ma; 32088b7b0f2SMatthew Ahrens int err; 32188b7b0f2SMatthew Ahrens 32288b7b0f2SMatthew Ahrens ma.ma_zv = zv; 32388b7b0f2SMatthew Ahrens ma.ma_blks = 0; 32488b7b0f2SMatthew Ahrens zvol_free_extents(zv); 32588b7b0f2SMatthew Ahrens 3263adc9019SEric Taylor /* commit any in-flight changes before traversing the dataset */ 3273adc9019SEric Taylor txg_wait_synced(dmu_objset_pool(os), 0); 3283adc9019SEric Taylor err = traverse_dataset(dmu_objset_ds(os), 0, 32988b7b0f2SMatthew Ahrens TRAVERSE_PRE | TRAVERSE_PREFETCH_METADATA, zvol_map_block, &ma); 33088b7b0f2SMatthew Ahrens if (err || ma.ma_blks != (zv->zv_volsize / zv->zv_volblocksize)) { 33188b7b0f2SMatthew Ahrens zvol_free_extents(zv); 33288b7b0f2SMatthew Ahrens return (err ? err : EIO); 333e7cbe64fSgw } 33488b7b0f2SMatthew Ahrens 335e7cbe64fSgw return (0); 336e7cbe64fSgw } 337e7cbe64fSgw 338ecd6cf80Smarks /* ARGSUSED */ 339fa9e4066Sahrens void 340ecd6cf80Smarks zvol_create_cb(objset_t *os, void *arg, cred_t *cr, dmu_tx_t *tx) 341fa9e4066Sahrens { 342da6c28aaSamw zfs_creat_t *zct = arg; 343da6c28aaSamw nvlist_t *nvprops = zct->zct_props; 344fa9e4066Sahrens int error; 345e9dbad6fSeschrock uint64_t volblocksize, volsize; 346fa9e4066Sahrens 347ecd6cf80Smarks VERIFY(nvlist_lookup_uint64(nvprops, 348e9dbad6fSeschrock zfs_prop_to_name(ZFS_PROP_VOLSIZE), &volsize) == 0); 349ecd6cf80Smarks if (nvlist_lookup_uint64(nvprops, 350e9dbad6fSeschrock zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE), &volblocksize) != 0) 351e9dbad6fSeschrock volblocksize = zfs_prop_default_numeric(ZFS_PROP_VOLBLOCKSIZE); 352e9dbad6fSeschrock 353e9dbad6fSeschrock /* 354e7cbe64fSgw * These properties must be removed from the list so the generic 355e9dbad6fSeschrock * property setting step won't apply to them. 356e9dbad6fSeschrock */ 357ecd6cf80Smarks VERIFY(nvlist_remove_all(nvprops, 358e9dbad6fSeschrock zfs_prop_to_name(ZFS_PROP_VOLSIZE)) == 0); 359ecd6cf80Smarks (void) nvlist_remove_all(nvprops, 360e9dbad6fSeschrock zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE)); 361e9dbad6fSeschrock 362e9dbad6fSeschrock error = dmu_object_claim(os, ZVOL_OBJ, DMU_OT_ZVOL, volblocksize, 363fa9e4066Sahrens DMU_OT_NONE, 0, tx); 364fa9e4066Sahrens ASSERT(error == 0); 365fa9e4066Sahrens 366fa9e4066Sahrens error = zap_create_claim(os, ZVOL_ZAP_OBJ, DMU_OT_ZVOL_PROP, 367fa9e4066Sahrens DMU_OT_NONE, 0, tx); 368fa9e4066Sahrens ASSERT(error == 0); 369fa9e4066Sahrens 370e9dbad6fSeschrock error = zap_update(os, ZVOL_ZAP_OBJ, "size", 8, 1, &volsize, tx); 371fa9e4066Sahrens ASSERT(error == 0); 372fa9e4066Sahrens } 373fa9e4066Sahrens 374b77b9231SDan McDonald /* 375b77b9231SDan McDonald * Replay a TX_TRUNCATE ZIL transaction if asked. TX_TRUNCATE is how we 376b77b9231SDan McDonald * implement DKIOCFREE/free-long-range. 377b77b9231SDan McDonald */ 378b77b9231SDan McDonald static int 3793f7978d0SAlan Somers zvol_replay_truncate(void *arg1, void *arg2, boolean_t byteswap) 380b77b9231SDan McDonald { 3813f7978d0SAlan Somers zvol_state_t *zv = arg1; 3823f7978d0SAlan Somers lr_truncate_t *lr = arg2; 383b77b9231SDan McDonald uint64_t offset, length; 384b77b9231SDan McDonald 385b77b9231SDan McDonald if (byteswap) 386b77b9231SDan McDonald byteswap_uint64_array(lr, sizeof (*lr)); 387b77b9231SDan McDonald 388b77b9231SDan McDonald offset = lr->lr_offset; 389b77b9231SDan McDonald length = lr->lr_length; 390b77b9231SDan McDonald 391b77b9231SDan McDonald return (dmu_free_long_range(zv->zv_objset, ZVOL_OBJ, offset, length)); 392b77b9231SDan McDonald } 393b77b9231SDan McDonald 39422ac5be4Sperrin /* 39522ac5be4Sperrin * Replay a TX_WRITE ZIL transaction that didn't get committed 39622ac5be4Sperrin * after a system failure 39722ac5be4Sperrin */ 39822ac5be4Sperrin static int 3993f7978d0SAlan Somers zvol_replay_write(void *arg1, void *arg2, boolean_t byteswap) 40022ac5be4Sperrin { 4013f7978d0SAlan Somers zvol_state_t *zv = arg1; 4023f7978d0SAlan Somers lr_write_t *lr = arg2; 40322ac5be4Sperrin objset_t *os = zv->zv_objset; 40422ac5be4Sperrin char *data = (char *)(lr + 1); /* data follows lr_write_t */ 405b24ab676SJeff Bonwick uint64_t offset, length; 40622ac5be4Sperrin dmu_tx_t *tx; 40722ac5be4Sperrin int error; 40822ac5be4Sperrin 40922ac5be4Sperrin if (byteswap) 41022ac5be4Sperrin byteswap_uint64_array(lr, sizeof (*lr)); 41122ac5be4Sperrin 412b24ab676SJeff Bonwick offset = lr->lr_offset; 413b24ab676SJeff Bonwick length = lr->lr_length; 414b24ab676SJeff Bonwick 415b24ab676SJeff Bonwick /* If it's a dmu_sync() block, write the whole block */ 416b24ab676SJeff Bonwick if (lr->lr_common.lrc_reclen == sizeof (lr_write_t)) { 417b24ab676SJeff Bonwick uint64_t blocksize = BP_GET_LSIZE(&lr->lr_blkptr); 418b24ab676SJeff Bonwick if (length < blocksize) { 419b24ab676SJeff Bonwick offset -= offset % blocksize; 420b24ab676SJeff Bonwick length = blocksize; 421b24ab676SJeff Bonwick } 422b24ab676SJeff Bonwick } 423975c32a0SNeil Perrin 42422ac5be4Sperrin tx = dmu_tx_create(os); 425b24ab676SJeff Bonwick dmu_tx_hold_write(tx, ZVOL_OBJ, offset, length); 4261209a471SNeil Perrin error = dmu_tx_assign(tx, TXG_WAIT); 42722ac5be4Sperrin if (error) { 42822ac5be4Sperrin dmu_tx_abort(tx); 42922ac5be4Sperrin } else { 430b24ab676SJeff Bonwick dmu_write(os, ZVOL_OBJ, offset, length, data, tx); 43122ac5be4Sperrin dmu_tx_commit(tx); 43222ac5be4Sperrin } 43322ac5be4Sperrin 43422ac5be4Sperrin return (error); 43522ac5be4Sperrin } 43622ac5be4Sperrin 43722ac5be4Sperrin /* ARGSUSED */ 43822ac5be4Sperrin static int 4393f7978d0SAlan Somers zvol_replay_err(void *arg1, void *arg2, boolean_t byteswap) 44022ac5be4Sperrin { 441be6fd75aSMatthew Ahrens return (SET_ERROR(ENOTSUP)); 44222ac5be4Sperrin } 44322ac5be4Sperrin 44422ac5be4Sperrin /* 44522ac5be4Sperrin * Callback vectors for replaying records. 446b77b9231SDan McDonald * Only TX_WRITE and TX_TRUNCATE are needed for zvol. 44722ac5be4Sperrin */ 44822ac5be4Sperrin zil_replay_func_t *zvol_replay_vector[TX_MAX_TYPE] = { 44922ac5be4Sperrin zvol_replay_err, /* 0 no such transaction type */ 45022ac5be4Sperrin zvol_replay_err, /* TX_CREATE */ 45122ac5be4Sperrin zvol_replay_err, /* TX_MKDIR */ 45222ac5be4Sperrin zvol_replay_err, /* TX_MKXATTR */ 45322ac5be4Sperrin zvol_replay_err, /* TX_SYMLINK */ 45422ac5be4Sperrin zvol_replay_err, /* TX_REMOVE */ 45522ac5be4Sperrin zvol_replay_err, /* TX_RMDIR */ 45622ac5be4Sperrin zvol_replay_err, /* TX_LINK */ 45722ac5be4Sperrin zvol_replay_err, /* TX_RENAME */ 45822ac5be4Sperrin zvol_replay_write, /* TX_WRITE */ 459b77b9231SDan McDonald zvol_replay_truncate, /* TX_TRUNCATE */ 46022ac5be4Sperrin zvol_replay_err, /* TX_SETATTR */ 46122ac5be4Sperrin zvol_replay_err, /* TX_ACL */ 462975c32a0SNeil Perrin zvol_replay_err, /* TX_CREATE_ACL */ 463975c32a0SNeil Perrin zvol_replay_err, /* TX_CREATE_ATTR */ 464975c32a0SNeil Perrin zvol_replay_err, /* TX_CREATE_ACL_ATTR */ 465975c32a0SNeil Perrin zvol_replay_err, /* TX_MKDIR_ACL */ 466975c32a0SNeil Perrin zvol_replay_err, /* TX_MKDIR_ATTR */ 467975c32a0SNeil Perrin zvol_replay_err, /* TX_MKDIR_ACL_ATTR */ 468975c32a0SNeil Perrin zvol_replay_err, /* TX_WRITE2 */ 46922ac5be4Sperrin }; 47022ac5be4Sperrin 471681d9761SEric Taylor int 472681d9761SEric Taylor zvol_name2minor(const char *name, minor_t *minor) 473681d9761SEric Taylor { 474681d9761SEric Taylor zvol_state_t *zv; 475681d9761SEric Taylor 476c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 477681d9761SEric Taylor zv = zvol_minor_lookup(name); 478681d9761SEric Taylor if (minor && zv) 479681d9761SEric Taylor *minor = zv->zv_minor; 480c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 481681d9761SEric Taylor return (zv ? 0 : -1); 482681d9761SEric Taylor } 483681d9761SEric Taylor 484e7cbe64fSgw /* 485e7cbe64fSgw * Create a minor node (plus a whole lot more) for the specified volume. 486fa9e4066Sahrens */ 487fa9e4066Sahrens int 488681d9761SEric Taylor zvol_create_minor(const char *name) 489fa9e4066Sahrens { 490c99e4bdcSChris Kirby zfs_soft_state_t *zs; 491fa9e4066Sahrens zvol_state_t *zv; 492fa9e4066Sahrens objset_t *os; 49367bd71c6Sperrin dmu_object_info_t doi; 494fa9e4066Sahrens minor_t minor = 0; 495fa9e4066Sahrens char chrbuf[30], blkbuf[30]; 496fa9e4066Sahrens int error; 497fa9e4066Sahrens 498c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 499fa9e4066Sahrens 5001195e687SMark J Musante if (zvol_minor_lookup(name) != NULL) { 501c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 502be6fd75aSMatthew Ahrens return (SET_ERROR(EEXIST)); 503fa9e4066Sahrens } 504fa9e4066Sahrens 505503ad85cSMatthew Ahrens /* lie and say we're read-only */ 5066e0cbcaaSMatthew Ahrens error = dmu_objset_own(name, DMU_OST_ZVOL, B_TRUE, FTAG, &os); 507fa9e4066Sahrens 508fa9e4066Sahrens if (error) { 509c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 510fa9e4066Sahrens return (error); 511fa9e4066Sahrens } 512fa9e4066Sahrens 513c99e4bdcSChris Kirby if ((minor = zfsdev_minor_alloc()) == 0) { 5146e0cbcaaSMatthew Ahrens dmu_objset_disown(os, FTAG); 515c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 516be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 517fa9e4066Sahrens } 518fa9e4066Sahrens 519c99e4bdcSChris Kirby if (ddi_soft_state_zalloc(zfsdev_state, minor) != DDI_SUCCESS) { 5206e0cbcaaSMatthew Ahrens dmu_objset_disown(os, FTAG); 521c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 522be6fd75aSMatthew Ahrens return (SET_ERROR(EAGAIN)); 523fa9e4066Sahrens } 524e9dbad6fSeschrock (void) ddi_prop_update_string(minor, zfs_dip, ZVOL_PROP_NAME, 525e9dbad6fSeschrock (char *)name); 526fa9e4066Sahrens 527681d9761SEric Taylor (void) snprintf(chrbuf, sizeof (chrbuf), "%u,raw", minor); 528fa9e4066Sahrens 529fa9e4066Sahrens if (ddi_create_minor_node(zfs_dip, chrbuf, S_IFCHR, 530fa9e4066Sahrens minor, DDI_PSEUDO, 0) == DDI_FAILURE) { 531c99e4bdcSChris Kirby ddi_soft_state_free(zfsdev_state, minor); 5326e0cbcaaSMatthew Ahrens dmu_objset_disown(os, FTAG); 533c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 534be6fd75aSMatthew Ahrens return (SET_ERROR(EAGAIN)); 535fa9e4066Sahrens } 536fa9e4066Sahrens 537681d9761SEric Taylor (void) snprintf(blkbuf, sizeof (blkbuf), "%u", minor); 538fa9e4066Sahrens 539fa9e4066Sahrens if (ddi_create_minor_node(zfs_dip, blkbuf, S_IFBLK, 540fa9e4066Sahrens minor, DDI_PSEUDO, 0) == DDI_FAILURE) { 541fa9e4066Sahrens ddi_remove_minor_node(zfs_dip, chrbuf); 542c99e4bdcSChris Kirby ddi_soft_state_free(zfsdev_state, minor); 5436e0cbcaaSMatthew Ahrens dmu_objset_disown(os, FTAG); 544c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 545be6fd75aSMatthew Ahrens return (SET_ERROR(EAGAIN)); 546fa9e4066Sahrens } 547fa9e4066Sahrens 548c99e4bdcSChris Kirby zs = ddi_get_soft_state(zfsdev_state, minor); 549c99e4bdcSChris Kirby zs->zss_type = ZSST_ZVOL; 550c99e4bdcSChris Kirby zv = zs->zss_data = kmem_zalloc(sizeof (zvol_state_t), KM_SLEEP); 551681d9761SEric Taylor (void) strlcpy(zv->zv_name, name, MAXPATHLEN); 552fa9e4066Sahrens zv->zv_min_bs = DEV_BSHIFT; 553fa9e4066Sahrens zv->zv_minor = minor; 554fa9e4066Sahrens zv->zv_objset = os; 555f9af39baSGeorge Wilson if (dmu_objset_is_snapshot(os) || !spa_writeable(dmu_objset_spa(os))) 556681d9761SEric Taylor zv->zv_flags |= ZVOL_RDONLY; 557*79315247SMatthew Ahrens rangelock_init(&zv->zv_rangelock, NULL, NULL); 55888b7b0f2SMatthew Ahrens list_create(&zv->zv_extents, sizeof (zvol_extent_t), 55988b7b0f2SMatthew Ahrens offsetof(zvol_extent_t, ze_node)); 56067bd71c6Sperrin /* get and cache the blocksize */ 56167bd71c6Sperrin error = dmu_object_info(os, ZVOL_OBJ, &doi); 56267bd71c6Sperrin ASSERT(error == 0); 56367bd71c6Sperrin zv->zv_volblocksize = doi.doi_data_block_size; 56422ac5be4Sperrin 565f9af39baSGeorge Wilson if (spa_writeable(dmu_objset_spa(os))) { 566f9af39baSGeorge Wilson if (zil_replay_disable) 567f9af39baSGeorge Wilson zil_destroy(dmu_objset_zil(os), B_FALSE); 568f9af39baSGeorge Wilson else 569f9af39baSGeorge Wilson zil_replay(os, zv, zvol_replay_vector); 570f9af39baSGeorge Wilson } 5716e0cbcaaSMatthew Ahrens dmu_objset_disown(os, FTAG); 572681d9761SEric Taylor zv->zv_objset = NULL; 573fa9e4066Sahrens 574fa9e4066Sahrens zvol_minors++; 575fa9e4066Sahrens 576c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 577fa9e4066Sahrens 578fa9e4066Sahrens return (0); 579fa9e4066Sahrens } 580fa9e4066Sahrens 581fa9e4066Sahrens /* 582fa9e4066Sahrens * Remove minor node for the specified volume. 583fa9e4066Sahrens */ 584681d9761SEric Taylor static int 585681d9761SEric Taylor zvol_remove_zv(zvol_state_t *zv) 586681d9761SEric Taylor { 587681d9761SEric Taylor char nmbuf[20]; 588c99e4bdcSChris Kirby minor_t minor = zv->zv_minor; 589681d9761SEric Taylor 590c99e4bdcSChris Kirby ASSERT(MUTEX_HELD(&zfsdev_state_lock)); 591681d9761SEric Taylor if (zv->zv_total_opens != 0) 592be6fd75aSMatthew Ahrens return (SET_ERROR(EBUSY)); 593681d9761SEric Taylor 594c99e4bdcSChris Kirby (void) snprintf(nmbuf, sizeof (nmbuf), "%u,raw", minor); 595681d9761SEric Taylor ddi_remove_minor_node(zfs_dip, nmbuf); 596681d9761SEric Taylor 597c99e4bdcSChris Kirby (void) snprintf(nmbuf, sizeof (nmbuf), "%u", minor); 598681d9761SEric Taylor ddi_remove_minor_node(zfs_dip, nmbuf); 599681d9761SEric Taylor 600*79315247SMatthew Ahrens rangelock_fini(&zv->zv_rangelock); 601681d9761SEric Taylor 602c99e4bdcSChris Kirby kmem_free(zv, sizeof (zvol_state_t)); 603c99e4bdcSChris Kirby 604c99e4bdcSChris Kirby ddi_soft_state_free(zfsdev_state, minor); 605681d9761SEric Taylor 606681d9761SEric Taylor zvol_minors--; 607681d9761SEric Taylor return (0); 608681d9761SEric Taylor } 609681d9761SEric Taylor 610fa9e4066Sahrens int 611e9dbad6fSeschrock zvol_remove_minor(const char *name) 612fa9e4066Sahrens { 613fa9e4066Sahrens zvol_state_t *zv; 614681d9761SEric Taylor int rc; 615fa9e4066Sahrens 616c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 617e9dbad6fSeschrock if ((zv = zvol_minor_lookup(name)) == NULL) { 618c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 619be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 620fa9e4066Sahrens } 621681d9761SEric Taylor rc = zvol_remove_zv(zv); 622c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 623681d9761SEric Taylor return (rc); 624681d9761SEric Taylor } 625fa9e4066Sahrens 626681d9761SEric Taylor int 627681d9761SEric Taylor zvol_first_open(zvol_state_t *zv) 628681d9761SEric Taylor { 629681d9761SEric Taylor objset_t *os; 630681d9761SEric Taylor uint64_t volsize; 631681d9761SEric Taylor int error; 632681d9761SEric Taylor uint64_t readonly; 633fa9e4066Sahrens 634681d9761SEric Taylor /* lie and say we're read-only */ 635681d9761SEric Taylor error = dmu_objset_own(zv->zv_name, DMU_OST_ZVOL, B_TRUE, 636681d9761SEric Taylor zvol_tag, &os); 637681d9761SEric Taylor if (error) 638681d9761SEric Taylor return (error); 639fa9e4066Sahrens 640c61ea566SGeorge Wilson zv->zv_objset = os; 641681d9761SEric Taylor error = zap_lookup(os, ZVOL_ZAP_OBJ, "size", 8, 1, &volsize); 642681d9761SEric Taylor if (error) { 643681d9761SEric Taylor ASSERT(error == 0); 644681d9761SEric Taylor dmu_objset_disown(os, zvol_tag); 645681d9761SEric Taylor return (error); 646681d9761SEric Taylor } 647c61ea566SGeorge Wilson 6488dfe5547SRichard Yao error = dnode_hold(os, ZVOL_OBJ, zvol_tag, &zv->zv_dn); 64994d1a210STim Haley if (error) { 65094d1a210STim Haley dmu_objset_disown(os, zvol_tag); 65194d1a210STim Haley return (error); 65294d1a210STim Haley } 653c61ea566SGeorge Wilson 654c61ea566SGeorge Wilson zvol_size_changed(zv, volsize); 655681d9761SEric Taylor zv->zv_zilog = zil_open(os, zvol_get_data); 656fa9e4066Sahrens 657681d9761SEric Taylor VERIFY(dsl_prop_get_integer(zv->zv_name, "readonly", &readonly, 658681d9761SEric Taylor NULL) == 0); 659f9af39baSGeorge Wilson if (readonly || dmu_objset_is_snapshot(os) || 660f9af39baSGeorge Wilson !spa_writeable(dmu_objset_spa(os))) 661681d9761SEric Taylor zv->zv_flags |= ZVOL_RDONLY; 662681d9761SEric Taylor else 663681d9761SEric Taylor zv->zv_flags &= ~ZVOL_RDONLY; 664681d9761SEric Taylor return (error); 665681d9761SEric Taylor } 666fa9e4066Sahrens 667681d9761SEric Taylor void 668681d9761SEric Taylor zvol_last_close(zvol_state_t *zv) 669681d9761SEric Taylor { 67022ac5be4Sperrin zil_close(zv->zv_zilog); 67122ac5be4Sperrin zv->zv_zilog = NULL; 6722e2c1355SMatthew Ahrens 6738dfe5547SRichard Yao dnode_rele(zv->zv_dn, zvol_tag); 6748dfe5547SRichard Yao zv->zv_dn = NULL; 6752e2c1355SMatthew Ahrens 6762e2c1355SMatthew Ahrens /* 6772e2c1355SMatthew Ahrens * Evict cached data 6782e2c1355SMatthew Ahrens */ 6792e2c1355SMatthew Ahrens if (dsl_dataset_is_dirty(dmu_objset_ds(zv->zv_objset)) && 6802e2c1355SMatthew Ahrens !(zv->zv_flags & ZVOL_RDONLY)) 6812e2c1355SMatthew Ahrens txg_wait_synced(dmu_objset_pool(zv->zv_objset), 0); 6823b2aab18SMatthew Ahrens dmu_objset_evict_dbufs(zv->zv_objset); 6832e2c1355SMatthew Ahrens 684503ad85cSMatthew Ahrens dmu_objset_disown(zv->zv_objset, zvol_tag); 685fa9e4066Sahrens zv->zv_objset = NULL; 686fa9e4066Sahrens } 687fa9e4066Sahrens 688e7cbe64fSgw int 689e7cbe64fSgw zvol_prealloc(zvol_state_t *zv) 690e7cbe64fSgw { 691e7cbe64fSgw objset_t *os = zv->zv_objset; 692e7cbe64fSgw dmu_tx_t *tx; 693e7cbe64fSgw uint64_t refd, avail, usedobjs, availobjs; 694e7cbe64fSgw uint64_t resid = zv->zv_volsize; 695e7cbe64fSgw uint64_t off = 0; 696e7cbe64fSgw 697e7cbe64fSgw /* Check the space usage before attempting to allocate the space */ 698e7cbe64fSgw dmu_objset_space(os, &refd, &avail, &usedobjs, &availobjs); 699e7cbe64fSgw if (avail < zv->zv_volsize) 700be6fd75aSMatthew Ahrens return (SET_ERROR(ENOSPC)); 701e7cbe64fSgw 702e7cbe64fSgw /* Free old extents if they exist */ 703e7cbe64fSgw zvol_free_extents(zv); 704e7cbe64fSgw 705e7cbe64fSgw while (resid != 0) { 706e7cbe64fSgw int error; 707b5152584SMatthew Ahrens uint64_t bytes = MIN(resid, SPA_OLD_MAXBLOCKSIZE); 708e7cbe64fSgw 709e7cbe64fSgw tx = dmu_tx_create(os); 710e7cbe64fSgw dmu_tx_hold_write(tx, ZVOL_OBJ, off, bytes); 711e7cbe64fSgw error = dmu_tx_assign(tx, TXG_WAIT); 712e7cbe64fSgw if (error) { 713e7cbe64fSgw dmu_tx_abort(tx); 714cdb0ab79Smaybee (void) dmu_free_long_range(os, ZVOL_OBJ, 0, off); 715e7cbe64fSgw return (error); 716e7cbe64fSgw } 71782c9918fSTim Haley dmu_prealloc(os, ZVOL_OBJ, off, bytes, tx); 718e7cbe64fSgw dmu_tx_commit(tx); 719e7cbe64fSgw off += bytes; 720e7cbe64fSgw resid -= bytes; 721e7cbe64fSgw } 722e7cbe64fSgw txg_wait_synced(dmu_objset_pool(os), 0); 723e7cbe64fSgw 724e7cbe64fSgw return (0); 725e7cbe64fSgw } 726e7cbe64fSgw 7273b2aab18SMatthew Ahrens static int 728681d9761SEric Taylor zvol_update_volsize(objset_t *os, uint64_t volsize) 729e7cbe64fSgw { 730e7cbe64fSgw dmu_tx_t *tx; 731e7cbe64fSgw int error; 732e7cbe64fSgw 733c99e4bdcSChris Kirby ASSERT(MUTEX_HELD(&zfsdev_state_lock)); 734e7cbe64fSgw 735681d9761SEric Taylor tx = dmu_tx_create(os); 736e7cbe64fSgw dmu_tx_hold_zap(tx, ZVOL_ZAP_OBJ, TRUE, NULL); 7374bb73804SMatthew Ahrens dmu_tx_mark_netfree(tx); 738e7cbe64fSgw error = dmu_tx_assign(tx, TXG_WAIT); 739e7cbe64fSgw if (error) { 740e7cbe64fSgw dmu_tx_abort(tx); 741e7cbe64fSgw return (error); 742e7cbe64fSgw } 743e7cbe64fSgw 744681d9761SEric Taylor error = zap_update(os, ZVOL_ZAP_OBJ, "size", 8, 1, 745e7cbe64fSgw &volsize, tx); 746e7cbe64fSgw dmu_tx_commit(tx); 747e7cbe64fSgw 748e7cbe64fSgw if (error == 0) 749681d9761SEric Taylor error = dmu_free_long_range(os, 750cdb0ab79Smaybee ZVOL_OBJ, volsize, DMU_OBJECT_END); 751681d9761SEric Taylor return (error); 752681d9761SEric Taylor } 753e7cbe64fSgw 754681d9761SEric Taylor void 755681d9761SEric Taylor zvol_remove_minors(const char *name) 756681d9761SEric Taylor { 757681d9761SEric Taylor zvol_state_t *zv; 758681d9761SEric Taylor char *namebuf; 759681d9761SEric Taylor minor_t minor; 760681d9761SEric Taylor 761681d9761SEric Taylor namebuf = kmem_zalloc(strlen(name) + 2, KM_SLEEP); 762681d9761SEric Taylor (void) strncpy(namebuf, name, strlen(name)); 763681d9761SEric Taylor (void) strcat(namebuf, "/"); 764c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 765c99e4bdcSChris Kirby for (minor = 1; minor <= ZFSDEV_MAX_MINOR; minor++) { 766681d9761SEric Taylor 767c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 768681d9761SEric Taylor if (zv == NULL) 769681d9761SEric Taylor continue; 770681d9761SEric Taylor if (strncmp(namebuf, zv->zv_name, strlen(namebuf)) == 0) 771681d9761SEric Taylor (void) zvol_remove_zv(zv); 772e7cbe64fSgw } 773681d9761SEric Taylor kmem_free(namebuf, strlen(name) + 2); 774681d9761SEric Taylor 775c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 776e7cbe64fSgw } 777e7cbe64fSgw 778c61ea566SGeorge Wilson static int 7793b2aab18SMatthew Ahrens zvol_update_live_volsize(zvol_state_t *zv, uint64_t volsize) 780fa9e4066Sahrens { 781e7cbe64fSgw uint64_t old_volsize = 0ULL; 7823b2aab18SMatthew Ahrens int error = 0; 783fa9e4066Sahrens 784c61ea566SGeorge Wilson ASSERT(MUTEX_HELD(&zfsdev_state_lock)); 785c61ea566SGeorge Wilson 786e7cbe64fSgw /* 787e7cbe64fSgw * Reinitialize the dump area to the new size. If we 788681d9761SEric Taylor * failed to resize the dump area then restore it back to 789c61ea566SGeorge Wilson * its original size. We must set the new volsize prior 790c61ea566SGeorge Wilson * to calling dumpvp_resize() to ensure that the devices' 791c61ea566SGeorge Wilson * size(9P) is not visible by the dump subsystem. 792e7cbe64fSgw */ 7933b2aab18SMatthew Ahrens old_volsize = zv->zv_volsize; 7943b2aab18SMatthew Ahrens zvol_size_changed(zv, volsize); 7953b2aab18SMatthew Ahrens 7963b2aab18SMatthew Ahrens if (zv->zv_flags & ZVOL_DUMPIFIED) { 7973b2aab18SMatthew Ahrens if ((error = zvol_dumpify(zv)) != 0 || 7983b2aab18SMatthew Ahrens (error = dumpvp_resize()) != 0) { 7993b2aab18SMatthew Ahrens int dumpify_error; 8003b2aab18SMatthew Ahrens 8013b2aab18SMatthew Ahrens (void) zvol_update_volsize(zv->zv_objset, old_volsize); 8023b2aab18SMatthew Ahrens zvol_size_changed(zv, old_volsize); 8033b2aab18SMatthew Ahrens dumpify_error = zvol_dumpify(zv); 8043b2aab18SMatthew Ahrens error = dumpify_error ? dumpify_error : error; 805681d9761SEric Taylor } 806fa9e4066Sahrens } 807fa9e4066Sahrens 808573ca77eSGeorge Wilson /* 809573ca77eSGeorge Wilson * Generate a LUN expansion event. 810573ca77eSGeorge Wilson */ 8113b2aab18SMatthew Ahrens if (error == 0) { 812573ca77eSGeorge Wilson sysevent_id_t eid; 813573ca77eSGeorge Wilson nvlist_t *attr; 814573ca77eSGeorge Wilson char *physpath = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 815573ca77eSGeorge Wilson 816681d9761SEric Taylor (void) snprintf(physpath, MAXPATHLEN, "%s%u", ZVOL_PSEUDO_DEV, 817573ca77eSGeorge Wilson zv->zv_minor); 818573ca77eSGeorge Wilson 819573ca77eSGeorge Wilson VERIFY(nvlist_alloc(&attr, NV_UNIQUE_NAME, KM_SLEEP) == 0); 820573ca77eSGeorge Wilson VERIFY(nvlist_add_string(attr, DEV_PHYS_PATH, physpath) == 0); 821573ca77eSGeorge Wilson 822573ca77eSGeorge Wilson (void) ddi_log_sysevent(zfs_dip, SUNW_VENDOR, EC_DEV_STATUS, 823573ca77eSGeorge Wilson ESC_DEV_DLE, attr, &eid, DDI_SLEEP); 824573ca77eSGeorge Wilson 825573ca77eSGeorge Wilson nvlist_free(attr); 826573ca77eSGeorge Wilson kmem_free(physpath, MAXPATHLEN); 827573ca77eSGeorge Wilson } 828c61ea566SGeorge Wilson return (error); 829c61ea566SGeorge Wilson } 830573ca77eSGeorge Wilson 831c61ea566SGeorge Wilson int 832c61ea566SGeorge Wilson zvol_set_volsize(const char *name, uint64_t volsize) 833c61ea566SGeorge Wilson { 834c61ea566SGeorge Wilson zvol_state_t *zv = NULL; 835c61ea566SGeorge Wilson objset_t *os; 836c61ea566SGeorge Wilson int error; 837c61ea566SGeorge Wilson dmu_object_info_t doi; 838c61ea566SGeorge Wilson uint64_t readonly; 8393b2aab18SMatthew Ahrens boolean_t owned = B_FALSE; 8403b2aab18SMatthew Ahrens 8413b2aab18SMatthew Ahrens error = dsl_prop_get_integer(name, 8423b2aab18SMatthew Ahrens zfs_prop_to_name(ZFS_PROP_READONLY), &readonly, NULL); 8433b2aab18SMatthew Ahrens if (error != 0) 8443b2aab18SMatthew Ahrens return (error); 8453b2aab18SMatthew Ahrens if (readonly) 846be6fd75aSMatthew Ahrens return (SET_ERROR(EROFS)); 847c61ea566SGeorge Wilson 848c61ea566SGeorge Wilson mutex_enter(&zfsdev_state_lock); 849c61ea566SGeorge Wilson zv = zvol_minor_lookup(name); 8503b2aab18SMatthew Ahrens 8513b2aab18SMatthew Ahrens if (zv == NULL || zv->zv_objset == NULL) { 8523b2aab18SMatthew Ahrens if ((error = dmu_objset_own(name, DMU_OST_ZVOL, B_FALSE, 8533b2aab18SMatthew Ahrens FTAG, &os)) != 0) { 8543b2aab18SMatthew Ahrens mutex_exit(&zfsdev_state_lock); 8553b2aab18SMatthew Ahrens return (error); 8563b2aab18SMatthew Ahrens } 8573b2aab18SMatthew Ahrens owned = B_TRUE; 8583b2aab18SMatthew Ahrens if (zv != NULL) 8593b2aab18SMatthew Ahrens zv->zv_objset = os; 8603b2aab18SMatthew Ahrens } else { 8613b2aab18SMatthew Ahrens os = zv->zv_objset; 862c61ea566SGeorge Wilson } 863c61ea566SGeorge Wilson 864c61ea566SGeorge Wilson if ((error = dmu_object_info(os, ZVOL_OBJ, &doi)) != 0 || 8653b2aab18SMatthew Ahrens (error = zvol_check_volsize(volsize, doi.doi_data_block_size)) != 0) 866c61ea566SGeorge Wilson goto out; 867c61ea566SGeorge Wilson 8683b2aab18SMatthew Ahrens error = zvol_update_volsize(os, volsize); 869c61ea566SGeorge Wilson 8703b2aab18SMatthew Ahrens if (error == 0 && zv != NULL) 8713b2aab18SMatthew Ahrens error = zvol_update_live_volsize(zv, volsize); 872bb0ade09Sahrens out: 8733b2aab18SMatthew Ahrens if (owned) { 8743b2aab18SMatthew Ahrens dmu_objset_disown(os, FTAG); 8753b2aab18SMatthew Ahrens if (zv != NULL) 8763b2aab18SMatthew Ahrens zv->zv_objset = NULL; 8773b2aab18SMatthew Ahrens } 878c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 879fa9e4066Sahrens return (error); 880fa9e4066Sahrens } 881fa9e4066Sahrens 882fa9e4066Sahrens /*ARGSUSED*/ 883fa9e4066Sahrens int 884fa9e4066Sahrens zvol_open(dev_t *devp, int flag, int otyp, cred_t *cr) 885fa9e4066Sahrens { 886fa9e4066Sahrens zvol_state_t *zv; 887681d9761SEric Taylor int err = 0; 888fa9e4066Sahrens 889c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 890fa9e4066Sahrens 891c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(getminor(*devp), ZSST_ZVOL); 892fa9e4066Sahrens if (zv == NULL) { 893c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 894be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 895fa9e4066Sahrens } 896fa9e4066Sahrens 897681d9761SEric Taylor if (zv->zv_total_opens == 0) 898681d9761SEric Taylor err = zvol_first_open(zv); 899681d9761SEric Taylor if (err) { 900c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 901681d9761SEric Taylor return (err); 902681d9761SEric Taylor } 903681d9761SEric Taylor if ((flag & FWRITE) && (zv->zv_flags & ZVOL_RDONLY)) { 904be6fd75aSMatthew Ahrens err = SET_ERROR(EROFS); 905681d9761SEric Taylor goto out; 906fa9e4066Sahrens } 907c7f714e2SEric Taylor if (zv->zv_flags & ZVOL_EXCL) { 908be6fd75aSMatthew Ahrens err = SET_ERROR(EBUSY); 909681d9761SEric Taylor goto out; 910c7f714e2SEric Taylor } 911c7f714e2SEric Taylor if (flag & FEXCL) { 912c7f714e2SEric Taylor if (zv->zv_total_opens != 0) { 913be6fd75aSMatthew Ahrens err = SET_ERROR(EBUSY); 914681d9761SEric Taylor goto out; 915c7f714e2SEric Taylor } 916c7f714e2SEric Taylor zv->zv_flags |= ZVOL_EXCL; 917c7f714e2SEric Taylor } 918fa9e4066Sahrens 919fa9e4066Sahrens if (zv->zv_open_count[otyp] == 0 || otyp == OTYP_LYR) { 920fa9e4066Sahrens zv->zv_open_count[otyp]++; 921fa9e4066Sahrens zv->zv_total_opens++; 922fa9e4066Sahrens } 923c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 924fa9e4066Sahrens 925681d9761SEric Taylor return (err); 926681d9761SEric Taylor out: 927681d9761SEric Taylor if (zv->zv_total_opens == 0) 928681d9761SEric Taylor zvol_last_close(zv); 929c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 930681d9761SEric Taylor return (err); 931fa9e4066Sahrens } 932fa9e4066Sahrens 933fa9e4066Sahrens /*ARGSUSED*/ 934fa9e4066Sahrens int 935fa9e4066Sahrens zvol_close(dev_t dev, int flag, int otyp, cred_t *cr) 936fa9e4066Sahrens { 937fa9e4066Sahrens minor_t minor = getminor(dev); 938fa9e4066Sahrens zvol_state_t *zv; 939681d9761SEric Taylor int error = 0; 940fa9e4066Sahrens 941c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 942fa9e4066Sahrens 943c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 944fa9e4066Sahrens if (zv == NULL) { 945c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 946be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 947fa9e4066Sahrens } 948fa9e4066Sahrens 949c7f714e2SEric Taylor if (zv->zv_flags & ZVOL_EXCL) { 950c7f714e2SEric Taylor ASSERT(zv->zv_total_opens == 1); 951c7f714e2SEric Taylor zv->zv_flags &= ~ZVOL_EXCL; 952fa9e4066Sahrens } 953fa9e4066Sahrens 954fa9e4066Sahrens /* 955fa9e4066Sahrens * If the open count is zero, this is a spurious close. 956fa9e4066Sahrens * That indicates a bug in the kernel / DDI framework. 957fa9e4066Sahrens */ 958fa9e4066Sahrens ASSERT(zv->zv_open_count[otyp] != 0); 959fa9e4066Sahrens ASSERT(zv->zv_total_opens != 0); 960fa9e4066Sahrens 961fa9e4066Sahrens /* 962fa9e4066Sahrens * You may get multiple opens, but only one close. 963fa9e4066Sahrens */ 964fa9e4066Sahrens zv->zv_open_count[otyp]--; 965fa9e4066Sahrens zv->zv_total_opens--; 966fa9e4066Sahrens 967681d9761SEric Taylor if (zv->zv_total_opens == 0) 968681d9761SEric Taylor zvol_last_close(zv); 969fa9e4066Sahrens 970c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 971681d9761SEric Taylor return (error); 972fa9e4066Sahrens } 973fa9e4066Sahrens 974cab3a55eSPrakash Surya /* ARGSUSED */ 975feb08c6bSbillm static void 976b24ab676SJeff Bonwick zvol_get_done(zgd_t *zgd, int error) 97767bd71c6Sperrin { 978b24ab676SJeff Bonwick if (zgd->zgd_db) 979b24ab676SJeff Bonwick dmu_buf_rele(zgd->zgd_db, zgd); 980b24ab676SJeff Bonwick 981*79315247SMatthew Ahrens rangelock_exit(zgd->zgd_lr); 982b24ab676SJeff Bonwick 98367bd71c6Sperrin kmem_free(zgd, sizeof (zgd_t)); 98467bd71c6Sperrin } 98567bd71c6Sperrin 98667bd71c6Sperrin /* 98767bd71c6Sperrin * Get data to generate a TX_WRITE intent log record. 98867bd71c6Sperrin */ 989feb08c6bSbillm static int 9901271e4b1SPrakash Surya zvol_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio) 99167bd71c6Sperrin { 99267bd71c6Sperrin zvol_state_t *zv = arg; 993b24ab676SJeff Bonwick uint64_t offset = lr->lr_offset; 994b24ab676SJeff Bonwick uint64_t size = lr->lr_length; /* length of user data */ 99567bd71c6Sperrin dmu_buf_t *db; 99667bd71c6Sperrin zgd_t *zgd; 99767bd71c6Sperrin int error; 99867bd71c6Sperrin 9991271e4b1SPrakash Surya ASSERT3P(lwb, !=, NULL); 10001271e4b1SPrakash Surya ASSERT3P(zio, !=, NULL); 10011271e4b1SPrakash Surya ASSERT3U(size, !=, 0); 1002b24ab676SJeff Bonwick 1003b24ab676SJeff Bonwick zgd = kmem_zalloc(sizeof (zgd_t), KM_SLEEP); 10041271e4b1SPrakash Surya zgd->zgd_lwb = lwb; 1005feb08c6bSbillm 1006c2e6a7d6Sperrin /* 1007c2e6a7d6Sperrin * Write records come in two flavors: immediate and indirect. 1008c2e6a7d6Sperrin * For small writes it's cheaper to store the data with the 1009c2e6a7d6Sperrin * log record (immediate); for large writes it's cheaper to 1010c2e6a7d6Sperrin * sync the data and get a pointer to it (indirect) so that 1011c2e6a7d6Sperrin * we don't have to write the data twice. 1012c2e6a7d6Sperrin */ 101342b14111SLOLi if (buf != NULL) { /* immediate write */ 1014*79315247SMatthew Ahrens zgd->zgd_lr = rangelock_enter(&zv->zv_rangelock, offset, size, 101542b14111SLOLi RL_READER); 10168dfe5547SRichard Yao error = dmu_read_by_dnode(zv->zv_dn, offset, size, buf, 1017b24ab676SJeff Bonwick DMU_READ_NO_PREFETCH); 101842b14111SLOLi } else { /* indirect write */ 101942b14111SLOLi /* 102042b14111SLOLi * Have to lock the whole block to ensure when it's written out 102142b14111SLOLi * and its checksum is being calculated that no one can change 102242b14111SLOLi * the data. Contrarily to zfs_get_data we need not re-check 102342b14111SLOLi * blocksize after we get the lock because it cannot be changed. 102442b14111SLOLi */ 1025b24ab676SJeff Bonwick size = zv->zv_volblocksize; 1026b24ab676SJeff Bonwick offset = P2ALIGN(offset, size); 1027*79315247SMatthew Ahrens zgd->zgd_lr = rangelock_enter(&zv->zv_rangelock, offset, size, 102842b14111SLOLi RL_READER); 10298dfe5547SRichard Yao error = dmu_buf_hold_by_dnode(zv->zv_dn, offset, zgd, &db, 103047cb52daSJeff Bonwick DMU_READ_NO_PREFETCH); 1031b24ab676SJeff Bonwick if (error == 0) { 1032b7edcb94SMatthew Ahrens blkptr_t *bp = &lr->lr_blkptr; 103380901aeaSGeorge Wilson 1034b24ab676SJeff Bonwick zgd->zgd_db = db; 1035b24ab676SJeff Bonwick zgd->zgd_bp = bp; 103667bd71c6Sperrin 1037b24ab676SJeff Bonwick ASSERT(db->db_offset == offset); 1038b24ab676SJeff Bonwick ASSERT(db->db_size == size); 103967bd71c6Sperrin 1040b24ab676SJeff Bonwick error = dmu_sync(zio, lr->lr_common.lrc_txg, 1041b24ab676SJeff Bonwick zvol_get_done, zgd); 1042975c32a0SNeil Perrin 1043b24ab676SJeff Bonwick if (error == 0) 1044b24ab676SJeff Bonwick return (0); 1045b24ab676SJeff Bonwick } 1046975c32a0SNeil Perrin } 1047975c32a0SNeil Perrin 1048b24ab676SJeff Bonwick zvol_get_done(zgd, error); 1049b24ab676SJeff Bonwick 105067bd71c6Sperrin return (error); 105167bd71c6Sperrin } 105267bd71c6Sperrin 1053a24e15ceSperrin /* 1054a24e15ceSperrin * zvol_log_write() handles synchronous writes using TX_WRITE ZIL transactions. 105522ac5be4Sperrin * 105622ac5be4Sperrin * We store data in the log buffers if it's small enough. 105767bd71c6Sperrin * Otherwise we will later flush the data out via dmu_sync(). 105822ac5be4Sperrin */ 105967bd71c6Sperrin ssize_t zvol_immediate_write_sz = 32768; 106022ac5be4Sperrin 1061feb08c6bSbillm static void 1062510b6c0eSNeil Perrin zvol_log_write(zvol_state_t *zv, dmu_tx_t *tx, offset_t off, ssize_t resid, 1063510b6c0eSNeil Perrin boolean_t sync) 106422ac5be4Sperrin { 1065feb08c6bSbillm uint32_t blocksize = zv->zv_volblocksize; 10661209a471SNeil Perrin zilog_t *zilog = zv->zv_zilog; 1067c5ee4681SAlexander Motin itx_wr_state_t write_state; 1068510b6c0eSNeil Perrin 1069b24ab676SJeff Bonwick if (zil_replaying(zilog, tx)) 10701209a471SNeil Perrin return; 10711209a471SNeil Perrin 1072c5ee4681SAlexander Motin if (zilog->zl_logbias == ZFS_LOGBIAS_THROUGHPUT) 1073c5ee4681SAlexander Motin write_state = WR_INDIRECT; 1074c5ee4681SAlexander Motin else if (!spa_has_slogs(zilog->zl_spa) && 1075c5ee4681SAlexander Motin resid >= blocksize && blocksize > zvol_immediate_write_sz) 1076c5ee4681SAlexander Motin write_state = WR_INDIRECT; 1077c5ee4681SAlexander Motin else if (sync) 1078c5ee4681SAlexander Motin write_state = WR_COPIED; 1079c5ee4681SAlexander Motin else 1080c5ee4681SAlexander Motin write_state = WR_NEED_COPY; 1081feb08c6bSbillm 1082510b6c0eSNeil Perrin while (resid) { 1083510b6c0eSNeil Perrin itx_t *itx; 1084510b6c0eSNeil Perrin lr_write_t *lr; 1085c5ee4681SAlexander Motin itx_wr_state_t wr_state = write_state; 1086c5ee4681SAlexander Motin ssize_t len = resid; 1087c5ee4681SAlexander Motin 1088c5ee4681SAlexander Motin if (wr_state == WR_COPIED && resid > ZIL_MAX_COPIED_DATA) 1089c5ee4681SAlexander Motin wr_state = WR_NEED_COPY; 1090c5ee4681SAlexander Motin else if (wr_state == WR_INDIRECT) 1091c5ee4681SAlexander Motin len = MIN(blocksize - P2PHASE(off, blocksize), resid); 1092510b6c0eSNeil Perrin 1093510b6c0eSNeil Perrin itx = zil_itx_create(TX_WRITE, sizeof (*lr) + 1094c5ee4681SAlexander Motin (wr_state == WR_COPIED ? len : 0)); 1095feb08c6bSbillm lr = (lr_write_t *)&itx->itx_lr; 10968dfe5547SRichard Yao if (wr_state == WR_COPIED && dmu_read_by_dnode(zv->zv_dn, 10978dfe5547SRichard Yao off, len, lr + 1, DMU_READ_NO_PREFETCH) != 0) { 1098b24ab676SJeff Bonwick zil_itx_destroy(itx); 1099510b6c0eSNeil Perrin itx = zil_itx_create(TX_WRITE, sizeof (*lr)); 1100510b6c0eSNeil Perrin lr = (lr_write_t *)&itx->itx_lr; 1101c5ee4681SAlexander Motin wr_state = WR_NEED_COPY; 1102510b6c0eSNeil Perrin } 1103510b6c0eSNeil Perrin 1104c5ee4681SAlexander Motin itx->itx_wr_state = wr_state; 1105feb08c6bSbillm lr->lr_foid = ZVOL_OBJ; 1106feb08c6bSbillm lr->lr_offset = off; 1107510b6c0eSNeil Perrin lr->lr_length = len; 1108b24ab676SJeff Bonwick lr->lr_blkoff = 0; 1109feb08c6bSbillm BP_ZERO(&lr->lr_blkptr); 1110feb08c6bSbillm 1111510b6c0eSNeil Perrin itx->itx_private = zv; 1112510b6c0eSNeil Perrin itx->itx_sync = sync; 1113510b6c0eSNeil Perrin 11145002558fSNeil Perrin zil_itx_assign(zilog, itx, tx); 1115510b6c0eSNeil Perrin 1116510b6c0eSNeil Perrin off += len; 1117510b6c0eSNeil Perrin resid -= len; 111822ac5be4Sperrin } 111922ac5be4Sperrin } 112022ac5be4Sperrin 112188b7b0f2SMatthew Ahrens static int 1122810e43b2SBill Pijewski zvol_dumpio_vdev(vdev_t *vd, void *addr, uint64_t offset, uint64_t origoffset, 1123810e43b2SBill Pijewski uint64_t size, boolean_t doread, boolean_t isdump) 1124e7cbe64fSgw { 1125e7cbe64fSgw vdev_disk_t *dvd; 1126e7cbe64fSgw int c; 1127e7cbe64fSgw int numerrors = 0; 1128e7cbe64fSgw 1129810e43b2SBill Pijewski if (vd->vdev_ops == &vdev_mirror_ops || 1130810e43b2SBill Pijewski vd->vdev_ops == &vdev_replacing_ops || 1131810e43b2SBill Pijewski vd->vdev_ops == &vdev_spare_ops) { 1132810e43b2SBill Pijewski for (c = 0; c < vd->vdev_children; c++) { 1133810e43b2SBill Pijewski int err = zvol_dumpio_vdev(vd->vdev_child[c], 1134810e43b2SBill Pijewski addr, offset, origoffset, size, doread, isdump); 1135810e43b2SBill Pijewski if (err != 0) { 1136810e43b2SBill Pijewski numerrors++; 1137810e43b2SBill Pijewski } else if (doread) { 1138810e43b2SBill Pijewski break; 1139810e43b2SBill Pijewski } 1140e7cbe64fSgw } 1141e7cbe64fSgw } 1142e7cbe64fSgw 1143810e43b2SBill Pijewski if (!vd->vdev_ops->vdev_op_leaf && vd->vdev_ops != &vdev_raidz_ops) 1144e7cbe64fSgw return (numerrors < vd->vdev_children ? 0 : EIO); 1145e7cbe64fSgw 1146dc0bb255SEric Taylor if (doread && !vdev_readable(vd)) 1147be6fd75aSMatthew Ahrens return (SET_ERROR(EIO)); 1148dc0bb255SEric Taylor else if (!doread && !vdev_writeable(vd)) 1149be6fd75aSMatthew Ahrens return (SET_ERROR(EIO)); 1150e7cbe64fSgw 1151810e43b2SBill Pijewski if (vd->vdev_ops == &vdev_raidz_ops) { 1152810e43b2SBill Pijewski return (vdev_raidz_physio(vd, 1153810e43b2SBill Pijewski addr, size, offset, origoffset, doread, isdump)); 1154810e43b2SBill Pijewski } 1155810e43b2SBill Pijewski 1156e7cbe64fSgw offset += VDEV_LABEL_START_SIZE; 1157e7cbe64fSgw 1158e7cbe64fSgw if (ddi_in_panic() || isdump) { 115988b7b0f2SMatthew Ahrens ASSERT(!doread); 116088b7b0f2SMatthew Ahrens if (doread) 1161be6fd75aSMatthew Ahrens return (SET_ERROR(EIO)); 1162810e43b2SBill Pijewski dvd = vd->vdev_tsd; 1163810e43b2SBill Pijewski ASSERT3P(dvd, !=, NULL); 1164e7cbe64fSgw return (ldi_dump(dvd->vd_lh, addr, lbtodb(offset), 1165e7cbe64fSgw lbtodb(size))); 1166e7cbe64fSgw } else { 1167810e43b2SBill Pijewski dvd = vd->vdev_tsd; 1168810e43b2SBill Pijewski ASSERT3P(dvd, !=, NULL); 1169810e43b2SBill Pijewski return (vdev_disk_ldi_physio(dvd->vd_lh, addr, size, 1170810e43b2SBill Pijewski offset, doread ? B_READ : B_WRITE)); 1171e7cbe64fSgw } 1172e7cbe64fSgw } 1173e7cbe64fSgw 117488b7b0f2SMatthew Ahrens static int 117588b7b0f2SMatthew Ahrens zvol_dumpio(zvol_state_t *zv, void *addr, uint64_t offset, uint64_t size, 117688b7b0f2SMatthew Ahrens boolean_t doread, boolean_t isdump) 1177e7cbe64fSgw { 1178e7cbe64fSgw vdev_t *vd; 1179e7cbe64fSgw int error; 118088b7b0f2SMatthew Ahrens zvol_extent_t *ze; 1181e7cbe64fSgw spa_t *spa = dmu_objset_spa(zv->zv_objset); 1182e7cbe64fSgw 118388b7b0f2SMatthew Ahrens /* Must be sector aligned, and not stradle a block boundary. */ 118488b7b0f2SMatthew Ahrens if (P2PHASE(offset, DEV_BSIZE) || P2PHASE(size, DEV_BSIZE) || 118588b7b0f2SMatthew Ahrens P2BOUNDARY(offset, size, zv->zv_volblocksize)) { 1186be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 118788b7b0f2SMatthew Ahrens } 118888b7b0f2SMatthew Ahrens ASSERT(size <= zv->zv_volblocksize); 1189e7cbe64fSgw 119088b7b0f2SMatthew Ahrens /* Locate the extent this belongs to */ 119188b7b0f2SMatthew Ahrens ze = list_head(&zv->zv_extents); 119288b7b0f2SMatthew Ahrens while (offset >= ze->ze_nblks * zv->zv_volblocksize) { 119388b7b0f2SMatthew Ahrens offset -= ze->ze_nblks * zv->zv_volblocksize; 119488b7b0f2SMatthew Ahrens ze = list_next(&zv->zv_extents, ze); 119588b7b0f2SMatthew Ahrens } 119624cc0e1cSGeorge Wilson 11973b2aab18SMatthew Ahrens if (ze == NULL) 1198be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 11993b2aab18SMatthew Ahrens 120024cc0e1cSGeorge Wilson if (!ddi_in_panic()) 120124cc0e1cSGeorge Wilson spa_config_enter(spa, SCL_STATE, FTAG, RW_READER); 120224cc0e1cSGeorge Wilson 120388b7b0f2SMatthew Ahrens vd = vdev_lookup_top(spa, DVA_GET_VDEV(&ze->ze_dva)); 120488b7b0f2SMatthew Ahrens offset += DVA_GET_OFFSET(&ze->ze_dva); 1205810e43b2SBill Pijewski error = zvol_dumpio_vdev(vd, addr, offset, DVA_GET_OFFSET(&ze->ze_dva), 1206810e43b2SBill Pijewski size, doread, isdump); 120724cc0e1cSGeorge Wilson 120824cc0e1cSGeorge Wilson if (!ddi_in_panic()) 120924cc0e1cSGeorge Wilson spa_config_exit(spa, SCL_STATE, FTAG); 121024cc0e1cSGeorge Wilson 1211e7cbe64fSgw return (error); 1212e7cbe64fSgw } 1213e7cbe64fSgw 1214fa9e4066Sahrens int 1215fa9e4066Sahrens zvol_strategy(buf_t *bp) 1216fa9e4066Sahrens { 1217c99e4bdcSChris Kirby zfs_soft_state_t *zs = NULL; 1218c99e4bdcSChris Kirby zvol_state_t *zv; 1219fa9e4066Sahrens uint64_t off, volsize; 122088b7b0f2SMatthew Ahrens size_t resid; 1221fa9e4066Sahrens char *addr; 122222ac5be4Sperrin objset_t *os; 1223fa9e4066Sahrens int error = 0; 122488b7b0f2SMatthew Ahrens boolean_t doread = bp->b_flags & B_READ; 1225810e43b2SBill Pijewski boolean_t is_dumpified; 1226510b6c0eSNeil Perrin boolean_t sync; 1227fa9e4066Sahrens 1228c99e4bdcSChris Kirby if (getminor(bp->b_edev) == 0) { 1229be6fd75aSMatthew Ahrens error = SET_ERROR(EINVAL); 1230c99e4bdcSChris Kirby } else { 1231c99e4bdcSChris Kirby zs = ddi_get_soft_state(zfsdev_state, getminor(bp->b_edev)); 1232c99e4bdcSChris Kirby if (zs == NULL) 1233be6fd75aSMatthew Ahrens error = SET_ERROR(ENXIO); 1234c99e4bdcSChris Kirby else if (zs->zss_type != ZSST_ZVOL) 1235be6fd75aSMatthew Ahrens error = SET_ERROR(EINVAL); 1236fa9e4066Sahrens } 1237fa9e4066Sahrens 1238c99e4bdcSChris Kirby if (error) { 1239c99e4bdcSChris Kirby bioerror(bp, error); 1240fa9e4066Sahrens biodone(bp); 1241fa9e4066Sahrens return (0); 1242fa9e4066Sahrens } 1243fa9e4066Sahrens 1244c99e4bdcSChris Kirby zv = zs->zss_data; 1245c99e4bdcSChris Kirby 1246681d9761SEric Taylor if (!(bp->b_flags & B_READ) && (zv->zv_flags & ZVOL_RDONLY)) { 1247fa9e4066Sahrens bioerror(bp, EROFS); 1248fa9e4066Sahrens biodone(bp); 1249fa9e4066Sahrens return (0); 1250fa9e4066Sahrens } 1251fa9e4066Sahrens 1252fa9e4066Sahrens off = ldbtob(bp->b_blkno); 1253fa9e4066Sahrens volsize = zv->zv_volsize; 1254fa9e4066Sahrens 125522ac5be4Sperrin os = zv->zv_objset; 125622ac5be4Sperrin ASSERT(os != NULL); 1257fa9e4066Sahrens 1258fa9e4066Sahrens bp_mapin(bp); 1259fa9e4066Sahrens addr = bp->b_un.b_addr; 1260fa9e4066Sahrens resid = bp->b_bcount; 1261fa9e4066Sahrens 126288b7b0f2SMatthew Ahrens if (resid > 0 && (off < 0 || off >= volsize)) { 126388b7b0f2SMatthew Ahrens bioerror(bp, EIO); 126488b7b0f2SMatthew Ahrens biodone(bp); 126588b7b0f2SMatthew Ahrens return (0); 126688b7b0f2SMatthew Ahrens } 126773ec3d9cSgw 1268810e43b2SBill Pijewski is_dumpified = zv->zv_flags & ZVOL_DUMPIFIED; 126955da60b9SMark J Musante sync = ((!(bp->b_flags & B_ASYNC) && 127055da60b9SMark J Musante !(zv->zv_flags & ZVOL_WCE)) || 127155da60b9SMark J Musante (zv->zv_objset->os_sync == ZFS_SYNC_ALWAYS)) && 1272810e43b2SBill Pijewski !doread && !is_dumpified; 1273510b6c0eSNeil Perrin 1274a24e15ceSperrin /* 1275a24e15ceSperrin * There must be no buffer changes when doing a dmu_sync() because 1276a24e15ceSperrin * we can't change the data whilst calculating the checksum. 1277a24e15ceSperrin */ 1278*79315247SMatthew Ahrens locked_range_t *lr = rangelock_enter(&zv->zv_rangelock, off, resid, 127988b7b0f2SMatthew Ahrens doread ? RL_READER : RL_WRITER); 1280fa9e4066Sahrens 1281e7cbe64fSgw while (resid != 0 && off < volsize) { 128288b7b0f2SMatthew Ahrens size_t size = MIN(resid, zvol_maxphys); 1283810e43b2SBill Pijewski if (is_dumpified) { 1284e7cbe64fSgw size = MIN(size, P2END(off, zv->zv_volblocksize) - off); 128588b7b0f2SMatthew Ahrens error = zvol_dumpio(zv, addr, off, size, 128688b7b0f2SMatthew Ahrens doread, B_FALSE); 128788b7b0f2SMatthew Ahrens } else if (doread) { 12887bfdf011SNeil Perrin error = dmu_read(os, ZVOL_OBJ, off, size, addr, 12897bfdf011SNeil Perrin DMU_READ_PREFETCH); 1290fa9e4066Sahrens } else { 129122ac5be4Sperrin dmu_tx_t *tx = dmu_tx_create(os); 1292fa9e4066Sahrens dmu_tx_hold_write(tx, ZVOL_OBJ, off, size); 1293fa9e4066Sahrens error = dmu_tx_assign(tx, TXG_WAIT); 1294fa9e4066Sahrens if (error) { 1295fa9e4066Sahrens dmu_tx_abort(tx); 1296fa9e4066Sahrens } else { 129722ac5be4Sperrin dmu_write(os, ZVOL_OBJ, off, size, addr, tx); 1298510b6c0eSNeil Perrin zvol_log_write(zv, tx, off, size, sync); 1299fa9e4066Sahrens dmu_tx_commit(tx); 1300fa9e4066Sahrens } 1301fa9e4066Sahrens } 1302b87f3af3Sperrin if (error) { 1303b87f3af3Sperrin /* convert checksum errors into IO errors */ 1304b87f3af3Sperrin if (error == ECKSUM) 1305be6fd75aSMatthew Ahrens error = SET_ERROR(EIO); 1306fa9e4066Sahrens break; 1307b87f3af3Sperrin } 1308fa9e4066Sahrens off += size; 1309fa9e4066Sahrens addr += size; 1310fa9e4066Sahrens resid -= size; 1311fa9e4066Sahrens } 1312*79315247SMatthew Ahrens rangelock_exit(lr); 1313fa9e4066Sahrens 1314fa9e4066Sahrens if ((bp->b_resid = resid) == bp->b_bcount) 1315fa9e4066Sahrens bioerror(bp, off > volsize ? EINVAL : error); 1316fa9e4066Sahrens 1317510b6c0eSNeil Perrin if (sync) 13185002558fSNeil Perrin zil_commit(zv->zv_zilog, ZVOL_OBJ); 1319feb08c6bSbillm biodone(bp); 132022ac5be4Sperrin 1321fa9e4066Sahrens return (0); 1322fa9e4066Sahrens } 1323fa9e4066Sahrens 132467bd71c6Sperrin /* 132567bd71c6Sperrin * Set the buffer count to the zvol maximum transfer. 132667bd71c6Sperrin * Using our own routine instead of the default minphys() 132767bd71c6Sperrin * means that for larger writes we write bigger buffers on X86 132867bd71c6Sperrin * (128K instead of 56K) and flush the disk write cache less often 132967bd71c6Sperrin * (every zvol_maxphys - currently 1MB) instead of minphys (currently 133067bd71c6Sperrin * 56K on X86 and 128K on sparc). 133167bd71c6Sperrin */ 133267bd71c6Sperrin void 133367bd71c6Sperrin zvol_minphys(struct buf *bp) 133467bd71c6Sperrin { 133567bd71c6Sperrin if (bp->b_bcount > zvol_maxphys) 133667bd71c6Sperrin bp->b_bcount = zvol_maxphys; 133767bd71c6Sperrin } 133867bd71c6Sperrin 1339e7cbe64fSgw int 1340e7cbe64fSgw zvol_dump(dev_t dev, caddr_t addr, daddr_t blkno, int nblocks) 1341e7cbe64fSgw { 1342e7cbe64fSgw minor_t minor = getminor(dev); 1343e7cbe64fSgw zvol_state_t *zv; 1344e7cbe64fSgw int error = 0; 1345e7cbe64fSgw uint64_t size; 1346e7cbe64fSgw uint64_t boff; 1347e7cbe64fSgw uint64_t resid; 1348e7cbe64fSgw 1349c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 1350e7cbe64fSgw if (zv == NULL) 1351be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 1352e7cbe64fSgw 13533b2aab18SMatthew Ahrens if ((zv->zv_flags & ZVOL_DUMPIFIED) == 0) 1354be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 13553b2aab18SMatthew Ahrens 1356e7cbe64fSgw boff = ldbtob(blkno); 1357e7cbe64fSgw resid = ldbtob(nblocks); 135888b7b0f2SMatthew Ahrens 135988b7b0f2SMatthew Ahrens VERIFY3U(boff + resid, <=, zv->zv_volsize); 136088b7b0f2SMatthew Ahrens 1361e7cbe64fSgw while (resid) { 1362e7cbe64fSgw size = MIN(resid, P2END(boff, zv->zv_volblocksize) - boff); 136388b7b0f2SMatthew Ahrens error = zvol_dumpio(zv, addr, boff, size, B_FALSE, B_TRUE); 1364e7cbe64fSgw if (error) 1365e7cbe64fSgw break; 1366e7cbe64fSgw boff += size; 1367e7cbe64fSgw addr += size; 1368e7cbe64fSgw resid -= size; 1369e7cbe64fSgw } 1370e7cbe64fSgw 1371e7cbe64fSgw return (error); 1372e7cbe64fSgw } 1373e7cbe64fSgw 1374fa9e4066Sahrens /*ARGSUSED*/ 1375fa9e4066Sahrens int 1376feb08c6bSbillm zvol_read(dev_t dev, uio_t *uio, cred_t *cr) 1377fa9e4066Sahrens { 1378c7ca1008Sgw minor_t minor = getminor(dev); 1379c7ca1008Sgw zvol_state_t *zv; 138073ec3d9cSgw uint64_t volsize; 1381feb08c6bSbillm int error = 0; 1382fa9e4066Sahrens 1383c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 1384c7ca1008Sgw if (zv == NULL) 1385be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 1386c7ca1008Sgw 138773ec3d9cSgw volsize = zv->zv_volsize; 138873ec3d9cSgw if (uio->uio_resid > 0 && 138973ec3d9cSgw (uio->uio_loffset < 0 || uio->uio_loffset >= volsize)) 1390be6fd75aSMatthew Ahrens return (SET_ERROR(EIO)); 139173ec3d9cSgw 139288b7b0f2SMatthew Ahrens if (zv->zv_flags & ZVOL_DUMPIFIED) { 139388b7b0f2SMatthew Ahrens error = physio(zvol_strategy, NULL, dev, B_READ, 139488b7b0f2SMatthew Ahrens zvol_minphys, uio); 139588b7b0f2SMatthew Ahrens return (error); 139688b7b0f2SMatthew Ahrens } 139788b7b0f2SMatthew Ahrens 1398*79315247SMatthew Ahrens locked_range_t *lr = rangelock_enter(&zv->zv_rangelock, 1399*79315247SMatthew Ahrens uio->uio_loffset, uio->uio_resid, RL_READER); 140073ec3d9cSgw while (uio->uio_resid > 0 && uio->uio_loffset < volsize) { 1401feb08c6bSbillm uint64_t bytes = MIN(uio->uio_resid, DMU_MAX_ACCESS >> 1); 1402fa9e4066Sahrens 140373ec3d9cSgw /* don't read past the end */ 140473ec3d9cSgw if (bytes > volsize - uio->uio_loffset) 140573ec3d9cSgw bytes = volsize - uio->uio_loffset; 140673ec3d9cSgw 1407feb08c6bSbillm error = dmu_read_uio(zv->zv_objset, ZVOL_OBJ, uio, bytes); 1408b87f3af3Sperrin if (error) { 1409b87f3af3Sperrin /* convert checksum errors into IO errors */ 1410b87f3af3Sperrin if (error == ECKSUM) 1411be6fd75aSMatthew Ahrens error = SET_ERROR(EIO); 1412feb08c6bSbillm break; 1413b87f3af3Sperrin } 1414feb08c6bSbillm } 1415*79315247SMatthew Ahrens rangelock_exit(lr); 1416*79315247SMatthew Ahrens 1417feb08c6bSbillm return (error); 1418fa9e4066Sahrens } 1419fa9e4066Sahrens 1420fa9e4066Sahrens /*ARGSUSED*/ 1421fa9e4066Sahrens int 1422feb08c6bSbillm zvol_write(dev_t dev, uio_t *uio, cred_t *cr) 1423fa9e4066Sahrens { 1424c7ca1008Sgw minor_t minor = getminor(dev); 1425c7ca1008Sgw zvol_state_t *zv; 142673ec3d9cSgw uint64_t volsize; 1427feb08c6bSbillm int error = 0; 1428510b6c0eSNeil Perrin boolean_t sync; 1429feb08c6bSbillm 1430c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 1431c7ca1008Sgw if (zv == NULL) 1432be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 1433c7ca1008Sgw 143473ec3d9cSgw volsize = zv->zv_volsize; 143573ec3d9cSgw if (uio->uio_resid > 0 && 143673ec3d9cSgw (uio->uio_loffset < 0 || uio->uio_loffset >= volsize)) 1437be6fd75aSMatthew Ahrens return (SET_ERROR(EIO)); 143873ec3d9cSgw 1439e7cbe64fSgw if (zv->zv_flags & ZVOL_DUMPIFIED) { 1440e7cbe64fSgw error = physio(zvol_strategy, NULL, dev, B_WRITE, 1441e7cbe64fSgw zvol_minphys, uio); 1442e7cbe64fSgw return (error); 1443e7cbe64fSgw } 1444e7cbe64fSgw 144555da60b9SMark J Musante sync = !(zv->zv_flags & ZVOL_WCE) || 144655da60b9SMark J Musante (zv->zv_objset->os_sync == ZFS_SYNC_ALWAYS); 1447510b6c0eSNeil Perrin 1448*79315247SMatthew Ahrens locked_range_t *lr = rangelock_enter(&zv->zv_rangelock, 1449*79315247SMatthew Ahrens uio->uio_loffset, uio->uio_resid, RL_WRITER); 145073ec3d9cSgw while (uio->uio_resid > 0 && uio->uio_loffset < volsize) { 1451feb08c6bSbillm uint64_t bytes = MIN(uio->uio_resid, DMU_MAX_ACCESS >> 1); 1452feb08c6bSbillm uint64_t off = uio->uio_loffset; 1453feb08c6bSbillm dmu_tx_t *tx = dmu_tx_create(zv->zv_objset); 145473ec3d9cSgw 145573ec3d9cSgw if (bytes > volsize - off) /* don't write past the end */ 145673ec3d9cSgw bytes = volsize - off; 145773ec3d9cSgw 1458feb08c6bSbillm dmu_tx_hold_write(tx, ZVOL_OBJ, off, bytes); 1459feb08c6bSbillm error = dmu_tx_assign(tx, TXG_WAIT); 1460feb08c6bSbillm if (error) { 1461feb08c6bSbillm dmu_tx_abort(tx); 1462feb08c6bSbillm break; 1463feb08c6bSbillm } 14648dfe5547SRichard Yao error = dmu_write_uio_dnode(zv->zv_dn, uio, bytes, tx); 1465feb08c6bSbillm if (error == 0) 1466510b6c0eSNeil Perrin zvol_log_write(zv, tx, off, bytes, sync); 1467feb08c6bSbillm dmu_tx_commit(tx); 1468feb08c6bSbillm 1469feb08c6bSbillm if (error) 1470feb08c6bSbillm break; 1471feb08c6bSbillm } 1472*79315247SMatthew Ahrens rangelock_exit(lr); 1473*79315247SMatthew Ahrens 1474510b6c0eSNeil Perrin if (sync) 14755002558fSNeil Perrin zil_commit(zv->zv_zilog, ZVOL_OBJ); 1476feb08c6bSbillm return (error); 1477fa9e4066Sahrens } 1478fa9e4066Sahrens 1479c7f714e2SEric Taylor int 1480c7f714e2SEric Taylor zvol_getefi(void *arg, int flag, uint64_t vs, uint8_t bs) 1481c7f714e2SEric Taylor { 1482c7f714e2SEric Taylor struct uuid uuid = EFI_RESERVED; 1483c7f714e2SEric Taylor efi_gpe_t gpe = { 0 }; 1484c7f714e2SEric Taylor uint32_t crc; 1485c7f714e2SEric Taylor dk_efi_t efi; 1486c7f714e2SEric Taylor int length; 1487c7f714e2SEric Taylor char *ptr; 1488c7f714e2SEric Taylor 1489c7f714e2SEric Taylor if (ddi_copyin(arg, &efi, sizeof (dk_efi_t), flag)) 1490be6fd75aSMatthew Ahrens return (SET_ERROR(EFAULT)); 1491c7f714e2SEric Taylor ptr = (char *)(uintptr_t)efi.dki_data_64; 1492c7f714e2SEric Taylor length = efi.dki_length; 1493c7f714e2SEric Taylor /* 1494c7f714e2SEric Taylor * Some clients may attempt to request a PMBR for the 1495c7f714e2SEric Taylor * zvol. Currently this interface will return EINVAL to 1496c7f714e2SEric Taylor * such requests. These requests could be supported by 1497c7f714e2SEric Taylor * adding a check for lba == 0 and consing up an appropriate 1498c7f714e2SEric Taylor * PMBR. 1499c7f714e2SEric Taylor */ 1500c7f714e2SEric Taylor if (efi.dki_lba < 1 || efi.dki_lba > 2 || length <= 0) 1501be6fd75aSMatthew Ahrens return (SET_ERROR(EINVAL)); 1502c7f714e2SEric Taylor 1503c7f714e2SEric Taylor gpe.efi_gpe_StartingLBA = LE_64(34ULL); 1504c7f714e2SEric Taylor gpe.efi_gpe_EndingLBA = LE_64((vs >> bs) - 1); 1505c7f714e2SEric Taylor UUID_LE_CONVERT(gpe.efi_gpe_PartitionTypeGUID, uuid); 1506c7f714e2SEric Taylor 1507c7f714e2SEric Taylor if (efi.dki_lba == 1) { 1508c7f714e2SEric Taylor efi_gpt_t gpt = { 0 }; 1509c7f714e2SEric Taylor 1510c7f714e2SEric Taylor gpt.efi_gpt_Signature = LE_64(EFI_SIGNATURE); 1511c7f714e2SEric Taylor gpt.efi_gpt_Revision = LE_32(EFI_VERSION_CURRENT); 1512c7f714e2SEric Taylor gpt.efi_gpt_HeaderSize = LE_32(sizeof (gpt)); 1513c7f714e2SEric Taylor gpt.efi_gpt_MyLBA = LE_64(1ULL); 1514c7f714e2SEric Taylor gpt.efi_gpt_FirstUsableLBA = LE_64(34ULL); 1515c7f714e2SEric Taylor gpt.efi_gpt_LastUsableLBA = LE_64((vs >> bs) - 1); 1516c7f714e2SEric Taylor gpt.efi_gpt_PartitionEntryLBA = LE_64(2ULL); 1517c7f714e2SEric Taylor gpt.efi_gpt_NumberOfPartitionEntries = LE_32(1); 1518c7f714e2SEric Taylor gpt.efi_gpt_SizeOfPartitionEntry = 1519c7f714e2SEric Taylor LE_32(sizeof (efi_gpe_t)); 1520c7f714e2SEric Taylor CRC32(crc, &gpe, sizeof (gpe), -1U, crc32_table); 1521c7f714e2SEric Taylor gpt.efi_gpt_PartitionEntryArrayCRC32 = LE_32(~crc); 1522c7f714e2SEric Taylor CRC32(crc, &gpt, sizeof (gpt), -1U, crc32_table); 1523c7f714e2SEric Taylor gpt.efi_gpt_HeaderCRC32 = LE_32(~crc); 1524c7f714e2SEric Taylor if (ddi_copyout(&gpt, ptr, MIN(sizeof (gpt), length), 1525c7f714e2SEric Taylor flag)) 1526be6fd75aSMatthew Ahrens return (SET_ERROR(EFAULT)); 1527c7f714e2SEric Taylor ptr += sizeof (gpt); 1528c7f714e2SEric Taylor length -= sizeof (gpt); 1529c7f714e2SEric Taylor } 1530c7f714e2SEric Taylor if (length > 0 && ddi_copyout(&gpe, ptr, MIN(sizeof (gpe), 1531c7f714e2SEric Taylor length), flag)) 1532be6fd75aSMatthew Ahrens return (SET_ERROR(EFAULT)); 1533c7f714e2SEric Taylor return (0); 1534c7f714e2SEric Taylor } 1535c7f714e2SEric Taylor 15363fb517f7SJames Moore /* 15373fb517f7SJames Moore * BEGIN entry points to allow external callers access to the volume. 15383fb517f7SJames Moore */ 15393fb517f7SJames Moore /* 15403fb517f7SJames Moore * Return the volume parameters needed for access from an external caller. 15413fb517f7SJames Moore * These values are invariant as long as the volume is held open. 15423fb517f7SJames Moore */ 15433fb517f7SJames Moore int 15443fb517f7SJames Moore zvol_get_volume_params(minor_t minor, uint64_t *blksize, 15453fb517f7SJames Moore uint64_t *max_xfer_len, void **minor_hdl, void **objset_hdl, void **zil_hdl, 15468dfe5547SRichard Yao void **rl_hdl, void **dnode_hdl) 15473fb517f7SJames Moore { 15483fb517f7SJames Moore zvol_state_t *zv; 15493fb517f7SJames Moore 1550c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(minor, ZSST_ZVOL); 1551c99e4bdcSChris Kirby if (zv == NULL) 1552be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 15533fb517f7SJames Moore if (zv->zv_flags & ZVOL_DUMPIFIED) 1554be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 15553fb517f7SJames Moore 15563fb517f7SJames Moore ASSERT(blksize && max_xfer_len && minor_hdl && 15578dfe5547SRichard Yao objset_hdl && zil_hdl && rl_hdl && dnode_hdl); 15583fb517f7SJames Moore 15593fb517f7SJames Moore *blksize = zv->zv_volblocksize; 15603fb517f7SJames Moore *max_xfer_len = (uint64_t)zvol_maxphys; 15613fb517f7SJames Moore *minor_hdl = zv; 15623fb517f7SJames Moore *objset_hdl = zv->zv_objset; 15633fb517f7SJames Moore *zil_hdl = zv->zv_zilog; 1564*79315247SMatthew Ahrens *rl_hdl = &zv->zv_rangelock; 15658dfe5547SRichard Yao *dnode_hdl = zv->zv_dn; 15663fb517f7SJames Moore return (0); 15673fb517f7SJames Moore } 15683fb517f7SJames Moore 15693fb517f7SJames Moore /* 15703fb517f7SJames Moore * Return the current volume size to an external caller. 15713fb517f7SJames Moore * The size can change while the volume is open. 15723fb517f7SJames Moore */ 15733fb517f7SJames Moore uint64_t 15743fb517f7SJames Moore zvol_get_volume_size(void *minor_hdl) 15753fb517f7SJames Moore { 15763fb517f7SJames Moore zvol_state_t *zv = minor_hdl; 15773fb517f7SJames Moore 15783fb517f7SJames Moore return (zv->zv_volsize); 15793fb517f7SJames Moore } 15803fb517f7SJames Moore 15813fb517f7SJames Moore /* 15823fb517f7SJames Moore * Return the current WCE setting to an external caller. 15833fb517f7SJames Moore * The WCE setting can change while the volume is open. 15843fb517f7SJames Moore */ 15853fb517f7SJames Moore int 15863fb517f7SJames Moore zvol_get_volume_wce(void *minor_hdl) 15873fb517f7SJames Moore { 15883fb517f7SJames Moore zvol_state_t *zv = minor_hdl; 15893fb517f7SJames Moore 15903fb517f7SJames Moore return ((zv->zv_flags & ZVOL_WCE) ? 1 : 0); 15913fb517f7SJames Moore } 15923fb517f7SJames Moore 15933fb517f7SJames Moore /* 15943fb517f7SJames Moore * Entry point for external callers to zvol_log_write 15953fb517f7SJames Moore */ 15963fb517f7SJames Moore void 15973fb517f7SJames Moore zvol_log_write_minor(void *minor_hdl, dmu_tx_t *tx, offset_t off, ssize_t resid, 15983fb517f7SJames Moore boolean_t sync) 15993fb517f7SJames Moore { 16003fb517f7SJames Moore zvol_state_t *zv = minor_hdl; 16013fb517f7SJames Moore 16023fb517f7SJames Moore zvol_log_write(zv, tx, off, resid, sync); 16033fb517f7SJames Moore } 16043fb517f7SJames Moore /* 16053fb517f7SJames Moore * END entry points to allow external callers access to the volume. 16063fb517f7SJames Moore */ 16073fb517f7SJames Moore 1608b77b9231SDan McDonald /* 1609b77b9231SDan McDonald * Log a DKIOCFREE/free-long-range to the ZIL with TX_TRUNCATE. 1610b77b9231SDan McDonald */ 1611b77b9231SDan McDonald static void 1612b77b9231SDan McDonald zvol_log_truncate(zvol_state_t *zv, dmu_tx_t *tx, uint64_t off, uint64_t len, 1613b77b9231SDan McDonald boolean_t sync) 1614b77b9231SDan McDonald { 1615b77b9231SDan McDonald itx_t *itx; 1616b77b9231SDan McDonald lr_truncate_t *lr; 1617b77b9231SDan McDonald zilog_t *zilog = zv->zv_zilog; 1618b77b9231SDan McDonald 1619b77b9231SDan McDonald if (zil_replaying(zilog, tx)) 1620b77b9231SDan McDonald return; 1621b77b9231SDan McDonald 1622b77b9231SDan McDonald itx = zil_itx_create(TX_TRUNCATE, sizeof (*lr)); 1623b77b9231SDan McDonald lr = (lr_truncate_t *)&itx->itx_lr; 1624b77b9231SDan McDonald lr->lr_foid = ZVOL_OBJ; 1625b77b9231SDan McDonald lr->lr_offset = off; 1626b77b9231SDan McDonald lr->lr_length = len; 1627b77b9231SDan McDonald 1628b77b9231SDan McDonald itx->itx_sync = sync; 1629b77b9231SDan McDonald zil_itx_assign(zilog, itx, tx); 1630b77b9231SDan McDonald } 1631b77b9231SDan McDonald 1632fa9e4066Sahrens /* 1633fa9e4066Sahrens * Dirtbag ioctls to support mkfs(1M) for UFS filesystems. See dkio(7I). 1634b77b9231SDan McDonald * Also a dirtbag dkio ioctl for unmap/free-block functionality. 1635fa9e4066Sahrens */ 1636fa9e4066Sahrens /*ARGSUSED*/ 1637fa9e4066Sahrens int 1638fa9e4066Sahrens zvol_ioctl(dev_t dev, int cmd, intptr_t arg, int flag, cred_t *cr, int *rvalp) 1639fa9e4066Sahrens { 1640fa9e4066Sahrens zvol_state_t *zv; 1641af2c4821Smaybee struct dk_callback *dkc; 1642fa9e4066Sahrens int error = 0; 1643*79315247SMatthew Ahrens locked_range_t *lr; 1644fa9e4066Sahrens 1645c99e4bdcSChris Kirby mutex_enter(&zfsdev_state_lock); 1646fa9e4066Sahrens 1647c99e4bdcSChris Kirby zv = zfsdev_get_soft_state(getminor(dev), ZSST_ZVOL); 1648fa9e4066Sahrens 1649fa9e4066Sahrens if (zv == NULL) { 1650c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 1651be6fd75aSMatthew Ahrens return (SET_ERROR(ENXIO)); 1652fa9e4066Sahrens } 1653701f66c4SEric Taylor ASSERT(zv->zv_total_opens > 0); 1654fa9e4066Sahrens 1655fa9e4066Sahrens switch (cmd) { 1656fa9e4066Sahrens 1657fa9e4066Sahrens case DKIOCINFO: 1658a0b60564SGeorge Wilson { 1659a0b60564SGeorge Wilson struct dk_cinfo dki; 1660a0b60564SGeorge Wilson 1661af2c4821Smaybee bzero(&dki, sizeof (dki)); 1662af2c4821Smaybee (void) strcpy(dki.dki_cname, "zvol"); 1663af2c4821Smaybee (void) strcpy(dki.dki_dname, "zvol"); 1664af2c4821Smaybee dki.dki_ctype = DKC_UNKNOWN; 16653adc9019SEric Taylor dki.dki_unit = getminor(dev); 1666b5152584SMatthew Ahrens dki.dki_maxtransfer = 1667b5152584SMatthew Ahrens 1 << (SPA_OLD_MAXBLOCKSHIFT - zv->zv_min_bs); 1668c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 1669af2c4821Smaybee if (ddi_copyout(&dki, (void *)arg, sizeof (dki), flag)) 1670be6fd75aSMatthew Ahrens error = SET_ERROR(EFAULT); 1671fa9e4066Sahrens return (error); 1672a0b60564SGeorge Wilson } 1673fa9e4066Sahrens 1674fa9e4066Sahrens case DKIOCGMEDIAINFO: 1675a0b60564SGeorge Wilson { 1676a0b60564SGeorge Wilson struct dk_minfo dkm; 1677a0b60564SGeorge Wilson 1678fa9e4066Sahrens bzero(&dkm, sizeof (dkm)); 1679fa9e4066Sahrens dkm.dki_lbsize = 1U << zv->zv_min_bs; 1680fa9e4066Sahrens dkm.dki_capacity = zv->zv_volsize >> zv->zv_min_bs; 1681fa9e4066Sahrens dkm.dki_media_type = DK_UNKNOWN; 1682c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 1683fa9e4066Sahrens if (ddi_copyout(&dkm, (void *)arg, sizeof (dkm), flag)) 1684be6fd75aSMatthew Ahrens error = SET_ERROR(EFAULT); 1685fa9e4066Sahrens return (error); 1686a0b60564SGeorge Wilson } 1687a0b60564SGeorge Wilson 1688a0b60564SGeorge Wilson case DKIOCGMEDIAINFOEXT: 1689a0b60564SGeorge Wilson { 1690a0b60564SGeorge Wilson struct dk_minfo_ext dkmext; 1691a0b60564SGeorge Wilson 1692a0b60564SGeorge Wilson bzero(&dkmext, sizeof (dkmext)); 1693a0b60564SGeorge Wilson dkmext.dki_lbsize = 1U << zv->zv_min_bs; 1694a0b60564SGeorge Wilson dkmext.dki_pbsize = zv->zv_volblocksize; 1695a0b60564SGeorge Wilson dkmext.dki_capacity = zv->zv_volsize >> zv->zv_min_bs; 1696a0b60564SGeorge Wilson dkmext.dki_media_type = DK_UNKNOWN; 1697a0b60564SGeorge Wilson mutex_exit(&zfsdev_state_lock); 1698a0b60564SGeorge Wilson if (ddi_copyout(&dkmext, (void *)arg, sizeof (dkmext), flag)) 1699a0b60564SGeorge Wilson error = SET_ERROR(EFAULT); 1700a0b60564SGeorge Wilson return (error); 1701a0b60564SGeorge Wilson } 1702fa9e4066Sahrens 1703fa9e4066Sahrens case DKIOCGETEFI: 1704a0b60564SGeorge Wilson { 1705a0b60564SGeorge Wilson uint64_t vs = zv->zv_volsize; 1706a0b60564SGeorge Wilson uint8_t bs = zv->zv_min_bs; 1707fa9e4066Sahrens 1708a0b60564SGeorge Wilson mutex_exit(&zfsdev_state_lock); 1709a0b60564SGeorge Wilson error = zvol_getefi((void *)arg, flag, vs, bs); 1710a0b60564SGeorge Wilson return (error); 1711a0b60564SGeorge Wilson } 1712fa9e4066Sahrens 1713feb08c6bSbillm case DKIOCFLUSHWRITECACHE: 1714af2c4821Smaybee dkc = (struct dk_callback *)arg; 1715c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 17165002558fSNeil Perrin zil_commit(zv->zv_zilog, ZVOL_OBJ); 1717af2c4821Smaybee if ((flag & FKIOCTL) && dkc != NULL && dkc->dkc_callback) { 1718af2c4821Smaybee (*dkc->dkc_callback)(dkc->dkc_cookie, error); 1719af2c4821Smaybee error = 0; 1720af2c4821Smaybee } 1721701f66c4SEric Taylor return (error); 1722701f66c4SEric Taylor 1723701f66c4SEric Taylor case DKIOCGETWCE: 1724a0b60564SGeorge Wilson { 1725a0b60564SGeorge Wilson int wce = (zv->zv_flags & ZVOL_WCE) ? 1 : 0; 1726a0b60564SGeorge Wilson if (ddi_copyout(&wce, (void *)arg, sizeof (int), 1727a0b60564SGeorge Wilson flag)) 1728a0b60564SGeorge Wilson error = SET_ERROR(EFAULT); 1729a0b60564SGeorge Wilson break; 1730a0b60564SGeorge Wilson } 1731a0b60564SGeorge Wilson case DKIOCSETWCE: 1732a0b60564SGeorge Wilson { 1733a0b60564SGeorge Wilson int wce; 1734a0b60564SGeorge Wilson if (ddi_copyin((void *)arg, &wce, sizeof (int), 1735a0b60564SGeorge Wilson flag)) { 1736a0b60564SGeorge Wilson error = SET_ERROR(EFAULT); 1737701f66c4SEric Taylor break; 1738701f66c4SEric Taylor } 1739a0b60564SGeorge Wilson if (wce) { 1740a0b60564SGeorge Wilson zv->zv_flags |= ZVOL_WCE; 1741a0b60564SGeorge Wilson mutex_exit(&zfsdev_state_lock); 1742a0b60564SGeorge Wilson } else { 1743a0b60564SGeorge Wilson zv->zv_flags &= ~ZVOL_WCE; 1744a0b60564SGeorge Wilson mutex_exit(&zfsdev_state_lock); 1745a0b60564SGeorge Wilson zil_commit(zv->zv_zilog, ZVOL_OBJ); 1746701f66c4SEric Taylor } 1747a0b60564SGeorge Wilson return (0); 1748a0b60564SGeorge Wilson } 1749feb08c6bSbillm 1750b6130eadSmaybee case DKIOCGGEOM: 1751b6130eadSmaybee case DKIOCGVTOC: 1752e7cbe64fSgw /* 1753e7cbe64fSgw * commands using these (like prtvtoc) expect ENOTSUP 1754e7cbe64fSgw * since we're emulating an EFI label 1755e7cbe64fSgw */ 1756be6fd75aSMatthew Ahrens error = SET_ERROR(ENOTSUP); 1757b6130eadSmaybee break; 1758b6130eadSmaybee 1759e7cbe64fSgw case DKIOCDUMPINIT: 1760*79315247SMatthew Ahrens lr = rangelock_enter(&zv->zv_rangelock, 0, zv->zv_volsize, 1761e7cbe64fSgw RL_WRITER); 1762e7cbe64fSgw error = zvol_dumpify(zv); 1763*79315247SMatthew Ahrens rangelock_exit(lr); 1764e7cbe64fSgw break; 1765e7cbe64fSgw 1766e7cbe64fSgw case DKIOCDUMPFINI: 176706d5ae10SEric Taylor if (!(zv->zv_flags & ZVOL_DUMPIFIED)) 176806d5ae10SEric Taylor break; 1769*79315247SMatthew Ahrens lr = rangelock_enter(&zv->zv_rangelock, 0, zv->zv_volsize, 1770e7cbe64fSgw RL_WRITER); 1771e7cbe64fSgw error = zvol_dump_fini(zv); 1772*79315247SMatthew Ahrens rangelock_exit(lr); 1773e7cbe64fSgw break; 1774e7cbe64fSgw 1775b77b9231SDan McDonald case DKIOCFREE: 1776b77b9231SDan McDonald { 1777047c81d3SSaso Kiselkov dkioc_free_list_t *dfl; 1778b77b9231SDan McDonald dmu_tx_t *tx; 1779b77b9231SDan McDonald 1780893c83baSGeorge Wilson if (!zvol_unmap_enabled) 1781893c83baSGeorge Wilson break; 1782893c83baSGeorge Wilson 1783047c81d3SSaso Kiselkov if (!(flag & FKIOCTL)) { 1784047c81d3SSaso Kiselkov error = dfl_copyin((void *)arg, &dfl, flag, KM_SLEEP); 1785047c81d3SSaso Kiselkov if (error != 0) 1786047c81d3SSaso Kiselkov break; 1787047c81d3SSaso Kiselkov } else { 1788047c81d3SSaso Kiselkov dfl = (dkioc_free_list_t *)arg; 1789047c81d3SSaso Kiselkov ASSERT3U(dfl->dfl_num_exts, <=, DFL_COPYIN_MAX_EXTS); 1790047c81d3SSaso Kiselkov if (dfl->dfl_num_exts > DFL_COPYIN_MAX_EXTS) { 1791047c81d3SSaso Kiselkov error = SET_ERROR(EINVAL); 1792047c81d3SSaso Kiselkov break; 1793047c81d3SSaso Kiselkov } 1794b77b9231SDan McDonald } 1795b77b9231SDan McDonald 1796574e2414SGeorge Wilson mutex_exit(&zfsdev_state_lock); 1797b77b9231SDan McDonald 1798047c81d3SSaso Kiselkov for (int i = 0; i < dfl->dfl_num_exts; i++) { 1799047c81d3SSaso Kiselkov uint64_t start = dfl->dfl_exts[i].dfle_start, 1800047c81d3SSaso Kiselkov length = dfl->dfl_exts[i].dfle_length, 1801047c81d3SSaso Kiselkov end = start + length; 1802047c81d3SSaso Kiselkov 1803047c81d3SSaso Kiselkov /* 1804047c81d3SSaso Kiselkov * Apply Postel's Law to length-checking. If they 1805047c81d3SSaso Kiselkov * overshoot, just blank out until the end, if there's 1806047c81d3SSaso Kiselkov * a need to blank out anything. 1807047c81d3SSaso Kiselkov */ 1808047c81d3SSaso Kiselkov if (start >= zv->zv_volsize) 1809047c81d3SSaso Kiselkov continue; /* No need to do anything... */ 1810047c81d3SSaso Kiselkov if (end > zv->zv_volsize) { 1811047c81d3SSaso Kiselkov end = DMU_OBJECT_END; 1812047c81d3SSaso Kiselkov length = end - start; 1813047c81d3SSaso Kiselkov } 1814b77b9231SDan McDonald 1815*79315247SMatthew Ahrens lr = rangelock_enter(&zv->zv_rangelock, start, length, 1816047c81d3SSaso Kiselkov RL_WRITER); 1817047c81d3SSaso Kiselkov tx = dmu_tx_create(zv->zv_objset); 1818047c81d3SSaso Kiselkov error = dmu_tx_assign(tx, TXG_WAIT); 1819047c81d3SSaso Kiselkov if (error != 0) { 1820047c81d3SSaso Kiselkov dmu_tx_abort(tx); 1821047c81d3SSaso Kiselkov } else { 1822047c81d3SSaso Kiselkov zvol_log_truncate(zv, tx, start, length, 1823047c81d3SSaso Kiselkov B_TRUE); 1824047c81d3SSaso Kiselkov dmu_tx_commit(tx); 1825047c81d3SSaso Kiselkov error = dmu_free_long_range(zv->zv_objset, 1826047c81d3SSaso Kiselkov ZVOL_OBJ, start, length); 1827047c81d3SSaso Kiselkov } 1828047c81d3SSaso Kiselkov 1829*79315247SMatthew Ahrens rangelock_exit(lr); 1830047c81d3SSaso Kiselkov 1831047c81d3SSaso Kiselkov if (error != 0) 1832047c81d3SSaso Kiselkov break; 1833047c81d3SSaso Kiselkov } 1834b77b9231SDan McDonald 18351c9272b8SStephen Blinick /* 18361c9272b8SStephen Blinick * If the write-cache is disabled, 'sync' property 18371c9272b8SStephen Blinick * is set to 'always', or if the caller is asking for 18381c9272b8SStephen Blinick * a synchronous free, commit this operation to the zil. 18391c9272b8SStephen Blinick * This will sync any previous uncommitted writes to the 18401c9272b8SStephen Blinick * zvol object. 18411c9272b8SStephen Blinick * Can be overridden by the zvol_unmap_sync_enabled tunable. 18421c9272b8SStephen Blinick */ 18431c9272b8SStephen Blinick if ((error == 0) && zvol_unmap_sync_enabled && 18441c9272b8SStephen Blinick (!(zv->zv_flags & ZVOL_WCE) || 18451c9272b8SStephen Blinick (zv->zv_objset->os_sync == ZFS_SYNC_ALWAYS) || 1846047c81d3SSaso Kiselkov (dfl->dfl_flags & DF_WAIT_SYNC))) { 18471c9272b8SStephen Blinick zil_commit(zv->zv_zilog, ZVOL_OBJ); 1848b77b9231SDan McDonald } 18491c9272b8SStephen Blinick 1850047c81d3SSaso Kiselkov if (!(flag & FKIOCTL)) 1851047c81d3SSaso Kiselkov dfl_free(dfl); 1852047c81d3SSaso Kiselkov 1853574e2414SGeorge Wilson return (error); 1854b77b9231SDan McDonald } 1855b77b9231SDan McDonald 1856fa9e4066Sahrens default: 1857be6fd75aSMatthew Ahrens error = SET_ERROR(ENOTTY); 1858fa9e4066Sahrens break; 1859fa9e4066Sahrens 1860fa9e4066Sahrens } 1861c99e4bdcSChris Kirby mutex_exit(&zfsdev_state_lock); 1862fa9e4066Sahrens return (error); 1863fa9e4066Sahrens } 1864fa9e4066Sahrens 1865fa9e4066Sahrens int 1866fa9e4066Sahrens zvol_busy(void) 1867fa9e4066Sahrens { 1868fa9e4066Sahrens return (zvol_minors != 0); 1869fa9e4066Sahrens } 1870fa9e4066Sahrens 1871fa9e4066Sahrens void 1872fa9e4066Sahrens zvol_init(void) 1873fa9e4066Sahrens { 1874c99e4bdcSChris Kirby VERIFY(ddi_soft_state_init(&zfsdev_state, sizeof (zfs_soft_state_t), 1875c99e4bdcSChris Kirby 1) == 0); 1876c99e4bdcSChris Kirby mutex_init(&zfsdev_state_lock, NULL, MUTEX_DEFAULT, NULL); 1877fa9e4066Sahrens } 1878fa9e4066Sahrens 1879fa9e4066Sahrens void 1880fa9e4066Sahrens zvol_fini(void) 1881fa9e4066Sahrens { 1882c99e4bdcSChris Kirby mutex_destroy(&zfsdev_state_lock); 1883c99e4bdcSChris Kirby ddi_soft_state_fini(&zfsdev_state); 1884fa9e4066Sahrens } 1885e7cbe64fSgw 1886810e43b2SBill Pijewski /*ARGSUSED*/ 1887810e43b2SBill Pijewski static int 1888810e43b2SBill Pijewski zfs_mvdev_dump_feature_check(void *arg, dmu_tx_t *tx) 1889810e43b2SBill Pijewski { 1890810e43b2SBill Pijewski spa_t *spa = dmu_tx_pool(tx)->dp_spa; 1891810e43b2SBill Pijewski 18922acef22dSMatthew Ahrens if (spa_feature_is_active(spa, SPA_FEATURE_MULTI_VDEV_CRASH_DUMP)) 1893810e43b2SBill Pijewski return (1); 1894810e43b2SBill Pijewski return (0); 1895810e43b2SBill Pijewski } 1896810e43b2SBill Pijewski 1897810e43b2SBill Pijewski /*ARGSUSED*/ 1898810e43b2SBill Pijewski static void 1899810e43b2SBill Pijewski zfs_mvdev_dump_activate_feature_sync(void *arg, dmu_tx_t *tx) 1900810e43b2SBill Pijewski { 1901810e43b2SBill Pijewski spa_t *spa = dmu_tx_pool(tx)->dp_spa; 1902810e43b2SBill Pijewski 19032acef22dSMatthew Ahrens spa_feature_incr(spa, SPA_FEATURE_MULTI_VDEV_CRASH_DUMP, tx); 1904810e43b2SBill Pijewski } 1905810e43b2SBill Pijewski 1906e7cbe64fSgw static int 1907e7cbe64fSgw zvol_dump_init(zvol_state_t *zv, boolean_t resize) 1908e7cbe64fSgw { 1909e7cbe64fSgw dmu_tx_t *tx; 1910810e43b2SBill Pijewski int error; 1911e7cbe64fSgw objset_t *os = zv->zv_objset; 1912810e43b2SBill Pijewski spa_t *spa = dmu_objset_spa(os); 1913810e43b2SBill Pijewski vdev_t *vd = spa->spa_root_vdev; 1914e7cbe64fSgw nvlist_t *nv = NULL; 1915810e43b2SBill Pijewski uint64_t version = spa_version(spa); 1916b10bba72SGeorge Wilson uint64_t checksum, compress, refresrv, vbs, dedup; 1917e7cbe64fSgw 1918c99e4bdcSChris Kirby ASSERT(MUTEX_HELD(&zfsdev_state_lock)); 1919810e43b2SBill Pijewski ASSERT(vd->vdev_ops == &vdev_root_ops); 1920810e43b2SBill Pijewski 1921681d9761SEric Taylor error = dmu_free_long_range(zv->zv_objset, ZVOL_OBJ, 0, 1922681d9761SEric Taylor DMU_OBJECT_END); 1923b10bba72SGeorge Wilson if (error != 0) 1924b10bba72SGeorge Wilson return (error); 1925681d9761SEric Taylor /* wait for dmu_free_long_range to actually free the blocks */ 1926681d9761SEric Taylor txg_wait_synced(dmu_objset_pool(zv->zv_objset), 0); 1927e7cbe64fSgw 1928810e43b2SBill Pijewski /* 1929810e43b2SBill Pijewski * If the pool on which the dump device is being initialized has more 1930810e43b2SBill Pijewski * than one child vdev, check that the MULTI_VDEV_CRASH_DUMP feature is 1931810e43b2SBill Pijewski * enabled. If so, bump that feature's counter to indicate that the 1932810e43b2SBill Pijewski * feature is active. We also check the vdev type to handle the 1933810e43b2SBill Pijewski * following case: 1934810e43b2SBill Pijewski * # zpool create test raidz disk1 disk2 disk3 1935810e43b2SBill Pijewski * Now have spa_root_vdev->vdev_children == 1 (the raidz vdev), 1936810e43b2SBill Pijewski * the raidz vdev itself has 3 children. 1937810e43b2SBill Pijewski */ 1938810e43b2SBill Pijewski if (vd->vdev_children > 1 || vd->vdev_ops == &vdev_raidz_ops) { 1939810e43b2SBill Pijewski if (!spa_feature_is_enabled(spa, 19402acef22dSMatthew Ahrens SPA_FEATURE_MULTI_VDEV_CRASH_DUMP)) 1941810e43b2SBill Pijewski return (SET_ERROR(ENOTSUP)); 1942810e43b2SBill Pijewski (void) dsl_sync_task(spa_name(spa), 1943810e43b2SBill Pijewski zfs_mvdev_dump_feature_check, 19447d46dc6cSMatthew Ahrens zfs_mvdev_dump_activate_feature_sync, NULL, 19457d46dc6cSMatthew Ahrens 2, ZFS_SPACE_CHECK_RESERVED); 1946810e43b2SBill Pijewski } 1947810e43b2SBill Pijewski 1948b10bba72SGeorge Wilson if (!resize) { 1949b10bba72SGeorge Wilson error = dsl_prop_get_integer(zv->zv_name, 1950b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_COMPRESSION), &compress, NULL); 1951b10bba72SGeorge Wilson if (error == 0) { 1952b10bba72SGeorge Wilson error = dsl_prop_get_integer(zv->zv_name, 1953b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_CHECKSUM), &checksum, 1954b10bba72SGeorge Wilson NULL); 1955b10bba72SGeorge Wilson } 1956b10bba72SGeorge Wilson if (error == 0) { 1957b10bba72SGeorge Wilson error = dsl_prop_get_integer(zv->zv_name, 1958b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_REFRESERVATION), 1959b10bba72SGeorge Wilson &refresrv, NULL); 1960b10bba72SGeorge Wilson } 1961b10bba72SGeorge Wilson if (error == 0) { 1962b10bba72SGeorge Wilson error = dsl_prop_get_integer(zv->zv_name, 1963b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE), &vbs, 1964b10bba72SGeorge Wilson NULL); 1965b10bba72SGeorge Wilson } 1966b10bba72SGeorge Wilson if (version >= SPA_VERSION_DEDUP && error == 0) { 1967b10bba72SGeorge Wilson error = dsl_prop_get_integer(zv->zv_name, 1968b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_DEDUP), &dedup, NULL); 1969b10bba72SGeorge Wilson } 1970b10bba72SGeorge Wilson } 1971b10bba72SGeorge Wilson if (error != 0) 1972b10bba72SGeorge Wilson return (error); 1973b10bba72SGeorge Wilson 1974e7cbe64fSgw tx = dmu_tx_create(os); 1975e7cbe64fSgw dmu_tx_hold_zap(tx, ZVOL_ZAP_OBJ, TRUE, NULL); 1976681d9761SEric Taylor dmu_tx_hold_bonus(tx, ZVOL_OBJ); 1977e7cbe64fSgw error = dmu_tx_assign(tx, TXG_WAIT); 1978b10bba72SGeorge Wilson if (error != 0) { 1979e7cbe64fSgw dmu_tx_abort(tx); 1980e7cbe64fSgw return (error); 1981e7cbe64fSgw } 1982e7cbe64fSgw 1983e7cbe64fSgw /* 1984e7cbe64fSgw * If we are resizing the dump device then we only need to 1985e7cbe64fSgw * update the refreservation to match the newly updated 1986e7cbe64fSgw * zvolsize. Otherwise, we save off the original state of the 1987e7cbe64fSgw * zvol so that we can restore them if the zvol is ever undumpified. 1988e7cbe64fSgw */ 1989e7cbe64fSgw if (resize) { 1990e7cbe64fSgw error = zap_update(os, ZVOL_ZAP_OBJ, 1991e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_REFRESERVATION), 8, 1, 1992e7cbe64fSgw &zv->zv_volsize, tx); 1993e7cbe64fSgw } else { 1994b10bba72SGeorge Wilson error = zap_update(os, ZVOL_ZAP_OBJ, 1995e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_COMPRESSION), 8, 1, 1996e7cbe64fSgw &compress, tx); 1997b10bba72SGeorge Wilson if (error == 0) { 1998b10bba72SGeorge Wilson error = zap_update(os, ZVOL_ZAP_OBJ, 1999b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_CHECKSUM), 8, 1, 2000b10bba72SGeorge Wilson &checksum, tx); 2001b10bba72SGeorge Wilson } 2002b10bba72SGeorge Wilson if (error == 0) { 2003b10bba72SGeorge Wilson error = zap_update(os, ZVOL_ZAP_OBJ, 2004b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_REFRESERVATION), 8, 1, 2005b10bba72SGeorge Wilson &refresrv, tx); 2006b10bba72SGeorge Wilson } 2007b10bba72SGeorge Wilson if (error == 0) { 2008b10bba72SGeorge Wilson error = zap_update(os, ZVOL_ZAP_OBJ, 2009b10bba72SGeorge Wilson zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE), 8, 1, 2010b10bba72SGeorge Wilson &vbs, tx); 2011b10bba72SGeorge Wilson } 2012b10bba72SGeorge Wilson if (error == 0) { 2013b10bba72SGeorge Wilson error = dmu_object_set_blocksize( 2014b10bba72SGeorge Wilson os, ZVOL_OBJ, SPA_OLD_MAXBLOCKSIZE, 0, tx); 2015b10bba72SGeorge Wilson } 2016b10bba72SGeorge Wilson if (version >= SPA_VERSION_DEDUP && error == 0) { 2017b10bba72SGeorge Wilson error = zap_update(os, ZVOL_ZAP_OBJ, 20188d265e66SGeorge Wilson zfs_prop_to_name(ZFS_PROP_DEDUP), 8, 1, 20198d265e66SGeorge Wilson &dedup, tx); 20208d265e66SGeorge Wilson } 2021681d9761SEric Taylor if (error == 0) 2022b5152584SMatthew Ahrens zv->zv_volblocksize = SPA_OLD_MAXBLOCKSIZE; 2023e7cbe64fSgw } 2024e7cbe64fSgw dmu_tx_commit(tx); 2025e7cbe64fSgw 2026e7cbe64fSgw /* 2027e7cbe64fSgw * We only need update the zvol's property if we are initializing 2028e7cbe64fSgw * the dump area for the first time. 2029e7cbe64fSgw */ 2030b10bba72SGeorge Wilson if (error == 0 && !resize) { 2031b10bba72SGeorge Wilson /* 2032b10bba72SGeorge Wilson * If MULTI_VDEV_CRASH_DUMP is active, use the NOPARITY checksum 2033b10bba72SGeorge Wilson * function. Otherwise, use the old default -- OFF. 2034b10bba72SGeorge Wilson */ 2035b10bba72SGeorge Wilson checksum = spa_feature_is_active(spa, 2036b10bba72SGeorge Wilson SPA_FEATURE_MULTI_VDEV_CRASH_DUMP) ? ZIO_CHECKSUM_NOPARITY : 2037b10bba72SGeorge Wilson ZIO_CHECKSUM_OFF; 2038b10bba72SGeorge Wilson 2039e7cbe64fSgw VERIFY(nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_SLEEP) == 0); 2040e7cbe64fSgw VERIFY(nvlist_add_uint64(nv, 2041e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_REFRESERVATION), 0) == 0); 2042e7cbe64fSgw VERIFY(nvlist_add_uint64(nv, 2043e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_COMPRESSION), 2044e7cbe64fSgw ZIO_COMPRESS_OFF) == 0); 2045e7cbe64fSgw VERIFY(nvlist_add_uint64(nv, 2046e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_CHECKSUM), 2047810e43b2SBill Pijewski checksum) == 0); 20488d265e66SGeorge Wilson if (version >= SPA_VERSION_DEDUP) { 20498d265e66SGeorge Wilson VERIFY(nvlist_add_uint64(nv, 20508d265e66SGeorge Wilson zfs_prop_to_name(ZFS_PROP_DEDUP), 20518d265e66SGeorge Wilson ZIO_CHECKSUM_OFF) == 0); 20528d265e66SGeorge Wilson } 2053e7cbe64fSgw 205492241e0bSTom Erickson error = zfs_set_prop_nvlist(zv->zv_name, ZPROP_SRC_LOCAL, 205592241e0bSTom Erickson nv, NULL); 2056e7cbe64fSgw nvlist_free(nv); 2057e7cbe64fSgw } 2058e7cbe64fSgw 2059e7cbe64fSgw /* Allocate the space for the dump */ 2060b10bba72SGeorge Wilson if (error == 0) 2061b10bba72SGeorge Wilson error = zvol_prealloc(zv); 2062e7cbe64fSgw return (error); 2063e7cbe64fSgw } 2064e7cbe64fSgw 2065e7cbe64fSgw static int 2066e7cbe64fSgw zvol_dumpify(zvol_state_t *zv) 2067e7cbe64fSgw { 2068e7cbe64fSgw int error = 0; 2069e7cbe64fSgw uint64_t dumpsize = 0; 2070e7cbe64fSgw dmu_tx_t *tx; 2071e7cbe64fSgw objset_t *os = zv->zv_objset; 2072e7cbe64fSgw 2073681d9761SEric Taylor if (zv->zv_flags & ZVOL_RDONLY) 2074be6fd75aSMatthew Ahrens return (SET_ERROR(EROFS)); 2075e7cbe64fSgw 2076e7cbe64fSgw if (zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, ZVOL_DUMPSIZE, 2077e7cbe64fSgw 8, 1, &dumpsize) != 0 || dumpsize != zv->zv_volsize) { 20784445fffbSMatthew Ahrens boolean_t resize = (dumpsize > 0); 2079e7cbe64fSgw 2080e7cbe64fSgw if ((error = zvol_dump_init(zv, resize)) != 0) { 2081e7cbe64fSgw (void) zvol_dump_fini(zv); 2082e7cbe64fSgw return (error); 2083e7cbe64fSgw } 2084e7cbe64fSgw } 2085e7cbe64fSgw 2086e7cbe64fSgw /* 2087e7cbe64fSgw * Build up our lba mapping. 2088e7cbe64fSgw */ 2089e7cbe64fSgw error = zvol_get_lbas(zv); 2090e7cbe64fSgw if (error) { 2091e7cbe64fSgw (void) zvol_dump_fini(zv); 2092e7cbe64fSgw return (error); 2093e7cbe64fSgw } 2094e7cbe64fSgw 2095e7cbe64fSgw tx = dmu_tx_create(os); 2096e7cbe64fSgw dmu_tx_hold_zap(tx, ZVOL_ZAP_OBJ, TRUE, NULL); 2097e7cbe64fSgw error = dmu_tx_assign(tx, TXG_WAIT); 2098e7cbe64fSgw if (error) { 2099e7cbe64fSgw dmu_tx_abort(tx); 2100e7cbe64fSgw (void) zvol_dump_fini(zv); 2101e7cbe64fSgw return (error); 2102e7cbe64fSgw } 2103e7cbe64fSgw 2104e7cbe64fSgw zv->zv_flags |= ZVOL_DUMPIFIED; 2105e7cbe64fSgw error = zap_update(os, ZVOL_ZAP_OBJ, ZVOL_DUMPSIZE, 8, 1, 2106e7cbe64fSgw &zv->zv_volsize, tx); 2107e7cbe64fSgw dmu_tx_commit(tx); 2108e7cbe64fSgw 2109e7cbe64fSgw if (error) { 2110e7cbe64fSgw (void) zvol_dump_fini(zv); 2111e7cbe64fSgw return (error); 2112e7cbe64fSgw } 2113e7cbe64fSgw 2114e7cbe64fSgw txg_wait_synced(dmu_objset_pool(os), 0); 2115e7cbe64fSgw return (0); 2116e7cbe64fSgw } 2117e7cbe64fSgw 2118e7cbe64fSgw static int 2119e7cbe64fSgw zvol_dump_fini(zvol_state_t *zv) 2120e7cbe64fSgw { 2121e7cbe64fSgw dmu_tx_t *tx; 2122e7cbe64fSgw objset_t *os = zv->zv_objset; 2123e7cbe64fSgw nvlist_t *nv; 2124e7cbe64fSgw int error = 0; 2125afee20e4SGeorge Wilson uint64_t checksum, compress, refresrv, vbs, dedup; 21268d265e66SGeorge Wilson uint64_t version = spa_version(dmu_objset_spa(zv->zv_objset)); 2127e7cbe64fSgw 2128b7e50089Smaybee /* 2129b7e50089Smaybee * Attempt to restore the zvol back to its pre-dumpified state. 2130b7e50089Smaybee * This is a best-effort attempt as it's possible that not all 2131b7e50089Smaybee * of these properties were initialized during the dumpify process 2132b7e50089Smaybee * (i.e. error during zvol_dump_init). 2133b7e50089Smaybee */ 2134b7e50089Smaybee 2135e7cbe64fSgw tx = dmu_tx_create(os); 2136e7cbe64fSgw dmu_tx_hold_zap(tx, ZVOL_ZAP_OBJ, TRUE, NULL); 2137e7cbe64fSgw error = dmu_tx_assign(tx, TXG_WAIT); 2138e7cbe64fSgw if (error) { 2139e7cbe64fSgw dmu_tx_abort(tx); 2140e7cbe64fSgw return (error); 2141e7cbe64fSgw } 2142b7e50089Smaybee (void) zap_remove(os, ZVOL_ZAP_OBJ, ZVOL_DUMPSIZE, tx); 2143b7e50089Smaybee dmu_tx_commit(tx); 2144e7cbe64fSgw 2145e7cbe64fSgw (void) zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, 2146e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_CHECKSUM), 8, 1, &checksum); 2147e7cbe64fSgw (void) zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, 2148e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_COMPRESSION), 8, 1, &compress); 2149e7cbe64fSgw (void) zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, 2150e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_REFRESERVATION), 8, 1, &refresrv); 215188b7b0f2SMatthew Ahrens (void) zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, 215288b7b0f2SMatthew Ahrens zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE), 8, 1, &vbs); 2153e7cbe64fSgw 2154e7cbe64fSgw VERIFY(nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_SLEEP) == 0); 2155e7cbe64fSgw (void) nvlist_add_uint64(nv, 2156e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_CHECKSUM), checksum); 2157e7cbe64fSgw (void) nvlist_add_uint64(nv, 2158e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_COMPRESSION), compress); 2159e7cbe64fSgw (void) nvlist_add_uint64(nv, 2160e7cbe64fSgw zfs_prop_to_name(ZFS_PROP_REFRESERVATION), refresrv); 21618d265e66SGeorge Wilson if (version >= SPA_VERSION_DEDUP && 21628d265e66SGeorge Wilson zap_lookup(zv->zv_objset, ZVOL_ZAP_OBJ, 21638d265e66SGeorge Wilson zfs_prop_to_name(ZFS_PROP_DEDUP), 8, 1, &dedup) == 0) { 21648d265e66SGeorge Wilson (void) nvlist_add_uint64(nv, 21658d265e66SGeorge Wilson zfs_prop_to_name(ZFS_PROP_DEDUP), dedup); 21668d265e66SGeorge Wilson } 216792241e0bSTom Erickson (void) zfs_set_prop_nvlist(zv->zv_name, ZPROP_SRC_LOCAL, 216892241e0bSTom Erickson nv, NULL); 2169e7cbe64fSgw nvlist_free(nv); 2170e7cbe64fSgw 2171b7e50089Smaybee zvol_free_extents(zv); 2172b7e50089Smaybee zv->zv_flags &= ~ZVOL_DUMPIFIED; 2173b7e50089Smaybee (void) dmu_free_long_range(os, ZVOL_OBJ, 0, DMU_OBJECT_END); 2174681d9761SEric Taylor /* wait for dmu_free_long_range to actually free the blocks */ 2175681d9761SEric Taylor txg_wait_synced(dmu_objset_pool(zv->zv_objset), 0); 2176681d9761SEric Taylor tx = dmu_tx_create(os); 2177681d9761SEric Taylor dmu_tx_hold_bonus(tx, ZVOL_OBJ); 2178681d9761SEric Taylor error = dmu_tx_assign(tx, TXG_WAIT); 2179681d9761SEric Taylor if (error) { 2180681d9761SEric Taylor dmu_tx_abort(tx); 2181681d9761SEric Taylor return (error); 2182681d9761SEric Taylor } 2183b24ab676SJeff Bonwick if (dmu_object_set_blocksize(os, ZVOL_OBJ, vbs, 0, tx) == 0) 2184b24ab676SJeff Bonwick zv->zv_volblocksize = vbs; 2185681d9761SEric Taylor dmu_tx_commit(tx); 2186b7e50089Smaybee 2187e7cbe64fSgw return (0); 2188e7cbe64fSgw } 2189