/illumos-gate/usr/src/boot/sys/cddl/boot/zfs/ |
H A D | zfsimpl.h | 134 #define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT) macro 485 #define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
|
/illumos-gate/usr/src/cmd/mdb/common/modules/zfs/ |
H A D | zfs.c | 4337 "[n*%llu, (n+1)*%llu)", SPA_MINBLOCKSIZE, in arc_compression_stats() 4338 SPA_MINBLOCKSIZE); in arc_compression_stats() 4341 "[2^(n-1)*%llu, 2^n*%llu)", SPA_MINBLOCKSIZE, in arc_compression_stats() 4342 SPA_MINBLOCKSIZE); in arc_compression_stats()
|
/illumos-gate/usr/src/cmd/zdb/ |
H A D | zdb.c | 3413 #define PSIZE_HISTO_SIZE (SPA_OLD_MAXBLOCKSIZE / SPA_MINBLOCKSIZE + 2) 3501 idx = MIN(idx, SPA_OLD_MAXBLOCKSIZE / SPA_MINBLOCKSIZE + 1); in zdb_count_block() 6161 lsize -= SPA_MINBLOCKSIZE) { in zdb_read_block() 6172 lsize -= SPA_MINBLOCKSIZE; in zdb_read_block()
|
/illumos-gate/usr/src/cmd/ztest/ |
H A D | ztest.c | 759 MAX(SPA_MINBLOCKSIZE << 1, value); in process_options() 4562 if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) { in ztest_dmu_read_write_zcopy() 4587 chunksize < (SPA_MINBLOCKSIZE * 2)) { in ztest_dmu_read_write_zcopy() 4631 if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) { in ztest_dmu_read_write_zcopy() 4648 if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) { in ztest_dmu_read_write_zcopy()
|
/illumos-gate/usr/src/common/zfs/ |
H A D | zfs_fletcher.c | 599 if (size < SPA_MINBLOCKSIZE) in fletcher_4_incremental_native() 612 if (size < SPA_MINBLOCKSIZE) in fletcher_4_incremental_byteswap()
|
/illumos-gate/usr/src/grub/grub-0.97/stage2/zfs-include/ |
H A D | spa.h | 65 #define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT) macro
|
H A D | zio.h | 42 #define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
|
/illumos-gate/usr/src/lib/libzfs/common/ |
H A D | libzfs_dataset.c | 1178 if (intval < SPA_MINBLOCKSIZE || in zfs_valid_proplist() 1208 (intval < SPA_MINBLOCKSIZE || in zfs_valid_proplist()
|
/illumos-gate/usr/src/uts/common/fs/zfs/ |
H A D | dbuf.c | 2550 BP_GET_LSIZE(blkptr) : SPA_MINBLOCKSIZE; in dbuf_create() 3056 blksz = SPA_MINBLOCKSIZE; in dbuf_spill_set_blksz() 3058 blksz = P2ROUNDUP(blksz, SPA_MINBLOCKSIZE); in dbuf_spill_set_blksz()
|
H A D | dmu.c | 2539 *nblk512 = ((DN_USED_BYTES(dn->dn_phys) + SPA_MINBLOCKSIZE/2) >> in dmu_object_size_from_db()
|
H A D | dmu_recv.c | 1128 P2PHASE(drro->drr_blksz, SPA_MINBLOCKSIZE) || in receive_object() 1129 drro->drr_blksz < SPA_MINBLOCKSIZE || in receive_object() 1698 if (drrs->drr_length < SPA_MINBLOCKSIZE || in receive_spill()
|
H A D | dnode.c | 423 ASSERT0(P2PHASE(size, SPA_MINBLOCKSIZE)); in dnode_setdblksz() 425 ASSERT3U(size, >=, SPA_MINBLOCKSIZE); in dnode_setdblksz() 582 blocksize = P2ROUNDUP(blocksize, SPA_MINBLOCKSIZE); in dnode_allocate() 669 ASSERT3U(blocksize, >=, SPA_MINBLOCKSIZE); in dnode_reallocate() 672 ASSERT0(blocksize % SPA_MINBLOCKSIZE); in dnode_reallocate() 1744 size = SPA_MINBLOCKSIZE; in dnode_set_blksz() 1746 size = P2ROUNDUP(size, SPA_MINBLOCKSIZE); in dnode_set_blksz()
|
H A D | dnode_sync.c | 694 SPA_MINBLOCKSIZE) == 0); in dnode_sync()
|
H A D | dsl_crypt.c | 1982 if (ret != 0 || blksz < SPA_MINBLOCKSIZE) in dsl_crypto_recv_raw_objset_check()
|
H A D | metaslab.c | 1120 ASSERT(weight >= SPA_MINBLOCKSIZE || weight == 0); in metaslab_group_sort() 1266 } else if (mc->mc_alloc_groups == 0 || psize == SPA_MINBLOCKSIZE) { in metaslab_group_allocatable() 2889 if (space > 0 && space < SPA_MINBLOCKSIZE) in metaslab_space_weight() 2890 space = SPA_MINBLOCKSIZE; in metaslab_space_weight() 3328 ASSERT(size >= SPA_MINBLOCKSIZE || in metaslab_passivate()
|
H A D | sa.c | 500 blocksize = SPA_MINBLOCKSIZE; in sa_resize_spill() 505 blocksize = P2ROUNDUP_TYPED(size, SPA_MINBLOCKSIZE, uint32_t); in sa_resize_spill()
|
H A D | spa.c | 8635 if (vd->vdev_deflate_ratio != SPA_MINBLOCKSIZE)
|
H A D | space_map.c | 257 uint64_t bufsz = MAX(sm->sm_blksz, SPA_MINBLOCKSIZE); in space_map_incremental_destroy()
|
H A D | vdev.c | 4063 ASSERT((space & (SPA_MINBLOCKSIZE-1)) == 0); in vdev_deflated_space()
|
H A D | vdev_disk.c | 708 *ashift = highbit64(MAX(pbsize, SPA_MINBLOCKSIZE)) - 1; in vdev_disk_open()
|
H A D | vdev_indirect.c | 1124 IS_P2ALIGNED(inner_size, 2 * SPA_MINBLOCKSIZE)) { in vdev_indirect_remap()
|
/illumos-gate/usr/src/uts/common/fs/zfs/sys/ |
H A D | arc_impl.h | 245 CTASSERT(sizeof (l2arc_dev_hdr_phys_t) == SPA_MINBLOCKSIZE); 302 CTASSERT(sizeof (l2arc_log_blk_phys_t) >= SPA_MINBLOCKSIZE);
|
H A D | spa.h | 87 #define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT) macro
|
H A D | zfs_fuid.h | 54 #define FUID_SIZE_ESTIMATE(z) ((z)->z_fuid_size + (SPA_MINBLOCKSIZE << 1))
|
H A D | zio.h | 60 #define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
|