Lines Matching refs:dn

169 dmu_buf_hold_noread_by_dnode(dnode_t *dn, uint64_t offset,  in dmu_buf_hold_noread_by_dnode()  argument
175 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_buf_hold_noread_by_dnode()
176 blkid = dbuf_whichblock(dn, 0, offset); in dmu_buf_hold_noread_by_dnode()
177 db = dbuf_hold(dn, blkid, tag); in dmu_buf_hold_noread_by_dnode()
178 rw_exit(&dn->dn_struct_rwlock); in dmu_buf_hold_noread_by_dnode()
192 dnode_t *dn; in dmu_buf_hold_noread() local
197 err = dnode_hold(os, object, FTAG, &dn); in dmu_buf_hold_noread()
200 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_buf_hold_noread()
201 blkid = dbuf_whichblock(dn, 0, offset); in dmu_buf_hold_noread()
202 db = dbuf_hold(dn, blkid, tag); in dmu_buf_hold_noread()
203 rw_exit(&dn->dn_struct_rwlock); in dmu_buf_hold_noread()
204 dnode_rele(dn, FTAG); in dmu_buf_hold_noread()
216 dmu_buf_hold_by_dnode(dnode_t *dn, uint64_t offset, in dmu_buf_hold_by_dnode() argument
227 err = dmu_buf_hold_noread_by_dnode(dn, offset, tag, dbp); in dmu_buf_hold_by_dnode()
275 dnode_t *dn; in dmu_set_bonus() local
279 dn = DB_DNODE(db); in dmu_set_bonus()
281 if (dn->dn_bonus != db) { in dmu_set_bonus()
286 dnode_setbonuslen(dn, newsize, tx); in dmu_set_bonus()
298 dnode_t *dn; in dmu_set_bonustype() local
302 dn = DB_DNODE(db); in dmu_set_bonustype()
306 } else if (dn->dn_bonus != db) { in dmu_set_bonustype()
309 dnode_setbonus_type(dn, type, tx); in dmu_set_bonustype()
321 dnode_t *dn; in dmu_get_bonustype() local
325 dn = DB_DNODE(db); in dmu_get_bonustype()
326 type = dn->dn_bonustype; in dmu_get_bonustype()
335 dnode_t *dn; in dmu_rm_spill() local
338 error = dnode_hold(os, object, FTAG, &dn); in dmu_rm_spill()
339 dbuf_rm_spill(dn, tx); in dmu_rm_spill()
340 rw_enter(&dn->dn_struct_rwlock, RW_WRITER); in dmu_rm_spill()
341 dnode_rm_spill(dn, tx); in dmu_rm_spill()
342 rw_exit(&dn->dn_struct_rwlock); in dmu_rm_spill()
343 dnode_rele(dn, FTAG); in dmu_rm_spill()
352 int dmu_bonus_hold_by_dnode(dnode_t *dn, void *tag, dmu_buf_t **dbp, in dmu_bonus_hold_by_dnode() argument
364 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_bonus_hold_by_dnode()
365 if (dn->dn_bonus == NULL) { in dmu_bonus_hold_by_dnode()
366 rw_exit(&dn->dn_struct_rwlock); in dmu_bonus_hold_by_dnode()
367 rw_enter(&dn->dn_struct_rwlock, RW_WRITER); in dmu_bonus_hold_by_dnode()
368 if (dn->dn_bonus == NULL) in dmu_bonus_hold_by_dnode()
369 dbuf_create_bonus(dn); in dmu_bonus_hold_by_dnode()
371 db = dn->dn_bonus; in dmu_bonus_hold_by_dnode()
375 VERIFY(dnode_add_ref(dn, db)); in dmu_bonus_hold_by_dnode()
376 atomic_inc_32(&dn->dn_dbufs_count); in dmu_bonus_hold_by_dnode()
384 rw_exit(&dn->dn_struct_rwlock); in dmu_bonus_hold_by_dnode()
388 dnode_evict_bonus(dn); in dmu_bonus_hold_by_dnode()
405 dnode_t *dn; in dmu_bonus_hold_impl() local
415 error = dnode_hold(os, object, FTAG, &dn); in dmu_bonus_hold_impl()
419 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_bonus_hold_impl()
420 if (dn->dn_bonus == NULL) { in dmu_bonus_hold_impl()
421 rw_exit(&dn->dn_struct_rwlock); in dmu_bonus_hold_impl()
422 rw_enter(&dn->dn_struct_rwlock, RW_WRITER); in dmu_bonus_hold_impl()
423 if (dn->dn_bonus == NULL) in dmu_bonus_hold_impl()
424 dbuf_create_bonus(dn); in dmu_bonus_hold_impl()
426 db = dn->dn_bonus; in dmu_bonus_hold_impl()
430 VERIFY(dnode_add_ref(dn, db)); in dmu_bonus_hold_impl()
431 atomic_inc_32(&dn->dn_dbufs_count); in dmu_bonus_hold_impl()
439 rw_exit(&dn->dn_struct_rwlock); in dmu_bonus_hold_impl()
441 dnode_rele(dn, FTAG); in dmu_bonus_hold_impl()
445 dnode_evict_bonus(dn); in dmu_bonus_hold_impl()
471 dmu_spill_hold_by_dnode(dnode_t *dn, uint32_t flags, void *tag, dmu_buf_t **dbp) in dmu_spill_hold_by_dnode() argument
477 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_spill_hold_by_dnode()
479 db = dbuf_hold(dn, DMU_SPILL_BLKID, tag); in dmu_spill_hold_by_dnode()
482 rw_exit(&dn->dn_struct_rwlock); in dmu_spill_hold_by_dnode()
497 dnode_t *dn; in dmu_spill_hold_existing() local
501 dn = DB_DNODE(db); in dmu_spill_hold_existing()
503 if (spa_version(dn->dn_objset->os_spa) < SPA_VERSION_SA) { in dmu_spill_hold_existing()
506 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_spill_hold_existing()
508 if (!dn->dn_have_spill) { in dmu_spill_hold_existing()
511 err = dmu_spill_hold_by_dnode(dn, in dmu_spill_hold_existing()
515 rw_exit(&dn->dn_struct_rwlock); in dmu_spill_hold_existing()
527 dnode_t *dn; in dmu_spill_hold_by_bonus() local
535 dn = DB_DNODE(db); in dmu_spill_hold_by_bonus()
536 err = dmu_spill_hold_by_dnode(dn, db_flags, tag, dbp); in dmu_spill_hold_by_bonus()
549 dmu_buf_hold_array_by_dnode(dnode_t *dn, uint64_t offset, uint64_t length, in dmu_buf_hold_array_by_dnode() argument
568 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_buf_hold_array_by_dnode()
569 if (dn->dn_datablkshift) { in dmu_buf_hold_array_by_dnode()
570 int blkshift = dn->dn_datablkshift; in dmu_buf_hold_array_by_dnode()
574 if (offset + length > dn->dn_datablksz) { in dmu_buf_hold_array_by_dnode()
577 (longlong_t)dn->dn_objset-> in dmu_buf_hold_array_by_dnode()
579 (longlong_t)dn->dn_object, dn->dn_datablksz, in dmu_buf_hold_array_by_dnode()
581 rw_exit(&dn->dn_struct_rwlock); in dmu_buf_hold_array_by_dnode()
588 zio = zio_root(dn->dn_objset->os_spa, NULL, NULL, ZIO_FLAG_CANFAIL); in dmu_buf_hold_array_by_dnode()
589 blkid = dbuf_whichblock(dn, 0, offset); in dmu_buf_hold_array_by_dnode()
591 dmu_buf_impl_t *db = dbuf_hold(dn, blkid + i, tag); in dmu_buf_hold_array_by_dnode()
593 rw_exit(&dn->dn_struct_rwlock); in dmu_buf_hold_array_by_dnode()
606 DNODE_META_IS_CACHEABLE(dn) && length <= zfetch_array_rd_sz) { in dmu_buf_hold_array_by_dnode()
607 dmu_zfetch(&dn->dn_zfetch, blkid, nblks, in dmu_buf_hold_array_by_dnode()
608 read && DNODE_IS_CACHEABLE(dn), B_TRUE); in dmu_buf_hold_array_by_dnode()
610 rw_exit(&dn->dn_struct_rwlock); in dmu_buf_hold_array_by_dnode()
646 dnode_t *dn; in dmu_buf_hold_array() local
649 err = dnode_hold(os, object, FTAG, &dn); in dmu_buf_hold_array()
653 err = dmu_buf_hold_array_by_dnode(dn, offset, length, read, tag, in dmu_buf_hold_array()
656 dnode_rele(dn, FTAG); in dmu_buf_hold_array()
667 dnode_t *dn; in dmu_buf_hold_array_by_bonus() local
671 dn = DB_DNODE(db); in dmu_buf_hold_array_by_bonus()
672 err = dmu_buf_hold_array_by_dnode(dn, offset, length, read, tag, in dmu_buf_hold_array_by_bonus()
708 dnode_t *dn; in dmu_prefetch() local
713 dn = DMU_META_DNODE(os); in dmu_prefetch()
718 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_prefetch()
719 blkid = dbuf_whichblock(dn, level, in dmu_prefetch()
721 dbuf_prefetch(dn, level, blkid, pri, 0); in dmu_prefetch()
722 rw_exit(&dn->dn_struct_rwlock); in dmu_prefetch()
736 err = dnode_hold(os, object, FTAG, &dn); in dmu_prefetch()
747 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_prefetch()
748 if (level > 0 || dn->dn_datablkshift != 0) { in dmu_prefetch()
749 nblks = dbuf_whichblock(dn, level, offset + len - 1) - in dmu_prefetch()
750 dbuf_whichblock(dn, level, offset) + 1; in dmu_prefetch()
752 nblks = (offset < dn->dn_datablksz); in dmu_prefetch()
756 blkid = dbuf_whichblock(dn, level, offset); in dmu_prefetch()
758 dbuf_prefetch(dn, level, blkid + i, pri, 0); in dmu_prefetch()
760 rw_exit(&dn->dn_struct_rwlock); in dmu_prefetch()
762 dnode_rele(dn, FTAG); in dmu_prefetch()
776 get_next_chunk(dnode_t *dn, uint64_t *start, uint64_t minimum) in get_next_chunk() argument
778 uint64_t maxblks = DMU_MAX_ACCESS >> (dn->dn_indblkshift + 1); in get_next_chunk()
781 dn->dn_datablksz * EPB(dn->dn_indblkshift, SPA_BLKPTRSHIFT); in get_next_chunk()
801 err = dnode_next_offset(dn, in get_next_chunk()
837 dmu_free_long_range_impl(objset_t *os, dnode_t *dn, uint64_t offset, in dmu_free_long_range_impl() argument
840 uint64_t object_size = (dn->dn_maxblkid + 1) * dn->dn_datablksz; in dmu_free_long_range_impl()
862 if (dmu_objset_zfs_unmounting(dn->dn_objset)) in dmu_free_long_range_impl()
868 err = get_next_chunk(dn, &chunk_begin, offset); in dmu_free_long_range_impl()
895 dmu_tx_hold_free(tx, dn->dn_object, chunk_begin, chunk_len); in dmu_free_long_range_impl()
915 dnode_free_range(dn, chunk_begin, chunk_len, tx); in dmu_free_long_range_impl()
928 dnode_t *dn; in dmu_free_long_range() local
931 err = dnode_hold(os, object, FTAG, &dn); in dmu_free_long_range()
934 err = dmu_free_long_range_impl(os, dn, offset, length); in dmu_free_long_range()
943 dn->dn_maxblkid = 0; in dmu_free_long_range()
945 dnode_rele(dn, FTAG); in dmu_free_long_range()
980 dnode_t *dn; in dmu_free_range() local
981 int err = dnode_hold(os, object, FTAG, &dn); in dmu_free_range()
986 dnode_free_range(dn, offset, size, tx); in dmu_free_range()
987 dnode_rele(dn, FTAG); in dmu_free_range()
992 dmu_read_impl(dnode_t *dn, uint64_t offset, uint64_t size, in dmu_read_impl() argument
1003 if (dn->dn_maxblkid == 0) { in dmu_read_impl()
1004 int newsz = offset > dn->dn_datablksz ? 0 : in dmu_read_impl()
1005 MIN(size, dn->dn_datablksz - offset); in dmu_read_impl()
1018 err = dmu_buf_hold_array_by_dnode(dn, offset, mylen, in dmu_read_impl()
1048 dnode_t *dn; in dmu_read() local
1051 err = dnode_hold(os, object, FTAG, &dn); in dmu_read()
1055 err = dmu_read_impl(dn, offset, size, buf, flags); in dmu_read()
1056 dnode_rele(dn, FTAG); in dmu_read()
1061 dmu_read_by_dnode(dnode_t *dn, uint64_t offset, uint64_t size, void *buf, in dmu_read_by_dnode() argument
1064 return (dmu_read_impl(dn, offset, size, buf, flags)); in dmu_read_by_dnode()
1118 dmu_write_by_dnode(dnode_t *dn, uint64_t offset, uint64_t size, in dmu_write_by_dnode() argument
1127 VERIFY0(dmu_buf_hold_array_by_dnode(dn, offset, size, in dmu_write_by_dnode()
1134 dmu_object_remap_one_indirect(objset_t *os, dnode_t *dn, in dmu_object_remap_one_indirect() argument
1137 uint64_t l1blkid = dbuf_whichblock(dn, 1, offset); in dmu_object_remap_one_indirect()
1140 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_object_remap_one_indirect()
1141 dmu_buf_impl_t *dbuf = dbuf_hold_level(dn, 1, l1blkid, FTAG); in dmu_object_remap_one_indirect()
1152 rw_exit(&dn->dn_struct_rwlock); in dmu_object_remap_one_indirect()
1162 dmu_tx_hold_remap_l1indirect(tx, dn->dn_object); in dmu_object_remap_one_indirect()
1196 dnode_t *dn; in dmu_object_remap_indirects() local
1198 err = dnode_hold(os, object, FTAG, &dn); in dmu_object_remap_indirects()
1203 if (dn->dn_nlevels <= 1) { in dmu_object_remap_indirects()
1213 if (err == 0 && dnode_needs_remap(dn)) { in dmu_object_remap_indirects()
1215 dmu_tx_hold_bonus(tx, dn->dn_object); in dmu_object_remap_indirects()
1217 dnode_setdirty(dn, tx); in dmu_object_remap_indirects()
1224 dnode_rele(dn, FTAG); in dmu_object_remap_indirects()
1229 l1span = 1ULL << (dn->dn_indblkshift - SPA_BLKPTRSHIFT + in dmu_object_remap_indirects()
1230 dn->dn_datablkshift); in dmu_object_remap_indirects()
1234 while (dnode_next_offset(dn, 0, &offset, 2, 1, 0) == 0) { in dmu_object_remap_indirects()
1239 if ((err = dmu_object_remap_one_indirect(os, dn, in dmu_object_remap_indirects()
1246 dnode_rele(dn, FTAG); in dmu_object_remap_indirects()
1415 dmu_read_uio_dnode(dnode_t *dn, uio_t *uio, uint64_t size) in dmu_read_uio_dnode() argument
1425 err = dmu_buf_hold_array_by_dnode(dn, uio->uio_loffset, size, in dmu_read_uio_dnode()
1484 dnode_t *dn; in dmu_read_uio_dbuf() local
1491 dn = DB_DNODE(db); in dmu_read_uio_dbuf()
1492 err = dmu_read_uio_dnode(dn, uio, size); in dmu_read_uio_dbuf()
1506 dnode_t *dn; in dmu_read_uio() local
1512 err = dnode_hold(os, object, FTAG, &dn); in dmu_read_uio()
1516 err = dmu_read_uio_dnode(dn, uio, size); in dmu_read_uio()
1518 dnode_rele(dn, FTAG); in dmu_read_uio()
1524 dmu_write_uio_dnode(dnode_t *dn, uio_t *uio, uint64_t size, dmu_tx_t *tx) in dmu_write_uio_dnode() argument
1531 err = dmu_buf_hold_array_by_dnode(dn, uio->uio_loffset, size, in dmu_write_uio_dnode()
1589 dnode_t *dn; in dmu_write_uio_dbuf() local
1596 dn = DB_DNODE(db); in dmu_write_uio_dbuf()
1597 err = dmu_write_uio_dnode(dn, uio, size, tx); in dmu_write_uio_dbuf()
1612 dnode_t *dn; in dmu_write_uio() local
1618 err = dnode_hold(os, object, FTAG, &dn); in dmu_write_uio()
1622 err = dmu_write_uio_dnode(dn, uio, size, tx); in dmu_write_uio()
1624 dnode_rele(dn, FTAG); in dmu_write_uio()
1762 dmu_assign_arcbuf_by_dnode(dnode_t *dn, uint64_t offset, arc_buf_t *buf, in dmu_assign_arcbuf_by_dnode() argument
1766 objset_t *os = dn->dn_objset; in dmu_assign_arcbuf_by_dnode()
1767 uint64_t object = dn->dn_object; in dmu_assign_arcbuf_by_dnode()
1771 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_assign_arcbuf_by_dnode()
1772 blkid = dbuf_whichblock(dn, 0, offset); in dmu_assign_arcbuf_by_dnode()
1773 db = dbuf_hold(dn, blkid, FTAG); in dmu_assign_arcbuf_by_dnode()
1776 rw_exit(&dn->dn_struct_rwlock); in dmu_assign_arcbuf_by_dnode()
1790 os = dn->dn_objset; in dmu_assign_arcbuf_by_dnode()
1791 object = dn->dn_object; in dmu_assign_arcbuf_by_dnode()
2037 dnode_t *dn; in dmu_sync() local
2046 dn = DB_DNODE(db); in dmu_sync()
2047 dmu_write_policy(os, dn, db->db_level, WP_DMU_SYNC, &zp); in dmu_sync()
2134 dn = DB_DNODE(db); in dmu_sync()
2135 if (dr->dr_next != NULL || dnode_block_freed(dn, db->db_blkid)) in dmu_sync()
2172 dnode_t *dn; in dmu_object_set_nlevels() local
2175 err = dnode_hold(os, object, FTAG, &dn); in dmu_object_set_nlevels()
2178 err = dnode_set_nlevels(dn, nlevels, tx); in dmu_object_set_nlevels()
2179 dnode_rele(dn, FTAG); in dmu_object_set_nlevels()
2187 dnode_t *dn; in dmu_object_set_blocksize() local
2190 err = dnode_hold(os, object, FTAG, &dn); in dmu_object_set_blocksize()
2193 err = dnode_set_blksz(dn, size, ibs, tx); in dmu_object_set_blocksize()
2194 dnode_rele(dn, FTAG); in dmu_object_set_blocksize()
2202 dnode_t *dn; in dmu_object_set_maxblkid() local
2205 err = dnode_hold(os, object, FTAG, &dn); in dmu_object_set_maxblkid()
2208 rw_enter(&dn->dn_struct_rwlock, RW_WRITER); in dmu_object_set_maxblkid()
2209 dnode_new_blkid(dn, maxblkid, tx, B_FALSE, B_TRUE); in dmu_object_set_maxblkid()
2210 rw_exit(&dn->dn_struct_rwlock); in dmu_object_set_maxblkid()
2211 dnode_rele(dn, FTAG); in dmu_object_set_maxblkid()
2219 dnode_t *dn; in dmu_object_set_checksum() local
2228 VERIFY0(dnode_hold(os, object, FTAG, &dn)); in dmu_object_set_checksum()
2230 dn->dn_checksum = checksum; in dmu_object_set_checksum()
2231 dnode_setdirty(dn, tx); in dmu_object_set_checksum()
2232 dnode_rele(dn, FTAG); in dmu_object_set_checksum()
2239 dnode_t *dn; in dmu_object_set_compress() local
2248 VERIFY0(dnode_hold(os, object, FTAG, &dn)); in dmu_object_set_compress()
2249 dn->dn_compress = compress; in dmu_object_set_compress()
2250 dnode_setdirty(dn, tx); in dmu_object_set_compress()
2251 dnode_rele(dn, FTAG); in dmu_object_set_compress()
2261 dmu_write_policy(objset_t *os, dnode_t *dn, int level, int wp, zio_prop_t *zp) in dmu_write_policy() argument
2263 dmu_object_type_t type = dn ? dn->dn_type : DMU_OT_OBJSET; in dmu_write_policy()
2322 compress = zio_compress_select(os->os_spa, dn->dn_compress, in dmu_write_policy()
2326 zio_checksum_select(dn->dn_checksum, checksum) : in dmu_write_policy()
2383 zp->zp_type = (wp & WP_SPILL) ? dn->dn_bonustype : type; in dmu_write_policy()
2401 dnode_t *dn; in dmu_offset_next() local
2413 err = dnode_hold(os, object, FTAG, &dn); in dmu_offset_next()
2418 err = dnode_next_offset(dn, (hole ? DNODE_FIND_HOLE : 0), off, 1, 1, 0); in dmu_offset_next()
2419 dnode_rele(dn, FTAG); in dmu_offset_next()
2434 dnode_t *dn; in dmu_object_wait_synced() local
2437 error = dnode_hold(os, object, FTAG, &dn); in dmu_object_wait_synced()
2442 mutex_enter(&dn->dn_mtx); in dmu_object_wait_synced()
2444 if (list_link_active(&dn->dn_dirty_link[i]) || in dmu_object_wait_synced()
2445 !list_is_empty(&dn->dn_dirty_records[i])) { in dmu_object_wait_synced()
2449 mutex_exit(&dn->dn_mtx); in dmu_object_wait_synced()
2451 dnode_rele(dn, FTAG); in dmu_object_wait_synced()
2460 dmu_object_info_from_dnode(dnode_t *dn, dmu_object_info_t *doi) in dmu_object_info_from_dnode() argument
2464 rw_enter(&dn->dn_struct_rwlock, RW_READER); in dmu_object_info_from_dnode()
2465 mutex_enter(&dn->dn_mtx); in dmu_object_info_from_dnode()
2467 dnp = dn->dn_phys; in dmu_object_info_from_dnode()
2469 doi->doi_data_block_size = dn->dn_datablksz; in dmu_object_info_from_dnode()
2470 doi->doi_metadata_block_size = dn->dn_indblkshift ? in dmu_object_info_from_dnode()
2471 1ULL << dn->dn_indblkshift : 0; in dmu_object_info_from_dnode()
2472 doi->doi_type = dn->dn_type; in dmu_object_info_from_dnode()
2473 doi->doi_bonus_type = dn->dn_bonustype; in dmu_object_info_from_dnode()
2474 doi->doi_bonus_size = dn->dn_bonuslen; in dmu_object_info_from_dnode()
2475 doi->doi_dnodesize = dn->dn_num_slots << DNODE_SHIFT; in dmu_object_info_from_dnode()
2476 doi->doi_indirection = dn->dn_nlevels; in dmu_object_info_from_dnode()
2477 doi->doi_checksum = dn->dn_checksum; in dmu_object_info_from_dnode()
2478 doi->doi_compress = dn->dn_compress; in dmu_object_info_from_dnode()
2479 doi->doi_nblkptr = dn->dn_nblkptr; in dmu_object_info_from_dnode()
2481 doi->doi_max_offset = (dn->dn_maxblkid + 1) * dn->dn_datablksz; in dmu_object_info_from_dnode()
2486 mutex_exit(&dn->dn_mtx); in dmu_object_info_from_dnode()
2487 rw_exit(&dn->dn_struct_rwlock); in dmu_object_info_from_dnode()
2497 dnode_t *dn; in dmu_object_info() local
2498 int err = dnode_hold(os, object, FTAG, &dn); in dmu_object_info()
2504 dmu_object_info_from_dnode(dn, doi); in dmu_object_info()
2506 dnode_rele(dn, FTAG); in dmu_object_info()
2532 dnode_t *dn; in dmu_object_size_from_db() local
2535 dn = DB_DNODE(db); in dmu_object_size_from_db()
2537 *blksize = dn->dn_datablksz; in dmu_object_size_from_db()
2539 *nblk512 = ((DN_USED_BYTES(dn->dn_phys) + SPA_MINBLOCKSIZE/2) >> in dmu_object_size_from_db()
2540 SPA_MINBLOCKSHIFT) + dn->dn_num_slots; in dmu_object_size_from_db()
2548 dnode_t *dn; in dmu_object_dnsize_from_db() local
2551 dn = DB_DNODE(db); in dmu_object_dnsize_from_db()
2552 *dnsize = dn->dn_num_slots << DNODE_SHIFT; in dmu_object_dnsize_from_db()