Lines Matching refs:size

205 hat_memload_alloc(vmem_t *vmp, size_t size, int flags)  in hat_memload_alloc()  argument
208 return (segkmem_alloc(vmp, size, flags)); in hat_memload_alloc()
218 segkmem_alloc_permanent(vmem_t *vmp, size_t size, int flags) in segkmem_alloc_permanent() argument
220 return (segkmem_alloc(vmp, size, flags | VM_NORELOC)); in segkmem_alloc_permanent()
367 boot_mapin(caddr_t addr, size_t size) in boot_mapin() argument
373 if (page_resv(btop(size), KM_NOSLEEP) == 0) in boot_mapin()
376 for (eaddr = addr + size; addr < eaddr; addr += PAGESIZE) { in boot_mapin()
420 boot_alloc(void *inaddr, size_t size, uint_t align) in boot_alloc() argument
428 size = ptob(btopr(size)); in boot_alloc()
430 if (bop_alloc_chunk(addr, size, align) != (caddr_t)addr) in boot_alloc()
433 if (BOP_ALLOC(bootops, addr, size, align) != addr) in boot_alloc()
436 boot_mapin((caddr_t)addr, size); in boot_alloc()
450 segkmem_fault(struct hat *hat, struct seg *seg, caddr_t addr, size_t size, in segkmem_fault() argument
460 if (seg->s_as != &kas || size > seg->s_size || in segkmem_fault()
461 addr < seg->s_base || addr + size > seg->s_base + seg->s_size) in segkmem_fault()
469 return (SEGOP_FAULT(hat, segkp, addr, size, type, rw)); in segkmem_fault()
474 npages = btopr(size); in segkmem_fault()
501 hat_reserve(seg->s_as, addr, size); in segkmem_fault()
518 segkmem_setprot(struct seg *seg, caddr_t addr, size_t size, uint_t prot) in segkmem_setprot() argument
522 if (seg->s_as != &kas || size > seg->s_size || in segkmem_setprot()
523 addr < seg->s_base || addr + size > seg->s_base + seg->s_size) in segkmem_setprot()
531 return (SEGOP_SETPROT(segkp, addr, size, prot)); in segkmem_setprot()
534 hat_unload(kas.a_hat, addr, size, HAT_UNLOAD); in segkmem_setprot()
536 hat_chgprot(kas.a_hat, addr, size, prot); in segkmem_setprot()
546 segkmem_checkprot(struct seg *seg, caddr_t addr, size_t size, uint_t prot) in segkmem_checkprot() argument
558 return (SEGOP_CHECKPROT(segkp, addr, size, prot)); in segkmem_checkprot()
589 segkmem_xdump_range(void *arg, void *start, size_t size) in segkmem_xdump_range() argument
593 caddr_t addr_end = addr + size; in segkmem_xdump_range()
605 segkmem_dump_range(void *arg, void *start, size_t size) in segkmem_dump_range() argument
608 caddr_t addr_end = addr + size; in segkmem_dump_range()
617 vmem_size(heap_lp_arena, VMEM_ALLOC) < size) { in segkmem_dump_range()
621 segkmem_xdump_range(arg, start, size); in segkmem_dump_range()
837 segkmem_page_create(void *addr, size_t size, int vmflag, void *arg) in segkmem_page_create() argument
861 return (page_create_va(vp, (u_offset_t)(uintptr_t)addr, size, in segkmem_page_create()
870 segkmem_xalloc(vmem_t *vmp, void *inaddr, size_t size, int vmflag, uint_t attr, in segkmem_xalloc() argument
875 pgcnt_t npages = btopr(size); in segkmem_xalloc()
878 if (inaddr == NULL && (addr = vmem_alloc(vmp, size, vmflag)) == NULL) in segkmem_xalloc()
885 vmem_free(vmp, addr, size); in segkmem_xalloc()
889 ppl = page_create_func(addr, size, vmflag, pcarg); in segkmem_xalloc()
892 vmem_free(vmp, addr, size); in segkmem_xalloc()
940 segkmem_alloc_vn(vmem_t *vmp, size_t size, int vmflag, struct vnode *vp) in segkmem_alloc_vn() argument
960 if (gcp->gc_arena == vmp && gcp->gc_size == size) { in segkmem_alloc_vn()
966 addr = vmem_alloc(vmp, size, vmflag | VM_PANIC); in segkmem_alloc_vn()
967 if (boot_alloc(addr, size, BO_NO_ALIGN) != addr) in segkmem_alloc_vn()
971 return (segkmem_xalloc(vmp, NULL, size, vmflag, 0, in segkmem_alloc_vn()
976 segkmem_alloc(vmem_t *vmp, size_t size, int vmflag) in segkmem_alloc() argument
978 return (segkmem_alloc_vn(vmp, size, vmflag, &kvp)); in segkmem_alloc()
982 segkmem_zio_alloc(vmem_t *vmp, size_t size, int vmflag) in segkmem_zio_alloc() argument
984 return (segkmem_alloc_vn(vmp, size, vmflag, &kvps[KV_ZVP])); in segkmem_zio_alloc()
994 segkmem_xfree(vmem_t *vmp, void *inaddr, size_t size, struct vnode *vp, in segkmem_xfree() argument
1000 pgcnt_t npages = btopr(size); in segkmem_xfree()
1008 gc->gc_size = size; in segkmem_xfree()
1014 hat_unload(kas.a_hat, addr, size, HAT_UNLOAD_UNLOCK); in segkmem_xfree()
1016 for (eaddr = addr + size; addr < eaddr; addr += PAGESIZE) { in segkmem_xfree()
1046 vmem_free(vmp, inaddr, size); in segkmem_xfree()
1051 segkmem_free(vmem_t *vmp, void *inaddr, size_t size) in segkmem_free() argument
1053 segkmem_xfree(vmp, inaddr, size, &kvp, NULL); in segkmem_free()
1057 segkmem_zio_free(vmem_t *vmp, void *inaddr, size_t size) in segkmem_zio_free() argument
1059 segkmem_xfree(vmp, inaddr, size, &kvps[KV_ZVP], NULL); in segkmem_zio_free()
1077 segkmem_mapin(struct seg *seg, void *addr, size_t size, uint_t vprot, in segkmem_mapin() argument
1080 hat_unload(seg->s_as->a_hat, addr, size, HAT_UNLOAD_UNLOCK); in segkmem_mapin()
1081 hat_devload(seg->s_as->a_hat, addr, size, pfn, vprot, in segkmem_mapin()
1086 segkmem_mapout(struct seg *seg, void *addr, size_t size) in segkmem_mapout() argument
1088 hat_unload(seg->s_as->a_hat, addr, size, HAT_UNLOAD_UNLOCK); in segkmem_mapout()
1110 segkmem_page_create_large(void *addr, size_t size, int vmflag, void *arg) in segkmem_page_create_large() argument
1125 return (page_create_va_large(&kvp, (u_offset_t)(uintptr_t)addr, size, in segkmem_page_create_large()
1135 segkmem_xalloc_lp(vmem_t *vmp, void *inaddr, size_t size, int vmflag, in segkmem_xalloc_lp() argument
1141 pgcnt_t npages = btopr(size); in segkmem_xalloc_lp()
1143 pgcnt_t nlpages = size >> segkmem_lpshift; in segkmem_xalloc_lp()
1163 if (inaddr == NULL && (addr = vmem_alloc(vmp, size, vmflag)) == NULL) in segkmem_xalloc_lp()
1215 vmem_free(vmp, addr, size); in segkmem_xalloc_lp()
1227 segkmem_free_one_lp(caddr_t addr, size_t size) in segkmem_free_one_lp() argument
1230 pgcnt_t pgs_left = btopr(size); in segkmem_free_one_lp()
1232 ASSERT(size == segkmem_lpsize); in segkmem_free_one_lp()
1234 hat_unload(kas.a_hat, addr, size, HAT_UNLOAD_UNLOCK); in segkmem_free_one_lp()
1264 size_t size; in segkmem_alloc_lp() local
1270 size = *sizep; in segkmem_alloc_lp()
1276 size_t asize = P2ROUNDUP(size, kmemlp_qnt); in segkmem_alloc_lp()
1284 ASSERT(asize >= size); in segkmem_alloc_lp()
1304 return (segkmem_alloc(vmp, size, vmflag)); in segkmem_alloc_lp()
1375 lpcb->alloc_bytes_failed += size; in segkmem_alloc_lp()
1382 return (segkmem_alloc(vmp, size, vmflag)); in segkmem_alloc_lp()
1386 segkmem_free_lp(vmem_t *vmp, void *inaddr, size_t size) in segkmem_free_lp() argument
1389 segkmem_free(vmp, inaddr, size); in segkmem_free_lp()
1391 vmem_free(kmem_lp_arena, inaddr, size); in segkmem_free_lp()
1402 segkmem_alloc_lpi(vmem_t *vmp, size_t size, int vmflag) in segkmem_alloc_lpi() argument
1407 ASSERT(size != 0); in segkmem_alloc_lpi()
1416 addr = segkmem_xalloc_lp(vmp, NULL, size, vmflag, 0, in segkmem_alloc_lpi()
1427 segkmem_free_lpi(vmem_t *vmp, void *inaddr, size_t size) in segkmem_free_lpi() argument
1429 pgcnt_t nlpages = size >> segkmem_lpshift; in segkmem_free_lpi()
1432 pgcnt_t npages = btopr(size); in segkmem_free_lpi()
1446 vmem_free(vmp, inaddr, size); in segkmem_free_lpi()
1548 segkmem_kvmm_init(void *base, size_t size) in segkmem_kvmm_init() argument
1551 ASSERT(size != 0); in segkmem_kvmm_init()
1553 kvmm_arena = vmem_create("kvmm_arena", base, size, 1024 * 1024, in segkmem_kvmm_init()
1562 segkmem_alloc_ppa(vmem_t *vmp, size_t size, int vmflag) in segkmem_alloc_ppa() argument
1568 return (segkmem_alloc(vmp, size, vmflag)); in segkmem_alloc_ppa()
1570 ASSERT((size & (ppaquantum - 1)) == 0); in segkmem_alloc_ppa()
1572 addr = vmem_xalloc(vmp, size, ppaquantum, 0, 0, NULL, NULL, vmflag); in segkmem_alloc_ppa()
1573 if (addr != NULL && segkmem_xalloc(vmp, addr, size, vmflag, 0, in segkmem_alloc_ppa()
1575 vmem_xfree(vmp, addr, size); in segkmem_alloc_ppa()
1583 segkmem_free_ppa(vmem_t *vmp, void *addr, size_t size) in segkmem_free_ppa() argument
1590 segkmem_free(vmp, addr, size); in segkmem_free_ppa()
1592 segkmem_free(NULL, addr, size); in segkmem_free_ppa()
1593 vmem_xfree(vmp, addr, size); in segkmem_free_ppa()