/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2008 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* * A quick way to generate a cache consistent address to map in a page. * users: ppcopy, pagezero, /proc, dev/mem * * The ppmapin/ppmapout routines provide a quick way of generating a cache * consistent address by reserving a given amount of kernel address space. * The base is PPMAPBASE and its size is PPMAPSIZE. This memory is divided * into x number of sets, where x is the number of colors for the virtual * cache. The number of colors is how many times a page can be mapped * simulatenously in the cache. For direct map caches this translates to * the number of pages in the cache. * Each set will be assigned a group of virtual pages from the reserved memory * depending on its virtual color. * When trying to assign a virtual address we will find out the color for the * physical page in question (if applicable). Then we will try to find an * available virtual page from the set of the appropiate color. */ #define clsettoarray(color, set) ((color * nsets) + set) int pp_slots = 4; /* small default, tuned by cpu module */ /* tuned by cpu module, default is "safe" */ int pp_consistent_coloring = PPAGE_STORES_POLLUTE | PPAGE_LOADS_POLLUTE; static caddr_t ppmap_vaddrs[PPMAPSIZE / MMU_PAGESIZE]; static int nsets; /* number of sets */ static int ppmap_pages; /* generate align mask */ static int ppmap_shift; /* set selector */ #ifdef PPDEBUG #define MAXCOLORS 16 /* for debug only */ static int ppalloc_noslot = 0; /* # of allocations from kernelmap */ static int align_hits[MAXCOLORS]; static int pp_allocs; /* # of ppmapin requests */ #endif /* PPDEBUG */ /* * There are only 64 TLB entries on spitfire, 16 on cheetah * (fully-associative TLB) so we allow the cpu module to tune the * number to use here via pp_slots. */ static struct ppmap_va { caddr_t ppmap_slots[MAXPP_SLOTS]; } ppmap_va[NCPU]; void ppmapinit(void) { int color, nset, setsize; caddr_t va; ASSERT(pp_slots <= MAXPP_SLOTS); va = (caddr_t)PPMAPBASE; if (cache & CACHE_VAC) { int a; ppmap_pages = mmu_btop(shm_alignment); nsets = PPMAPSIZE / shm_alignment; setsize = shm_alignment; ppmap_shift = MMU_PAGESHIFT; a = ppmap_pages; while (a >>= 1) ppmap_shift++; } else { /* * If we do not have a virtual indexed cache we simply * have only one set containing all pages. */ ppmap_pages = 1; nsets = mmu_btop(PPMAPSIZE); setsize = MMU_PAGESIZE; ppmap_shift = MMU_PAGESHIFT; } for (color = 0; color < ppmap_pages; color++) { for (nset = 0; nset < nsets; nset++) { ppmap_vaddrs[clsettoarray(color, nset)] = (caddr_t)((uintptr_t)va + (nset * setsize)); } va += MMU_PAGESIZE; } } /* * Allocate a cache consistent virtual address to map a page, pp, * with protection, vprot; and map it in the MMU, using the most * efficient means possible. The argument avoid is a virtual address * hint which when masked yields an offset into a virtual cache * that should be avoided when allocating an address to map in a * page. An avoid arg of -1 means you don't care, for instance pagezero. * * machine dependent, depends on virtual address space layout, * understands that all kernel addresses have bit 31 set. * * NOTE: For sun4 platforms the meaning of the hint argument is opposite from * that found in other architectures. In other architectures the hint * (called avoid) was used to ask ppmapin to NOT use the specified cache color. * This was used to avoid virtual cache trashing in the bcopy. Unfortunately * in the case of a COW, this later on caused a cache aliasing conflict. In * sun4, the bcopy routine uses the block ld/st instructions so we don't have * to worry about virtual cache trashing. Actually, by using the hint to choose * the right color we can almost guarantee a cache conflict will not occur. */ caddr_t ppmapin(page_t *pp, uint_t vprot, caddr_t hint) { int color, nset, index, start; caddr_t va; #ifdef PPDEBUG pp_allocs++; #endif /* PPDEBUG */ if (cache & CACHE_VAC) { color = sfmmu_get_ppvcolor(pp); if (color == -1) { if ((intptr_t)hint != -1L) { color = addr_to_vcolor(hint); } else { color = addr_to_vcolor(mmu_ptob(pp->p_pagenum)); } } } else { /* * For physical caches, we can pick any address we want. */ color = 0; } start = color; do { for (nset = 0; nset < nsets; nset++) { index = clsettoarray(color, nset); va = ppmap_vaddrs[index]; if (va != NULL) { #ifdef PPDEBUG align_hits[color]++; #endif /* PPDEBUG */ if (atomic_cas_ptr(&ppmap_vaddrs[index], va, NULL) == va) { hat_memload(kas.a_hat, va, pp, vprot | HAT_NOSYNC, HAT_LOAD_LOCK); return (va); } } } /* * first pick didn't succeed, try another */ if (++color == ppmap_pages) color = 0; } while (color != start); #ifdef PPDEBUG ppalloc_noslot++; #endif /* PPDEBUG */ /* * No free slots; get a random one from the kernel heap area. */ va = vmem_alloc(heap_arena, PAGESIZE, VM_SLEEP); hat_memload(kas.a_hat, va, pp, vprot | HAT_NOSYNC, HAT_LOAD_LOCK); return (va); } void ppmapout(caddr_t va) { int color, nset, index; if (va >= kernelheap && va < ekernelheap) { /* * Space came from kernelmap, flush the page and * return the space. */ hat_unload(kas.a_hat, va, PAGESIZE, (HAT_UNLOAD_NOSYNC | HAT_UNLOAD_UNLOCK)); vmem_free(heap_arena, va, PAGESIZE); } else { /* * Space came from ppmap_vaddrs[], give it back. */ color = addr_to_vcolor(va); ASSERT((cache & CACHE_VAC)? (color < ppmap_pages) : 1); nset = ((uintptr_t)va >> ppmap_shift) & (nsets - 1); index = clsettoarray(color, nset); hat_unload(kas.a_hat, va, PAGESIZE, (HAT_UNLOAD_NOSYNC | HAT_UNLOAD_UNLOCK)); ASSERT(ppmap_vaddrs[index] == NULL); ppmap_vaddrs[index] = va; } } #ifdef DEBUG #define PP_STAT_ADD(stat) (stat)++ uint_t pload, ploadfail; uint_t ppzero, ppzero_short; #else #define PP_STAT_ADD(stat) #endif /* DEBUG */ /* * Find a slot in per CPU page copy area. Load up a locked TLB in the * running cpu. We don't call hat layer to load up the tte since the * mapping is only temporary. If the thread migrates it'll get a TLB * miss trap and TLB/TSB miss handler will panic since there is no * official hat record of this mapping. */ static caddr_t pp_load_tlb(processorid_t cpu, caddr_t **pslot, page_t *pp, uint_t prot) { struct ppmap_va *ppmap; tte_t tte; caddr_t *myslot; caddr_t va; long i, start, stride; int vcolor; uint_t flags, strict_flag; PP_STAT_ADD(pload); ppmap = &ppmap_va[cpu]; va = (caddr_t)(PPMAP_FAST_BASE + (MMU_PAGESIZE * MAXPP_SLOTS) * cpu); myslot = ppmap->ppmap_slots; ASSERT(addr_to_vcolor(va) == 0); if (prot & TTE_HWWR_INT) { flags = PPAGE_STORE_VCOLORING | PPAGE_STORES_POLLUTE; strict_flag = PPAGE_STORES_POLLUTE; } else { flags = PPAGE_LOAD_VCOLORING | PPAGE_LOADS_POLLUTE; strict_flag = PPAGE_LOADS_POLLUTE; } /* * If consistent handling is required then keep the current * vcolor of the page. Furthermore, if loads or stores can * pollute the VAC then using a "new" page (unassigned vcolor) * won't work and we have to return a failure. */ if (pp_consistent_coloring & flags) { vcolor = sfmmu_get_ppvcolor(pp); if ((vcolor == -1) && (pp_consistent_coloring & strict_flag)) return (NULL); /* else keep the current vcolor of the page */ } else { vcolor = -1; } if (vcolor != -1) { va += MMU_PAGESIZE * vcolor; start = vcolor; stride = ppmap_pages; /* number of colors */ myslot += vcolor; } else { start = 0; stride = 1; } for (i = start; i < pp_slots; i += stride) { if (*myslot == NULL) { if (atomic_cas_ptr(myslot, NULL, va) == NULL) break; } myslot += stride; va += MMU_PAGESIZE * stride; } if (i >= pp_slots) { PP_STAT_ADD(ploadfail); return (NULL); } ASSERT(vcolor == -1 || addr_to_vcolor(va) == vcolor); /* * Now we have a slot we can use, make the tte. */ tte.tte_inthi = TTE_VALID_INT | TTE_PFN_INTHI(pp->p_pagenum); tte.tte_intlo = TTE_PFN_INTLO(pp->p_pagenum) | TTE_CP_INT | TTE_CV_INT | TTE_PRIV_INT | TTE_LCK_INT | prot; ASSERT(CPU->cpu_id == cpu); sfmmu_dtlb_ld_kva(va, &tte); *pslot = myslot; /* Return ptr to the slot we used. */ return (va); } static void pp_unload_tlb(caddr_t *pslot, caddr_t va) { ASSERT(*pslot == va); vtag_flushpage(va, (uint64_t)ksfmmup); *pslot = NULL; /* release the slot */ } /* * Common copy routine which attempts to use hwblkpagecopy. If this routine * can't be used, failure (0) will be returned. Otherwise, a PAGESIZE page * will be copied and success (1) will be returned. */ int ppcopy_common(page_t *fm_pp, page_t *to_pp) { caddr_t fm_va, to_va; caddr_t *fm_slot, *to_slot; processorid_t cpu; label_t ljb; int ret = 1; ASSERT(fm_pp != NULL && PAGE_LOCKED(fm_pp)); ASSERT(to_pp != NULL && PAGE_LOCKED(to_pp)); /* * If we can't use VIS block loads and stores we can't use * pp_load_tlb/pp_unload_tlb due to the possibility of * d$ aliasing. */ if (!use_hw_bcopy && (cache & CACHE_VAC)) return (0); kpreempt_disable(); cpu = CPU->cpu_id; fm_va = pp_load_tlb(cpu, &fm_slot, fm_pp, 0); if (fm_va == NULL) { kpreempt_enable(); return (0); } to_va = pp_load_tlb(cpu, &to_slot, to_pp, TTE_HWWR_INT); if (to_va == NULL) { pp_unload_tlb(fm_slot, fm_va); kpreempt_enable(); return (0); } if (on_fault(&ljb)) { ret = 0; goto faulted; } hwblkpagecopy(fm_va, to_va); no_fault(); faulted: ASSERT(CPU->cpu_id == cpu); pp_unload_tlb(fm_slot, fm_va); pp_unload_tlb(to_slot, to_va); kpreempt_enable(); return (ret); } /* * Routine to copy kernel pages during relocation. It will copy one * PAGESIZE page to another PAGESIZE page. This function may be called * above LOCK_LEVEL so it should not grab any locks. */ void ppcopy_kernel__relocatable(page_t *fm_pp, page_t *to_pp) { uint64_t fm_pa, to_pa; size_t nbytes; fm_pa = (uint64_t)(fm_pp->p_pagenum) << MMU_PAGESHIFT; to_pa = (uint64_t)(to_pp->p_pagenum) << MMU_PAGESHIFT; nbytes = MMU_PAGESIZE; for (; nbytes > 0; fm_pa += 32, to_pa += 32, nbytes -= 32) hw_pa_bcopy32(fm_pa, to_pa); } /* * Copy the data from the physical page represented by "frompp" to * that represented by "topp". * * Try to use per cpu mapping first, if that fails then call pp_mapin * to load it. * * Returns one on success or zero on some sort of fault while doing the copy. */ int ppcopy(page_t *fm_pp, page_t *to_pp) { caddr_t fm_va, to_va; label_t ljb; int ret = 1; boolean_t use_kpm = B_FALSE; /* Try the fast path first */ if (ppcopy_common(fm_pp, to_pp)) return (1); /* * Try to map using KPM if enabled and we are the cageout thread. * If it fails, fall back to ppmapin/ppmaput */ if (kpm_enable) { if (curthread == kcage_cageout_thread) use_kpm = B_TRUE; } if (use_kpm) { if ((fm_va = hat_kpm_mapin(fm_pp, NULL)) == NULL || (to_va = hat_kpm_mapin(to_pp, NULL)) == NULL) { if (fm_va != NULL) hat_kpm_mapout(fm_pp, NULL, fm_va); use_kpm = B_FALSE; } } if (use_kpm == B_FALSE) { /* do the slow path */ fm_va = ppmapin(fm_pp, PROT_READ, (caddr_t)-1); to_va = ppmapin(to_pp, PROT_READ | PROT_WRITE, fm_va); if (on_fault(&ljb)) { ret = 0; goto faulted; } } bcopy(fm_va, to_va, PAGESIZE); no_fault(); faulted: /* unmap */ if (use_kpm == B_TRUE) { hat_kpm_mapout(fm_pp, NULL, fm_va); hat_kpm_mapout(to_pp, NULL, to_va); } else { ppmapout(fm_va); ppmapout(to_va); } return (ret); } /* * Zero the physical page from off to off + len given by `pp' * without changing the reference and modified bits of page. * * Again, we'll try per cpu mapping first. */ void pagezero(page_t *pp, uint_t off, uint_t len) { caddr_t va; caddr_t *slot; int fast = 1; processorid_t cpu; extern int hwblkclr(void *, size_t); extern int use_hw_bzero; ASSERT((int)len > 0 && (int)off >= 0 && off + len <= PAGESIZE); ASSERT(PAGE_LOCKED(pp)); PP_STAT_ADD(ppzero); if (len != MMU_PAGESIZE || !use_hw_bzero) { /* * Since the fast path doesn't do anything about * VAC coloring, we make sure bcopy h/w will be used. */ fast = 0; va = NULL; PP_STAT_ADD(ppzero_short); } kpreempt_disable(); if (fast) { cpu = CPU->cpu_id; va = pp_load_tlb(cpu, &slot, pp, TTE_HWWR_INT); } if (va == NULL) { /* * We are here either length != MMU_PAGESIZE or pp_load_tlb() * returns NULL or use_hw_bzero is disabled. */ va = ppmapin(pp, PROT_READ | PROT_WRITE, (caddr_t)-1); fast = 0; } if (hwblkclr(va + off, len)) { /* * We may not have used block commit asi. * So flush the I-$ manually */ ASSERT(fast == 0); sync_icache(va + off, len); } else { /* * We have used blk commit, and flushed the I-$. However we * still may have an instruction in the pipeline. Only a flush * instruction will invalidate that. */ doflush(va); } if (fast) { ASSERT(CPU->cpu_id == cpu); pp_unload_tlb(slot, va); } else { ppmapout(va); } kpreempt_enable(); }