xref: /illumos-gate/usr/src/uts/common/io/physmem.c (revision 12e9eaaed8879c4b56117e0ea40df0a2d2e16b0e)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 
27 #include <sys/types.h>
28 #include <sys/modctl.h>
29 #include <sys/conf.h>
30 #include <sys/ddi.h>
31 #include <sys/sunddi.h>
32 #include <sys/devops.h>
33 #include <sys/stat.h>
34 #include <sys/file.h>
35 #include <sys/cred.h>
36 #include <sys/policy.h>
37 #include <sys/errno.h>
38 #include <vm/seg_dev.h>
39 #include <vm/seg_vn.h>
40 #include <vm/page.h>
41 #include <sys/fs/swapnode.h>
42 #include <sys/sysmacros.h>
43 #include <sys/fcntl.h>
44 #include <sys/vmsystm.h>
45 #include <sys/physmem.h>
46 #include <sys/vfs_opreg.h>
47 
48 static dev_info_t		*physmem_dip = NULL;
49 
50 /*
51  * Linked list element hanging off physmem_proc_hash below, which holds all
52  * the information for a given segment which has been setup for this process.
53  * This is a simple linked list as we are assuming that for a given process
54  * the setup ioctl will only be called a handful of times.  If this assumption
55  * changes in the future, a quicker to traverse data structure should be used.
56  */
57 struct physmem_hash {
58 	struct physmem_hash *ph_next;
59 	uint64_t ph_base_pa;
60 	caddr_t ph_base_va;
61 	size_t ph_seg_len;
62 	struct vnode *ph_vnode;
63 };
64 
65 /*
66  * Hash of all of the processes which have setup mappings with the driver with
67  * pointers to per process data.
68  */
69 struct physmem_proc_hash {
70 	struct proc *pph_proc;
71 	struct physmem_hash *pph_hash;
72 	struct physmem_proc_hash *pph_next;
73 };
74 
75 
76 /* Needs to be a power of two for simple hash algorithm */
77 #define	PPH_SIZE	8
78 struct physmem_proc_hash *pph[PPH_SIZE];
79 
80 /*
81  * Lock which protects the pph hash above.  To add an element (either a new
82  * process or a new segment) the WRITE lock must be held.  To traverse the
83  * list, only a READ lock is needed.
84  */
85 krwlock_t pph_rwlock;
86 
87 #define	PHYSMEM_HASH(procp) ((int)((((uintptr_t)procp) >> 8) & (PPH_SIZE - 1)))
88 
89 /*
90  * Need to keep a reference count of how many processes have the driver
91  * open to prevent it from disappearing.
92  */
93 uint64_t physmem_vnodecnt;
94 kmutex_t physmem_mutex;		/* protects phsymem_vnodecnt */
95 
96 static int physmem_getpage(struct vnode *vp, offset_t off, size_t len,
97     uint_t *protp, page_t *pl[], size_t plsz, struct seg *seg, caddr_t addr,
98     enum seg_rw rw, struct cred *cr, caller_context_t *ct);
99 
100 static int physmem_addmap(struct vnode *vp, offset_t off, struct as *as,
101     caddr_t addr, size_t len, uchar_t prot, uchar_t maxprot, uint_t flags,
102     struct cred *cred, caller_context_t *ct);
103 
104 static int physmem_delmap(struct vnode *vp, offset_t off, struct as *as,
105     caddr_t addr, size_t len, uint_t prot, uint_t maxprot, uint_t flags,
106     struct cred *cred, caller_context_t *ct);
107 
108 static void physmem_inactive(vnode_t *vp, cred_t *crp, caller_context_t *ct);
109 
110 const fs_operation_def_t physmem_vnodeops_template[] = {
111 	VOPNAME_GETPAGE,	{ .vop_getpage = physmem_getpage },
112 	VOPNAME_ADDMAP,		{ .vop_addmap = physmem_addmap },
113 	VOPNAME_DELMAP,		{ .vop_delmap = physmem_delmap },
114 	VOPNAME_INACTIVE,	{ .vop_inactive = physmem_inactive },
115 	NULL,			NULL
116 };
117 
118 vnodeops_t *physmem_vnodeops = NULL;
119 
120 /*
121  * Removes the current process from the hash if the process has no more
122  * physmem segments active.
123  */
124 void
125 physmem_remove_hash_proc()
126 {
127 	int index;
128 	struct physmem_proc_hash **walker;
129 	struct physmem_proc_hash *victim = NULL;
130 
131 	index = PHYSMEM_HASH(curproc);
132 	rw_enter(&pph_rwlock, RW_WRITER);
133 	walker = &pph[index];
134 	while (*walker != NULL) {
135 		if ((*walker)->pph_proc == curproc &&
136 		    (*walker)->pph_hash == NULL) {
137 			victim = *walker;
138 			*walker = victim->pph_next;
139 			break;
140 		}
141 		walker = &((*walker)->pph_next);
142 	}
143 	rw_exit(&pph_rwlock);
144 	if (victim != NULL)
145 		kmem_free(victim, sizeof (struct physmem_proc_hash));
146 }
147 
148 /*
149  * Add a new entry to the hash for the given process to cache the
150  * address ranges that it is working on.  If this is the first hash
151  * item to be added for this process, we will create the head pointer
152  * for this process.
153  * Returns 0 on success, ERANGE when the physical address is already in the
154  * hash.
155  */
156 int
157 physmem_add_hash(struct physmem_hash *php)
158 {
159 	int index;
160 	struct physmem_proc_hash *iterator;
161 	struct physmem_proc_hash *newp = NULL;
162 	struct physmem_hash *temp;
163 	int ret = 0;
164 
165 	index = PHYSMEM_HASH(curproc);
166 
167 insert:
168 	rw_enter(&pph_rwlock, RW_WRITER);
169 	iterator = pph[index];
170 	while (iterator != NULL) {
171 		if (iterator->pph_proc == curproc) {
172 			/*
173 			 * check to make sure a single process does not try to
174 			 * map the same region twice.
175 			 */
176 			for (temp = iterator->pph_hash; temp != NULL;
177 			    temp = temp->ph_next) {
178 				if ((php->ph_base_pa >= temp->ph_base_pa &&
179 				    php->ph_base_pa < temp->ph_base_pa +
180 				    temp->ph_seg_len) ||
181 				    (temp->ph_base_pa >= php->ph_base_pa &&
182 				    temp->ph_base_pa < php->ph_base_pa +
183 				    php->ph_seg_len)) {
184 					ret = ERANGE;
185 					break;
186 				}
187 			}
188 			if (ret == 0) {
189 				php->ph_next = iterator->pph_hash;
190 				iterator->pph_hash = php;
191 			}
192 			rw_exit(&pph_rwlock);
193 			/* Need to check for two threads in sync */
194 			if (newp != NULL)
195 				kmem_free(newp, sizeof (*newp));
196 			return (ret);
197 		}
198 		iterator = iterator->pph_next;
199 	}
200 
201 	if (newp != NULL) {
202 		newp->pph_proc = curproc;
203 		newp->pph_next = pph[index];
204 		newp->pph_hash = php;
205 		php->ph_next = NULL;
206 		pph[index] = newp;
207 		rw_exit(&pph_rwlock);
208 		return (0);
209 	}
210 
211 	rw_exit(&pph_rwlock);
212 	/* Dropped the lock so we could use KM_SLEEP */
213 	newp = kmem_zalloc(sizeof (struct physmem_proc_hash), KM_SLEEP);
214 	goto insert;
215 }
216 
217 /*
218  * Will return the pointer to the physmem_hash struct if the setup routine
219  * has previously been called for this memory.
220  * Returns NULL on failure.
221  */
222 struct physmem_hash *
223 physmem_get_hash(uint64_t req_paddr, size_t len, proc_t *procp)
224 {
225 	int index;
226 	struct physmem_proc_hash *proc_hp;
227 	struct physmem_hash *php;
228 
229 	ASSERT(rw_lock_held(&pph_rwlock));
230 
231 	index = PHYSMEM_HASH(procp);
232 	proc_hp = pph[index];
233 	while (proc_hp != NULL) {
234 		if (proc_hp->pph_proc == procp) {
235 			php = proc_hp->pph_hash;
236 			while (php != NULL) {
237 				if ((req_paddr >= php->ph_base_pa) &&
238 				    (req_paddr + len <=
239 				    php->ph_base_pa + php->ph_seg_len)) {
240 					return (php);
241 				}
242 				php = php->ph_next;
243 			}
244 		}
245 		proc_hp = proc_hp->pph_next;
246 	}
247 	return (NULL);
248 }
249 
250 int
251 physmem_validate_cookie(uint64_t p_cookie)
252 {
253 	int index;
254 	struct physmem_proc_hash *proc_hp;
255 	struct physmem_hash *php;
256 
257 	ASSERT(rw_lock_held(&pph_rwlock));
258 
259 	index = PHYSMEM_HASH(curproc);
260 	proc_hp = pph[index];
261 	while (proc_hp != NULL) {
262 		if (proc_hp->pph_proc == curproc) {
263 			php = proc_hp->pph_hash;
264 			while (php != NULL) {
265 				if ((uint64_t)(uintptr_t)php == p_cookie) {
266 					return (1);
267 				}
268 				php = php->ph_next;
269 			}
270 		}
271 		proc_hp = proc_hp->pph_next;
272 	}
273 	return (0);
274 }
275 
276 /*
277  * Remove the given vnode from the pph hash.  If it exists in the hash the
278  * process still has to be around as the vnode is obviously still around and
279  * since it's a physmem vnode, it must be in the hash.
280  * If it is not in the hash that must mean that the setup ioctl failed.
281  * Return 0 in this instance, 1 if it is in the hash.
282  */
283 int
284 physmem_remove_vnode_hash(vnode_t *vp)
285 {
286 	int index;
287 	struct physmem_proc_hash *proc_hp;
288 	struct physmem_hash **phpp;
289 	struct physmem_hash *victim;
290 
291 	index = PHYSMEM_HASH(curproc);
292 	/* synchronize with the map routine */
293 	rw_enter(&pph_rwlock, RW_WRITER);
294 	proc_hp = pph[index];
295 	while (proc_hp != NULL) {
296 		if (proc_hp->pph_proc == curproc) {
297 			phpp = &proc_hp->pph_hash;
298 			while (*phpp != NULL) {
299 				if ((*phpp)->ph_vnode == vp) {
300 					victim = *phpp;
301 					*phpp = victim->ph_next;
302 
303 					rw_exit(&pph_rwlock);
304 					kmem_free(victim, sizeof (*victim));
305 					return (1);
306 				}
307 				phpp = &(*phpp)->ph_next;
308 			}
309 		}
310 		proc_hp = proc_hp->pph_next;
311 	}
312 	rw_exit(&pph_rwlock);
313 
314 	/* not found */
315 	return (0);
316 }
317 
318 int
319 physmem_setup_vnops()
320 {
321 	int error;
322 	char *name = "physmem";
323 	if (physmem_vnodeops != NULL)
324 		cmn_err(CE_PANIC, "physmem vnodeops already set\n");
325 	error = vn_make_ops(name, physmem_vnodeops_template, &physmem_vnodeops);
326 	if (error != 0) {
327 		cmn_err(CE_WARN, "physmem_setup_vnops: bad vnode ops template");
328 	}
329 	return (error);
330 }
331 
332 /*
333  * The guts of the PHYSMEM_SETUP ioctl.
334  * Create a segment in the address space with the specified parameters.
335  * If pspp->user_va is NULL, as_gap will be used to find an appropriate VA.
336  * We do not do bounds checking on the requested physical addresses, if they
337  * do not exist in the system, they will not be mappable.
338  * Returns 0 on success with the following error codes on failure:
339  *	ENOMEM - The VA range requested was already mapped if pspp->user_va is
340  *		non-NULL or the system was unable to find enough VA space for
341  *		the desired length if user_va was NULL>
342  *	EINVAL - The requested PA, VA, or length was not PAGESIZE aligned.
343  */
344 int
345 physmem_setup_addrs(struct physmem_setup_param *pspp)
346 {
347 	struct as *as = curproc->p_as;
348 	struct segvn_crargs vn_a;
349 	int ret = 0;
350 	uint64_t base_pa;
351 	size_t len;
352 	caddr_t uvaddr;
353 	struct vnode *vp;
354 	struct physmem_hash *php;
355 
356 	ASSERT(pspp != NULL);
357 	base_pa = pspp->req_paddr;
358 	len = pspp->len;
359 	uvaddr = (caddr_t)(uintptr_t)pspp->user_va;
360 
361 	/* Sanity checking */
362 	if (!IS_P2ALIGNED(base_pa, PAGESIZE))
363 		return (EINVAL);
364 	if (!IS_P2ALIGNED(len, PAGESIZE))
365 		return (EINVAL);
366 	if (uvaddr != NULL && !IS_P2ALIGNED(uvaddr, PAGESIZE))
367 		return (EINVAL);
368 
369 	php = kmem_zalloc(sizeof (struct physmem_hash), KM_SLEEP);
370 
371 	/* Need to bump vnode count so that the driver can not be unloaded */
372 	mutex_enter(&physmem_mutex);
373 	physmem_vnodecnt++;
374 	mutex_exit(&physmem_mutex);
375 
376 	vp = vn_alloc(KM_SLEEP);
377 	ASSERT(vp != NULL);	/* SLEEP can't return NULL */
378 	vn_setops(vp, physmem_vnodeops);
379 
380 	php->ph_vnode = vp;
381 
382 	vn_a.vp = vp;
383 	vn_a.offset = (u_offset_t)base_pa;
384 	vn_a.type = MAP_SHARED;
385 	vn_a.prot = PROT_ALL;
386 	vn_a.maxprot = PROT_ALL;
387 	vn_a.flags = 0;
388 	vn_a.cred = NULL;
389 	vn_a.amp = NULL;
390 	vn_a.szc = 0;
391 	vn_a.lgrp_mem_policy_flags = 0;
392 
393 	as_rangelock(as);
394 	if (uvaddr != NULL) {
395 		if (as_gap(as, len, &uvaddr, &len, AH_LO, NULL) == -1) {
396 			ret = ENOMEM;
397 fail:
398 			as_rangeunlock(as);
399 			vn_free(vp);
400 			kmem_free(php, sizeof (*php));
401 			mutex_enter(&physmem_mutex);
402 			physmem_vnodecnt--;
403 			mutex_exit(&physmem_mutex);
404 			return (ret);
405 		}
406 	} else {
407 		/* We pick the address for the user */
408 		map_addr(&uvaddr, len, 0, 1, 0);
409 		if (uvaddr == NULL) {
410 			ret = ENOMEM;
411 			goto fail;
412 		}
413 	}
414 	ret = as_map(as, uvaddr, len, segvn_create, &vn_a);
415 
416 	if (ret == 0) {
417 		as_rangeunlock(as);
418 		php->ph_base_pa = base_pa;
419 		php->ph_base_va = uvaddr;
420 		php->ph_seg_len = len;
421 		pspp->user_va = (uint64_t)(uintptr_t)uvaddr;
422 		pspp->cookie = (uint64_t)(uintptr_t)php;
423 		ret = physmem_add_hash(php);
424 		if (ret == 0)
425 			return (0);
426 
427 		/* Note that the call to as_unmap will free the vnode */
428 		(void) as_unmap(as, uvaddr, len);
429 		kmem_free(php, sizeof (*php));
430 		return (ret);
431 	}
432 
433 	goto fail;
434 	/*NOTREACHED*/
435 }
436 
437 /*
438  * The guts of the PHYSMEM_MAP ioctl.
439  * Map the given PA to the appropriate VA if PHYSMEM_SETUP ioctl has already
440  * been called for this PA range.
441  * Returns 0 on success with the following error codes on failure:
442  *	EPERM - The requested page is long term locked, and thus repeated
443  *		requests to allocate this page will likely fail.
444  *	EAGAIN - The requested page could not be allocated, but it is believed
445  *		that future attempts could succeed.
446  *	ENOMEM - There was not enough free memory in the system to safely
447  *		map the requested page.
448  *	EINVAL - The requested paddr was not PAGESIZE aligned or the
449  *		PHYSMEM_SETUP ioctl was not called for this page.
450  *	ENOENT - The requested page was iniside the kernel cage, and the
451  *		PHYSMEM_CAGE flag was not set.
452  *	EBUSY - The requested page is retired and the PHYSMEM_RETIRE flag
453  *		was not set.
454  */
455 static int
456 physmem_map_addrs(struct physmem_map_param *pmpp)
457 {
458 	caddr_t uvaddr;
459 	page_t *pp;
460 	uint64_t req_paddr;
461 	struct vnode *vp;
462 	int ret = 0;
463 	struct physmem_hash *php;
464 	uint_t flags = 0;
465 
466 	ASSERT(pmpp != NULL);
467 	req_paddr = pmpp->req_paddr;
468 
469 	if (!IS_P2ALIGNED(req_paddr, PAGESIZE))
470 		return (EINVAL);
471 	/* Find the vnode for this map request */
472 	rw_enter(&pph_rwlock, RW_READER);
473 	php = physmem_get_hash(req_paddr, PAGESIZE, curproc);
474 	if (php == NULL) {
475 		rw_exit(&pph_rwlock);
476 		return (EINVAL);
477 	}
478 	vp = php->ph_vnode;
479 	uvaddr = php->ph_base_va + (req_paddr - php->ph_base_pa);
480 	rw_exit(&pph_rwlock);
481 
482 	pp = page_numtopp_nolock(btop((size_t)req_paddr));
483 	if (pp == NULL) {
484 		pmpp->ret_va = 0;
485 		return (EPERM);
486 	}
487 
488 	/*
489 	 * Check to see if page already mapped correctly.  This can happen
490 	 * when we failed to capture a page previously and it was captured
491 	 * asynchronously for us.  Return success in this case.
492 	 */
493 	if (pp->p_vnode == vp) {
494 		ASSERT(pp->p_offset == (u_offset_t)req_paddr);
495 		pmpp->ret_va = (uint64_t)(uintptr_t)uvaddr;
496 		return (0);
497 	}
498 
499 	/*
500 	 * physmem should be responsible for checking for cage
501 	 * and prom pages.
502 	 */
503 	if (pmpp->flags & PHYSMEM_CAGE)
504 		flags = CAPTURE_GET_CAGE;
505 	if (pmpp->flags & PHYSMEM_RETIRED)
506 		flags |= CAPTURE_GET_RETIRED;
507 
508 	ret = page_trycapture(pp, 0, flags | CAPTURE_PHYSMEM, curproc);
509 
510 	if (ret != 0) {
511 		pmpp->ret_va = 0;
512 		return (ret);
513 	} else {
514 		pmpp->ret_va = (uint64_t)(uintptr_t)uvaddr;
515 		return (0);
516 	}
517 }
518 
519 /*
520  * Map the given page into the process's address space if possible.
521  * We actually only hash the page in on the correct vnode as the page
522  * will be mapped via segvn_pagefault.
523  * returns 0 on success
524  * returns 1 if there is no need to map this page anymore (process exited)
525  * returns -1 if we failed to map the page.
526  */
527 int
528 map_page_proc(page_t *pp, void *arg, uint_t flags)
529 {
530 	struct vnode *vp;
531 	proc_t *procp = (proc_t *)arg;
532 	int ret;
533 	u_offset_t paddr = (u_offset_t)ptob(pp->p_pagenum);
534 	struct physmem_hash *php;
535 
536 	ASSERT(pp != NULL);
537 
538 	/*
539 	 * Check against availrmem to make sure that we're not low on memory.
540 	 * We check again here as ASYNC requests do not do this check elsewhere.
541 	 * We return 1 as we don't want the page to have the PR_CAPTURE bit
542 	 * set or be on the page capture hash.
543 	 */
544 	if (swapfs_minfree > availrmem + 1) {
545 		page_free(pp, 1);
546 		return (1);
547 	}
548 
549 	/*
550 	 * If this is an asynchronous request for the current process,
551 	 * we can not map the page as it's possible that we are also in the
552 	 * process of unmapping the page which could result in a deadlock
553 	 * with the as lock.
554 	 */
555 	if ((flags & CAPTURE_ASYNC) && (curproc == procp)) {
556 		page_free(pp, 1);
557 		return (-1);
558 	}
559 
560 	/* only return zeroed out pages */
561 	pagezero(pp, 0, PAGESIZE);
562 
563 	rw_enter(&pph_rwlock, RW_READER);
564 	php = physmem_get_hash(paddr, PAGESIZE, procp);
565 	if (php == NULL) {
566 		rw_exit(&pph_rwlock);
567 		/*
568 		 * Free the page as there is no longer a valid outstanding
569 		 * request for this page.
570 		 */
571 		page_free(pp, 1);
572 		return (1);
573 	}
574 
575 	vp = php->ph_vnode;
576 
577 	/*
578 	 * We need to protect against a possible deadlock here where we own
579 	 * the vnode page hash mutex and want to acquire it again as there
580 	 * are locations in the code, where we unlock a page while holding
581 	 * the mutex which can lead to the page being captured and eventually
582 	 * end up here.
583 	 */
584 	if (mutex_owned(page_vnode_mutex(vp))) {
585 		rw_exit(&pph_rwlock);
586 		page_free(pp, 1);
587 		return (-1);
588 	}
589 
590 	ret = page_hashin(pp, vp, paddr, NULL);
591 	rw_exit(&pph_rwlock);
592 	if (ret == 0) {
593 		page_free(pp, 1);
594 		return (-1);
595 	}
596 
597 	page_downgrade(pp);
598 
599 	mutex_enter(&freemem_lock);
600 	availrmem--;
601 	mutex_exit(&freemem_lock);
602 
603 	return (0);
604 }
605 
606 /*
607  * The guts of the PHYSMEM_DESTROY ioctl.
608  * The cookie passed in will provide all of the information needed to
609  * free up the address space and physical memory associated with the
610  * corresponding PHSYMEM_SETUP ioctl.
611  * Returns 0 on success with the following error codes on failure:
612  *	EINVAL - The cookie supplied is not valid.
613  */
614 int
615 physmem_destroy_addrs(uint64_t p_cookie)
616 {
617 	struct as *as = curproc->p_as;
618 	size_t len;
619 	caddr_t uvaddr;
620 
621 	rw_enter(&pph_rwlock, RW_READER);
622 	if (physmem_validate_cookie(p_cookie) == 0) {
623 		rw_exit(&pph_rwlock);
624 		return (EINVAL);
625 	}
626 
627 	len = ((struct physmem_hash *)(uintptr_t)p_cookie)->ph_seg_len;
628 	uvaddr = ((struct physmem_hash *)(uintptr_t)p_cookie)->ph_base_va;
629 	rw_exit(&pph_rwlock);
630 
631 	(void) as_unmap(as, uvaddr, len);
632 
633 	return (0);
634 }
635 
636 /*
637  * If the page has been hashed into the physmem vnode, then just look it up
638  * and return it via pl, otherwise return ENOMEM as the map ioctl has not
639  * succeeded on the given page.
640  */
641 /*ARGSUSED*/
642 static int
643 physmem_getpage(struct vnode *vp, offset_t off, size_t len, uint_t *protp,
644     page_t *pl[], size_t plsz, struct seg *seg, caddr_t addr, enum seg_rw rw,
645     struct cred *cr, caller_context_t *ct)
646 {
647 	page_t *pp;
648 
649 	ASSERT(len == PAGESIZE);
650 	ASSERT(AS_READ_HELD(seg->s_as));
651 
652 	/*
653 	 * If the page is in the hash, then we successfully claimed this
654 	 * page earlier, so return it to the caller.
655 	 */
656 	pp = page_lookup(vp, off, SE_SHARED);
657 	if (pp != NULL) {
658 		pl[0] = pp;
659 		pl[1] = NULL;
660 		*protp = PROT_ALL;
661 		return (0);
662 	}
663 	return (ENOMEM);
664 }
665 
666 /*
667  * We can not allow a process mapping /dev/physmem pages to fork as there can
668  * only be a single mapping to a /dev/physmem page at a given time.  Thus, the
669  * return of EINVAL when we are not working on our own address space.
670  * Otherwise we return zero as this function is required for normal operation.
671  */
672 /*ARGSUSED*/
673 static int
674 physmem_addmap(struct vnode *vp, offset_t off, struct as *as,
675     caddr_t addr, size_t len, uchar_t prot, uchar_t maxprot, uint_t flags,
676     struct cred *cred, caller_context_t *ct)
677 {
678 	if (curproc->p_as != as) {
679 		return (EINVAL);
680 	}
681 	return (0);
682 }
683 
684 /* Will always get called for removing a whole segment. */
685 /*ARGSUSED*/
686 static int
687 physmem_delmap(struct vnode *vp, offset_t off, struct as *as,
688     caddr_t addr, size_t len, uint_t prot, uint_t maxprot, uint_t flags,
689     struct cred *cred, caller_context_t *ct)
690 {
691 	/*
692 	 * Release our hold on the vnode so that the final VN_RELE will
693 	 * call physmem_inactive to clean things up.
694 	 */
695 	VN_RELE(vp);
696 
697 	return (0);
698 }
699 
700 /*
701  * Clean up all the pages belonging to this vnode and then free it.
702  */
703 /*ARGSUSED*/
704 static void
705 physmem_inactive(vnode_t *vp, cred_t *crp, caller_context_t *ct)
706 {
707 	page_t *pp;
708 
709 	/*
710 	 * Remove the vnode from the hash now, to prevent asynchronous
711 	 * attempts to map into this vnode.  This avoids a deadlock
712 	 * where two threads try to get into this logic at the same
713 	 * time and try to map the pages they are destroying into the
714 	 * other's address space.
715 	 * If it's not in the hash, just free it.
716 	 */
717 	if (physmem_remove_vnode_hash(vp) == 0) {
718 		ASSERT(vp->v_pages == NULL);
719 		vn_free(vp);
720 		physmem_remove_hash_proc();
721 		mutex_enter(&physmem_mutex);
722 		physmem_vnodecnt--;
723 		mutex_exit(&physmem_mutex);
724 		return;
725 	}
726 
727 	/*
728 	 * At this point in time, no other logic can be adding or removing
729 	 * pages from the vnode, otherwise the v_pages list could be inaccurate.
730 	 */
731 
732 	while ((pp = vp->v_pages) != NULL) {
733 		page_t *rpp;
734 		if (page_tryupgrade(pp)) {
735 			/*
736 			 * set lckcnt for page_destroy to do availrmem
737 			 * accounting
738 			 */
739 			pp->p_lckcnt = 1;
740 			page_destroy(pp, 0);
741 		} else {
742 			/* failure to lock should be transient */
743 			rpp = page_lookup(vp, ptob(pp->p_pagenum), SE_SHARED);
744 			if (rpp != pp) {
745 				page_unlock(rpp);
746 				continue;
747 			}
748 			page_unlock(pp);
749 		}
750 	}
751 	vn_free(vp);
752 	physmem_remove_hash_proc();
753 	mutex_enter(&physmem_mutex);
754 	physmem_vnodecnt--;
755 	mutex_exit(&physmem_mutex);
756 }
757 
758 /*ARGSUSED*/
759 static int
760 physmem_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *credp,
761     int *rvalp)
762 {
763 	int ret;
764 
765 	switch (cmd) {
766 	case PHYSMEM_SETUP:
767 		{
768 			struct physmem_setup_param psp;
769 			if (ddi_copyin((void *)arg, &psp,
770 			    sizeof (struct physmem_setup_param), 0))
771 				return (EFAULT);
772 			ret = physmem_setup_addrs(&psp);
773 			if (ddi_copyout(&psp, (void *)arg, sizeof (psp), 0))
774 				return (EFAULT);
775 		}
776 		break;
777 	case PHYSMEM_MAP:
778 		{
779 			struct physmem_map_param pmp;
780 			if (ddi_copyin((void *)arg, &pmp,
781 			    sizeof (struct physmem_map_param), 0))
782 				return (EFAULT);
783 			ret = physmem_map_addrs(&pmp);
784 			if (ddi_copyout(&pmp, (void *)arg, sizeof (pmp), 0))
785 				return (EFAULT);
786 		}
787 		break;
788 	case PHYSMEM_DESTROY:
789 		{
790 			uint64_t cookie;
791 			if (ddi_copyin((void *)arg, &cookie,
792 			    sizeof (uint64_t), 0))
793 				return (EFAULT);
794 			ret = physmem_destroy_addrs(cookie);
795 		}
796 		break;
797 	default:
798 		return (ENOTSUP);
799 	}
800 	return (ret);
801 }
802 
803 /*ARGSUSED*/
804 static int
805 physmem_open(dev_t *devp, int flag, int otyp, cred_t *credp)
806 {
807 	int ret;
808 	static int msg_printed = 0;
809 
810 	if ((flag & (FWRITE | FREAD)) != (FWRITE | FREAD)) {
811 		return (EINVAL);
812 	}
813 
814 	/* need to make sure we have the right privileges */
815 	if ((ret = secpolicy_resource(credp)) != 0)
816 		return (ret);
817 	if ((ret = secpolicy_lock_memory(credp)) != 0)
818 		return (ret);
819 
820 	if (msg_printed == 0) {
821 		cmn_err(CE_NOTE, "!driver has been opened. This driver may "
822 		    "take out long term locks on pages which may impact "
823 		    "dynamic reconfiguration events");
824 		msg_printed = 1;
825 	}
826 
827 	return (0);
828 }
829 
830 /*ARGSUSED*/
831 static int
832 physmem_close(dev_t dev, int flag, int otyp, cred_t *credp)
833 {
834 	return (0);
835 }
836 
837 /*ARGSUSED*/
838 static int
839 physmem_getinfo(dev_info_t *dip, ddi_info_cmd_t infocmd,
840     void *arg, void **resultp)
841 {
842 	switch (infocmd) {
843 	case DDI_INFO_DEVT2DEVINFO:
844 		*resultp = physmem_dip;
845 		return (DDI_SUCCESS);
846 
847 	case DDI_INFO_DEVT2INSTANCE:
848 		*resultp = (void *)(ulong_t)getminor((dev_t)arg);
849 		return (DDI_SUCCESS);
850 
851 	default:
852 		return (DDI_FAILURE);
853 	}
854 }
855 
856 static int
857 physmem_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
858 {
859 	int i;
860 
861 	if (cmd == DDI_RESUME) {
862 		return (DDI_SUCCESS);
863 	}
864 
865 	if (cmd != DDI_ATTACH)
866 		return (DDI_FAILURE);
867 
868 	if (ddi_create_minor_node(dip, ddi_get_name(dip), S_IFCHR,
869 	    ddi_get_instance(dip), DDI_PSEUDO, 0) != DDI_SUCCESS)
870 		return (DDI_FAILURE);
871 
872 	physmem_dip = dip;
873 
874 	/* Initialize driver specific data */
875 	if (physmem_setup_vnops()) {
876 		ddi_remove_minor_node(dip, ddi_get_name(dip));
877 		return (DDI_FAILURE);
878 	}
879 
880 	for (i = 0; i < PPH_SIZE; i++)
881 		pph[i] = NULL;
882 
883 	page_capture_register_callback(PC_PHYSMEM, 10000,
884 	    map_page_proc);
885 
886 	return (DDI_SUCCESS);
887 }
888 
889 static int
890 physmem_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
891 {
892 	int ret = DDI_SUCCESS;
893 
894 	if (cmd == DDI_SUSPEND) {
895 		return (DDI_SUCCESS);
896 	}
897 
898 	if (cmd != DDI_DETACH)
899 		return (DDI_FAILURE);
900 
901 	ASSERT(physmem_dip == dip);
902 
903 	mutex_enter(&physmem_mutex);
904 	if (physmem_vnodecnt == 0) {
905 		if (physmem_vnodeops != NULL) {
906 			vn_freevnodeops(physmem_vnodeops);
907 			physmem_vnodeops = NULL;
908 			page_capture_unregister_callback(PC_PHYSMEM);
909 		}
910 	} else {
911 		ret = EBUSY;
912 	}
913 	mutex_exit(&physmem_mutex);
914 	if (ret == DDI_SUCCESS)
915 		ddi_remove_minor_node(dip, ddi_get_name(dip));
916 	return (ret);
917 }
918 
919 static struct cb_ops physmem_cb_ops = {
920 	physmem_open,	/* open */
921 	physmem_close,	/* close */
922 	nodev,		/* strategy */
923 	nodev,		/* print */
924 	nodev,		/* dump */
925 	nodev,		/* read */
926 	nodev,		/* write */
927 	physmem_ioctl,	/* ioctl */
928 	nodev,		/* devmap */
929 	nodev,		/* mmap */
930 	nodev,		/* segmap */
931 	nochpoll,	/* chpoll */
932 	ddi_prop_op,	/* prop_op */
933 	NULL,		/* cb_str */
934 	D_NEW | D_MP | D_DEVMAP,
935 	CB_REV,
936 	NULL,
937 	NULL
938 };
939 
940 static struct dev_ops physmem_ops = {
941 	DEVO_REV,
942 	0,
943 	physmem_getinfo,
944 	nulldev,
945 	nulldev,
946 	physmem_attach,
947 	physmem_detach,
948 	nodev,
949 	&physmem_cb_ops,
950 	NULL,
951 	NULL,
952 	ddi_quiesce_not_needed,		/* quiesce */
953 };
954 
955 static struct modldrv modldrv = {
956 	&mod_driverops,
957 	"physmem driver",
958 	&physmem_ops
959 };
960 
961 static struct modlinkage modlinkage = {
962 	MODREV_1,
963 	&modldrv,
964 	NULL
965 };
966 
967 int
968 _init(void)
969 {
970 	return (mod_install(&modlinkage));
971 }
972 
973 int
974 _info(struct modinfo *modinfop)
975 {
976 	return (mod_info(&modlinkage, modinfop));
977 }
978 
979 int
980 _fini(void)
981 {
982 	return (mod_remove(&modlinkage));
983 }
984