Searched refs:locks (Results 1 - 16 of 16) sorted by relevance

H A Dwhatlock.d30 @locks["adaptive"] = count();
36 @locks["spin"] = count();
H A Duucp35 /usr/bin/rm -rf /usr/spool/locks/* >/dev/null 2>&1 &
H A Dunlocked_paths.pl42 my $locks = quotemeta $2;
46 if (!($locks =~ /$lock/)) {
H A Drrwlock.c46 * then grab a write lock without first dropping all read locks, and
55 * to enter multiple and unique rrwlock_ts for read locks at the same time.
311 * A reader-mostly lock implementation, tuning above reader-writer locks
314 * The idea is to split single busy lock into array of locks, so that
317 * Writer same time has to sequentially aquire write on all the locks.
329 rrw_init(&rrl->locks[i], track_all);
338 rrw_destroy(&rrl->locks[i]);
363 rrw_enter_read(&rrl->locks[RRM_TD_LOCK()], tag);
372 rrw_enter_write(&rrl->locks[i]);
380 if (rrl->locks[
H A Dspa_misc.c67 * There are three basic locks for managing spa_t structures:
157 * To read the configuration, it suffices to hold one of these locks as reader.
158 * To modify the configuration, you must hold all locks as writer. To modify
162 * We use these distinct config locks to avoid recursive lock entry.
167 * The spa config locks cannot be normal rwlocks because we need the
204 * other locks, and lower than all of them, to ensure that it's safe
440 spa_config_tryenter(spa_t *spa, int locks, void *tag, krw_t rw) argument
444 if (!(locks & (1 << i)))
450 spa_config_exit(spa, locks & ((1 << i) - 1),
458 spa_config_exit(spa, locks
471 spa_config_enter(spa_t *spa, int locks, void *tag, krw_t rw) argument
504 spa_config_exit(spa_t *spa, int locks, void *tag) argument
523 spa_config_held(spa_t *spa, int locks, krw_t rw) argument
1329 int locks = SCL_STATE_ALL | oplocks; local
H A Dmetaslab.c420 * Must hold one of the spa_config locks.
905 int locks = spa_config_held(spa, SCL_ALL, RW_WRITER); local
926 * config locks. However, the metaslab group's taskq might be trying
928 * lower locks to allow the I/O to complete. At a minimum,
932 spa_config_exit(spa, locks & ~(SCL_ZIO - 1), spa);
934 spa_config_enter(spa, locks & ~(SCL_ZIO - 1), spa, RW_WRITER);
1241 * racy since we can't hold the locks for all metaslab
2016 * cannot do these assertions because we hold spa config locks and
H A Drrwlock.h85 * A reader-mostly lock implementation, tuning above reader-writer locks
93 rrwlock_t locks[RRM_NUM_LOCKS]; member in struct:rrmlock
H A Dspa.h912 /* Pool configuration locks */
913 extern int spa_config_tryenter(spa_t *spa, int locks, void *tag, krw_t rw);
914 extern void spa_config_enter(spa_t *spa, int locks, void *tag, krw_t rw);
915 extern void spa_config_exit(spa_t *spa, int locks, void *tag);
916 extern int spa_config_held(spa_t *spa, int locks, krw_t rw);
H A Dcheck_locking.c434 static struct tracker_list *locks; variable in typeref:struct:tracker_list
695 add_tracker(&locks, my_id, name, sym);
706 add_tracker(&locks, my_id, name, sym);
720 add_tracker(&locks, my_id, name, sym);
743 add_tracker(&locks, my_id, name, sym);
1073 FOR_EACH_PTR(locks, tracker) {
H A Duudemon.cleanup113 LOCKS=/var/spool/locks # needs a comment in parms.h on USRSPOOLOCKS
H A DMakefile78 ROOTSPLOCKS = $(ROOTSPOOL)/locks
H A Derrlist85 46 No record locks available
H A Dtypegraph.c2965 foundlock_t *locks; local
2973 locks = mdb_zalloc(size, UM_SLEEP);
2976 bcopy(fl->fl_locks, locks, osize);
2980 fl->fl_locks = locks;
3058 * We have our type. Now iterate for locks. Note that we don't yet
3059 * deal with locks in flexible array members.
H A Dnfs4_recovery.c717 * There must be a corresponding call to nfs4_end_op() to free up any locks
2195 * Reclaim all the active (acquired) locks for the given file.
2209 locklist_t *locks, *llp; local
2225 * and mark its remaining locks (if any) as belonging to NOPID, so
2230 locks = flk_active_locks_for_vp(vp);
2231 for (llp = locks; llp != NULL; llp = llp->ll_next) {
2289 if (locks != NULL)
2290 flk_free_locklist(locks);
2478 * Reopen all the files for the given filesystem and reclaim any locks.
3287 * locks tha
H A Dnfs4_vnops.c1990 * (in the absence of locks).
2284 * We could attempt to clean up locks, except we're sure
2285 * that the current process didn't acquire any locks on
2304 * release all of the SYSV style record locks. Otherwise,
2306 * of the network locks. All of the locks held by this
10515 * Don't allow concurrent locks and mapping if mandatory locking is
10922 * locks, this is so that the process gets the latest bits from the
10926 * allow unlocks to succeed, to avoid having orphan locks on the
11901 * structure without requiring any other locks
15844 locklist_t *locks, *llp, *ri_llp, *tmp_llp; local
H A Dfd.c194 int locks);
1229 * If the locks parameter is nonzero, then all mutexes, semaphores and
1236 fd_cleanup(dev_info_t *dip, struct fdctlr *fdc, int hard, int locks) argument
1329 if (locks) {
1363 * Similarly, the locks parameter is also set to 1.

Completed in 126 milliseconds