lk 307 kern/dev/generic/console.c struct lock *lk;
lk 312 kern/dev/generic/console.c lk = con_userlock_read;
lk 315 kern/dev/generic/console.c lk = con_userlock_write;
lk 318 kern/dev/generic/console.c KASSERT(lk != NULL);
lk 319 kern/dev/generic/console.c lock_acquire(lk);
lk 329 kern/dev/generic/console.c lock_release(lk);
lk 339 kern/dev/generic/console.c lock_release(lk);
lk 348 kern/dev/generic/console.c lock_release(lk);
lk 79 kern/include/spinlock.h void spinlock_init(struct spinlock *lk);
lk 80 kern/include/spinlock.h void spinlock_cleanup(struct spinlock *lk);
lk 82 kern/include/spinlock.h void spinlock_acquire(struct spinlock *lk);
lk 83 kern/include/spinlock.h void spinlock_release(struct spinlock *lk);
lk 85 kern/include/spinlock.h bool spinlock_do_i_hold(struct spinlock *lk);
lk 49 kern/thread/spinlock.c spinlock_init(struct spinlock *lk)
lk 51 kern/thread/spinlock.c spinlock_data_set(&lk->lk_lock, 0);
lk 52 kern/thread/spinlock.c lk->lk_holder = NULL;
lk 59 kern/thread/spinlock.c spinlock_cleanup(struct spinlock *lk)
lk 61 kern/thread/spinlock.c KASSERT(lk->lk_holder == NULL);
lk 62 kern/thread/spinlock.c KASSERT(spinlock_data_get(&lk->lk_lock) == 0);
lk 73 kern/thread/spinlock.c spinlock_acquire(struct spinlock *lk)
lk 82 kern/thread/spinlock.c if (lk->lk_holder == mycpu) {
lk 83 kern/thread/spinlock.c panic("Deadlock on spinlock %p\n", lk);
lk 101 kern/thread/spinlock.c if (spinlock_data_get(&lk->lk_lock) != 0) {
lk 104 kern/thread/spinlock.c if (spinlock_data_testandset(&lk->lk_lock) != 0) {
lk 110 kern/thread/spinlock.c lk->lk_holder = mycpu;
lk 117 kern/thread/spinlock.c spinlock_release(struct spinlock *lk)
lk 121 kern/thread/spinlock.c KASSERT(lk->lk_holder == curcpu->c_self);
lk 124 kern/thread/spinlock.c lk->lk_holder = NULL;
lk 125 kern/thread/spinlock.c spinlock_data_set(&lk->lk_lock, 0);
lk 133 kern/thread/spinlock.c spinlock_do_i_hold(struct spinlock *lk)
lk 140 kern/thread/spinlock.c return (lk->lk_holder == curcpu->c_self);