Searched refs:locks (Results 1 – 19 of 19) sorted by relevance
87 struct bpf_res_spin_lock *locks[48] = {}; in res_spin_lock_test_held_lock_max() local92 _Static_assert(ARRAY_SIZE(((struct rqspinlock_held){}).locks) == 31, in res_spin_lock_test_held_lock_max()104 locks[i] = &e->lock; in res_spin_lock_test_held_lock_max()116 locks[i] = &e->lock; in res_spin_lock_test_held_lock_max()121 if (bpf_res_spin_lock(locks[i])) in res_spin_lock_test_held_lock_max()128 ret = bpf_res_spin_lock(locks[34]); in res_spin_lock_test_held_lock_max()130 bpf_res_spin_unlock(locks[34]); in res_spin_lock_test_held_lock_max()139 bpf_res_spin_unlock(locks[i]); in res_spin_lock_test_held_lock_max()
26 } locks SEC(".maps");52 lock = bpf_map_lookup_elem(&locks, &key); in spin_lock_work()
98 locks_clock_cnt=`function_count locks clock`99 clock_locks_cnt=`function_count clock locks`167 check_cnt $locks_clock_cnt locks clock170 check_cnt $clock_locks_cnt clock locks
58 unmatched-locks = Rcu-lock \ domain(matched)60 and unmatched = unmatched-locks | unmatched-unlocks62 and unmatched-locks-to-unlocks =63 [unmatched-locks] ; po ; [unmatched-unlocks]64 and matched = matched | (unmatched-locks-to-unlocks \
17 * File "lock.cat" handles locks and is experimental.18 * It can be replaced by include "cos.cat" for tests that do not use locks.
83 empty ([UNMATCHED-LKW] ; loc ; [UNMATCHED-LKW]) \ id as unmatched-locks
6 * This test shows that write-write ordering provided by locks
8 * the critical sections are protected by different locks.
8 * critical sections are protected by different locks. Note: Even when a
80 --combine-locks::194 Note that RW-variant of locks have :R and :W suffix. Names without the230 Note that it will busy-wait after it gets the lock. Delaying locks can
55 u32 locks; /* count of 'lock' transactions */ member
660 if (lock & P(LOCK, LOCKED)) stats->locks++; in c2c_decode_stats()770 stats->locks += add->locks; in c2c_add_stats()
15 variables outside of their corresponding locks should continue reading.
531 as reader-writer locks or sequence locks.
45 adding new "little kernel locks".
499 Potentially infinite loops, such as those used to wait for locks to562 attention only to those executions in which both locks are actually
423 * Wait for pre-existing local locks. One at
2086 the two locking operations use different locks. Thus we cannot have
2373 fprintf(out, " Locked Load/Store Operations : %10d\n", stats->locks); in print_c2c__display_stats()2424 fprintf(out, " Locked Access on shared lines : %10d\n", stats->locks); in print_shared_cacheline_info()
Completed in 27 milliseconds