| /fs/btrfs/ |
| A D | locking.c | 222 if (!path->locks[i]) in btrfs_unlock_up_safe() 224 btrfs_tree_unlock_rw(path->nodes[i], path->locks[i]); in btrfs_unlock_up_safe() 225 path->locks[i] = 0; in btrfs_unlock_up_safe()
|
| A D | ctree.c | 168 if (p->locks[i]) { in btrfs_release_path() 170 p->locks[i] = 0; in btrfs_release_path() 927 path->locks[level] = 0; in balance_level() 1408 if (!path->locks[i]) in unlock_up() 1431 path->locks[i] = 0; in unlock_up() 1764 p->locks[level] = root_lock; in btrfs_search_slot_get_root() 2125 p->locks[u] = 0; in btrfs_search_slot() 2268 p->locks[level] = BTRFS_READ_LOCK; in btrfs_search_old_slot() 2323 p->locks[level] = BTRFS_READ_LOCK; in btrfs_search_old_slot() 4880 if (path->locks[level]) { in btrfs_next_old_leaf() [all …]
|
| A D | extent-tree.c | 5443 ASSERT(path->locks[level]); in walk_down_proc() 5464 path->locks[level] = 0; in walk_down_proc() 5471 ASSERT(path->locks[level]); in walk_down_proc() 5496 path->locks[level] = 0; in walk_down_proc() 5844 if (!path->locks[level]) { in walk_up_proc() 5856 path->locks[level] = 0; in walk_up_proc() 5867 path->locks[level] = 0; in walk_up_proc() 5901 if (!path->locks[level]) { in walk_up_proc() 6028 if (path->locks[level]) { in walk_up_tree() 6031 path->locks[level] = 0; in walk_up_tree() [all …]
|
| A D | ref-verify.c | 561 path->locks[level-1] = BTRFS_READ_LOCK; in walk_down_tree() 589 btrfs_tree_unlock_rw(path->nodes[l], path->locks[l]); in walk_up_tree() 593 path->locks[l] = 0; in walk_up_tree() 992 path->locks[level] = BTRFS_READ_LOCK; in btrfs_build_ref_tree()
|
| A D | qgroup.c | 2216 btrfs_tree_unlock_rw(eb, path->locks[level]); in adjust_slots_upwards() 2217 path->locks[level] = 0; in adjust_slots_upwards() 2321 src_path->locks[root_level] = 0; in qgroup_trace_extent_swap() 2344 src_path->locks[cur_level] = BTRFS_READ_LOCK; in qgroup_trace_extent_swap() 2474 dst_path->locks[cur_level] = BTRFS_READ_LOCK; in qgroup_trace_new_subtree_blocks() 2507 dst_path->locks[cur_level]); in qgroup_trace_new_subtree_blocks() 2511 dst_path->locks[cur_level] = 0; in qgroup_trace_new_subtree_blocks() 2554 dst_path->locks[level] = 0; in qgroup_trace_subtree_swap() 2646 path->locks[root_level] = 0; /* so release_path doesn't try to unlock */ in btrfs_qgroup_trace_subtree() 2672 path->locks[level] = BTRFS_READ_LOCK; in btrfs_qgroup_trace_subtree()
|
| A D | ctree.h | 62 u8 locks[BTRFS_MAX_LEVEL]; member
|
| A D | defrag.c | 505 ASSERT(path->locks[1] != 0); in btrfs_defrag_leaves()
|
| A D | backref.c | 2169 path->locks[0] = 0; in btrfs_ref_to_path()
|
| A D | file.c | 481 path->locks[0] == BTRFS_WRITE_LOCK && in btrfs_drop_extents()
|
| A D | relocation.c | 2267 path->locks[upper->level] = 0; in do_relocation()
|
| /fs/smb/client/ |
| A D | smb2file.c | 287 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) { in smb2_unlock_range() 329 &cfile->llist->locks); in smb2_unlock_range() 347 cifs_move_llist(&tmp_llist, &cfile->llist->locks); in smb2_unlock_range() 369 list_for_each_entry(li, &fdlocks->locks, llist) { in smb2_push_mand_fdlocks()
|
| A D | file.c | 637 if (!list_empty(&cur->locks)) { in cifs_has_mand_locks() 687 INIT_LIST_HEAD(&fdlocks->locks); in cifs_new_fileinfo() 769 list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) { in cifsFileInfo_put_final() 1545 list_for_each_entry(li, &fdlocks->locks, llist) { in cifs_find_fid_lock_conflict() 1635 list_add_tail(&lock->llist, &cfile->llist->locks); in cifs_lock_add() 1662 list_add_tail(&lock->llist, &cfile->llist->locks); in cifs_lock_add_if() 1787 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) { in cifs_push_mandatory_locks() 2162 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) { in cifs_unlock_range() 2203 &cfile->llist->locks); in cifs_unlock_range() 2221 &cfile->llist->locks); in cifs_unlock_range()
|
| A D | cifsglob.h | 1474 struct list_head locks; /* locks held by fid above */ member
|
| /fs/ceph/ |
| A D | Makefile | 8 ceph-y := super.o inode.o dir.o file.o locks.o addr.o ioctl.o \
|
| /fs/dlm/ |
| A D | Kconfig | 16 list of resource and locks the local node knows about.
|
| A D | dlm_internal.h | 718 struct list_head locks; member
|
| A D | lock.c | 5805 list_add_tail(&lkb->lkb_ownqueue, &ua->proc->locks); in dlm_user_request() 5937 list_add_tail(&lkb->lkb_ownqueue, &ua->proc->locks); in dlm_user_adopt_orphan() 6138 if (list_empty(&proc->locks)) in del_proc_lock() 6141 lkb = list_entry(proc->locks.next, struct dlm_lkb, lkb_ownqueue); in del_proc_lock() 6212 if (!list_empty(&proc->locks)) { in purge_proc_locks() 6213 lkb = list_entry(proc->locks.next, struct dlm_lkb, in purge_proc_locks()
|
| A D | user.c | 656 INIT_LIST_HEAD(&proc->locks); in device_open()
|
| /fs/ocfs2/ |
| A D | Makefile | 27 locks.o \
|
| /fs/jffs2/ |
| A D | README.Locking | 95 GC thread locks it, sends the signal, then unlocks it - while the GC 96 thread itself locks it, zeroes c->gc_task, then unlocks on the exit path.
|
| /fs/ |
| A D | Makefile | 35 obj-$(CONFIG_FILE_LOCKING) += locks.o
|
| /fs/bcachefs/ |
| A D | Kconfig | 90 bool "Optimistic spinning for six locks"
|
| /fs/smb/server/ |
| A D | smb2misc.c | 175 *off = offsetof(struct smb2_lock_req, locks); in smb2_get_data_area_len()
|
| /fs/affs/ |
| A D | Changes | 150 - Removed some unnecessary locks as Bill
|
| /fs/smb/common/ |
| A D | smb2pdu.h | 857 DECLARE_FLEX_ARRAY(struct smb2_lock_element, locks);
|