| /fs/btrfs/ |
| A D | ctree.h | 652 const struct btrfs_item_batch *batch); 660 struct btrfs_item_batch batch; in btrfs_insert_empty_item() local 662 batch.keys = key; in btrfs_insert_empty_item() 663 batch.data_sizes = &data_size; in btrfs_insert_empty_item() 664 batch.total_data_size = data_size; in btrfs_insert_empty_item() 665 batch.nr = 1; in btrfs_insert_empty_item() 667 return btrfs_insert_empty_items(trans, root, path, &batch); in btrfs_insert_empty_item()
|
| A D | extent_io.c | 1969 batch->ebs[batch->nr++] = eb; in eb_batch_add() 1975 batch->nr = 0; in eb_batch_init() 1976 batch->cur = 0; in eb_batch_init() 1981 if (batch->cur >= batch->nr) in eb_batch_next() 1983 return batch->ebs[batch->cur++]; in eb_batch_next() 1990 eb_batch_init(batch); in eb_batch_release() 2043 return batch->nr; in buffer_tree_get_ebs_tag() 2156 struct eb_batch batch; in btrfs_btree_wait_writeback_range() local 2160 eb_batch_init(&batch); in btrfs_btree_wait_writeback_range() 2185 struct eb_batch batch; in btree_write_cache_pages() local [all …]
|
| A D | delayed-inode.c | 640 struct btrfs_item_batch batch; in btrfs_insert_delayed_item() local 671 batch.total_data_size = first_data_size; in btrfs_insert_delayed_item() 672 batch.nr = 1; in btrfs_insert_delayed_item() 697 batch.nr++; in btrfs_insert_delayed_item() 699 batch.total_data_size += next->data_len; in btrfs_insert_delayed_item() 703 if (batch.nr == 1) { in btrfs_insert_delayed_item() 707 batch.keys = &first_key; in btrfs_insert_delayed_item() 708 batch.data_sizes = &first_data_size; in btrfs_insert_delayed_item() 714 ins_data = kmalloc(batch.nr * sizeof(u32) + in btrfs_insert_delayed_item() 722 batch.keys = ins_keys; in btrfs_insert_delayed_item() [all …]
|
| A D | tree-log.c | 3654 batch.nr = count; in flush_dir_items_batch() 3659 batch.keys = &key; in flush_dir_items_batch() 4437 batch.keys = ins_keys; in copy_items() 4440 batch.nr = 0; in copy_items() 4535 batch.nr++; in copy_items() 4543 if (batch.nr == 0) in copy_items() 6143 batch.keys = ins_keys; in log_delayed_insertion_items() 6152 &batch, first); in log_delayed_insertion_items() 6156 batch.nr = 0; in log_delayed_insertion_items() 6168 batch.nr++; in log_delayed_insertion_items() [all …]
|
| A D | ctree.c | 4157 const struct btrfs_item_batch *batch) in setup_items_for_insert() argument 4184 total_size = batch->total_data_size + (batch->nr * sizeof(struct btrfs_item)); in setup_items_for_insert() 4212 ioff - batch->total_data_size); in setup_items_for_insert() 4224 for (i = 0; i < batch->nr; i++) { in setup_items_for_insert() 4227 data_end -= batch->data_sizes[i]; in setup_items_for_insert() 4256 struct btrfs_item_batch batch; in btrfs_setup_item_for_insert() local 4258 batch.keys = key; in btrfs_setup_item_for_insert() 4259 batch.data_sizes = &data_size; in btrfs_setup_item_for_insert() 4260 batch.total_data_size = data_size; in btrfs_setup_item_for_insert() 4261 batch.nr = 1; in btrfs_setup_item_for_insert() [all …]
|
| A D | compression.c | 1120 compr_pool.shrinker->batch = 32; in btrfs_init_compress()
|
| A D | inode.c | 6411 struct btrfs_item_batch batch; in btrfs_create_new_inode() local 6522 batch.keys = &key[0]; in btrfs_create_new_inode() 6523 batch.data_sizes = &sizes[0]; in btrfs_create_new_inode() 6524 batch.total_data_size = sizes[0] + (args->orphan ? 0 : sizes[1]); in btrfs_create_new_inode() 6525 batch.nr = args->orphan ? 1 : 2; in btrfs_create_new_inode() 6526 ret = btrfs_insert_empty_items(trans, root, path, &batch); in btrfs_create_new_inode()
|
| /fs/xfs/ |
| A D | xfs_discard.c | 183 int batch = XFS_DISCARD_MAX_EXAMINE; in xfs_trim_gather_extents() local 246 if (--batch <= 0) { in xfs_trim_gather_extents() 632 int batch; member 649 if (--tr->batch <= 0) { in xfs_trim_gather_rtgroup_extent() 714 tr.batch = XFS_DISCARD_MAX_EXAMINE; in xfs_trim_rtgroup_extents()
|
| A D | xfs_icache.c | 1766 struct xfs_inode *batch[XFS_LOOKUP_BATCH]; in xfs_icwalk_ag() local 1773 (void **) batch, first_index, in xfs_icwalk_ag() 1786 struct xfs_inode *ip = batch[i]; in xfs_icwalk_ag() 1789 batch[i] = NULL; in xfs_icwalk_ag() 1814 if (!batch[i]) in xfs_icwalk_ag() 1816 error = xfs_icwalk_process_inode(goal, batch[i], pag, in xfs_icwalk_ag() 2346 mp->m_inodegc_shrinker->batch = XFS_INODEGC_SHRINKER_BATCH; in xfs_inodegc_register_shrinker()
|
| A D | xfs_mount.c | 1456 s32 batch; in xfs_dec_freecounter() local 1470 batch = 1; in xfs_dec_freecounter() 1472 batch = XFS_FDBLOCKS_BATCH; in xfs_dec_freecounter() 1485 percpu_counter_add_batch(&counter->count, -((int64_t)delta), batch); in xfs_dec_freecounter()
|
| A D | xfs_mount.h | 741 enum xfs_free_counter ctr, s64 rhs, s32 batch) in xfs_compare_freecounter() argument 743 return __percpu_counter_compare(&mp->m_free[ctr].count, rhs, batch); in xfs_compare_freecounter()
|
| A D | xfs_qm.c | 73 struct xfs_dquot *batch[XFS_DQ_LOOKUP_BATCH]; in xfs_qm_dquot_walk() local 78 nr_found = radix_tree_gang_lookup(tree, (void **)batch, in xfs_qm_dquot_walk() 86 struct xfs_dquot *dqp = batch[i]; in xfs_qm_dquot_walk() 90 error = execute(batch[i], data); in xfs_qm_dquot_walk()
|
| /fs/nfs/ |
| A D | nfs42xattr.c | 983 scan_objects_cb scan, long batch, int seeks) in nfs4_xattr_shrinker_init() argument 999 (*shrinker)->batch = batch; in nfs4_xattr_shrinker_init()
|
| /fs/bcachefs/ |
| A D | btree_key_cache.c | 842 shrink->batch = 1 << 14; in bch2_fs_btree_key_cache_init()
|
| /fs/f2fs/ |
| A D | node.c | 2718 unsigned int batch = SHRINK_NID_BATCH_SIZE; in f2fs_try_to_free_nids() local 2722 if (!nr_shrink || !batch || in f2fs_try_to_free_nids() 2728 batch--; in f2fs_try_to_free_nids()
|
| /fs/ |
| A D | super.c | 386 s->s_shrink->batch = 1024; in alloc_super()
|