| /fs/ext4/ |
| A D | extents.c | 3412 if (allocated) { in ext4_split_extent() 3496 *allocated = 0; in ext4_ext_convert_to_initialized() 3541 *allocated = map_len; in ext4_ext_convert_to_initialized() 3587 *allocated = map_len; in ext4_ext_convert_to_initialized() 3590 if (*allocated) { in ext4_ext_convert_to_initialized() 3744 allocated); in ext4_split_convert_extents() 3883 map->m_len = *allocated; in convert_initialized_extent() 3901 *allocated); in ext4_ext_handle_unwritten_extents() 4390 if (ar.len > allocated) in ext4_ext_map_blocks() 4391 ar.len = allocated; in ext4_ext_map_blocks() [all …]
|
| /fs/ |
| A D | kernel_read_file.c | 41 void *allocated = NULL; in kernel_read_file() local 80 *buf = allocated = vmalloc(i_size); in kernel_read_file() 115 if (allocated) { in kernel_read_file()
|
| /fs/ntfs3/ |
| A D | run.c | 256 run->allocated = 0; in run_truncate_head() 293 run->allocated = 0; in run_truncate() 370 if (run->allocated < used + sizeof(struct ntfs_run)) { in run_add_entry() 378 if (is_power_of_2(run->allocated)) in run_add_entry() 379 bytes = run->allocated << 1; in run_add_entry() 384 bytes = run->allocated + (16 * PAGE_SIZE); in run_add_entry() 402 run->allocated = bytes; in run_add_entry() 1177 if (bytes > new_run->allocated) { in run_clone() 1185 new_run->allocated = bytes; in run_clone()
|
| A D | ntfs_fs.h | 120 size_t allocated; /* Currently allocated ntfs_run storage size. */ member 851 size_t flags, size_t *allocated); 928 run->allocated = 0; in run_init()
|
| /fs/vboxsf/ |
| A D | utils.c | 53 s64 allocated; in vboxsf_init_inode() local 125 allocated = info->allocated + 511; in vboxsf_init_inode() 126 do_div(allocated, 512); in vboxsf_init_inode() 127 inode->i_blocks = allocated; in vboxsf_init_inode()
|
| A D | shfl_hostintf.h | 265 s64 allocated; member
|
| /fs/udf/ |
| A D | balloc.c | 693 int allocated; in udf_prealloc_blocks() local 696 allocated = udf_bitmap_prealloc_blocks(sb, in udf_prealloc_blocks() 701 allocated = udf_table_prealloc_blocks(sb, in udf_prealloc_blocks() 708 if (inode && allocated > 0) in udf_prealloc_blocks() 709 inode_add_bytes(inode, allocated << sb->s_blocksize_bits); in udf_prealloc_blocks() 710 return allocated; in udf_prealloc_blocks()
|
| /fs/xfs/libxfs/ |
| A D | xfs_bmap_btree.c | 201 dst->bc_bmap.allocated += src->bc_bmap.allocated; in xfs_bmbt_update_cursor() 204 src->bc_bmap.allocated = 0; in xfs_bmbt_update_cursor() 259 cur->bc_bmap.allocated++; in xfs_bmbt_alloc_block() 680 cur->bc_bmap.allocated = 0; in xfs_bmbt_init_cursor()
|
| A D | xfs_ialloc.c | 2846 unsigned int *allocated) in xfs_ialloc_count_ondisk() argument 2882 *allocated = ret; in xfs_ialloc_count_ondisk() 2896 unsigned int allocated; in xfs_ialloc_has_inodes_at_extent() local 2902 error = xfs_ialloc_count_ondisk(cur, agino, last_agino, &allocated); in xfs_ialloc_has_inodes_at_extent() 2906 if (allocated == 0) in xfs_ialloc_has_inodes_at_extent() 2908 else if (allocated == last_agino - agino + 1) in xfs_ialloc_has_inodes_at_extent()
|
| A D | xfs_bmap.c | 697 cur->bc_bmap.allocated++; in xfs_bmap_extents_to_btree() 916 cur->bc_bmap.allocated = 0; in xfs_bmap_add_attrfork_btree() 944 cur->bc_bmap.allocated = 0; in xfs_bmap_add_attrfork_extents() 1777 (bma->cur ? bma->cur->bc_bmap.allocated : 0)); in xfs_bmap_add_extent_delay_real() 1868 (bma->cur ? bma->cur->bc_bmap.allocated : 0)); in xfs_bmap_add_extent_delay_real() 1992 da_new += bma->cur->bc_bmap.allocated; in xfs_bmap_add_extent_delay_real() 1993 bma->cur->bc_bmap.allocated = 0; in xfs_bmap_add_extent_delay_real() 2559 cur->bc_bmap.allocated = 0; in xfs_bmap_add_extent_unwritten_real() 2809 cur->bc_bmap.allocated = 0; in xfs_bmap_add_extent_hole_real() 5470 cur->bc_bmap.allocated = 0; in __xfs_bunmapi() [all …]
|
| /fs/xfs/scrub/ |
| A D | alloc_repair.c | 440 unsigned int allocated = 0; in xrep_abt_reserve_space() local 467 if (allocated >= required) in xrep_abt_reserve_space() 470 desired = required - allocated; in xrep_abt_reserve_space() 493 allocated += len; in xrep_abt_reserve_space()
|
| /fs/bcachefs/ |
| A D | btree_journal_iter_types.h | 14 bool allocated:1; member
|
| A D | recovery.c | 292 if (!k->allocated) { in bch2_journal_replay_key() 370 if (!(k->k->k.type == KEY_TYPE_accounting && !k->allocated)) in bch2_journal_replay() 402 if (k->allocated) in bch2_journal_replay() 411 (!k->allocated ? BCH_TRANS_COMMIT_no_journal_res : 0), in bch2_journal_replay() 443 (!k->allocated in bch2_journal_replay()
|
| A D | fs-io-pagecache.h | 44 x(allocated)
|
| A D | ec.c | 1336 BUG_ON(!s->allocated); in ec_stripe_create() 1441 BUG_ON(!s->allocated && !s->err); in ec_stripe_new_set_pending() 1627 if (h->s && !h->s->allocated && dev_mask_nr(&devs_leaving)) in ec_stripe_head_devs_update() 1659 h->s->allocated && in bch2_ec_stripe_head_put() 2038 if (s->allocated) in bch2_ec_stripe_head_get() 2039 goto allocated; in bch2_ec_stripe_head_get() 2097 s->allocated = true; in bch2_ec_stripe_head_get() 2098 allocated: in bch2_ec_stripe_head_get()
|
| A D | btree_journal_iter.c | 267 .allocated = true, in bch2_journal_key_insert_take() 281 if (keys->data[idx].allocated) in bch2_journal_key_insert_take() 678 if (i->allocated) in bch2_journal_keys_put()
|
| A D | util.c | 956 unsigned i, nr, allocated = 1 << 12; 957 u16 *test_array = kmalloc_array(allocated, sizeof(test_array[0]), GFP_KERNEL); 959 for (nr = 1; nr < allocated; nr++) {
|
| A D | ec.h | 211 bool allocated; member
|
| /fs/affs/ |
| A D | Changes | 26 last allocated block is not cleared. Well, 148 when it allocated a new block. 262 - Extension block caches are now allocated on
|
| /fs/xfs/ |
| A D | xfs_zone_alloc.c | 757 xfs_rgblock_t allocated; in xfs_zone_alloc_blocks() local 766 allocated = oz->oz_allocated; in xfs_zone_alloc_blocks() 770 trace_xfs_zone_alloc_blocks(oz, allocated, count_fsb); in xfs_zone_alloc_blocks() 775 *sector += XFS_FSB_TO_BB(mp, allocated); in xfs_zone_alloc_blocks()
|
| /fs/ext2/ |
| A D | balloc.c | 1308 goto allocated; in ext2_new_blocks() 1351 goto allocated; in ext2_new_blocks() 1370 allocated: in ext2_new_blocks()
|
| /fs/ubifs/ |
| A D | file.c | 815 int err = 0, allocated = 0; in ubifs_bulk_read() local 856 allocated = 1; in ubifs_bulk_read() 864 if (!allocated) in ubifs_bulk_read()
|
| A D | journal.c | 861 int dlen = COMPRESSED_DATA_NODE_BUF_SZ, allocated = 1; in ubifs_jnl_write_data() local 885 allocated = 0; in ubifs_jnl_write_data() 947 if (!allocated) in ubifs_jnl_write_data() 959 if (!allocated) in ubifs_jnl_write_data()
|
| /fs/btrfs/ |
| A D | extent_io.c | 596 unsigned int allocated; in btrfs_alloc_page_array() local 598 for (allocated = 0; allocated < nr_pages;) { in btrfs_alloc_page_array() 599 unsigned int last = allocated; in btrfs_alloc_page_array() 601 allocated = alloc_pages_bulk(gfp, nr_pages, page_array); in btrfs_alloc_page_array() 602 if (unlikely(allocated == last)) { in btrfs_alloc_page_array() 604 for (int i = 0; i < allocated; i++) { in btrfs_alloc_page_array()
|
| /fs/proc/ |
| A D | Kconfig | 72 If the elfcore hdr is allocated and prepared by the dump kernel
|