| /fs/ocfs2/ |
| A D | move_extents.c | 733 if ((*alloc_size + *len_defraged) < threshold) { in ocfs2_calc_extent_defrag_len() 737 *len_defraged += *alloc_size; in ocfs2_calc_extent_defrag_len() 752 *alloc_size = threshold - *len_defraged; in ocfs2_calc_extent_defrag_len() 825 if (alloc_size > len_to_move) in __ocfs2_move_extents_range() 826 alloc_size = len_to_move; in __ocfs2_move_extents_range() 857 &alloc_size, flags); in __ocfs2_move_extents_range() 860 &new_phys_cpos, alloc_size, in __ocfs2_move_extents_range() 863 new_phys_cpos += alloc_size; in __ocfs2_move_extents_range() 871 context->clusters_moved += alloc_size; in __ocfs2_move_extents_range() 873 cpos += alloc_size; in __ocfs2_move_extents_range() [all …]
|
| A D | file.c | 1441 u32 cpos, phys_cpos, clusters, alloc_size; in ocfs2_allocate_unwritten_extents() local 1475 &alloc_size, NULL); in ocfs2_allocate_unwritten_extents() 1485 if (alloc_size > clusters) in ocfs2_allocate_unwritten_extents() 1486 alloc_size = clusters; in ocfs2_allocate_unwritten_extents() 1496 ret = ocfs2_extend_allocation(inode, cpos, alloc_size, 1); in ocfs2_allocate_unwritten_extents() 1504 cpos += alloc_size; in ocfs2_allocate_unwritten_extents() 1505 clusters -= alloc_size; in ocfs2_allocate_unwritten_extents()
|
| A D | xattr.c | 812 u32 trunc_len, cpos, phys_cpos, alloc_size; in ocfs2_xattr_shrink_size() local 822 &alloc_size, in ocfs2_xattr_shrink_size() 829 if (alloc_size > trunc_len) in ocfs2_xattr_shrink_size() 830 alloc_size = trunc_len; in ocfs2_xattr_shrink_size() 833 phys_cpos, alloc_size, in ocfs2_xattr_shrink_size() 842 block, alloc_size); in ocfs2_xattr_shrink_size() 843 cpos += alloc_size; in ocfs2_xattr_shrink_size() 844 trunc_len -= alloc_size; in ocfs2_xattr_shrink_size()
|
| /fs/ntfs3/ |
| A D | attrib.c | 1885 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_collapse_range() 1892 total_size = alloc_size; in attr_collapse_range() 2087 attr_b->nres.alloc_size = cpu_to_le64(alloc_size - bytes); in attr_collapse_range() 2153 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_punch_hole() 2156 if (vbo >= alloc_size) { in attr_punch_hole() 2164 if (bytes > alloc_size) in attr_punch_hole() 2165 bytes = alloc_size; in attr_punch_hole() 2361 alloc_size = data_size; in attr_insert_range() 2365 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_insert_range() 2423 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_insert_range() [all …]
|
| A D | record.c | 205 u64 data_size, alloc_size, tot_size; in mi_enum_attr() local 321 alloc_size = le64_to_cpu(attr->nres.alloc_size); in mi_enum_attr() 322 if (data_size > alloc_size) in mi_enum_attr() 326 if (alloc_size & t32) in mi_enum_attr() 339 if (tot_size > alloc_size) in mi_enum_attr() 345 if (alloc_size > mi->sbi->volume.size) in mi_enum_attr()
|
| A D | frecord.c | 498 alloc = le64_to_cpu(attr->nres.alloc_size) >> in ni_repack() 1447 attr->nres.alloc_size = in ni_insert_nonresident() 1449 attr->nres.data_size = attr->nres.alloc_size; in ni_insert_nonresident() 1869 u64 end, alloc_size; in ni_fiemap() local 1905 alloc_size = le64_to_cpu(attr->nres.alloc_size); in ni_fiemap() 1906 if (end > alloc_size) in ni_fiemap() 1907 end = alloc_size; in ni_fiemap() 3086 dup->alloc_size = 0; in ni_update_parent() 3094 dup->alloc_size = dup->data_size = 0; in ni_update_parent() 3105 dup->alloc_size = is_attr_ext(attr) ? in ni_update_parent() [all …]
|
| A D | ntfs.h | 352 __le64 alloc_size; // 0x28: The allocated size of attribute in bytes. member 397 le64_to_cpu(attr->nres.alloc_size)) in attr_ondisk_size() 561 __le64 alloc_size; // 0x20: Data attribute allocated size, multiple of cluster size. member
|
| A D | inode.c | 151 t64 = le64_to_cpu(attr->nres.alloc_size); in ntfs_read_mft() 267 if (!attr->nres.alloc_size) in ntfs_read_mft() 310 inode_set_bytes(inode, le64_to_cpu(attr->nres.alloc_size)); in ntfs_read_mft() 321 if (le64_to_cpu(attr->nres.alloc_size) >= 0x20000000) in ntfs_read_mft() 1375 fname->dup.alloc_size = fname->dup.data_size = 0; in ntfs_create_inode() 1530 attr->nres.alloc_size = in ntfs_create_inode()
|
| A D | fslog.c | 216 __le64 alloc_size; member 2725 dsize > le64_to_cpu(attr->nres.alloc_size)) { in check_attr() 3008 attr->nres.alloc_size = cpu_to_le64(ntfs_up_cluster(sbi, size)); in attr_create_nonres_log() 3372 attr->nres.alloc_size = new_sz->alloc_size; in do_action() 4859 if (size > le64_to_cpu(attr->nres.alloc_size)) { in log_replay() 4861 attr->nres.alloc_size = cpu_to_le64(size); in log_replay() 5091 attr->nres.alloc_size = in log_replay()
|
| A D | index.c | 1437 u32 alloc_size = ntfs_up_cluster(sbi, data_size); in indx_create_allocate() local 1438 CLST len = alloc_size >> sbi->cluster_bits; in indx_create_allocate() 1464 inode_set_bytes(&ni->vfs_inode, alloc_size); in indx_create_allocate()
|
| A D | fsntfs.c | 521 new_mft_total = le64_to_cpu(attr->nres.alloc_size) >> sbi->record_bits; in ntfs_extend_mft()
|
| /fs/smb/server/ |
| A D | xattr.h | 43 __u64 alloc_size; member
|
| A D | ndr.c | 213 ret = ndr_write_int64(n, da->alloc_size); in ndr_encode_dos_attr()
|
| A D | smb2pdu.c | 3529 loff_t alloc_size; in smb2_open() local 3538 alloc_size = le64_to_cpu(az_req->AllocationSize); in smb2_open() 3541 alloc_size); in smb2_open() 3544 alloc_size); in smb2_open()
|
| /fs/xfs/ |
| A D | xfs_log_cil.c | 278 int alloc_size; in xlog_cil_alloc_shadow_bufs() local 319 alloc_size = nbytes + xlog_cil_iovec_space(niovecs); in xlog_cil_alloc_shadow_bufs() 326 alloc_size > lip->li_lv_shadow->lv_alloc_size) { in xlog_cil_alloc_shadow_bufs() 335 lv = xlog_kvmalloc(alloc_size); in xlog_cil_alloc_shadow_bufs() 341 lv->lv_alloc_size = alloc_size; in xlog_cil_alloc_shadow_bufs()
|
| /fs/ |
| A D | select.c | 627 size_t size, alloc_size; in core_sys_select() local 657 alloc_size = 6 * size; in core_sys_select() 658 bits = kvmalloc(alloc_size, GFP_KERNEL); in core_sys_select()
|
| /fs/gfs2/ |
| A D | bmap.c | 943 u64 alloc_size; in __gfs2_iomap_get() local 949 alloc_size = len << inode->i_blkbits; in __gfs2_iomap_get() 950 if (alloc_size < iomap->length) in __gfs2_iomap_get() 951 iomap->length = alloc_size; in __gfs2_iomap_get()
|
| /fs/btrfs/ |
| A D | super.c | 1692 u64 alloc_size; in btrfs_calc_avail_data_space() local 1695 alloc_size = devices_info[i].max_avail; in btrfs_calc_avail_data_space() 1697 devices_info[j].max_avail -= alloc_size; in btrfs_calc_avail_data_space()
|
| A D | ioctl.c | 2969 int alloc_size; in btrfs_ioctl_space_info() local 3014 alloc_size = sizeof(*dest) * slot_count; in btrfs_ioctl_space_info() 3019 if (alloc_size > PAGE_SIZE) in btrfs_ioctl_space_info() 3023 dest = kmalloc(alloc_size, GFP_KERNEL); in btrfs_ioctl_space_info() 3079 if (copy_to_user(user_dest, dest_orig, alloc_size)) in btrfs_ioctl_space_info()
|
| A D | send.c | 8051 size_t alloc_size; in btrfs_ioctl_send() local 8191 alloc_size = array_size(sizeof(*arg->clone_sources), in btrfs_ioctl_send() 8195 clone_sources_tmp = kvmalloc(alloc_size, GFP_KERNEL); in btrfs_ioctl_send() 8202 alloc_size); in btrfs_ioctl_send()
|