| /linux/fs/ntfs3/ |
| A D | run.c | 49 if (vcn < r->vcn) { in run_lookup() 54 if (vcn < r->vcn + r->len) { in run_lookup() 65 if (vcn >= r->vcn) { in run_lookup() 74 if (vcn < r->vcn) { in run_lookup() 211 gap = vcn - r->vcn; in run_lookup_entry() 236 if (vcn > r->vcn) { in run_truncate_head() 239 r->vcn = vcn; in run_truncate_head() 276 r->len = vcn - r->vcn; in run_truncate() 414 r->vcn = vcn; in run_add_entry() 502 if (vcn > r->vcn) { in run_collapse_range() [all …]
|
| A D | attrib.c | 99 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 105 vcn ? *vcn : svcn, Add2Ptr(attr, run_off), in attr_load_runs() 206 vcn += flen; in attr_allocate_clusters() 472 if (svcn <= vcn && vcn <= evcn) { in attr_set_size() 877 if (le_b && (vcn < svcn || evcn1 <= vcn)) { in attr_data_get_block() 1154 if (evcn < vcn || vcn < svcn) { in attr_load_runs_vcn() 1181 for (vcn = from >> cluster_bits; vcn <= vcn_last; vcn += clen) { in attr_load_runs_range() 1184 vcn); in attr_load_runs_range() 1477 if (svcn <= vcn && vcn < evcn1) { in attr_allocate_frame() 1747 if (svcn <= vcn && vcn < evcn1) { in attr_collapse_range() [all …]
|
| A D | attrlist.c | 172 u8 name_len, const CLST *vcn) in al_find_ex() argument 191 le_vcn = le64_to_cpu(le->vcn); in al_find_ex() 206 if (!vcn) in al_find_ex() 209 if (*vcn == le_vcn) in al_find_ex() 212 if (*vcn < le_vcn) in al_find_ex() 229 u8 name_len, CLST vcn) in al_find_le_to_insert() argument 244 if (!le->vcn) { in al_find_le_to_insert() 258 if (le64_to_cpu(le->vcn) >= vcn) in al_find_le_to_insert() 318 le->vcn = cpu_to_le64(svcn); in al_add_le() 382 le = al_find_ex(ni, NULL, type, name, name_len, &vcn); in al_delete_le() [all …]
|
| A D | file.c | 120 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local 124 vcn = pos >> bits; in ntfs_extend_initialized_size() 132 loff_t vbo = (loff_t)vcn << bits; in ntfs_extend_initialized_size() 285 u64 vbo = (u64)vcn << sbi->cluster_bits; in ntfs_sparse_cluster() 385 CLST vcn = from >> sbi->cluster_bits; in ntfs_file_mmap() local 389 for (; vcn < end; vcn += len) { in ntfs_file_mmap() 671 CLST vcn = vbo >> sbi->cluster_bits; in ntfs_fallocate() local 682 for (; vcn < cend; vcn += clen) { in ntfs_fallocate() 683 err = attr_data_get_block(ni, vcn, cend - vcn, in ntfs_fallocate() 687 if (!new || vcn >= vcn_v) in ntfs_fallocate() [all …]
|
| A D | ntfs_fs.h | 413 struct runs_tree *run, const CLST *vcn); 415 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, 432 CLST vcn); 458 u8 name_len, const CLST *vcn); 513 u8 name_len, const CLST *vcn, 519 const __le16 *name, u8 name_len, CLST vcn, 776 void run_truncate(struct runs_tree *run, CLST vcn); 777 void run_truncate_head(struct runs_tree *run, CLST vcn); 795 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, 1127 struct runs_tree *run, CLST vcn) in attr_load_runs_attr() argument [all …]
|
| A D | frecord.c | 223 if (vcn && *vcn) in ni_find_attr() 225 } else if (!vcn) { in ni_find_attr() 304 if (vcn) { in ni_load_attr() 307 if (!next || le64_to_cpu(next->vcn) > vcn) in ni_load_attr() 676 if (le->vcn) in ni_try_remove_attr_list() 804 le->vcn = 0; in ni_create_attr_list() 1909 vcn = vcn_next; in ni_fiemap() 1940 vcn += clen; in ni_fiemap() 1965 vcn += clen; in ni_fiemap() 2162 for (vcn = vbo >> sbi->cluster_bits; vcn < end; vcn += clen) { in ni_decompress_file() [all …]
|
| /linux/fs/ntfs/ |
| A D | runlist.c | 151 if ((dst->vcn + dst->length) != src->vcn) in ntfs_are_rl_mergeable() 682 drl[ds].vcn = drl[ds - 1].vcn + in ntfs_runlists_merge() 804 rl[rlpos].vcn = vcn; in ntfs_mapping_pairs_decompress() 924 rl[rlpos].vcn = vcn; in ntfs_mapping_pairs_decompress() 944 rl[rlpos].vcn = vcn; in ntfs_mapping_pairs_decompress() 1004 if (unlikely(vcn < rl[0].vcn)) in ntfs_rl_vcn_to_lcn() 1008 if (unlikely(vcn < rl[i+1].vcn)) { in ntfs_rl_vcn_to_lcn() 1010 return rl[i].lcn + (vcn - rl[i].vcn); in ntfs_rl_vcn_to_lcn() 1042 if (unlikely(!rl || vcn < rl[0].vcn)) in ntfs_rl_find_vcn_nolock() 1045 if (unlikely(vcn < rl[1].vcn)) { in ntfs_rl_find_vcn_nolock() [all …]
|
| A D | index.c | 108 VCN vcn, old_vcn; in ntfs_index_lookup() local 264 page = ntfs_map_page(ia_mapping, vcn << in ntfs_index_lookup() 276 ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << in ntfs_index_lookup() 288 (long long)vcn, idx_ni->mft_no); in ntfs_index_lookup() 297 (unsigned long long)vcn, idx_ni->mft_no); in ntfs_index_lookup() 305 "driver bug.", (unsigned long long)vcn, in ntfs_index_lookup() 316 "driver.", (unsigned long long)vcn, in ntfs_index_lookup() 324 (unsigned long long)vcn, idx_ni->mft_no); in ntfs_index_lookup() 408 old_vcn = vcn; in ntfs_index_lookup() 410 if (vcn >= 0) { in ntfs_index_lookup() [all …]
|
| A D | dir.c | 80 VCN vcn, old_vcn; in ntfs_lookup_inode_by_name() local 308 page = ntfs_map_page(ia_mapping, vcn << in ntfs_lookup_inode_by_name() 360 "driver.", (unsigned long long)vcn, in ntfs_lookup_inode_by_name() 542 old_vcn = vcn; in ntfs_lookup_inode_by_name() 545 if (vcn >= 0) { in ntfs_lookup_inode_by_name() 549 PAGE_SHIFT == vcn << in ntfs_lookup_inode_by_name() 634 VCN vcn, old_vcn; 838 "driver.", (unsigned long long)vcn, 952 old_vcn = vcn; 954 if (vcn >= 0) { [all …]
|
| A D | logfile.c | 714 VCN vcn, end_vcn; in ntfs_empty_logfile() local 738 vcn = 0; in ntfs_empty_logfile() 746 if (unlikely(!rl || vcn < rl->vcn || !rl->length)) { in ntfs_empty_logfile() 748 err = ntfs_map_runlist_nolock(log_ni, vcn, NULL); in ntfs_empty_logfile() 755 BUG_ON(!rl || vcn < rl->vcn || !rl->length); in ntfs_empty_logfile() 758 while (rl->length && vcn >= rl[1].vcn) in ntfs_empty_logfile() 771 vcn = rl->vcn; in ntfs_empty_logfile() 782 if (rl[1].vcn > end_vcn) in ntfs_empty_logfile() 783 len = end_vcn - rl->vcn; in ntfs_empty_logfile() 819 } while ((++rl)->vcn < end_vcn); in ntfs_empty_logfile()
|
| A D | aops.c | 167 VCN vcn; in ntfs_read_block() local 253 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_block() 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 283 err = ntfs_map_runlist(ni, vcn); in ntfs_read_block() 532 VCN vcn; in ntfs_write_block() local 701 vcn >>= vol->cluster_size_bits; in ntfs_write_block() 709 while (rl->length && rl[1].vcn <= vcn) in ntfs_write_block() 761 err = ntfs_map_runlist(ni, vcn); in ntfs_write_block() 1002 VCN vcn; in ntfs_write_mst_block() local 1010 vcn >>= vol->cluster_size_bits; in ntfs_write_mst_block() [all …]
|
| A D | attrib.h | 49 extern int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, 51 extern int ntfs_map_runlist(ntfs_inode *ni, VCN vcn); 53 extern LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, 57 const VCN vcn, ntfs_attr_search_ctx *ctx);
|
| A D | attrib.c | 84 (unsigned long long)vcn); in ntfs_map_runlist_nolock() 125 <= vcn && end_vcn >= vcn)) in ntfs_map_runlist_nolock() 169 if (unlikely(vcn && vcn >= end_vcn)) { in ntfs_map_runlist_nolock() 339 BUG_ON(vcn < 0); in ntfs_attr_vcn_to_lcn_nolock() 462 BUG_ON(vcn < 0); in ntfs_attr_find_vcn_nolock() 473 if (likely(rl && vcn >= rl[0].vcn)) { in ntfs_attr_find_vcn_nolock() 475 if (unlikely(vcn < rl[1].vcn)) { in ntfs_attr_find_vcn_nolock() 732 (unsigned long long)rl->vcn, in load_attribute_list() 1893 VCN vcn; in ntfs_attr_extend_allocation() local 2014 vcn = 0; in ntfs_attr_extend_allocation() [all …]
|
| A D | compress.c | 478 VCN vcn; in ntfs_read_compressed_block() local 596 for (vcn = start_vcn, start_vcn += cb_clusters; vcn < start_vcn; in ntfs_read_compressed_block() 597 vcn++) { in ntfs_read_compressed_block() 607 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_compressed_block() 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 613 (unsigned long long)vcn, in ntfs_read_compressed_block() 630 if (!ntfs_map_runlist(ni, vcn)) in ntfs_read_compressed_block() 730 if (vcn == start_vcn - cb_clusters) { in ntfs_read_compressed_block() 773 } else if (vcn == start_vcn) { in ntfs_read_compressed_block()
|
| A D | runlist.h | 29 VCN vcn; /* vcn = Starting virtual cluster number. */ member 65 extern LCN ntfs_rl_vcn_to_lcn(const runlist_element *rl, const VCN vcn); 70 const VCN vcn);
|
| A D | lcnalloc.c | 388 rl[rlpos].vcn = rl[rlpos - 1].vcn + in ntfs_cluster_alloc() 393 rl[rlpos].vcn = start_vcn; in ntfs_cluster_alloc() 726 rl[rlpos].vcn = rl[rlpos - 1].vcn + rl[rlpos - 1].length; in ntfs_cluster_alloc() 883 delta = start_vcn - rl->vcn; in __ntfs_cluster_free() 916 VCN vcn; in __ntfs_cluster_free() local 919 vcn = rl->vcn; in __ntfs_cluster_free() 920 rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx); in __ntfs_cluster_free()
|
| A D | mft.c | 525 VCN vcn; in ntfs_sync_mft_mirror() local 546 while (rl->length && rl[1].vcn <= vcn) in ntfs_sync_mft_mirror() 718 VCN vcn; in write_mft_record_nolock() local 734 while (rl->length && rl[1].vcn <= vcn) in write_mft_record_nolock() 1344 rl[1].vcn++; in ntfs_mft_bitmap_extend_allocation_nolock() 1409 if (ll >= rl2->vcn) in ntfs_mft_bitmap_extend_allocation_nolock() 1412 BUG_ON(ll < rl2->vcn); in ntfs_mft_bitmap_extend_allocation_nolock() 1518 rl[1].vcn--; in ntfs_mft_bitmap_extend_allocation_nolock() 1775 old_last_vcn = rl[1].vcn; in ntfs_mft_data_extend_allocation_nolock() 1841 if (ll >= rl2->vcn) in ntfs_mft_data_extend_allocation_nolock() [all …]
|
| A D | file.c | 626 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write() 704 cdelta = bh_cpos - vcn; in ntfs_prepare_pages_for_non_resident_write() 843 vcn = bh_cpos; in ntfs_prepare_pages_for_non_resident_write() 844 vcn_len = rl[1].vcn - vcn; in ntfs_prepare_pages_for_non_resident_write() 1157 vcn = bh_cpos; in ntfs_prepare_pages_for_non_resident_write() 1166 if (likely(vcn + vcn_len >= cend)) { in ntfs_prepare_pages_for_non_resident_write() 1302 vcn, highest_vcn, NULL)) { in ntfs_prepare_pages_for_non_resident_write() 1774 VCN vcn; in ntfs_perform_write() local 1784 vcn = pos >> vol->cluster_size_bits; in ntfs_perform_write() 1785 if (vcn != last_vcn) { in ntfs_perform_write() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_vcn.c | 103 adev->vcn.indirect_sram = true; in amdgpu_vcn_sw_init() 183 release_firmware(adev->vcn.fw); in amdgpu_vcn_sw_init() 184 adev->vcn.fw = NULL; in amdgpu_vcn_sw_init() 229 &adev->vcn.inst[i].gpu_addr, &adev->vcn.inst[i].cpu_addr); in amdgpu_vcn_sw_init() 235 adev->vcn.inst[i].fw_shared_cpu_addr = adev->vcn.inst[i].cpu_addr + in amdgpu_vcn_sw_init() 237 adev->vcn.inst[i].fw_shared_gpu_addr = adev->vcn.inst[i].gpu_addr + in amdgpu_vcn_sw_init() 240 if (adev->vcn.indirect_sram) { in amdgpu_vcn_sw_init() 243 &adev->vcn.inst[i].dpg_sram_gpu_addr, &adev->vcn.inst[i].dpg_sram_cpu_addr); in amdgpu_vcn_sw_init() 262 if (adev->vcn.indirect_sram) { in amdgpu_vcn_sw_fini() 279 release_firmware(adev->vcn.fw); in amdgpu_vcn_sw_fini() [all …]
|
| A D | vcn_v2_5.c | 80 adev->vcn.num_vcn_inst = 2; in vcn_v2_5_early_init() 81 adev->vcn.harvest_config = 0; in vcn_v2_5_early_init() 82 adev->vcn.num_enc_rings = 1; in vcn_v2_5_early_init() 90 adev->vcn.harvest_config |= 1 << i; in vcn_v2_5_early_init() 97 adev->vcn.num_enc_rings = 2; in vcn_v2_5_early_init() 171 ring = &adev->vcn.inst[j].ring_dec; in vcn_v2_5_sw_init() 193 &adev->vcn.inst[j].irq, 0, in vcn_v2_5_sw_init() 278 ring = &adev->vcn.inst[j].ring_dec; in vcn_v2_5_hw_init() 1795 if(state == adev->vcn.cur_state) in vcn_v2_5_set_powergating_state() 1804 adev->vcn.cur_state = state; in vcn_v2_5_set_powergating_state() [all …]
|
| A D | vcn_v2_0.c | 73 adev->vcn.num_enc_rings = 1; in vcn_v2_0_early_init() 75 adev->vcn.num_enc_rings = 2; in vcn_v2_0_early_init() 101 &adev->vcn.inst->irq); in vcn_v2_0_sw_init() 109 &adev->vcn.inst->irq); in vcn_v2_0_sw_init() 124 ring = &adev->vcn.inst->ring_dec; in vcn_v2_0_sw_init() 156 ring = &adev->vcn.inst->ring_enc[i]; in vcn_v2_0_sw_init() 237 ring = &adev->vcn.inst->ring_enc[i]; in vcn_v2_0_hw_init() 800 adev->vcn.inst->dpg_sram_curr_addr = (uint32_t *)adev->vcn.inst->dpg_sram_cpu_addr; in vcn_v2_0_start_dpg_mode() 1765 if (state == adev->vcn.cur_state) in vcn_v2_0_set_powergating_state() 1774 adev->vcn.cur_state = state; in vcn_v2_0_set_powergating_state() [all …]
|
| A D | vcn_v3_0.c | 88 adev->vcn.harvest_config = 0; in vcn_v3_0_early_init() 89 adev->vcn.num_enc_rings = 1; in vcn_v3_0_early_init() 98 adev->vcn.num_enc_rings = 0; in vcn_v3_0_early_init() 100 adev->vcn.num_enc_rings = 2; in vcn_v3_0_early_init() 179 ring = &adev->vcn.inst[i].ring_dec; in vcn_v3_0_sw_init() 295 ring = &adev->vcn.inst[i].ring_dec; in vcn_v3_0_hw_init() 324 ring = &adev->vcn.inst[i].ring_dec; in vcn_v3_0_hw_init() 2095 if (adev->vcn.num_enc_rings > 0) in vcn_v3_0_set_enc_ring_funcs() 2171 if(state == adev->vcn.cur_state) in vcn_v3_0_set_powergating_state() 2180 adev->vcn.cur_state = state; in vcn_v3_0_set_powergating_state() [all …]
|
| A D | vcn_v1_0.c | 69 adev->vcn.num_enc_rings = 2; in vcn_v1_0_early_init() 102 &adev->vcn.inst->irq); in vcn_v1_0_sw_init() 120 ring = &adev->vcn.inst->ring_dec; in vcn_v1_0_sw_init() 127 adev->vcn.internal.scratch9 = adev->vcn.inst->external.scratch9 = in vcn_v1_0_sw_init() 129 adev->vcn.internal.data0 = adev->vcn.inst->external.data0 = in vcn_v1_0_sw_init() 131 adev->vcn.internal.data1 = adev->vcn.inst->external.data1 = in vcn_v1_0_sw_init() 133 adev->vcn.internal.cmd = adev->vcn.inst->external.cmd = in vcn_v1_0_sw_init() 135 adev->vcn.internal.nop = adev->vcn.inst->external.nop = in vcn_v1_0_sw_init() 1769 if(state == adev->vcn.cur_state) in vcn_v1_0_set_powergating_state() 1778 adev->vcn.cur_state = state; in vcn_v1_0_set_powergating_state() [all …]
|
| A D | vega10_reg_init.c | 83 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_DOORBELL64_VCN0_1; in vega10_doorbell_index_init() 84 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_DOORBELL64_VCN2_3; in vega10_doorbell_index_init() 85 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_DOORBELL64_VCN4_5; in vega10_doorbell_index_init() 86 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_DOORBELL64_VCN6_7; in vega10_doorbell_index_init()
|
| A D | vega20_reg_init.c | 89 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_VEGA20_DOORBELL64_VCN0_1; in vega20_doorbell_index_init() 90 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_VEGA20_DOORBELL64_VCN2_3; in vega20_doorbell_index_init() 91 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_VEGA20_DOORBELL64_VCN4_5; in vega20_doorbell_index_init() 92 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_VEGA20_DOORBELL64_VCN6_7; in vega20_doorbell_index_init()
|