| /drivers/gpu/drm/amd/display/dc/sspl/ |
| A D | dc_spl_scl_easf_filters.c | 2383 unsigned int num_entries) in spl_easf_get_scale_ratio_to_reg_value() argument 2430 unsigned int num_entries; in spl_get_reducer_gain6() local 2449 unsigned int num_entries; in spl_get_reducer_gain4() local 2468 unsigned int num_entries; in spl_get_gainRing6() local 2487 unsigned int num_entries; in spl_get_gainRing4() local 2506 unsigned int num_entries; in spl_get_3tap_dntilt_uptilt_offset() local 2520 unsigned int num_entries; in spl_get_3tap_uptilt_maxval() local 2534 unsigned int num_entries; in spl_get_3tap_dntilt_slope() local 2548 unsigned int num_entries; in spl_get_3tap_uptilt1_slope() local 2562 unsigned int num_entries; in spl_get_3tap_uptilt2_slope() local [all …]
|
| /drivers/gpu/drm/i915/display/ |
| A D | intel_ddi_buf_trans.c | 32 .num_entries = ARRAY_SIZE(_hsw_trans_dp), 49 .num_entries = ARRAY_SIZE(_hsw_trans_fdi), 70 .num_entries = ARRAY_SIZE(_hsw_trans_hdmi), 88 .num_entries = ARRAY_SIZE(_bdw_trans_edp), 105 .num_entries = ARRAY_SIZE(_bdw_trans_dp), 122 .num_entries = ARRAY_SIZE(_bdw_trans_fdi), 141 .num_entries = ARRAY_SIZE(_bdw_trans_hdmi), 160 .num_entries = ARRAY_SIZE(_skl_trans_dp), 214 .num_entries = ARRAY_SIZE(_kbl_trans_dp), 377 .num_entries = ARRAY_SIZE(_bxt_trans_dp), [all …]
|
| /drivers/char/agp/ |
| A D | generic.c | 328 num_entries = A_SIZE_8(temp)->num_entries; in agp_num_entries() 331 num_entries = A_SIZE_16(temp)->num_entries; in agp_num_entries() 334 num_entries = A_SIZE_32(temp)->num_entries; in agp_num_entries() 337 num_entries = A_SIZE_LVL2(temp)->num_entries; in agp_num_entries() 340 num_entries = A_SIZE_FIX(temp)->num_entries; in agp_num_entries() 880 num_entries = A_SIZE_16(temp)->num_entries; in agp_generic_create_gatt_table() 1045 num_entries = A_SIZE_8(temp)->num_entries; in agp_generic_insert_memory() 1048 num_entries = A_SIZE_16(temp)->num_entries; in agp_generic_insert_memory() 1051 num_entries = A_SIZE_32(temp)->num_entries; in agp_generic_insert_memory() 1054 num_entries = A_SIZE_FIX(temp)->num_entries; in agp_generic_insert_memory() [all …]
|
| A D | efficeon-agp.c | 198 int num_entries, l1_pages; in efficeon_create_gatt_table() local 200 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_create_gatt_table() 202 printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries); in efficeon_create_gatt_table() 205 BUG_ON(num_entries & 0x3ff); in efficeon_create_gatt_table() 206 l1_pages = num_entries >> 10; in efficeon_create_gatt_table() 238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() local 245 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_insert_memory() 246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() local 291 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_remove_memory() [all …]
|
| A D | alpha-agp.c | 88 int num_entries, status; in alpha_core_agp_insert_memory() local 95 num_entries = A_SIZE_FIX(temp)->num_entries; in alpha_core_agp_insert_memory() 96 if ((pg_start + mem->page_count) > num_entries) in alpha_core_agp_insert_memory() 169 aper_size->num_entries = agp->aperture.size / PAGE_SIZE; in alpha_core_agp_setup() 170 aper_size->page_order = __ffs(aper_size->num_entries / 1024); in alpha_core_agp_setup()
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn321/ |
| A D | dcn321_fpu.c | 211 unsigned int *num_entries, in dcn321_insert_entry_into_table_sorted() argument 219 if (*num_entries == 0) { in dcn321_insert_entry_into_table_sorted() 221 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 225 if (index >= *num_entries) in dcn321_insert_entry_into_table_sorted() 229 for (i = *num_entries; i > index; i--) in dcn321_insert_entry_into_table_sorted() 233 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 242 if (*num_entries == 0) in remove_entry_from_table_at_index() 424 *num_entries = 0; in build_synthetic_soc_states() 492 for (i = *num_entries - 1; i >= 0 ; i--) { in build_synthetic_soc_states() 522 for (i = *num_entries - 1; i >= 0 ; i--) { in build_synthetic_soc_states() [all …]
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn351/ |
| A D | dcn351_fpu.c | 278 ASSERT(clk_table->num_entries); in dcn351_update_bw_bounding_box_fpu() 350 if (clk_table->num_entries) in dcn351_update_bw_bounding_box_fpu() 387 if (clk_table->num_entries > 2) { in dcn351_update_bw_bounding_box_fpu() 391 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 408 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 410 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 412 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 414 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 416 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() 418 clk_table->num_entries; in dcn351_update_bw_bounding_box_fpu() [all …]
|
| /drivers/gpu/drm/amd/display/dc/dml/dcn35/ |
| A D | dcn35_fpu.c | 244 ASSERT(clk_table->num_entries); in dcn35_update_bw_bounding_box_fpu() 316 if (clk_table->num_entries) in dcn35_update_bw_bounding_box_fpu() 353 if (clk_table->num_entries > 2) { in dcn35_update_bw_bounding_box_fpu() 357 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 375 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 377 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 379 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 381 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 383 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 385 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() [all …]
|
| /drivers/net/ethernet/netronome/nfp/ |
| A D | nfp_shared_buf.c | 81 unsigned int i, num_entries, entry_sz; in nfp_shared_buf_register() local 92 num_entries = n; in nfp_shared_buf_register() 95 num_entries * sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 100 entry_sz = nfp_cpp_area_size(sb_desc_area) / num_entries; in nfp_shared_buf_register() 102 pf->shared_bufs = kmalloc_array(num_entries, sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 109 for (i = 0; i < num_entries; i++) { in nfp_shared_buf_register() 125 pf->num_shared_bufs = num_entries; in nfp_shared_buf_register()
|
| /drivers/net/wwan/iosm/ |
| A D | iosm_ipc_coredump.c | 66 u32 byte_read, num_entries, file_size; in ipc_coredump_get_list() local 95 num_entries = le32_to_cpu(cd_table->list.num_entries); in ipc_coredump_get_list() 96 if (num_entries == 0 || num_entries > IOSM_NOF_CD_REGION) { in ipc_coredump_get_list() 101 for (i = 0; i < num_entries; i++) { in ipc_coredump_get_list()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_bo_list.h | 53 unsigned num_entries; member 59 struct amdgpu_bo_list_entry entries[] __counted_by(num_entries); 71 size_t num_entries, 76 e != &list->entries[list->num_entries]; \ 81 e != &list->entries[list->num_entries]; \
|
| A D | amdgpu_bo_list.c | 72 size_t num_entries, struct amdgpu_bo_list **result) in amdgpu_bo_list_create() argument 74 unsigned last_entry = 0, first_userptr = num_entries; in amdgpu_bo_list_create() 81 list = kvzalloc(struct_size(list, entries, num_entries), GFP_KERNEL); in amdgpu_bo_list_create() 87 list->num_entries = num_entries; in amdgpu_bo_list_create() 90 for (i = 0; i < num_entries; ++i) { in amdgpu_bo_list_create() 136 trace_amdgpu_cs_bo_status(list->num_entries, total_size); in amdgpu_bo_list_create() 145 for (i = first_userptr; i < num_entries; ++i) in amdgpu_bo_list_create()
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_irq.c | 18 pf->irq_tracker.num_entries = max_vectors; in ice_init_irq_tracker() 24 ice_init_virt_irq_tracker(struct ice_pf *pf, u32 base, u32 num_entries) in ice_init_virt_irq_tracker() argument 26 pf->virt_irq_tracker.bm = bitmap_zalloc(num_entries, GFP_KERNEL); in ice_init_virt_irq_tracker() 30 pf->virt_irq_tracker.num_entries = num_entries; in ice_init_virt_irq_tracker() 77 struct xa_limit limit = { .max = pf->irq_tracker.num_entries - 1, in ice_get_irq_res() 253 pf->virt_irq_tracker.num_entries, in ice_virt_get_irqs() 256 if (res >= pf->virt_irq_tracker.num_entries) in ice_virt_get_irqs()
|
| /drivers/net/dsa/sja1105/ |
| A D | sja1105_tas.c | 172 int num_entries = 0; in sja1105_init_scheduling() local 213 num_entries += tas_data->offload[port]->num_entries; in sja1105_init_scheduling() 219 num_entries += gating_cfg->num_entries; in sja1105_init_scheduling() 235 table->entry_count = num_entries; in sja1105_init_scheduling() 283 schedule_end_idx = k + offload->num_entries - 1; in sja1105_init_scheduling() 310 for (i = 0; i < offload->num_entries; i++, k++) { in sja1105_init_scheduling() 425 i < offload->num_entries; in sja1105_tas_check_conflicts() 431 j < admin->num_entries; in sja1105_tas_check_conflicts() 470 size_t num_entries = gating_cfg->num_entries; in sja1105_gating_check_conflicts() local 486 dummy->num_entries = num_entries; in sja1105_gating_check_conflicts() [all …]
|
| /drivers/net/ethernet/engleder/ |
| A D | tsnep_selftests.c | 381 qopt->num_entries = 7; in tsnep_test_taprio() 405 qopt->num_entries = 8; in tsnep_test_taprio() 468 qopt->num_entries = 2; in tsnep_test_taprio_change() 501 qopt->num_entries = 3; in tsnep_test_taprio_change() 513 qopt->num_entries = 2; in tsnep_test_taprio_change() 527 qopt->num_entries = 4; in tsnep_test_taprio_change() 539 qopt->num_entries = 2; in tsnep_test_taprio_change() 551 qopt->num_entries = 3; in tsnep_test_taprio_change() 567 qopt->num_entries = 4; in tsnep_test_taprio_change() 583 qopt->num_entries = 5; in tsnep_test_taprio_change() [all …]
|
| /drivers/fwctl/pds/ |
| A D | main.c | 114 num_endpoints = le32_to_cpu(pdsfc->endpoints->num_entries); in pdsfc_free_endpoints() 131 num_endpoints = le32_to_cpu(pdsfc->endpoints->num_entries); in pdsfc_free_operations() 219 u32 num_entries; in pdsfc_get_operations() local 252 num_entries = le32_to_cpu(data->num_entries); in pdsfc_get_operations() 253 dev_dbg(dev, "num_entries %d\n", num_entries); in pdsfc_get_operations() 254 for (i = 0; i < num_entries; i++) { in pdsfc_get_operations() 287 u32 num_entries; in pdsfc_validate_rpc() local 305 num_entries = le32_to_cpu(pdsfc->endpoints->num_entries); in pdsfc_validate_rpc() 306 for (i = 0; i < num_entries; i++) { in pdsfc_validate_rpc() 335 num_entries = le32_to_cpu(ep_info->operations->num_entries); in pdsfc_validate_rpc() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_pt.c | 985 u32 num_entries) in xe_pt_cancel_bind() argument 1130 u32 num_entries) in xe_pt_free_bind() argument 1146 *num_entries = 0; in xe_pt_prepare_bind() 1163 num_entries); in xe_vm_dbg_print_entries() 1760 u32 num_entries) in xe_pt_abort_unbind() argument 1786 u32 num_entries) in xe_pt_commit_prepare_unbind() argument 1876 pt_op->num_entries, true); in bind_op_prepare() 2012 if (pt_op->num_entries > 1) in xe_pt_op_check_range_skip_invalidation() 2058 pt_op->num_entries); in unbind_range_prepare() 2612 pt_op->num_entries, in xe_pt_update_ops_abort() [all …]
|
| /drivers/iommu/ |
| A D | io-pgtable-arm-v7s.c | 328 for (i = 0; i < num_entries; i++) in __arm_v7s_set_pte() 331 __arm_v7s_pte_sync(ptep, num_entries, cfg); in __arm_v7s_set_pte() 402 for (i = 0; i < num_entries; i++) in arm_v7s_init_pte() 422 if (num_entries > 1) in arm_v7s_init_pte() 427 __arm_v7s_set_pte(ptep, pte, num_entries, cfg); in arm_v7s_init_pte() 466 int num_entries = size >> ARM_V7S_LVL_SHIFT(lvl); in __arm_v7s_map() local 472 if (num_entries) in __arm_v7s_map() 474 lvl, num_entries, ptep); in __arm_v7s_map() 576 } while (++i < num_entries); in __arm_v7s_unmap() 594 if (num_entries) { in __arm_v7s_unmap() [all …]
|
| A D | io-pgtable-dart.c | 112 dart_iopte prot, int num_entries, in dart_init_pte() argument 119 for (i = 0; i < num_entries; i++) in dart_init_pte() 132 for (i = 0; i < num_entries; i++) in dart_init_pte() 230 int ret = 0, tbl, num_entries, max_entries, map_idx_start; in dart_map_pages() local 269 num_entries = min_t(int, pgcount, max_entries); in dart_map_pages() 271 ret = dart_init_pte(data, iova, paddr, prot, num_entries, ptep); in dart_map_pages() 273 *mapped += num_entries * pgsize; in dart_map_pages() 290 int i = 0, num_entries, max_entries, unmap_idx_start; in dart_unmap_pages() local 306 num_entries = min_t(int, pgcount, max_entries); in dart_unmap_pages() 308 while (i < num_entries) { in dart_unmap_pages()
|
| /drivers/soc/qcom/ |
| A D | smsm.c | 83 u32 num_entries; member 484 u32 num_entries; in smsm_get_size_info() member 495 smsm->num_entries = SMSM_DEFAULT_NUM_ENTRIES; in smsm_get_size_info() 500 smsm->num_entries = info->num_entries; in smsm_get_size_info() 505 smsm->num_entries, smsm->num_hosts); in smsm_get_size_info() 533 smsm->num_entries, in qcom_smsm_probe() 576 smsm->num_entries * sizeof(u32)); in qcom_smsm_probe() 590 size = smsm->num_entries * smsm->num_hosts * sizeof(u32); in qcom_smsm_probe() 622 if (ret || id >= smsm->num_entries) { in qcom_smsm_probe() 649 for (id = 0; id < smsm->num_entries; id++) in qcom_smsm_probe() [all …]
|
| /drivers/gpu/drm/amd/display/dc/dml2/dml21/src/dml2_mcg/ |
| A D | dml2_mcg_dcn4.c | 67 min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values; in build_min_clk_table_fine_grained() 70 for (i = min_table->dram_bw_table.num_entries - 1; i > 0; i--) { in build_min_clk_table_fine_grained() 82 for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) { in build_min_clk_table_fine_grained() 106 for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) { in build_min_clk_table_fine_grained() 109 min_table->dram_bw_table.num_entries = i; in build_min_clk_table_fine_grained() 115 for (i = 0; i < (int)min_table->dram_bw_table.num_entries - 1; i++) { in build_min_clk_table_fine_grained() 121 for (j = i + 1; j < min_table->dram_bw_table.num_entries; j++) { in build_min_clk_table_fine_grained() 126 min_table->dram_bw_table.num_entries--; in build_min_clk_table_fine_grained() 142 min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values; in build_min_clk_table_coarse_grained()
|
| /drivers/pci/ |
| A D | tph.c | 428 int num_entries, i, offset; in pci_restore_tph_state() local 447 num_entries = pcie_tph_get_st_table_size(pdev); in pci_restore_tph_state() 448 for (i = 0; i < num_entries; i++) { in pci_restore_tph_state() 458 int num_entries, i, offset; in pci_save_tph_state() local 479 num_entries = pcie_tph_get_st_table_size(pdev); in pci_save_tph_state() 480 for (i = 0; i < num_entries; i++) { in pci_save_tph_state() 496 int num_entries; in pci_tph_init() local 503 num_entries = pcie_tph_get_st_table_size(pdev); in pci_tph_init() 504 save_size = sizeof(u32) + num_entries * sizeof(u16); in pci_tph_init()
|
| /drivers/parisc/ |
| A D | iosapic.c | 259 unsigned long num_entries = 0UL; in iosapic_load_irt() local 266 status = pdc_pat_get_irt_size(&num_entries, cell_num); in iosapic_load_irt() 270 BUG_ON(num_entries == 0); in iosapic_load_irt() 278 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 299 status = pdc_pci_irt_size(&num_entries, 0); in iosapic_load_irt() 307 BUG_ON(num_entries == 0); in iosapic_load_irt() 309 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 317 status = pdc_pci_irt(num_entries, 0, table); in iosapic_load_irt() 332 num_entries, in iosapic_load_irt() 335 for (i = 0 ; i < num_entries ; i++, p++) { in iosapic_load_irt() [all …]
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_nve.c | 65 unsigned int num_entries; member 311 WARN_ON(mc_record->num_entries); in mlxsw_sp_nve_mc_record_destroy() 323 unsigned int num_entries = mc_record->num_entries; in mlxsw_sp_nve_mc_record_get() local 337 if (mc_record->num_entries != 0) in mlxsw_sp_nve_mc_record_put() 368 unsigned int num_entries = 0; in mlxsw_sp_nve_mc_record_refresh() local 393 num_entries++); in mlxsw_sp_nve_mc_record_refresh() 396 WARN_ON(num_entries != mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 449 mc_record->num_entries++; in mlxsw_sp_nve_mc_record_ip_add() 472 mc_record->num_entries--; in mlxsw_sp_nve_mc_record_ip_add() 484 mc_record->num_entries--; in mlxsw_sp_nve_mc_record_entry_del() [all …]
|
| /drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
| A D | smu7_smumgr.c | 340 toc->num_entries = 0; in smu7_request_smu_load_fw() 344 UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 347 UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 350 UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 353 UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 356 UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 359 UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 362 UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 365 UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 368 UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() [all …]
|