| /drivers/md/persistent-data/ |
| A D | dm-array.c | 26 __le32 max_entries; member 153 uint32_t max_entries, in alloc_ablock() argument 163 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock() 350 uint32_t max_entries, in insert_new_ablock() argument 407 unsigned int max_entries; member 524 resize->max_entries, in grow_add_tail_block() 765 size -= max_entries; in dm_array_new() 794 entry = index % max_entries; in dm_array_get_value() 813 unsigned int max_entries; in array_set_value() local 826 entry = index % max_entries; in array_set_value() [all …]
|
| A D | dm-btree-remove.c | 197 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local 200 if (max_entries != r_max_entries) { in shift() 205 if (nr_left - count > max_entries) { in shift() 210 if (nr_right + count > max_entries) { in shift() 316 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local 319 if (nr_left + shift > max_entries) { in delete_center_node() 357 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in redistribute3() local 363 BUG_ON(target_left > max_entries); in redistribute3() 364 BUG_ON(target_right > max_entries); in redistribute3() 431 if ((left->header.max_entries != center->header.max_entries) || in __rebalance3() [all …]
|
| A D | dm-btree-spine.c | 40 uint32_t flags, nr_entries, max_entries; in node_check() local 58 max_entries = le32_to_cpu(h->max_entries); in node_check() 62 (sizeof(__le64) + value_size) * max_entries > block_size) { in node_check() 67 if (nr_entries > max_entries) { in node_check()
|
| A D | dm-btree.c | 91 uint32_t max_entries = le32_to_cpu(node->header.max_entries); in insert_at() local 95 index >= max_entries || in insert_at() 96 nr_entries >= max_entries) { in insert_at() 135 uint32_t max_entries; in dm_btree_empty() local 148 n->header.max_entries = cpu_to_le32(max_entries); in dm_btree_empty() 673 rn->header.max_entries = ln->header.max_entries; in split_one_into_two() 773 mn->header.max_entries = ln->header.max_entries; in split_two_into_three() 863 ln->header.max_entries = pn->header.max_entries; in btree_split_beneath() 880 rn->header.max_entries = pn->header.max_entries; in btree_split_beneath() 889 pn->header.max_entries = cpu_to_le32( in btree_split_beneath() [all …]
|
| A D | dm-btree-internal.h | 35 __le32 max_entries; member 116 return &n->keys[le32_to_cpu(n->header.max_entries)]; in value_base()
|
| /drivers/acpi/ |
| A D | tables.c | 244 int proc_num, unsigned int max_entries) in acpi_table_parse_entries_array() argument 270 0, proc, proc_num, max_entries); in acpi_table_parse_entries_array() 279 void *arg, unsigned int max_entries) in __acpi_table_parse_entries() argument 289 max_entries); in __acpi_table_parse_entries() 305 unsigned int max_entries) in acpi_table_parse_entries() argument 308 NULL, NULL, max_entries); in acpi_table_parse_entries() 312 acpi_tbl_entry_handler handler, unsigned int max_entries) in acpi_table_parse_madt() argument 316 handler, max_entries); in acpi_table_parse_madt()
|
| /drivers/iommu/ |
| A D | io-pgtable-dart.c | 230 int ret = 0, tbl, num_entries, max_entries, map_idx_start; in dart_map_pages() local 268 max_entries = DART_PTES_PER_TABLE(data) - map_idx_start; in dart_map_pages() 269 num_entries = min_t(int, pgcount, max_entries); in dart_map_pages() 290 int i = 0, num_entries, max_entries, unmap_idx_start; in dart_unmap_pages() local 305 max_entries = DART_PTES_PER_TABLE(data) - unmap_idx_start; in dart_unmap_pages() 306 num_entries = min_t(int, pgcount, max_entries); in dart_unmap_pages()
|
| A D | io-pgtable-arm.c | 433 int ret = 0, num_entries, max_entries, map_idx_start; in __arm_lpae_map() local 441 max_entries = arm_lpae_max_entries(map_idx_start, data); in __arm_lpae_map() 442 num_entries = min_t(int, pgcount, max_entries); in __arm_lpae_map() 631 int i = 0, num_entries, max_entries, unmap_idx_start; in __arm_lpae_unmap() local 647 max_entries = arm_lpae_max_entries(unmap_idx_start, data); in __arm_lpae_unmap() 648 num_entries = min_t(int, pgcount, max_entries); in __arm_lpae_unmap() 810 int max_entries, ret; in __arm_lpae_iopte_walk() local 816 max_entries = ARM_LPAE_PGD_SIZE(data) / sizeof(arm_lpae_iopte); in __arm_lpae_iopte_walk() 818 max_entries = ARM_LPAE_PTES_PER_TABLE(data); in __arm_lpae_iopte_walk() 821 (idx < max_entries) && (walk_data->addr < walk_data->end); ++idx) { in __arm_lpae_iopte_walk()
|
| /drivers/infiniband/core/ |
| A D | uverbs_std_types_device.c | 321 int max_entries; in UVERBS_HANDLER() local 338 max_entries = uverbs_attr_ptr_get_array_size( in UVERBS_HANDLER() 341 if (max_entries <= 0) in UVERBS_HANDLER() 342 return max_entries ?: -EINVAL; in UVERBS_HANDLER() 349 entries = uverbs_kcalloc(attrs, max_entries, sizeof(*entries)); in UVERBS_HANDLER() 353 num_entries = rdma_query_gid_table(ib_dev, entries, max_entries); in UVERBS_HANDLER()
|
| /drivers/net/ethernet/broadcom/bnge/ |
| A D | bnge_rmem.c | 244 entries = clamp_t(u32, entries, ctxm->min_entries, ctxm->max_entries); in bnge_setup_ctxm_pg_tbls() 346 max_qps = ctxm->max_entries; in bnge_alloc_ctx_mem() 349 max_srqs = ctxm->max_entries; in bnge_alloc_ctx_mem() 379 rc = bnge_setup_ctxm_pg_tbls(bd, ctxm, ctxm->max_entries, 1); in bnge_alloc_ctx_mem() 384 rc = bnge_setup_ctxm_pg_tbls(bd, ctxm, ctxm->max_entries, 1); in bnge_alloc_ctx_mem() 395 num_mr = min_t(u32, ctxm->max_entries / 2, 1024 * 256); in bnge_alloc_ctx_mem()
|
| A D | bnge_resc.c | 195 u16 max_entries, pad; in bnge_set_dflt_rss_indir_tbl() local 199 max_entries = bnge_get_rxfh_indir_size(bd); in bnge_set_dflt_rss_indir_tbl() 202 for (i = 0; i < max_entries; i++) in bnge_set_dflt_rss_indir_tbl() 206 pad = bd->rss_indir_tbl_entries - max_entries; in bnge_set_dflt_rss_indir_tbl()
|
| /drivers/net/wireguard/ |
| A D | ratelimiter.c | 18 static unsigned int max_entries, table_size; variable 135 if (atomic_inc_return(&total_entries) > max_entries) in wg_ratelimiter_allow() 177 max_entries = table_size * 8; in wg_ratelimiter_init()
|
| /drivers/ufs/core/ |
| A D | ufs-mcq.c | 242 hwq->max_entries; in ufshcd_mcq_memory_alloc() 251 cqe_size = sizeof(struct cq_entry) * hwq->max_entries; in ufshcd_mcq_memory_alloc() 319 u32 entries = hwq->max_entries; in ufshcd_mcq_compl_all_cqes_lock() 364 qsize = hwq->max_entries * MCQ_ENTRY_SIZE_IN_DWORD - 1; in ufshcd_mcq_make_queues_operational() 477 hwq->max_entries = hba->nutrs + 1; in ufshcd_mcq_init() 654 if (sq_head_slot == hwq->max_entries) in ufshcd_mcq_sqe_search()
|
| /drivers/bus/ |
| A D | stm32_etzpc.c | 48 if (firewall_id >= ctrl->max_entries) { in stm32_etzpc_grant_access() 102 etzpc_controller->max_entries = nb_per + nb_master; in stm32_etzpc_probe()
|
| A D | stm32_rifsc.c | 114 if (firewall_id >= rifsc_controller->max_entries) { in stm32_rifsc_grant_access() 213 rifsc_controller->max_entries = nb_risup + nb_rimu + nb_risal; in stm32_rifsc_probe()
|
| A D | stm32_firewall.h | 47 unsigned int max_entries; member
|
| /drivers/gpu/drm/i915/gt/ |
| A D | intel_ggtt.c | 583 const int max_entries = ggtt_total_entries(ggtt) - first_entry; in gen8_ggtt_clear_range() local 586 if (WARN(num_entries > max_entries, in gen8_ggtt_clear_range() 588 first_entry, num_entries, max_entries)) in gen8_ggtt_clear_range() 589 num_entries = max_entries; in gen8_ggtt_clear_range() 602 const int max_entries = ggtt_total_entries(ggtt) - first_entry; in gen8_ggtt_scratch_range_bind() local 604 if (WARN(num_entries > max_entries, in gen8_ggtt_scratch_range_bind() 606 first_entry, num_entries, max_entries)) in gen8_ggtt_scratch_range_bind() 607 num_entries = max_entries; in gen8_ggtt_scratch_range_bind() 766 if (WARN(num_entries > max_entries, in gen6_ggtt_clear_range() 768 first_entry, num_entries, max_entries)) in gen6_ggtt_clear_range() [all …]
|
| A D | intel_sseu.h | 48 #define GEN_SSEU_STRIDE(max_entries) DIV_ROUND_UP(max_entries, BITS_PER_BYTE) argument
|
| /drivers/net/wireguard/selftest/ |
| A D | ratelimiter.c | 109 for (i = 0; i <= max_entries; ++i) { in capacity_test() 111 if (wg_ratelimiter_allow(skb4, &init_net) != (i != max_entries)) in capacity_test()
|
| /drivers/media/platform/renesas/vsp1/ |
| A D | vsp1_dl.c | 118 unsigned int max_entries; member 298 dlb->max_entries = num_entries; in vsp1_dl_body_pool_create() 389 if (WARN_ONCE(dlb->num_entries >= dlb->max_entries, in vsp1_dl_body_write() 390 "DLB size exceeded (max %u)", dlb->max_entries)) in vsp1_dl_body_write() 574 header_offset = dl->body0->max_entries * sizeof(*dl->body0->entries); in vsp1_dl_list_alloc()
|
| /drivers/net/ethernet/netronome/nfp/bpf/ |
| A D | offload.c | 356 offmap->map.max_entries) { in nfp_bpf_map_alloc() 358 offmap->map.max_entries, in nfp_bpf_map_alloc() 403 bpf->map_elems_in_use += offmap->map.max_entries; in nfp_bpf_map_alloc() 418 bpf->map_elems_in_use -= offmap->map.max_entries; in nfp_bpf_map_free()
|
| A D | fw.h | 82 __be32 max_entries; member
|
| /drivers/gpu/drm/ |
| A D | drm_prime.c | 1060 int max_entries) in drm_prime_sg_to_page_array() argument 1066 if (WARN_ON(p - pages >= max_entries)) in drm_prime_sg_to_page_array() 1086 int max_entries) in drm_prime_sg_to_dma_addr_array() argument 1092 if (WARN_ON(a - addrs >= max_entries)) in drm_prime_sg_to_dma_addr_array()
|
| /drivers/gpu/drm/i915/display/ |
| A D | intel_dmc.c | 1113 u32 num_entries, max_entries; in parse_dmc_fw_package() local 1120 max_entries = PACKAGE_MAX_FW_INFO_ENTRIES; in parse_dmc_fw_package() 1122 max_entries = PACKAGE_V2_MAX_FW_INFO_ENTRIES; in parse_dmc_fw_package() 1133 package_size += max_entries * sizeof(struct intel_fw_info); in parse_dmc_fw_package() 1144 if (WARN_ON(package_header->num_entries > max_entries)) in parse_dmc_fw_package() 1145 num_entries = max_entries; in parse_dmc_fw_package()
|
| /drivers/s390/scsi/ |
| A D | zfcp_fc.c | 735 struct zfcp_adapter *adapter, int max_entries) in zfcp_fc_eval_gpn_ft() argument 764 for (x = 1; x < max_entries && !last; x++) { in zfcp_fc_eval_gpn_ft() 814 int chain, max_entries, buf_num, max_bytes; in zfcp_fc_scan_ports() local 820 max_entries = chain ? ZFCP_FC_GPN_FT_MAX_ENT : ZFCP_FC_GPN_FT_ENT_PAGE; in zfcp_fc_scan_ports() 837 ret = zfcp_fc_eval_gpn_ft(fc_req, adapter, max_entries); in zfcp_fc_scan_ports()
|