| /drivers/firmware/efi/ |
| A D | unaccepted_memory.c | 36 unsigned long range_start, range_end; in accept_memory() local 127 for_each_set_bitrange_from(range_start, range_end, unaccepted->bitmap, in accept_memory() 130 unsigned long len = range_end - range_start; in accept_memory() 133 phys_end = range_end * unit_size + unaccepted->phys_base; in accept_memory()
|
| /drivers/firmware/efi/libstub/ |
| A D | unaccepted_memory.c | 182 unsigned long range_start, range_end; in accept_memory() local 212 for_each_set_bitrange_from(range_start, range_end, in accept_memory() 217 phys_end = range_end * unit_size + unaccepted_table->phys_base; in accept_memory() 221 range_start, range_end - range_start); in accept_memory()
|
| /drivers/infiniband/hw/hfi1/ |
| A D | fault.c | 111 unsigned long range_start, range_end, i; in fault_opcodes_write() local 129 if (kstrtoul(token, 0, &range_end)) in fault_opcodes_write() 132 range_end = range_start; in fault_opcodes_write() 134 if (range_start == range_end && range_start == -1UL) { in fault_opcodes_write() 140 if (range_start >= bound || range_end >= bound) in fault_opcodes_write() 143 for (i = range_start; i <= range_end; i++) { in fault_opcodes_write()
|
| /drivers/gpu/drm/ |
| A D | drm_mm.c | 518 u64 range_start, u64 range_end, in drm_mm_insert_node_in_range() argument 525 DRM_MM_BUG_ON(range_start > range_end); in drm_mm_insert_node_in_range() 527 if (unlikely(size == 0 || range_end - range_start < size)) in drm_mm_insert_node_in_range() 540 for (hole = first_hole(mm, range_start, range_end, size, mode); in drm_mm_insert_node_in_range() 548 if (mode == DRM_MM_INSERT_LOW && hole_start >= range_end) in drm_mm_insert_node_in_range() 560 adj_end = min(col_end, range_end); in drm_mm_insert_node_in_range() 581 min(col_end, range_end) - adj_start < size) in drm_mm_insert_node_in_range() 727 scan->range_end = end; in drm_mm_scan_init_with_range() 778 adj_end = min(col_end, scan->range_end); in drm_mm_scan_add_block() 797 min(col_end, scan->range_end) - adj_start < scan->size) in drm_mm_scan_add_block()
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_doorbell.c | 213 int range_end = dev->shared_resources.non_cp_doorbells_end; in init_doorbell_bitmap() local 219 pr_debug("reserved doorbell 0x%03x - 0x%03x\n", range_start, range_end); in init_doorbell_bitmap() 222 range_end + KFD_QUEUE_DOORBELL_MIRROR_OFFSET); in init_doorbell_bitmap() 225 if (i >= range_start && i <= range_end) { in init_doorbell_bitmap()
|
| A D | kfd_device.c | 1324 (*mem_obj)->range_end = found; in kfd_gtt_sa_allocate() 1347 (*mem_obj)->range_end = in kfd_gtt_sa_allocate() 1355 if ((*mem_obj)->range_end != found) { in kfd_gtt_sa_allocate() 1375 (*mem_obj)->range_start, (*mem_obj)->range_end); in kfd_gtt_sa_allocate() 1379 (*mem_obj)->range_end - (*mem_obj)->range_start + 1); in kfd_gtt_sa_allocate() 1401 mem_obj, mem_obj->range_start, mem_obj->range_end); in kfd_gtt_sa_free() 1407 mem_obj->range_end - mem_obj->range_start + 1); in kfd_gtt_sa_free()
|
| A D | kfd_priv.h | 245 uint32_t range_end; member
|
| /drivers/gpu/drm/xe/ |
| A D | xe_reg_whitelist.c | 159 u32 range_start, range_end; in xe_reg_whitelist_print_entry() local 177 range_end = range_start | REG_GENMASK(range_bit, 0); in xe_reg_whitelist_print_entry() 192 range_start, range_end, in xe_reg_whitelist_print_entry()
|
| A D | xe_vm.c | 2307 u64 range_end = addr + range; in vm_bind_ioctl_ops_create() local 2395 u64 ret = xe_svm_find_vma_start(vm, addr, range_end, vma); in vm_bind_ioctl_ops_create() 2425 if (range_end > xe_svm_range_end(svm_range) && in vm_bind_ioctl_ops_create()
|
| /drivers/gpu/drm/msm/ |
| A D | msm_gem.c | 461 u64 range_end) in get_vma_locked() argument 470 vma = msm_gem_vma_new(vm, obj, 0, range_start, range_end); in get_vma_locked() 473 GEM_WARN_ON((vma->va.addr + obj->size) > range_end); in get_vma_locked() 548 u64 range_start, u64 range_end) in get_and_pin_iova_range_locked() argument 558 vma = get_vma_locked(obj, vm, range_start, range_end); in get_and_pin_iova_range_locked() 577 u64 range_start, u64 range_end) in msm_gem_get_and_pin_iova_range() argument 583 ret = get_and_pin_iova_range_locked(obj, vm, iova, range_start, range_end); in msm_gem_get_and_pin_iova_range()
|
| A D | msm_gem.h | 190 u64 offset, u64 range_start, u64 range_end); 278 u64 range_start, u64 range_end);
|
| A D | msm_gem_vma.c | 372 u64 offset, u64 range_start, u64 range_end) in msm_gem_vma_new() argument 390 range_start, range_end, 0); in msm_gem_vma_new() 396 range_end = range_start + obj->size; in msm_gem_vma_new() 400 GEM_WARN_ON((range_end - range_start) > obj->size); in msm_gem_vma_new() 402 drm_gpuva_init(&vma->base, range_start, range_end - range_start, obj, offset); in msm_gem_vma_new()
|
| /drivers/mmc/host/ |
| A D | sdhci-sprd.c | 594 int range_end = SDHCI_SPRD_MAX_RANGE; in sdhci_sprd_get_best_clk_sample() local 608 range_end = i - 1; in sdhci_sprd_get_best_clk_sample() 619 range_end = i - 1; in sdhci_sprd_get_best_clk_sample() 622 middle_range = range_end - (range_length - 1) / 2; in sdhci_sprd_get_best_clk_sample()
|
| /drivers/gpu/drm/amd/display/dc/dml2/dml21/src/dml2_top/ |
| A D | dml2_top_soc15.c | 338 int range_end; in calculate_first_second_splitting() local 351 range_end = mcache_boundaries[left_cache_id] - shift - 1; in calculate_first_second_splitting() 353 if (range_start <= pipe_h_vp_start && pipe_h_vp_start <= range_end) in calculate_first_second_splitting() 356 range_start = range_end + 1; in calculate_first_second_splitting() 359 range_end = MAX_VP; in calculate_first_second_splitting() 366 if (range_start <= pipe_h_vp_end && pipe_h_vp_end <= range_end) { in calculate_first_second_splitting() 369 range_end = range_start - 1; in calculate_first_second_splitting()
|
| /drivers/accel/habanalabs/goya/ |
| A D | goya_coresight.c | 371 u64 range_start, range_end; in goya_etr_validate_address() local 380 range_end = prop->dmmu.end_addr; in goya_etr_validate_address() 382 return hl_mem_area_inside_range(addr, size, range_start, range_end); in goya_etr_validate_address()
|
| /drivers/net/ethernet/netronome/nfp/bpf/ |
| A D | jit.c | 2734 s16 range_end = meta->pkt_cache.range_end; in mem_ldx_data_init_pktcache() local 2741 len = range_end - range_start; in mem_ldx_data_init_pktcache() 2866 if (meta->pkt_cache.range_end) { in mem_ldx() 4279 s16 range_start = 0, range_end = 0; in nfp_bpf_opt_pkt_cache() local 4332 s16 new_end = range_end; in nfp_bpf_opt_pkt_cache() 4341 if (end > range_end) { in nfp_bpf_opt_pkt_cache() 4352 range_end = new_end; in nfp_bpf_opt_pkt_cache() 4359 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache() 4371 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache() 4381 range_end = meta->pkt_cache.range_end; in nfp_bpf_opt_pkt_cache() [all …]
|
| A D | main.h | 311 s16 range_end; member
|
| /drivers/net/ethernet/cisco/enic/ |
| A D | enic.h | 73 u32 range_end; member
|
| A D | enic_ethtool.c | 385 ecmd->rx_coalesce_usecs_high = rxcoal->range_end; in enic_get_coalesce() 459 rxcoal->range_end = rx_coalesce_usecs_high; in enic_set_coalesce()
|
| A D | enic_main.c | 400 rx_coal->range_end = ENIC_RX_COALESCE_RANGE_END; in enic_set_rx_coal_setting() 1314 timer = range_start + ((rx_coal->range_end - range_start) * in enic_calc_int_moderation()
|
| /drivers/gpu/drm/msm/adreno/ |
| A D | a6xx_gmu.c | 1281 u64 range_start, range_end; in a6xx_gmu_memory_alloc() local 1288 range_end = 0x80000000; in a6xx_gmu_memory_alloc() 1292 range_end = iova + size; in a6xx_gmu_memory_alloc() 1302 range_start, range_end); in a6xx_gmu_memory_alloc()
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_shmem.c | 305 .range_end = LLONG_MAX, in __shmem_writeback()
|