Lines Matching refs:range

692 		 struct xe_svm_range *range,  in xe_pt_stage_bind()  argument
710 .va_curs_start = range ? range->base.itree.start : in xe_pt_stage_bind()
719 if (range) { in xe_pt_stage_bind()
722 if (!xe_svm_range_pages_valid(range)) { in xe_pt_stage_bind()
723 xe_svm_range_debug(range, "BIND PREPARE - RETRY"); in xe_pt_stage_bind()
727 if (xe_svm_range_has_dma_mapping(range)) { in xe_pt_stage_bind()
728 xe_res_first_dma(range->base.dma_addr, 0, in xe_pt_stage_bind()
729 range->base.itree.last + 1 - range->base.itree.start, in xe_pt_stage_bind()
731 xe_svm_range_debug(range, "BIND PREPARE - MIXED"); in xe_pt_stage_bind()
754 if (!range) in xe_pt_stage_bind()
757 if (!xe_vma_is_null(vma) && !range) { in xe_pt_stage_bind()
767 } else if (!range) { in xe_pt_stage_bind()
773 range ? range->base.itree.start : xe_vma_start(vma), in xe_pt_stage_bind()
774 range ? range->base.itree.last + 1 : xe_vma_end(vma), in xe_pt_stage_bind()
940 struct xe_svm_range *range) in xe_pt_zap_ptes_range() argument
951 u8 pt_mask = (range->tile_present & ~range->tile_invalidated); in xe_pt_zap_ptes_range()
958 (void)xe_pt_walk_shared(&pt->base, pt->level, range->base.itree.start, in xe_pt_zap_ptes_range()
959 range->base.itree.last + 1, &xe_walk.base); in xe_pt_zap_ptes_range()
1140 struct xe_svm_range *range, in xe_pt_prepare_bind() argument
1147 err = xe_pt_stage_bind(tile, vma, range, entries, num_entries, in xe_pt_prepare_bind()
1476 struct xe_svm_range *range = NULL; in xe_pt_svm_pre_commit() local
1484 xa_for_each(&op->prefetch_range.range, i, range) { in xe_pt_svm_pre_commit()
1485 xe_svm_range_debug(range, "PRE-COMMIT"); in xe_pt_svm_pre_commit()
1487 if (!xe_svm_range_pages_valid(range)) { in xe_pt_svm_pre_commit()
1488 xe_svm_range_debug(range, "PRE-COMMIT - RETRY"); in xe_pt_svm_pre_commit()
1496 range = op->map_range.range; in xe_pt_svm_pre_commit()
1498 xe_svm_range_debug(range, "PRE-COMMIT"); in xe_pt_svm_pre_commit()
1500 if (!xe_svm_range_pages_valid(range)) { in xe_pt_svm_pre_commit()
1501 xe_svm_range_debug(range, "PRE-COMMIT - RETRY"); in xe_pt_svm_pre_commit()
1712 struct xe_svm_range *range, in xe_pt_stage_unbind() argument
1715 u64 start = range ? range->base.itree.start : xe_vma_start(vma); in xe_pt_stage_unbind()
1716 u64 end = range ? range->base.itree.last + 1 : xe_vma_end(vma); in xe_pt_stage_unbind()
1916 struct xe_vma *vma, struct xe_svm_range *range) in bind_range_prepare() argument
1926 range->base.itree.start, range->base.itree.last); in bind_range_prepare()
1930 pt_op->rebind = BIT(tile->id) & range->tile_present; in bind_range_prepare()
1932 err = xe_pt_prepare_bind(tile, vma, range, pt_op->entries, in bind_range_prepare()
1941 range->base.itree.start, in bind_range_prepare()
1942 range->base.itree.last + 1); in bind_range_prepare()
2000 struct xe_svm_range *range) in xe_pt_op_check_range_skip_invalidation() argument
2019 return xe_svm_range_size(range) >= SZ_2M; in xe_pt_op_check_range_skip_invalidation()
2027 struct xe_svm_range *range) in unbind_range_prepare() argument
2032 if (!(range->tile_present & BIT(tile->id))) in unbind_range_prepare()
2037 range->base.itree.start, range->base.itree.last); in unbind_range_prepare()
2043 pt_op->num_entries = xe_pt_stage_unbind(tile, vm, NULL, range, in unbind_range_prepare()
2048 xe_pt_update_ops_rfence_interval(pt_update_ops, range->base.itree.start, in unbind_range_prepare()
2049 range->base.itree.last + 1); in unbind_range_prepare()
2053 xe_vm_has_valid_gpu_mapping(tile, range->tile_present, in unbind_range_prepare()
2054 range->tile_invalidated) || in unbind_range_prepare()
2055 !xe_pt_op_check_range_skip_invalidation(pt_op, range); in unbind_range_prepare()
2119 struct xe_svm_range *range; in op_prepare() local
2122 xa_for_each(&op->prefetch_range.range, i, range) { in op_prepare()
2124 vma, range); in op_prepare()
2140 op->map_range.range); in op_prepare()
2143 op->unmap_range.range); in op_prepare()
2284 struct xe_svm_range *range, in range_present_and_invalidated_tile() argument
2291 WRITE_ONCE(range->tile_present, range->tile_present | BIT(tile_id)); in range_present_and_invalidated_tile()
2292 WRITE_ONCE(range->tile_invalidated, range->tile_invalidated & ~BIT(tile_id)); in range_present_and_invalidated_tile()
2343 struct xe_svm_range *range = NULL; in op_commit() local
2346 xa_for_each(&op->prefetch_range.range, i, range) in op_commit()
2347 range_present_and_invalidated_tile(vm, range, tile->id); in op_commit()
2358 range_present_and_invalidated_tile(vm, op->map_range.range, tile->id); in op_commit()
2360 WRITE_ONCE(op->unmap_range.range->tile_present, in op_commit()
2361 op->unmap_range.range->tile_present & in op_commit()