Lines Matching refs:access_flags

56 				     u64 iova, int access_flags,
668 res = key1.access_flags - key2.access_flags; in cache_ent_key_cmp()
735 smallest->rb_key.access_flags == rb_key.access_flags && in mkey_cache_ent_from_rb_key()
743 int access_flags) in _mlx5_mr_cache_alloc() argument
779 int access_flags) in get_unchangeable_access_flags() argument
783 if ((access_flags & IB_ACCESS_REMOTE_ATOMIC) && in get_unchangeable_access_flags()
788 if ((access_flags & IB_ACCESS_RELAXED_ORDERING) && in get_unchangeable_access_flags()
793 if ((access_flags & IB_ACCESS_RELAXED_ORDERING) && in get_unchangeable_access_flags()
802 int access_flags, int access_mode, in mlx5_mr_cache_alloc() argument
808 .access_flags = get_unchangeable_access_flags(dev, access_flags) in mlx5_mr_cache_alloc()
815 return _mlx5_mr_cache_alloc(dev, ent, access_flags); in mlx5_mr_cache_alloc()
1116 u64 length, int access_flags, u64 iova) in set_mr_fields() argument
1123 mr->access_flags = access_flags; in set_mr_fields()
1139 int access_flags) in alloc_cacheable_mr() argument
1158 rb_key.ats = mlx5_umem_needs_ats(dev, umem, access_flags); in alloc_cacheable_mr()
1159 rb_key.access_flags = get_unchangeable_access_flags(dev, access_flags); in alloc_cacheable_mr()
1167 mr = reg_create(pd, umem, iova, access_flags, page_size, false); in alloc_cacheable_mr()
1175 mr = _mlx5_mr_cache_alloc(dev, ent, access_flags); in alloc_cacheable_mr()
1182 set_mr_fields(dev, mr, umem->length, access_flags, iova); in alloc_cacheable_mr()
1192 u64 iova, int access_flags, in reg_create() argument
1211 mr->access_flags = access_flags; in reg_create()
1225 if (WARN_ON(access_flags & IB_ACCESS_ON_DEMAND)) { in reg_create()
1238 set_mkc_access_pd_addr_fields(mkc, access_flags, iova, in reg_create()
1249 if (mlx5_umem_needs_ats(dev, umem, access_flags)) in reg_create()
1264 set_mr_fields(dev, mr, umem->length, access_flags, iova); in reg_create()
1351 if (attr->access_flags & ~MLX5_IB_DM_MEMIC_ALLOWED_ACCESS) in mlx5_ib_reg_dm_mr()
1360 if (attr->access_flags & ~MLX5_IB_DM_SW_ICM_ALLOWED_ACCESS) in mlx5_ib_reg_dm_mr()
1370 attr->access_flags, mode); in mlx5_ib_reg_dm_mr()
1374 u64 iova, int access_flags) in create_real_mr() argument
1383 mr = alloc_cacheable_mr(pd, umem, iova, access_flags); in create_real_mr()
1389 mr = reg_create(pd, umem, iova, access_flags, page_size, true); in create_real_mr()
1417 u64 iova, int access_flags, in create_user_odp_mr() argument
1437 mr = mlx5_ib_alloc_implicit_mr(to_mpd(pd), access_flags); in create_user_odp_mr()
1447 odp = ib_umem_odp_get(&dev->ib_dev, start, length, access_flags, in create_user_odp_mr()
1452 mr = alloc_cacheable_mr(pd, &odp->umem, iova, access_flags); in create_user_odp_mr()
1475 u64 iova, int access_flags, in mlx5_ib_reg_user_mr() argument
1485 start, iova, length, access_flags); in mlx5_ib_reg_user_mr()
1487 if (access_flags & IB_ACCESS_ON_DEMAND) in mlx5_ib_reg_user_mr()
1488 return create_user_odp_mr(pd, start, length, iova, access_flags, in mlx5_ib_reg_user_mr()
1490 umem = ib_umem_get(&dev->ib_dev, start, length, access_flags); in mlx5_ib_reg_user_mr()
1493 return create_real_mr(pd, umem, iova, access_flags); in mlx5_ib_reg_user_mr()
1517 int fd, int access_flags, in mlx5_ib_reg_user_mr_dmabuf() argument
1531 offset, virt_addr, length, fd, access_flags); in mlx5_ib_reg_user_mr_dmabuf()
1538 access_flags, in mlx5_ib_reg_user_mr_dmabuf()
1547 access_flags); in mlx5_ib_reg_user_mr_dmabuf()
1610 int access_flags, int flags, struct ib_umem *new_umem, in umr_rereg_pas() argument
1632 mr->access_flags = access_flags; in umr_rereg_pas()
1677 new_access_flags = mr->access_flags; in mlx5_ib_rereg_user_mr()
1685 if (can_use_umr_rereg_access(dev, mr->access_flags, in mlx5_ib_rereg_user_mr()
1720 can_use_umr_rereg_access(dev, mr->access_flags, new_access_flags)) { in mlx5_ib_rereg_user_mr()