| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_pgt.c | 32 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc() 42 mutex_unlock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc() 52 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_free() 63 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc_range() 313 struct mlxsw_sp_pgt *pgt; in mlxsw_sp_pgt_init() local 318 pgt = kzalloc(sizeof(*mlxsw_sp->pgt), GFP_KERNEL); in mlxsw_sp_pgt_init() 319 if (!pgt) in mlxsw_sp_pgt_init() 322 idr_init(&pgt->pgt_idr); in mlxsw_sp_pgt_init() 324 mutex_init(&pgt->lock); in mlxsw_sp_pgt_init() 326 mlxsw_sp->pgt = pgt; in mlxsw_sp_pgt_init() [all …]
|
| A D | spectrum.h | 210 struct mlxsw_sp_pgt *pgt; member
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| A D | vmm.c | 31 if (pgt) { in nvkm_vmm_pt_del() 32 kvfree(pgt->pde); in nvkm_vmm_pt_del() 33 kfree(pgt); in nvkm_vmm_pt_del() 56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new() 62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new() 63 if (!pgt->pde) { in nvkm_vmm_pt_new() 64 kfree(pgt); in nvkm_vmm_pt_new() 69 return pgt; in nvkm_vmm_pt_new() 280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes() 489 pgt = nvkm_vmm_pt_new(desc, NVKM_VMM_PDE_SPARSED(pgt), it->page); in nvkm_vmm_ref_swpt() [all …]
|
| A D | vmmgh100.c | 140 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gh100_vmm_pd0_pde() local 144 if (pgt->pt[0] && !gh100_vmm_pde(pgt->pt[0], &data[0])) in gh100_vmm_pd0_pde() 146 if (pgt->pt[1] && !gh100_vmm_pde(pgt->pt[1], &data[1])) in gh100_vmm_pd0_pde() 181 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gh100_vmm_pd1_pde() local 185 if (!gh100_vmm_pde(pgt->pt[0], &data)) in gh100_vmm_pd1_pde()
|
| A D | vmmgp100.c | 238 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd0_pde() local 242 if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0])) in gp100_vmm_pd0_pde() 244 if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1])) in gp100_vmm_pd0_pde() 365 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd1_pde() local 369 if (!gp100_vmm_pde(pgt->pt[0], &data)) in gp100_vmm_pd1_pde()
|
| A D | vmmnv50.c | 106 nv50_vmm_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgt, u64 *pdata) in nv50_vmm_pde() argument 110 if (pgt && (pt = pgt->pt[0])) { in nv50_vmm_pde() 111 switch (pgt->page) { in nv50_vmm_pde()
|
| A D | vmmgf100.c | 108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gf100_vmm_pgd_pde() local 113 if ((pt = pgt->pt[0])) { in gf100_vmm_pgd_pde() 127 if ((pt = pgt->pt[1])) { in gf100_vmm_pgd_pde()
|
| /drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
| A D | usernv04.c | 52 struct nvkm_memory *pgt = in nv04_dmaobj_bind() local 55 return nvkm_gpuobj_wrap(pgt, pgpuobj); in nv04_dmaobj_bind() 56 nvkm_kmap(pgt); in nv04_dmaobj_bind() 57 offset = nvkm_ro32(pgt, 8 + (offset >> 10)); in nv04_dmaobj_bind() 59 nvkm_done(pgt); in nv04_dmaobj_bind()
|
| /drivers/firmware/efi/libstub/ |
| A D | x86-5lvl.c | 68 u64 *pgt = (void *)la57_toggle + PAGE_SIZE; in efi_5level_switch() local 81 new_cr3 = memset(pgt, 0, PAGE_SIZE); in efi_5level_switch() 89 new_cr3 = memcpy(pgt, new_cr3, PAGE_SIZE); in efi_5level_switch()
|
| /drivers/accel/habanalabs/common/mmu/ |
| A D | mmu.c | 953 u64 hl_mmu_hr_pte_phys_to_virt(struct hl_ctx *ctx, struct pgt_info *pgt, in hl_mmu_hr_pte_phys_to_virt() argument 959 return pgt->virt_addr + pte_offset; in hl_mmu_hr_pte_phys_to_virt()
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_mmu.c | 822 const struct io_pgtable *pgt = io_pgtable_ops_to_pgtable(vm->pgtbl_ops); in panthor_vm_page_size() local 823 u32 pg_shift = ffs(pgt->cfg.pgsize_bitmap) - 1; in panthor_vm_page_size()
|
| /drivers/net/ethernet/broadcom/bnx2x/ |
| A D | bnx2x_self_test.c | 78 static int pgt(struct st_pred_args *args) in pgt() function 395 NA, 1, 0, pgt,
|
| /drivers/accel/habanalabs/common/ |
| A D | habanalabs.h | 3900 u64 hl_mmu_hr_pte_phys_to_virt(struct hl_ctx *ctx, struct pgt_info *pgt, u64 phys_pte_addr,
|