Home
last modified time | relevance | path

Searched refs:pgt (Results 1 – 13 of 13) sorted by relevance

/drivers/net/ethernet/mellanox/mlxsw/
A Dspectrum_pgt.c32 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc()
42 mutex_unlock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc()
52 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_free()
63 mutex_lock(&mlxsw_sp->pgt->lock); in mlxsw_sp_pgt_mid_alloc_range()
313 struct mlxsw_sp_pgt *pgt; in mlxsw_sp_pgt_init() local
318 pgt = kzalloc(sizeof(*mlxsw_sp->pgt), GFP_KERNEL); in mlxsw_sp_pgt_init()
319 if (!pgt) in mlxsw_sp_pgt_init()
322 idr_init(&pgt->pgt_idr); in mlxsw_sp_pgt_init()
324 mutex_init(&pgt->lock); in mlxsw_sp_pgt_init()
326 mlxsw_sp->pgt = pgt; in mlxsw_sp_pgt_init()
[all …]
A Dspectrum.h210 struct mlxsw_sp_pgt *pgt; member
/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
A Dvmm.c31 if (pgt) { in nvkm_vmm_pt_del()
32 kvfree(pgt->pde); in nvkm_vmm_pt_del()
33 kfree(pgt); in nvkm_vmm_pt_del()
56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new()
62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new()
63 if (!pgt->pde) { in nvkm_vmm_pt_new()
64 kfree(pgt); in nvkm_vmm_pt_new()
69 return pgt; in nvkm_vmm_pt_new()
280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes()
489 pgt = nvkm_vmm_pt_new(desc, NVKM_VMM_PDE_SPARSED(pgt), it->page); in nvkm_vmm_ref_swpt()
[all …]
A Dvmmgh100.c140 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gh100_vmm_pd0_pde() local
144 if (pgt->pt[0] && !gh100_vmm_pde(pgt->pt[0], &data[0])) in gh100_vmm_pd0_pde()
146 if (pgt->pt[1] && !gh100_vmm_pde(pgt->pt[1], &data[1])) in gh100_vmm_pd0_pde()
181 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gh100_vmm_pd1_pde() local
185 if (!gh100_vmm_pde(pgt->pt[0], &data)) in gh100_vmm_pd1_pde()
A Dvmmgp100.c238 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd0_pde() local
242 if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0])) in gp100_vmm_pd0_pde()
244 if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1])) in gp100_vmm_pd0_pde()
365 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd1_pde() local
369 if (!gp100_vmm_pde(pgt->pt[0], &data)) in gp100_vmm_pd1_pde()
A Dvmmnv50.c106 nv50_vmm_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgt, u64 *pdata) in nv50_vmm_pde() argument
110 if (pgt && (pt = pgt->pt[0])) { in nv50_vmm_pde()
111 switch (pgt->page) { in nv50_vmm_pde()
A Dvmmgf100.c108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gf100_vmm_pgd_pde() local
113 if ((pt = pgt->pt[0])) { in gf100_vmm_pgd_pde()
127 if ((pt = pgt->pt[1])) { in gf100_vmm_pgd_pde()
/drivers/gpu/drm/nouveau/nvkm/engine/dma/
A Dusernv04.c52 struct nvkm_memory *pgt = in nv04_dmaobj_bind() local
55 return nvkm_gpuobj_wrap(pgt, pgpuobj); in nv04_dmaobj_bind()
56 nvkm_kmap(pgt); in nv04_dmaobj_bind()
57 offset = nvkm_ro32(pgt, 8 + (offset >> 10)); in nv04_dmaobj_bind()
59 nvkm_done(pgt); in nv04_dmaobj_bind()
/drivers/firmware/efi/libstub/
A Dx86-5lvl.c68 u64 *pgt = (void *)la57_toggle + PAGE_SIZE; in efi_5level_switch() local
81 new_cr3 = memset(pgt, 0, PAGE_SIZE); in efi_5level_switch()
89 new_cr3 = memcpy(pgt, new_cr3, PAGE_SIZE); in efi_5level_switch()
/drivers/accel/habanalabs/common/mmu/
A Dmmu.c953 u64 hl_mmu_hr_pte_phys_to_virt(struct hl_ctx *ctx, struct pgt_info *pgt, in hl_mmu_hr_pte_phys_to_virt() argument
959 return pgt->virt_addr + pte_offset; in hl_mmu_hr_pte_phys_to_virt()
/drivers/gpu/drm/panthor/
A Dpanthor_mmu.c822 const struct io_pgtable *pgt = io_pgtable_ops_to_pgtable(vm->pgtbl_ops); in panthor_vm_page_size() local
823 u32 pg_shift = ffs(pgt->cfg.pgsize_bitmap) - 1; in panthor_vm_page_size()
/drivers/net/ethernet/broadcom/bnx2x/
A Dbnx2x_self_test.c78 static int pgt(struct st_pred_args *args) in pgt() function
395 NA, 1, 0, pgt,
/drivers/accel/habanalabs/common/
A Dhabanalabs.h3900 u64 hl_mmu_hr_pte_phys_to_virt(struct hl_ctx *ctx, struct pgt_info *pgt, u64 phys_pte_addr,

Completed in 55 milliseconds