/linux-6.3-rc2/drivers/gpu/drm/tegra/ |
A D | gem.c | 62 struct drm_gem_object *gem = &obj->gem; in tegra_bo_pin() local 78 if (gem->import_attach) { in tegra_bo_pin() 96 map->size = gem->size; in tegra_bo_pin() 148 map->size = gem->size; in tegra_bo_pin() 581 gem->size); in __tegra_gem_mmap() 631 gem->size) < 0) in tegra_gem_prime_map_dma_buf() 698 err = drm_gem_mmap_obj(gem, gem->size, vma); in tegra_gem_prime_mmap() 749 exp_info.priv = gem; in tegra_gem_prime_export() 764 return gem; in tegra_gem_prime_import() 772 return &bo->gem; in tegra_gem_prime_import() [all …]
|
A D | gem.h | 36 struct drm_gem_object gem; member 52 static inline struct tegra_bo *to_tegra_bo(struct drm_gem_object *gem) in to_tegra_bo() argument 54 return container_of(gem, struct tegra_bo, gem); in to_tegra_bo() 69 void tegra_bo_free_object(struct drm_gem_object *gem); 75 int __tegra_gem_mmap(struct drm_gem_object *gem, struct vm_area_struct *vma); 78 struct dma_buf *tegra_gem_prime_export(struct drm_gem_object *gem,
|
A D | fb.c | 127 fb->obj[i] = &planes[i]->gem; in tegra_fb_alloc() 146 struct drm_gem_object *gem; in tegra_fb_create() local 156 gem = drm_gem_object_lookup(file, cmd->handles[i]); in tegra_fb_create() 157 if (!gem) { in tegra_fb_create() 167 if (gem->size < size) { in tegra_fb_create() 172 planes[i] = to_tegra_bo(gem); in tegra_fb_create() 185 drm_gem_object_put(&planes[i]->gem); in tegra_fb_create() 199 err = drm_gem_mmap_obj(&bo->gem, bo->gem.size, vma); in tegra_fb_mmap() 203 return __tegra_gem_mmap(&bo->gem, vma); in tegra_fb_mmap() 251 drm_gem_object_put(&bo->gem); in tegra_fbdev_probe() [all …]
|
A D | drm.c | 374 if (!gem) in tegra_gem_mmap() 377 bo = to_tegra_bo(gem); in tegra_gem_mmap() 381 drm_gem_object_put(gem); in tegra_gem_mmap() 648 if (!gem) in tegra_gem_set_tiling() 651 bo = to_tegra_bo(gem); in tegra_gem_set_tiling() 670 if (!gem) in tegra_gem_get_tiling() 673 bo = to_tegra_bo(gem); in tegra_gem_get_tiling() 712 if (!gem) in tegra_gem_set_flags() 715 bo = to_tegra_bo(gem); in tegra_gem_set_flags() 734 if (!gem) in tegra_gem_get_flags() [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/ |
A D | drm_gem_ttm_helper.c | 26 const struct drm_gem_object *gem) in drm_gem_ttm_print_info() argument 65 int drm_gem_ttm_vmap(struct drm_gem_object *gem, in drm_gem_ttm_vmap() argument 68 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gem); in drm_gem_ttm_vmap() 82 void drm_gem_ttm_vunmap(struct drm_gem_object *gem, in drm_gem_ttm_vunmap() argument 99 int drm_gem_ttm_mmap(struct drm_gem_object *gem, in drm_gem_ttm_mmap() argument 113 drm_gem_object_put(gem); in drm_gem_ttm_mmap() 138 struct drm_gem_object *gem; in drm_gem_ttm_dumb_map_offset() local 140 gem = drm_gem_object_lookup(file, handle); in drm_gem_ttm_dumb_map_offset() 141 if (!gem) in drm_gem_ttm_dumb_map_offset() 144 *offset = drm_vma_node_offset_addr(&gem->vma_node); in drm_gem_ttm_dumb_map_offset() [all …]
|
A D | drm_gem_vram_helper.c | 192 struct drm_gem_object *gem; in drm_gem_vram_create() local 201 gem = dev->driver->gem_create_object(dev, size); in drm_gem_vram_create() 202 if (IS_ERR(gem)) in drm_gem_vram_create() 203 return ERR_CAST(gem); in drm_gem_vram_create() 204 gbo = drm_gem_vram_of_gem(gem); in drm_gem_vram_create() 209 gem = &gbo->bo.base; in drm_gem_vram_create() 212 if (!gem->funcs) in drm_gem_vram_create() 213 gem->funcs = &drm_gem_vram_object_funcs; in drm_gem_vram_create() 215 ret = drm_gem_object_init(dev, gem, size); in drm_gem_vram_create() 584 struct drm_gem_vram_object *gbo = drm_gem_vram_of_gem(gem); in drm_gem_vram_object_free() [all …]
|
A D | drm_fb_dma_helper.c | 49 struct drm_gem_object *gem; in drm_fb_dma_get_gem_obj() local 51 gem = drm_gem_fb_get_obj(fb, plane); in drm_fb_dma_get_gem_obj() 52 if (!gem) in drm_fb_dma_get_gem_obj() 55 return to_drm_gem_dma_obj(gem); in drm_fb_dma_get_gem_obj()
|
A D | drm_client.c | 246 if (buffer->gem) { in drm_client_buffer_delete() 247 drm_gem_vunmap_unlocked(buffer->gem, &buffer->map); in drm_client_buffer_delete() 248 drm_gem_object_put(buffer->gem); in drm_client_buffer_delete() 285 buffer->gem = obj; in drm_client_buffer_create() 331 ret = drm_gem_vmap_unlocked(buffer->gem, map); in drm_client_buffer_vmap() 353 drm_gem_vunmap_unlocked(buffer->gem, map); in drm_client_buffer_vunmap()
|
/linux-6.3-rc2/drivers/gpu/drm/i915/ |
A D | Makefile | 150 gem-y += \ 151 gem/i915_gem_busy.o \ 160 gem/i915_gem_lmem.o \ 161 gem/i915_gem_mman.o \ 162 gem/i915_gem_pages.o \ 163 gem/i915_gem_phys.o \ 164 gem/i915_gem_pm.o \ 171 gem/i915_gem_ttm.o \ 175 gem/i915_gem_wait.o \ 176 gem/i915_gemfs.o [all …]
|
A D | i915_sysfs.c | 80 spin_lock(&i915->gem.contexts.lock); in i915_l3_read() 85 spin_unlock(&i915->gem.contexts.lock); in i915_l3_read() 113 spin_lock(&i915->gem.contexts.lock); in i915_l3_write() 126 list_for_each_entry(ctx, &i915->gem.contexts.list, link) in i915_l3_write() 129 spin_unlock(&i915->gem.contexts.lock); in i915_l3_write()
|
/linux-6.3-rc2/Documentation/devicetree/bindings/net/ |
A D | cdns,macb.yaml | 23 - cdns,zynq-gem # Xilinx Zynq-7xxx SoC 24 - cdns,zynqmp-gem # Xilinx Zynq Ultrascale+ MPSoC 25 - const: cdns,gem # Generic 30 - xlnx,versal-gem # Xilinx Versal 31 - xlnx,zynq-gem # Xilinx Zynq-7xxx SoC 32 - xlnx,zynqmp-gem # Xilinx Zynq Ultrascale+ MPSoC 33 - const: cdns,gem # Generic 57 - sifive,fu540-c000-gem # SiFive FU540-C000 SoC 59 - cdns,gem # Generic 155 const: sifive,fu540-c000-gem [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/nouveau/ |
A D | nouveau_gem.c | 90 if (gem->import_attach) in nouveau_gem_object_del() 438 struct drm_gem_object *gem; in validate_init() local 442 if (!gem) { in validate_init() 450 drm_gem_object_put(gem); in validate_init() 457 drm_gem_object_put(gem); in validate_init() 965 if (!gem) in nouveau_gem_ioctl_cpu_prep() 980 drm_gem_object_put(gem); in nouveau_gem_ioctl_cpu_prep() 994 if (!gem) in nouveau_gem_ioctl_cpu_fini() 999 drm_gem_object_put(gem); in nouveau_gem_ioctl_cpu_fini() 1012 if (!gem) in nouveau_gem_ioctl_info() [all …]
|
A D | nouveau_gem.h | 11 nouveau_gem_object(struct drm_gem_object *gem) in nouveau_gem_object() argument 13 return gem ? container_of(gem, struct nouveau_bo, bo.base) : NULL; in nouveau_gem_object()
|
A D | nouveau_ttm.c | 192 drm->gem.vram_available >> PAGE_SHIFT); in nouveau_ttm_init_vram() 198 drm->gem.vram_available >> PAGE_SHIFT); in nouveau_ttm_init_vram() 221 unsigned long size_pages = drm->gem.gart_available >> PAGE_SHIFT; in nouveau_ttm_init_gtt() 317 drm->gem.vram_available = drm->client.device.info.ram_user; in nouveau_ttm_init() 333 drm->gem.gart_available = drm->client.vmm.vmm.limit; in nouveau_ttm_init() 335 drm->gem.gart_available = drm->agp.size; in nouveau_ttm_init() 347 NV_INFO(drm, "VRAM: %d MiB\n", (u32)(drm->gem.vram_available >> 20)); in nouveau_ttm_init() 348 NV_INFO(drm, "GART: %d MiB\n", (u32)(drm->gem.gart_available >> 20)); in nouveau_ttm_init()
|
/linux-6.3-rc2/drivers/net/ethernet/sun/ |
A D | sungem.c | 144 struct gem *gp = netdev_priv(dev); in _sungem_phy_read() 177 struct gem *gp = netdev_priv(dev); in _sungem_phy_write() 875 struct gem *gp = container_of(napi, struct gem, napi); in gem_poll() 926 struct gem *gp = netdev_priv(dev); in gem_interrupt() 954 struct gem *gp = netdev_priv(dev); in gem_poll_controller() 964 struct gem *gp = netdev_priv(dev); in gem_tx_timeout() 992 struct gem *gp = netdev_priv(dev); in gem_start_xmit() 2141 struct gem *gp = netdev_priv(dev); in gem_do_start() 2180 struct gem *gp = netdev_priv(dev); in gem_do_stop() 2225 struct gem *gp = container_of(work, struct gem, reset_task); in gem_reset_task() [all …]
|
/linux-6.3-rc2/include/drm/ |
A D | drm_gem_ttm_helper.h | 18 const struct drm_gem_object *gem); 19 int drm_gem_ttm_vmap(struct drm_gem_object *gem, 21 void drm_gem_ttm_vunmap(struct drm_gem_object *gem, 23 int drm_gem_ttm_mmap(struct drm_gem_object *gem,
|
/linux-6.3-rc2/drivers/gpu/drm/qxl/ |
A D | qxl_gem.c | 68 mutex_lock(&qdev->gem.mutex); in qxl_gem_object_create() 69 list_add_tail(&qbo->list, &qdev->gem.objects); in qxl_gem_object_create() 70 mutex_unlock(&qdev->gem.mutex); in qxl_gem_object_create() 116 INIT_LIST_HEAD(&qdev->gem.objects); in qxl_gem_init()
|
A D | qxl_object.c | 45 mutex_lock(&qdev->gem.mutex); in qxl_ttm_bo_destroy() 47 mutex_unlock(&qdev->gem.mutex); in qxl_ttm_bo_destroy() 359 if (list_empty(&qdev->gem.objects)) in qxl_bo_force_delete() 362 list_for_each_entry_safe(bo, n, &qdev->gem.objects, list) { in qxl_bo_force_delete() 366 mutex_lock(&qdev->gem.mutex); in qxl_bo_force_delete() 368 mutex_unlock(&qdev->gem.mutex); in qxl_bo_force_delete()
|
/linux-6.3-rc2/drivers/gpu/drm/radeon/ |
A D | radeon_prime.c | 63 mutex_lock(&rdev->gem.mutex); in radeon_gem_prime_import_sg_table() 64 list_add_tail(&bo->list, &rdev->gem.objects); in radeon_gem_prime_import_sg_table() 65 mutex_unlock(&rdev->gem.mutex); in radeon_gem_prime_import_sg_table()
|
/linux-6.3-rc2/drivers/gpu/drm/i915/pxp/ |
A D | intel_pxp.c | 357 spin_lock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 358 list_for_each_entry_safe(ctx, cn, &i915->gem.contexts.list, link) { in intel_pxp_invalidate() 370 spin_unlock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 396 spin_lock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate() 400 spin_unlock_irq(&i915->gem.contexts.lock); in intel_pxp_invalidate()
|
/linux-6.3-rc2/drivers/gpu/drm/rcar-du/ |
A D | rcar_du_vsp.c | 242 struct drm_gem_dma_object *gem = drm_fb_dma_get_gem_obj(fb, i); in rcar_du_vsp_map_fb() local 245 if (gem->sgt) { in rcar_du_vsp_map_fb() 256 ret = sg_alloc_table(sgt, gem->sgt->orig_nents, in rcar_du_vsp_map_fb() 261 src = gem->sgt->sgl; in rcar_du_vsp_map_fb() 263 for (j = 0; j < gem->sgt->orig_nents; ++j) { in rcar_du_vsp_map_fb() 270 ret = dma_get_sgtable(rcdu->dev, sgt, gem->vaddr, in rcar_du_vsp_map_fb() 271 gem->dma_addr, gem->base.size); in rcar_du_vsp_map_fb()
|
/linux-6.3-rc2/drivers/gpu/drm/shmobile/ |
A D | shmob_drm_plane.c | 43 struct drm_gem_dma_object *gem; in shmob_drm_plane_compute_base() local 47 gem = drm_fb_dma_get_gem_obj(fb, 0); in shmob_drm_plane_compute_base() 48 splane->dma[0] = gem->dma_addr + fb->offsets[0] in shmob_drm_plane_compute_base() 53 gem = drm_fb_dma_get_gem_obj(fb, 1); in shmob_drm_plane_compute_base() 54 splane->dma[1] = gem->dma_addr + fb->offsets[1] in shmob_drm_plane_compute_base()
|
/linux-6.3-rc2/Documentation/devicetree/bindings/interconnect/ |
A D | qcom,rpmh.yaml | 36 - qcom,sc7180-gem-noc 48 - qcom,sc8180x-gem-noc 81 - qcom,sm8150-gem-noc 90 - qcom,sm8250-gem-noc 99 - qcom,sm8350-gem-noc
|
/linux-6.3-rc2/drivers/gpu/drm/aspeed/ |
A D | aspeed_gfx_crtc.c | 170 struct drm_gem_dma_object *gem; in aspeed_gfx_pipe_update() local 187 gem = drm_fb_dma_get_gem_obj(fb, 0); in aspeed_gfx_pipe_update() 188 if (!gem) in aspeed_gfx_pipe_update() 190 writel(gem->dma_addr, priv->base + CRT_ADDR); in aspeed_gfx_pipe_update()
|
/linux-6.3-rc2/drivers/gpu/drm/fsl-dcu/ |
A D | fsl_dcu_drm_plane.c | 87 struct drm_gem_dma_object *gem; in fsl_dcu_drm_plane_atomic_update() local 98 gem = drm_fb_dma_get_gem_obj(fb, 0); in fsl_dcu_drm_plane_atomic_update() 139 DCU_CTRLDESCLN(index, 3), gem->dma_addr); in fsl_dcu_drm_plane_atomic_update()
|