| /drivers/macintosh/ |
| A D | rack-meter.c | 162 out_le32(&rm->dma_regs->cmdptr, rm->dma_buf_p); in rackmeter_do_pause() 216 struct rackmeter *rm = rcpu->rm; in rackmeter_do_timer() local 246 pause = (rm->cpu[0].zero && rm->cpu[1].zero); in rackmeter_do_timer() 249 pause = (rm->cpu[0].zero && rm->cpu[1].zero); in rackmeter_do_timer() 267 rm->cpu[0].rm = rm; in rackmeter_init_cpu_sniffer() 269 rm->cpu[1].rm = rm; in rackmeter_init_cpu_sniffer() 399 rm = kzalloc(sizeof(*rm), GFP_KERNEL); in rackmeter_probe() 510 rm->dma_buf_v, rm->dma_buf_p); in rackmeter_probe() 518 kfree(rm); in rackmeter_probe() 540 free_irq(rm->irq, rm); in rackmeter_remove() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ |
| A D | vmm.c | 30 if (vmm->rm.external) { in r535_mmu_vaspace_del() 42 vmm->rm.external = false; in r535_mmu_vaspace_del() 59 &vmm->rm.client, &vmm->rm.device); in r535_mmu_vaspace_new() 94 &vmm->rm.rsvd); in r535_mmu_vaspace_new() 112 ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1; in r535_mmu_vaspace_new() 174 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_mmu_new() 179 rm->mmu = hw->mmu; in r535_mmu_new() 180 rm->mem = hw->mem; in r535_mmu_new() 181 rm->vmm = hw->vmm; in r535_mmu_new() 182 rm->kind = hw->kind; in r535_mmu_new() [all …]
|
| A D | disp.c | 116 const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api; in r535_curs_init() 187 const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api; in r535_dmac_init() 768 const struct nvkm_rm_api *rmapi = disp->rm.objcom.client->gsp->rm->api; in r535_outp_inherit() 889 const struct nvkm_rm_api *rmapi = outp->disp->rm.objcom.client->gsp->rm->api; in r535_outp_detect() 1064 struct nvkm_rm *rm = outp->disp->rm.objcom.client->gsp->rm; in r535_dp_rates() local 1257 const struct nvkm_rm_api *rmapi = disp->rm.objcom.client->gsp->rm->api; in r535_outp_new() 1540 ret = nvkm_gsp_client_device_ctor(gsp, &disp->rm.client, &disp->rm.device); in r535_disp_oneinit() 1545 &disp->rm.objcom); in r535_disp_oneinit() 1737 if (!(rm = kzalloc(sizeof(*rm) + 6 * sizeof(rm->user[0]), GFP_KERNEL))) in r535_disp_new() 1744 rm->uevent = hw->uevent; in r535_disp_new() [all …]
|
| A D | Kbuild | 5 nvkm-y += nvkm/subdev/gsp/rm/r535/rm.o 6 nvkm-y += nvkm/subdev/gsp/rm/r535/gsp.o 7 nvkm-y += nvkm/subdev/gsp/rm/r535/rpc.o 8 nvkm-y += nvkm/subdev/gsp/rm/r535/ctrl.o 9 nvkm-y += nvkm/subdev/gsp/rm/r535/alloc.o 13 nvkm-y += nvkm/subdev/gsp/rm/r535/bar.o 14 nvkm-y += nvkm/subdev/gsp/rm/r535/fbsr.o 15 nvkm-y += nvkm/subdev/gsp/rm/r535/vmm.o 17 nvkm-y += nvkm/subdev/gsp/rm/r535/disp.o 20 nvkm-y += nvkm/subdev/gsp/rm/r535/ce.o [all …]
|
| A D | fifo.c | 67 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear() 180 if (!chan->rm.mthdbuf.ptr) in r535_chan_ramfc_write() 187 &chan->rm.object); in r535_chan_ramfc_write() 266 chan->rm.grctx = vctx; in r535_gr_ctor() 386 switch (rm) { in r535_fifo_xlat_rm_engine_type() 444 engn->rm.size = in r535_fifo_ectx_size() 462 struct nvkm_rm *rm = gsp->rm; in r535_fifo_runl_ctor() local 592 struct nvkm_fifo_func *rm; in r535_fifo_new() local 594 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_fifo_new() 597 rm->dtor = r535_fifo_dtor; in r535_fifo_new() [all …]
|
| A D | bar.c | 168 struct nvkm_bar_func *rm; in r535_bar_new_() local 172 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_bar_new_() 175 rm->dtor = r535_bar_dtor; in r535_bar_new_() 176 rm->oneinit = hw->oneinit; in r535_bar_new_() 177 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_() 178 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_() 179 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_() 180 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_() 184 rm->bar2.vmm = hw->bar2.vmm; in r535_bar_new_() 185 rm->flush = r535_bar_flush; in r535_bar_new_() [all …]
|
| A D | fbsr.c | 309 struct nvkm_instmem_func *rm; in r535_instmem_new() local 312 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_instmem_new() 315 rm->dtor = r535_instmem_dtor; in r535_instmem_new() 316 rm->fini = hw->fini; in r535_instmem_new() 317 rm->memory_new = hw->memory_new; in r535_instmem_new() 318 rm->memory_wrap = hw->memory_wrap; in r535_instmem_new() 319 rm->zero = false; in r535_instmem_new() 320 rm->set_bar0_window_addr = hw->set_bar0_window_addr; in r535_instmem_new() 322 ret = nv50_instmem_new_(rm, device, type, inst, pinstmem); in r535_instmem_new() 324 kfree(rm); in r535_instmem_new()
|
| A D | gr.c | 68 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.subdevice, in r535_gr_promote_ctx() 74 ctrl->hChanClient = vmm->rm.client.object.handle; in r535_gr_promote_ctx() 139 return nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.subdevice, ctrl); in r535_gr_promote_ctx() 159 ret = r535_gr_promote_ctx(gr, false, grc->vmm, grc->mem, grc->vma, &chan->rm.object); in r535_gr_chan_new() 278 struct nvkm_rm *rm = gsp->rm; in r535_gr_oneinit() local 301 ret = rm->api->fifo->chan.alloc(&golden.vmm->rm.device, NVKM_RM_CHAN(0), in r535_gr_oneinit() 302 1, 0, true, rm->api->fifo->rsvd_chids, in r535_gr_oneinit() 315 ret = rm->api->gr->get_ctxbufs_info(gr); in r535_gr_oneinit() 325 ret = nvkm_gsp_rm_alloc(&golden.chan, NVKM_RM_THREED, rm->gpu->gr.class.threed, 0, &threed); in r535_gr_oneinit()
|
| /drivers/gpu/drm/msm/disp/dpu1/ |
| A D | dpu_rm.c | 41 struct dpu_rm *rm, in dpu_rm_init() argument 48 if (!rm || !cat || !mmio) { in dpu_rm_init() 54 memset(rm, 0, sizeof(*rm)); in dpu_rm_init() 430 struct dpu_rm *rm, in _dpu_rm_reserve_ctls() argument 458 if (!rm->ctl_blks[j]) in _dpu_rm_reserve_ctls() 646 if (!rm->cdm_blk) { in _dpu_rm_reserve_cdm() 667 struct dpu_rm *rm, in _dpu_rm_make_reservation() argument 761 struct dpu_rm *rm, in dpu_rm_reserve() argument 797 if (!rm->hw_sspp[i]) in dpu_rm_try_sspp() 985 const struct dpu_rm *rm = global_state->rm; in dpu_rm_print_state() local [all …]
|
| A D | dpu_rm.h | 70 struct dpu_rm *rm, 75 int dpu_rm_reserve(struct dpu_rm *rm, 83 struct dpu_hw_sspp *dpu_rm_reserve_sspp(struct dpu_rm *rm, 91 int dpu_rm_get_assigned_resources(struct dpu_rm *rm, 103 static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx) in dpu_rm_get_intf() argument 105 return rm->hw_intf[intf_idx - INTF_0]; in dpu_rm_get_intf() 113 static inline struct dpu_hw_wb *dpu_rm_get_wb(struct dpu_rm *rm, enum dpu_wb wb_idx) in dpu_rm_get_wb() argument 115 return rm->hw_wb[wb_idx - WB_0]; in dpu_rm_get_wb() 123 static inline struct dpu_hw_sspp *dpu_rm_get_sspp(struct dpu_rm *rm, enum dpu_sspp sspp_idx) in dpu_rm_get_sspp() argument 125 return rm->hw_sspp[sspp_idx - SSPP_NONE]; in dpu_rm_get_sspp()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/ |
| A D | engine.c | 19 struct nvkm_gsp_object rm; member 27 nvkm_gsp_rm_free(&obj->rm); in nvkm_rm_engine_obj_dtor() 40 struct nvkm_rm *rm = chan->client->gsp->rm; in nvkm_rm_engine_obj_new() local 53 ret = rm->api->ce->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 59 ret = rm->api->nvdec->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 62 ret = rm->api->nvenc->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 65 ret = rm->api->nvjpg->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 68 ret = rm->api->ofa->alloc(chan, handle, class, inst, &obj->rm); in nvkm_rm_engine_obj_new() 163 return nvkm_rm_gr_new(rm); in nvkm_rm_engine_new() 168 return nvkm_rm_nvdec_new(rm, inst); in nvkm_rm_engine_new() [all …]
|
| A D | Kbuild | 4 nvkm-y += nvkm/subdev/gsp/rm/client.o 5 nvkm-y += nvkm/subdev/gsp/rm/engine.o 6 nvkm-y += nvkm/subdev/gsp/rm/gr.o 7 nvkm-y += nvkm/subdev/gsp/rm/nvdec.o 8 nvkm-y += nvkm/subdev/gsp/rm/nvenc.o 10 nvkm-y += nvkm/subdev/gsp/rm/tu1xx.o 11 nvkm-y += nvkm/subdev/gsp/rm/ga100.o 12 nvkm-y += nvkm/subdev/gsp/rm/ga1xx.o 13 nvkm-y += nvkm/subdev/gsp/rm/ad10x.o 14 nvkm-y += nvkm/subdev/gsp/rm/gh100.o [all …]
|
| A D | gr.c | 22 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_fini() local 25 if (rm->api->gr->scrubber.fini) in nvkm_rm_gr_fini() 26 rm->api->gr->scrubber.fini(gr); in nvkm_rm_gr_fini() 34 struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm; in nvkm_rm_gr_init() local 38 if (rm->api->gr->scrubber.init) { in nvkm_rm_gr_init() 48 nvkm_rm_gr_new(struct nvkm_rm *rm) in nvkm_rm_gr_new() argument 51 rm->gpu->gr.class.i2m, in nvkm_rm_gr_new() 52 rm->gpu->gr.class.twod, in nvkm_rm_gr_new() 53 rm->gpu->gr.class.threed, in nvkm_rm_gr_new() 54 rm->gpu->gr.class.compute, in nvkm_rm_gr_new() [all …]
|
| A D | nvdec.c | 15 nvkm_rm_nvdec_new(struct nvkm_rm *rm, int inst) in nvkm_rm_nvdec_new() argument 24 ret = nvkm_rm_engine_ctor(nvkm_rm_nvdec_dtor, rm, NVKM_ENGINE_NVDEC, inst, in nvkm_rm_nvdec_new() 25 &rm->gpu->nvdec.class, 1, &nvdec->engine); in nvkm_rm_nvdec_new() 31 rm->device->nvdec[inst] = nvdec; in nvkm_rm_nvdec_new()
|
| A D | nvenc.c | 15 nvkm_rm_nvenc_new(struct nvkm_rm *rm, int inst) in nvkm_rm_nvenc_new() argument 24 ret = nvkm_rm_engine_ctor(nvkm_rm_nvenc_dtor, rm, NVKM_ENGINE_NVENC, inst, in nvkm_rm_nvenc_new() 25 &rm->gpu->nvenc.class, 1, &nvenc->engine); in nvkm_rm_nvenc_new() 31 rm->device->nvenc[inst] = nvenc; in nvkm_rm_nvenc_new()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/vfn/ |
| A D | r535.c | 37 const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu; in r535_vfn_new() 38 struct nvkm_vfn_func *rm; in r535_vfn_new() local 41 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_vfn_new() 44 rm->dtor = r535_vfn_dtor; in r535_vfn_new() 45 rm->intr = &tu102_vfn_intr; in r535_vfn_new() 46 rm->user.addr = 0x030000; in r535_vfn_new() 47 rm->user.size = 0x010000; in r535_vfn_new() 48 rm->user.base.minver = -1; in r535_vfn_new() 49 rm->user.base.maxver = -1; in r535_vfn_new() 50 rm->user.base.oclass = gpu->usermode.class; in r535_vfn_new() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/ |
| A D | disp.c | 86 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_dp_set_indexed_link_rates() 104 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_dp_get_caps() 113 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_dp_get_caps() 137 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_dp_get_caps() 163 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_bl_ctrl() 173 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_disp_get_active() 188 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_disp_get_active() 197 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r570_disp_get_connect_state() 209 nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl); in r570_disp_get_connect_state() 218 ctrl = nvkm_gsp_rm_ctrl_rd(&disp->rm.objcom, in r570_disp_get_supported() [all …]
|
| A D | Kbuild | 2 nvkm-y += nvkm/subdev/gsp/rm/r570/rm.o 3 nvkm-y += nvkm/subdev/gsp/rm/r570/gsp.o 4 nvkm-y += nvkm/subdev/gsp/rm/r570/client.o 5 nvkm-y += nvkm/subdev/gsp/rm/r570/fbsr.o 6 nvkm-y += nvkm/subdev/gsp/rm/r570/disp.o 7 nvkm-y += nvkm/subdev/gsp/rm/r570/fifo.o 8 nvkm-y += nvkm/subdev/gsp/rm/r570/gr.o 9 nvkm-y += nvkm/subdev/gsp/rm/r570/ofa.o
|
| A D | gr.c | 59 ctrl = nvkm_gsp_rm_ctrl_get(&gr->scrubber.vmm->rm.device.subdevice, in r570_gr_scrubber_ctrl() 67 return nvkm_gsp_rm_ctrl_wr(&gr->scrubber.vmm->rm.device.subdevice, ctrl); in r570_gr_scrubber_ctrl() 98 struct nvkm_rm *rm = gsp->rm; in r570_gr_scrubber_init() local 119 0x2000 + device->fifo->rm.mthdbuf_size, 0, true, in r570_gr_scrubber_init() 133 ret = rm->api->fifo->chan.alloc(&gr->scrubber.vmm->rm.device, KGRAPHICS_SCRUBBER_HANDLE_CHANNEL, in r570_gr_scrubber_init() 148 rm->gpu->gr.class.threed, 0, &gr->scrubber.threed); in r570_gr_scrubber_init()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ |
| A D | base.c | 94 nvkm_firmware_put(gsp->fws.rm); in nvkm_gsp_dtor_fws() 95 gsp->fws.rm = NULL; in nvkm_gsp_dtor_fws() 107 kfree(gsp->rm); in nvkm_gsp_dtor() 146 if (fwif->rm) { in nvkm_gsp_new_() 149 gsp->rm = kzalloc(sizeof(*gsp->rm), GFP_KERNEL); in nvkm_gsp_new_() 150 if (!gsp->rm) in nvkm_gsp_new_() 153 gsp->rm->device = device; in nvkm_gsp_new_() 154 gsp->rm->gpu = fwif->func->rm.gpu; in nvkm_gsp_new_() 155 gsp->rm->wpr = fwif->rm->wpr; in nvkm_gsp_new_() 156 gsp->rm->api = fwif->rm->api; in nvkm_gsp_new_()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
| A D | r535.c | 36 struct nvkm_devinit_func *rm; in r535_devinit_new() local 39 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_devinit_new() 42 rm->dtor = r535_devinit_dtor; in r535_devinit_new() 43 rm->post = hw->post; in r535_devinit_new() 44 rm->disable = hw->disable; in r535_devinit_new() 46 ret = nv50_devinit_new_(rm, device, type, inst, pdevinit); in r535_devinit_new() 48 kfree(rm); in r535_devinit_new()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
| A D | r535.c | 71 struct nvkm_fb_func *rm; in r535_fb_new() local 74 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_fb_new() 77 rm->dtor = r535_fb_dtor; in r535_fb_new() 78 rm->sysmem.flush_page_init = hw->sysmem.flush_page_init; in r535_fb_new() 79 rm->vidmem.size = hw->vidmem.size; in r535_fb_new() 80 rm->ram_new = r535_fb_ram_new; in r535_fb_new() 82 ret = nvkm_fb_new_(rm, device, type, inst, pfb); in r535_fb_new() 84 kfree(rm); in r535_fb_new()
|
| /drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
| A D | gsp.h | 75 const struct firmware *rm; member 228 struct nvkm_rm *rm; member 249 struct dentry *rm; member 263 return gsp && (gsp->fws.rm || gsp->fw.img); in nvkm_gsp_rm() 271 return gsp->rm->api->rpc->get(gsp, fn, argc); in nvkm_gsp_rpc_get() 278 return gsp->rm->api->rpc->push(gsp, argv, policy, repc); in nvkm_gsp_rpc_push() 307 gsp->rm->api->rpc->done(gsp, repv); in nvkm_gsp_rpc_done() 350 object->client->gsp->rm->api->ctrl->done(object, repv); in nvkm_gsp_rm_ctrl_done() 365 argv = gsp->rm->api->alloc->get(object, oclass, argc); in nvkm_gsp_rm_alloc_get() 399 object->client->gsp->rm->api->alloc->done(object, repv); in nvkm_gsp_rm_alloc_done() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb/ |
| A D | common.h | 65 #define t1_rx_mode_promisc(rm) (rm->dev->flags & IFF_PROMISC) argument 66 #define t1_rx_mode_allmulti(rm) (rm->dev->flags & IFF_ALLMULTI) argument 67 #define t1_rx_mode_mc_cnt(rm) (netdev_mc_count(rm->dev)) argument 68 #define t1_get_netdev(rm) (rm->dev) argument
|
| /drivers/media/i2c/ |
| A D | ov4689.c | 418 struct regmap *rm = ov4689->regmap; in ov4689_setup_timings() local 421 cci_write(rm, OV4689_REG_H_CROP_START, 8, &ret); in ov4689_setup_timings() 422 cci_write(rm, OV4689_REG_V_CROP_START, 8, &ret); in ov4689_setup_timings() 423 cci_write(rm, OV4689_REG_H_CROP_END, 2711, &ret); in ov4689_setup_timings() 424 cci_write(rm, OV4689_REG_V_CROP_END, 1531, &ret); in ov4689_setup_timings() 429 cci_write(rm, OV4689_REG_H_WIN_OFF, 8, &ret); in ov4689_setup_timings() 430 cci_write(rm, OV4689_REG_V_WIN_OFF, 4, &ret); in ov4689_setup_timings() 432 cci_write(rm, OV4689_REG_VFIFO_CTRL_01, 167, &ret); in ov4689_setup_timings() 439 struct regmap *rm = ov4689->regmap; in ov4689_setup_blc_anchors() local 442 cci_write(rm, OV4689_REG_ANCHOR_LEFT_START, 16, &ret); in ov4689_setup_blc_anchors() [all …]
|