| /drivers/staging/media/atomisp/pci/runtime/isys/src/ |
| A D | virtual_isys.c | 298 if (!acquire_sid(me->stream2mmio_id, &me->stream2mmio_sid_id)) { in create_input_system_channel() 312 &me->ib_buffer)) { in create_input_system_channel() 313 release_sid(me->stream2mmio_id, &me->stream2mmio_sid_id); in create_input_system_channel() 317 if (!acquire_dma_channel(me->dma_id, &me->dma_channel)) { in create_input_system_channel() 318 release_sid(me->stream2mmio_id, &me->stream2mmio_sid_id); in create_input_system_channel() 334 release_dma_channel(me->dma_id, &me->dma_channel); in destroy_input_system_channel() 353 me->csi_rx.backend_id, in create_input_system_input_port() 368 me->csi_rx.backend_id, in create_input_system_input_port() 384 me->csi_rx.backend_id, in create_input_system_input_port() 404 me->csi_rx.backend_id, in create_input_system_input_port() [all …]
|
| /drivers/staging/media/atomisp/pci/ |
| A D | sh_css_params.c | 1372 me = kvmalloc(sizeof(*me), GFP_KERNEL); in ia_css_morph_table_allocate() 1519 me = kvmalloc(sizeof(*me), GFP_KERNEL); in ia_css_isp_3a_statistics_map_allocate() 2139 me = kvcalloc(1, sizeof(*me), GFP_KERNEL); in ia_css_isp_3a_statistics_allocate() 2158 me->size = me->dmem_size + me->vmem_size * 2 + me->hmem_size; in ia_css_isp_3a_statistics_allocate() 2159 me->data_ptr = hmm_alloc(me->size); in ia_css_isp_3a_statistics_allocate() 2166 me->data.dmem.s3a_tbl = me->data_ptr; in ia_css_isp_3a_statistics_allocate() 2168 me->data.vmem.s3a_tbl_hi = me->data_ptr + me->dmem_size; in ia_css_isp_3a_statistics_allocate() 2169 me->data.vmem.s3a_tbl_lo = me->data_ptr + me->dmem_size + me->vmem_size; in ia_css_isp_3a_statistics_allocate() 2172 me->data_hmem.rgby_tbl = me->data_ptr + me->dmem_size + 2 * me->vmem_size; in ia_css_isp_3a_statistics_allocate() 4164 me = kvcalloc(1, sizeof(*me), GFP_KERNEL); in ia_css_3a_statistics_allocate() [all …]
|
| A D | sh_css_host_data.c | 13 struct ia_css_host_data *me; in ia_css_host_data_allocate() local 16 if (!me) in ia_css_host_data_allocate() 18 me->size = (uint32_t)size; in ia_css_host_data_allocate() 19 me->address = kvmalloc(size, GFP_KERNEL); in ia_css_host_data_allocate() 20 if (!me->address) { in ia_css_host_data_allocate() 21 kfree(me); in ia_css_host_data_allocate() 24 return me; in ia_css_host_data_allocate() 29 if (me) { in ia_css_host_data_free() 30 kvfree(me->address); in ia_css_host_data_free() 31 me->address = NULL; in ia_css_host_data_free() [all …]
|
| A D | sh_css_param_shading.c | 331 me = kmalloc(sizeof(*me), GFP_KERNEL); in ia_css_shading_table_alloc() 332 if (!me) in ia_css_shading_table_alloc() 333 return me; in ia_css_shading_table_alloc() 337 me->sensor_width = 0; in ia_css_shading_table_alloc() 338 me->sensor_height = 0; in ia_css_shading_table_alloc() 339 me->fraction_bits = 0; in ia_css_shading_table_alloc() 341 me->data[i] = in ia_css_shading_table_alloc() 344 if (!me->data[i]) { in ia_css_shading_table_alloc() 349 me->data[j] = NULL; in ia_css_shading_table_alloc() 351 kfree(me); in ia_css_shading_table_alloc() [all …]
|
| A D | ia_css_dvs.h | 170 ia_css_isp_dvs_statistics_free(struct ia_css_isp_dvs_statistics *me); 184 ia_css_isp_dvs2_statistics_free(struct ia_css_isp_dvs_statistics *me); 198 ia_css_dvs_statistics_free(struct ia_css_dvs_statistics *me); 212 ia_css_dvs_coefficients_free(struct ia_css_dvs_coefficients *me); 226 ia_css_dvs2_statistics_free(struct ia_css_dvs2_statistics *me); 240 ia_css_dvs2_coefficients_free(struct ia_css_dvs2_coefficients *me); 286 ia_css_isp_dvs_statistics_map_free(struct ia_css_isp_dvs_statistics_map *me);
|
| A D | sh_css.c | 1170 me->mode, in start_pipe() 1883 me = kmalloc(sizeof(*me), GFP_KERNEL); in create_pipe() 1884 if (!me) in create_pipe() 1889 kfree(me); in create_pipe() 1893 err = pipe_generate_pipe_num(me, &me->pipe_num); in create_pipe() 1895 kfree(me); in create_pipe() 1899 *pipe = me; in create_pipe() 2653 assert(me); in add_yuv_scaler_stage() 2710 assert(me); in add_capture_pp_stage() 6706 me->num_stages = 1; in create_host_copy_pipeline() [all …]
|
| A D | ia_css_3a.h | 138 ia_css_isp_3a_statistics_free(struct ia_css_isp_3a_statistics *me); 152 ia_css_3a_statistics_free(struct ia_css_3a_statistics *me); 184 ia_css_isp_3a_statistics_map_free(struct ia_css_isp_3a_statistics_map *me);
|
| A D | atomisp_csi2.c | 161 struct media_entity *me = &sd->entity; in mipi_csi2_init_entities() local 173 me->ops = &csi2_media_ops; in mipi_csi2_init_entities() 174 me->function = MEDIA_ENT_F_VID_IF_BRIDGE; in mipi_csi2_init_entities() 175 ret = media_entity_pads_init(me, CSI2_PADS_NUM, pads); in mipi_csi2_init_entities()
|
| /drivers/staging/media/atomisp/pci/isp/kernels/sdis/sdis_1.0/ |
| A D | ia_css_sdis.host.c | 312 me = kvcalloc(1, sizeof(*me), GFP_KERNEL); in ia_css_isp_dvs_statistics_allocate() 313 if (!me) in ia_css_isp_dvs_statistics_allocate() 324 me->data_ptr = hmm_alloc(me->size); in ia_css_isp_dvs_statistics_allocate() 328 me->hor_proj = me->data_ptr; in ia_css_isp_dvs_statistics_allocate() 330 me->ver_proj = me->data_ptr + hor_size; in ia_css_isp_dvs_statistics_allocate() 334 return me; in ia_css_isp_dvs_statistics_allocate() 353 me = kvmalloc(sizeof(*me), GFP_KERNEL); in ia_css_isp_dvs_statistics_map_allocate() 354 if (!me) { in ia_css_isp_dvs_statistics_map_allocate() 377 return me; in ia_css_isp_dvs_statistics_map_allocate() 386 if (me) { in ia_css_isp_dvs_statistics_map_free() [all …]
|
| /drivers/staging/media/atomisp/pci/isp/kernels/sdis/sdis_2/ |
| A D | ia_css_sdis2.host.c | 277 me = kvcalloc(1, sizeof(*me), GFP_KERNEL); in ia_css_isp_dvs2_statistics_allocate() 278 if (!me) in ia_css_isp_dvs2_statistics_allocate() 287 me->size = 2 * size; in ia_css_isp_dvs2_statistics_allocate() 288 me->data_ptr = hmm_alloc(me->size); in ia_css_isp_dvs2_statistics_allocate() 291 me->hor_proj = me->data_ptr; in ia_css_isp_dvs2_statistics_allocate() 292 me->hor_size = size; in ia_css_isp_dvs2_statistics_allocate() 293 me->ver_proj = me->data_ptr + size; in ia_css_isp_dvs2_statistics_allocate() 294 me->ver_size = size; in ia_css_isp_dvs2_statistics_allocate() 297 return me; in ia_css_isp_dvs2_statistics_allocate() 308 if (me) { in ia_css_isp_dvs2_statistics_free() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | msix.c | 81 struct hfi1_msix_entry *me; in msix_request_irq() local 113 me = &dd->msix_info.msix_entries[nr]; in msix_request_irq() 114 me->irq = irq; in msix_request_irq() 115 me->arg = arg; in msix_request_irq() 116 me->type = type; in msix_request_irq() 119 ret = hfi1_get_irq_affinity(dd, me); in msix_request_irq() 288 struct hfi1_msix_entry *me; in msix_free_irq() local 298 hfi1_put_irq_affinity(dd, me); in msix_free_irq() 301 me->arg = NULL; in msix_free_irq() 342 struct hfi1_msix_entry *me; in msix_netdev_synchronize_irq() local [all …]
|
| /drivers/staging/media/atomisp/pci/runtime/frame/src/ |
| A D | frame.c | 363 if (!me) in ia_css_frame_allocate_with_buffer_size() 372 kvfree(me); in ia_css_frame_allocate_with_buffer_size() 373 me = NULL; in ia_css_frame_allocate_with_buffer_size() 376 *frame = me; in ia_css_frame_allocate_with_buffer_size() 628 if (!me) in frame_allocate_with_data() 637 kvfree(me); in frame_allocate_with_data() 640 *frame = me; in frame_allocate_with_data() 653 struct ia_css_frame *me = kvmalloc(sizeof(*me), GFP_KERNEL); in frame_create() local 655 if (!me) in frame_create() 658 memset(me, 0, sizeof(*me)); in frame_create() [all …]
|
| /drivers/misc/mei/ |
| A D | Makefile | 17 obj-$(CONFIG_INTEL_MEI_ME) += mei-me.o 18 mei-me-objs := pci-me.o 19 mei-me-objs += hw-me.o 22 mei-gsc-objs := gsc-me.o
|
| /drivers/staging/media/atomisp/pci/isp/kernels/dvs/dvs_1.0/ |
| A D | ia_css_dvs.host.c | 225 struct ia_css_host_data *me; in convert_allocate_dvs_6axis_config() local 231 me = ia_css_host_data_allocate((size_t)((DVS_6AXIS_BYTES(binary) / 2) * 3)); in convert_allocate_dvs_6axis_config() 233 if (!me) in convert_allocate_dvs_6axis_config() 246 convert_coords_to_ispparams(me, dvs_6axis_config, in convert_allocate_dvs_6axis_config() 255 convert_coords_to_ispparams(me, dvs_6axis_config, in convert_allocate_dvs_6axis_config() 258 return me; in convert_allocate_dvs_6axis_config() 267 struct ia_css_host_data *me; in store_dvs_6axis_config() local 273 me = convert_allocate_dvs_6axis_config(dvs_6axis_config, in store_dvs_6axis_config() 277 if (!me) in store_dvs_6axis_config() 285 me); in store_dvs_6axis_config() [all …]
|
| /drivers/staging/media/atomisp/pci/hive_isp_css_common/host/ |
| A D | irq.c | 80 unsigned int me = 1U << irq_id; in irq_enable_channel() local 85 mask |= me; in irq_enable_channel() 86 enable |= me; in irq_enable_channel() 87 edge_in |= me; /* rising edge */ in irq_enable_channel() 93 _HRT_IRQ_CONTROLLER_MASK_REG_IDX, mask & ~me); in irq_enable_channel() 102 _HRT_IRQ_CONTROLLER_CLEAR_REG_IDX, me); in irq_enable_channel() 135 unsigned int me = 1U << irq_id; in irq_disable_channel() local 140 mask &= ~me; in irq_disable_channel() 141 enable &= ~me; in irq_disable_channel() 151 _HRT_IRQ_CONTROLLER_CLEAR_REG_IDX, me); in irq_disable_channel()
|
| /drivers/net/ |
| A D | vrf.c | 161 me = kmalloc(sizeof(*me), flags); in vrf_map_elem_alloc() 162 if (!me) in vrf_map_elem_alloc() 165 return me; in vrf_map_elem_alloc() 170 kfree(me); in vrf_map_elem_free() 178 me->users = users; in vrf_map_elem_init() 191 return me; in vrf_map_lookup_elem() 245 if (!me) { in vrf_map_register_dev() 246 me = new_me; in vrf_map_register_dev() 263 users = ++me->users; in vrf_map_register_dev() 293 if (!me) in vrf_map_unregister_dev() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | uvd_v7_0.c | 89 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_get_rptr() 90 return RREG32_SOC15(UVD, ring->me, mmUVD_RB_RPTR); in uvd_v7_0_enc_ring_get_rptr() 123 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_get_wptr() 124 return RREG32_SOC15(UVD, ring->me, mmUVD_RB_WPTR); in uvd_v7_0_enc_ring_get_wptr() 161 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_set_wptr() 162 WREG32_SOC15(UVD, ring->me, mmUVD_RB_WPTR, in uvd_v7_0_enc_ring_set_wptr() 165 WREG32_SOC15(UVD, ring->me, mmUVD_RB_WPTR2, in uvd_v7_0_enc_ring_set_wptr() 448 sprintf(ring->name, "uvd_%d", ring->me); in uvd_v7_0_sw_init() 1295 if (!ring->me) in uvd_v7_0_ring_patch_cs_in_place() 1606 adev->uvd.inst[i].ring.me = i; in uvd_v7_0_set_ring_funcs() [all …]
|
| A D | gfx_v12_0.c | 316 me = 1; in gfx_v12_0_kiq_map_queues() 320 me = 0; in gfx_v12_0_kiq_map_queues() 324 me = 2; in gfx_v12_0_kiq_map_queues() 972 ring->me = me; in gfx_v12_0_gfx_ring_init() 1005 ring->me = mec + 1; in gfx_v12_0_compute_ring_init() 1380 inst = adev->gfx.me.num_me * adev->gfx.me.num_pipe_per_me * in gfx_v12_0_alloc_ip_dump() 1839 if (me != 0) in gfx_v12_0_get_cpg_int_cntl() 1858 if (me != 1) in gfx_v12_0_get_cpc_int_cntl() 4390 switch (ring->me) { in gfx_v12_0_ring_emit_hdp_flush() 4710 if (!me) { in gfx_v12_0_set_gfx_eop_interrupt_state() [all …]
|
| A D | gfx_v11_0.c | 376 me = 1; in gfx11_kiq_map_queues() 380 me = 0; in gfx11_kiq_map_queues() 384 me = 2; in gfx11_kiq_map_queues() 1134 ring->me = me; in gfx_v11_0_gfx_ring_init() 1170 ring->me = mec + 1; in gfx_v11_0_compute_ring_init() 1562 inst = adev->gfx.me.num_me * adev->gfx.me.num_pipe_per_me * in gfx_v11_0_alloc_ip_dump() 2151 if (me != 0) in gfx_v11_0_get_cpg_int_cntl() 2172 if (me != 1) in gfx_v11_0_get_cpc_int_cntl() 5818 switch (ring->me) { in gfx_v11_0_ring_emit_hdp_flush() 6298 if (!me) { in gfx_v11_0_set_gfx_eop_interrupt_state() [all …]
|
| A D | amdgpu_vcn.c | 520 !adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_ring_end_use() 779 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_dec_sw_send_msg() 811 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_dec_sw_send_msg() 909 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_create_msg() 923 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_create_msg() 945 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_create_msg() 976 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_destroy_msg() 990 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_destroy_msg() 1012 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_enc_get_destroy_msg() 1525 if (adev->vcn.inst[ring->me].using_unified_queue) in amdgpu_vcn_ring_reset() [all …]
|
| A D | vcn_v2_0.c | 1486 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.data0, 0)); in vcn_v2_0_dec_ring_insert_start() 1488 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_insert_start() 1523 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.nop, 0)); in vcn_v2_0_dec_ring_insert_nop() 1547 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.data0, 0)); in vcn_v2_0_dec_ring_emit_fence() 1553 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_emit_fence() 1562 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_emit_fence() 1610 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_emit_reg_wait() 1641 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_emit_wreg() 1838 WREG32(adev->vcn.inst[ring->me].external.scratch9, 0xCAFEDEAD); in vcn_v2_0_dec_ring_test_ring() 1842 amdgpu_ring_write(ring, PACKET0(adev->vcn.inst[ring->me].internal.cmd, 0)); in vcn_v2_0_dec_ring_test_ring() [all …]
|
| A D | amdgpu_rlc.c | 193 int me, i, max_me; in amdgpu_gfx_rlc_setup_cp_table() local 201 for (me = 0; me < max_me; me++) { in amdgpu_gfx_rlc_setup_cp_table() 202 if (me == 0) { in amdgpu_gfx_rlc_setup_cp_table() 210 } else if (me == 1) { in amdgpu_gfx_rlc_setup_cp_table() 218 } else if (me == 2) { in amdgpu_gfx_rlc_setup_cp_table() 226 } else if (me == 3) { in amdgpu_gfx_rlc_setup_cp_table() 234 } else if (me == 4) { in amdgpu_gfx_rlc_setup_cp_table()
|
| A D | jpeg_v2_5.c | 446 return RREG32_SOC15(JPEG, ring->me, mmUVD_JRBC_RB_RPTR); in jpeg_v2_5_dec_ring_get_rptr() 463 return RREG32_SOC15(JPEG, ring->me, mmUVD_JRBC_RB_WPTR); in jpeg_v2_5_dec_ring_get_wptr() 481 WREG32_SOC15(JPEG, ring->me, mmUVD_JRBC_RB_WPTR, lower_32_bits(ring->wptr)); in jpeg_v2_5_dec_ring_set_wptr() 500 amdgpu_ring_write(ring, 0x80000000 | (1 << (ring->me * 2 + 14))); in jpeg_v2_6_dec_ring_insert_start() 518 amdgpu_ring_write(ring, (1 << (ring->me * 2 + 14))); in jpeg_v2_6_dec_ring_insert_end() 654 jpeg_v2_5_stop_inst(ring->adev, ring->me); in jpeg_v2_5_ring_reset() 655 jpeg_v2_5_start_inst(ring->adev, ring->me); in jpeg_v2_5_ring_reset() 766 adev->jpeg.inst[i].ring_dec->me = i; in jpeg_v2_5_set_dec_ring_funcs()
|
| /drivers/staging/media/atomisp/pci/runtime/pipeline/src/ |
| A D | pipeline.c | 719 me->pipe_id, continuous); in ia_css_pipeline_configure_inout_port() 720 switch (me->pipe_id) { in ia_css_pipeline_configure_inout_port() 723 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 726 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 731 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 735 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 748 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 752 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 757 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() 760 SH_CSS_PIPE_PORT_CONFIG_SET(me->inout_port_config, in ia_css_pipeline_configure_inout_port() [all …]
|
| /drivers/media/platform/samsung/exynos4-is/ |
| A D | media-dev.h | 155 static inline struct fimc_md *entity_to_fimc_mdev(struct media_entity *me) in entity_to_fimc_mdev() argument 157 return me->graph_obj.mdev == NULL ? NULL : in entity_to_fimc_mdev() 158 container_of(me->graph_obj.mdev, struct fimc_md, media_dev); in entity_to_fimc_mdev()
|