| /drivers/gpu/drm/imagination/ |
| A D | pvr_drv.c | 1181 if (usr_stride > obj_size && in pvr_set_uobj() 1182 clear_user(u64_to_user_ptr(usr_ptr + obj_size), usr_stride - obj_size)) { in pvr_set_uobj() 1205 if (obj_size == in->stride) { in pvr_get_uobj_array() 1207 (unsigned long)obj_size * in->count)) in pvr_get_uobj_array() 1218 out_ptr += obj_size; in pvr_get_uobj_array() 1242 if (obj_size == out->stride) { in pvr_set_uobj_array() 1244 (unsigned long)obj_size * out->count)) in pvr_set_uobj_array() 1247 u32 cpy_elem_size = min_t(u32, out->stride, obj_size); in pvr_set_uobj_array() 1255 out_ptr += obj_size; in pvr_set_uobj_array() 1259 if (out->stride > obj_size && in pvr_set_uobj_array() [all …]
|
| A D | pvr_drv.h | 21 int pvr_get_uobj(u64 usr_ptr, u32 usr_size, u32 min_size, u32 obj_size, void *out); 22 int pvr_set_uobj(u64 usr_ptr, u32 usr_size, u32 min_size, u32 obj_size, const void *in); 23 int pvr_get_uobj_array(const struct drm_pvr_obj_array *in, u32 min_stride, u32 obj_size, 25 int pvr_set_uobj_array(const struct drm_pvr_obj_array *out, u32 min_stride, u32 obj_size,
|
| /drivers/net/ethernet/mellanox/mlx4/ |
| A D | icm.c | 261 (MLX4_TABLE_CHUNK_SIZE / table->obj_size); in mlx4_table_get() 299 i = (obj & (table->num_obj - 1)) / (MLX4_TABLE_CHUNK_SIZE / table->obj_size); in mlx4_table_put() 328 idx = (u64) (obj & (table->num_obj - 1)) * table->obj_size; in mlx4_table_find() 385 int inc = MLX4_TABLE_CHUNK_SIZE / table->obj_size; in mlx4_table_get_range() 411 for (i = start; i <= end; i += MLX4_TABLE_CHUNK_SIZE / table->obj_size) in mlx4_table_put_range() 416 u64 virt, int obj_size, u32 nobj, int reserved, in mlx4_init_icm_table() argument 425 obj_per_chunk = MLX4_TABLE_CHUNK_SIZE / obj_size; in mlx4_init_icm_table() 436 table->obj_size = obj_size; in mlx4_init_icm_table() 441 size = (u64) nobj * obj_size; in mlx4_init_icm_table() 442 for (i = 0; i * MLX4_TABLE_CHUNK_SIZE < reserved * obj_size; ++i) { in mlx4_init_icm_table()
|
| A D | icm.h | 91 u64 virt, int obj_size, u32 nobj, int reserved,
|
| /drivers/net/can/spi/mcp251xfd/ |
| A D | mcp251xfd-ring.c | 349 priv->rx_obj_num_coalesce_irq, rx_ring->obj_size, in mcp251xfd_ring_init() 350 priv->rx_obj_num_coalesce_irq * rx_ring->obj_size); in mcp251xfd_ring_init() 360 rx_ring->obj_size, in mcp251xfd_ring_init() 362 rx_ring->obj_size); in mcp251xfd_ring_init() 368 rx_ring->obj_num, rx_ring->obj_size, in mcp251xfd_ring_init() 369 rx_ring->obj_num * rx_ring->obj_size); in mcp251xfd_ring_init() 377 priv->tx->obj_num, priv->tx->obj_size, in mcp251xfd_ring_init() 378 priv->tx->obj_num * priv->tx->obj_size); in mcp251xfd_ring_init() 515 tx_ring->obj_size = tx_obj_size; in mcp251xfd_ring_alloc() 539 rx_ring->obj_size = rx_obj_size; in mcp251xfd_ring_alloc()
|
| A D | mcp251xfd-rx.c | 43 *rx_tail = fifo_ua / ring->obj_size; in mcp251xfd_rx_tail_get_from_chip() 223 len * ring->obj_size / val_bytes); in mcp251xfd_rx_obj_read() 284 i * ring->obj_size); in mcp251xfd_handle_rxif_ring()
|
| A D | mcp251xfd.h | 547 u8 obj_size; member 565 u8 obj_size; member 850 return ring->base + ring->obj_size * n; in mcp251xfd_get_tx_obj_addr() 856 return ring->base + ring->obj_size * n; in mcp251xfd_get_rx_obj_addr() 916 tx_ring->obj_size; in mcp251xfd_get_tx_nr_by_addr()
|
| A D | mcp251xfd-dump.c | 176 .val = rx->obj_size, in mcp251xfd_dump_rx_ring_one() 219 .val = tx->obj_size, in mcp251xfd_dump_tx_ring()
|
| /drivers/infiniband/hw/mthca/ |
| A D | mthca_memfree.c | 224 int i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; in mthca_table_get() 264 i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; in mthca_table_put() 290 idx = (obj & (table->num_obj - 1)) * table->obj_size; in mthca_table_find() 324 int inc = MTHCA_TABLE_CHUNK_SIZE / table->obj_size; in mthca_table_get_range() 352 for (i = start; i <= end; i += MTHCA_TABLE_CHUNK_SIZE / table->obj_size) in mthca_table_put_range() 357 u64 virt, int obj_size, in mthca_alloc_icm_table() argument 367 obj_per_chunk = MTHCA_TABLE_CHUNK_SIZE / obj_size; in mthca_alloc_icm_table() 377 table->obj_size = obj_size; in mthca_alloc_icm_table() 385 for (i = 0; i * MTHCA_TABLE_CHUNK_SIZE < reserved * obj_size; ++i) { in mthca_alloc_icm_table() 387 if ((i + 1) * MTHCA_TABLE_CHUNK_SIZE > nobj * obj_size) in mthca_alloc_icm_table() [all …]
|
| A D | mthca_memfree.h | 67 int obj_size; member 87 u64 virt, int obj_size,
|
| /drivers/infiniband/hw/hns/ |
| A D | hns_roce_hem.c | 226 table_idx = *obj / (chunk_size / table->obj_size); in hns_roce_calc_hem_mhop() 534 i = obj / (table->table_chunk_size / table->obj_size); in hns_roce_table_get() 659 i = obj / (table->table_chunk_size / table->obj_size); in hns_roce_table_put() 693 obj_per_chunk = table->table_chunk_size / table->obj_size; in hns_roce_table_find() 696 dma_offset = offset = idx_offset * table->obj_size; in hns_roce_table_find() 730 unsigned long obj_size, unsigned long nobj) in hns_roce_init_hem_table() argument 737 obj_per_chunk = table->table_chunk_size / obj_size; in hns_roce_init_hem_table() 759 obj_per_chunk = buf_chunk_size / obj_size; in hns_roce_init_hem_table() 806 table->obj_size = obj_size; in hns_roce_init_hem_table() 845 obj = i * buf_chunk_size / table->obj_size; in hns_roce_cleanup_mhop_hem_table() [all …]
|
| A D | hns_roce_hem.h | 98 unsigned long obj_size, unsigned long nobj);
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | qat_uclo.c | 1844 unsigned int *obj_size) in qat_uclo_seek_obj_inside_mof() argument 1853 *obj_size = obj_hdr[i].obj_size; in qat_uclo_seek_obj_inside_mof() 1995 char **obj_ptr, unsigned int *obj_size) in qat_uclo_map_mof_obj() argument 2007 if (obj_size) in qat_uclo_map_mof_obj() 2008 *obj_size = mof_size; in qat_uclo_map_mof_obj() 2041 obj_ptr, obj_size); in qat_uclo_map_mof_obj() 2048 u32 obj_size; in qat_uclo_map_obj() local 2059 &obj_addr, &obj_size); in qat_uclo_map_obj() 2064 obj_size = mem_size; in qat_uclo_map_obj() 2068 qat_uclo_map_suof_obj(handle, obj_addr, obj_size) : in qat_uclo_map_obj() [all …]
|
| A D | icp_qat_uclo.h | 577 unsigned int obj_size; member
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_drv.c | 90 u32 obj_size) in panthor_get_uobj_array() argument 104 out_alloc = kvmalloc_array(in->count, obj_size, GFP_KERNEL); in panthor_get_uobj_array() 108 if (obj_size == in->stride) { in panthor_get_uobj_array() 111 (unsigned long)obj_size * in->count)) in panthor_get_uobj_array() 119 ret = copy_struct_from_user(out_ptr, obj_size, in_ptr, in->stride); in panthor_get_uobj_array() 123 out_ptr += obj_size; in panthor_get_uobj_array()
|
| /drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| A D | chtls_cm.h | 127 chtls_tcp_ops->obj_size = sizeof(struct tcp_request_sock); in chtls_init_rsk_ops()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgv_sriovmsg.h | 425 unsigned int amd_sriov_msg_checksum(void *obj, unsigned long obj_size, unsigned int key,
|
| A D | amdgpu_virt.c | 253 unsigned long obj_size, in amd_sriov_msg_checksum() argument 263 for (i = 0; i < obj_size; ++i) in amd_sriov_msg_checksum()
|
| /drivers/gpu/drm/ |
| A D | drm_gem.c | 1118 int drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size, in drm_gem_mmap_obj() argument 1124 if (obj_size < vma->vm_end - vma->vm_start) in drm_gem_mmap_obj()
|
| /drivers/gpu/drm/radeon/ |
| A D | radeon_asic.h | 92 uint32_t offset, uint32_t obj_size); 340 uint32_t offset, uint32_t obj_size);
|
| /drivers/net/ppp/ |
| A D | pptp.c | 619 .obj_size = sizeof(struct pppox_sock),
|
| A D | pppoe.c | 528 .obj_size = sizeof(struct pppox_sock),
|
| /drivers/infiniband/core/ |
| A D | rdma_core.c | 275 uobj = kzalloc(obj->type_attrs->obj_size, GFP_KERNEL); in alloc_uobj()
|
| /drivers/isdn/mISDN/ |
| A D | socket.c | 19 .obj_size = sizeof(struct mISDN_sock)
|
| /drivers/net/ |
| A D | tap.c | 36 .obj_size = sizeof(struct tap_queue),
|