| /drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| A D | crypto.c | 23 ({ typeof(bulk) _bulk = (bulk); \ 296 bulk = kzalloc(sizeof(*bulk), GFP_KERNEL); in mlx5_crypto_dek_bulk_create() 297 if (!bulk) in mlx5_crypto_dek_bulk_create() 324 return bulk; in mlx5_crypto_dek_bulk_create() 329 kfree(bulk); in mlx5_crypto_dek_bulk_create() 379 if (bulk) { in mlx5_crypto_dek_pool_pop() 408 bulk->avail_start = bulk->num_deks; in mlx5_crypto_dek_pool_pop() 429 struct mlx5_crypto_dek_bulk *bulk = dek->bulk; in mlx5_crypto_dek_free_locked() local 443 if (!bulk->avail_deks && !bulk->in_use_deks) in mlx5_crypto_dek_free_locked() 572 bulk->avail_deks = bulk->num_deks; in mlx5_crypto_dek_pool_reset_synced() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | fs_pool.c | 31 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount() 78 struct mlx5_fs_bulk *bulk; in mlx5_fs_pool_cleanup() local 82 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 84 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 86 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 107 fs_pool->available_units -= bulk->bulk_len; in mlx5_fs_pool_free_bulk() 108 fs_pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_free_bulk() 184 if (bulk_free_amount == bulk->bulk_len) { in mlx5_fs_pool_release_index() 185 list_del(&bulk->pool_list); in mlx5_fs_pool_release_index() 187 mlx5_fs_pool_free_bulk(fs_pool, bulk); in mlx5_fs_pool_release_index() [all …]
|
| A D | fs_counters.c | 166 if (counter->bulk) in mlx5_fc_release() 417 static void mlx5_fc_init(struct mlx5_fc *counter, struct mlx5_fc_bulk *bulk, in mlx5_fc_init() argument 420 counter->bulk = bulk; in mlx5_fc_init() 426 return counter->bulk->base_id; in mlx5_fc_get_base_id() 526 struct mlx5_fs_bulk *fs_bulk = &fc->bulk->fs_bulk; in mlx5_fc_pool_release_counter() 531 pool_index.index = fc->id - fc->bulk->base_id; in mlx5_fc_pool_release_counter() 565 counter->bulk = fc_bulk; in mlx5_fc_local_create() 575 kfree(counter->bulk); in mlx5_fc_local_destroy()
|
| A D | fs_pool.h | 23 int (*bulk_destroy)(struct mlx5_core_dev *dev, struct mlx5_fs_bulk *bulk); 45 int mlx5_fs_bulk_get_free_amount(struct mlx5_fs_bulk *bulk);
|
| /drivers/gpu/drm/ttm/ |
| A D | ttm_resource.c | 45 cursor->bulk = NULL; in ttm_resource_cursor_clear_bulk() 57 if (WARN_ON_ONCE(bulk != cursor->bulk)) { in ttm_resource_cursor_move_bulk_tail() 126 memset(bulk, 0, sizeof(*bulk)); in ttm_lru_bulk_move_init() 127 INIT_LIST_HEAD(&bulk->cursor_list); in ttm_lru_bulk_move_init() 140 struct ttm_lru_bulk_move *bulk) in ttm_lru_bulk_move_fini() argument 143 ttm_bulk_move_drop_cursors(bulk); in ttm_lru_bulk_move_fini() 160 ttm_bulk_move_adjust_cursors(bulk); in ttm_lru_bulk_move_tail() 624 bulk = bo->bulk_move; in ttm_resource_cursor_check_bulk() 626 if (cursor->bulk != bulk) { in ttm_resource_cursor_check_bulk() 627 if (bulk) { in ttm_resource_cursor_check_bulk() [all …]
|
| /drivers/staging/vc04_services/vchiq-mmal/ |
| A D | mmal-vchiq.c | 150 } bulk; /* bulk data */ member 270 msg_context->u.bulk.port->buffer_cb(msg_context->u.bulk.instance, in buffer_work_cb() 271 msg_context->u.bulk.port, in buffer_work_cb() 367 msg_context->u.bulk.buffer_used = in inline_receive() 397 msg_context->u.bulk.port = port; in buffer_from_host() 398 msg_context->u.bulk.buffer = buf; in buffer_from_host() 477 msg_context->u.bulk.mmal_flags = in buffer_to_host_cb() 490 msg_context->u.bulk.status = in buffer_to_host_cb() 498 msg_context->u.bulk.status = 0; in buffer_to_host_cb() 503 msg_context->u.bulk.status = in buffer_to_host_cb() [all …]
|
| /drivers/staging/vc04_services/interface/ |
| A D | TESTING | 49 Testing bulk transfer for alignment. 50 Testing bulk transfer at PAGE_SIZE. 61 vchi bulk (size 0, 0 async, 0 oneway) -> 546.000000us 62 vchi bulk (size 0, 0 oneway) -> 230.000000us 65 vchi bulk (size 0, 0 async, 0 oneway) -> 296.000000us 66 vchi bulk (size 0, 0 oneway) -> 266.000000us 68 vchi bulk (size 0, 0 oneway) -> 456.000000us 70 vchi bulk (size 0, 0 oneway) -> 640.000000us 72 vchi bulk (size 0, 0 oneway) -> 2309.000000us 78 vchi bulk (size 0, 0 oneway) -> nanus [all …]
|
| /drivers/staging/vc04_services/interface/vchiq_arm/ |
| A D | vchiq_core.c | 468 if (bulk) { in make_service_callback() 1330 bulk->actual); in service_notify_bulk() 1334 bulk->actual); in service_notify_bulk() 1518 if (bulk->offset) in create_pagelist() 1569 if (bulk->offset) { in create_pagelist() 1778 if (bulk && bulk->remote_data && bulk->actual) in vchiq_complete_bulk() 1817 service->remoteport, bulk->size, bulk->remote_size); in abort_outstanding_bulks() 1821 bulk->size = 0; in abort_outstanding_bulks() 2125 localport, bulk->actual, &bulk->dma_addr); in parse_message() 3096 dir_char, bulk->size, &bulk->dma_addr, bulk->cb_data); in vchiq_bulk_xfer_queue_msg_killable() [all …]
|
| A D | vchiq_arm.c | 598 struct vchiq_bulk *bulk = waiter->bulk_waiter.bulk; in vchiq_blocking_bulk_transfer() local 600 if (bulk) { in vchiq_blocking_bulk_transfer() 603 if ((bulk->dma_addr != (dma_addr_t)(uintptr_t)bulk_params->dma_addr) || in vchiq_blocking_bulk_transfer() 604 (bulk->size != bulk_params->size)) { in vchiq_blocking_bulk_transfer() 610 bulk->waiter = NULL; in vchiq_blocking_bulk_transfer() 623 if ((ret != -EAGAIN) || fatal_signal_pending(current) || !waiter->bulk_waiter.bulk) { in vchiq_blocking_bulk_transfer() 624 struct vchiq_bulk *bulk = waiter->bulk_waiter.bulk; in vchiq_blocking_bulk_transfer() local 626 if (bulk) { in vchiq_blocking_bulk_transfer() 629 bulk->waiter = NULL; in vchiq_blocking_bulk_transfer()
|
| A D | vchiq_core.h | 445 struct vchiq_bulk *bulk; member 506 struct vchiq_bulk *bulk); 510 struct vchiq_bulk *bulk);
|
| /drivers/net/ethernet/intel/libeth/ |
| A D | tx.c | 12 DEFINE_STATIC_CALL_NULL(bulk, libeth_xdp_return_buff_bulk); 27 __libeth_xdp_complete_tx(sqe, cp, static_call(bulk), in libeth_tx_complete_any() 38 static_call_update(bulk, ops ? ops->bulk : NULL); in libeth_attach_xdp()
|
| A D | xdp.c | 126 const struct libeth_xdp_tx_frame *pos = &bq->bulk[sent]; in libeth_xdp_tx_exception() 133 memmove(bq->bulk, pos, left * sizeof(*bq->bulk)); in libeth_xdp_tx_exception() 431 .bulk = libeth_xdp_return_buff_bulk,
|
| /drivers/gpu/drm/msm/ |
| A D | msm_io_utils.c | 17 struct clk *msm_clk_bulk_get_clock(struct clk_bulk_data *bulk, int count, in msm_clk_bulk_get_clock() argument 25 for (i = 0; bulk && i < count; i++) { in msm_clk_bulk_get_clock() 26 if (!strcmp(bulk[i].id, name) || !strcmp(bulk[i].id, n)) in msm_clk_bulk_get_clock() 27 return bulk[i].clk; in msm_clk_bulk_get_clock()
|
| A D | msm_mdss.c | 373 struct clk_bulk_data *bulk; in mdp5_mdss_parse_clock() local 380 bulk = devm_kcalloc(&pdev->dev, MDP5_MDSS_NUM_CLOCKS, sizeof(struct clk_bulk_data), GFP_KERNEL); in mdp5_mdss_parse_clock() 381 if (!bulk) in mdp5_mdss_parse_clock() 384 bulk[num_clocks++].id = "iface"; in mdp5_mdss_parse_clock() 385 bulk[num_clocks++].id = "bus"; in mdp5_mdss_parse_clock() 386 bulk[num_clocks++].id = "vsync"; in mdp5_mdss_parse_clock() 388 ret = devm_clk_bulk_get_optional(&pdev->dev, num_clocks, bulk); in mdp5_mdss_parse_clock() 392 *clocks = bulk; in mdp5_mdss_parse_clock()
|
| /drivers/media/usb/uvc/ |
| A D | uvc_video.c | 1392 nbytes = min(stream->bulk.max_payload_size - stream->bulk.payload_size, in uvc_video_encode_data() 1592 if (stream->bulk.header_size == 0 && !stream->bulk.skip_payload) { in uvc_video_decode_bulk() 1628 stream->bulk.payload_size >= stream->bulk.max_payload_size) { in uvc_video_decode_bulk() 1631 stream->bulk.payload_size); in uvc_video_decode_bulk() 1636 stream->bulk.header_size = 0; in uvc_video_decode_bulk() 1637 stream->bulk.skip_payload = 0; in uvc_video_decode_bulk() 1672 stream->bulk.payload_size == stream->bulk.max_payload_size) { in uvc_video_encode_bulk() 1681 stream->bulk.header_size = 0; in uvc_video_encode_bulk() 2009 stream->bulk.header_size = 0; in uvc_video_start_transfer() 2010 stream->bulk.skip_payload = 0; in uvc_video_start_transfer() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | fs_hws_pools.c | 128 pr_bulk->prs_data[i].bulk = pr_bulk; in mlx5_fs_hws_pr_bulk_create() 225 struct mlx5_fs_bulk *fs_bulk = &pr_data->bulk->fs_bulk; in mlx5_fs_hws_pr_pool_release_pr() 237 return pr_data->bulk->hws_action; in mlx5_fs_hws_pr_get_action() 282 mh_bulk->mhs_data[i].bulk = mh_bulk; in mlx5_fs_hws_mh_bulk_create() 374 struct mlx5_fs_bulk *fs_bulk = &mh_data->bulk->fs_bulk; in mlx5_fs_hws_mh_pool_release_mh() 409 struct mlx5_fc_bulk *fc_bulk = counter->bulk; in mlx5_fc_get_hws_action() 420 mlx5_fs_put_hws_action(&counter->bulk->hws_data); in mlx5_fc_put_hws_action()
|
| A D | fs_hws_pools.h | 21 struct mlx5_fs_hws_pr_bulk *bulk; member 40 struct mlx5_fs_hws_mh_bulk *bulk; member
|
| /drivers/remoteproc/ |
| A D | qcom_wcnss.c | 447 struct regulator_bulk_data *bulk; in wcnss_init_regulators() local 465 bulk = devm_kcalloc(wcnss->dev, in wcnss_init_regulators() 468 if (!bulk) in wcnss_init_regulators() 472 bulk[i].supply = info[i].name; in wcnss_init_regulators() 474 ret = devm_regulator_bulk_get(wcnss->dev, num_vregs, bulk); in wcnss_init_regulators() 480 regulator_set_voltage(bulk[i].consumer, in wcnss_init_regulators() 485 regulator_set_load(bulk[i].consumer, info[i].load_uA); in wcnss_init_regulators() 488 wcnss->vregs = bulk; in wcnss_init_regulators()
|
| /drivers/media/usb/dvb-usb-v2/ |
| A D | usb_urb.c | 155 stream->props.u.bulk.buffersize, in usb_urb_alloc_bulk_urbs() 260 buf_size = stream->props.u.bulk.buffersize; in usb_urb_reconfig() 281 props->u.bulk.buffersize == in usb_urb_reconfig() 282 stream->props.u.bulk.buffersize) in usb_urb_reconfig() 326 stream->props.u.bulk.buffersize); in usb_urb_initv2()
|
| /drivers/media/usb/dvb-usb/ |
| A D | dtt200u.c | 213 .bulk = { 265 .bulk = { 317 .bulk = { 369 .bulk = {
|
| A D | dibusb-mb.c | 224 .bulk = { 314 .bulk = { 383 .bulk = { 445 .bulk = {
|
| A D | cxusb.c | 1743 .bulk = { 1795 .bulk = { 1854 .bulk = { 1922 .bulk = { 1981 .bulk = { 2033 .bulk = { 2087 .bulk = { 2144 .bulk = { 2198 .bulk = { 2244 .bulk = { [all …]
|
| A D | m920x.c | 963 .bulk = { 1011 .bulk = { 1074 .bulk = { 1098 .bulk = { 1140 .bulk = { 1251 .bulk = {
|
| /drivers/usb/core/ |
| A D | devio.c | 1300 len1 = bulk->len; in do_proc_bulk() 1304 if (bulk->ep & USB_DIR_IN) in do_proc_bulk() 1336 tmo = bulk->timeout; in do_proc_bulk() 1337 if (bulk->ep & 0x80) { in do_proc_bulk() 1375 struct usbdevfs_bulktransfer bulk; in proc_bulk() local 1377 if (copy_from_user(&bulk, arg, sizeof(bulk))) in proc_bulk() 1379 return do_proc_bulk(ps, &bulk); in proc_bulk() 2154 struct usbdevfs_bulktransfer bulk; in proc_bulk_compat() local 2157 if (get_user(bulk.ep, &p32->ep) || in proc_bulk_compat() 2162 bulk.data = compat_ptr(addr); in proc_bulk_compat() [all …]
|
| /drivers/iio/magnetometer/ |
| A D | ak8974.c | 210 __le16 bulk; in ak8974_get_u16_val() local 212 ret = regmap_bulk_read(ak8974->map, reg, &bulk, 2); in ak8974_get_u16_val() 215 *val = le16_to_cpu(bulk); in ak8974_get_u16_val() 222 __le16 bulk = cpu_to_le16(val); in ak8974_set_u16_val() local 224 return regmap_bulk_write(ak8974->map, reg, &bulk, 2); in ak8974_set_u16_val()
|