| /drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_binding.c | 762 cbs->bind_cmd_buffer[cbs->bind_cmd_count++] = in vmw_collect_view_ids() 798 cbs->bind_cmd_buffer[cbs->bind_cmd_count++] = in vmw_collect_dirty_view_ids() 849 cbs->bind_first_slot, cbs->bind_cmd_count); in vmw_emit_set_sr() 879 if (cbs->ds_view.bi.ctx && !cbs->ds_view.bi.scrubbed) in vmw_emit_set_rt() 1061 vmw_collect_dirty_vbs(cbs, loc, cbs->dirty_vb, in vmw_emit_set_vb() 1080 cbs->bind_first_slot, cbs->bind_cmd_count); in vmw_emit_set_vb() 1095 vmw_collect_view_ids(cbs, loc, vmw_max_num_uavs(cbs->dev_priv)); in vmw_emit_set_uav() 1363 cbs = vzalloc(sizeof(*cbs)); in vmw_binding_state_alloc() 1364 if (!cbs) { in vmw_binding_state_alloc() 1371 return cbs; in vmw_binding_state_alloc() [all …]
|
| A D | vmwgfx_binding.h | 217 extern void vmw_binding_add(struct vmw_ctx_binding_state *cbs, 220 extern void vmw_binding_cb_offset_update(struct vmw_ctx_binding_state *cbs, 222 extern void vmw_binding_add_uav_index(struct vmw_ctx_binding_state *cbs, 229 extern int vmw_binding_rebind_all(struct vmw_ctx_binding_state *cbs); 230 extern void vmw_binding_state_kill(struct vmw_ctx_binding_state *cbs); 231 extern void vmw_binding_state_scrub(struct vmw_ctx_binding_state *cbs); 234 extern void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs); 236 vmw_binding_state_list(struct vmw_ctx_binding_state *cbs); 237 extern void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs);
|
| A D | vmwgfx_context.c | 38 struct vmw_ctx_binding_state *cbs; member 156 vmw_binding_state_kill(uctx->cbs); in vmw_hw_context_destroy() 207 uctx->cbs = vmw_binding_state_alloc(dev_priv); in vmw_gb_context_init() 208 if (IS_ERR(uctx->cbs)) { in vmw_gb_context_init() 209 ret = PTR_ERR(uctx->cbs); in vmw_gb_context_init() 393 vmw_binding_state_scrub(uctx->cbs); in vmw_gb_context_unbind() 559 vmw_binding_state_scrub(uctx->cbs); in vmw_dx_context_scrub_cotables() 692 if (ctx->cbs) in vmw_user_context_free() 693 vmw_binding_state_free(ctx->cbs); in vmw_user_context_free() 811 return vmw_binding_state_list(uctx->cbs); in vmw_context_binding_list() [all …]
|
| /drivers/net/ethernet/microchip/lan966x/ |
| A D | lan966x_cbs.c | 9 u32 cir, cbs; in lan966x_cbs_add() local 20 cbs = (qopt->idleslope - qopt->sendslope) * in lan966x_cbs_add() 29 cbs = DIV_ROUND_UP(cbs, 4096); in lan966x_cbs_add() 31 cbs = cbs ?: 1; in lan966x_cbs_add() 35 cbs > GENMASK(6, 0)) in lan966x_cbs_add() 45 QSYS_CIR_CFG_CIR_BURST_SET(cbs), in lan966x_cbs_add()
|
| A D | lan966x_tbf.c | 11 u32 cir, cbs; in lan966x_tbf_add() local 26 cbs = qopt->replace_params.max_size; in lan966x_tbf_add() 33 cbs = DIV_ROUND_UP(cbs, 4096); in lan966x_tbf_add() 35 cbs = cbs ?: 1; in lan966x_tbf_add() 39 cbs > GENMASK(6, 0)) in lan966x_tbf_add() 49 QSYS_CIR_CFG_CIR_BURST_SET(cbs), in lan966x_tbf_add()
|
| /drivers/net/ethernet/mscc/ |
| A D | ocelot_police.c | 27 u32 cir = 0, cbs = 0, pir = 0, pbs = 0; in qos_policer_conf_set() local 48 cbs = conf->cbs; in qos_policer_conf_set() 49 if (cir == 0 && cbs == 0) { in qos_policer_conf_set() 55 cbs = DIV_ROUND_UP(cbs, 4096); in qos_policer_conf_set() 56 cbs = (cbs ? cbs : 1); /* No zero burst size */ in qos_policer_conf_set() 125 if (cbs > cbs_max) { in qos_policer_conf_set() 128 pol_ix, cbs, cbs_max); in qos_policer_conf_set() 151 ANA_POL_CIR_CFG_CIR_BURST(cbs), in qos_policer_conf_set()
|
| A D | ocelot_police.h | 28 u32 cbs; /* CBS in bytes/frames (ignored in SLB mode) */ member
|
| /drivers/net/ethernet/qlogic/qed/ |
| A D | qed_ll2.c | 106 if (cdev->ll2->cbs && cdev->ll2->cbs->tx_cb) in qed_ll2b_complete_tx_packet() 225 if (cdev->ll2->cbs && cdev->ll2->cbs->rx_cb) { in qed_ll2b_complete_rx_packet() 401 p_ll2_conn->cbs.tx_comp_cb(p_ll2_conn->cbs.cookie, in qed_ll2_txq_completion() 471 p_ll2_conn->cbs.slowpath_cb(p_ll2_conn->cbs.cookie, in qed_ll2_handle_slowpath() 1329 if (!cbs || (!cbs->rx_comp_cb || in qed_ll2_set_cbs() 1331 !cbs->tx_comp_cb || !cbs->tx_release_cb || !cbs->cookie)) in qed_ll2_set_cbs() 1334 p_ll2_info->cbs.rx_comp_cb = cbs->rx_comp_cb; in qed_ll2_set_cbs() 1335 p_ll2_info->cbs.rx_release_cb = cbs->rx_release_cb; in qed_ll2_set_cbs() 1336 p_ll2_info->cbs.tx_comp_cb = cbs->tx_comp_cb; in qed_ll2_set_cbs() 1338 p_ll2_info->cbs.slowpath_cb = cbs->slowpath_cb; in qed_ll2_set_cbs() [all …]
|
| A D | qed_iwarp.c | 2641 struct qed_ll2_cbs cbs; in qed_iwarp_ll2_start() local 2661 cbs.rx_comp_cb = qed_iwarp_ll2_comp_syn_pkt; in qed_iwarp_ll2_start() 2662 cbs.rx_release_cb = qed_iwarp_ll2_rel_rx_pkt; in qed_iwarp_ll2_start() 2663 cbs.tx_comp_cb = qed_iwarp_ll2_comp_tx_pkt; in qed_iwarp_ll2_start() 2664 cbs.tx_release_cb = qed_iwarp_ll2_rel_tx_pkt; in qed_iwarp_ll2_start() 2665 cbs.slowpath_cb = NULL; in qed_iwarp_ll2_start() 2666 cbs.cookie = p_hwfn; in qed_iwarp_ll2_start() 2679 data.cbs = &cbs; in qed_iwarp_ll2_start() 2728 cbs.rx_comp_cb = qed_iwarp_ll2_comp_mpa_pkt; in qed_iwarp_ll2_start() 2729 cbs.slowpath_cb = qed_iwarp_ll2_slowpath; in qed_iwarp_ll2_start() [all …]
|
| A D | qed_ll2.h | 113 struct qed_ll2_cbs cbs; member
|
| /drivers/misc/sgi-gru/ |
| A D | gru_instructions.h | 628 struct gru_control_block_status *cbs = (void *)cb; in gru_get_cb_status() local 630 return cbs->istatus; in gru_get_cb_status() 636 struct gru_control_block_status *cbs = (void *)cb; in gru_get_cb_message_queue_substatus() local 638 return cbs->isubstatus & CBSS_MSG_QUEUE_MASK; in gru_get_cb_message_queue_substatus() 644 struct gru_control_block_status *cbs = (void *)cb; in gru_get_cb_substatus() local 646 return cbs->isubstatus; in gru_get_cb_substatus() 657 struct gru_control_block_status *cbs = (void *)cb; in gru_check_status() local 660 ret = cbs->istatus; in gru_check_status()
|
| /drivers/net/dsa/sja1105/ |
| A D | sja1105_main.c | 2158 if (priv->cbs[i].port == port && priv->cbs[i].prio == prio) in sja1105_find_cbs_shaper() 2172 if (!priv->cbs[i].idle_slope && !priv->cbs[i].send_slope) in sja1105_find_unused_cbs_shaper() 2184 struct sja1105_cbs_entry *cbs = &priv->cbs[i]; in sja1105_delete_cbs_shaper() local 2186 if (cbs->port == port && cbs->prio == prio) { in sja1105_delete_cbs_shaper() 2187 memset(cbs, 0, sizeof(*cbs)); in sja1105_delete_cbs_shaper() 2216 cbs = &priv->cbs[index]; in sja1105_setup_tc_cbs() 2217 cbs->port = port; in sja1105_setup_tc_cbs() 2253 if (!priv->cbs) in sja1105_reload_cbs() 2257 struct sja1105_cbs_entry *cbs = &priv->cbs[i]; in sja1105_reload_cbs() local 2259 if (!cbs->idle_slope && !cbs->send_slope) in sja1105_reload_cbs() [all …]
|
| /drivers/gpu/drm/nouveau/dispnv04/ |
| A D | arb.c | 58 int found, mclk_extra, mclk_loop, cbs, m1, p1; in nv04_calc_arb() local 69 cbs = 128; in nv04_calc_arb() 92 m1 = clwm + cbs - 512; in nv04_calc_arb() 103 fifo->burst = cbs; in nv04_calc_arb()
|
| /drivers/net/ethernet/intel/ |
| A D | e100.c | 533 struct param_range cbs; member 555 struct cb *cbs; member 1046 nic->params.cbs = cbs; in e100_get_defaults() 1849 if (nic->cbs) { in e100_clean_cbs() 1863 nic->cbs = NULL; in e100_clean_cbs() 1868 nic->cbs; in e100_clean_cbs() 1882 if (!nic->cbs) in e100_alloc_cbs() 2561 struct param_range *cbs = &nic->params.cbs; in e100_get_ringparam() local 2576 struct param_range *cbs = &nic->params.cbs; in e100_set_ringparam() local 2585 cbs->count = max(ring->tx_pending, cbs->min); in e100_set_ringparam() [all …]
|
| /drivers/infiniband/hw/qedr/ |
| A D | qedr_roce_cm.c | 266 struct qed_ll2_cbs cbs; in qedr_ll2_start() local 270 cbs.rx_comp_cb = qedr_ll2_complete_rx_packet; in qedr_ll2_start() 271 cbs.tx_comp_cb = qedr_ll2_complete_tx_packet; in qedr_ll2_start() 272 cbs.rx_release_cb = qedr_ll2_release_rx_packet; in qedr_ll2_start() 273 cbs.tx_release_cb = qedr_ll2_complete_tx_packet; in qedr_ll2_start() 274 cbs.cookie = dev; in qedr_ll2_start() 289 data.cbs = &cbs; in qedr_ll2_start()
|
| /drivers/video/fbdev/nvidia/ |
| A D | nv_hw.c | 263 cbs = 128; in nv4CalcArbitration() 307 cbs * 1000 * 1000 / (8 * width) / in nv4CalcArbitration() 341 m1 = clwm + cbs - 512; in nv4CalcArbitration() 420 int found, mclk_extra, mclk_loop, cbs, m1; in nv10CalcArbitration() local 441 cbs = 512; in nv10CalcArbitration() 559 cbs = 512; in nv10CalcArbitration() 583 if (cbs <= 32) { in nv10CalcArbitration() 588 cbs = cbs / 2; in nv10CalcArbitration() 605 if (clwm < (1024 - cbs + 8)) in nv10CalcArbitration() 606 clwm = 1024 - cbs + 8; in nv10CalcArbitration() [all …]
|
| /drivers/net/ethernet/freescale/enetc/ |
| A D | enetc_qos.c | 237 struct tc_cbs_qopt_offload *cbs = type_data; in enetc_setup_tc_cbs() local 244 u8 tc = cbs->queue; in enetc_setup_tc_cbs() 256 if (!cbs->enable) { in enetc_setup_tc_cbs() 274 if (cbs->idleslope - cbs->sendslope != port_transmit_rate * 1000L || in enetc_setup_tc_cbs() 275 cbs->idleslope < 0 || cbs->sendslope > 0) in enetc_setup_tc_cbs() 280 bw = cbs->idleslope / (port_transmit_rate * 10UL); in enetc_setup_tc_cbs() 456 u32 cbs; member 884 fmi_config->cbs = cpu_to_le32(fmi->cbs); in enetc_flowmeter_hw_set() 1293 fmi->cbs = entryp->police.burst; in enetc_psfp_parse_clsflower()
|
| /drivers/video/fbdev/riva/ |
| A D | riva_hw.c | 660 int found, mclk_extra, mclk_loop, cbs, m1, p1; in nv4CalcArbitration() local 677 cbs = 128; in nv4CalcArbitration() 718 video_fill_us = cbs*1000*1000 / 16 / nvclk_freq ; in nv4CalcArbitration() 751 m1 = clwm + cbs - 512; in nv4CalcArbitration() 842 int found, mclk_extra, mclk_loop, cbs, m1; in nv10CalcArbitration() local 863 cbs = 512; in nv10CalcArbitration() 986 cbs = 512; in nv10CalcArbitration() 1001 m1 = clwm + cbs - 1024; /* Amount of overfill */ in nv10CalcArbitration() 1012 if(cbs <= 32) { in nv10CalcArbitration() 1015 cbs = cbs/2; /* reduce the burst size */ in nv10CalcArbitration() [all …]
|
| /drivers/net/ethernet/amd/xgbe/ |
| A D | xgbe-dcb.c | 23 ets->cbs = pdata->ets->cbs; in xgbe_dcb_ieee_getets()
|
| /drivers/dma/ppc4xx/ |
| A D | xor.h | 70 u32 cbs; /* status */ member
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_dcb_nl.c | 46 ets->cbs = dcbxcfg->etscfg.cbs; in i40e_dcbnl_ieee_getets() 123 pf->tmp_cfg.etscfg.cbs = ets->cbs; in i40e_dcbnl_ieee_setets()
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/tc/ |
| A D | meter.c | 82 mlx5e_flow_meter_cbs_calc(u64 cbs, u8 *man, u8 *exp) in mlx5e_flow_meter_cbs_calc() argument 89 m = cbs >> e; in mlx5e_flow_meter_cbs_calc() 93 _delta = cbs - _cbs; in mlx5e_flow_meter_cbs_calc()
|
| /drivers/net/ethernet/intel/fm10k/ |
| A D | fm10k_dcbnl.c | 17 ets->cbs = 0; in fm10k_dcbnl_ieee_getets()
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_dcb_nl.c | 42 ets->cbs = dcbxcfg->etscfg.cbs; in ice_dcbnl_getets() 83 new_cfg->etscfg.cbs = ets->cbs; in ice_dcbnl_setets()
|
| /drivers/net/ethernet/broadcom/genet/ |
| A D | bcmgenet.h | 544 struct enet_cb *cbs; /* tx ring buffer control block*/ member 568 struct enet_cb *cbs; /* Rx ring buffer control block */ member
|