| /drivers/infiniband/hw/cxgb4/ |
| A D | t4.h | 344 u16 in_use; member 374 u16 in_use; member 417 u16 in_use; member 432 srq->in_use++; in t4_srq_produce() 444 srq->in_use++; in t4_srq_produce_pending_wr() 452 srq->in_use--; in t4_srq_consume_pending_wr() 459 srq->in_use--; in t4_srq_produce_ooo() 474 srq->in_use--; in t4_srq_consume() 497 wq->rq.in_use++; in t4_rq_produce() 507 wq->rq.in_use--; in t4_rq_consume() [all …]
|
| A D | cq.c | 206 int in_use = wq->rq.in_use - count; in c4iw_flush_rq() local 209 wq, cq, wq->rq.in_use, count); in c4iw_flush_rq() 210 while (in_use--) { in c4iw_flush_rq() 475 srq->in_use, srq->size, in post_pending_srq_wrs() 502 srq->wq_pidx, srq->in_use, srq->size, in reap_srq_cqe() 508 srq->wq_pidx, srq->in_use, in reap_srq_cqe() 520 srq->in_use, srq->size, in reap_srq_cqe() 711 wq->sq.in_use -= wq->sq.size + idx - wq->sq.cidx; in poll_cq() 713 wq->sq.in_use -= idx - wq->sq.cidx; in poll_cq() 778 srq->wq.in_use < srq->srq_limit) in __c4iw_poll_cq_one()
|
| /drivers/md/dm-vdo/ |
| A D | action-manager.c | 29 bool in_use; member 140 return manager->current_action->in_use ? manager->current_action->context : NULL; in vdo_get_current_action_context() 257 manager->current_action->in_use = false; in finish_action_callback() 265 (manager->current_action->in_use || vdo_schedule_default_action(manager)); in finish_action_callback() 362 if (!manager->current_action->in_use) { in vdo_schedule_operation_with_context() 364 } else if (!manager->current_action->next->in_use) { in vdo_schedule_operation_with_context() 374 .in_use = true, in vdo_schedule_operation_with_context()
|
| /drivers/md/bcache/ |
| A D | writeback.h | 105 unsigned int in_use = dc->disk.c->gc_stats.in_use; in should_writeback() local 109 in_use > bch_cutoff_writeback_sync) in should_writeback() 125 in_use <= bch_cutoff_writeback); in should_writeback()
|
| A D | writeback.c | 23 c->gc_stats.in_use < BCH_AUTO_GC_DIRTY_THRESHOLD) in update_gc_after_writeback() 106 c->gc_stats.in_use > BCH_WRITEBACK_FRAGMENT_THRESHOLD_LOW && dirty > 0) { in __update_writeback_rate() 112 if (c->gc_stats.in_use <= BCH_WRITEBACK_FRAGMENT_THRESHOLD_MID) { in __update_writeback_rate() 114 (c->gc_stats.in_use - BCH_WRITEBACK_FRAGMENT_THRESHOLD_LOW); in __update_writeback_rate() 115 } else if (c->gc_stats.in_use <= BCH_WRITEBACK_FRAGMENT_THRESHOLD_HIGH) { in __update_writeback_rate() 117 (c->gc_stats.in_use - BCH_WRITEBACK_FRAGMENT_THRESHOLD_MID); in __update_writeback_rate() 120 (c->gc_stats.in_use - BCH_WRITEBACK_FRAGMENT_THRESHOLD_HIGH); in __update_writeback_rate()
|
| /drivers/net/ethernet/stmicro/stmmac/ |
| A D | stmmac_tc.c | 17 entry->in_use = true; in tc_fill_all_pass_entry() 48 first->in_use = true; in tc_find_entry() 166 frag->in_use = false; in tc_fill_entry() 167 entry->in_use = false; in tc_fill_entry() 180 entry->in_use = false; in tc_unfill_entry() 184 entry->in_use = false; in tc_unfill_entry() 603 if (!entry->in_use) in tc_add_flow() 627 entry->in_use = false; in tc_del_flow() 694 entry->in_use = true; in tc_add_vlan_flow() 714 entry->in_use = false; in tc_del_vlan_flow() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| A D | crypto.c | 65 unsigned long *in_use; member 307 bulk->in_use = bitmap_zalloc(num_deks, GFP_KERNEL); in mlx5_crypto_dek_bulk_create() 308 if (!bulk->in_use) { in mlx5_crypto_dek_bulk_create() 327 bitmap_free(bulk->in_use); in mlx5_crypto_dek_bulk_create() 353 bitmap_free(bulk->in_use); in mlx5_crypto_dek_bulk_free() 403 bitmap_set(bulk->in_use, pos, 1); in mlx5_crypto_dek_pool_pop() 435 old_val = test_and_clear_bit(obj_offset, bulk->in_use); in mlx5_crypto_dek_free_locked() 483 unsigned long *in_use = bulk->in_use; in mlx5_crypto_dek_bulk_reset_synced() local 490 i++, need_sync++, in_use++) { in mlx5_crypto_dek_bulk_reset_synced() 491 reused = hweight_long((*need_sync) ^ (*in_use)); in mlx5_crypto_dek_bulk_reset_synced() [all …]
|
| /drivers/gpu/drm/amd/display/dc/core/ |
| A D | dc_vm_helper.c | 65 int in_use = 0; in dc_get_vmid_use_vector() local 68 in_use |= dc->vm_helper->hubp_vmid_usage[i].vmid_usage[0] in dc_get_vmid_use_vector() 70 return in_use; in dc_get_vmid_use_vector()
|
| /drivers/gpu/drm/imx/dcss/ |
| A D | dcss-ss.c | 70 bool in_use; member 75 if (!ss->in_use) in dcss_ss_write() 167 ss->in_use = true; in dcss_ss_enable() 173 ss->in_use = false; in dcss_ss_shutoff()
|
| A D | dcss-ctxld.c | 82 bool in_use; member 97 !(irq_status & CTXLD_ENABLE) && ctxld->in_use) { in dcss_ctxld_irq_handler() 98 ctxld->in_use = false; in dcss_ctxld_irq_handler() 296 ctxld->in_use = true; in dcss_ctxld_enable_locked() 325 if (ctxld->armed && !ctxld->in_use) { in dcss_ctxld_kick() 388 while (!time_after(jiffies, timeout) && ctxld->in_use) in dcss_ctxld_suspend()
|
| A D | dcss-dtg.c | 85 bool in_use; member 100 if (!dtg->in_use) in dcss_dtg_write() 304 dtg->in_use = true; in dcss_dtg_enable() 314 dtg->in_use = false; in dcss_dtg_shutoff() 319 return dtg->in_use; in dcss_dtg_is_enabled()
|
| /drivers/media/v4l2-core/ |
| A D | v4l2-event.c | 52 kev->sev->in_use--; in __v4l2_event_dequeue() 118 if (sev->in_use == sev->elems) { in __v4l2_event_queue_fh() 122 sev->in_use--; in __v4l2_event_queue_fh() 138 kev = sev->events + sev_pos(sev, sev->in_use); in __v4l2_event_queue_fh() 145 sev->in_use++; in __v4l2_event_queue_fh() 216 for (i = 0; i < sev->in_use; i++) { in __v4l2_event_unsubscribe()
|
| /drivers/ptp/ |
| A D | ptp_private.h | 106 bool in_use = false; in ptp_vclock_in_use() local 122 in_use = true; in ptp_vclock_in_use() 126 return in_use; in ptp_vclock_in_use()
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_flex_type.h | 168 u8 in_use; member 196 u8 in_use; member 210 u8 in_use; member 298 u8 in_use; /* non-zero if used */ member
|
| A D | ice_ptp.c | 401 clear_bit(idx, tx->in_use); in ice_ptp_req_tx_single_tstamp() 484 clear_bit(idx, tx->in_use); in ice_ptp_complete_tx_single_tstamp() 638 clear_bit(idx, tx->in_use); in ice_ptp_process_tx_tstamp() 742 unsigned long *in_use, *stale; in ice_ptp_alloc_tx_tracker() local 749 if (!tstamps || !in_use || !stale) { in ice_ptp_alloc_tx_tracker() 751 bitmap_free(in_use); in ice_ptp_alloc_tx_tracker() 758 tx->in_use = in_use; in ice_ptp_alloc_tx_tracker() 806 clear_bit(idx, tx->in_use); in ice_ptp_flush_tx_tracker() 879 bitmap_free(tx->in_use); in ice_ptp_release_tx_tracker() 880 tx->in_use = NULL; in ice_ptp_release_tx_tracker() [all …]
|
| A D | ice_flex_pipe.c | 705 if (!hw->blk[blk].xlt1.ptg_tbl[ptg].in_use) in ice_ptg_remove_ptype() 935 if (!hw->blk[blk].xlt2.vsig_tbl[i].in_use) in ice_vsig_alloc() 964 if (xlt2->vsig_tbl[i].in_use && in ice_find_dup_props_vsig() 1525 if (hw->blk[blk].masks.masks[i].in_use) { in ice_alloc_prof_mask() 3227 if (prof->tcam[i].in_use) { in ice_rem_prof_id() 3228 prof->tcam[i].in_use = false; in ice_rem_prof_id() 3635 tcam->in_use = 0; in ice_prof_tcam_ena_dis() 3660 tcam->in_use = 1; in ice_prof_tcam_ena_dis() 3716 t->tcam[i].in_use) { in ice_adj_prof_priorities() 3728 !t->tcam[i].in_use) { in ice_adj_prof_priorities() [all …]
|
| /drivers/media/platform/mediatek/vcodec/decoder/vdec/ |
| A D | vdec_h264_req_common.c | 245 DECLARE_BITMAP(in_use, ARRAY_SIZE(dec_param->dpb)) = { 0, }; in mtk_vdec_h264_update_dpb() 252 set_bit(i, in_use); in mtk_vdec_h264_update_dpb() 267 for_each_set_bit(j, in_use, ARRAY_SIZE(dec_param->dpb)) { in mtk_vdec_h264_update_dpb() 277 clear_bit(j, in_use); in mtk_vdec_h264_update_dpb()
|
| /drivers/gpu/drm/vc4/ |
| A D | vc4_kms.c | 414 if (!old_hvs_state->fifo_state[channel].in_use) in vc4_atomic_commit_tail() 521 if (!hvs_state->fifo_state[channel].in_use) in vc4_atomic_commit_setup() 761 state->fifo_state[i].in_use = old_state->fifo_state[i].in_use; in vc4_hvs_channels_duplicate_state() 797 drm_printf(p, "\t\tin use=%d\n", hvs_state->fifo_state[i].in_use); in vc4_hvs_channels_print_state() 889 if (!hvs_new_state->fifo_state[i].in_use) in vc4_pv_muxing_atomic_check() 968 hvs_new_state->fifo_state[channel].in_use = false; in vc4_pv_muxing_atomic_check() 984 hvs_new_state->fifo_state[channel].in_use = true; in vc4_pv_muxing_atomic_check() 1045 if (!hvs_new_state->fifo_state[i].in_use) in vc4_core_clock_atomic_check()
|
| /drivers/gpu/ipu-v3/ |
| A D | ipu-pre.c | 102 bool in_use; member 146 if (pre->cur.in_use) in ipu_pre_get() 159 pre->cur.in_use = true; in ipu_pre_get() 167 pre->cur.in_use = false; in ipu_pre_put()
|
| A D | ipu-dp.c | 47 bool in_use; member 329 if (dp->in_use) in ipu_dp_get() 332 dp->in_use = true; in ipu_dp_get() 340 dp->in_use = false; in ipu_dp_put()
|
| A D | ipu-dc.c | 95 bool in_use; member 320 if (dc->in_use) { in ipu_dc_get() 325 dc->in_use = true; in ipu_dc_get() 338 dc->in_use = false; in ipu_dc_put()
|
| /drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
| A D | chcr_common.h | 83 return q->size - 1 - q->in_use; in chcr_txq_avail() 88 q->in_use += n; in chcr_txq_advance()
|
| /drivers/hid/ |
| A D | hid-prodikeys.c | 36 unsigned long in_use; member 233 pms->in_use = 0; in pcmidi_sustained_note_release() 243 pms->in_use = 0; in init_sustain_timers() 256 pms->in_use = 1; in stop_sustain_timers() 351 if (!pms->in_use) { in pcmidi_handle_report3() 355 pms->in_use = 1; in pcmidi_handle_report3()
|
| /drivers/gpu/drm/vc4/tests/ |
| A D | vc4_test_pv_muxing.c | 36 if (!hvs_state->fifo_state[i].in_use) in check_fifo_conflict() 161 KUNIT_EXPECT_TRUE(test, new_hvs_state->fifo_state[channel].in_use); in check_channel_for_encoder() 875 KUNIT_ASSERT_TRUE(test, new_hvs_state->fifo_state[hdmi0_channel].in_use); in drm_test_vc5_pv_muxing_bugs_subsequent_crtc_enable() 911 KUNIT_ASSERT_TRUE(test, new_hvs_state->fifo_state[hdmi1_channel].in_use); in drm_test_vc5_pv_muxing_bugs_subsequent_crtc_enable() 984 KUNIT_ASSERT_TRUE(test, new_hvs_state->fifo_state[old_hdmi0_channel].in_use); in drm_test_vc5_pv_muxing_bugs_stable_fifo() 992 KUNIT_ASSERT_TRUE(test, new_hvs_state->fifo_state[old_hdmi1_channel].in_use); in drm_test_vc5_pv_muxing_bugs_stable_fifo() 1030 KUNIT_ASSERT_TRUE(test, new_hvs_state->fifo_state[hdmi1_channel].in_use); in drm_test_vc5_pv_muxing_bugs_stable_fifo()
|
| /drivers/gpu/drm/i915/display/ |
| A D | skl_scaler.c | 190 scaler_state->scalers[*scaler_id].in_use = false; in skl_update_scaler() 313 if (scaler_state->scalers[i].in_use) in intel_allocate_scaler() 316 scaler_state->scalers[i].in_use = true; in intel_allocate_scaler() 413 scaler_state->scalers[*scaler_id].in_use = false; in intel_atomic_setup_scaler() 415 scaler_state->scalers[0].in_use = true; in intel_atomic_setup_scaler() 874 if (!scaler_state->scalers[i].in_use) in skl_detach_scalers() 916 scaler_state->scalers[i].in_use = true; in skl_scaler_get_config()
|