| /linux/drivers/gpu/drm/i915/gt/ |
| A D | intel_timeline.c | 175 return tl; in intel_timeline_create_from_engine() 182 return tl; in intel_timeline_create_from_engine() 212 tl->fence_context, tl->hwsp_offset); in intel_timeline_pin() 302 GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb); in timeline_advance() 304 return tl->seqno += 1 + tl->has_initial_breadcrumb; in timeline_advance() 318 tl->hwsp_seqno = tl->hwsp_map + next_ofs; in __intel_timeline_get_seqno() 350 tl = NULL; in intel_timeline_read_hwsp() 352 if (tl) { in intel_timeline_read_hwsp() 361 tl = NULL; in intel_timeline_read_hwsp() 365 if (!tl) in intel_timeline_read_hwsp() [all …]
|
| A D | selftest_timeline.c | 81 tl = xchg(&state->history[idx], tl); in __mock_hwsp_record() 82 if (tl) { in __mock_hwsp_record() 101 if (IS_ERR(tl)) in __mock_hwsp_timeline() 496 if (READ_ONCE(*tl->hwsp_seqno) != tl->seqno) { in checked_tl_write() 498 *tl->hwsp_seqno, tl->seqno); in checked_tl_write() 586 n, tl->fence_context, tl->hwsp_offset, *tl->hwsp_seqno); in live_hwsp_engine() 658 n, tl->fence_context, tl->hwsp_offset, *tl->hwsp_seqno); in live_hwsp_alternate() 684 if (IS_ERR(tl)) in live_hwsp_wrap() 1205 WRITE_ONCE(*(u32 *)tl->hwsp_seqno, tl->seqno); in live_hwsp_rollover_kernel() 1292 WRITE_ONCE(*(u32 *)tl->hwsp_seqno, tl->seqno); in live_hwsp_rollover_user() [all …]
|
| A D | intel_gt_requests.c | 78 retire_requests(tl); in engine_retire() 79 mutex_unlock(&tl->mutex); in engine_retire() 81 intel_timeline_put(tl); in engine_retire() 84 tl = ptr_mask_bits(next, 1); in engine_retire() 85 } while (tl); in engine_retire() 103 intel_timeline_get(tl); in add_retire() 118 if (add_retire(engine, tl)) in intel_engine_add_retire() 149 intel_timeline_get(tl); in intel_gt_retire_requests_timeout() 174 if (!retire_requests(tl)) in intel_gt_retire_requests_timeout() 176 mutex_unlock(&tl->mutex); in intel_gt_retire_requests_timeout() [all …]
|
| A D | intel_timeline.h | 48 return i915_syncmap_set(&tl->sync, context, seqno); in __intel_timeline_sync_set() 51 static inline int intel_timeline_sync_set(struct intel_timeline *tl, in intel_timeline_sync_set() argument 60 return i915_syncmap_is_later(&tl->sync, context, seqno); in __intel_timeline_sync_is_later() 69 void __intel_timeline_pin(struct intel_timeline *tl); 71 void intel_timeline_enter(struct intel_timeline *tl); 72 int intel_timeline_get_seqno(struct intel_timeline *tl, 75 void intel_timeline_exit(struct intel_timeline *tl); 76 void intel_timeline_unpin(struct intel_timeline *tl); 78 void intel_timeline_reset_seqno(const struct intel_timeline *tl); 95 intel_timeline_is_last(const struct intel_timeline *tl, in intel_timeline_is_last() argument [all …]
|
| A D | selftest_context.c | 17 struct intel_timeline *tl = i915_request_timeline(rq); in request_sync() local 21 intel_timeline_get(tl); in request_sync() 35 lockdep_unpin_lock(&tl->mutex, rq->cookie); in request_sync() 36 mutex_unlock(&tl->mutex); in request_sync() 39 intel_timeline_put(tl); in request_sync() 46 struct intel_timeline *tl = ce->timeline; in context_sync() local 49 mutex_lock(&tl->mutex); in context_sync() 54 if (list_empty(&tl->requests)) in context_sync() 57 rq = list_last_entry(&tl->requests, typeof(*rq), link); in context_sync() 68 mutex_unlock(&tl->mutex); in context_sync()
|
| A D | intel_context.h | 251 struct intel_timeline *tl = ce->timeline; in intel_context_timeline_lock() local 255 err = mutex_lock_interruptible_nested(&tl->mutex, 0); in intel_context_timeline_lock() 257 err = mutex_lock_interruptible_nested(&tl->mutex, in intel_context_timeline_lock() 260 err = mutex_lock_interruptible(&tl->mutex); in intel_context_timeline_lock() 264 return tl; in intel_context_timeline_lock() 267 static inline void intel_context_timeline_unlock(struct intel_timeline *tl) in intel_context_timeline_unlock() argument 268 __releases(&tl->mutex) in intel_context_timeline_unlock() 270 mutex_unlock(&tl->mutex); in intel_context_timeline_unlock()
|
| A D | mock_engine.c | 16 static int mock_timeline_pin(struct intel_timeline *tl) in mock_timeline_pin() argument 20 if (WARN_ON(!i915_gem_object_trylock(tl->hwsp_ggtt->obj, NULL))) in mock_timeline_pin() 23 err = intel_timeline_pin_map(tl); in mock_timeline_pin() 24 i915_gem_object_unlock(tl->hwsp_ggtt->obj); in mock_timeline_pin() 28 atomic_inc(&tl->pin_count); in mock_timeline_pin() 32 static void mock_timeline_unpin(struct intel_timeline *tl) in mock_timeline_unpin() argument 34 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in mock_timeline_unpin() 35 atomic_dec(&tl->pin_count); in mock_timeline_unpin()
|
| /linux/fs/smb/client/ |
| A D | dfs_cache.h | 55 dfs_cache_get_next_tgt(struct dfs_cache_tgt_list *tl, in dfs_cache_get_next_tgt() argument 58 if (!tl || !tl->tl_numtgts || list_empty(&tl->tl_list) || in dfs_cache_get_next_tgt() 59 !it || list_is_last(&it->it_list, &tl->tl_list)) in dfs_cache_get_next_tgt() 65 dfs_cache_get_tgt_iterator(struct dfs_cache_tgt_list *tl) in dfs_cache_get_tgt_iterator() argument 67 if (!tl) in dfs_cache_get_tgt_iterator() 69 return list_first_entry_or_null(&tl->tl_list, in dfs_cache_get_tgt_iterator() 78 if (!tl || !tl->tl_numtgts || list_empty(&tl->tl_list)) in dfs_cache_free_tgts() 80 list_for_each_entry_safe(it, nit, &tl->tl_list, it_list) { in dfs_cache_free_tgts() 85 tl->tl_numtgts = 0; in dfs_cache_free_tgts() 95 dfs_cache_get_nr_tgts(const struct dfs_cache_tgt_list *tl) in dfs_cache_get_nr_tgts() argument [all …]
|
| A D | dfs.h | 23 struct dfs_cache_tgt_list tl; member 41 #define ref_walk_tl(w) (&ref_walk_cur(w)->tl) 64 dfs_cache_free_tgts(&ref->tl); in __ref_walk_free() 100 tit = dfs_cache_get_tgt_iterator(&ref->tl); in ref_walk_next_tgt() 102 tit = dfs_cache_get_next_tgt(&ref->tl, ref->tit); in ref_walk_next_tgt() 149 struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) in dfs_get_referral() argument 156 cifs_remap(cifs_sb), path, ref, tl); in dfs_get_referral()
|
| A D | dfs.c | 441 struct dfs_cache_tgt_list *tl) in __tree_connect_dfs_target() argument 451 tit = dfs_cache_get_tgt_iterator(tl); in __tree_connect_dfs_target() 458 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) { in __tree_connect_dfs_target() 506 dfs_cache_free_tgts(tl); in __tree_connect_dfs_target() 510 list_replace_init(&ntl.tl_list, &tl->tl_list); in __tree_connect_dfs_target() 526 struct dfs_cache_tgt_list *tl) in tree_connect_dfs_target() argument 534 rc = __tree_connect_dfs_target(xid, tcon, cifs_sb, tree, islink, tl); in tree_connect_dfs_target() 545 dfs_cache_free_tgts(tl); in tree_connect_dfs_target() 554 DFS_CACHE_TGT_LIST(tl); in cifs_tree_connect() 604 dfs_cache_noreq_find(server->leaf_fullpath + 1, &ref, &tl)) { in cifs_tree_connect() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| A D | nfp_net_debugdump.c | 63 struct nfp_dump_tl_hdr tl; member 69 struct nfp_dump_tl_hdr tl; member 75 struct nfp_dump_tl_hdr tl; member 83 struct nfp_dump_tl_hdr tl; member 92 struct nfp_dump_tl_hdr tl; member 125 struct nfp_dump_tl *tl; in nfp_traverse_tlvs() local 131 tl = p; in nfp_traverse_tlvs() 132 if (!tl->type && !tl->length) in nfp_traverse_tlvs() 135 if (be32_to_cpu(tl->length) > remaining - sizeof(*tl)) in nfp_traverse_tlvs() 138 total_tlv_size = sizeof(*tl) + be32_to_cpu(tl->length); in nfp_traverse_tlvs() [all …]
|
| /linux/crypto/ |
| A D | vmac.c | 151 int i; u64 th, tl; \ 156 ADD128(rh, rl, th, tl); \ 162 int i; u64 th, tl; \ 167 ADD128(rh, rl, th, tl); \ 177 int i; u64 th, tl; \ 182 ADD128(rh, rl, th, tl); \ 185 ADD128(rh, rl, th, tl); \ 188 ADD128(rh, rl, th, tl); \ 191 ADD128(rh, rl, th, tl); \ 197 int i; u64 th, tl; \ [all …]
|
| A D | camellia_generic.c | 359 u32 dw, tl, tr; in camellia_setup_tail() local 458 tl = subL[10] ^ (subR[10] & ~subR[8]); in camellia_setup_tail() 459 dw = tl & subL[8]; /* FL(kl1) */ in camellia_setup_tail() 467 tl = subL[7] ^ (subR[7] & ~subR[9]); in camellia_setup_tail() 468 dw = tl & subL[9]; /* FLinv(kl2) */ in camellia_setup_tail() 480 tl = subL[18] ^ (subR[18] & ~subR[16]); in camellia_setup_tail() 481 dw = tl & subL[16]; /* FL(kl3) */ in camellia_setup_tail() 489 tl = subL[15] ^ (subR[15] & ~subR[17]); in camellia_setup_tail() 490 dw = tl & subL[17]; /* FLinv(kl4) */ in camellia_setup_tail() 509 dw = tl & subL[24]; /* FL(kl5) */ in camellia_setup_tail() [all …]
|
| /linux/drivers/isdn/mISDN/ |
| A D | fsm.c | 98 struct FsmTimer *ft = from_timer(ft, t, tl); in FsmExpireTimer() 114 timer_setup(&ft->tl, FsmExpireTimer, 0); in mISDN_FsmInitTimer() 126 del_timer(&ft->tl); in mISDN_FsmDelTimer() 141 if (timer_pending(&ft->tl)) { in mISDN_FsmAddTimer() 152 ft->tl.expires = jiffies + (millisec * HZ) / 1000; in mISDN_FsmAddTimer() 153 add_timer(&ft->tl); in mISDN_FsmAddTimer() 169 if (timer_pending(&ft->tl)) in mISDN_FsmRestartTimer() 170 del_timer(&ft->tl); in mISDN_FsmRestartTimer() 173 ft->tl.expires = jiffies + (millisec * HZ) / 1000; in mISDN_FsmRestartTimer() 174 add_timer(&ft->tl); in mISDN_FsmRestartTimer()
|
| A D | timerdev.c | 39 struct timer_list tl; member 77 timer_shutdown_sync(&timer->tl); in mISDN_close() 158 struct mISDNtimer *timer = from_timer(timer, t, tl); in dev_expire_timer() 183 timer_setup(&timer->tl, dev_expire_timer, 0); in misdn_add_timer() 189 timer->tl.expires = jiffies + ((HZ * (u_long)timeout) / 1000); in misdn_add_timer() 190 add_timer(&timer->tl); in misdn_add_timer() 207 timer_shutdown_sync(&timer->tl); in misdn_del_timer()
|
| A D | dsp_tones.c | 462 struct dsp *dsp = from_timer(dsp, t, tone.tl); in dsp_tone_timeout() 481 tone->tl.expires = jiffies + (pat->seq[index] * HZ) / 8000; in dsp_tone_timeout() 482 add_timer(&tone->tl); in dsp_tone_timeout() 507 if (dsp->features.hfc_loops && timer_pending(&tonet->tl)) in dsp_tone() 508 del_timer(&tonet->tl); in dsp_tone() 541 if (timer_pending(&tonet->tl)) in dsp_tone() 542 del_timer(&tonet->tl); in dsp_tone() 543 tonet->tl.expires = jiffies + (pat->seq[0] * HZ) / 8000; in dsp_tone() 544 add_timer(&tonet->tl); in dsp_tone()
|
| /linux/drivers/s390/net/ |
| A D | fsm.c | 135 fsm_timer *this = from_timer(this, t, tl); in fsm_expire_timer() 151 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_settimer() 161 del_timer(&this->tl); in fsm_deltimer() 173 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_addtimer() 176 this->tl.expires = jiffies + (millisec * HZ) / 1000; in fsm_addtimer() 177 add_timer(&this->tl); in fsm_addtimer() 191 del_timer(&this->tl); in fsm_modtimer() 192 timer_setup(&this->tl, fsm_expire_timer, 0); in fsm_modtimer() 195 this->tl.expires = jiffies + (millisec * HZ) / 1000; in fsm_modtimer() 196 add_timer(&this->tl); in fsm_modtimer()
|
| /linux/fs/ext4/ |
| A D | fast_commit.c | 707 struct ext4_fc_tl tl; in ext4_fc_reserve_space() local 772 struct ext4_fc_tl tl; in ext4_fc_write_tail() local 815 struct ext4_fc_tl tl; in ext4_fc_add_tlv() local 836 struct ext4_fc_tl tl; in ext4_fc_add_dentry_tlv() local 868 struct ext4_fc_tl tl; in ext4_fc_write_inode() local 2068 struct ext4_fc_tl_mem tl; in ext4_fc_replay_scan() local 2103 !ext4_fc_value_len_isvalid(sbi, tl.fc_tag, tl.fc_len)) { in ext4_fc_replay_scan() 2110 switch (tl.fc_tag) { in ext4_fc_replay_scan() 2231 switch (tl.fc_tag) { in ext4_fc_replay() 2252 tl.fc_len, 0); in ext4_fc_replay() [all …]
|
| /linux/arch/x86/crypto/ |
| A D | camellia_glue.c | 800 u32 dw, tl, tr; in camellia_setup_tail() local 912 dw = tl & (subRL[8] >> 32); /* FL(kl1) */ in camellia_setup_tail() 914 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail() 920 tl = (subRL[7] >> 32) ^ (subRL[7] & ~subRL[9]); in camellia_setup_tail() 923 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail() 932 dw = tl & (subRL[16] >> 32); /* FL(kl3) */ in camellia_setup_tail() 934 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail() 943 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail() 956 dw = tl & (subRL[24] >> 32); /* FL(kl5) */ in camellia_setup_tail() 958 tt = (tr | ((u64)tl << 32)); in camellia_setup_tail() [all …]
|
| /linux/drivers/gpu/drm/i915/ |
| A D | i915_request.c | 849 if (list_empty(&tl->requests)) in request_alloc_slow() 866 retire_requests(tl); in request_alloc_slow() 956 tl->fence_context, seqno); in __i915_request_create() 958 RCU_INIT_POINTER(rq->timeline, tl); in __i915_request_create() 959 rq->hwsp_seqno = tl->hwsp_seqno; in __i915_request_create() 1031 struct intel_timeline *tl; in i915_request_create() local 1034 if (IS_ERR(tl)) in i915_request_create() 1035 return ERR_CAST(tl); in i915_request_create() 1054 intel_context_timeline_unlock(tl); in i915_request_create() 1850 lockdep_assert_held(&tl->mutex); in i915_request_add() [all …]
|
| /linux/arch/s390/include/asm/ |
| A D | dat-bits.h | 24 unsigned long tl : 2; /* Region- or Segment-Table Length */ member 46 unsigned long tl : 2; /* Region-Second-Table Length */ member 61 unsigned long tl : 2; /* Region-Third-Table Length */ member 75 unsigned long tl : 2; /* Segment-Table Length */ member
|
| /linux/lib/ |
| A D | inflate.c | 618 if ((e = (t = tl + ((unsigned)b & ml))->e) > 16) in inflate_codes() 804 huft_free(tl); in inflate_fixed() 813 if (inflate_codes(tl, td, bl, bd)) { in inflate_fixed() 820 huft_free(tl); in inflate_fixed() 904 huft_free(tl); in inflate_dynamic() 918 j = (td = tl + ((unsigned)b & m))->b; in inflate_dynamic() 966 huft_free(tl); in inflate_dynamic() 983 huft_free(tl); in inflate_dynamic() 1001 huft_free(tl); in inflate_dynamic() 1010 if (inflate_codes(tl, td, bl, bd)) { [all …]
|
| /linux/kernel/sched/ |
| A D | topology.c | 1606 if (tl->sd_flags) in sd_init() 1639 .name = tl->name, in sd_init() 1723 for (tl = sched_domain_topology; tl->mask; tl++) 1730 sched_domain_topology = tl; in set_sched_topology() 1963 if (!tl) in sched_init_numa() 1995 sched_domain_topology = tl; in sched_init_numa() 2227 for_each_sd_topology(tl) { in __sdt_alloc() 2296 for_each_sd_topology(tl) { in __sdt_free() 2383 if (!cpumask_equal(tl->mask(cpu), tl->mask(i)) && in topology_span_sane() 2384 cpumask_intersects(tl->mask(cpu), tl->mask(i))) in topology_span_sane() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| A D | qed_vf.h | 55 struct channel_tlv tl; member 62 struct channel_tlv tl; member 74 struct channel_tlv tl; member 123 struct channel_tlv tl; member 232 struct channel_tlv tl; member 345 struct channel_tlv tl; member 353 struct channel_tlv tl; member 359 struct channel_tlv tl; member 365 struct channel_tlv tl; member 377 struct channel_tlv tl; member [all …]
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| A D | 4965-rs.c | 240 while (tl->queue_count && tl->time_stamp < oldest_time) { in il4965_rs_tl_rm_old_stats() 241 tl->total -= tl->packet_count[tl->head]; in il4965_rs_tl_rm_old_stats() 242 tl->packet_count[tl->head] = 0; in il4965_rs_tl_rm_old_stats() 244 tl->queue_count--; in il4965_rs_tl_rm_old_stats() 245 tl->head++; in il4965_rs_tl_rm_old_stats() 247 tl->head = 0; in il4965_rs_tl_rm_old_stats() 279 tl->total = 1; in il4965_rs_tl_add_packet() 282 tl->head = 0; in il4965_rs_tl_add_packet() 296 tl->packet_count[idx] = tl->packet_count[idx] + 1; in il4965_rs_tl_add_packet() 297 tl->total = tl->total + 1; in il4965_rs_tl_add_packet() [all …]
|