| /linux/drivers/net/wireless/ath/ath5k/ |
| A D | qcu.c | 286 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_set_tx_retry_limits() local 326 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() local 333 (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) in ath5k_hw_reset_tx_queue() 366 if (tq->tqi_cbr_period) { in ath5k_hw_reset_tx_queue() 369 AR5K_REG_SM(tq->tqi_cbr_overflow_limit, in ath5k_hw_reset_tx_queue() 376 if (tq->tqi_cbr_overflow_limit) in ath5k_hw_reset_tx_queue() 382 if (tq->tqi_ready_time && (tq->tqi_type != AR5K_TX_QUEUE_CAB)) in ath5k_hw_reset_tx_queue() 388 if (tq->tqi_burst_time) { in ath5k_hw_reset_tx_queue() 412 switch (tq->tqi_type) { in ath5k_hw_reset_tx_queue() 434 ath5k_hw_reg_write(ah, ((tq->tqi_ready_time - in ath5k_hw_reset_tx_queue() [all …]
|
| /linux/drivers/net/vmxnet3/ |
| A D | vmxnet3_drv.c | 408 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup() 439 tq->tx_ring.base, tq->tx_ring.basePA); in vmxnet3_tq_destroy() 444 tq->data_ring.size * tq->txdata_desc_size, in vmxnet3_tq_destroy() 477 memset(tq->tx_ring.base, 0, tq->tx_ring.size * in vmxnet3_tq_init() 492 memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size); in vmxnet3_tq_init() 505 tq->comp_ring.base || tq->buf_info); in vmxnet3_tq_create() 516 tq->data_ring.size * tq->txdata_desc_size, in vmxnet3_tq_create() 531 tq->buf_info = kcalloc_node(tq->tx_ring.size, sizeof(tq->buf_info[0]), in vmxnet3_tq_create() 694 tbi = tq->buf_info + tq->tx_ring.next2fill; in vmxnet3_map_pkt() 722 tbi = tq->buf_info + tq->tx_ring.next2fill; in vmxnet3_map_pkt() [all …]
|
| A D | vmxnet3_ethtool.c | 468 buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); in vmxnet3_get_regs() 470 buf[j++] = tq->tx_ring.size; in vmxnet3_get_regs() 471 buf[j++] = tq->tx_ring.next2fill; in vmxnet3_get_regs() 472 buf[j++] = tq->tx_ring.next2comp; in vmxnet3_get_regs() 473 buf[j++] = tq->tx_ring.gen; in vmxnet3_get_regs() 477 buf[j++] = tq->data_ring.size; in vmxnet3_get_regs() 478 buf[j++] = tq->txdata_desc_size; in vmxnet3_get_regs() 482 buf[j++] = tq->comp_ring.size; in vmxnet3_get_regs() 483 buf[j++] = tq->comp_ring.next2proc; in vmxnet3_get_regs() 484 buf[j++] = tq->comp_ring.gen; in vmxnet3_get_regs() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| A D | sge.c | 235 return tq->size - 1 - tq->in_use; in txq_avail() 405 tq->cidx = cidx; in free_tx_desc() 1144 tq->in_use += n; in txq_advance() 1145 tq->pidx += n; in txq_advance() 1146 if (tq->pidx >= tq->size) in txq_advance() 1147 tq->pidx -= tq->size; in txq_advance() 1432 end = ((void *)tq->desc + ((void *)end - (void *)tq->stat)); in t4vf_eth_xmit() 2519 tq->size * sizeof(*tq->desc) + s->stat_len, in free_txq() 2520 tq->desc, tq->phys_addr); in free_txq() 2521 tq->cntxt_id = 0; in free_txq() [all …]
|
| /linux/drivers/media/v4l2-core/ |
| A D | v4l2-jpeg.c | 212 int c, h_v, tq; in jpeg_parse_frame_header() local 233 tq = jpeg_get_byte(stream); in jpeg_parse_frame_header() 234 if (tq < 0) in jpeg_parse_frame_header() 235 return tq; in jpeg_parse_frame_header() 242 component->quantization_table_selector = tq; in jpeg_parse_frame_header() 318 u8 pq, tq, *qk; in jpeg_parse_quantization_tables() local 336 tq = pq_tq & 0xf; in jpeg_parse_quantization_tables() 337 if (tq > 3) in jpeg_parse_quantization_tables() 347 tables[tq].start = qk; in jpeg_parse_quantization_tables() 348 tables[tq].length = pq ? 128 : 64; in jpeg_parse_quantization_tables()
|
| /linux/net/tipc/ |
| A D | trace.h | 284 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 286 TP_ARGS(r, f, t, tq), 301 __entry->len = skb_queue_len(tq); 303 msg_seqno(buf_msg(skb_peek(tq))) : 0; 305 msg_seqno(buf_msg(skb_peek_tail(tq))) : 0; 314 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 315 TP_ARGS(r, f, t, tq), 320 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 321 TP_ARGS(r, f, t, tq),
|
| /linux/drivers/net/ |
| A D | ifb.c | 52 struct sk_buff_head tq; member 70 skb = skb_peek(&txp->tq); in ifb_ri_tasklet() 74 skb_queue_splice_tail_init(&txp->rq, &txp->tq); in ifb_ri_tasklet() 78 while ((skb = __skb_dequeue(&txp->tq)) != NULL) { in ifb_ri_tasklet() 97 if (skb_queue_len(&txp->tq) != 0) in ifb_ri_tasklet() 175 __skb_queue_head_init(&txp->tq); in ifb_dev_init() 207 __skb_queue_purge(&txp->tq); in ifb_dev_free()
|
| /linux/arch/arm/boot/dts/ |
| A D | imx7s-mba7.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 17 compatible = "tq,imx7s-mba7", "tq,imx7s-tqma7", "fsl,imx7s";
|
| A D | imx7d-mba7.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 17 compatible = "tq,imx7d-mba7", "tq,imx7d-tqma7", "fsl,imx7d";
|
| A D | imx53-mba53.dts | 12 compatible = "tq,mba53", "tq,tqma53", "fsl,imx53"; 65 compatible = "tq,imx53-mba53-sgtl5000",
|
| A D | imx6dl-tqma6a.dtsi | 4 * Copyright 2013-2017 Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx6dl-tqma6b.dtsi | 4 * Copyright 2013-2017 Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx6q-tqma6a.dtsi | 4 * Copyright 2013-2017 Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx7s-tqma7.dtsi | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx6qdl-tqma6a.dtsi | 4 * Copyright 2013-2017 Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx6qdl-tqma6b.dtsi | 4 * Copyright 2013-2017 Markus Niebel <Markus.Niebel@tq-group.com>
|
| A D | imx7d-tqma7.dtsi | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com>
|
| /linux/drivers/input/serio/ |
| A D | hp_sdc.c | 190 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_take() 412 curr = hp_sdc.tq[curridx]; in hp_sdc_put() 416 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 430 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 573 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 609 if (hp_sdc.tq[i] == this) in __hp_sdc_enqueue_transaction() 617 if (hp_sdc.tq[i] == NULL) { in __hp_sdc_enqueue_transaction() 618 hp_sdc.tq[i] = this; in __hp_sdc_enqueue_transaction() 652 if (hp_sdc.tq[i] == this) in hp_sdc_dequeue_transaction() 653 hp_sdc.tq[i] = NULL; in hp_sdc_dequeue_transaction() [all …]
|
| /linux/drivers/net/can/dev/ |
| A D | bittiming.c | 152 bt->tq = (u32)v64; in can_calc_bittiming() 225 brp64 = (u64)priv->clock.freq * (u64)bt->tq; in can_fixup_bittiming() 276 if (!bt->tq && bt->bitrate && btc) in can_get_bittiming() 278 else if (bt->tq && !bt->bitrate && btc) in can_get_bittiming() 280 else if (!bt->tq && bt->bitrate && bitrate_const) in can_get_bittiming()
|
| /linux/drivers/input/keyboard/ |
| A D | sunkbd.c | 63 struct work_struct tq; member 103 schedule_work(&sunkbd->tq); in sunkbd_interrupt() 230 struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq); in sunkbd_reinit() 253 cancel_work_sync(&sunkbd->tq); in sunkbd_enable() 277 INIT_WORK(&sunkbd->tq, sunkbd_reinit); in sunkbd_connect()
|
| A D | lkkbd.c | 270 struct work_struct tq; member 457 schedule_work(&lk->tq); in lkkbd_interrupt() 568 struct lkkbd *lk = container_of(work, struct lkkbd, tq); in lkkbd_reinit() 623 INIT_WORK(&lk->tq, lkkbd_reinit); in lkkbd_connect()
|
| /linux/drivers/gpu/drm/i915/gem/selftests/ |
| A D | i915_gem_context.c | 650 struct i915_request *tq[5] = {}; in igt_ctx_exec() local 708 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 743 throttle_release(tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 760 struct i915_request *tq[5] = {}; in igt_shared_ctx_exec() local 842 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 878 throttle_release(tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 1345 struct i915_request *tq[5] = {}; in igt_ctx_readonly() local 1420 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_ctx_readonly() 1456 throttle_release(tq, ARRAY_SIZE(tq)); in igt_ctx_readonly()
|
| /linux/net/batman-adv/ |
| A D | bat_iv_ogm.c | 219 batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE; in batadv_iv_ogm_iface_enable() 300 new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty); in batadv_hop_penalty() 374 batadv_ogm_packet->tq, batadv_ogm_packet->ttl, in batadv_iv_ogm_send_to_if() 717 batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq, in batadv_iv_ogm_forward() 722 batadv_ogm_packet->tq, batadv_ogm_packet->ttl); in batadv_iv_ogm_forward() 999 batadv_ogm_packet->tq); in batadv_iv_ogm_orig_update() 1167 combined_tq = batadv_ogm_packet->tq * in batadv_iv_ogm_calc_tq() 1174 batadv_ogm_packet->tq = combined_tq; in batadv_iv_ogm_calc_tq() 1180 tq_iface_hop_penalty, batadv_ogm_packet->tq, in batadv_iv_ogm_calc_tq() 1353 if (ogm_packet->tq == 0) { in batadv_iv_ogm_process_per_outif() [all …]
|
| /linux/include/uapi/linux/can/ |
| A D | netlink.h | 34 __u32 tq; /* Time quanta (TQ) in nanoseconds */ member
|
| /linux/ipc/ |
| A D | sem.c | 285 struct sem_queue *q, *tq; in unmerge_queues() local 295 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { in unmerge_queues() 1147 struct sem_queue *q, *tq; in freeary() local 1164 list_for_each_entry_safe(q, tq, &sma->pending_const, list) { in freeary() 1169 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { in freeary() 1175 list_for_each_entry_safe(q, tq, &sem->pending_const, list) { in freeary() 1179 list_for_each_entry_safe(q, tq, &sem->pending_alter, list) { in freeary()
|