/linux-6.3-rc2/drivers/mmc/core/ |
A D | queue.c | 34 if ((mq->cqe_busy & MMC_CQE_DCMD_BUSY) && !mmc_cqe_dcmd_busy(mq)) in mmc_cqe_check_busy() 143 mmc_get_card(mq->card, &mq->ctx); in mmc_mq_recovery_handler() 161 mmc_put_card(mq->card, &mq->ctx); in mmc_mq_recovery_handler() 249 if (mq->recovery_needed || mq->busy) { in mmc_mq_queue_rq() 285 mq->busy = true; in mmc_mq_queue_rq() 329 mq->busy = false; in mmc_mq_queue_rq() 416 mq->card = card; in mmc_init_queue() 420 memset(&mq->tag_set, 0, sizeof(mq->tag_set)); in mmc_init_queue() 435 mq->tag_set.driver_data = mq; in mmc_init_queue() 454 disk = blk_mq_alloc_disk(&mq->tag_set, mq); in mmc_init_queue() [all …]
|
A D | block.c | 260 mq = &md->queue; in power_ro_lock_store() 646 mq = &md->queue; in mmc_blk_ioctl_cmd() 1495 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_cqe_complete_rq() 1574 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_hsq_issue_rw_rq() 2100 mq->in_flight[mmc_issue_type(mq, req)] -= 1; in mmc_blk_mq_dec_in_flight() 2107 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_mq_dec_in_flight() 2165 mmc_blk_mq_poll_completion(mq, mq->complete_req); in mmc_blk_mq_complete_prev_req() 2170 mmc_blk_mq_post_req(mq, mq->complete_req, true); in mmc_blk_mq_complete_prev_req() 2220 queue_work(mq->card->complete_wq, &mq->complete_work); in mmc_blk_mq_req_done() 2272 wait_event(mq->wait, mmc_blk_rw_wait_cond(mq, &err)); in mmc_blk_rw_wait() [all …]
|
A D | queue.h | 97 struct gendisk *mmc_init_queue(struct mmc_queue *mq, struct mmc_card *card); 104 void mmc_cqe_check_busy(struct mmc_queue *mq); 107 enum mmc_issue_type mmc_issue_type(struct mmc_queue *mq, struct request *req); 109 static inline int mmc_tot_in_flight(struct mmc_queue *mq) in mmc_tot_in_flight() argument 111 return mq->in_flight[MMC_ISSUE_SYNC] + in mmc_tot_in_flight() 112 mq->in_flight[MMC_ISSUE_DCMD] + in mmc_tot_in_flight() 113 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_tot_in_flight() 116 static inline int mmc_cqe_qcnt(struct mmc_queue *mq) in mmc_cqe_qcnt() argument 118 return mq->in_flight[MMC_ISSUE_DCMD] + in mmc_cqe_qcnt() 119 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_cqe_qcnt()
|
A D | block.h | 8 void mmc_blk_cqe_recovery(struct mmc_queue *mq); 12 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req); 14 void mmc_blk_mq_recovery(struct mmc_queue *mq);
|
/linux-6.3-rc2/drivers/md/ |
A D | dm-cache-policy-smq.c | 955 q_del(e->dirty ? &mq->dirty : &mq->clean, e); in del_queue() 1185 e = q_peek(&mq->dirty, mq->dirty.nr_levels, idle); in queue_writeback() 1365 kfree(mq); in smq_destroy() 1736 if (!mq) in __smq_create() 1746 mq->cache_blocks_per_hotspot_block = div64_u64(mq->hotspot_block_size, mq->cache_block_size); in __smq_create() 1764 init_allocator(&mq->cache_alloc, &mq->es, in __smq_create() 1788 q_init(&mq->hotspot, &mq->es, NR_HOTSPOT_LEVELS); in __smq_create() 1793 q_init(&mq->clean, &mq->es, NR_CACHE_LEVELS); in __smq_create() 1794 q_init(&mq->dirty, &mq->es, NR_CACHE_LEVELS); in __smq_create() 1802 if (h_init(&mq->hotspot_table, &mq->es, mq->nr_hotspot_blocks)) in __smq_create() [all …]
|
/linux-6.3-rc2/drivers/scsi/arm/ |
A D | msgqueue.c | 29 return mq; in mqe_alloc() 40 if (mq) { in mqe_free() 42 msgq->free = mq; in mqe_free() 85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() 102 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); in msgqueue_getmsg() 104 return mq ? &mq->msg : NULL; in msgqueue_getmsg() 120 if (mq) { in msgqueue_addmsg() 130 mq->msg.fifo = 0; in msgqueue_addmsg() 131 mq->next = NULL; in msgqueue_addmsg() 137 *mqp = mq; in msgqueue_addmsg() [all …]
|
/linux-6.3-rc2/drivers/sh/maple/ |
A D | maple.c | 142 mq = mdev->mq; in maple_release_device() 144 kfree(mq); in maple_release_device() 189 mq = kzalloc(sizeof(*mq), GFP_KERNEL); in maple_allocq() 190 if (!mq) in maple_allocq() 194 mq->dev = mdev; in maple_allocq() 196 if (!mq->recvbuf) in maple_allocq() 198 mq->recvbuf->buf = &((mq->recvbuf->bufx)[0]); in maple_allocq() 200 return mq; in maple_allocq() 203 kfree(mq); in maple_allocq() 226 if (!mdev->mq) { in maple_alloc_dev() [all …]
|
/linux-6.3-rc2/block/ |
A D | Makefile | 9 blk-lib.o blk-mq.o blk-mq-tag.o blk-stat.o \ 10 blk-mq-sysfs.o blk-mq-cpumap.o blk-mq-sched.o ioctl.o \ 24 obj-$(CONFIG_MQ_IOSCHED_DEADLINE) += mq-deadline.o 31 obj-$(CONFIG_BLK_MQ_PCI) += blk-mq-pci.o 32 obj-$(CONFIG_BLK_MQ_VIRTIO) += blk-mq-virtio.o 33 obj-$(CONFIG_BLK_MQ_RDMA) += blk-mq-rdma.o 36 obj-$(CONFIG_BLK_DEBUG_FS) += blk-mq-debugfs.o 37 obj-$(CONFIG_BLK_DEBUG_FS_ZONED)+= blk-mq-debugfs-zoned.o
|
/linux-6.3-rc2/drivers/misc/sgi-xp/ |
A D | xpc_uv.c | 117 mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset, in xpc_get_gru_mq_irq_uv() 119 if (mq->irq < 0) in xpc_get_gru_mq_irq_uv() 120 return mq->irq; in xpc_get_gru_mq_irq_uv() 122 mq->mmr_value = uv_read_global_mmr64(mmr_pnode, mq->mmr_offset); in xpc_get_gru_mq_irq_uv() 133 uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mq->mmr_value); in xpc_get_gru_mq_irq_uv() 169 mq->order, &mq->mmr_offset); in xpc_gru_mq_watchlist_alloc_uv() 177 mq->order, &mq->mmr_offset); in xpc_gru_mq_watchlist_alloc_uv() 222 if (mq == NULL) { in xpc_create_gru_mq_uv() 291 return mq; in xpc_create_gru_mq_uv() 305 kfree(mq); in xpc_create_gru_mq_uv() [all …]
|
/linux-6.3-rc2/drivers/misc/sgi-gru/ |
A D | grukservices.c | 553 mq->start = &mq->data; in gru_create_message_queue() 554 mq->start2 = &mq->data + (qlines / 2 - 1) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 555 mq->next = &mq->data; in gru_create_message_queue() 556 mq->limit = &mq->data + (qlines - 2) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 561 mqd->mq = mq; in gru_create_message_queue() 839 struct message_queue *mq = mqd->mq; in gru_free_message() local 849 pnext = mq->next; in gru_free_message() 854 } else if (pnext < mq->start2 && next >= mq->start2) { in gru_free_message() 860 mq->next = next; in gru_free_message() 871 struct message_queue *mq = mqd->mq; in gru_get_next_message() local [all …]
|
/linux-6.3-rc2/drivers/scsi/elx/efct/ |
A D | efct_hw_queues.c | 52 if (!mq) { in efct_hw_init_queues() 247 struct hw_mq *mq = kzalloc(sizeof(*mq), GFP_KERNEL); in efct_hw_new_mq() local 249 if (!mq) in efct_hw_new_mq() 257 mq->queue = &hw->mq[mq->instance]; in efct_hw_new_mq() 262 kfree(mq); in efct_hw_new_mq() 266 hw->hw_mq[mq->instance] = mq; in efct_hw_new_mq() 270 mq->queue->id, mq->entry_count); in efct_hw_new_mq() 271 return mq; in efct_hw_new_mq() 435 if (!mq) in efct_hw_del_mq() 439 mq->cq->eq->hw->hw_mq[mq->instance] = NULL; in efct_hw_del_mq() [all …]
|
/linux-6.3-rc2/drivers/mailbox/ |
A D | omap-mailbox.c | 269 spin_lock_irq(&mq->lock); in mbox_rx_work() 270 if (mq->full) { in mbox_rx_work() 271 mq->full = false; in mbox_rx_work() 297 mq->full = true; in __mbox_rx_interrupt() 334 mq = kzalloc(sizeof(*mq), GFP_KERNEL); in mbox_queue_alloc() 335 if (!mq) in mbox_queue_alloc() 344 return mq; in mbox_queue_alloc() 347 kfree(mq); in mbox_queue_alloc() 363 if (!mq) in omap_mbox_startup() 365 mbox->rxq = mq; in omap_mbox_startup() [all …]
|
/linux-6.3-rc2/Documentation/admin-guide/device-mapper/ |
A D | cache-policies.rst | 29 multiqueue (mq) 48 with the multiqueue (mq) policy. 50 The smq policy (vs mq) offers the promise of less memory utilization, 54 Users may switch from "mq" to "smq" simply by appropriately reloading a 56 mq policy's hints to be dropped. Also, performance of the cache may 63 The mq policy used a lot of memory; 88 bytes per cache block on a 64 78 mq placed entries in different levels of the multiqueue structures 91 The mq policy maintained a hit count for each cache block. For a 105 Testing smq shows substantially better performance than mq. 129 /dev/sdd 512 0 mq 4 sequential_threshold 1024 random_threshold 8"
|
/linux-6.3-rc2/include/linux/ |
A D | maple.h | 70 struct mapleq *mq; member 71 void (*callback) (struct mapleq * mq); 90 void (*callback) (struct mapleq * mq),
|
/linux-6.3-rc2/drivers/mtd/maps/ |
A D | vmu-flash.c | 89 static void vmu_blockread(struct mapleq *mq) in vmu_blockread() argument 94 mdev = mq->dev; in vmu_blockread() 101 memcpy(card->blockread, mq->recvbuf->buf + 12, in vmu_blockread() 191 list_del_init(&(mdev->mq->list)); in maple_vmu_read_block() 192 kfree(mdev->mq->sendbuf); in maple_vmu_read_block() 193 mdev->mq->sendbuf = NULL; in maple_vmu_read_block() 283 kfree(mdev->mq->sendbuf); in maple_vmu_write_block() 284 mdev->mq->sendbuf = NULL; in maple_vmu_write_block() 285 list_del_init(&(mdev->mq->list)); in maple_vmu_write_block() 511 mdev = mq->dev; in vmu_queryblocks() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/netronome/nfp/abm/ |
A D | qdisc.c | 304 struct nfp_qdisc *mq = nfp_abm_qdisc_tree_deref_slot(slot); in nfp_abm_qdisc_clear_mq() local 307 if (mq->type != NFP_QDISC_MQ || mq->netdev != netdev) in nfp_abm_qdisc_clear_mq() 309 for (i = 0; i < mq->num_children; i++) in nfp_abm_qdisc_clear_mq() 310 if (mq->children[i] == qdisc) { in nfp_abm_qdisc_clear_mq() 311 mq->children[i] = NULL; in nfp_abm_qdisc_clear_mq() 791 memset(&qdisc->mq.stats, 0, sizeof(qdisc->mq.stats)); in nfp_abm_mq_stats() 792 memset(&qdisc->mq.prev_stats, 0, sizeof(qdisc->mq.prev_stats)); in nfp_abm_mq_stats() 803 nfp_abm_stats_propagate(&qdisc->mq.stats, in nfp_abm_mq_stats() 805 nfp_abm_stats_propagate(&qdisc->mq.prev_stats, in nfp_abm_mq_stats() 810 nfp_abm_stats_calculate(&qdisc->mq.stats, &qdisc->mq.prev_stats, in nfp_abm_mq_stats()
|
/linux-6.3-rc2/Documentation/block/ |
A D | switching-sched.rst | 17 the fly to select one of mq-deadline, none, bfq, or kyber schedulers - 32 [mq-deadline] kyber bfq none 35 [none] mq-deadline kyber bfq
|
A D | blk-mq.rst | 4 Multi-Queue Block IO Queueing Mechanism (blk-mq) 36 to different CPUs) wanted to perform block IO. Instead of this, the blk-mq API 45 for instance), blk-mq takes action: it will store and manage IO requests to 49 blk-mq has two group of queues: software staging queues and hardware dispatch 59 resources to accept more requests, blk-mq will places requests on a temporary 142 … Block IO: Introducing Multi-queue SSD Access on Multi-core Systems <http://kernel.dk/blk-mq.pdf>`_ 151 .. kernel-doc:: include/linux/blk-mq.h 153 .. kernel-doc:: block/blk-mq.c
|
A D | null_blk.rst | 114 0 nullb* use default blk-mq io scheduler 122 0 Register as a non-blocking blk-mq driver device. 123 1 Register as a blocking blk-mq driver device, null_blk will set 133 1 Tag set shared between devices for blk-mq. Only makes sense with
|
/linux-6.3-rc2/arch/riscv/boot/dts/allwinner/ |
A D | Makefile | 7 dtb-$(CONFIG_ARCH_SUNXI) += sun20i-d1-mangopi-mq-pro.dtb 9 dtb-$(CONFIG_ARCH_SUNXI) += sun20i-d1s-mangopi-mq.dtb
|
/linux-6.3-rc2/drivers/input/mouse/ |
A D | maplemouse.c | 27 static void dc_mouse_callback(struct mapleq *mq) in dc_mouse_callback() argument 30 struct maple_device *mapledev = mq->dev; in dc_mouse_callback() 33 unsigned char *res = mq->recvbuf->buf; in dc_mouse_callback()
|
/linux-6.3-rc2/Documentation/devicetree/bindings/powerpc/4xx/ |
A D | ppc440spe-adma.txt | 82 - compatible : "ibm,mq-440spe"; 87 MQ0: mq { 88 compatible = "ibm,mq-440spe";
|
/linux-6.3-rc2/drivers/net/wireless/intel/iwlwifi/dvm/ |
A D | main.c | 2033 int mq = priv->queue_to_mac80211[queue]; in iwl_stop_sw_queue() local 2041 queue, mq); in iwl_stop_sw_queue() 2045 set_bit(mq, &priv->transport_queue_stop); in iwl_stop_sw_queue() 2046 ieee80211_stop_queue(priv->hw, mq); in iwl_stop_sw_queue() 2052 int mq = priv->queue_to_mac80211[queue]; in iwl_wake_sw_queue() local 2060 queue, mq); in iwl_wake_sw_queue() 2064 clear_bit(mq, &priv->transport_queue_stop); in iwl_wake_sw_queue() 2067 ieee80211_wake_queue(priv->hw, mq); in iwl_wake_sw_queue() 2072 int mq; in iwlagn_lift_passive_no_rx() local 2077 for (mq = 0; mq < IWLAGN_FIRST_AMPDU_QUEUE; mq++) { in iwlagn_lift_passive_no_rx() [all …]
|
/linux-6.3-rc2/arch/ia64/include/asm/sn/ |
A D | sn_sal.h | 92 sn_mq_watchlist_alloc(int blade, void *mq, unsigned int mq_size, in sn_mq_watchlist_alloc() argument 100 addr = (unsigned long)mq; in sn_mq_watchlist_alloc()
|
/linux-6.3-rc2/drivers/input/joystick/ |
A D | maplecontrol.c | 27 static void dc_pad_callback(struct mapleq *mq) in dc_pad_callback() argument 30 struct maple_device *mapledev = mq->dev; in dc_pad_callback() 33 unsigned char *res = mq->recvbuf->buf; in dc_pad_callback()
|