Home
last modified time | relevance | path

Searched refs:opf (Results 1 – 25 of 57) sorted by relevance

123

/linux/arch/sparc/kernel/
A Dvisemul.c307 switch (opf) { in edge()
352 switch (opf) { in edge()
394 switch (opf) { in array()
484 switch (opf) { in pformat()
594 switch (opf) { in pmul()
715 switch (opf) { in pcmp()
807 unsigned int opf; in vis_emul() local
822 switch (opf) { in vis_emul()
843 pmul(regs, insn, opf); in vis_emul()
855 pcmp(regs, insn, opf); in vis_emul()
[all …]
/linux/drivers/md/
A Ddm-io.c306 static void do_region(const blk_opf_t opf, unsigned int region, in do_region() argument
319 const enum req_op op = opf & REQ_OP_MASK; in do_region()
353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, in do_region()
385 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, in dispatch_io() argument
400 if (where[i].count || (opf & REQ_PREFLUSH)) in dispatch_io()
401 do_region(opf, i, where + i, dp, io, ioprio); in dispatch_io()
412 struct dm_io_region *where, blk_opf_t opf, in async_io() argument
428 dispatch_io(opf, num_regions, where, dp, io, ioprio); in async_io()
445 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, in sync_io() argument
452 async_io(client, num_regions, where, opf | REQ_SYNC, dp, in sync_io()
A Ddm-snap-persistent.c232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, in chunk_io() argument
241 .bi_opf = opf, in chunk_io()
288 static int area_io(struct pstore *ps, blk_opf_t opf) in area_io() argument
292 return chunk_io(ps, ps->area, chunk, opf, 0); in area_io()
/linux/block/
A Dblk-cgroup-rwstat.h62 blk_opf_t opf, uint64_t val) in blkg_rwstat_add() argument
66 if (op_is_discard(opf)) in blkg_rwstat_add()
68 else if (op_is_write(opf)) in blkg_rwstat_add()
75 if (op_is_sync(opf)) in blkg_rwstat_add()
A Dblk-wbt.c533 static inline unsigned int get_limit(struct rq_wb *rwb, blk_opf_t opf) in get_limit() argument
537 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD) in get_limit()
548 if ((opf & REQ_HIPRIO) || wb_recent_wait(rwb)) in get_limit()
550 else if ((opf & REQ_BACKGROUND) || close_io(rwb)) { in get_limit()
565 blk_opf_t opf; member
571 return rq_wait_inc_below(rqw, get_limit(data->rwb, data->opf)); in wbt_inflight_cb()
585 blk_opf_t opf) in __wbt_wait() argument
591 .opf = opf, in __wbt_wait()
A Dbio.c246 unsigned short max_vecs, blk_opf_t opf) in bio_init() argument
250 bio->bi_opf = opf; in bio_init()
308 bio->bi_opf = opf; in bio_reset()
367 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) in blk_next_bio() argument
444 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, in bio_alloc_percpu_cache() argument
504 blk_opf_t opf, gfp_t gfp_mask, in bio_alloc_bioset() argument
515 if (opf & REQ_ALLOC_CACHE) { in bio_alloc_bioset()
526 opf &= ~REQ_ALLOC_CACHE; in bio_alloc_bioset()
563 opf &= ~REQ_ALLOC_CACHE; in bio_alloc_bioset()
578 bio_init(bio, bdev, bvl, nr_vecs, opf); in bio_alloc_bioset()
[all …]
A Dblk-mq.h87 static inline enum hctx_type blk_mq_get_hctx_type(blk_opf_t opf) in blk_mq_get_hctx_type() argument
94 if (opf & REQ_POLLED) in blk_mq_get_hctx_type()
96 else if ((opf & REQ_OP_MASK) == REQ_OP_READ) in blk_mq_get_hctx_type()
108 blk_opf_t opf, in blk_mq_map_queue() argument
111 return ctx->hctxs[blk_mq_get_hctx_type(opf)]; in blk_mq_map_queue()
A Dfops.c30 blk_opf_t opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in dio_bio_write_op() local
34 opf |= REQ_FUA; in dio_bio_write_op()
35 return opf; in dio_bio_write_op()
173 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO() local
178 opf |= REQ_ALLOC_CACHE; in __blkdev_direct_IO()
179 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO()
252 bio = bio_alloc(bdev, nr_pages, opf, GFP_KERNEL); in __blkdev_direct_IO()
308 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO_async() local
315 opf |= REQ_ALLOC_CACHE; in __blkdev_direct_IO_async()
316 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO_async()
A Dbfq-cgroup.c223 blk_opf_t opf) in bfqg_stats_update_io_add() argument
225 blkg_rwstat_add(&bfqg->stats.queued, opf, 1); in bfqg_stats_update_io_add()
231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_remove() argument
233 blkg_rwstat_add(&bfqg->stats.queued, opf, -1); in bfqg_stats_update_io_remove()
236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_merged() argument
238 blkg_rwstat_add(&bfqg->stats.merged, opf, 1); in bfqg_stats_update_io_merged()
242 u64 io_start_time_ns, blk_opf_t opf) in bfqg_stats_update_completion() argument
248 blkg_rwstat_add(&stats->service_time, opf, in bfqg_stats_update_completion()
251 blkg_rwstat_add(&stats->wait_time, opf, in bfqg_stats_update_completion()
257 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { } in bfqg_stats_update_io_remove() argument
[all …]
A Dbfq-iosched.h1068 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf);
1069 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf);
1071 u64 io_start_time_ns, blk_opf_t opf);
1079 blk_opf_t opf);
/linux/drivers/nvme/target/
A Dio-cmd-bdev.c251 blk_opf_t opf; in nvmet_bdev_execute_rw() local
266 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_rw()
268 opf |= REQ_FUA; in nvmet_bdev_execute_rw()
271 opf = REQ_OP_READ; in nvmet_bdev_execute_rw()
276 opf |= REQ_NOMERGE; in nvmet_bdev_execute_rw()
283 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_rw()
285 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf, in nvmet_bdev_execute_rw()
312 opf, GFP_KERNEL); in nvmet_bdev_execute_rw()
/linux/drivers/scsi/device_handler/
A Dscsi_dh_hp_sw.c83 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_tur() local
103 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_tur()
132 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_start_stop() local
157 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_start_stop()
A Dscsi_dh_emc.c242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_trespass_cmd() local
269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len, in send_trespass_cmd()
/linux/fs/
A Ddirect-io.c169 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_refill_pages()
246 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_complete()
336 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_end_aio()
427 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_submit()
502 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_complete()
606 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in get_more_blocks()
789 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in submit_page_section()
906 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in do_direct_IO()
1182 dio->opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in __blockdev_direct_IO()
1184 dio->opf |= REQ_NOWAIT; in __blockdev_direct_IO()
[all …]
A Dmpage.c168 blk_opf_t opf = REQ_OP_READ; in do_mpage_readpage() local
177 opf |= REQ_RAHEAD; in do_mpage_readpage()
290 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf, in do_mpage_readpage()
/linux/fs/hfsplus/
A Dwrapper.c49 void *buf, void **data, blk_opf_t opf) in hfsplus_submit_bio() argument
51 const enum req_op op = opf & REQ_OP_MASK; in hfsplus_submit_bio()
68 bio = bio_alloc(sb->s_bdev, 1, opf, GFP_NOIO); in hfsplus_submit_bio()
/linux/drivers/block/
A Dbrd.c193 unsigned int len, unsigned int off, blk_opf_t opf, in brd_do_bvec() argument
199 if (op_is_write(opf)) { in brd_do_bvec()
204 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO; in brd_do_bvec()
212 if (!op_is_write(opf)) { in brd_do_bvec()
/linux/fs/nilfs2/
A Dbtnode.c87 sector_t pblocknr, blk_opf_t opf, in nilfs_btnode_submit_block() argument
120 if (opf & REQ_RAHEAD) { in nilfs_btnode_submit_block()
138 submit_bh(opf, bh); in nilfs_btnode_submit_block()
A Dmdt.c113 nilfs_mdt_submit_block(struct inode *inode, unsigned long blkoff, blk_opf_t opf, in nilfs_mdt_submit_block() argument
128 if (opf & REQ_RAHEAD) { in nilfs_mdt_submit_block()
150 submit_bh(opf, bh); in nilfs_mdt_submit_block()
154 opf & REQ_OP_MASK); in nilfs_mdt_submit_block()
/linux/include/linux/
A Dbio.h359 blk_opf_t opf, gfp_t gfp_mask,
372 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) in bio_alloc() argument
374 return bio_alloc_bioset(bdev, nr_vecs, opf, gfp_mask, &fs_bio_set); in bio_alloc()
410 unsigned short max_vecs, blk_opf_t opf);
412 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf);
681 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp);
A Dblktrace_api.h113 void blk_fill_rwbs(char *rwbs, blk_opf_t opf);
/linux/drivers/target/
A Dtarget_core_iblock.c361 blk_opf_t opf) in iblock_get_bio() argument
370 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, in iblock_get_bio()
740 blk_opf_t opf; in iblock_execute_rw() local
753 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in iblock_execute_rw()
761 opf |= REQ_FUA; in iblock_execute_rw()
763 opf |= REQ_FUA; in iblock_execute_rw()
766 opf = REQ_OP_READ; in iblock_execute_rw()
781 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); in iblock_execute_rw()
814 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); in iblock_execute_rw()
/linux/kernel/trace/
A Dblktrace.c236 what |= MASK_TC_BIT(opf, SYNC); in __blk_add_trace()
237 what |= MASK_TC_BIT(opf, RAHEAD); in __blk_add_trace()
238 what |= MASK_TC_BIT(opf, META); in __blk_add_trace()
239 what |= MASK_TC_BIT(opf, PREFLUSH); in __blk_add_trace()
240 what |= MASK_TC_BIT(opf, FUA); in __blk_add_trace()
1879 if (opf & REQ_PREFLUSH) in blk_fill_rwbs()
1882 switch (opf & REQ_OP_MASK) { in blk_fill_rwbs()
1903 if (opf & REQ_FUA) in blk_fill_rwbs()
1905 if (opf & REQ_RAHEAD) in blk_fill_rwbs()
1907 if (opf & REQ_SYNC) in blk_fill_rwbs()
[all …]
/linux/fs/btrfs/
A Dbio.h102 struct btrfs_bio *btrfs_bio_alloc(unsigned int nr_vecs, blk_opf_t opf,
/linux/fs/gfs2/
A Dlops.h20 void gfs2_log_submit_bio(struct bio **biop, blk_opf_t opf);

Completed in 59 milliseconds

123