Lines Matching refs:io
34 struct io { struct
93 static void store_io_and_region_in_bio(struct bio *bio, struct io *io, in store_io_and_region_in_bio() argument
96 if (unlikely(!IS_ALIGNED((unsigned long)io, DM_IO_MAX_REGIONS))) { in store_io_and_region_in_bio()
97 DMCRIT("Unaligned struct io pointer %p", io); in store_io_and_region_in_bio()
101 bio->bi_private = (void *)((unsigned long)io | region); in store_io_and_region_in_bio()
104 static void retrieve_io_and_region_from_bio(struct bio *bio, struct io **io, in retrieve_io_and_region_from_bio() argument
109 *io = (void *)(val & -(unsigned long)DM_IO_MAX_REGIONS); in retrieve_io_and_region_from_bio()
119 static void complete_io(struct io *io) in complete_io() argument
121 unsigned long error_bits = io->error_bits; in complete_io()
122 io_notify_fn fn = io->callback; in complete_io()
123 void *context = io->context; in complete_io()
125 if (io->vma_invalidate_size) in complete_io()
126 invalidate_kernel_vmap_range(io->vma_invalidate_address, in complete_io()
127 io->vma_invalidate_size); in complete_io()
129 mempool_free(io, &io->client->pool); in complete_io()
133 static void dec_count(struct io *io, unsigned int region, blk_status_t error) in dec_count() argument
136 set_bit(region, &io->error_bits); in dec_count()
138 if (atomic_dec_and_test(&io->count)) in dec_count()
139 complete_io(io); in dec_count()
144 struct io *io; in endio() local
154 retrieve_io_and_region_from_bio(bio, &io, ®ion); in endio()
159 dec_count(io, region, error); in endio()
308 struct io *io) in do_region() argument
330 atomic_inc(&io->count); in do_region()
331 dec_count(io, region, BLK_STS_NOTSUPP); in do_region()
354 &io->client->bios); in do_region()
357 store_io_and_region_in_bio(bio, io, region); in do_region()
379 atomic_inc(&io->count); in do_region()
386 struct io *io, int sync) in dispatch_io() argument
403 do_region(opf, i, where + i, dp, io); in dispatch_io()
410 dec_count(io, 0, 0); in dispatch_io()
430 struct io *io; in sync_io() local
440 io = mempool_alloc(&client->pool, GFP_NOIO); in sync_io()
441 io->error_bits = 0; in sync_io()
442 atomic_set(&io->count, 1); /* see dispatch_io() */ in sync_io()
443 io->client = client; in sync_io()
444 io->callback = sync_io_complete; in sync_io()
445 io->context = &sio; in sync_io()
447 io->vma_invalidate_address = dp->vma_invalidate_address; in sync_io()
448 io->vma_invalidate_size = dp->vma_invalidate_size; in sync_io()
450 dispatch_io(opf, num_regions, where, dp, io, 1); in sync_io()
464 struct io *io; in async_io() local
472 io = mempool_alloc(&client->pool, GFP_NOIO); in async_io()
473 io->error_bits = 0; in async_io()
474 atomic_set(&io->count, 1); /* see dispatch_io() */ in async_io()
475 io->client = client; in async_io()
476 io->callback = fn; in async_io()
477 io->context = context; in async_io()
479 io->vma_invalidate_address = dp->vma_invalidate_address; in async_io()
480 io->vma_invalidate_size = dp->vma_invalidate_size; in async_io()
482 dispatch_io(opf, num_regions, where, dp, io, 0); in async_io()
545 _dm_io_cache = KMEM_CACHE(io, 0); in dm_io_init()