Searched refs:chunk_end (Results 1 – 3 of 3) sorted by relevance
470 size_t chunk_end = tcg_ctx->gen_insn_end_off[insn]; in tb_gen_code() local471 if (chunk_end > chunk_start) { in tb_gen_code()475 chunk_end - chunk_start); in tb_gen_code()476 chunk_start = chunk_end; in tb_gen_code()
1264 uint8_t *chunk_start, uint8_t *chunk_end) in qemu_rdma_register_and_get_keys() argument1287 uint64_t len = chunk_end - chunk_start; in qemu_rdma_register_and_get_keys()2003 uint8_t *chunk_start, *chunk_end; in qemu_rdma_write_one() local2040 chunk_end = ram_chunk_end(block, chunk + chunks); in qemu_rdma_write_one()2128 chunk_start, chunk_end)) { in qemu_rdma_write_one()2262 uint8_t *chunk_end; in qemu_rdma_buffer_mergeable() local2274 chunk_end = ram_chunk_end(block, rdma->current_chunk); in qemu_rdma_buffer_mergeable()2295 if ((host_addr + len) > chunk_end) { in qemu_rdma_buffer_mergeable()3670 uint8_t *chunk_start, *chunk_end; in rdma_registration_handle() local3712 chunk_end = ram_chunk_end(block, chunk + reg->chunks); in rdma_registration_handle()[all …]
770 unsigned long chunk_end = QEMU_ALIGN_UP(start + npages, chunk_pages); in migration_clear_memory_region_dirty_bitmap_range() local776 for (i = chunk_start; i < chunk_end; i += chunk_pages) { in migration_clear_memory_region_dirty_bitmap_range()
Completed in 20 milliseconds