Lines Matching refs:i
249 int i, me; in kvm_make_vcpus_request_mask() local
257 for_each_set_bit(i, vcpu_bitmap, KVM_MAX_VCPUS) { in kvm_make_vcpus_request_mask()
258 vcpu = kvm_get_vcpu(kvm, i); in kvm_make_vcpus_request_mask()
274 unsigned long i; in kvm_make_all_cpus_request() local
283 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_make_all_cpus_request()
484 unsigned long i; in kvm_destroy_vcpus() local
487 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_destroy_vcpus()
489 xa_erase(&kvm->vcpu_array, i); in kvm_destroy_vcpus()
497 WARN_ON_ONCE(xa_load(&kvm->vcpu_array, i) || kvm_get_vcpu(kvm, i)); in kvm_destroy_vcpus()
571 int i, idx; in kvm_handle_hva_range() local
587 for (i = 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { in kvm_handle_hva_range()
590 slots = __kvm_memslots(kvm, i); in kvm_handle_hva_range()
1000 int i; in kvm_destroy_vm_debugfs() local
1010 for (i = 0; i < kvm_debugfs_num_entries; i++) in kvm_destroy_vm_debugfs()
1011 kfree(kvm->debugfs_stat_data[i]); in kvm_destroy_vm_debugfs()
1023 int i, ret = -ENOMEM; in kvm_create_vm_debugfs() local
1051 for (i = 0; i < kvm_vm_stats_header.num_desc; ++i) { in kvm_create_vm_debugfs()
1052 pdesc = &kvm_vm_stats_desc[i]; in kvm_create_vm_debugfs()
1060 kvm->debugfs_stat_data[i] = stat_data; in kvm_create_vm_debugfs()
1066 for (i = 0; i < kvm_vcpu_stats_header.num_desc; ++i) { in kvm_create_vm_debugfs()
1067 pdesc = &kvm_vcpu_stats_desc[i]; in kvm_create_vm_debugfs()
1075 kvm->debugfs_stat_data[i + kvm_vm_stats_header.num_desc] = stat_data; in kvm_create_vm_debugfs()
1110 int r, i, j; in kvm_create_vm() local
1159 for (i = 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { in kvm_create_vm()
1161 slots = &kvm->__memslots[i][j]; in kvm_create_vm()
1170 slots->generation = i; in kvm_create_vm()
1173 rcu_assign_pointer(kvm->memslots[i], &kvm->__memslots[i][0]); in kvm_create_vm()
1177 for (i = 0; i < KVM_NR_BUSES; i++) { in kvm_create_vm()
1178 rcu_assign_pointer(kvm->buses[i], in kvm_create_vm()
1180 if (!kvm->buses[i]) in kvm_create_vm()
1230 for (i = 0; i < KVM_NR_BUSES; i++) in kvm_create_vm()
1231 kfree(kvm_get_bus(kvm, i)); in kvm_create_vm()
1266 int i; in kvm_destroy_vm() local
1278 for (i = 0; i < KVM_NR_BUSES; i++) { in kvm_destroy_vm()
1279 struct kvm_io_bus *bus = kvm_get_bus(kvm, i); in kvm_destroy_vm()
1283 kvm->buses[i] = NULL; in kvm_destroy_vm()
1310 for (i = 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { in kvm_destroy_vm()
1311 kvm_free_memslots(kvm, &kvm->__memslots[i][0]); in kvm_destroy_vm()
1312 kvm_free_memslots(kvm, &kvm->__memslots[i][1]); in kvm_destroy_vm()
1374 unsigned long i, j; in kvm_trylock_all_vcpus() local
1378 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_trylock_all_vcpus()
1385 if (i == j) in kvm_trylock_all_vcpus()
1396 unsigned long i, j; in kvm_lock_all_vcpus() local
1401 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_lock_all_vcpus()
1410 if (i == j) in kvm_lock_all_vcpus()
1421 unsigned long i; in kvm_unlock_all_vcpus() local
1425 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_unlock_all_vcpus()
2160 int i, as_id, id; in kvm_get_dirty_log() local
2185 for (i = 0; !any && i < n/sizeof(long); ++i) in kvm_get_dirty_log()
2186 any = (*memslot)->dirty_bitmap[i]; in kvm_get_dirty_log()
2223 int i, as_id, id; in kvm_get_dirty_log_protect() local
2264 for (i = 0; i < n / sizeof(long); i++) { in kvm_get_dirty_log_protect()
2268 if (!dirty_bitmap[i]) in kvm_get_dirty_log_protect()
2272 mask = xchg(&dirty_bitmap[i], 0); in kvm_get_dirty_log_protect()
2273 dirty_bitmap_buffer[i] = mask; in kvm_get_dirty_log_protect()
2275 offset = i * BITS_PER_LONG; in kvm_get_dirty_log_protect()
2336 unsigned long i, n; in kvm_clear_dirty_log_protect() local
2375 for (offset = log->first_page, i = offset / BITS_PER_LONG, in kvm_clear_dirty_log_protect()
2377 i++, offset += BITS_PER_LONG) { in kvm_clear_dirty_log_protect()
2379 atomic_long_t *p = (atomic_long_t *) &dirty_bitmap[i]; in kvm_clear_dirty_log_protect()
2472 int i; in kvm_handle_gfn_range() local
2485 for (i = 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { in kvm_handle_gfn_range()
2486 slots = __kvm_memslots(kvm, i); in kvm_handle_gfn_range()
2555 unsigned long i; in kvm_vm_set_mem_attributes() local
2573 for (i = start; i < end; i++) { in kvm_vm_set_mem_attributes()
2574 r = xa_reserve(&kvm->mem_attr_array, i, GFP_KERNEL_ACCOUNT); in kvm_vm_set_mem_attributes()
2583 for (i = start; i < end; i++) { in kvm_vm_set_mem_attributes()
2584 r = xa_err(xa_store(&kvm->mem_attr_array, i, entry, in kvm_vm_set_mem_attributes()
3958 int nr_vcpus, start, i, idx, yielded; in kvm_vcpu_on_spin() local
3992 for (i = 0; i < nr_vcpus; i++) { in kvm_vcpu_on_spin()
3993 idx = (start + i) % nr_vcpus; in kvm_vcpu_on_spin()
4019 WRITE_ONCE(kvm->last_boosted_vcpu, i); in kvm_vcpu_on_spin()
4968 unsigned long i; in kvm_vm_ioctl_reset_dirty_pages() local
4977 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_vm_ioctl_reset_dirty_pages()
4999 int i; in kvm_are_all_memslots_empty() local
5003 for (i = 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { in kvm_are_all_memslots_empty()
5004 if (!kvm_memslots_empty(__kvm_memslots(kvm, i))) in kvm_are_all_memslots_empty()
5762 int i; in kvm_io_bus_destroy() local
5764 for (i = 0; i < bus->dev_count; i++) { in kvm_io_bus_destroy()
5765 struct kvm_io_device *pos = bus->range[i].dev; in kvm_io_bus_destroy()
5938 int i; in kvm_io_bus_register_dev() local
5963 for (i = 0; i < bus->dev_count; i++) in kvm_io_bus_register_dev()
5964 if (kvm_io_bus_cmp(&bus->range[i], &range) > 0) in kvm_io_bus_register_dev()
5967 memcpy(new_bus, bus, sizeof(*bus) + i * sizeof(struct kvm_io_range)); in kvm_io_bus_register_dev()
5969 new_bus->range[i] = range; in kvm_io_bus_register_dev()
5970 memcpy(new_bus->range + i + 1, bus->range + i, in kvm_io_bus_register_dev()
5971 (bus->dev_count - i) * sizeof(struct kvm_io_range)); in kvm_io_bus_register_dev()
5982 int i; in kvm_io_bus_unregister_dev() local
5991 for (i = 0; i < bus->dev_count; i++) { in kvm_io_bus_unregister_dev()
5992 if (bus->range[i].dev == dev) { in kvm_io_bus_unregister_dev()
5997 if (i == bus->dev_count) in kvm_io_bus_unregister_dev()
6003 memcpy(new_bus, bus, struct_size(bus, range, i)); in kvm_io_bus_unregister_dev()
6005 memcpy(new_bus->range + i, bus->range + i + 1, in kvm_io_bus_unregister_dev()
6006 flex_array_size(new_bus, range, new_bus->dev_count - i)); in kvm_io_bus_unregister_dev()
6103 unsigned long i; in kvm_get_stat_per_vcpu() local
6108 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_get_stat_per_vcpu()
6116 unsigned long i; in kvm_clear_stat_per_vcpu() local
6119 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_clear_stat_per_vcpu()
6308 int i; in kvm_init_debug() local
6312 for (i = 0; i < kvm_vm_stats_header.num_desc; ++i) { in kvm_init_debug()
6313 pdesc = &kvm_vm_stats_desc[i]; in kvm_init_debug()
6323 for (i = 0; i < kvm_vcpu_stats_header.num_desc; ++i) { in kvm_init_debug()
6324 pdesc = &kvm_vcpu_stats_desc[i]; in kvm_init_debug()