Lines Matching refs:offset

2266 			gfn_t offset;  in kvm_get_dirty_log_protect()  local
2275 offset = i * BITS_PER_LONG; in kvm_get_dirty_log_protect()
2277 offset, mask); in kvm_get_dirty_log_protect()
2335 gfn_t offset; in kvm_clear_dirty_log_protect() local
2375 for (offset = log->first_page, i = offset / BITS_PER_LONG, in kvm_clear_dirty_log_protect()
2377 i++, offset += BITS_PER_LONG) { in kvm_clear_dirty_log_protect()
2394 offset, mask); in kvm_clear_dirty_log_protect()
3169 static int next_segment(unsigned long len, int offset) in next_segment() argument
3171 if (len > PAGE_SIZE - offset) in next_segment()
3172 return PAGE_SIZE - offset; in next_segment()
3179 void *data, int offset, int len) in __kvm_read_guest_page() argument
3184 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_read_guest_page()
3190 r = __copy_from_user(data, (void __user *)addr + offset, len); in __kvm_read_guest_page()
3196 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, in kvm_read_guest_page() argument
3201 return __kvm_read_guest_page(slot, gfn, data, offset, len); in kvm_read_guest_page()
3206 int offset, int len) in kvm_vcpu_read_guest_page() argument
3210 return __kvm_read_guest_page(slot, gfn, data, offset, len); in kvm_vcpu_read_guest_page()
3218 int offset = offset_in_page(gpa); in kvm_read_guest() local
3221 while ((seg = next_segment(len, offset)) != 0) { in kvm_read_guest()
3222 ret = kvm_read_guest_page(kvm, gfn, data, offset, seg); in kvm_read_guest()
3225 offset = 0; in kvm_read_guest()
3238 int offset = offset_in_page(gpa); in kvm_vcpu_read_guest() local
3241 while ((seg = next_segment(len, offset)) != 0) { in kvm_vcpu_read_guest()
3242 ret = kvm_vcpu_read_guest_page(vcpu, gfn, data, offset, seg); in kvm_vcpu_read_guest()
3245 offset = 0; in kvm_vcpu_read_guest()
3255 void *data, int offset, unsigned long len) in __kvm_read_guest_atomic() argument
3260 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_read_guest_atomic()
3267 r = __copy_from_user_inatomic(data, (void __user *)addr + offset, len); in __kvm_read_guest_atomic()
3279 int offset = offset_in_page(gpa); in kvm_vcpu_read_guest_atomic() local
3281 return __kvm_read_guest_atomic(slot, gfn, data, offset, len); in kvm_vcpu_read_guest_atomic()
3288 const void *data, int offset, int len) in __kvm_write_guest_page() argument
3293 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_write_guest_page()
3299 r = __copy_to_user((void __user *)addr + offset, data, len); in __kvm_write_guest_page()
3307 const void *data, int offset, int len) in kvm_write_guest_page() argument
3311 return __kvm_write_guest_page(kvm, slot, gfn, data, offset, len); in kvm_write_guest_page()
3316 const void *data, int offset, int len) in kvm_vcpu_write_guest_page() argument
3320 return __kvm_write_guest_page(vcpu->kvm, slot, gfn, data, offset, len); in kvm_vcpu_write_guest_page()
3329 int offset = offset_in_page(gpa); in kvm_write_guest() local
3332 while ((seg = next_segment(len, offset)) != 0) { in kvm_write_guest()
3333 ret = kvm_write_guest_page(kvm, gfn, data, offset, seg); in kvm_write_guest()
3336 offset = 0; in kvm_write_guest()
3350 int offset = offset_in_page(gpa); in kvm_vcpu_write_guest() local
3353 while ((seg = next_segment(len, offset)) != 0) { in kvm_vcpu_write_guest()
3354 ret = kvm_vcpu_write_guest_page(vcpu, gfn, data, offset, seg); in kvm_vcpu_write_guest()
3357 offset = 0; in kvm_vcpu_write_guest()
3370 int offset = offset_in_page(gpa); in __kvm_gfn_to_hva_cache_init() local
3398 ghc->hva += offset; in __kvm_gfn_to_hva_cache_init()
3416 void *data, unsigned int offset, in kvm_write_guest_offset_cached() argument
3421 gpa_t gpa = ghc->gpa + offset; in kvm_write_guest_offset_cached()
3423 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_write_guest_offset_cached()
3437 r = __copy_to_user((void __user *)ghc->hva + offset, data, len); in kvm_write_guest_offset_cached()
3454 void *data, unsigned int offset, in kvm_read_guest_offset_cached() argument
3459 gpa_t gpa = ghc->gpa + offset; in kvm_read_guest_offset_cached()
3461 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_read_guest_offset_cached()
3475 r = __copy_from_user(data, (void __user *)ghc->hva + offset, len); in kvm_read_guest_offset_cached()
3495 int offset = offset_in_page(gpa); in kvm_clear_guest() local
3498 while ((seg = next_segment(len, offset)) != 0) { in kvm_clear_guest()
3499 ret = kvm_write_guest_page(kvm, gfn, zero_page, offset, seg); in kvm_clear_guest()
3502 offset = 0; in kvm_clear_guest()
4280 size_t size, loff_t *offset) in kvm_vcpu_stats_read() argument
4286 sizeof(vcpu->stat), user_buffer, size, offset); in kvm_vcpu_stats_read()
5081 size_t size, loff_t *offset) in kvm_vm_stats_read() argument
5087 sizeof(kvm->stat), user_buffer, size, offset); in kvm_vm_stats_read()
6087 static int kvm_get_stat_per_vm(struct kvm *kvm, size_t offset, u64 *val) in kvm_get_stat_per_vm() argument
6089 *val = *(u64 *)((void *)(&kvm->stat) + offset); in kvm_get_stat_per_vm()
6094 static int kvm_clear_stat_per_vm(struct kvm *kvm, size_t offset) in kvm_clear_stat_per_vm() argument
6096 *(u64 *)((void *)(&kvm->stat) + offset) = 0; in kvm_clear_stat_per_vm()
6101 static int kvm_get_stat_per_vcpu(struct kvm *kvm, size_t offset, u64 *val) in kvm_get_stat_per_vcpu() argument
6109 *val += *(u64 *)((void *)(&vcpu->stat) + offset); in kvm_get_stat_per_vcpu()
6114 static int kvm_clear_stat_per_vcpu(struct kvm *kvm, size_t offset) in kvm_clear_stat_per_vcpu() argument
6120 *(u64 *)((void *)(&vcpu->stat) + offset) = 0; in kvm_clear_stat_per_vcpu()
6133 stat_data->desc->desc.offset, val); in kvm_stat_data_get()
6137 stat_data->desc->desc.offset, val); in kvm_stat_data_get()
6155 stat_data->desc->desc.offset); in kvm_stat_data_clear()
6159 stat_data->desc->desc.offset); in kvm_stat_data_clear()
6183 unsigned offset = (long)_offset; in vm_stat_get() local
6190 kvm_get_stat_per_vm(kvm, offset, &tmp_val); in vm_stat_get()
6199 unsigned offset = (long)_offset; in vm_stat_clear() local
6207 kvm_clear_stat_per_vm(kvm, offset); in vm_stat_clear()
6219 unsigned offset = (long)_offset; in vcpu_stat_get() local
6226 kvm_get_stat_per_vcpu(kvm, offset, &tmp_val); in vcpu_stat_get()
6235 unsigned offset = (long)_offset; in vcpu_stat_clear() local
6243 kvm_clear_stat_per_vcpu(kvm, offset); in vcpu_stat_clear()
6320 (void *)(long)pdesc->desc.offset, fops); in kvm_init_debug()
6331 (void *)(long)pdesc->desc.offset, fops); in kvm_init_debug()