Lines Matching refs:len

3169 static int next_segment(unsigned long len, int offset)  in next_segment()  argument
3171 if (len > PAGE_SIZE - offset) in next_segment()
3174 return len; in next_segment()
3179 void *data, int offset, int len) in __kvm_read_guest_page() argument
3184 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_read_guest_page()
3190 r = __copy_from_user(data, (void __user *)addr + offset, len); in __kvm_read_guest_page()
3197 int len) in kvm_read_guest_page() argument
3201 return __kvm_read_guest_page(slot, gfn, data, offset, len); in kvm_read_guest_page()
3206 int offset, int len) in kvm_vcpu_read_guest_page() argument
3210 return __kvm_read_guest_page(slot, gfn, data, offset, len); in kvm_vcpu_read_guest_page()
3214 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len) in kvm_read_guest() argument
3221 while ((seg = next_segment(len, offset)) != 0) { in kvm_read_guest()
3226 len -= seg; in kvm_read_guest()
3234 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len) in kvm_vcpu_read_guest() argument
3241 while ((seg = next_segment(len, offset)) != 0) { in kvm_vcpu_read_guest()
3246 len -= seg; in kvm_vcpu_read_guest()
3255 void *data, int offset, unsigned long len) in __kvm_read_guest_atomic() argument
3260 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_read_guest_atomic()
3267 r = __copy_from_user_inatomic(data, (void __user *)addr + offset, len); in __kvm_read_guest_atomic()
3275 void *data, unsigned long len) in kvm_vcpu_read_guest_atomic() argument
3281 return __kvm_read_guest_atomic(slot, gfn, data, offset, len); in kvm_vcpu_read_guest_atomic()
3288 const void *data, int offset, int len) in __kvm_write_guest_page() argument
3293 if (WARN_ON_ONCE(offset + len > PAGE_SIZE)) in __kvm_write_guest_page()
3299 r = __copy_to_user((void __user *)addr + offset, data, len); in __kvm_write_guest_page()
3307 const void *data, int offset, int len) in kvm_write_guest_page() argument
3311 return __kvm_write_guest_page(kvm, slot, gfn, data, offset, len); in kvm_write_guest_page()
3316 const void *data, int offset, int len) in kvm_vcpu_write_guest_page() argument
3320 return __kvm_write_guest_page(vcpu->kvm, slot, gfn, data, offset, len); in kvm_vcpu_write_guest_page()
3325 unsigned long len) in kvm_write_guest() argument
3332 while ((seg = next_segment(len, offset)) != 0) { in kvm_write_guest()
3337 len -= seg; in kvm_write_guest()
3346 unsigned long len) in kvm_vcpu_write_guest() argument
3353 while ((seg = next_segment(len, offset)) != 0) { in kvm_vcpu_write_guest()
3358 len -= seg; in kvm_vcpu_write_guest()
3368 gpa_t gpa, unsigned long len) in __kvm_gfn_to_hva_cache_init() argument
3372 gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; in __kvm_gfn_to_hva_cache_init()
3403 ghc->len = len; in __kvm_gfn_to_hva_cache_init()
3408 gpa_t gpa, unsigned long len) in kvm_gfn_to_hva_cache_init() argument
3411 return __kvm_gfn_to_hva_cache_init(slots, ghc, gpa, len); in kvm_gfn_to_hva_cache_init()
3417 unsigned long len) in kvm_write_guest_offset_cached() argument
3423 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_write_guest_offset_cached()
3427 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_write_guest_offset_cached()
3435 return kvm_write_guest(kvm, gpa, data, len); in kvm_write_guest_offset_cached()
3437 r = __copy_to_user((void __user *)ghc->hva + offset, data, len); in kvm_write_guest_offset_cached()
3447 void *data, unsigned long len) in kvm_write_guest_cached() argument
3449 return kvm_write_guest_offset_cached(kvm, ghc, data, 0, len); in kvm_write_guest_cached()
3455 unsigned long len) in kvm_read_guest_offset_cached() argument
3461 if (WARN_ON_ONCE(len + offset > ghc->len)) in kvm_read_guest_offset_cached()
3465 if (__kvm_gfn_to_hva_cache_init(slots, ghc, ghc->gpa, ghc->len)) in kvm_read_guest_offset_cached()
3473 return kvm_read_guest(kvm, gpa, data, len); in kvm_read_guest_offset_cached()
3475 r = __copy_from_user(data, (void __user *)ghc->hva + offset, len); in kvm_read_guest_offset_cached()
3484 void *data, unsigned long len) in kvm_read_guest_cached() argument
3486 return kvm_read_guest_offset_cached(kvm, ghc, data, 0, len); in kvm_read_guest_cached()
3490 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len) in kvm_clear_guest() argument
3498 while ((seg = next_segment(len, offset)) != 0) { in kvm_clear_guest()
3503 len -= seg; in kvm_clear_guest()
4583 if (kvm_sigmask.len != sizeof(sigset)) in kvm_vcpu_ioctl()
4669 if (kvm_sigmask.len != sizeof(compat_sigset_t)) in kvm_vcpu_compat_ioctl()
5786 if (r2->len) { in kvm_io_bus_cmp()
5787 addr1 += r1->len; in kvm_io_bus_cmp()
5788 addr2 += r2->len; in kvm_io_bus_cmp()
5803 gpa_t addr, int len) in kvm_io_bus_get_first_dev() argument
5810 .len = len, in kvm_io_bus_get_first_dev()
5831 idx = kvm_io_bus_get_first_dev(bus, range->addr, range->len); in __kvm_io_bus_write()
5838 range->len, val)) in __kvm_io_bus_write()
5847 int len, const void *val) in kvm_io_bus_write() argument
5855 .len = len, in kvm_io_bus_write()
5867 gpa_t addr, int len, const void *val, long cookie) in kvm_io_bus_write_cookie() argument
5874 .len = len, in kvm_io_bus_write_cookie()
5884 if (!kvm_iodevice_write(vcpu, bus->range[cookie].dev, addr, len, in kvm_io_bus_write_cookie()
5900 idx = kvm_io_bus_get_first_dev(bus, range->addr, range->len); in __kvm_io_bus_read()
5907 range->len, val)) in __kvm_io_bus_read()
5916 int len, void *val) in kvm_io_bus_read() argument
5924 .len = len, in kvm_io_bus_read()
5936 int len, struct kvm_io_device *dev) in kvm_io_bus_register_dev() argument
5959 .len = len, in kvm_io_bus_register_dev()