Lines Matching refs:gfn

293 	kvm_pfn_t gfn;  member
1144 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
1147 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
1148 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
1149 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
1150 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
1151 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
1156 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
1157 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
1159 kvm_pfn_t gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn);
1160 kvm_pfn_t gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn);
1161 kvm_pfn_t __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn,
1171 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
1179 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
1191 #define __kvm_get_guest(kvm, gfn, offset, v) \ argument
1193 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1211 #define __kvm_put_guest(kvm, gfn, offset, v) \ argument
1213 unsigned long __addr = gfn_to_hva(kvm, gfn); \
1220 mark_page_dirty(kvm, gfn); \
1234 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
1235 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
1236 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1237 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
1238 void mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn);
1239 void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
1242 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
1243 kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
1244 kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
1247 unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
1248 unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
1249 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
1255 int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
1259 void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
1613 try_get_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in try_get_memslot() argument
1618 if (gfn >= slot->base_gfn && gfn < slot->base_gfn + slot->npages) in try_get_memslot()
1632 search_memslots(struct kvm_memslots *slots, gfn_t gfn, bool approx) in search_memslots() argument
1641 if (gfn >= slot->base_gfn) { in search_memslots()
1642 if (gfn < slot->base_gfn + slot->npages) in search_memslots()
1653 ____gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn, bool approx) in ____gfn_to_memslot() argument
1658 slot = try_get_memslot(slot, gfn); in ____gfn_to_memslot()
1662 slot = search_memslots(slots, gfn, approx); in ____gfn_to_memslot()
1677 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) in __gfn_to_memslot() argument
1679 return ____gfn_to_memslot(slots, gfn, false); in __gfn_to_memslot()
1683 __gfn_to_hva_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
1691 unsigned long offset = gfn - slot->base_gfn; in __gfn_to_hva_memslot()
1696 static inline int memslot_id(struct kvm *kvm, gfn_t gfn) in memslot_id() argument
1698 return gfn_to_memslot(kvm, gfn)->id; in memslot_id()
1709 static inline gpa_t gfn_to_gpa(gfn_t gfn) in gfn_to_gpa() argument
1711 return (gpa_t)gfn << PAGE_SHIFT; in gfn_to_gpa()