| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
| A D | base.c | 29 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_fini() local 30 fault->func->buffer.intr(fault->buffer[index], false); in nvkm_fault_ntfy_fini() 36 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_init() local 37 fault->func->buffer.intr(fault->buffer[index], true); in nvkm_fault_ntfy_init() 50 return fault->func->intr(fault); in nvkm_fault_intr() 58 fault->func->fini(fault); in nvkm_fault_fini() 67 fault->func->init(fault); in nvkm_fault_init() 81 buffer->fault = fault; in nvkm_fault_oneinit_buffer() 122 ret = fault->func->oneinit(fault); in nvkm_fault_oneinit() 142 return fault; in nvkm_fault_dtor() [all …]
|
| A D | gv100.c | 33 struct nvkm_fault *fault = container_of(work, typeof(*fault), nrpfb_work); in gv100_fault_buffer_process() local 122 struct nvkm_fault *fault = container_of(ntfy, typeof(*fault), nrpfb); in gv100_fault_ntfy_nrpfb() local 124 schedule_work(&fault->nrpfb_work); in gv100_fault_ntfy_nrpfb() 162 gv100_fault_intr_fault(fault); in gv100_fault_intr() 168 if (fault->buffer[0]) { in gv100_fault_intr() 175 if (fault->buffer[1]) { in gv100_fault_intr() 189 nvkm_event_ntfy_block(&fault->nrpfb); in gv100_fault_fini() 190 flush_work(&fault->nrpfb_work); in gv100_fault_fini() 192 if (fault->buffer[0]) in gv100_fault_fini() 193 fault->func->buffer.fini(fault->buffer[0]); in gv100_fault_fini() [all …]
|
| A D | tu102.c | 88 struct nvkm_fault *fault = container_of(inth, typeof(*fault), info_fault); in tu102_fault_info_fault() local 118 nvkm_event_ntfy_block(&fault->nrpfb); in tu102_fault_fini() 119 flush_work(&fault->nrpfb_work); in tu102_fault_fini() 121 if (fault->buffer[0]) in tu102_fault_fini() 122 fault->func->buffer.fini(fault->buffer[0]); in tu102_fault_fini() 124 nvkm_inth_block(&fault->info_fault); in tu102_fault_fini() 130 nvkm_inth_allow(&fault->info_fault); in tu102_fault_init() 132 fault->func->buffer.init(fault->buffer[0]); in tu102_fault_init() 133 nvkm_event_ntfy_allow(&fault->nrpfb); in tu102_fault_init() 145 &fault->info_fault); in tu102_fault_oneinit() [all …]
|
| A D | Kbuild | 2 nvkm-y += nvkm/subdev/fault/base.o 3 nvkm-y += nvkm/subdev/fault/user.o 4 nvkm-y += nvkm/subdev/fault/gp100.o 5 nvkm-y += nvkm/subdev/fault/gp10b.o 6 nvkm-y += nvkm/subdev/fault/gv100.o 7 nvkm-y += nvkm/subdev/fault/tu102.o
|
| A D | user.c | 42 return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id, in nvkm_ufault_uevent() 51 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map() 62 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini() 70 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init() 96 struct nvkm_fault *fault = device->fault; in nvkm_ufault_new() local 97 struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; in nvkm_ufault_new()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| A D | fault.c | 69 if (!ibd->fault->n_rxfaults[i] && !ibd->fault->n_txfaults[i]) in _fault_stats_seq_show() 96 struct fault *fault = file->private_data; in fault_opcodes_write() local 138 bitmap_zero(fault->opcodes, sizeof(fault->opcodes) * in fault_opcodes_write() 170 struct fault *fault = file->private_data; in fault_opcodes_read() local 210 if (ibd->fault) in hfi1_fault_exit_debugfs() 221 ibd->fault = kzalloc(sizeof(*ibd->fault), GFP_KERNEL); in hfi1_fault_init_debugfs() 222 if (!ibd->fault) in hfi1_fault_init_debugfs() 268 if (ibd->fault) in hfi1_dbg_fault_suppress_err() 278 if (!ibd->fault || !ibd->fault->enable) in __hfi1_should_fault() 293 if (ibd->fault->fault_skip && ibd->fault->skip) { in __hfi1_should_fault() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| A D | fault.c | 82 if (!hwpt->fault) in iommufd_fault_domain_attach_dev() 99 struct iommufd_fault *fault = hwpt->fault; in iommufd_auto_response_faults() local 103 if (!fault) in iommufd_auto_response_faults() 156 if (old->fault) in __fault_domain_replace_dev() 159 if (hwpt->fault) { in __fault_domain_replace_dev() 184 bool iopf_off = !hwpt->fault && old->fault; in iommufd_fault_domain_replace_dev() 185 bool iopf_on = hwpt->fault && !old->fault; in iommufd_fault_domain_replace_dev() 376 fault = iommufd_object_alloc(ucmd->ictx, fault, IOMMUFD_OBJ_FAULT); in iommufd_fault_alloc() 377 if (IS_ERR(fault)) in iommufd_fault_alloc() 387 fault, O_RDWR); in iommufd_fault_alloc() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| A D | mmu_internal.h | 286 struct kvm_page_fault *fault) in kvm_mmu_prepare_memory_fault_exit() argument 289 PAGE_SIZE, fault->write, fault->exec, in kvm_mmu_prepare_memory_fault_exit() 290 fault->is_private); in kvm_mmu_prepare_memory_fault_exit() 297 struct kvm_page_fault fault = { in kvm_mmu_do_page_fault() local 321 fault.gfn = fault.addr >> PAGE_SHIFT; in kvm_mmu_do_page_fault() 322 fault.slot = kvm_vcpu_gfn_to_memslot(vcpu, fault.gfn); in kvm_mmu_do_page_fault() 326 r = kvm_tdp_page_fault(vcpu, &fault); in kvm_mmu_do_page_fault() 328 r = vcpu->arch.mmu->page_fault(vcpu, &fault); in kvm_mmu_do_page_fault() 335 if (r == RET_PF_EMULATE && fault.is_private) { in kvm_mmu_do_page_fault() 337 kvm_mmu_prepare_memory_fault_exit(vcpu, &fault); in kvm_mmu_do_page_fault() [all …]
|
| A D | paging_tmpl.h | 92 struct x86_exception fault; member 517 walker->fault.address = addr; in FNAME() 634 gfn_t base_gfn = fault->gfn; in FNAME() 713 if (fault->write && table_gfn == fault->gfn) in FNAME() 756 base_gfn, fault->pfn, fault); in FNAME() 783 WARN_ON_ONCE(fault->is_tdp); in FNAME() 797 if (!fault->prefetch) in FNAME() 803 fault->gfn = walker.gfn; in FNAME() 805 fault->slot = kvm_vcpu_gfn_to_memslot(vcpu, fault->gfn); in FNAME() 825 !is_cr0_wp(vcpu->arch.mmu) && !fault->user && fault->slot) { in FNAME() [all …]
|
| /linux/arch/mips/kernel/ |
| A D | unaligned.c | 175 goto fault; in emulate_load_store_insn() 184 goto fault; in emulate_load_store_insn() 299 goto fault; in emulate_load_store_insn() 314 goto fault; in emulate_load_store_insn() 329 goto fault; in emulate_load_store_insn() 348 goto fault; in emulate_load_store_insn() 371 goto fault; in emulate_load_store_insn() 584 fault: in emulate_load_store_insn() 656 goto fault; in emulate_load_store_microMIPS() 1211 fault: in emulate_load_store_microMIPS() [all …]
|
| /linux/arch/nios2/kernel/ |
| A D | misaligned.c | 72 unsigned int fault; in handle_unaligned_c() local 85 fault = 0; in handle_unaligned_c() 98 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 99 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 116 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 117 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 140 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 141 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 142 fault |= __get_user(d2, (u8 *)(addr+2)); in handle_unaligned_c() 143 fault |= __get_user(d3, (u8 *)(addr+3)); in handle_unaligned_c() [all …]
|
| /linux/arch/x86/kvm/svm/ |
| A D | svm_ops.h | 12 _ASM_EXTABLE(1b, %l[fault]) \ 13 ::: clobber : fault); \ 15 fault: \ 22 _ASM_EXTABLE(1b, %l[fault]) \ 23 :: op1 : clobber : fault); \ 25 fault: \ 32 _ASM_EXTABLE(1b, %l[fault]) \ 33 :: op1, op2 : clobber : fault); \ 35 fault: \
|
| /linux/drivers/gpu/drm/nouveau/ |
| A D | nouveau_svm.c | 66 u8 fault; member 68 } **fault; member 415 fault->hub, in nouveau_svm_fault_cancel_fault() 416 fault->gpc, in nouveau_svm_fault_cancel_fault() 423 switch (fault) { in nouveau_svm_fault_priority() 478 fault = kmalloc(sizeof(*fault), GFP_KERNEL); in nouveau_svm_fault_cache() 483 buffer->fault[buffer->fault_nr] = fault; in nouveau_svm_fault_cache() 486 fault = buffer->fault[buffer->fault_nr++]; in nouveau_svm_fault_cache() 495 fault->fault = (info & 0x0000001f); in nouveau_svm_fault_cache() 498 fault->inst, fault->addr, fault->access); in nouveau_svm_fault_cache() [all …]
|
| /linux/drivers/iommu/ |
| A D | io-pgfault.c | 64 struct iommu_fault *fault) in report_partial_fault() argument 72 iopf->fault = *fault; in report_partial_fault() 98 group->last_fault.fault = evt->fault; in iopf_group_alloc() 106 if (iopf->fault.prm.grpid == evt->fault.prm.grpid) in iopf_group_alloc() 121 struct iommu_fault *fault = &evt->fault; in find_fault_handler() local 126 fault->prm.pasid, 0); in find_fault_handler() 161 struct iommu_fault *fault = &evt->fault; in iopf_error_response() local 163 .pasid = fault->prm.pasid, in iopf_error_response() 164 .grpid = fault->prm.grpid, in iopf_error_response() 217 struct iommu_fault *fault = &evt->fault; in iommu_report_device_fault() local [all …]
|
| /linux/arch/mips/loongson64/ |
| A D | cop2-ex.c | 76 goto fault; in loongson_cu2_call() 80 goto fault; in loongson_cu2_call() 92 goto fault; in loongson_cu2_call() 96 goto fault; in loongson_cu2_call() 118 goto fault; in loongson_cu2_call() 123 goto fault; in loongson_cu2_call() 135 goto fault; in loongson_cu2_call() 141 goto fault; in loongson_cu2_call() 165 goto fault; in loongson_cu2_call() 176 goto fault; in loongson_cu2_call() [all …]
|
| /linux/arch/arc/kernel/ |
| A D | unaligned.c | 51 goto fault; \ 66 goto fault; \ 93 goto fault; \ 126 goto fault; \ 160 fault: state->fault = 1; in fixup_load() 180 goto fault; in fixup_store() 192 fault: state->fault = 1; in fixup_store() 227 goto fault; in misaligned_fixup() 231 goto fault; in misaligned_fixup() 239 goto fault; in misaligned_fixup() [all …]
|
| /linux/arch/powerpc/mm/ |
| A D | fault.c | 140 vm_fault_t fault) in do_sigbus() argument 153 if (fault & VM_FAULT_HWPOISON_LARGE) in do_sigbus() 155 if (fault & VM_FAULT_HWPOISON) in do_sigbus() 168 vm_fault_t fault) in mm_fault_error() argument 178 if (fault & VM_FAULT_OOM) { in mm_fault_error() 190 else if (fault & VM_FAULT_SIGSEGV) in mm_fault_error() 422 vm_fault_t fault, major = 0; in ___do_page_fault() local 502 if (!(fault & VM_FAULT_RETRY)) { in ___do_page_fault() 507 if (fault & VM_FAULT_MAJOR) in ___do_page_fault() 543 major |= fault & VM_FAULT_MAJOR; in ___do_page_fault() [all …]
|
| /linux/arch/riscv/mm/ |
| A D | fault.c | 69 if (fault & VM_FAULT_OOM) { in mm_fault_error() 80 } else if (fault & VM_FAULT_SIGSEGV) { in mm_fault_error() 231 vm_fault_t fault; in handle_page_fault() local 306 if (!(fault & VM_FAULT_RETRY)) { in handle_page_fault() 311 if (fault & VM_FAULT_MAJOR) in handle_page_fault() 314 if (fault_signal_pending(fault, regs)) { in handle_page_fault() 353 if (fault_signal_pending(fault, regs)) { in handle_page_fault() 360 if (fault & VM_FAULT_COMPLETED) in handle_page_fault() 363 if (unlikely(fault & VM_FAULT_RETRY)) { in handle_page_fault() 377 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault() [all …]
|
| /linux/arch/powerpc/lib/ |
| A D | checksum_32.S | 240 fault: label 244 EX_TABLE(70b, fault); 245 EX_TABLE(71b, fault); 246 EX_TABLE(72b, fault); 247 EX_TABLE(73b, fault); 248 EX_TABLE(54b, fault); 269 EX_TABLE(30b, fault); 270 EX_TABLE(31b, fault); 271 EX_TABLE(40b, fault); 272 EX_TABLE(41b, fault); [all …]
|
| /linux/arch/m68k/mm/ |
| A D | fault.c | 75 vm_fault_t fault; in do_page_fault() local 141 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 142 pr_debug("handle_mm_fault returns %x\n", fault); in do_page_fault() 144 if (fault_signal_pending(fault, regs)) { in do_page_fault() 151 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 154 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 155 if (fault & VM_FAULT_OOM) in do_page_fault() 157 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 159 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 164 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
| /linux/arch/hexagon/mm/ |
| A D | vm_fault.c | 43 vm_fault_t fault; in do_page_fault() local 84 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 86 if (fault_signal_pending(fault, regs)) { in do_page_fault() 93 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 97 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault() 98 if (fault & VM_FAULT_RETRY) { in do_page_fault() 113 if (fault & VM_FAULT_OOM) { in do_page_fault() 121 if (fault & VM_FAULT_SIGBUS) { in do_page_fault()
|
| /linux/arch/alpha/mm/ |
| A D | fault.c | 92 vm_fault_t fault; in do_page_fault() local 145 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 147 if (fault_signal_pending(fault, regs)) { in do_page_fault() 154 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 157 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 158 if (fault & VM_FAULT_OOM) in do_page_fault() 160 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 162 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 167 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
| /linux/arch/microblaze/mm/ |
| A D | fault.c | 93 vm_fault_t fault; in do_page_fault() local 221 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 223 if (fault_signal_pending(fault, regs)) { in do_page_fault() 230 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 233 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 234 if (fault & VM_FAULT_OOM) in do_page_fault() 236 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 238 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 243 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
| /linux/arch/nios2/mm/ |
| A D | fault.c | 50 vm_fault_t fault; in do_page_fault() local 124 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 126 if (fault_signal_pending(fault, regs)) { in do_page_fault() 133 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 136 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 137 if (fault & VM_FAULT_OOM) in do_page_fault() 139 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 141 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 146 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
| /linux/arch/parisc/mm/ |
| A D | fault.c | 274 vm_fault_t fault = 0; in do_page_fault() local 318 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault() 320 if (fault_signal_pending(fault, regs)) { in do_page_fault() 329 if (fault & VM_FAULT_COMPLETED) in do_page_fault() 332 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 338 if (fault & VM_FAULT_OOM) in do_page_fault() 340 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 342 else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| in do_page_fault() 347 if (fault & VM_FAULT_RETRY) { in do_page_fault() 411 if (fault & VM_FAULT_HWPOISON_LARGE) in do_page_fault() [all …]
|