Searched refs:csr_read (Results 1 – 17 of 17) sorted by relevance
13 context->scratch = csr_read(CSR_SCRATCH); in suspend_save_csrs()14 context->tvec = csr_read(CSR_TVEC); in suspend_save_csrs()15 context->ie = csr_read(CSR_IE); in suspend_save_csrs()28 context->satp = csr_read(CSR_SATP); in suspend_save_csrs()
35 cpu_mfr_info->vendor_id = csr_read(CSR_MVENDORID); in riscv_fill_cpu_mfr_info()36 cpu_mfr_info->arch_id = csr_read(CSR_MARCHID); in riscv_fill_cpu_mfr_info()37 cpu_mfr_info->imp_id = csr_read(CSR_MIMPID); in riscv_fill_cpu_mfr_info()
113 ci->mvendorid = csr_read(CSR_MVENDORID); in riscv_cpuinfo_starting()114 ci->marchid = csr_read(CSR_MARCHID); in riscv_cpuinfo_starting()115 ci->mimpid = csr_read(CSR_MIMPID); in riscv_cpuinfo_starting()
98 unsigned long tmp = csr_read(CSR_STATUS); in compat_mode_detect()102 (csr_read(CSR_STATUS) & SR_UXL) == SR_UXL_32; in compat_mode_detect()
243 unsigned long addr = csr_read(mtval); in handle_misaligned_load()326 unsigned long addr = csr_read(mtval); in handle_misaligned_store()
747 csr->vsie = csr_read(CSR_VSIE); in kvm_riscv_vcpu_sync_interrupts()750 hvip = csr_read(CSR_HVIP); in kvm_riscv_vcpu_sync_interrupts()914 csr->vsie = csr_read(CSR_VSIE); in kvm_arch_vcpu_put()915 csr->vstvec = csr_read(CSR_VSTVEC); in kvm_arch_vcpu_put()917 csr->vsepc = csr_read(CSR_VSEPC); in kvm_arch_vcpu_put()919 csr->vstval = csr_read(CSR_VSTVAL); in kvm_arch_vcpu_put()920 csr->hvip = csr_read(CSR_HVIP); in kvm_arch_vcpu_put()921 csr->vsatp = csr_read(CSR_VSATP); in kvm_arch_vcpu_put()1099 trap.scause = csr_read(CSR_SCAUSE); in kvm_arch_vcpu_ioctl_run()1100 trap.stval = csr_read(CSR_STVAL); in kvm_arch_vcpu_ioctl_run()[all …]
28 old = csr_read(CSR_HGATP); in kvm_riscv_gstage_vmid_detect()30 vmid_bits = csr_read(CSR_HGATP); in kvm_riscv_gstage_vmid_detect()
138 unsigned long vsstatus = csr_read(CSR_VSSTATUS); in kvm_riscv_vcpu_trap_redirect()162 vcpu->arch.guest_context.sepc = csr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect()
331 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()332 t->next_cycles |= (u64)csr_read(CSR_VSTIMECMPH) << 32; in kvm_riscv_vcpu_timer_sync()334 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()
766 if ((csr_read(CSR_HGATP) >> HGATP_MODE_SHIFT) == HGATP_MODE_SV57X4) { in kvm_riscv_gstage_mode_detect()774 if ((csr_read(CSR_HGATP) >> HGATP_MODE_SHIFT) == HGATP_MODE_SV48X4) { in kvm_riscv_gstage_mode_detect()
53 return csr_read(CSR_TIME); in get_cycles()59 return csr_read(CSR_TIMEH); in get_cycles_hi()
16 return csr_read(CSR_STATUS); in arch_local_save_flags()
355 #define csr_read(csr) \ macro
231 old = csr_read(CSR_SATP); in asids_init()234 asid_bits = (csr_read(CSR_SATP) >> SATP_ASID_SHIFT) & SATP_ASID_MASK; in asids_init()
125 pfn = csr_read(CSR_SATP) & SATP_PPN; in vmalloc_fault()
71 return csr_read(CSR_TIME); in __arch_get_hw_counter()
24 __val = csr_read(__csr_num); \ in csr_read_num()
Completed in 21 milliseconds