TVM worlds switch takes a different path from regular VM world switch as
it needs to make an ecall to TSM and TSM actually does the world switch.
The host doesn't need to save/restore any context as TSM is expected
to do that on behalf of the host. The TSM updatess the trap
information in the shared memory which host uses to figure out the
cause of the guest exit.
Signed-off-by: Atish Patra <atishp@rivosinc.com>
---
arch/riscv/kvm/cove.c | 31 +++++++++++++++++++++++++++++--
arch/riscv/kvm/vcpu.c | 11 +++++++++++
arch/riscv/kvm/vcpu_exit.c | 10 ++++++++++
3 files changed, 50 insertions(+), 2 deletions(-)
@@ -275,9 +275,36 @@ int kvm_riscv_cove_gstage_map(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned long hv
return rc;
}
-void kvm_riscv_cove_vcpu_switchto(struct kvm_vcpu *vcpu, struct kvm_cpu_trap *trap)
+void noinstr kvm_riscv_cove_vcpu_switchto(struct kvm_vcpu *vcpu, struct kvm_cpu_trap *trap)
{
- /* TODO */
+ int rc;
+ struct kvm *kvm = vcpu->kvm;
+ struct kvm_cove_tvm_context *tvmc;
+ struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
+ void *nshmem;
+
+ if (!kvm->arch.tvmc)
+ return;
+
+ tvmc = kvm->arch.tvmc;
+
+ nshmem = nacl_shmem();
+ /* Invoke finalize to mark TVM is ready run for the first time */
+ if (unlikely(!tvmc->finalized_done)) {
+
+ rc = sbi_covh_tsm_finalize_tvm(tvmc->tvm_guest_id, cntx->sepc, cntx->a1);
+ if (rc) {
+ kvm_err("TVM Finalized failed with %d\n", rc);
+ return;
+ }
+ tvmc->finalized_done = true;
+ }
+
+ rc = sbi_covh_run_tvm_vcpu(tvmc->tvm_guest_id, vcpu->vcpu_idx);
+ if (rc) {
+ trap->scause = EXC_CUSTOM_KVM_COVE_RUN_FAIL;
+ return;
+ }
}
void kvm_riscv_cove_vcpu_destroy(struct kvm_vcpu *vcpu)
@@ -1042,6 +1042,11 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
kvm_riscv_vcpu_timer_save(vcpu);
if (kvm_riscv_nacl_available()) {
+ /**
+ * For TVMs, we don't need a separate case as TSM only updates
+ * the required CSRs during the world switch. All other CSR
+ * value should be zeroed out by TSM anyways.
+ */
nshmem = nacl_shmem();
csr->vsstatus = nacl_shmem_csr_read(nshmem, CSR_VSSTATUS);
csr->vsie = nacl_shmem_csr_read(nshmem, CSR_VSIE);
@@ -1191,6 +1196,12 @@ static void noinstr kvm_riscv_vcpu_enter_exit(struct kvm_vcpu *vcpu,
gcntx->hstatus = csr_swap(CSR_HSTATUS, hcntx->hstatus);
}
+ trap->htval = nacl_shmem_csr_read(nshmem, CSR_HTVAL);
+ trap->htinst = nacl_shmem_csr_read(nshmem, CSR_HTINST);
+ } else if (is_cove_vcpu(vcpu)) {
+ nshmem = nacl_shmem();
+ kvm_riscv_cove_vcpu_switchto(vcpu, trap);
+
trap->htval = nacl_shmem_csr_read(nshmem, CSR_HTVAL);
trap->htinst = nacl_shmem_csr_read(nshmem, CSR_HTINST);
} else {
@@ -218,6 +218,15 @@ int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
else if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV)
ret = kvm_riscv_vcpu_sbi_ecall(vcpu, run);
break;
+ case EXC_CUSTOM_KVM_COVE_RUN_FAIL:
+ if (likely(is_cove_vcpu(vcpu))) {
+ ret = -EACCES;
+ run->fail_entry.hardware_entry_failure_reason =
+ KVM_EXIT_FAIL_ENTRY_COVE_RUN_VCPU;
+ run->fail_entry.cpu = vcpu->cpu;
+ run->exit_reason = KVM_EXIT_FAIL_ENTRY;
+ }
+ break;
default:
break;
}
@@ -225,6 +234,7 @@ int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
/* Print details in-case of error */
if (ret < 0) {
kvm_err("VCPU exit error %d\n", ret);
+ //TODO: These values are bogus/stale for a TVM. Improve it
kvm_err("SEPC=0x%lx SSTATUS=0x%lx HSTATUS=0x%lx\n",
vcpu->arch.guest_context.sepc,
vcpu->arch.guest_context.sstatus,