GNU Linux-libre 6.7.9-gnu
[releases.git] / arch / riscv / kvm / vcpu_sbi_replace.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2021 Western Digital Corporation or its affiliates.
4  *
5  * Authors:
6  *     Atish Patra <atish.patra@wdc.com>
7  */
8
9 #include <linux/errno.h>
10 #include <linux/err.h>
11 #include <linux/kvm_host.h>
12 #include <asm/sbi.h>
13 #include <asm/kvm_vcpu_timer.h>
14 #include <asm/kvm_vcpu_pmu.h>
15 #include <asm/kvm_vcpu_sbi.h>
16
17 static int kvm_sbi_ext_time_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
18                                     struct kvm_vcpu_sbi_return *retdata)
19 {
20         struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
21         u64 next_cycle;
22
23         if (cp->a6 != SBI_EXT_TIME_SET_TIMER) {
24                 retdata->err_val = SBI_ERR_INVALID_PARAM;
25                 return 0;
26         }
27
28         kvm_riscv_vcpu_pmu_incr_fw(vcpu, SBI_PMU_FW_SET_TIMER);
29 #if __riscv_xlen == 32
30         next_cycle = ((u64)cp->a1 << 32) | (u64)cp->a0;
31 #else
32         next_cycle = (u64)cp->a0;
33 #endif
34         kvm_riscv_vcpu_timer_next_event(vcpu, next_cycle);
35
36         return 0;
37 }
38
39 const struct kvm_vcpu_sbi_extension vcpu_sbi_ext_time = {
40         .extid_start = SBI_EXT_TIME,
41         .extid_end = SBI_EXT_TIME,
42         .handler = kvm_sbi_ext_time_handler,
43 };
44
45 static int kvm_sbi_ext_ipi_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
46                                    struct kvm_vcpu_sbi_return *retdata)
47 {
48         int ret = 0;
49         unsigned long i;
50         struct kvm_vcpu *tmp;
51         struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
52         unsigned long hmask = cp->a0;
53         unsigned long hbase = cp->a1;
54
55         if (cp->a6 != SBI_EXT_IPI_SEND_IPI) {
56                 retdata->err_val = SBI_ERR_INVALID_PARAM;
57                 return 0;
58         }
59
60         kvm_riscv_vcpu_pmu_incr_fw(vcpu, SBI_PMU_FW_IPI_SENT);
61         kvm_for_each_vcpu(i, tmp, vcpu->kvm) {
62                 if (hbase != -1UL) {
63                         if (tmp->vcpu_id < hbase)
64                                 continue;
65                         if (!(hmask & (1UL << (tmp->vcpu_id - hbase))))
66                                 continue;
67                 }
68                 ret = kvm_riscv_vcpu_set_interrupt(tmp, IRQ_VS_SOFT);
69                 if (ret < 0)
70                         break;
71                 kvm_riscv_vcpu_pmu_incr_fw(tmp, SBI_PMU_FW_IPI_RCVD);
72         }
73
74         return ret;
75 }
76
77 const struct kvm_vcpu_sbi_extension vcpu_sbi_ext_ipi = {
78         .extid_start = SBI_EXT_IPI,
79         .extid_end = SBI_EXT_IPI,
80         .handler = kvm_sbi_ext_ipi_handler,
81 };
82
83 static int kvm_sbi_ext_rfence_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
84                                       struct kvm_vcpu_sbi_return *retdata)
85 {
86         struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
87         unsigned long hmask = cp->a0;
88         unsigned long hbase = cp->a1;
89         unsigned long funcid = cp->a6;
90
91         switch (funcid) {
92         case SBI_EXT_RFENCE_REMOTE_FENCE_I:
93                 kvm_riscv_fence_i(vcpu->kvm, hbase, hmask);
94                 kvm_riscv_vcpu_pmu_incr_fw(vcpu, SBI_PMU_FW_FENCE_I_SENT);
95                 break;
96         case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA:
97                 if (cp->a2 == 0 && cp->a3 == 0)
98                         kvm_riscv_hfence_vvma_all(vcpu->kvm, hbase, hmask);
99                 else
100                         kvm_riscv_hfence_vvma_gva(vcpu->kvm, hbase, hmask,
101                                                   cp->a2, cp->a3, PAGE_SHIFT);
102                 kvm_riscv_vcpu_pmu_incr_fw(vcpu, SBI_PMU_FW_HFENCE_VVMA_SENT);
103                 break;
104         case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA_ASID:
105                 if (cp->a2 == 0 && cp->a3 == 0)
106                         kvm_riscv_hfence_vvma_asid_all(vcpu->kvm,
107                                                        hbase, hmask, cp->a4);
108                 else
109                         kvm_riscv_hfence_vvma_asid_gva(vcpu->kvm,
110                                                        hbase, hmask,
111                                                        cp->a2, cp->a3,
112                                                        PAGE_SHIFT, cp->a4);
113                 kvm_riscv_vcpu_pmu_incr_fw(vcpu, SBI_PMU_FW_HFENCE_VVMA_ASID_SENT);
114                 break;
115         case SBI_EXT_RFENCE_REMOTE_HFENCE_GVMA:
116         case SBI_EXT_RFENCE_REMOTE_HFENCE_GVMA_VMID:
117         case SBI_EXT_RFENCE_REMOTE_HFENCE_VVMA:
118         case SBI_EXT_RFENCE_REMOTE_HFENCE_VVMA_ASID:
119                 /*
120                  * Until nested virtualization is implemented, the
121                  * SBI HFENCE calls should be treated as NOPs
122                  */
123                 break;
124         default:
125                 retdata->err_val = SBI_ERR_NOT_SUPPORTED;
126         }
127
128         return 0;
129 }
130
131 const struct kvm_vcpu_sbi_extension vcpu_sbi_ext_rfence = {
132         .extid_start = SBI_EXT_RFENCE,
133         .extid_end = SBI_EXT_RFENCE,
134         .handler = kvm_sbi_ext_rfence_handler,
135 };
136
137 static int kvm_sbi_ext_srst_handler(struct kvm_vcpu *vcpu,
138                                     struct kvm_run *run,
139                                     struct kvm_vcpu_sbi_return *retdata)
140 {
141         struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
142         unsigned long funcid = cp->a6;
143         u32 reason = cp->a1;
144         u32 type = cp->a0;
145
146         switch (funcid) {
147         case SBI_EXT_SRST_RESET:
148                 switch (type) {
149                 case SBI_SRST_RESET_TYPE_SHUTDOWN:
150                         kvm_riscv_vcpu_sbi_system_reset(vcpu, run,
151                                                 KVM_SYSTEM_EVENT_SHUTDOWN,
152                                                 reason);
153                         retdata->uexit = true;
154                         break;
155                 case SBI_SRST_RESET_TYPE_COLD_REBOOT:
156                 case SBI_SRST_RESET_TYPE_WARM_REBOOT:
157                         kvm_riscv_vcpu_sbi_system_reset(vcpu, run,
158                                                 KVM_SYSTEM_EVENT_RESET,
159                                                 reason);
160                         retdata->uexit = true;
161                         break;
162                 default:
163                         retdata->err_val = SBI_ERR_NOT_SUPPORTED;
164                 }
165                 break;
166         default:
167                 retdata->err_val = SBI_ERR_NOT_SUPPORTED;
168         }
169
170         return 0;
171 }
172
173 const struct kvm_vcpu_sbi_extension vcpu_sbi_ext_srst = {
174         .extid_start = SBI_EXT_SRST,
175         .extid_end = SBI_EXT_SRST,
176         .handler = kvm_sbi_ext_srst_handler,
177 };
178
179 static int kvm_sbi_ext_dbcn_handler(struct kvm_vcpu *vcpu,
180                                     struct kvm_run *run,
181                                     struct kvm_vcpu_sbi_return *retdata)
182 {
183         struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
184         unsigned long funcid = cp->a6;
185
186         switch (funcid) {
187         case SBI_EXT_DBCN_CONSOLE_WRITE:
188         case SBI_EXT_DBCN_CONSOLE_READ:
189         case SBI_EXT_DBCN_CONSOLE_WRITE_BYTE:
190                 /*
191                  * The SBI debug console functions are unconditionally
192                  * forwarded to the userspace.
193                  */
194                 kvm_riscv_vcpu_sbi_forward(vcpu, run);
195                 retdata->uexit = true;
196                 break;
197         default:
198                 retdata->err_val = SBI_ERR_NOT_SUPPORTED;
199         }
200
201         return 0;
202 }
203
204 const struct kvm_vcpu_sbi_extension vcpu_sbi_ext_dbcn = {
205         .extid_start = SBI_EXT_DBCN,
206         .extid_end = SBI_EXT_DBCN,
207         .default_unavail = true,
208         .handler = kvm_sbi_ext_dbcn_handler,
209 };