1 /******************************************************************************
4 * Linux-specific hypervisor handling.
6 * Copyright (c) 2002-2004, K A Fraser
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License version 2
10 * as published by the Free Software Foundation; or, when distributed
11 * separately from the Linux kernel or incorporated into other
12 * software packages, subject to the following license:
14 * Permission is hereby granted, free of charge, to any person obtaining a copy
15 * of this source file (the "Software"), to deal in the Software without
16 * restriction, including without limitation the rights to use, copy, modify,
17 * merge, publish, distribute, sublicense, and/or sell copies of the Software,
18 * and to permit persons to whom the Software is furnished to do so, subject to
19 * the following conditions:
21 * The above copyright notice and this permission notice shall be included in
22 * all copies or substantial portions of the Software.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
25 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
26 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
27 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
28 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
29 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
33 #ifndef _ASM_X86_XEN_HYPERCALL_H
34 #define _ASM_X86_XEN_HYPERCALL_H
36 #include <linux/kernel.h>
37 #include <linux/spinlock.h>
38 #include <linux/errno.h>
39 #include <linux/string.h>
40 #include <linux/types.h>
42 #include <trace/events/xen.h>
45 #include <asm/pgtable.h>
47 #include <asm/nospec-branch.h>
49 #include <xen/interface/xen.h>
50 #include <xen/interface/sched.h>
51 #include <xen/interface/physdev.h>
52 #include <xen/interface/platform.h>
53 #include <xen/interface/xen-mca.h>
56 * The hypercall asms have to meet several constraints:
57 * - Work on 32- and 64-bit.
58 * The two architectures put their arguments in different sets of
61 * - Work around asm syntax quirks
62 * It isn't possible to specify one of the rNN registers in a
63 * constraint, so we use explicit register variables to get the
64 * args into the right place.
66 * - Mark all registers as potentially clobbered
67 * Even unused parameters can be clobbered by the hypervisor, so we
68 * need to make sure gcc knows it.
70 * - Avoid compiler bugs.
71 * This is the tricky part. Because x86_32 has such a constrained
72 * register set, gcc versions below 4.3 have trouble generating
73 * code when all the arg registers and memory are trashed by the
74 * asm. There are syntactically simpler ways of achieving the
75 * semantics below, but they cause the compiler to crash.
77 * The only combination I found which works is:
78 * - assign the __argX variables first
79 * - list all actually used parameters as "+r" (__argX)
82 * The result certainly isn't pretty, and it really shows up cpp's
83 * weakness as as macro language. Sorry. (But let's just give thanks
84 * there aren't more than 5 arguments...)
87 extern struct { char _entry[32]; } hypercall_page[];
89 #define __HYPERCALL "call hypercall_page+%c[offset]"
90 #define __HYPERCALL_ENTRY(x) \
91 [offset] "i" (__HYPERVISOR_##x * sizeof(hypercall_page[0]))
94 #define __HYPERCALL_RETREG "eax"
95 #define __HYPERCALL_ARG1REG "ebx"
96 #define __HYPERCALL_ARG2REG "ecx"
97 #define __HYPERCALL_ARG3REG "edx"
98 #define __HYPERCALL_ARG4REG "esi"
99 #define __HYPERCALL_ARG5REG "edi"
101 #define __HYPERCALL_RETREG "rax"
102 #define __HYPERCALL_ARG1REG "rdi"
103 #define __HYPERCALL_ARG2REG "rsi"
104 #define __HYPERCALL_ARG3REG "rdx"
105 #define __HYPERCALL_ARG4REG "r10"
106 #define __HYPERCALL_ARG5REG "r8"
109 #define __HYPERCALL_DECLS \
110 register unsigned long __res asm(__HYPERCALL_RETREG); \
111 register unsigned long __arg1 asm(__HYPERCALL_ARG1REG) = __arg1; \
112 register unsigned long __arg2 asm(__HYPERCALL_ARG2REG) = __arg2; \
113 register unsigned long __arg3 asm(__HYPERCALL_ARG3REG) = __arg3; \
114 register unsigned long __arg4 asm(__HYPERCALL_ARG4REG) = __arg4; \
115 register unsigned long __arg5 asm(__HYPERCALL_ARG5REG) = __arg5; \
116 register void *__sp asm(_ASM_SP);
118 #define __HYPERCALL_0PARAM "=r" (__res), "+r" (__sp)
119 #define __HYPERCALL_1PARAM __HYPERCALL_0PARAM, "+r" (__arg1)
120 #define __HYPERCALL_2PARAM __HYPERCALL_1PARAM, "+r" (__arg2)
121 #define __HYPERCALL_3PARAM __HYPERCALL_2PARAM, "+r" (__arg3)
122 #define __HYPERCALL_4PARAM __HYPERCALL_3PARAM, "+r" (__arg4)
123 #define __HYPERCALL_5PARAM __HYPERCALL_4PARAM, "+r" (__arg5)
125 #define __HYPERCALL_0ARG()
126 #define __HYPERCALL_1ARG(a1) \
127 __HYPERCALL_0ARG() __arg1 = (unsigned long)(a1);
128 #define __HYPERCALL_2ARG(a1,a2) \
129 __HYPERCALL_1ARG(a1) __arg2 = (unsigned long)(a2);
130 #define __HYPERCALL_3ARG(a1,a2,a3) \
131 __HYPERCALL_2ARG(a1,a2) __arg3 = (unsigned long)(a3);
132 #define __HYPERCALL_4ARG(a1,a2,a3,a4) \
133 __HYPERCALL_3ARG(a1,a2,a3) __arg4 = (unsigned long)(a4);
134 #define __HYPERCALL_5ARG(a1,a2,a3,a4,a5) \
135 __HYPERCALL_4ARG(a1,a2,a3,a4) __arg5 = (unsigned long)(a5);
137 #define __HYPERCALL_CLOBBER5 "memory"
138 #define __HYPERCALL_CLOBBER4 __HYPERCALL_CLOBBER5, __HYPERCALL_ARG5REG
139 #define __HYPERCALL_CLOBBER3 __HYPERCALL_CLOBBER4, __HYPERCALL_ARG4REG
140 #define __HYPERCALL_CLOBBER2 __HYPERCALL_CLOBBER3, __HYPERCALL_ARG3REG
141 #define __HYPERCALL_CLOBBER1 __HYPERCALL_CLOBBER2, __HYPERCALL_ARG2REG
142 #define __HYPERCALL_CLOBBER0 __HYPERCALL_CLOBBER1, __HYPERCALL_ARG1REG
144 #define _hypercall0(type, name) \
147 __HYPERCALL_0ARG(); \
148 asm volatile (__HYPERCALL \
149 : __HYPERCALL_0PARAM \
150 : __HYPERCALL_ENTRY(name) \
151 : __HYPERCALL_CLOBBER0); \
155 #define _hypercall1(type, name, a1) \
158 __HYPERCALL_1ARG(a1); \
159 asm volatile (__HYPERCALL \
160 : __HYPERCALL_1PARAM \
161 : __HYPERCALL_ENTRY(name) \
162 : __HYPERCALL_CLOBBER1); \
166 #define _hypercall2(type, name, a1, a2) \
169 __HYPERCALL_2ARG(a1, a2); \
170 asm volatile (__HYPERCALL \
171 : __HYPERCALL_2PARAM \
172 : __HYPERCALL_ENTRY(name) \
173 : __HYPERCALL_CLOBBER2); \
177 #define _hypercall3(type, name, a1, a2, a3) \
180 __HYPERCALL_3ARG(a1, a2, a3); \
181 asm volatile (__HYPERCALL \
182 : __HYPERCALL_3PARAM \
183 : __HYPERCALL_ENTRY(name) \
184 : __HYPERCALL_CLOBBER3); \
188 #define _hypercall4(type, name, a1, a2, a3, a4) \
191 __HYPERCALL_4ARG(a1, a2, a3, a4); \
192 asm volatile (__HYPERCALL \
193 : __HYPERCALL_4PARAM \
194 : __HYPERCALL_ENTRY(name) \
195 : __HYPERCALL_CLOBBER4); \
199 #define _hypercall5(type, name, a1, a2, a3, a4, a5) \
202 __HYPERCALL_5ARG(a1, a2, a3, a4, a5); \
203 asm volatile (__HYPERCALL \
204 : __HYPERCALL_5PARAM \
205 : __HYPERCALL_ENTRY(name) \
206 : __HYPERCALL_CLOBBER5); \
211 privcmd_call(unsigned call,
212 unsigned long a1, unsigned long a2,
213 unsigned long a3, unsigned long a4,
217 __HYPERCALL_5ARG(a1, a2, a3, a4, a5);
219 if (call >= PAGE_SIZE / sizeof(hypercall_page[0]))
223 asm volatile(CALL_NOSPEC
225 : [thunk_target] "a" (&hypercall_page[call])
226 : __HYPERCALL_CLOBBER5);
233 HYPERVISOR_set_trap_table(struct trap_info *table)
235 return _hypercall1(int, set_trap_table, table);
239 HYPERVISOR_mmu_update(struct mmu_update *req, int count,
240 int *success_count, domid_t domid)
242 return _hypercall4(int, mmu_update, req, count, success_count, domid);
246 HYPERVISOR_mmuext_op(struct mmuext_op *op, int count,
247 int *success_count, domid_t domid)
249 return _hypercall4(int, mmuext_op, op, count, success_count, domid);
253 HYPERVISOR_set_gdt(unsigned long *frame_list, int entries)
255 return _hypercall2(int, set_gdt, frame_list, entries);
259 HYPERVISOR_stack_switch(unsigned long ss, unsigned long esp)
261 return _hypercall2(int, stack_switch, ss, esp);
266 HYPERVISOR_set_callbacks(unsigned long event_selector,
267 unsigned long event_address,
268 unsigned long failsafe_selector,
269 unsigned long failsafe_address)
271 return _hypercall4(int, set_callbacks,
272 event_selector, event_address,
273 failsafe_selector, failsafe_address);
275 #else /* CONFIG_X86_64 */
277 HYPERVISOR_set_callbacks(unsigned long event_address,
278 unsigned long failsafe_address,
279 unsigned long syscall_address)
281 return _hypercall3(int, set_callbacks,
282 event_address, failsafe_address,
285 #endif /* CONFIG_X86_{32,64} */
288 HYPERVISOR_callback_op(int cmd, void *arg)
290 return _hypercall2(int, callback_op, cmd, arg);
294 HYPERVISOR_fpu_taskswitch(int set)
296 return _hypercall1(int, fpu_taskswitch, set);
300 HYPERVISOR_sched_op(int cmd, void *arg)
302 return _hypercall2(int, sched_op, cmd, arg);
306 HYPERVISOR_set_timer_op(u64 timeout)
308 unsigned long timeout_hi = (unsigned long)(timeout>>32);
309 unsigned long timeout_lo = (unsigned long)timeout;
310 return _hypercall2(long, set_timer_op, timeout_lo, timeout_hi);
314 HYPERVISOR_mca(struct xen_mc *mc_op)
316 mc_op->interface_version = XEN_MCA_INTERFACE_VERSION;
317 return _hypercall1(int, mca, mc_op);
321 HYPERVISOR_platform_op(struct xen_platform_op *op)
323 op->interface_version = XENPF_INTERFACE_VERSION;
324 return _hypercall1(int, platform_op, op);
328 HYPERVISOR_set_debugreg(int reg, unsigned long value)
330 return _hypercall2(int, set_debugreg, reg, value);
333 static inline unsigned long
334 HYPERVISOR_get_debugreg(int reg)
336 return _hypercall1(unsigned long, get_debugreg, reg);
340 HYPERVISOR_update_descriptor(u64 ma, u64 desc)
342 if (sizeof(u64) == sizeof(long))
343 return _hypercall2(int, update_descriptor, ma, desc);
344 return _hypercall4(int, update_descriptor, ma, ma>>32, desc, desc>>32);
348 HYPERVISOR_memory_op(unsigned int cmd, void *arg)
350 return _hypercall2(long, memory_op, cmd, arg);
354 HYPERVISOR_multicall(void *call_list, uint32_t nr_calls)
356 return _hypercall2(int, multicall, call_list, nr_calls);
360 HYPERVISOR_update_va_mapping(unsigned long va, pte_t new_val,
363 if (sizeof(new_val) == sizeof(long))
364 return _hypercall3(int, update_va_mapping, va,
367 return _hypercall4(int, update_va_mapping, va,
368 new_val.pte, new_val.pte >> 32, flags);
370 extern int __must_check xen_event_channel_op_compat(int, void *);
373 HYPERVISOR_event_channel_op(int cmd, void *arg)
375 int rc = _hypercall2(int, event_channel_op, cmd, arg);
376 if (unlikely(rc == -ENOSYS))
377 rc = xen_event_channel_op_compat(cmd, arg);
382 HYPERVISOR_xen_version(int cmd, void *arg)
384 return _hypercall2(int, xen_version, cmd, arg);
388 HYPERVISOR_console_io(int cmd, int count, char *str)
390 return _hypercall3(int, console_io, cmd, count, str);
393 extern int __must_check xen_physdev_op_compat(int, void *);
396 HYPERVISOR_physdev_op(int cmd, void *arg)
398 int rc = _hypercall2(int, physdev_op, cmd, arg);
399 if (unlikely(rc == -ENOSYS))
400 rc = xen_physdev_op_compat(cmd, arg);
405 HYPERVISOR_grant_table_op(unsigned int cmd, void *uop, unsigned int count)
407 return _hypercall3(int, grant_table_op, cmd, uop, count);
411 HYPERVISOR_update_va_mapping_otherdomain(unsigned long va, pte_t new_val,
412 unsigned long flags, domid_t domid)
414 if (sizeof(new_val) == sizeof(long))
415 return _hypercall4(int, update_va_mapping_otherdomain, va,
416 new_val.pte, flags, domid);
418 return _hypercall5(int, update_va_mapping_otherdomain, va,
419 new_val.pte, new_val.pte >> 32,
424 HYPERVISOR_vm_assist(unsigned int cmd, unsigned int type)
426 return _hypercall2(int, vm_assist, cmd, type);
430 HYPERVISOR_vcpu_op(int cmd, int vcpuid, void *extra_args)
432 return _hypercall3(int, vcpu_op, cmd, vcpuid, extra_args);
437 HYPERVISOR_set_segment_base(int reg, unsigned long value)
439 return _hypercall2(int, set_segment_base, reg, value);
444 HYPERVISOR_suspend(unsigned long start_info_mfn)
446 struct sched_shutdown r = { .reason = SHUTDOWN_suspend };
449 * For a PV guest the tools require that the start_info mfn be
450 * present in rdx/edx when the hypercall is made. Per the
451 * hypercall calling convention this is the third hypercall
452 * argument, which is start_info_mfn here.
454 return _hypercall3(int, sched_op, SCHEDOP_shutdown, &r, start_info_mfn);
458 HYPERVISOR_nmi_op(unsigned long op, unsigned long arg)
460 return _hypercall2(int, nmi_op, op, arg);
463 static inline unsigned long __must_check
464 HYPERVISOR_hvm_op(int op, void *arg)
466 return _hypercall2(unsigned long, hvm_op, op, arg);
473 return _hypercall1(int, tmem_op, op);
477 HYPERVISOR_xenpmu_op(unsigned int op, void *arg)
479 return _hypercall2(int, xenpmu_op, op, arg);
483 MULTI_fpu_taskswitch(struct multicall_entry *mcl, int set)
485 mcl->op = __HYPERVISOR_fpu_taskswitch;
488 trace_xen_mc_entry(mcl, 1);
492 MULTI_update_va_mapping(struct multicall_entry *mcl, unsigned long va,
493 pte_t new_val, unsigned long flags)
495 mcl->op = __HYPERVISOR_update_va_mapping;
497 if (sizeof(new_val) == sizeof(long)) {
498 mcl->args[1] = new_val.pte;
499 mcl->args[2] = flags;
501 mcl->args[1] = new_val.pte;
502 mcl->args[2] = new_val.pte >> 32;
503 mcl->args[3] = flags;
506 trace_xen_mc_entry(mcl, sizeof(new_val) == sizeof(long) ? 3 : 4);
510 MULTI_grant_table_op(struct multicall_entry *mcl, unsigned int cmd,
511 void *uop, unsigned int count)
513 mcl->op = __HYPERVISOR_grant_table_op;
515 mcl->args[1] = (unsigned long)uop;
516 mcl->args[2] = count;
518 trace_xen_mc_entry(mcl, 3);
522 MULTI_update_va_mapping_otherdomain(struct multicall_entry *mcl, unsigned long va,
523 pte_t new_val, unsigned long flags,
526 mcl->op = __HYPERVISOR_update_va_mapping_otherdomain;
528 if (sizeof(new_val) == sizeof(long)) {
529 mcl->args[1] = new_val.pte;
530 mcl->args[2] = flags;
531 mcl->args[3] = domid;
533 mcl->args[1] = new_val.pte;
534 mcl->args[2] = new_val.pte >> 32;
535 mcl->args[3] = flags;
536 mcl->args[4] = domid;
539 trace_xen_mc_entry(mcl, sizeof(new_val) == sizeof(long) ? 4 : 5);
543 MULTI_update_descriptor(struct multicall_entry *mcl, u64 maddr,
544 struct desc_struct desc)
546 mcl->op = __HYPERVISOR_update_descriptor;
547 if (sizeof(maddr) == sizeof(long)) {
548 mcl->args[0] = maddr;
549 mcl->args[1] = *(unsigned long *)&desc;
551 mcl->args[0] = maddr;
552 mcl->args[1] = maddr >> 32;
553 mcl->args[2] = desc.a;
554 mcl->args[3] = desc.b;
557 trace_xen_mc_entry(mcl, sizeof(maddr) == sizeof(long) ? 2 : 4);
561 MULTI_memory_op(struct multicall_entry *mcl, unsigned int cmd, void *arg)
563 mcl->op = __HYPERVISOR_memory_op;
565 mcl->args[1] = (unsigned long)arg;
567 trace_xen_mc_entry(mcl, 2);
571 MULTI_mmu_update(struct multicall_entry *mcl, struct mmu_update *req,
572 int count, int *success_count, domid_t domid)
574 mcl->op = __HYPERVISOR_mmu_update;
575 mcl->args[0] = (unsigned long)req;
576 mcl->args[1] = count;
577 mcl->args[2] = (unsigned long)success_count;
578 mcl->args[3] = domid;
580 trace_xen_mc_entry(mcl, 4);
584 MULTI_mmuext_op(struct multicall_entry *mcl, struct mmuext_op *op, int count,
585 int *success_count, domid_t domid)
587 mcl->op = __HYPERVISOR_mmuext_op;
588 mcl->args[0] = (unsigned long)op;
589 mcl->args[1] = count;
590 mcl->args[2] = (unsigned long)success_count;
591 mcl->args[3] = domid;
593 trace_xen_mc_entry(mcl, 4);
597 MULTI_set_gdt(struct multicall_entry *mcl, unsigned long *frames, int entries)
599 mcl->op = __HYPERVISOR_set_gdt;
600 mcl->args[0] = (unsigned long)frames;
601 mcl->args[1] = entries;
603 trace_xen_mc_entry(mcl, 2);
607 MULTI_stack_switch(struct multicall_entry *mcl,
608 unsigned long ss, unsigned long esp)
610 mcl->op = __HYPERVISOR_stack_switch;
614 trace_xen_mc_entry(mcl, 2);
617 #endif /* _ASM_X86_XEN_HYPERCALL_H */