1  /* SPDX-License-Identifier: GPL-2.0 */
2  #ifndef __KVM_X86_SVM_OPS_H
3  #define __KVM_X86_SVM_OPS_H
4  
5  #include <linux/compiler_types.h>
6  
7  #include "x86.h"
8  
9  #define svm_asm(insn, clobber...)				\
10  do {								\
11  	asm_volatile_goto("1: " __stringify(insn) "\n\t"	\
12  			  _ASM_EXTABLE(1b, %l[fault])		\
13  			  ::: clobber : fault);			\
14  	return;							\
15  fault:								\
16  	kvm_spurious_fault();					\
17  } while (0)
18  
19  #define svm_asm1(insn, op1, clobber...)				\
20  do {								\
21  	asm_volatile_goto("1: "  __stringify(insn) " %0\n\t"	\
22  			  _ASM_EXTABLE(1b, %l[fault])		\
23  			  :: op1 : clobber : fault);		\
24  	return;							\
25  fault:								\
26  	kvm_spurious_fault();					\
27  } while (0)
28  
29  #define svm_asm2(insn, op1, op2, clobber...)				\
30  do {									\
31  	asm_volatile_goto("1: "  __stringify(insn) " %1, %0\n\t"	\
32  			  _ASM_EXTABLE(1b, %l[fault])			\
33  			  :: op1, op2 : clobber : fault);		\
34  	return;								\
35  fault:									\
36  	kvm_spurious_fault();						\
37  } while (0)
38  
clgi(void)39  static inline void clgi(void)
40  {
41  	svm_asm(clgi);
42  }
43  
stgi(void)44  static inline void stgi(void)
45  {
46  	svm_asm(stgi);
47  }
48  
invlpga(unsigned long addr,u32 asid)49  static inline void invlpga(unsigned long addr, u32 asid)
50  {
51  	svm_asm2(invlpga, "c"(asid), "a"(addr));
52  }
53  
54  /*
55   * Despite being a physical address, the portion of rAX that is consumed by
56   * VMSAVE, VMLOAD, etc... is still controlled by the effective address size,
57   * hence 'unsigned long' instead of 'hpa_t'.
58   */
vmsave(unsigned long pa)59  static __always_inline void vmsave(unsigned long pa)
60  {
61  	svm_asm1(vmsave, "a" (pa), "memory");
62  }
63  
64  #endif /* __KVM_X86_SVM_OPS_H */
65