MN10300: Kill MN10300's own profiling Kconfig
[linux-2.6.git] / arch / x86 / include / asm / paravirt.h
1 #ifndef _ASM_X86_PARAVIRT_H
2 #define _ASM_X86_PARAVIRT_H
3 /* Various instructions on x86 need to be replaced for
4  * para-virtualization: those hooks are defined here. */
5
6 #ifdef CONFIG_PARAVIRT
7 #include <asm/pgtable_types.h>
8 #include <asm/asm.h>
9
10 /* Bitmask of what can be clobbered: usually at least eax. */
11 #define CLBR_NONE 0
12 #define CLBR_EAX  (1 << 0)
13 #define CLBR_ECX  (1 << 1)
14 #define CLBR_EDX  (1 << 2)
15 #define CLBR_EDI  (1 << 3)
16
17 #ifdef CONFIG_X86_32
18 /* CLBR_ANY should match all regs platform has. For i386, that's just it */
19 #define CLBR_ANY  ((1 << 4) - 1)
20
21 #define CLBR_ARG_REGS   (CLBR_EAX | CLBR_EDX | CLBR_ECX)
22 #define CLBR_RET_REG    (CLBR_EAX | CLBR_EDX)
23 #define CLBR_SCRATCH    (0)
24 #else
25 #define CLBR_RAX  CLBR_EAX
26 #define CLBR_RCX  CLBR_ECX
27 #define CLBR_RDX  CLBR_EDX
28 #define CLBR_RDI  CLBR_EDI
29 #define CLBR_RSI  (1 << 4)
30 #define CLBR_R8   (1 << 5)
31 #define CLBR_R9   (1 << 6)
32 #define CLBR_R10  (1 << 7)
33 #define CLBR_R11  (1 << 8)
34
35 #define CLBR_ANY  ((1 << 9) - 1)
36
37 #define CLBR_ARG_REGS   (CLBR_RDI | CLBR_RSI | CLBR_RDX | \
38                          CLBR_RCX | CLBR_R8 | CLBR_R9)
39 #define CLBR_RET_REG    (CLBR_RAX)
40 #define CLBR_SCRATCH    (CLBR_R10 | CLBR_R11)
41
42 #include <asm/desc_defs.h>
43 #endif /* X86_64 */
44
45 #define CLBR_CALLEE_SAVE ((CLBR_ARG_REGS | CLBR_SCRATCH) & ~CLBR_RET_REG)
46
47 #ifndef __ASSEMBLY__
48 #include <linux/types.h>
49 #include <linux/cpumask.h>
50 #include <asm/kmap_types.h>
51 #include <asm/desc_defs.h>
52
53 struct page;
54 struct thread_struct;
55 struct desc_ptr;
56 struct tss_struct;
57 struct mm_struct;
58 struct desc_struct;
59
60 /*
61  * Wrapper type for pointers to code which uses the non-standard
62  * calling convention.  See PV_CALL_SAVE_REGS_THUNK below.
63  */
64 struct paravirt_callee_save {
65         void *func;
66 };
67
68 /* general info */
69 struct pv_info {
70         unsigned int kernel_rpl;
71         int shared_kernel_pmd;
72         int paravirt_enabled;
73         const char *name;
74 };
75
76 struct pv_init_ops {
77         /*
78          * Patch may replace one of the defined code sequences with
79          * arbitrary code, subject to the same register constraints.
80          * This generally means the code is not free to clobber any
81          * registers other than EAX.  The patch function should return
82          * the number of bytes of code generated, as we nop pad the
83          * rest in generic code.
84          */
85         unsigned (*patch)(u8 type, u16 clobber, void *insnbuf,
86                           unsigned long addr, unsigned len);
87
88         /* Basic arch-specific setup */
89         void (*arch_setup)(void);
90         char *(*memory_setup)(void);
91         void (*post_allocator_init)(void);
92
93         /* Print a banner to identify the environment */
94         void (*banner)(void);
95 };
96
97
98 struct pv_lazy_ops {
99         /* Set deferred update mode, used for batching operations. */
100         void (*enter)(void);
101         void (*leave)(void);
102 };
103
104 struct pv_time_ops {
105         void (*time_init)(void);
106
107         /* Set and set time of day */
108         unsigned long (*get_wallclock)(void);
109         int (*set_wallclock)(unsigned long);
110
111         unsigned long long (*sched_clock)(void);
112         unsigned long (*get_tsc_khz)(void);
113 };
114
115 struct pv_cpu_ops {
116         /* hooks for various privileged instructions */
117         unsigned long (*get_debugreg)(int regno);
118         void (*set_debugreg)(int regno, unsigned long value);
119
120         void (*clts)(void);
121
122         unsigned long (*read_cr0)(void);
123         void (*write_cr0)(unsigned long);
124
125         unsigned long (*read_cr4_safe)(void);
126         unsigned long (*read_cr4)(void);
127         void (*write_cr4)(unsigned long);
128
129 #ifdef CONFIG_X86_64
130         unsigned long (*read_cr8)(void);
131         void (*write_cr8)(unsigned long);
132 #endif
133
134         /* Segment descriptor handling */
135         void (*load_tr_desc)(void);
136         void (*load_gdt)(const struct desc_ptr *);
137         void (*load_idt)(const struct desc_ptr *);
138         void (*store_gdt)(struct desc_ptr *);
139         void (*store_idt)(struct desc_ptr *);
140         void (*set_ldt)(const void *desc, unsigned entries);
141         unsigned long (*store_tr)(void);
142         void (*load_tls)(struct thread_struct *t, unsigned int cpu);
143 #ifdef CONFIG_X86_64
144         void (*load_gs_index)(unsigned int idx);
145 #endif
146         void (*write_ldt_entry)(struct desc_struct *ldt, int entrynum,
147                                 const void *desc);
148         void (*write_gdt_entry)(struct desc_struct *,
149                                 int entrynum, const void *desc, int size);
150         void (*write_idt_entry)(gate_desc *,
151                                 int entrynum, const gate_desc *gate);
152         void (*alloc_ldt)(struct desc_struct *ldt, unsigned entries);
153         void (*free_ldt)(struct desc_struct *ldt, unsigned entries);
154
155         void (*load_sp0)(struct tss_struct *tss, struct thread_struct *t);
156
157         void (*set_iopl_mask)(unsigned mask);
158
159         void (*wbinvd)(void);
160         void (*io_delay)(void);
161
162         /* cpuid emulation, mostly so that caps bits can be disabled */
163         void (*cpuid)(unsigned int *eax, unsigned int *ebx,
164                       unsigned int *ecx, unsigned int *edx);
165
166         /* MSR, PMC and TSR operations.
167            err = 0/-EFAULT.  wrmsr returns 0/-EFAULT. */
168         u64 (*read_msr_amd)(unsigned int msr, int *err);
169         u64 (*read_msr)(unsigned int msr, int *err);
170         int (*write_msr)(unsigned int msr, unsigned low, unsigned high);
171
172         u64 (*read_tsc)(void);
173         u64 (*read_pmc)(int counter);
174         unsigned long long (*read_tscp)(unsigned int *aux);
175
176         /*
177          * Atomically enable interrupts and return to userspace.  This
178          * is only ever used to return to 32-bit processes; in a
179          * 64-bit kernel, it's used for 32-on-64 compat processes, but
180          * never native 64-bit processes.  (Jump, not call.)
181          */
182         void (*irq_enable_sysexit)(void);
183
184         /*
185          * Switch to usermode gs and return to 64-bit usermode using
186          * sysret.  Only used in 64-bit kernels to return to 64-bit
187          * processes.  Usermode register state, including %rsp, must
188          * already be restored.
189          */
190         void (*usergs_sysret64)(void);
191
192         /*
193          * Switch to usermode gs and return to 32-bit usermode using
194          * sysret.  Used to return to 32-on-64 compat processes.
195          * Other usermode register state, including %esp, must already
196          * be restored.
197          */
198         void (*usergs_sysret32)(void);
199
200         /* Normal iret.  Jump to this with the standard iret stack
201            frame set up. */
202         void (*iret)(void);
203
204         void (*swapgs)(void);
205
206         struct pv_lazy_ops lazy_mode;
207 };
208
209 struct pv_irq_ops {
210         void (*init_IRQ)(void);
211
212         /*
213          * Get/set interrupt state.  save_fl and restore_fl are only
214          * expected to use X86_EFLAGS_IF; all other bits
215          * returned from save_fl are undefined, and may be ignored by
216          * restore_fl.
217          *
218          * NOTE: These functions callers expect the callee to preserve
219          * more registers than the standard C calling convention.
220          */
221         struct paravirt_callee_save save_fl;
222         struct paravirt_callee_save restore_fl;
223         struct paravirt_callee_save irq_disable;
224         struct paravirt_callee_save irq_enable;
225
226         void (*safe_halt)(void);
227         void (*halt)(void);
228
229 #ifdef CONFIG_X86_64
230         void (*adjust_exception_frame)(void);
231 #endif
232 };
233
234 struct pv_apic_ops {
235 #ifdef CONFIG_X86_LOCAL_APIC
236         void (*setup_boot_clock)(void);
237         void (*setup_secondary_clock)(void);
238
239         void (*startup_ipi_hook)(int phys_apicid,
240                                  unsigned long start_eip,
241                                  unsigned long start_esp);
242 #endif
243 };
244
245 struct pv_mmu_ops {
246         /*
247          * Called before/after init_mm pagetable setup. setup_start
248          * may reset %cr3, and may pre-install parts of the pagetable;
249          * pagetable setup is expected to preserve any existing
250          * mapping.
251          */
252         void (*pagetable_setup_start)(pgd_t *pgd_base);
253         void (*pagetable_setup_done)(pgd_t *pgd_base);
254
255         unsigned long (*read_cr2)(void);
256         void (*write_cr2)(unsigned long);
257
258         unsigned long (*read_cr3)(void);
259         void (*write_cr3)(unsigned long);
260
261         /*
262          * Hooks for intercepting the creation/use/destruction of an
263          * mm_struct.
264          */
265         void (*activate_mm)(struct mm_struct *prev,
266                             struct mm_struct *next);
267         void (*dup_mmap)(struct mm_struct *oldmm,
268                          struct mm_struct *mm);
269         void (*exit_mmap)(struct mm_struct *mm);
270
271
272         /* TLB operations */
273         void (*flush_tlb_user)(void);
274         void (*flush_tlb_kernel)(void);
275         void (*flush_tlb_single)(unsigned long addr);
276         void (*flush_tlb_others)(const struct cpumask *cpus,
277                                  struct mm_struct *mm,
278                                  unsigned long va);
279
280         /* Hooks for allocating and freeing a pagetable top-level */
281         int  (*pgd_alloc)(struct mm_struct *mm);
282         void (*pgd_free)(struct mm_struct *mm, pgd_t *pgd);
283
284         /*
285          * Hooks for allocating/releasing pagetable pages when they're
286          * attached to a pagetable
287          */
288         void (*alloc_pte)(struct mm_struct *mm, unsigned long pfn);
289         void (*alloc_pmd)(struct mm_struct *mm, unsigned long pfn);
290         void (*alloc_pmd_clone)(unsigned long pfn, unsigned long clonepfn, unsigned long start, unsigned long count);
291         void (*alloc_pud)(struct mm_struct *mm, unsigned long pfn);
292         void (*release_pte)(unsigned long pfn);
293         void (*release_pmd)(unsigned long pfn);
294         void (*release_pud)(unsigned long pfn);
295
296         /* Pagetable manipulation functions */
297         void (*set_pte)(pte_t *ptep, pte_t pteval);
298         void (*set_pte_at)(struct mm_struct *mm, unsigned long addr,
299                            pte_t *ptep, pte_t pteval);
300         void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval);
301         void (*pte_update)(struct mm_struct *mm, unsigned long addr,
302                            pte_t *ptep);
303         void (*pte_update_defer)(struct mm_struct *mm,
304                                  unsigned long addr, pte_t *ptep);
305
306         pte_t (*ptep_modify_prot_start)(struct mm_struct *mm, unsigned long addr,
307                                         pte_t *ptep);
308         void (*ptep_modify_prot_commit)(struct mm_struct *mm, unsigned long addr,
309                                         pte_t *ptep, pte_t pte);
310
311         struct paravirt_callee_save pte_val;
312         struct paravirt_callee_save make_pte;
313
314         struct paravirt_callee_save pgd_val;
315         struct paravirt_callee_save make_pgd;
316
317 #if PAGETABLE_LEVELS >= 3
318 #ifdef CONFIG_X86_PAE
319         void (*set_pte_atomic)(pte_t *ptep, pte_t pteval);
320         void (*pte_clear)(struct mm_struct *mm, unsigned long addr,
321                           pte_t *ptep);
322         void (*pmd_clear)(pmd_t *pmdp);
323
324 #endif  /* CONFIG_X86_PAE */
325
326         void (*set_pud)(pud_t *pudp, pud_t pudval);
327
328         struct paravirt_callee_save pmd_val;
329         struct paravirt_callee_save make_pmd;
330
331 #if PAGETABLE_LEVELS == 4
332         struct paravirt_callee_save pud_val;
333         struct paravirt_callee_save make_pud;
334
335         void (*set_pgd)(pgd_t *pudp, pgd_t pgdval);
336 #endif  /* PAGETABLE_LEVELS == 4 */
337 #endif  /* PAGETABLE_LEVELS >= 3 */
338
339 #ifdef CONFIG_HIGHPTE
340         void *(*kmap_atomic_pte)(struct page *page, enum km_type type);
341 #endif
342
343         struct pv_lazy_ops lazy_mode;
344
345         /* dom0 ops */
346
347         /* Sometimes the physical address is a pfn, and sometimes its
348            an mfn.  We can tell which is which from the index. */
349         void (*set_fixmap)(unsigned /* enum fixed_addresses */ idx,
350                            unsigned long phys, pgprot_t flags);
351 };
352
353 struct raw_spinlock;
354 struct pv_lock_ops {
355         int (*spin_is_locked)(struct raw_spinlock *lock);
356         int (*spin_is_contended)(struct raw_spinlock *lock);
357         void (*spin_lock)(struct raw_spinlock *lock);
358         void (*spin_lock_flags)(struct raw_spinlock *lock, unsigned long flags);
359         int (*spin_trylock)(struct raw_spinlock *lock);
360         void (*spin_unlock)(struct raw_spinlock *lock);
361 };
362
363 /* This contains all the paravirt structures: we get a convenient
364  * number for each function using the offset which we use to indicate
365  * what to patch. */
366 struct paravirt_patch_template {
367         struct pv_init_ops pv_init_ops;
368         struct pv_time_ops pv_time_ops;
369         struct pv_cpu_ops pv_cpu_ops;
370         struct pv_irq_ops pv_irq_ops;
371         struct pv_apic_ops pv_apic_ops;
372         struct pv_mmu_ops pv_mmu_ops;
373         struct pv_lock_ops pv_lock_ops;
374 };
375
376 extern struct pv_info pv_info;
377 extern struct pv_init_ops pv_init_ops;
378 extern struct pv_time_ops pv_time_ops;
379 extern struct pv_cpu_ops pv_cpu_ops;
380 extern struct pv_irq_ops pv_irq_ops;
381 extern struct pv_apic_ops pv_apic_ops;
382 extern struct pv_mmu_ops pv_mmu_ops;
383 extern struct pv_lock_ops pv_lock_ops;
384
385 #define PARAVIRT_PATCH(x)                                       \
386         (offsetof(struct paravirt_patch_template, x) / sizeof(void *))
387
388 #define paravirt_type(op)                               \
389         [paravirt_typenum] "i" (PARAVIRT_PATCH(op)),    \
390         [paravirt_opptr] "i" (&(op))
391 #define paravirt_clobber(clobber)               \
392         [paravirt_clobber] "i" (clobber)
393
394 /*
395  * Generate some code, and mark it as patchable by the
396  * apply_paravirt() alternate instruction patcher.
397  */
398 #define _paravirt_alt(insn_string, type, clobber)       \
399         "771:\n\t" insn_string "\n" "772:\n"            \
400         ".pushsection .parainstructions,\"a\"\n"        \
401         _ASM_ALIGN "\n"                                 \
402         _ASM_PTR " 771b\n"                              \
403         "  .byte " type "\n"                            \
404         "  .byte 772b-771b\n"                           \
405         "  .short " clobber "\n"                        \
406         ".popsection\n"
407
408 /* Generate patchable code, with the default asm parameters. */
409 #define paravirt_alt(insn_string)                                       \
410         _paravirt_alt(insn_string, "%c[paravirt_typenum]", "%c[paravirt_clobber]")
411
412 /* Simple instruction patching code. */
413 #define DEF_NATIVE(ops, name, code)                                     \
414         extern const char start_##ops##_##name[], end_##ops##_##name[]; \
415         asm("start_" #ops "_" #name ": " code "; end_" #ops "_" #name ":")
416
417 unsigned paravirt_patch_nop(void);
418 unsigned paravirt_patch_ident_32(void *insnbuf, unsigned len);
419 unsigned paravirt_patch_ident_64(void *insnbuf, unsigned len);
420 unsigned paravirt_patch_ignore(unsigned len);
421 unsigned paravirt_patch_call(void *insnbuf,
422                              const void *target, u16 tgt_clobbers,
423                              unsigned long addr, u16 site_clobbers,
424                              unsigned len);
425 unsigned paravirt_patch_jmp(void *insnbuf, const void *target,
426                             unsigned long addr, unsigned len);
427 unsigned paravirt_patch_default(u8 type, u16 clobbers, void *insnbuf,
428                                 unsigned long addr, unsigned len);
429
430 unsigned paravirt_patch_insns(void *insnbuf, unsigned len,
431                               const char *start, const char *end);
432
433 unsigned native_patch(u8 type, u16 clobbers, void *ibuf,
434                       unsigned long addr, unsigned len);
435
436 int paravirt_disable_iospace(void);
437
438 /*
439  * This generates an indirect call based on the operation type number.
440  * The type number, computed in PARAVIRT_PATCH, is derived from the
441  * offset into the paravirt_patch_template structure, and can therefore be
442  * freely converted back into a structure offset.
443  */
444 #define PARAVIRT_CALL   "call *%c[paravirt_opptr];"
445
446 /*
447  * These macros are intended to wrap calls through one of the paravirt
448  * ops structs, so that they can be later identified and patched at
449  * runtime.
450  *
451  * Normally, a call to a pv_op function is a simple indirect call:
452  * (pv_op_struct.operations)(args...).
453  *
454  * Unfortunately, this is a relatively slow operation for modern CPUs,
455  * because it cannot necessarily determine what the destination
456  * address is.  In this case, the address is a runtime constant, so at
457  * the very least we can patch the call to e a simple direct call, or
458  * ideally, patch an inline implementation into the callsite.  (Direct
459  * calls are essentially free, because the call and return addresses
460  * are completely predictable.)
461  *
462  * For i386, these macros rely on the standard gcc "regparm(3)" calling
463  * convention, in which the first three arguments are placed in %eax,
464  * %edx, %ecx (in that order), and the remaining arguments are placed
465  * on the stack.  All caller-save registers (eax,edx,ecx) are expected
466  * to be modified (either clobbered or used for return values).
467  * X86_64, on the other hand, already specifies a register-based calling
468  * conventions, returning at %rax, with parameteres going on %rdi, %rsi,
469  * %rdx, and %rcx. Note that for this reason, x86_64 does not need any
470  * special handling for dealing with 4 arguments, unlike i386.
471  * However, x86_64 also have to clobber all caller saved registers, which
472  * unfortunately, are quite a bit (r8 - r11)
473  *
474  * The call instruction itself is marked by placing its start address
475  * and size into the .parainstructions section, so that
476  * apply_paravirt() in arch/i386/kernel/alternative.c can do the
477  * appropriate patching under the control of the backend pv_init_ops
478  * implementation.
479  *
480  * Unfortunately there's no way to get gcc to generate the args setup
481  * for the call, and then allow the call itself to be generated by an
482  * inline asm.  Because of this, we must do the complete arg setup and
483  * return value handling from within these macros.  This is fairly
484  * cumbersome.
485  *
486  * There are 5 sets of PVOP_* macros for dealing with 0-4 arguments.
487  * It could be extended to more arguments, but there would be little
488  * to be gained from that.  For each number of arguments, there are
489  * the two VCALL and CALL variants for void and non-void functions.
490  *
491  * When there is a return value, the invoker of the macro must specify
492  * the return type.  The macro then uses sizeof() on that type to
493  * determine whether its a 32 or 64 bit value, and places the return
494  * in the right register(s) (just %eax for 32-bit, and %edx:%eax for
495  * 64-bit). For x86_64 machines, it just returns at %rax regardless of
496  * the return value size.
497  *
498  * 64-bit arguments are passed as a pair of adjacent 32-bit arguments
499  * i386 also passes 64-bit arguments as a pair of adjacent 32-bit arguments
500  * in low,high order
501  *
502  * Small structures are passed and returned in registers.  The macro
503  * calling convention can't directly deal with this, so the wrapper
504  * functions must do this.
505  *
506  * These PVOP_* macros are only defined within this header.  This
507  * means that all uses must be wrapped in inline functions.  This also
508  * makes sure the incoming and outgoing types are always correct.
509  */
510 #ifdef CONFIG_X86_32
511 #define PVOP_VCALL_ARGS                         \
512         unsigned long __eax = __eax, __edx = __edx, __ecx = __ecx
513 #define PVOP_CALL_ARGS                  PVOP_VCALL_ARGS
514
515 #define PVOP_CALL_ARG1(x)               "a" ((unsigned long)(x))
516 #define PVOP_CALL_ARG2(x)               "d" ((unsigned long)(x))
517 #define PVOP_CALL_ARG3(x)               "c" ((unsigned long)(x))
518
519 #define PVOP_VCALL_CLOBBERS             "=a" (__eax), "=d" (__edx),     \
520                                         "=c" (__ecx)
521 #define PVOP_CALL_CLOBBERS              PVOP_VCALL_CLOBBERS
522
523 #define PVOP_VCALLEE_CLOBBERS           "=a" (__eax), "=d" (__edx)
524 #define PVOP_CALLEE_CLOBBERS            PVOP_VCALLEE_CLOBBERS
525
526 #define EXTRA_CLOBBERS
527 #define VEXTRA_CLOBBERS
528 #else  /* CONFIG_X86_64 */
529 #define PVOP_VCALL_ARGS                                 \
530         unsigned long __edi = __edi, __esi = __esi,     \
531                 __edx = __edx, __ecx = __ecx
532 #define PVOP_CALL_ARGS          PVOP_VCALL_ARGS, __eax
533
534 #define PVOP_CALL_ARG1(x)               "D" ((unsigned long)(x))
535 #define PVOP_CALL_ARG2(x)               "S" ((unsigned long)(x))
536 #define PVOP_CALL_ARG3(x)               "d" ((unsigned long)(x))
537 #define PVOP_CALL_ARG4(x)               "c" ((unsigned long)(x))
538
539 #define PVOP_VCALL_CLOBBERS     "=D" (__edi),                           \
540                                 "=S" (__esi), "=d" (__edx),             \
541                                 "=c" (__ecx)
542 #define PVOP_CALL_CLOBBERS      PVOP_VCALL_CLOBBERS, "=a" (__eax)
543
544 #define PVOP_VCALLEE_CLOBBERS   "=a" (__eax)
545 #define PVOP_CALLEE_CLOBBERS    PVOP_VCALLEE_CLOBBERS
546
547 #define EXTRA_CLOBBERS   , "r8", "r9", "r10", "r11"
548 #define VEXTRA_CLOBBERS  , "rax", "r8", "r9", "r10", "r11"
549 #endif  /* CONFIG_X86_32 */
550
551 #ifdef CONFIG_PARAVIRT_DEBUG
552 #define PVOP_TEST_NULL(op)      BUG_ON(op == NULL)
553 #else
554 #define PVOP_TEST_NULL(op)      ((void)op)
555 #endif
556
557 #define ____PVOP_CALL(rettype, op, clbr, call_clbr, extra_clbr,         \
558                       pre, post, ...)                                   \
559         ({                                                              \
560                 rettype __ret;                                          \
561                 PVOP_CALL_ARGS;                                         \
562                 PVOP_TEST_NULL(op);                                     \
563                 /* This is 32-bit specific, but is okay in 64-bit */    \
564                 /* since this condition will never hold */              \
565                 if (sizeof(rettype) > sizeof(unsigned long)) {          \
566                         asm volatile(pre                                \
567                                      paravirt_alt(PARAVIRT_CALL)        \
568                                      post                               \
569                                      : call_clbr                        \
570                                      : paravirt_type(op),               \
571                                        paravirt_clobber(clbr),          \
572                                        ##__VA_ARGS__                    \
573                                      : "memory", "cc" extra_clbr);      \
574                         __ret = (rettype)((((u64)__edx) << 32) | __eax); \
575                 } else {                                                \
576                         asm volatile(pre                                \
577                                      paravirt_alt(PARAVIRT_CALL)        \
578                                      post                               \
579                                      : call_clbr                        \
580                                      : paravirt_type(op),               \
581                                        paravirt_clobber(clbr),          \
582                                        ##__VA_ARGS__                    \
583                                      : "memory", "cc" extra_clbr);      \
584                         __ret = (rettype)__eax;                         \
585                 }                                                       \
586                 __ret;                                                  \
587         })
588
589 #define __PVOP_CALL(rettype, op, pre, post, ...)                        \
590         ____PVOP_CALL(rettype, op, CLBR_ANY, PVOP_CALL_CLOBBERS,        \
591                       EXTRA_CLOBBERS, pre, post, ##__VA_ARGS__)
592
593 #define __PVOP_CALLEESAVE(rettype, op, pre, post, ...)                  \
594         ____PVOP_CALL(rettype, op.func, CLBR_RET_REG,                   \
595                       PVOP_CALLEE_CLOBBERS, ,                           \
596                       pre, post, ##__VA_ARGS__)
597
598
599 #define ____PVOP_VCALL(op, clbr, call_clbr, extra_clbr, pre, post, ...) \
600         ({                                                              \
601                 PVOP_VCALL_ARGS;                                        \
602                 PVOP_TEST_NULL(op);                                     \
603                 asm volatile(pre                                        \
604                              paravirt_alt(PARAVIRT_CALL)                \
605                              post                                       \
606                              : call_clbr                                \
607                              : paravirt_type(op),                       \
608                                paravirt_clobber(clbr),                  \
609                                ##__VA_ARGS__                            \
610                              : "memory", "cc" extra_clbr);              \
611         })
612
613 #define __PVOP_VCALL(op, pre, post, ...)                                \
614         ____PVOP_VCALL(op, CLBR_ANY, PVOP_VCALL_CLOBBERS,               \
615                        VEXTRA_CLOBBERS,                                 \
616                        pre, post, ##__VA_ARGS__)
617
618 #define __PVOP_VCALLEESAVE(rettype, op, pre, post, ...)                 \
619         ____PVOP_CALL(rettype, op.func, CLBR_RET_REG,                   \
620                       PVOP_VCALLEE_CLOBBERS, ,                          \
621                       pre, post, ##__VA_ARGS__)
622
623
624
625 #define PVOP_CALL0(rettype, op)                                         \
626         __PVOP_CALL(rettype, op, "", "")
627 #define PVOP_VCALL0(op)                                                 \
628         __PVOP_VCALL(op, "", "")
629
630 #define PVOP_CALLEE0(rettype, op)                                       \
631         __PVOP_CALLEESAVE(rettype, op, "", "")
632 #define PVOP_VCALLEE0(op)                                               \
633         __PVOP_VCALLEESAVE(op, "", "")
634
635
636 #define PVOP_CALL1(rettype, op, arg1)                                   \
637         __PVOP_CALL(rettype, op, "", "", PVOP_CALL_ARG1(arg1))
638 #define PVOP_VCALL1(op, arg1)                                           \
639         __PVOP_VCALL(op, "", "", PVOP_CALL_ARG1(arg1))
640
641 #define PVOP_CALLEE1(rettype, op, arg1)                                 \
642         __PVOP_CALLEESAVE(rettype, op, "", "", PVOP_CALL_ARG1(arg1))
643 #define PVOP_VCALLEE1(op, arg1)                                         \
644         __PVOP_VCALLEESAVE(op, "", "", PVOP_CALL_ARG1(arg1))
645
646
647 #define PVOP_CALL2(rettype, op, arg1, arg2)                             \
648         __PVOP_CALL(rettype, op, "", "", PVOP_CALL_ARG1(arg1),          \
649                     PVOP_CALL_ARG2(arg2))
650 #define PVOP_VCALL2(op, arg1, arg2)                                     \
651         __PVOP_VCALL(op, "", "", PVOP_CALL_ARG1(arg1),                  \
652                      PVOP_CALL_ARG2(arg2))
653
654 #define PVOP_CALLEE2(rettype, op, arg1, arg2)                           \
655         __PVOP_CALLEESAVE(rettype, op, "", "", PVOP_CALL_ARG1(arg1),    \
656                           PVOP_CALL_ARG2(arg2))
657 #define PVOP_VCALLEE2(op, arg1, arg2)                                   \
658         __PVOP_VCALLEESAVE(op, "", "", PVOP_CALL_ARG1(arg1),            \
659                            PVOP_CALL_ARG2(arg2))
660
661
662 #define PVOP_CALL3(rettype, op, arg1, arg2, arg3)                       \
663         __PVOP_CALL(rettype, op, "", "", PVOP_CALL_ARG1(arg1),          \
664                     PVOP_CALL_ARG2(arg2), PVOP_CALL_ARG3(arg3))
665 #define PVOP_VCALL3(op, arg1, arg2, arg3)                               \
666         __PVOP_VCALL(op, "", "", PVOP_CALL_ARG1(arg1),                  \
667                      PVOP_CALL_ARG2(arg2), PVOP_CALL_ARG3(arg3))
668
669 /* This is the only difference in x86_64. We can make it much simpler */
670 #ifdef CONFIG_X86_32
671 #define PVOP_CALL4(rettype, op, arg1, arg2, arg3, arg4)                 \
672         __PVOP_CALL(rettype, op,                                        \
673                     "push %[_arg4];", "lea 4(%%esp),%%esp;",            \
674                     PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2),         \
675                     PVOP_CALL_ARG3(arg3), [_arg4] "mr" ((u32)(arg4)))
676 #define PVOP_VCALL4(op, arg1, arg2, arg3, arg4)                         \
677         __PVOP_VCALL(op,                                                \
678                     "push %[_arg4];", "lea 4(%%esp),%%esp;",            \
679                     "0" ((u32)(arg1)), "1" ((u32)(arg2)),               \
680                     "2" ((u32)(arg3)), [_arg4] "mr" ((u32)(arg4)))
681 #else
682 #define PVOP_CALL4(rettype, op, arg1, arg2, arg3, arg4)                 \
683         __PVOP_CALL(rettype, op, "", "",                                \
684                     PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2),         \
685                     PVOP_CALL_ARG3(arg3), PVOP_CALL_ARG4(arg4))
686 #define PVOP_VCALL4(op, arg1, arg2, arg3, arg4)                         \
687         __PVOP_VCALL(op, "", "",                                        \
688                      PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2),        \
689                      PVOP_CALL_ARG3(arg3), PVOP_CALL_ARG4(arg4))
690 #endif
691
692 static inline int paravirt_enabled(void)
693 {
694         return pv_info.paravirt_enabled;
695 }
696
697 static inline void load_sp0(struct tss_struct *tss,
698                              struct thread_struct *thread)
699 {
700         PVOP_VCALL2(pv_cpu_ops.load_sp0, tss, thread);
701 }
702
703 #define ARCH_SETUP                      pv_init_ops.arch_setup();
704 static inline unsigned long get_wallclock(void)
705 {
706         return PVOP_CALL0(unsigned long, pv_time_ops.get_wallclock);
707 }
708
709 static inline int set_wallclock(unsigned long nowtime)
710 {
711         return PVOP_CALL1(int, pv_time_ops.set_wallclock, nowtime);
712 }
713
714 static inline void (*choose_time_init(void))(void)
715 {
716         return pv_time_ops.time_init;
717 }
718
719 /* The paravirtualized CPUID instruction. */
720 static inline void __cpuid(unsigned int *eax, unsigned int *ebx,
721                            unsigned int *ecx, unsigned int *edx)
722 {
723         PVOP_VCALL4(pv_cpu_ops.cpuid, eax, ebx, ecx, edx);
724 }
725
726 /*
727  * These special macros can be used to get or set a debugging register
728  */
729 static inline unsigned long paravirt_get_debugreg(int reg)
730 {
731         return PVOP_CALL1(unsigned long, pv_cpu_ops.get_debugreg, reg);
732 }
733 #define get_debugreg(var, reg) var = paravirt_get_debugreg(reg)
734 static inline void set_debugreg(unsigned long val, int reg)
735 {
736         PVOP_VCALL2(pv_cpu_ops.set_debugreg, reg, val);
737 }
738
739 static inline void clts(void)
740 {
741         PVOP_VCALL0(pv_cpu_ops.clts);
742 }
743
744 static inline unsigned long read_cr0(void)
745 {
746         return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr0);
747 }
748
749 static inline void write_cr0(unsigned long x)
750 {
751         PVOP_VCALL1(pv_cpu_ops.write_cr0, x);
752 }
753
754 static inline unsigned long read_cr2(void)
755 {
756         return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr2);
757 }
758
759 static inline void write_cr2(unsigned long x)
760 {
761         PVOP_VCALL1(pv_mmu_ops.write_cr2, x);
762 }
763
764 static inline unsigned long read_cr3(void)
765 {
766         return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr3);
767 }
768
769 static inline void write_cr3(unsigned long x)
770 {
771         PVOP_VCALL1(pv_mmu_ops.write_cr3, x);
772 }
773
774 static inline unsigned long read_cr4(void)
775 {
776         return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr4);
777 }
778 static inline unsigned long read_cr4_safe(void)
779 {
780         return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr4_safe);
781 }
782
783 static inline void write_cr4(unsigned long x)
784 {
785         PVOP_VCALL1(pv_cpu_ops.write_cr4, x);
786 }
787
788 #ifdef CONFIG_X86_64
789 static inline unsigned long read_cr8(void)
790 {
791         return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr8);
792 }
793
794 static inline void write_cr8(unsigned long x)
795 {
796         PVOP_VCALL1(pv_cpu_ops.write_cr8, x);
797 }
798 #endif
799
800 static inline void raw_safe_halt(void)
801 {
802         PVOP_VCALL0(pv_irq_ops.safe_halt);
803 }
804
805 static inline void halt(void)
806 {
807         PVOP_VCALL0(pv_irq_ops.safe_halt);
808 }
809
810 static inline void wbinvd(void)
811 {
812         PVOP_VCALL0(pv_cpu_ops.wbinvd);
813 }
814
815 #define get_kernel_rpl()  (pv_info.kernel_rpl)
816
817 static inline u64 paravirt_read_msr(unsigned msr, int *err)
818 {
819         return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err);
820 }
821 static inline u64 paravirt_read_msr_amd(unsigned msr, int *err)
822 {
823         return PVOP_CALL2(u64, pv_cpu_ops.read_msr_amd, msr, err);
824 }
825 static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high)
826 {
827         return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high);
828 }
829
830 /* These should all do BUG_ON(_err), but our headers are too tangled. */
831 #define rdmsr(msr, val1, val2)                  \
832 do {                                            \
833         int _err;                               \
834         u64 _l = paravirt_read_msr(msr, &_err); \
835         val1 = (u32)_l;                         \
836         val2 = _l >> 32;                        \
837 } while (0)
838
839 #define wrmsr(msr, val1, val2)                  \
840 do {                                            \
841         paravirt_write_msr(msr, val1, val2);    \
842 } while (0)
843
844 #define rdmsrl(msr, val)                        \
845 do {                                            \
846         int _err;                               \
847         val = paravirt_read_msr(msr, &_err);    \
848 } while (0)
849
850 #define wrmsrl(msr, val)        wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32)
851 #define wrmsr_safe(msr, a, b)   paravirt_write_msr(msr, a, b)
852
853 /* rdmsr with exception handling */
854 #define rdmsr_safe(msr, a, b)                   \
855 ({                                              \
856         int _err;                               \
857         u64 _l = paravirt_read_msr(msr, &_err); \
858         (*a) = (u32)_l;                         \
859         (*b) = _l >> 32;                        \
860         _err;                                   \
861 })
862
863 static inline int rdmsrl_safe(unsigned msr, unsigned long long *p)
864 {
865         int err;
866
867         *p = paravirt_read_msr(msr, &err);
868         return err;
869 }
870 static inline int rdmsrl_amd_safe(unsigned msr, unsigned long long *p)
871 {
872         int err;
873
874         *p = paravirt_read_msr_amd(msr, &err);
875         return err;
876 }
877
878 static inline u64 paravirt_read_tsc(void)
879 {
880         return PVOP_CALL0(u64, pv_cpu_ops.read_tsc);
881 }
882
883 #define rdtscl(low)                             \
884 do {                                            \
885         u64 _l = paravirt_read_tsc();           \
886         low = (int)_l;                          \
887 } while (0)
888
889 #define rdtscll(val) (val = paravirt_read_tsc())
890
891 static inline unsigned long long paravirt_sched_clock(void)
892 {
893         return PVOP_CALL0(unsigned long long, pv_time_ops.sched_clock);
894 }
895 #define calibrate_tsc() (pv_time_ops.get_tsc_khz())
896
897 static inline unsigned long long paravirt_read_pmc(int counter)
898 {
899         return PVOP_CALL1(u64, pv_cpu_ops.read_pmc, counter);
900 }
901
902 #define rdpmc(counter, low, high)               \
903 do {                                            \
904         u64 _l = paravirt_read_pmc(counter);    \
905         low = (u32)_l;                          \
906         high = _l >> 32;                        \
907 } while (0)
908
909 static inline unsigned long long paravirt_rdtscp(unsigned int *aux)
910 {
911         return PVOP_CALL1(u64, pv_cpu_ops.read_tscp, aux);
912 }
913
914 #define rdtscp(low, high, aux)                          \
915 do {                                                    \
916         int __aux;                                      \
917         unsigned long __val = paravirt_rdtscp(&__aux);  \
918         (low) = (u32)__val;                             \
919         (high) = (u32)(__val >> 32);                    \
920         (aux) = __aux;                                  \
921 } while (0)
922
923 #define rdtscpll(val, aux)                              \
924 do {                                                    \
925         unsigned long __aux;                            \
926         val = paravirt_rdtscp(&__aux);                  \
927         (aux) = __aux;                                  \
928 } while (0)
929
930 static inline void paravirt_alloc_ldt(struct desc_struct *ldt, unsigned entries)
931 {
932         PVOP_VCALL2(pv_cpu_ops.alloc_ldt, ldt, entries);
933 }
934
935 static inline void paravirt_free_ldt(struct desc_struct *ldt, unsigned entries)
936 {
937         PVOP_VCALL2(pv_cpu_ops.free_ldt, ldt, entries);
938 }
939
940 static inline void load_TR_desc(void)
941 {
942         PVOP_VCALL0(pv_cpu_ops.load_tr_desc);
943 }
944 static inline void load_gdt(const struct desc_ptr *dtr)
945 {
946         PVOP_VCALL1(pv_cpu_ops.load_gdt, dtr);
947 }
948 static inline void load_idt(const struct desc_ptr *dtr)
949 {
950         PVOP_VCALL1(pv_cpu_ops.load_idt, dtr);
951 }
952 static inline void set_ldt(const void *addr, unsigned entries)
953 {
954         PVOP_VCALL2(pv_cpu_ops.set_ldt, addr, entries);
955 }
956 static inline void store_gdt(struct desc_ptr *dtr)
957 {
958         PVOP_VCALL1(pv_cpu_ops.store_gdt, dtr);
959 }
960 static inline void store_idt(struct desc_ptr *dtr)
961 {
962         PVOP_VCALL1(pv_cpu_ops.store_idt, dtr);
963 }
964 static inline unsigned long paravirt_store_tr(void)
965 {
966         return PVOP_CALL0(unsigned long, pv_cpu_ops.store_tr);
967 }
968 #define store_tr(tr)    ((tr) = paravirt_store_tr())
969 static inline void load_TLS(struct thread_struct *t, unsigned cpu)
970 {
971         PVOP_VCALL2(pv_cpu_ops.load_tls, t, cpu);
972 }
973
974 #ifdef CONFIG_X86_64
975 static inline void load_gs_index(unsigned int gs)
976 {
977         PVOP_VCALL1(pv_cpu_ops.load_gs_index, gs);
978 }
979 #endif
980
981 static inline void write_ldt_entry(struct desc_struct *dt, int entry,
982                                    const void *desc)
983 {
984         PVOP_VCALL3(pv_cpu_ops.write_ldt_entry, dt, entry, desc);
985 }
986
987 static inline void write_gdt_entry(struct desc_struct *dt, int entry,
988                                    void *desc, int type)
989 {
990         PVOP_VCALL4(pv_cpu_ops.write_gdt_entry, dt, entry, desc, type);
991 }
992
993 static inline void write_idt_entry(gate_desc *dt, int entry, const gate_desc *g)
994 {
995         PVOP_VCALL3(pv_cpu_ops.write_idt_entry, dt, entry, g);
996 }
997 static inline void set_iopl_mask(unsigned mask)
998 {
999         PVOP_VCALL1(pv_cpu_ops.set_iopl_mask, mask);
1000 }
1001
1002 /* The paravirtualized I/O functions */
1003 static inline void slow_down_io(void)
1004 {
1005         pv_cpu_ops.io_delay();
1006 #ifdef REALLY_SLOW_IO
1007         pv_cpu_ops.io_delay();
1008         pv_cpu_ops.io_delay();
1009         pv_cpu_ops.io_delay();
1010 #endif
1011 }
1012
1013 #ifdef CONFIG_X86_LOCAL_APIC
1014 static inline void setup_boot_clock(void)
1015 {
1016         PVOP_VCALL0(pv_apic_ops.setup_boot_clock);
1017 }
1018
1019 static inline void setup_secondary_clock(void)
1020 {
1021         PVOP_VCALL0(pv_apic_ops.setup_secondary_clock);
1022 }
1023 #endif
1024
1025 static inline void paravirt_post_allocator_init(void)
1026 {
1027         if (pv_init_ops.post_allocator_init)
1028                 (*pv_init_ops.post_allocator_init)();
1029 }
1030
1031 static inline void paravirt_pagetable_setup_start(pgd_t *base)
1032 {
1033         (*pv_mmu_ops.pagetable_setup_start)(base);
1034 }
1035
1036 static inline void paravirt_pagetable_setup_done(pgd_t *base)
1037 {
1038         (*pv_mmu_ops.pagetable_setup_done)(base);
1039 }
1040
1041 #ifdef CONFIG_SMP
1042 static inline void startup_ipi_hook(int phys_apicid, unsigned long start_eip,
1043                                     unsigned long start_esp)
1044 {
1045         PVOP_VCALL3(pv_apic_ops.startup_ipi_hook,
1046                     phys_apicid, start_eip, start_esp);
1047 }
1048 #endif
1049
1050 static inline void paravirt_activate_mm(struct mm_struct *prev,
1051                                         struct mm_struct *next)
1052 {
1053         PVOP_VCALL2(pv_mmu_ops.activate_mm, prev, next);
1054 }
1055
1056 static inline void arch_dup_mmap(struct mm_struct *oldmm,
1057                                  struct mm_struct *mm)
1058 {
1059         PVOP_VCALL2(pv_mmu_ops.dup_mmap, oldmm, mm);
1060 }
1061
1062 static inline void arch_exit_mmap(struct mm_struct *mm)
1063 {
1064         PVOP_VCALL1(pv_mmu_ops.exit_mmap, mm);
1065 }
1066
1067 static inline void __flush_tlb(void)
1068 {
1069         PVOP_VCALL0(pv_mmu_ops.flush_tlb_user);
1070 }
1071 static inline void __flush_tlb_global(void)
1072 {
1073         PVOP_VCALL0(pv_mmu_ops.flush_tlb_kernel);
1074 }
1075 static inline void __flush_tlb_single(unsigned long addr)
1076 {
1077         PVOP_VCALL1(pv_mmu_ops.flush_tlb_single, addr);
1078 }
1079
1080 static inline void flush_tlb_others(const struct cpumask *cpumask,
1081                                     struct mm_struct *mm,
1082                                     unsigned long va)
1083 {
1084         PVOP_VCALL3(pv_mmu_ops.flush_tlb_others, cpumask, mm, va);
1085 }
1086
1087 static inline int paravirt_pgd_alloc(struct mm_struct *mm)
1088 {
1089         return PVOP_CALL1(int, pv_mmu_ops.pgd_alloc, mm);
1090 }
1091
1092 static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd)
1093 {
1094         PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd);
1095 }
1096
1097 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn)
1098 {
1099         PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn);
1100 }
1101 static inline void paravirt_release_pte(unsigned long pfn)
1102 {
1103         PVOP_VCALL1(pv_mmu_ops.release_pte, pfn);
1104 }
1105
1106 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn)
1107 {
1108         PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn);
1109 }
1110
1111 static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn,
1112                                             unsigned long start, unsigned long count)
1113 {
1114         PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count);
1115 }
1116 static inline void paravirt_release_pmd(unsigned long pfn)
1117 {
1118         PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn);
1119 }
1120
1121 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn)
1122 {
1123         PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn);
1124 }
1125 static inline void paravirt_release_pud(unsigned long pfn)
1126 {
1127         PVOP_VCALL1(pv_mmu_ops.release_pud, pfn);
1128 }
1129
1130 #ifdef CONFIG_HIGHPTE
1131 static inline void *kmap_atomic_pte(struct page *page, enum km_type type)
1132 {
1133         unsigned long ret;
1134         ret = PVOP_CALL2(unsigned long, pv_mmu_ops.kmap_atomic_pte, page, type);
1135         return (void *)ret;
1136 }
1137 #endif
1138
1139 static inline void pte_update(struct mm_struct *mm, unsigned long addr,
1140                               pte_t *ptep)
1141 {
1142         PVOP_VCALL3(pv_mmu_ops.pte_update, mm, addr, ptep);
1143 }
1144
1145 static inline void pte_update_defer(struct mm_struct *mm, unsigned long addr,
1146                                     pte_t *ptep)
1147 {
1148         PVOP_VCALL3(pv_mmu_ops.pte_update_defer, mm, addr, ptep);
1149 }
1150
1151 static inline pte_t __pte(pteval_t val)
1152 {
1153         pteval_t ret;
1154
1155         if (sizeof(pteval_t) > sizeof(long))
1156                 ret = PVOP_CALLEE2(pteval_t,
1157                                    pv_mmu_ops.make_pte,
1158                                    val, (u64)val >> 32);
1159         else
1160                 ret = PVOP_CALLEE1(pteval_t,
1161                                    pv_mmu_ops.make_pte,
1162                                    val);
1163
1164         return (pte_t) { .pte = ret };
1165 }
1166
1167 static inline pteval_t pte_val(pte_t pte)
1168 {
1169         pteval_t ret;
1170
1171         if (sizeof(pteval_t) > sizeof(long))
1172                 ret = PVOP_CALLEE2(pteval_t, pv_mmu_ops.pte_val,
1173                                    pte.pte, (u64)pte.pte >> 32);
1174         else
1175                 ret = PVOP_CALLEE1(pteval_t, pv_mmu_ops.pte_val,
1176                                    pte.pte);
1177
1178         return ret;
1179 }
1180
1181 static inline pgd_t __pgd(pgdval_t val)
1182 {
1183         pgdval_t ret;
1184
1185         if (sizeof(pgdval_t) > sizeof(long))
1186                 ret = PVOP_CALLEE2(pgdval_t, pv_mmu_ops.make_pgd,
1187                                    val, (u64)val >> 32);
1188         else
1189                 ret = PVOP_CALLEE1(pgdval_t, pv_mmu_ops.make_pgd,
1190                                    val);
1191
1192         return (pgd_t) { ret };
1193 }
1194
1195 static inline pgdval_t pgd_val(pgd_t pgd)
1196 {
1197         pgdval_t ret;
1198
1199         if (sizeof(pgdval_t) > sizeof(long))
1200                 ret =  PVOP_CALLEE2(pgdval_t, pv_mmu_ops.pgd_val,
1201                                     pgd.pgd, (u64)pgd.pgd >> 32);
1202         else
1203                 ret =  PVOP_CALLEE1(pgdval_t, pv_mmu_ops.pgd_val,
1204                                     pgd.pgd);
1205
1206         return ret;
1207 }
1208
1209 #define  __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION
1210 static inline pte_t ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr,
1211                                            pte_t *ptep)
1212 {
1213         pteval_t ret;
1214
1215         ret = PVOP_CALL3(pteval_t, pv_mmu_ops.ptep_modify_prot_start,
1216                          mm, addr, ptep);
1217
1218         return (pte_t) { .pte = ret };
1219 }
1220
1221 static inline void ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr,
1222                                            pte_t *ptep, pte_t pte)
1223 {
1224         if (sizeof(pteval_t) > sizeof(long))
1225                 /* 5 arg words */
1226                 pv_mmu_ops.ptep_modify_prot_commit(mm, addr, ptep, pte);
1227         else
1228                 PVOP_VCALL4(pv_mmu_ops.ptep_modify_prot_commit,
1229                             mm, addr, ptep, pte.pte);
1230 }
1231
1232 static inline void set_pte(pte_t *ptep, pte_t pte)
1233 {
1234         if (sizeof(pteval_t) > sizeof(long))
1235                 PVOP_VCALL3(pv_mmu_ops.set_pte, ptep,
1236                             pte.pte, (u64)pte.pte >> 32);
1237         else
1238                 PVOP_VCALL2(pv_mmu_ops.set_pte, ptep,
1239                             pte.pte);
1240 }
1241
1242 static inline void set_pte_at(struct mm_struct *mm, unsigned long addr,
1243                               pte_t *ptep, pte_t pte)
1244 {
1245         if (sizeof(pteval_t) > sizeof(long))
1246                 /* 5 arg words */
1247                 pv_mmu_ops.set_pte_at(mm, addr, ptep, pte);
1248         else
1249                 PVOP_VCALL4(pv_mmu_ops.set_pte_at, mm, addr, ptep, pte.pte);
1250 }
1251
1252 static inline void set_pmd(pmd_t *pmdp, pmd_t pmd)
1253 {
1254         pmdval_t val = native_pmd_val(pmd);
1255
1256         if (sizeof(pmdval_t) > sizeof(long))
1257                 PVOP_VCALL3(pv_mmu_ops.set_pmd, pmdp, val, (u64)val >> 32);
1258         else
1259                 PVOP_VCALL2(pv_mmu_ops.set_pmd, pmdp, val);
1260 }
1261
1262 #if PAGETABLE_LEVELS >= 3
1263 static inline pmd_t __pmd(pmdval_t val)
1264 {
1265         pmdval_t ret;
1266
1267         if (sizeof(pmdval_t) > sizeof(long))
1268                 ret = PVOP_CALLEE2(pmdval_t, pv_mmu_ops.make_pmd,
1269                                    val, (u64)val >> 32);
1270         else
1271                 ret = PVOP_CALLEE1(pmdval_t, pv_mmu_ops.make_pmd,
1272                                    val);
1273
1274         return (pmd_t) { ret };
1275 }
1276
1277 static inline pmdval_t pmd_val(pmd_t pmd)
1278 {
1279         pmdval_t ret;
1280
1281         if (sizeof(pmdval_t) > sizeof(long))
1282                 ret =  PVOP_CALLEE2(pmdval_t, pv_mmu_ops.pmd_val,
1283                                     pmd.pmd, (u64)pmd.pmd >> 32);
1284         else
1285                 ret =  PVOP_CALLEE1(pmdval_t, pv_mmu_ops.pmd_val,
1286                                     pmd.pmd);
1287
1288         return ret;
1289 }
1290
1291 static inline void set_pud(pud_t *pudp, pud_t pud)
1292 {
1293         pudval_t val = native_pud_val(pud);
1294
1295         if (sizeof(pudval_t) > sizeof(long))
1296                 PVOP_VCALL3(pv_mmu_ops.set_pud, pudp,
1297                             val, (u64)val >> 32);
1298         else
1299                 PVOP_VCALL2(pv_mmu_ops.set_pud, pudp,
1300                             val);
1301 }
1302 #if PAGETABLE_LEVELS == 4
1303 static inline pud_t __pud(pudval_t val)
1304 {
1305         pudval_t ret;
1306
1307         if (sizeof(pudval_t) > sizeof(long))
1308                 ret = PVOP_CALLEE2(pudval_t, pv_mmu_ops.make_pud,
1309                                    val, (u64)val >> 32);
1310         else
1311                 ret = PVOP_CALLEE1(pudval_t, pv_mmu_ops.make_pud,
1312                                    val);
1313
1314         return (pud_t) { ret };
1315 }
1316
1317 static inline pudval_t pud_val(pud_t pud)
1318 {
1319         pudval_t ret;
1320
1321         if (sizeof(pudval_t) > sizeof(long))
1322                 ret =  PVOP_CALLEE2(pudval_t, pv_mmu_ops.pud_val,
1323                                     pud.pud, (u64)pud.pud >> 32);
1324         else
1325                 ret =  PVOP_CALLEE1(pudval_t, pv_mmu_ops.pud_val,
1326                                     pud.pud);
1327
1328         return ret;
1329 }
1330
1331 static inline void set_pgd(pgd_t *pgdp, pgd_t pgd)
1332 {
1333         pgdval_t val = native_pgd_val(pgd);
1334
1335         if (sizeof(pgdval_t) > sizeof(long))
1336                 PVOP_VCALL3(pv_mmu_ops.set_pgd, pgdp,
1337                             val, (u64)val >> 32);
1338         else
1339                 PVOP_VCALL2(pv_mmu_ops.set_pgd, pgdp,
1340                             val);
1341 }
1342
1343 static inline void pgd_clear(pgd_t *pgdp)
1344 {
1345         set_pgd(pgdp, __pgd(0));
1346 }
1347
1348 static inline void pud_clear(pud_t *pudp)
1349 {
1350         set_pud(pudp, __pud(0));
1351 }
1352
1353 #endif  /* PAGETABLE_LEVELS == 4 */
1354
1355 #endif  /* PAGETABLE_LEVELS >= 3 */
1356
1357 #ifdef CONFIG_X86_PAE
1358 /* Special-case pte-setting operations for PAE, which can't update a
1359    64-bit pte atomically */
1360 static inline void set_pte_atomic(pte_t *ptep, pte_t pte)
1361 {
1362         PVOP_VCALL3(pv_mmu_ops.set_pte_atomic, ptep,
1363                     pte.pte, pte.pte >> 32);
1364 }
1365
1366 static inline void pte_clear(struct mm_struct *mm, unsigned long addr,
1367                              pte_t *ptep)
1368 {
1369         PVOP_VCALL3(pv_mmu_ops.pte_clear, mm, addr, ptep);
1370 }
1371
1372 static inline void pmd_clear(pmd_t *pmdp)
1373 {
1374         PVOP_VCALL1(pv_mmu_ops.pmd_clear, pmdp);
1375 }
1376 #else  /* !CONFIG_X86_PAE */
1377 static inline void set_pte_atomic(pte_t *ptep, pte_t pte)
1378 {
1379         set_pte(ptep, pte);
1380 }
1381
1382 static inline void pte_clear(struct mm_struct *mm, unsigned long addr,
1383                              pte_t *ptep)
1384 {
1385         set_pte_at(mm, addr, ptep, __pte(0));
1386 }
1387
1388 static inline void pmd_clear(pmd_t *pmdp)
1389 {
1390         set_pmd(pmdp, __pmd(0));
1391 }
1392 #endif  /* CONFIG_X86_PAE */
1393
1394 /* Lazy mode for batching updates / context switch */
1395 enum paravirt_lazy_mode {
1396         PARAVIRT_LAZY_NONE,
1397         PARAVIRT_LAZY_MMU,
1398         PARAVIRT_LAZY_CPU,
1399 };
1400
1401 enum paravirt_lazy_mode paravirt_get_lazy_mode(void);
1402 void paravirt_enter_lazy_cpu(void);
1403 void paravirt_leave_lazy_cpu(void);
1404 void paravirt_enter_lazy_mmu(void);
1405 void paravirt_leave_lazy_mmu(void);
1406 void paravirt_leave_lazy(enum paravirt_lazy_mode mode);
1407
1408 #define  __HAVE_ARCH_ENTER_LAZY_CPU_MODE
1409 static inline void arch_enter_lazy_cpu_mode(void)
1410 {
1411         PVOP_VCALL0(pv_cpu_ops.lazy_mode.enter);
1412 }
1413
1414 static inline void arch_leave_lazy_cpu_mode(void)
1415 {
1416         PVOP_VCALL0(pv_cpu_ops.lazy_mode.leave);
1417 }
1418
1419 void arch_flush_lazy_cpu_mode(void);
1420
1421 #define  __HAVE_ARCH_ENTER_LAZY_MMU_MODE
1422 static inline void arch_enter_lazy_mmu_mode(void)
1423 {
1424         PVOP_VCALL0(pv_mmu_ops.lazy_mode.enter);
1425 }
1426
1427 static inline void arch_leave_lazy_mmu_mode(void)
1428 {
1429         PVOP_VCALL0(pv_mmu_ops.lazy_mode.leave);
1430 }
1431
1432 void arch_flush_lazy_mmu_mode(void);
1433
1434 static inline void __set_fixmap(unsigned /* enum fixed_addresses */ idx,
1435                                 unsigned long phys, pgprot_t flags)
1436 {
1437         pv_mmu_ops.set_fixmap(idx, phys, flags);
1438 }
1439
1440 void _paravirt_nop(void);
1441 u32 _paravirt_ident_32(u32);
1442 u64 _paravirt_ident_64(u64);
1443
1444 #define paravirt_nop    ((void *)_paravirt_nop)
1445
1446 #ifdef CONFIG_SMP
1447
1448 static inline int __raw_spin_is_locked(struct raw_spinlock *lock)
1449 {
1450         return PVOP_CALL1(int, pv_lock_ops.spin_is_locked, lock);
1451 }
1452
1453 static inline int __raw_spin_is_contended(struct raw_spinlock *lock)
1454 {
1455         return PVOP_CALL1(int, pv_lock_ops.spin_is_contended, lock);
1456 }
1457 #define __raw_spin_is_contended __raw_spin_is_contended
1458
1459 static __always_inline void __raw_spin_lock(struct raw_spinlock *lock)
1460 {
1461         PVOP_VCALL1(pv_lock_ops.spin_lock, lock);
1462 }
1463
1464 static __always_inline void __raw_spin_lock_flags(struct raw_spinlock *lock,
1465                                                   unsigned long flags)
1466 {
1467         PVOP_VCALL2(pv_lock_ops.spin_lock_flags, lock, flags);
1468 }
1469
1470 static __always_inline int __raw_spin_trylock(struct raw_spinlock *lock)
1471 {
1472         return PVOP_CALL1(int, pv_lock_ops.spin_trylock, lock);
1473 }
1474
1475 static __always_inline void __raw_spin_unlock(struct raw_spinlock *lock)
1476 {
1477         PVOP_VCALL1(pv_lock_ops.spin_unlock, lock);
1478 }
1479
1480 #endif
1481
1482 /* These all sit in the .parainstructions section to tell us what to patch. */
1483 struct paravirt_patch_site {
1484         u8 *instr;              /* original instructions */
1485         u8 instrtype;           /* type of this instruction */
1486         u8 len;                 /* length of original instruction */
1487         u16 clobbers;           /* what registers you may clobber */
1488 };
1489
1490 extern struct paravirt_patch_site __parainstructions[],
1491         __parainstructions_end[];
1492
1493 #ifdef CONFIG_X86_32
1494 #define PV_SAVE_REGS "pushl %ecx; pushl %edx;"
1495 #define PV_RESTORE_REGS "popl %edx; popl %ecx;"
1496
1497 /* save and restore all caller-save registers, except return value */
1498 #define PV_SAVE_ALL_CALLER_REGS         "pushl %ecx;"
1499 #define PV_RESTORE_ALL_CALLER_REGS      "popl  %ecx;"
1500
1501 #define PV_FLAGS_ARG "0"
1502 #define PV_EXTRA_CLOBBERS
1503 #define PV_VEXTRA_CLOBBERS
1504 #else
1505 /* save and restore all caller-save registers, except return value */
1506 #define PV_SAVE_ALL_CALLER_REGS                                         \
1507         "push %rcx;"                                                    \
1508         "push %rdx;"                                                    \
1509         "push %rsi;"                                                    \
1510         "push %rdi;"                                                    \
1511         "push %r8;"                                                     \
1512         "push %r9;"                                                     \
1513         "push %r10;"                                                    \
1514         "push %r11;"
1515 #define PV_RESTORE_ALL_CALLER_REGS                                      \
1516         "pop %r11;"                                                     \
1517         "pop %r10;"                                                     \
1518         "pop %r9;"                                                      \
1519         "pop %r8;"                                                      \
1520         "pop %rdi;"                                                     \
1521         "pop %rsi;"                                                     \
1522         "pop %rdx;"                                                     \
1523         "pop %rcx;"
1524
1525 /* We save some registers, but all of them, that's too much. We clobber all
1526  * caller saved registers but the argument parameter */
1527 #define PV_SAVE_REGS "pushq %%rdi;"
1528 #define PV_RESTORE_REGS "popq %%rdi;"
1529 #define PV_EXTRA_CLOBBERS EXTRA_CLOBBERS, "rcx" , "rdx", "rsi"
1530 #define PV_VEXTRA_CLOBBERS EXTRA_CLOBBERS, "rdi", "rcx" , "rdx", "rsi"
1531 #define PV_FLAGS_ARG "D"
1532 #endif
1533
1534 /*
1535  * Generate a thunk around a function which saves all caller-save
1536  * registers except for the return value.  This allows C functions to
1537  * be called from assembler code where fewer than normal registers are
1538  * available.  It may also help code generation around calls from C
1539  * code if the common case doesn't use many registers.
1540  *
1541  * When a callee is wrapped in a thunk, the caller can assume that all
1542  * arg regs and all scratch registers are preserved across the
1543  * call. The return value in rax/eax will not be saved, even for void
1544  * functions.
1545  */
1546 #define PV_CALLEE_SAVE_REGS_THUNK(func)                                 \
1547         extern typeof(func) __raw_callee_save_##func;                   \
1548         static void *__##func##__ __used = func;                        \
1549                                                                         \
1550         asm(".pushsection .text;"                                       \
1551             "__raw_callee_save_" #func ": "                             \
1552             PV_SAVE_ALL_CALLER_REGS                                     \
1553             "call " #func ";"                                           \
1554             PV_RESTORE_ALL_CALLER_REGS                                  \
1555             "ret;"                                                      \
1556             ".popsection")
1557
1558 /* Get a reference to a callee-save function */
1559 #define PV_CALLEE_SAVE(func)                                            \
1560         ((struct paravirt_callee_save) { __raw_callee_save_##func })
1561
1562 /* Promise that "func" already uses the right calling convention */
1563 #define __PV_IS_CALLEE_SAVE(func)                       \
1564         ((struct paravirt_callee_save) { func })
1565
1566 static inline unsigned long __raw_local_save_flags(void)
1567 {
1568         unsigned long f;
1569
1570         asm volatile(paravirt_alt(PARAVIRT_CALL)
1571                      : "=a"(f)
1572                      : paravirt_type(pv_irq_ops.save_fl),
1573                        paravirt_clobber(CLBR_EAX)
1574                      : "memory", "cc");
1575         return f;
1576 }
1577
1578 static inline void raw_local_irq_restore(unsigned long f)
1579 {
1580         asm volatile(paravirt_alt(PARAVIRT_CALL)
1581                      : "=a"(f)
1582                      : PV_FLAGS_ARG(f),
1583                        paravirt_type(pv_irq_ops.restore_fl),
1584                        paravirt_clobber(CLBR_EAX)
1585                      : "memory", "cc");
1586 }
1587
1588 static inline void raw_local_irq_disable(void)
1589 {
1590         asm volatile(paravirt_alt(PARAVIRT_CALL)
1591                      :
1592                      : paravirt_type(pv_irq_ops.irq_disable),
1593                        paravirt_clobber(CLBR_EAX)
1594                      : "memory", "eax", "cc");
1595 }
1596
1597 static inline void raw_local_irq_enable(void)
1598 {
1599         asm volatile(paravirt_alt(PARAVIRT_CALL)
1600                      :
1601                      : paravirt_type(pv_irq_ops.irq_enable),
1602                        paravirt_clobber(CLBR_EAX)
1603                      : "memory", "eax", "cc");
1604 }
1605
1606 static inline unsigned long __raw_local_irq_save(void)
1607 {
1608         unsigned long f;
1609
1610         f = __raw_local_save_flags();
1611         raw_local_irq_disable();
1612         return f;
1613 }
1614
1615
1616 /* Make sure as little as possible of this mess escapes. */
1617 #undef PARAVIRT_CALL
1618 #undef __PVOP_CALL
1619 #undef __PVOP_VCALL
1620 #undef PVOP_VCALL0
1621 #undef PVOP_CALL0
1622 #undef PVOP_VCALL1
1623 #undef PVOP_CALL1
1624 #undef PVOP_VCALL2
1625 #undef PVOP_CALL2
1626 #undef PVOP_VCALL3
1627 #undef PVOP_CALL3
1628 #undef PVOP_VCALL4
1629 #undef PVOP_CALL4
1630
1631 #else  /* __ASSEMBLY__ */
1632
1633 #define _PVSITE(ptype, clobbers, ops, word, algn)       \
1634 771:;                                           \
1635         ops;                                    \
1636 772:;                                           \
1637         .pushsection .parainstructions,"a";     \
1638          .align algn;                           \
1639          word 771b;                             \
1640          .byte ptype;                           \
1641          .byte 772b-771b;                       \
1642          .short clobbers;                       \
1643         .popsection
1644
1645
1646 #define COND_PUSH(set, mask, reg)                       \
1647         .if ((~(set)) & mask); push %reg; .endif
1648 #define COND_POP(set, mask, reg)                        \
1649         .if ((~(set)) & mask); pop %reg; .endif
1650
1651 #ifdef CONFIG_X86_64
1652
1653 #define PV_SAVE_REGS(set)                       \
1654         COND_PUSH(set, CLBR_RAX, rax);          \
1655         COND_PUSH(set, CLBR_RCX, rcx);          \
1656         COND_PUSH(set, CLBR_RDX, rdx);          \
1657         COND_PUSH(set, CLBR_RSI, rsi);          \
1658         COND_PUSH(set, CLBR_RDI, rdi);          \
1659         COND_PUSH(set, CLBR_R8, r8);            \
1660         COND_PUSH(set, CLBR_R9, r9);            \
1661         COND_PUSH(set, CLBR_R10, r10);          \
1662         COND_PUSH(set, CLBR_R11, r11)
1663 #define PV_RESTORE_REGS(set)                    \
1664         COND_POP(set, CLBR_R11, r11);           \
1665         COND_POP(set, CLBR_R10, r10);           \
1666         COND_POP(set, CLBR_R9, r9);             \
1667         COND_POP(set, CLBR_R8, r8);             \
1668         COND_POP(set, CLBR_RDI, rdi);           \
1669         COND_POP(set, CLBR_RSI, rsi);           \
1670         COND_POP(set, CLBR_RDX, rdx);           \
1671         COND_POP(set, CLBR_RCX, rcx);           \
1672         COND_POP(set, CLBR_RAX, rax)
1673
1674 #define PARA_PATCH(struct, off)        ((PARAVIRT_PATCH_##struct + (off)) / 8)
1675 #define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .quad, 8)
1676 #define PARA_INDIRECT(addr)     *addr(%rip)
1677 #else
1678 #define PV_SAVE_REGS(set)                       \
1679         COND_PUSH(set, CLBR_EAX, eax);          \
1680         COND_PUSH(set, CLBR_EDI, edi);          \
1681         COND_PUSH(set, CLBR_ECX, ecx);          \
1682         COND_PUSH(set, CLBR_EDX, edx)
1683 #define PV_RESTORE_REGS(set)                    \
1684         COND_POP(set, CLBR_EDX, edx);           \
1685         COND_POP(set, CLBR_ECX, ecx);           \
1686         COND_POP(set, CLBR_EDI, edi);           \
1687         COND_POP(set, CLBR_EAX, eax)
1688
1689 #define PARA_PATCH(struct, off)        ((PARAVIRT_PATCH_##struct + (off)) / 4)
1690 #define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .long, 4)
1691 #define PARA_INDIRECT(addr)     *%cs:addr
1692 #endif
1693
1694 #define INTERRUPT_RETURN                                                \
1695         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_iret), CLBR_NONE,       \
1696                   jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_iret))
1697
1698 #define DISABLE_INTERRUPTS(clobbers)                                    \
1699         PARA_SITE(PARA_PATCH(pv_irq_ops, PV_IRQ_irq_disable), clobbers, \
1700                   PV_SAVE_REGS(clobbers | CLBR_CALLEE_SAVE);            \
1701                   call PARA_INDIRECT(pv_irq_ops+PV_IRQ_irq_disable);    \
1702                   PV_RESTORE_REGS(clobbers | CLBR_CALLEE_SAVE);)
1703
1704 #define ENABLE_INTERRUPTS(clobbers)                                     \
1705         PARA_SITE(PARA_PATCH(pv_irq_ops, PV_IRQ_irq_enable), clobbers,  \
1706                   PV_SAVE_REGS(clobbers | CLBR_CALLEE_SAVE);            \
1707                   call PARA_INDIRECT(pv_irq_ops+PV_IRQ_irq_enable);     \
1708                   PV_RESTORE_REGS(clobbers | CLBR_CALLEE_SAVE);)
1709
1710 #define USERGS_SYSRET32                                                 \
1711         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_usergs_sysret32),       \
1712                   CLBR_NONE,                                            \
1713                   jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_usergs_sysret32))
1714
1715 #ifdef CONFIG_X86_32
1716 #define GET_CR0_INTO_EAX                                \
1717         push %ecx; push %edx;                           \
1718         call PARA_INDIRECT(pv_cpu_ops+PV_CPU_read_cr0); \
1719         pop %edx; pop %ecx
1720
1721 #define ENABLE_INTERRUPTS_SYSEXIT                                       \
1722         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_irq_enable_sysexit),    \
1723                   CLBR_NONE,                                            \
1724                   jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_irq_enable_sysexit))
1725
1726
1727 #else   /* !CONFIG_X86_32 */
1728
1729 /*
1730  * If swapgs is used while the userspace stack is still current,
1731  * there's no way to call a pvop.  The PV replacement *must* be
1732  * inlined, or the swapgs instruction must be trapped and emulated.
1733  */
1734 #define SWAPGS_UNSAFE_STACK                                             \
1735         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_swapgs), CLBR_NONE,     \
1736                   swapgs)
1737
1738 /*
1739  * Note: swapgs is very special, and in practise is either going to be
1740  * implemented with a single "swapgs" instruction or something very
1741  * special.  Either way, we don't need to save any registers for
1742  * it.
1743  */
1744 #define SWAPGS                                                          \
1745         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_swapgs), CLBR_NONE,     \
1746                   call PARA_INDIRECT(pv_cpu_ops+PV_CPU_swapgs)          \
1747                  )
1748
1749 #define GET_CR2_INTO_RCX                                \
1750         call PARA_INDIRECT(pv_mmu_ops+PV_MMU_read_cr2); \
1751         movq %rax, %rcx;                                \
1752         xorq %rax, %rax;
1753
1754 #define PARAVIRT_ADJUST_EXCEPTION_FRAME                                 \
1755         PARA_SITE(PARA_PATCH(pv_irq_ops, PV_IRQ_adjust_exception_frame), \
1756                   CLBR_NONE,                                            \
1757                   call PARA_INDIRECT(pv_irq_ops+PV_IRQ_adjust_exception_frame))
1758
1759 #define USERGS_SYSRET64                                                 \
1760         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_usergs_sysret64),       \
1761                   CLBR_NONE,                                            \
1762                   jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_usergs_sysret64))
1763
1764 #define ENABLE_INTERRUPTS_SYSEXIT32                                     \
1765         PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_irq_enable_sysexit),    \
1766                   CLBR_NONE,                                            \
1767                   jmp PARA_INDIRECT(pv_cpu_ops+PV_CPU_irq_enable_sysexit))
1768 #endif  /* CONFIG_X86_32 */
1769
1770 #endif /* __ASSEMBLY__ */
1771 #endif /* CONFIG_PARAVIRT */
1772 #endif /* _ASM_X86_PARAVIRT_H */