blob: 8aa05583bc42dec102b3fffb63d2fa6a828eee4c [file] [log] [blame]
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +01001#include <asm/paravirt.h>
2#include <asm/asm-offsets.h>
Glauber de Oliveira Costa8a650ce2008-01-30 13:33:19 +01003#include <linux/stringify.h>
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +01004
5DEF_NATIVE(pv_irq_ops, irq_disable, "cli");
6DEF_NATIVE(pv_irq_ops, irq_enable, "sti");
7DEF_NATIVE(pv_irq_ops, restore_fl, "pushq %rdi; popfq");
8DEF_NATIVE(pv_irq_ops, save_fl, "pushfq; popq %rax");
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +01009DEF_NATIVE(pv_mmu_ops, read_cr2, "movq %cr2, %rax");
10DEF_NATIVE(pv_mmu_ops, read_cr3, "movq %cr3, %rax");
11DEF_NATIVE(pv_mmu_ops, write_cr3, "movq %rdi, %cr3");
12DEF_NATIVE(pv_mmu_ops, flush_tlb_single, "invlpg (%rdi)");
13DEF_NATIVE(pv_cpu_ops, clts, "clts");
14DEF_NATIVE(pv_cpu_ops, wbinvd, "wbinvd");
15
Jeremy Fitzhardinge2be29982008-06-25 00:19:28 -040016DEF_NATIVE(pv_cpu_ops, irq_enable_sysexit, "swapgs; sti; sysexit");
17DEF_NATIVE(pv_cpu_ops, usergs_sysret64, "swapgs; sysretq");
18DEF_NATIVE(pv_cpu_ops, usergs_sysret32, "swapgs; sysretl");
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +010019DEF_NATIVE(pv_cpu_ops, swapgs, "swapgs");
20
Jeremy Fitzhardinge41edafd2009-01-28 14:35:02 -080021DEF_NATIVE(, mov32, "mov %edi, %eax");
22DEF_NATIVE(, mov64, "mov %rdi, %rax");
23
Ingo Molnar62c7a1e2015-05-11 09:47:23 +020024#if defined(CONFIG_PARAVIRT_SPINLOCKS) && defined(CONFIG_QUEUED_SPINLOCKS)
Peter Zijlstra (Intel)f233f7f2015-04-24 14:56:38 -040025DEF_NATIVE(pv_lock_ops, queued_spin_unlock, "movb $0, (%rdi)");
26#endif
27
Jeremy Fitzhardinge41edafd2009-01-28 14:35:02 -080028unsigned paravirt_patch_ident_32(void *insnbuf, unsigned len)
29{
30 return paravirt_patch_insns(insnbuf, len,
31 start__mov32, end__mov32);
32}
33
34unsigned paravirt_patch_ident_64(void *insnbuf, unsigned len)
35{
36 return paravirt_patch_insns(insnbuf, len,
37 start__mov64, end__mov64);
38}
39
Peter Zijlstra (Intel)f233f7f2015-04-24 14:56:38 -040040extern bool pv_is_native_spin_unlock(void);
41
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +010042unsigned native_patch(u8 type, u16 clobbers, void *ibuf,
43 unsigned long addr, unsigned len)
44{
45 const unsigned char *start, *end;
46 unsigned ret;
47
48#define PATCH_SITE(ops, x) \
49 case PARAVIRT_PATCH(ops.x): \
50 start = start_##ops##_##x; \
51 end = end_##ops##_##x; \
52 goto patch_site
53 switch(type) {
54 PATCH_SITE(pv_irq_ops, restore_fl);
55 PATCH_SITE(pv_irq_ops, save_fl);
56 PATCH_SITE(pv_irq_ops, irq_enable);
57 PATCH_SITE(pv_irq_ops, irq_disable);
Jeremy Fitzhardinge2be29982008-06-25 00:19:28 -040058 PATCH_SITE(pv_cpu_ops, usergs_sysret32);
59 PATCH_SITE(pv_cpu_ops, usergs_sysret64);
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +010060 PATCH_SITE(pv_cpu_ops, swapgs);
61 PATCH_SITE(pv_mmu_ops, read_cr2);
62 PATCH_SITE(pv_mmu_ops, read_cr3);
63 PATCH_SITE(pv_mmu_ops, write_cr3);
64 PATCH_SITE(pv_cpu_ops, clts);
65 PATCH_SITE(pv_mmu_ops, flush_tlb_single);
66 PATCH_SITE(pv_cpu_ops, wbinvd);
Ingo Molnar62c7a1e2015-05-11 09:47:23 +020067#if defined(CONFIG_PARAVIRT_SPINLOCKS) && defined(CONFIG_QUEUED_SPINLOCKS)
Peter Zijlstra (Intel)f233f7f2015-04-24 14:56:38 -040068 case PARAVIRT_PATCH(pv_lock_ops.queued_spin_unlock):
69 if (pv_is_native_spin_unlock()) {
70 start = start_pv_lock_ops_queued_spin_unlock;
71 end = end_pv_lock_ops_queued_spin_unlock;
72 goto patch_site;
73 }
74#endif
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +010075
76 default:
77 ret = paravirt_patch_default(type, clobbers, ibuf, addr, len);
78 break;
Peter Zijlstra (Intel)f233f7f2015-04-24 14:56:38 -040079
80patch_site:
81 ret = paravirt_patch_insns(ibuf, len, start, end);
82 break;
Glauber de Oliveira Costa53fd13cf2008-01-30 13:32:10 +010083 }
84#undef PATCH_SITE
85 return ret;
86}