// SPDX-License-Identifier: GPL-2.0 #include #include #include #define PSTART(d, m) \ patch_data_##d.m #define PEND(d, m) \ (PSTART(d, m) + sizeof(patch_data_##d.m)) #define PATCH(d, m, insn_buff, len) \ paravirt_patch_insns(insn_buff, len, PSTART(d, m), PEND(d, m)) #define PATCH_CASE(ops, m, data, insn_buff, len) \ case PARAVIRT_PATCH(ops.m): \ return PATCH(data, ops##_##m, insn_buff, len) #ifdef CONFIG_PARAVIRT_XXL struct patch_xxl { const unsigned char irq_irq_disable[1]; const unsigned char irq_irq_enable[1]; const unsigned char irq_save_fl[2]; const unsigned char mmu_read_cr2[3]; const unsigned char mmu_read_cr3[3]; const unsigned char mmu_write_cr3[3]; const unsigned char irq_restore_fl[2]; # ifdef CONFIG_X86_64 const unsigned char cpu_wbinvd[2]; const unsigned char cpu_usergs_sysret64[6]; const unsigned char cpu_swapgs[3]; const unsigned char mov64[3]; # else const unsigned char cpu_iret[1]; # endif }; static const struct patch_xxl patch_data_xxl = { .irq_irq_disable = { 0xfa }, // cli .irq_irq_enable = { 0xfb }, // sti .irq_save_fl = { 0x9c, 0x58 }, // pushf; pop %[re]ax .mmu_read_cr2 = { 0x0f, 0x20, 0xd0 }, // mov %cr2, %[re]ax .mmu_read_cr3 = { 0x0f, 0x20, 0xd8 }, // mov %cr3, %[re]ax # ifdef CONFIG_X86_64 .mmu_write_cr3 = { 0x0f, 0x22, 0xdf }, // mov %rdi, %cr3 .irq_restore_fl = { 0x57, 0x9d }, // push %rdi; popfq .cpu_wbinvd = { 0x0f, 0x09 }, // wbinvd .cpu_usergs_sysret64 = { 0x0f, 0x01, 0xf8, 0x48, 0x0f, 0x07 }, // swapgs; sysretq .cpu_swapgs = { 0x0f, 0x01, 0xf8 }, // swapgs .mov64 = { 0x48, 0x89, 0xf8 }, // mov %rdi, %rax # else .mmu_write_cr3 = { 0x0f, 0x22, 0xd8 }, // mov %eax, %cr3 .irq_restore_fl = { 0x50, 0x9d }, // push %eax; popf .cpu_iret = { 0xcf }, // iret # endif }; unsigned int paravirt_patch_ident_64(void *insn_buff, unsigned int len) { #ifdef CONFIG_X86_64 return PATCH(xxl, mov64, insn_buff, len); #endif return 0; } # endif /* CONFIG_PARAVIRT_XXL */ #ifdef CONFIG_PARAVIRT_SPINLOCKS struct patch_lock { unsigned char queued_spin_unlock[3]; unsigned char vcpu_is_preempted[2]; }; static const struct patch_lock patch_data_lock = { .vcpu_is_preempted = { 0x31, 0xc0 }, // xor %eax, %eax # ifdef CONFIG_X86_64 .queued_spin_unlock = { 0xc6, 0x07, 0x00 }, // movb $0, (%rdi) # else .queued_spin_unlock = { 0xc6, 0x00, 0x00 }, // movb $0, (%eax) # endif }; #endif /* CONFIG_PARAVIRT_SPINLOCKS */ unsigned int native_patch(u8 type, void *insn_buff, unsigned long addr, unsigned int len) { switch (type) { #ifdef CONFIG_PARAVIRT_XXL PATCH_CASE(irq, restore_fl, xxl, insn_buff, len); PATCH_CASE(irq, save_fl, xxl, insn_buff, len); PATCH_CASE(irq, irq_enable, xxl, insn_buff, len); PATCH_CASE(irq, irq_disable, xxl, insn_buff, len); PATCH_CASE(mmu, read_cr2, xxl, insn_buff, len); PATCH_CASE(mmu, read_cr3, xxl, insn_buff, len); PATCH_CASE(mmu, write_cr3, xxl, insn_buff, len); # ifdef CONFIG_X86_64 PATCH_CASE(cpu, usergs_sysret64, xxl, insn_buff, len); PATCH_CASE(cpu, swapgs, xxl, insn_buff, len); PATCH_CASE(cpu, wbinvd, xxl, insn_buff, len); # else PATCH_CASE(cpu, iret, xxl, insn_buff, len); # endif #endif #ifdef CONFIG_PARAVIRT_SPINLOCKS case PARAVIRT_PATCH(lock.queued_spin_unlock): if (pv_is_native_spin_unlock()) return PATCH(lock, queued_spin_unlock, insn_buff, len); break; case PARAVIRT_PATCH(lock.vcpu_is_preempted): if (pv_is_native_vcpu_is_preempted()) return PATCH(lock, vcpu_is_preempted, insn_buff, len); break; #endif default: break; } return paravirt_patch_default(type, insn_buff, addr, len); }