| /****************************************************************************** |
| * arch/ia64/xen/xen_pv_ops.c |
| * |
| * Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp> |
| * VA Linux Systems Japan K.K. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation; either version 2 of the License, or |
| * (at your option) any later version. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA |
| * |
| */ |
| |
| #include <linux/console.h> |
| #include <linux/irq.h> |
| #include <linux/kernel.h> |
| #include <linux/pm.h> |
| #include <linux/unistd.h> |
| |
| #include <asm/xen/hypervisor.h> |
| #include <asm/xen/xencomm.h> |
| #include <asm/xen/privop.h> |
| |
| #include "irq_xen.h" |
| #include "time.h" |
| |
| /*************************************************************************** |
| * general info |
| */ |
| static struct pv_info xen_info __initdata = { |
| .kernel_rpl = 2, /* or 1: determin at runtime */ |
| .paravirt_enabled = 1, |
| .name = "Xen/ia64", |
| }; |
| |
| #define IA64_RSC_PL_SHIFT 2 |
| #define IA64_RSC_PL_BIT_SIZE 2 |
| #define IA64_RSC_PL_MASK \ |
| (((1UL << IA64_RSC_PL_BIT_SIZE) - 1) << IA64_RSC_PL_SHIFT) |
| |
| static void __init |
| xen_info_init(void) |
| { |
| /* Xenified Linux/ia64 may run on pl = 1 or 2. |
| * determin at run time. */ |
| unsigned long rsc = ia64_getreg(_IA64_REG_AR_RSC); |
| unsigned int rpl = (rsc & IA64_RSC_PL_MASK) >> IA64_RSC_PL_SHIFT; |
| xen_info.kernel_rpl = rpl; |
| } |
| |
| /*************************************************************************** |
| * pv_init_ops |
| * initialization hooks. |
| */ |
| |
| static void |
| xen_panic_hypercall(struct unw_frame_info *info, void *arg) |
| { |
| current->thread.ksp = (__u64)info->sw - 16; |
| HYPERVISOR_shutdown(SHUTDOWN_crash); |
| /* we're never actually going to get here... */ |
| } |
| |
| static int |
| xen_panic_event(struct notifier_block *this, unsigned long event, void *ptr) |
| { |
| unw_init_running(xen_panic_hypercall, NULL); |
| /* we're never actually going to get here... */ |
| return NOTIFY_DONE; |
| } |
| |
| static struct notifier_block xen_panic_block = { |
| xen_panic_event, NULL, 0 /* try to go last */ |
| }; |
| |
| static void xen_pm_power_off(void) |
| { |
| local_irq_disable(); |
| HYPERVISOR_shutdown(SHUTDOWN_poweroff); |
| } |
| |
| static void __init |
| xen_banner(void) |
| { |
| printk(KERN_INFO |
| "Running on Xen! pl = %d start_info_pfn=0x%lx nr_pages=%ld " |
| "flags=0x%x\n", |
| xen_info.kernel_rpl, |
| HYPERVISOR_shared_info->arch.start_info_pfn, |
| xen_start_info->nr_pages, xen_start_info->flags); |
| } |
| |
| static int __init |
| xen_reserve_memory(struct rsvd_region *region) |
| { |
| region->start = (unsigned long)__va( |
| (HYPERVISOR_shared_info->arch.start_info_pfn << PAGE_SHIFT)); |
| region->end = region->start + PAGE_SIZE; |
| return 1; |
| } |
| |
| static void __init |
| xen_arch_setup_early(void) |
| { |
| struct shared_info *s; |
| BUG_ON(!xen_pv_domain()); |
| |
| s = HYPERVISOR_shared_info; |
| xen_start_info = __va(s->arch.start_info_pfn << PAGE_SHIFT); |
| |
| /* Must be done before any hypercall. */ |
| xencomm_initialize(); |
| |
| xen_setup_features(); |
| /* Register a call for panic conditions. */ |
| atomic_notifier_chain_register(&panic_notifier_list, |
| &xen_panic_block); |
| pm_power_off = xen_pm_power_off; |
| |
| xen_ia64_enable_opt_feature(); |
| } |
| |
| static void __init |
| xen_arch_setup_console(char **cmdline_p) |
| { |
| add_preferred_console("xenboot", 0, NULL); |
| add_preferred_console("tty", 0, NULL); |
| /* use hvc_xen */ |
| add_preferred_console("hvc", 0, NULL); |
| |
| #if !defined(CONFIG_VT) || !defined(CONFIG_DUMMY_CONSOLE) |
| conswitchp = NULL; |
| #endif |
| } |
| |
| static int __init |
| xen_arch_setup_nomca(void) |
| { |
| return 1; |
| } |
| |
| static void __init |
| xen_post_smp_prepare_boot_cpu(void) |
| { |
| xen_setup_vcpu_info_placement(); |
| } |
| |
| #ifdef ASM_SUPPORTED |
| static unsigned long __init_or_module |
| xen_patch_bundle(void *sbundle, void *ebundle, unsigned long type); |
| #endif |
| static void __init |
| xen_patch_branch(unsigned long tag, unsigned long type); |
| |
| static const struct pv_init_ops xen_init_ops __initconst = { |
| .banner = xen_banner, |
| |
| .reserve_memory = xen_reserve_memory, |
| |
| .arch_setup_early = xen_arch_setup_early, |
| .arch_setup_console = xen_arch_setup_console, |
| .arch_setup_nomca = xen_arch_setup_nomca, |
| |
| .post_smp_prepare_boot_cpu = xen_post_smp_prepare_boot_cpu, |
| #ifdef ASM_SUPPORTED |
| .patch_bundle = xen_patch_bundle, |
| #endif |
| .patch_branch = xen_patch_branch, |
| }; |
| |
| /*************************************************************************** |
| * pv_fsys_data |
| * addresses for fsys |
| */ |
| |
| extern unsigned long xen_fsyscall_table[NR_syscalls]; |
| extern char xen_fsys_bubble_down[]; |
| struct pv_fsys_data xen_fsys_data __initdata = { |
| .fsyscall_table = (unsigned long *)xen_fsyscall_table, |
| .fsys_bubble_down = (void *)xen_fsys_bubble_down, |
| }; |
| |
| /*************************************************************************** |
| * pv_patchdata |
| * patchdata addresses |
| */ |
| |
| #define DECLARE(name) \ |
| extern unsigned long __xen_start_gate_##name##_patchlist[]; \ |
| extern unsigned long __xen_end_gate_##name##_patchlist[] |
| |
| DECLARE(fsyscall); |
| DECLARE(brl_fsys_bubble_down); |
| DECLARE(vtop); |
| DECLARE(mckinley_e9); |
| |
| extern unsigned long __xen_start_gate_section[]; |
| |
| #define ASSIGN(name) \ |
| .start_##name##_patchlist = \ |
| (unsigned long)__xen_start_gate_##name##_patchlist, \ |
| .end_##name##_patchlist = \ |
| (unsigned long)__xen_end_gate_##name##_patchlist |
| |
| static struct pv_patchdata xen_patchdata __initdata = { |
| ASSIGN(fsyscall), |
| ASSIGN(brl_fsys_bubble_down), |
| ASSIGN(vtop), |
| ASSIGN(mckinley_e9), |
| |
| .gate_section = (void*)__xen_start_gate_section, |
| }; |
| |
| /*************************************************************************** |
| * pv_cpu_ops |
| * intrinsics hooks. |
| */ |
| |
| #ifndef ASM_SUPPORTED |
| static void |
| xen_set_itm_with_offset(unsigned long val) |
| { |
| /* ia64_cpu_local_tick() calls this with interrupt enabled. */ |
| /* WARN_ON(!irqs_disabled()); */ |
| xen_set_itm(val - XEN_MAPPEDREGS->itc_offset); |
| } |
| |
| static unsigned long |
| xen_get_itm_with_offset(void) |
| { |
| /* unused at this moment */ |
| printk(KERN_DEBUG "%s is called.\n", __func__); |
| |
| WARN_ON(!irqs_disabled()); |
| return ia64_native_getreg(_IA64_REG_CR_ITM) + |
| XEN_MAPPEDREGS->itc_offset; |
| } |
| |
| /* ia64_set_itc() is only called by |
| * cpu_init() with ia64_set_itc(0) and ia64_sync_itc(). |
| * So XEN_MAPPEDRESG->itc_offset cal be considered as almost constant. |
| */ |
| static void |
| xen_set_itc(unsigned long val) |
| { |
| unsigned long mitc; |
| |
| WARN_ON(!irqs_disabled()); |
| mitc = ia64_native_getreg(_IA64_REG_AR_ITC); |
| XEN_MAPPEDREGS->itc_offset = val - mitc; |
| XEN_MAPPEDREGS->itc_last = val; |
| } |
| |
| static unsigned long |
| xen_get_itc(void) |
| { |
| unsigned long res; |
| unsigned long itc_offset; |
| unsigned long itc_last; |
| unsigned long ret_itc_last; |
| |
| itc_offset = XEN_MAPPEDREGS->itc_offset; |
| do { |
| itc_last = XEN_MAPPEDREGS->itc_last; |
| res = ia64_native_getreg(_IA64_REG_AR_ITC); |
| res += itc_offset; |
| if (itc_last >= res) |
| res = itc_last + 1; |
| ret_itc_last = cmpxchg(&XEN_MAPPEDREGS->itc_last, |
| itc_last, res); |
| } while (unlikely(ret_itc_last != itc_last)); |
| return res; |
| |
| #if 0 |
| /* ia64_itc_udelay() calls ia64_get_itc() with interrupt enabled. |
| Should it be paravirtualized instead? */ |
| WARN_ON(!irqs_disabled()); |
| itc_offset = XEN_MAPPEDREGS->itc_offset; |
| itc_last = XEN_MAPPEDREGS->itc_last; |
| res = ia64_native_getreg(_IA64_REG_AR_ITC); |
| res += itc_offset; |
| if (itc_last >= res) |
| res = itc_last + 1; |
| XEN_MAPPEDREGS->itc_last = res; |
| return res; |
| #endif |
| } |
| |
| static void xen_setreg(int regnum, unsigned long val) |
| { |
| switch (regnum) { |
| case _IA64_REG_AR_KR0 ... _IA64_REG_AR_KR7: |
| xen_set_kr(regnum - _IA64_REG_AR_KR0, val); |
| break; |
| case _IA64_REG_AR_ITC: |
| xen_set_itc(val); |
| break; |
| case _IA64_REG_CR_TPR: |
| xen_set_tpr(val); |
| break; |
| case _IA64_REG_CR_ITM: |
| xen_set_itm_with_offset(val); |
| break; |
| case _IA64_REG_CR_EOI: |
| xen_eoi(val); |
| break; |
| default: |
| ia64_native_setreg_func(regnum, val); |
| break; |
| } |
| } |
| |
| static unsigned long xen_getreg(int regnum) |
| { |
| unsigned long res; |
| |
| switch (regnum) { |
| case _IA64_REG_PSR: |
| res = xen_get_psr(); |
| break; |
| case _IA64_REG_AR_ITC: |
| res = xen_get_itc(); |
| break; |
| case _IA64_REG_CR_ITM: |
| res = xen_get_itm_with_offset(); |
| break; |
| case _IA64_REG_CR_IVR: |
| res = xen_get_ivr(); |
| break; |
| case _IA64_REG_CR_TPR: |
| res = xen_get_tpr(); |
| break; |
| default: |
| res = ia64_native_getreg_func(regnum); |
| break; |
| } |
| return res; |
| } |
| |
| /* turning on interrupts is a bit more complicated.. write to the |
| * memory-mapped virtual psr.i bit first (to avoid race condition), |
| * then if any interrupts were pending, we have to execute a hyperprivop |
| * to ensure the pending interrupt gets delivered; else we're done! */ |
| static void |
| xen_ssm_i(void) |
| { |
| int old = xen_get_virtual_psr_i(); |
| xen_set_virtual_psr_i(1); |
| barrier(); |
| if (!old && xen_get_virtual_pend()) |
| xen_hyper_ssm_i(); |
| } |
| |
| /* turning off interrupts can be paravirtualized simply by writing |
| * to a memory-mapped virtual psr.i bit (implemented as a 16-bit bool) */ |
| static void |
| xen_rsm_i(void) |
| { |
| xen_set_virtual_psr_i(0); |
| barrier(); |
| } |
| |
| static unsigned long |
| xen_get_psr_i(void) |
| { |
| return xen_get_virtual_psr_i() ? IA64_PSR_I : 0; |
| } |
| |
| static void |
| xen_intrin_local_irq_restore(unsigned long mask) |
| { |
| if (mask & IA64_PSR_I) |
| xen_ssm_i(); |
| else |
| xen_rsm_i(); |
| } |
| #else |
| #define __DEFINE_FUNC(name, code) \ |
| extern const char xen_ ## name ## _direct_start[]; \ |
| extern const char xen_ ## name ## _direct_end[]; \ |
| asm (".align 32\n" \ |
| ".proc xen_" #name "\n" \ |
| "xen_" #name ":\n" \ |
| "xen_" #name "_direct_start:\n" \ |
| code \ |
| "xen_" #name "_direct_end:\n" \ |
| "br.cond.sptk.many b6\n" \ |
| ".endp xen_" #name "\n") |
| |
| #define DEFINE_VOID_FUNC0(name, code) \ |
| extern void \ |
| xen_ ## name (void); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define DEFINE_VOID_FUNC1(name, code) \ |
| extern void \ |
| xen_ ## name (unsigned long arg); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define DEFINE_VOID_FUNC1_VOID(name, code) \ |
| extern void \ |
| xen_ ## name (void *arg); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define DEFINE_VOID_FUNC2(name, code) \ |
| extern void \ |
| xen_ ## name (unsigned long arg0, \ |
| unsigned long arg1); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define DEFINE_FUNC0(name, code) \ |
| extern unsigned long \ |
| xen_ ## name (void); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define DEFINE_FUNC1(name, type, code) \ |
| extern unsigned long \ |
| xen_ ## name (type arg); \ |
| __DEFINE_FUNC(name, code) |
| |
| #define XEN_PSR_I_ADDR_ADDR (XSI_BASE + XSI_PSR_I_ADDR_OFS) |
| |
| /* |
| * static void xen_set_itm_with_offset(unsigned long val) |
| * xen_set_itm(val - XEN_MAPPEDREGS->itc_offset); |
| */ |
| /* 2 bundles */ |
| DEFINE_VOID_FUNC1(set_itm_with_offset, |
| "mov r2 = " __stringify(XSI_BASE) " + " |
| __stringify(XSI_ITC_OFFSET_OFS) "\n" |
| ";;\n" |
| "ld8 r3 = [r2]\n" |
| ";;\n" |
| "sub r8 = r8, r3\n" |
| "break " __stringify(HYPERPRIVOP_SET_ITM) "\n"); |
| |
| /* |
| * static unsigned long xen_get_itm_with_offset(void) |
| * return ia64_native_getreg(_IA64_REG_CR_ITM) + XEN_MAPPEDREGS->itc_offset; |
| */ |
| /* 2 bundles */ |
| DEFINE_FUNC0(get_itm_with_offset, |
| "mov r2 = " __stringify(XSI_BASE) " + " |
| __stringify(XSI_ITC_OFFSET_OFS) "\n" |
| ";;\n" |
| "ld8 r3 = [r2]\n" |
| "mov r8 = cr.itm\n" |
| ";;\n" |
| "add r8 = r8, r2\n"); |
| |
| /* |
| * static void xen_set_itc(unsigned long val) |
| * unsigned long mitc; |
| * |
| * WARN_ON(!irqs_disabled()); |
| * mitc = ia64_native_getreg(_IA64_REG_AR_ITC); |
| * XEN_MAPPEDREGS->itc_offset = val - mitc; |
| * XEN_MAPPEDREGS->itc_last = val; |
| */ |
| /* 2 bundles */ |
| DEFINE_VOID_FUNC1(set_itc, |
| "mov r2 = " __stringify(XSI_BASE) " + " |
| __stringify(XSI_ITC_LAST_OFS) "\n" |
| "mov r3 = ar.itc\n" |
| ";;\n" |
| "sub r3 = r8, r3\n" |
| "st8 [r2] = r8, " |
| __stringify(XSI_ITC_LAST_OFS) " - " |
| __stringify(XSI_ITC_OFFSET_OFS) "\n" |
| ";;\n" |
| "st8 [r2] = r3\n"); |
| |
| /* |
| * static unsigned long xen_get_itc(void) |
| * unsigned long res; |
| * unsigned long itc_offset; |
| * unsigned long itc_last; |
| * unsigned long ret_itc_last; |
| * |
| * itc_offset = XEN_MAPPEDREGS->itc_offset; |
| * do { |
| * itc_last = XEN_MAPPEDREGS->itc_last; |
| * res = ia64_native_getreg(_IA64_REG_AR_ITC); |
| * res += itc_offset; |
| * if (itc_last >= res) |
| * res = itc_last + 1; |
| * ret_itc_last = cmpxchg(&XEN_MAPPEDREGS->itc_last, |
| * itc_last, res); |
| * } while (unlikely(ret_itc_last != itc_last)); |
| * return res; |
| */ |
| /* 5 bundles */ |
| DEFINE_FUNC0(get_itc, |
| "mov r2 = " __stringify(XSI_BASE) " + " |
| __stringify(XSI_ITC_OFFSET_OFS) "\n" |
| ";;\n" |
| "ld8 r9 = [r2], " __stringify(XSI_ITC_LAST_OFS) " - " |
| __stringify(XSI_ITC_OFFSET_OFS) "\n" |
| /* r9 = itc_offset */ |
| /* r2 = XSI_ITC_OFFSET */ |
| "888:\n" |
| "mov r8 = ar.itc\n" /* res = ar.itc */ |
| ";;\n" |
| "ld8 r3 = [r2]\n" /* r3 = itc_last */ |
| "add r8 = r8, r9\n" /* res = ar.itc + itc_offset */ |
| ";;\n" |
| "cmp.gtu p6, p0 = r3, r8\n" |
| ";;\n" |
| "(p6) add r8 = 1, r3\n" /* if (itc_last > res) itc_last + 1 */ |
| ";;\n" |
| "mov ar.ccv = r8\n" |
| ";;\n" |
| "cmpxchg8.acq r10 = [r2], r8, ar.ccv\n" |
| ";;\n" |
| "cmp.ne p6, p0 = r10, r3\n" |
| "(p6) hint @pause\n" |
| "(p6) br.cond.spnt 888b\n"); |
| |
| DEFINE_VOID_FUNC1_VOID(fc, |
| "break " __stringify(HYPERPRIVOP_FC) "\n"); |
| |
| /* |
| * psr_i_addr_addr = XEN_PSR_I_ADDR_ADDR |
| * masked_addr = *psr_i_addr_addr |
| * pending_intr_addr = masked_addr - 1 |
| * if (val & IA64_PSR_I) { |
| * masked = *masked_addr |
| * *masked_addr = 0:xen_set_virtual_psr_i(1) |
| * compiler barrier |
| * if (masked) { |
| * uint8_t pending = *pending_intr_addr; |
| * if (pending) |
| * XEN_HYPER_SSM_I |
| * } |
| * } else { |
| * *masked_addr = 1:xen_set_virtual_psr_i(0) |
| * } |
| */ |
| /* 6 bundles */ |
| DEFINE_VOID_FUNC1(intrin_local_irq_restore, |
| /* r8 = input value: 0 or IA64_PSR_I |
| * p6 = (flags & IA64_PSR_I) |
| * = if clause |
| * p7 = !(flags & IA64_PSR_I) |
| * = else clause |
| */ |
| "cmp.ne p6, p7 = r8, r0\n" |
| "mov r9 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" |
| ";;\n" |
| /* r9 = XEN_PSR_I_ADDR */ |
| "ld8 r9 = [r9]\n" |
| ";;\n" |
| |
| /* r10 = masked previous value */ |
| "(p6) ld1.acq r10 = [r9]\n" |
| ";;\n" |
| |
| /* p8 = !masked interrupt masked previously? */ |
| "(p6) cmp.ne.unc p8, p0 = r10, r0\n" |
| |
| /* p7 = else clause */ |
| "(p7) mov r11 = 1\n" |
| ";;\n" |
| /* masked = 1 */ |
| "(p7) st1.rel [r9] = r11\n" |
| |
| /* p6 = if clause */ |
| /* masked = 0 |
| * r9 = masked_addr - 1 |
| * = pending_intr_addr |
| */ |
| "(p8) st1.rel [r9] = r0, -1\n" |
| ";;\n" |
| /* r8 = pending_intr */ |
| "(p8) ld1.acq r11 = [r9]\n" |
| ";;\n" |
| /* p9 = interrupt pending? */ |
| "(p8) cmp.ne.unc p9, p10 = r11, r0\n" |
| ";;\n" |
| "(p10) mf\n" |
| /* issue hypercall to trigger interrupt */ |
| "(p9) break " __stringify(HYPERPRIVOP_SSM_I) "\n"); |
| |
| DEFINE_VOID_FUNC2(ptcga, |
| "break " __stringify(HYPERPRIVOP_PTC_GA) "\n"); |
| DEFINE_VOID_FUNC2(set_rr, |
| "break " __stringify(HYPERPRIVOP_SET_RR) "\n"); |
| |
| /* |
| * tmp = XEN_MAPPEDREGS->interrupt_mask_addr = XEN_PSR_I_ADDR_ADDR; |
| * tmp = *tmp |
| * tmp = *tmp; |
| * psr_i = tmp? 0: IA64_PSR_I; |
| */ |
| /* 4 bundles */ |
| DEFINE_FUNC0(get_psr_i, |
| "mov r9 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" |
| ";;\n" |
| "ld8 r9 = [r9]\n" /* r9 = XEN_PSR_I_ADDR */ |
| "mov r8 = 0\n" /* psr_i = 0 */ |
| ";;\n" |
| "ld1.acq r9 = [r9]\n" /* r9 = XEN_PSR_I */ |
| ";;\n" |
| "cmp.eq.unc p6, p0 = r9, r0\n" /* p6 = (XEN_PSR_I != 0) */ |
| ";;\n" |
| "(p6) mov r8 = " __stringify(1 << IA64_PSR_I_BIT) "\n"); |
| |
| DEFINE_FUNC1(thash, unsigned long, |
| "break " __stringify(HYPERPRIVOP_THASH) "\n"); |
| DEFINE_FUNC1(get_cpuid, int, |
| "break " __stringify(HYPERPRIVOP_GET_CPUID) "\n"); |
| DEFINE_FUNC1(get_pmd, int, |
| "break " __stringify(HYPERPRIVOP_GET_PMD) "\n"); |
| DEFINE_FUNC1(get_rr, unsigned long, |
| "break " __stringify(HYPERPRIVOP_GET_RR) "\n"); |
| |
| /* |
| * void xen_privop_ssm_i(void) |
| * |
| * int masked = !xen_get_virtual_psr_i(); |
| * // masked = *(*XEN_MAPPEDREGS->interrupt_mask_addr) |
| * xen_set_virtual_psr_i(1) |
| * // *(*XEN_MAPPEDREGS->interrupt_mask_addr) = 0 |
| * // compiler barrier |
| * if (masked) { |
| * uint8_t* pend_int_addr = |
| * (uint8_t*)(*XEN_MAPPEDREGS->interrupt_mask_addr) - 1; |
| * uint8_t pending = *pend_int_addr; |
| * if (pending) |
| * XEN_HYPER_SSM_I |
| * } |
| */ |
| /* 4 bundles */ |
| DEFINE_VOID_FUNC0(ssm_i, |
| "mov r8 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" |
| ";;\n" |
| "ld8 r8 = [r8]\n" /* r8 = XEN_PSR_I_ADDR */ |
| ";;\n" |
| "ld1.acq r9 = [r8]\n" /* r9 = XEN_PSR_I */ |
| ";;\n" |
| "st1.rel [r8] = r0, -1\n" /* psr_i = 0. enable interrupt |
| * r8 = XEN_PSR_I_ADDR - 1 |
| * = pend_int_addr |
| */ |
| "cmp.eq.unc p0, p6 = r9, r0\n"/* p6 = !XEN_PSR_I |
| * previously interrupt |
| * masked? |
| */ |
| ";;\n" |
| "(p6) ld1.acq r8 = [r8]\n" /* r8 = xen_pend_int */ |
| ";;\n" |
| "(p6) cmp.eq.unc p6, p7 = r8, r0\n" /*interrupt pending?*/ |
| ";;\n" |
| /* issue hypercall to get interrupt */ |
| "(p7) break " __stringify(HYPERPRIVOP_SSM_I) "\n" |
| ";;\n"); |
| |
| /* |
| * psr_i_addr_addr = XEN_MAPPEDREGS->interrupt_mask_addr |
| * = XEN_PSR_I_ADDR_ADDR; |
| * psr_i_addr = *psr_i_addr_addr; |
| * *psr_i_addr = 1; |
| */ |
| /* 2 bundles */ |
| DEFINE_VOID_FUNC0(rsm_i, |
| "mov r8 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" |
| /* r8 = XEN_PSR_I_ADDR */ |
| "mov r9 = 1\n" |
| ";;\n" |
| "ld8 r8 = [r8]\n" /* r8 = XEN_PSR_I */ |
| ";;\n" |
| "st1.rel [r8] = r9\n"); /* XEN_PSR_I = 1 */ |
| |
| extern void |
| xen_set_rr0_to_rr4(unsigned long val0, unsigned long val1, |
| unsigned long val2, unsigned long val3, |
| unsigned long val4); |
| __DEFINE_FUNC(set_rr0_to_rr4, |
| "break " __stringify(HYPERPRIVOP_SET_RR0_TO_RR4) "\n"); |
| |
| |
| extern unsigned long xen_getreg(int regnum); |
| #define __DEFINE_GET_REG(id, privop) \ |
| "mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \ |
| ";;\n" \ |
| "cmp.eq p6, p0 = r2, r8\n" \ |
| ";;\n" \ |
| "(p6) break " __stringify(HYPERPRIVOP_GET_ ## privop) "\n" \ |
| "(p6) br.cond.sptk.many b6\n" \ |
| ";;\n" |
| |
| __DEFINE_FUNC(getreg, |
| __DEFINE_GET_REG(PSR, PSR) |
| |
| /* get_itc */ |
| "mov r2 = " __stringify(_IA64_REG_AR_ITC) "\n" |
| ";;\n" |
| "cmp.eq p6, p0 = r2, r8\n" |
| ";;\n" |
| "(p6) br.cond.spnt xen_get_itc\n" |
| ";;\n" |
| |
| /* get itm */ |
| "mov r2 = " __stringify(_IA64_REG_CR_ITM) "\n" |
| ";;\n" |
| "cmp.eq p6, p0 = r2, r8\n" |
| ";;\n" |
| "(p6) br.cond.spnt xen_get_itm_with_offset\n" |
| ";;\n" |
| |
| __DEFINE_GET_REG(CR_IVR, IVR) |
| __DEFINE_GET_REG(CR_TPR, TPR) |
| |
| /* fall back */ |
| "movl r2 = ia64_native_getreg_func\n" |
| ";;\n" |
| "mov b7 = r2\n" |
| ";;\n" |
| "br.cond.sptk.many b7\n"); |
| |
| extern void xen_setreg(int regnum, unsigned long val); |
| #define __DEFINE_SET_REG(id, privop) \ |
| "mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \ |
| ";;\n" \ |
| "cmp.eq p6, p0 = r2, r9\n" \ |
| ";;\n" \ |
| "(p6) break " __stringify(HYPERPRIVOP_ ## privop) "\n" \ |
| "(p6) br.cond.sptk.many b6\n" \ |
| ";;\n" |
| |
| __DEFINE_FUNC(setreg, |
| /* kr0 .. kr 7*/ |
| /* |
| * if (_IA64_REG_AR_KR0 <= regnum && |
| * regnum <= _IA64_REG_AR_KR7) { |
| * register __index asm ("r8") = regnum - _IA64_REG_AR_KR0 |
| * register __val asm ("r9") = val |
| * "break HYPERPRIVOP_SET_KR" |
| * } |
| */ |
| "mov r17 = r9\n" |
| "mov r2 = " __stringify(_IA64_REG_AR_KR0) "\n" |
| ";;\n" |
| "cmp.ge p6, p0 = r9, r2\n" |
| "sub r17 = r17, r2\n" |
| ";;\n" |
| "(p6) cmp.ge.unc p7, p0 = " |
| __stringify(_IA64_REG_AR_KR7) " - " __stringify(_IA64_REG_AR_KR0) |
| ", r17\n" |
| ";;\n" |
| "(p7) mov r9 = r8\n" |
| ";;\n" |
| "(p7) mov r8 = r17\n" |
| "(p7) break " __stringify(HYPERPRIVOP_SET_KR) "\n" |
| |
| /* set itm */ |
| "mov r2 = " __stringify(_IA64_REG_CR_ITM) "\n" |
| ";;\n" |
| "cmp.eq p6, p0 = r2, r8\n" |
| ";;\n" |
| "(p6) br.cond.spnt xen_set_itm_with_offset\n" |
| |
| /* set itc */ |
| "mov r2 = " __stringify(_IA64_REG_AR_ITC) "\n" |
| ";;\n" |
| "cmp.eq p6, p0 = r2, r8\n" |
| ";;\n" |
| "(p6) br.cond.spnt xen_set_itc\n" |
| |
| __DEFINE_SET_REG(CR_TPR, SET_TPR) |
| __DEFINE_SET_REG(CR_EOI, EOI) |
| |
| /* fall back */ |
| "movl r2 = ia64_native_setreg_func\n" |
| ";;\n" |
| "mov b7 = r2\n" |
| ";;\n" |
| "br.cond.sptk.many b7\n"); |
| #endif |
| |
| static const struct pv_cpu_ops xen_cpu_ops __initconst = { |
| .fc = xen_fc, |
| .thash = xen_thash, |
| .get_cpuid = xen_get_cpuid, |
| .get_pmd = xen_get_pmd, |
| .getreg = xen_getreg, |
| .setreg = xen_setreg, |
| .ptcga = xen_ptcga, |
| .get_rr = xen_get_rr, |
| .set_rr = xen_set_rr, |
| .set_rr0_to_rr4 = xen_set_rr0_to_rr4, |
| .ssm_i = xen_ssm_i, |
| .rsm_i = xen_rsm_i, |
| .get_psr_i = xen_get_psr_i, |
| .intrin_local_irq_restore |
| = xen_intrin_local_irq_restore, |
| }; |
| |
| /****************************************************************************** |
| * replacement of hand written assembly codes. |
| */ |
| |
| extern char xen_switch_to; |
| extern char xen_leave_syscall; |
| extern char xen_work_processed_syscall; |
| extern char xen_leave_kernel; |
| |
| const struct pv_cpu_asm_switch xen_cpu_asm_switch = { |
| .switch_to = (unsigned long)&xen_switch_to, |
| .leave_syscall = (unsigned long)&xen_leave_syscall, |
| .work_processed_syscall = (unsigned long)&xen_work_processed_syscall, |
| .leave_kernel = (unsigned long)&xen_leave_kernel, |
| }; |
| |
| /*************************************************************************** |
| * pv_iosapic_ops |
| * iosapic read/write hooks. |
| */ |
| static void |
| xen_pcat_compat_init(void) |
| { |
| /* nothing */ |
| } |
| |
| static struct irq_chip* |
| xen_iosapic_get_irq_chip(unsigned long trigger) |
| { |
| return NULL; |
| } |
| |
| static unsigned int |
| xen_iosapic_read(char __iomem *iosapic, unsigned int reg) |
| { |
| struct physdev_apic apic_op; |
| int ret; |
| |
| apic_op.apic_physbase = (unsigned long)iosapic - |
| __IA64_UNCACHED_OFFSET; |
| apic_op.reg = reg; |
| ret = HYPERVISOR_physdev_op(PHYSDEVOP_apic_read, &apic_op); |
| if (ret) |
| return ret; |
| return apic_op.value; |
| } |
| |
| static void |
| xen_iosapic_write(char __iomem *iosapic, unsigned int reg, u32 val) |
| { |
| struct physdev_apic apic_op; |
| |
| apic_op.apic_physbase = (unsigned long)iosapic - |
| __IA64_UNCACHED_OFFSET; |
| apic_op.reg = reg; |
| apic_op.value = val; |
| HYPERVISOR_physdev_op(PHYSDEVOP_apic_write, &apic_op); |
| } |
| |
| static struct pv_iosapic_ops xen_iosapic_ops __initdata = { |
| .pcat_compat_init = xen_pcat_compat_init, |
| .__get_irq_chip = xen_iosapic_get_irq_chip, |
| |
| .__read = xen_iosapic_read, |
| .__write = xen_iosapic_write, |
| }; |
| |
| /*************************************************************************** |
| * pv_ops initialization |
| */ |
| |
| void __init |
| xen_setup_pv_ops(void) |
| { |
| xen_info_init(); |
| pv_info = xen_info; |
| pv_init_ops = xen_init_ops; |
| pv_fsys_data = xen_fsys_data; |
| pv_patchdata = xen_patchdata; |
| pv_cpu_ops = xen_cpu_ops; |
| pv_iosapic_ops = xen_iosapic_ops; |
| pv_irq_ops = xen_irq_ops; |
| pv_time_ops = xen_time_ops; |
| |
| paravirt_cpu_asm_init(&xen_cpu_asm_switch); |
| } |
| |
| #ifdef ASM_SUPPORTED |
| /*************************************************************************** |
| * binary pacthing |
| * pv_init_ops.patch_bundle |
| */ |
| |
| #define DEFINE_FUNC_GETREG(name, privop) \ |
| DEFINE_FUNC0(get_ ## name, \ |
| "break "__stringify(HYPERPRIVOP_GET_ ## privop) "\n") |
| |
| DEFINE_FUNC_GETREG(psr, PSR); |
| DEFINE_FUNC_GETREG(eflag, EFLAG); |
| DEFINE_FUNC_GETREG(ivr, IVR); |
| DEFINE_FUNC_GETREG(tpr, TPR); |
| |
| #define DEFINE_FUNC_SET_KR(n) \ |
| DEFINE_VOID_FUNC0(set_kr ## n, \ |
| ";;\n" \ |
| "mov r9 = r8\n" \ |
| "mov r8 = " #n "\n" \ |
| "break " __stringify(HYPERPRIVOP_SET_KR) "\n") |
| |
| DEFINE_FUNC_SET_KR(0); |
| DEFINE_FUNC_SET_KR(1); |
| DEFINE_FUNC_SET_KR(2); |
| DEFINE_FUNC_SET_KR(3); |
| DEFINE_FUNC_SET_KR(4); |
| DEFINE_FUNC_SET_KR(5); |
| DEFINE_FUNC_SET_KR(6); |
| DEFINE_FUNC_SET_KR(7); |
| |
| #define __DEFINE_FUNC_SETREG(name, privop) \ |
| DEFINE_VOID_FUNC0(name, \ |
| "break "__stringify(HYPERPRIVOP_ ## privop) "\n") |
| |
| #define DEFINE_FUNC_SETREG(name, privop) \ |
| __DEFINE_FUNC_SETREG(set_ ## name, SET_ ## privop) |
| |
| DEFINE_FUNC_SETREG(eflag, EFLAG); |
| DEFINE_FUNC_SETREG(tpr, TPR); |
| __DEFINE_FUNC_SETREG(eoi, EOI); |
| |
| extern const char xen_check_events[]; |
| extern const char __xen_intrin_local_irq_restore_direct_start[]; |
| extern const char __xen_intrin_local_irq_restore_direct_end[]; |
| extern const unsigned long __xen_intrin_local_irq_restore_direct_reloc; |
| |
| asm ( |
| ".align 32\n" |
| ".proc xen_check_events\n" |
| "xen_check_events:\n" |
| /* masked = 0 |
| * r9 = masked_addr - 1 |
| * = pending_intr_addr |
| */ |
| "st1.rel [r9] = r0, -1\n" |
| ";;\n" |
| /* r8 = pending_intr */ |
| "ld1.acq r11 = [r9]\n" |
| ";;\n" |
| /* p9 = interrupt pending? */ |
| "cmp.ne p9, p10 = r11, r0\n" |
| ";;\n" |
| "(p10) mf\n" |
| /* issue hypercall to trigger interrupt */ |
| "(p9) break " __stringify(HYPERPRIVOP_SSM_I) "\n" |
| "br.cond.sptk.many b6\n" |
| ".endp xen_check_events\n" |
| "\n" |
| ".align 32\n" |
| ".proc __xen_intrin_local_irq_restore_direct\n" |
| "__xen_intrin_local_irq_restore_direct:\n" |
| "__xen_intrin_local_irq_restore_direct_start:\n" |
| "1:\n" |
| "{\n" |
| "cmp.ne p6, p7 = r8, r0\n" |
| "mov r17 = ip\n" /* get ip to calc return address */ |
| "mov r9 = "__stringify(XEN_PSR_I_ADDR_ADDR) "\n" |
| ";;\n" |
| "}\n" |
| "{\n" |
| /* r9 = XEN_PSR_I_ADDR */ |
| "ld8 r9 = [r9]\n" |
| ";;\n" |
| /* r10 = masked previous value */ |
| "(p6) ld1.acq r10 = [r9]\n" |
| "adds r17 = 1f - 1b, r17\n" /* calculate return address */ |
| ";;\n" |
| "}\n" |
| "{\n" |
| /* p8 = !masked interrupt masked previously? */ |
| "(p6) cmp.ne.unc p8, p0 = r10, r0\n" |
| "\n" |
| /* p7 = else clause */ |
| "(p7) mov r11 = 1\n" |
| ";;\n" |
| "(p8) mov b6 = r17\n" /* set return address */ |
| "}\n" |
| "{\n" |
| /* masked = 1 */ |
| "(p7) st1.rel [r9] = r11\n" |
| "\n" |
| "[99:]\n" |
| "(p8) brl.cond.dptk.few xen_check_events\n" |
| "}\n" |
| /* pv calling stub is 5 bundles. fill nop to adjust return address */ |
| "{\n" |
| "nop 0\n" |
| "nop 0\n" |
| "nop 0\n" |
| "}\n" |
| "1:\n" |
| "__xen_intrin_local_irq_restore_direct_end:\n" |
| ".endp __xen_intrin_local_irq_restore_direct\n" |
| "\n" |
| ".align 8\n" |
| "__xen_intrin_local_irq_restore_direct_reloc:\n" |
| "data8 99b\n" |
| ); |
| |
| static struct paravirt_patch_bundle_elem xen_patch_bundle_elems[] |
| __initdata_or_module = |
| { |
| #define XEN_PATCH_BUNDLE_ELEM(name, type) \ |
| { \ |
| (void*)xen_ ## name ## _direct_start, \ |
| (void*)xen_ ## name ## _direct_end, \ |
| PARAVIRT_PATCH_TYPE_ ## type, \ |
| } |
| |
| XEN_PATCH_BUNDLE_ELEM(fc, FC), |
| XEN_PATCH_BUNDLE_ELEM(thash, THASH), |
| XEN_PATCH_BUNDLE_ELEM(get_cpuid, GET_CPUID), |
| XEN_PATCH_BUNDLE_ELEM(get_pmd, GET_PMD), |
| XEN_PATCH_BUNDLE_ELEM(ptcga, PTCGA), |
| XEN_PATCH_BUNDLE_ELEM(get_rr, GET_RR), |
| XEN_PATCH_BUNDLE_ELEM(set_rr, SET_RR), |
| XEN_PATCH_BUNDLE_ELEM(set_rr0_to_rr4, SET_RR0_TO_RR4), |
| XEN_PATCH_BUNDLE_ELEM(ssm_i, SSM_I), |
| XEN_PATCH_BUNDLE_ELEM(rsm_i, RSM_I), |
| XEN_PATCH_BUNDLE_ELEM(get_psr_i, GET_PSR_I), |
| { |
| (void*)__xen_intrin_local_irq_restore_direct_start, |
| (void*)__xen_intrin_local_irq_restore_direct_end, |
| PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE, |
| }, |
| |
| #define XEN_PATCH_BUNDLE_ELEM_GETREG(name, reg) \ |
| { \ |
| xen_get_ ## name ## _direct_start, \ |
| xen_get_ ## name ## _direct_end, \ |
| PARAVIRT_PATCH_TYPE_GETREG + _IA64_REG_ ## reg, \ |
| } |
| |
| XEN_PATCH_BUNDLE_ELEM_GETREG(psr, PSR), |
| XEN_PATCH_BUNDLE_ELEM_GETREG(eflag, AR_EFLAG), |
| |
| XEN_PATCH_BUNDLE_ELEM_GETREG(ivr, CR_IVR), |
| XEN_PATCH_BUNDLE_ELEM_GETREG(tpr, CR_TPR), |
| |
| XEN_PATCH_BUNDLE_ELEM_GETREG(itc, AR_ITC), |
| XEN_PATCH_BUNDLE_ELEM_GETREG(itm_with_offset, CR_ITM), |
| |
| |
| #define __XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \ |
| { \ |
| xen_ ## name ## _direct_start, \ |
| xen_ ## name ## _direct_end, \ |
| PARAVIRT_PATCH_TYPE_SETREG + _IA64_REG_ ## reg, \ |
| } |
| |
| #define XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \ |
| __XEN_PATCH_BUNDLE_ELEM_SETREG(set_ ## name, reg) |
| |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr0, AR_KR0), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr1, AR_KR1), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr2, AR_KR2), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr3, AR_KR3), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr4, AR_KR4), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr5, AR_KR5), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr6, AR_KR6), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(kr7, AR_KR7), |
| |
| XEN_PATCH_BUNDLE_ELEM_SETREG(eflag, AR_EFLAG), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(tpr, CR_TPR), |
| __XEN_PATCH_BUNDLE_ELEM_SETREG(eoi, CR_EOI), |
| |
| XEN_PATCH_BUNDLE_ELEM_SETREG(itc, AR_ITC), |
| XEN_PATCH_BUNDLE_ELEM_SETREG(itm_with_offset, CR_ITM), |
| }; |
| |
| static unsigned long __init_or_module |
| xen_patch_bundle(void *sbundle, void *ebundle, unsigned long type) |
| { |
| const unsigned long nelems = sizeof(xen_patch_bundle_elems) / |
| sizeof(xen_patch_bundle_elems[0]); |
| unsigned long used; |
| const struct paravirt_patch_bundle_elem *found; |
| |
| used = __paravirt_patch_apply_bundle(sbundle, ebundle, type, |
| xen_patch_bundle_elems, nelems, |
| &found); |
| |
| if (found == NULL) |
| /* fallback */ |
| return ia64_native_patch_bundle(sbundle, ebundle, type); |
| if (used == 0) |
| return used; |
| |
| /* relocation */ |
| switch (type) { |
| case PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE: { |
| unsigned long reloc = |
| __xen_intrin_local_irq_restore_direct_reloc; |
| unsigned long reloc_offset = reloc - (unsigned long) |
| __xen_intrin_local_irq_restore_direct_start; |
| unsigned long tag = (unsigned long)sbundle + reloc_offset; |
| paravirt_patch_reloc_brl(tag, xen_check_events); |
| break; |
| } |
| default: |
| /* nothing */ |
| break; |
| } |
| return used; |
| } |
| #endif /* ASM_SUPPOTED */ |
| |
| const struct paravirt_patch_branch_target xen_branch_target[] |
| __initconst = { |
| #define PARAVIRT_BR_TARGET(name, type) \ |
| { \ |
| &xen_ ## name, \ |
| PARAVIRT_PATCH_TYPE_BR_ ## type, \ |
| } |
| PARAVIRT_BR_TARGET(switch_to, SWITCH_TO), |
| PARAVIRT_BR_TARGET(leave_syscall, LEAVE_SYSCALL), |
| PARAVIRT_BR_TARGET(work_processed_syscall, WORK_PROCESSED_SYSCALL), |
| PARAVIRT_BR_TARGET(leave_kernel, LEAVE_KERNEL), |
| }; |
| |
| static void __init |
| xen_patch_branch(unsigned long tag, unsigned long type) |
| { |
| __paravirt_patch_apply_branch(tag, type, xen_branch_target, |
| ARRAY_SIZE(xen_branch_target)); |
| } |