/****************************************************************************** * arch/ia64/xen/xen_pv_ops.c * * Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp> * VA Linux Systems Japan K.K. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA * */ #include <linux/console.h> #include <linux/irq.h> #include <linux/kernel.h> #include <linux/pm.h> #include <linux/unistd.h> #include <asm/xen/hypervisor.h> #include <asm/xen/xencomm.h> #include <asm/xen/privop.h> #include "irq_xen.h" #include "time.h" /*************************************************************************** * general info */ static struct pv_info xen_info __initdata = { .kernel_rpl = 2, /* or 1: determin at runtime */ .paravirt_enabled = 1, .name = "Xen/ia64", }; #define IA64_RSC_PL_SHIFT 2 #define IA64_RSC_PL_BIT_SIZE 2 #define IA64_RSC_PL_MASK \ (((1UL << IA64_RSC_PL_BIT_SIZE) - 1) << IA64_RSC_PL_SHIFT) static void __init xen_info_init(void) { /* Xenified Linux/ia64 may run on pl = 1 or 2. * determin at run time. */ unsigned long rsc = ia64_getreg(_IA64_REG_AR_RSC); unsigned int rpl = (rsc & IA64_RSC_PL_MASK) >> IA64_RSC_PL_SHIFT; xen_info.kernel_rpl = rpl; } /*************************************************************************** * pv_init_ops * initialization hooks. */ static void xen_panic_hypercall(struct unw_frame_info *info, void *arg) { current->thread.ksp = (__u64)info->sw - 16; HYPERVISOR_shutdown(SHUTDOWN_crash); /* we're never actually going to get here... */ } static int xen_panic_event(struct notifier_block *this, unsigned long event, void *ptr) { unw_init_running(xen_panic_hypercall, NULL); /* we're never actually going to get here... */ return NOTIFY_DONE; } static struct notifier_block xen_panic_block = { xen_panic_event, NULL, 0 /* try to go last */ }; static void xen_pm_power_off(void) { local_irq_disable(); HYPERVISOR_shutdown(SHUTDOWN_poweroff); } static void __init xen_banner(void) { printk(KERN_INFO "Running on Xen! pl = %d start_info_pfn=0x%lx nr_pages=%ld " "flags=0x%x\n", xen_info.kernel_rpl, HYPERVISOR_shared_info->arch.start_info_pfn, xen_start_info->nr_pages, xen_start_info->flags); } static int __init xen_reserve_memory(struct rsvd_region *region) { region->start = (unsigned long)__va( (HYPERVISOR_shared_info->arch.start_info_pfn << PAGE_SHIFT)); region->end = region->start + PAGE_SIZE; return 1; } static void __init xen_arch_setup_early(void) { struct shared_info *s; BUG_ON(!xen_pv_domain()); s = HYPERVISOR_shared_info; xen_start_info = __va(s->arch.start_info_pfn << PAGE_SHIFT); /* Must be done before any hypercall. */ xencomm_initialize(); xen_setup_features(); /* Register a call for panic conditions. */ atomic_notifier_chain_register(&panic_notifier_list, &xen_panic_block); pm_power_off = xen_pm_power_off; xen_ia64_enable_opt_feature(); } static void __init xen_arch_setup_console(char **cmdline_p) { add_preferred_console("xenboot", 0, NULL); add_preferred_console("tty", 0, NULL); /* use hvc_xen */ add_preferred_console("hvc", 0, NULL); #if !defined(CONFIG_VT) || !defined(CONFIG_DUMMY_CONSOLE) conswitchp = NULL; #endif } static int __init xen_arch_setup_nomca(void) { return 1; } static void __init xen_post_smp_prepare_boot_cpu(void) { xen_setup_vcpu_info_placement(); } #ifdef ASM_SUPPORTED static unsigned long __init_or_module xen_patch_bundle(void *sbundle, void *ebundle, unsigned long type); #endif static void __init xen_patch_branch(unsigned long tag, unsigned long type); static const struct pv_init_ops xen_init_ops __initconst = { .banner = xen_banner, .reserve_memory = xen_reserve_memory, .arch_setup_early = xen_arch_setup_early, .arch_setup_console = xen_arch_setup_console, .arch_setup_nomca = xen_arch_setup_nomca, .post_smp_prepare_boot_cpu = xen_post_smp_prepare_boot_cpu, #ifdef ASM_SUPPORTED .patch_bundle = xen_patch_bundle, #endif .patch_branch = xen_patch_branch, }; /*************************************************************************** * pv_fsys_data * addresses for fsys */ extern unsigned long xen_fsyscall_table[NR_syscalls]; extern char xen_fsys_bubble_down[]; struct pv_fsys_data xen_fsys_data __initdata = { .fsyscall_table = (unsigned long *)xen_fsyscall_table, .fsys_bubble_down = (void *)xen_fsys_bubble_down, }; /*************************************************************************** * pv_patchdata * patchdata addresses */ #define DECLARE(name) \ extern unsigned long __xen_start_gate_##name##_patchlist[]; \ extern unsigned long __xen_end_gate_##name##_patchlist[] DECLARE(fsyscall); DECLARE(brl_fsys_bubble_down); DECLARE(vtop); DECLARE(mckinley_e9); extern unsigned long __xen_start_gate_section[]; #define ASSIGN(name) \ .start_##name##_patchlist = \ (unsigned long)__xen_start_gate_##name##_patchlist, \ .end_##name##_patchlist = \ (unsigned long)__xen_end_gate_##name##_patchlist static struct pv_patchdata xen_patchdata __initdata = { ASSIGN(fsyscall), ASSIGN(brl_fsys_bubble_down), ASSIGN(vtop), ASSIGN(mckinley_e9), .gate_section = (void*)__xen_start_gate_section, }; /*************************************************************************** * pv_cpu_ops * intrinsics hooks. */ #ifndef ASM_SUPPORTED static void xen_set_itm_with_offset(unsigned long val) { /* ia64_cpu_local_tick() calls this with interrupt enabled. */ /* WARN_ON(!irqs_disabled()); */ xen_set_itm(val - XEN_MAPPEDREGS->itc_offset); } static unsigned long xen_get_itm_with_offset(void) { /* unused at this moment */ printk(KERN_DEBUG "%s is called.\n", __func__); WARN_ON(!irqs_disabled()); return ia64_native_getreg(_IA64_REG_CR_ITM) + XEN_MAPPEDREGS->itc_offset; } /* ia64_set_itc() is only called by * cpu_init() with ia64_set_itc(0) and ia64_sync_itc(). * So XEN_MAPPEDRESG->itc_offset cal be considered as almost constant. */ static void xen_set_itc(unsigned long val) { unsigned long mitc; WARN_ON(!irqs_disabled()); mitc = ia64_native_getreg(_IA64_REG_AR_ITC); XEN_MAPPEDREGS->itc_offset = val - mitc; XEN_MAPPEDREGS->itc_last = val; } static unsigned long xen_get_itc(void) { unsigned long res; unsigned long itc_offset; unsigned long itc_last; unsigned long ret_itc_last; itc_offset = XEN_MAPPEDREGS->itc_offset; do { itc_last = XEN_MAPPEDREGS->itc_last; res = ia64_native_getreg(_IA64_REG_AR_ITC); res += itc_offset; if (itc_last >= res) res = itc_last + 1; ret_itc_last = cmpxchg(&XEN_MAPPEDREGS->itc_last, itc_last, res); } while (unlikely(ret_itc_last != itc_last)); return res; #if 0 /* ia64_itc_udelay() calls ia64_get_itc() with interrupt enabled. Should it be paravirtualized instead? */ WARN_ON(!irqs_disabled()); itc_offset = XEN_MAPPEDREGS->itc_offset; itc_last = XEN_MAPPEDREGS->itc_last; res = ia64_native_getreg(_IA64_REG_AR_ITC); res += itc_offset; if (itc_last >= res) res = itc_last + 1; XEN_MAPPEDREGS->itc_last = res; return res; #endif } static void xen_setreg(int regnum, unsigned long val) { switch (regnum) { case _IA64_REG_AR_KR0 ... _IA64_REG_AR_KR7: xen_set_kr(regnum - _IA64_REG_AR_KR0, val); break; case _IA64_REG_AR_ITC: xen_set_itc(val); break; case _IA64_REG_CR_TPR: xen_set_tpr(val); break; case _IA64_REG_CR_ITM: xen_set_itm_with_offset(val); break; case _IA64_REG_CR_EOI: xen_eoi(val); break; default: ia64_native_setreg_func(regnum, val); break; } } static unsigned long xen_getreg(int regnum) { unsigned long res; switch (regnum) { case _IA64_REG_PSR: res = xen_get_psr(); break; case _IA64_REG_AR_ITC: res = xen_get_itc(); break; case _IA64_REG_CR_ITM: res = xen_get_itm_with_offset(); break; case _IA64_REG_CR_IVR: res = xen_get_ivr(); break; case _IA64_REG_CR_TPR: res = xen_get_tpr(); break; default: res = ia64_native_getreg_func(regnum); break; } return res; } /* turning on interrupts is a bit more complicated.. write to the * memory-mapped virtual psr.i bit first (to avoid race condition), * then if any interrupts were pending, we have to execute a hyperprivop * to ensure the pending interrupt gets delivered; else we're done! */ static void xen_ssm_i(void) { int old = xen_get_virtual_psr_i(); xen_set_virtual_psr_i(1); barrier(); if (!old && xen_get_virtual_pend()) xen_hyper_ssm_i(); } /* turning off interrupts can be paravirtualized simply by writing * to a memory-mapped virtual psr.i bit (implemented as a 16-bit bool) */ static void xen_rsm_i(void) { xen_set_virtual_psr_i(0); barrier(); } static unsigned long xen_get_psr_i(void) { return xen_get_virtual_psr_i() ? IA64_PSR_I : 0; } static void xen_intrin_local_irq_restore(unsigned long mask) { if (mask & IA64_PSR_I) xen_ssm_i(); else xen_rsm_i(); } #else #define __DEFINE_FUNC(name, code) \ extern const char xen_ ## name ## _direct_start[]; \ extern const char xen_ ## name ## _direct_end[]; \ asm (".align 32\n" \ ".proc xen_" #name "\n" \ "xen_" #name ":\n" \ "xen_" #name "_direct_start:\n" \ code \ "xen_" #name "_direct_end:\n" \ "br.cond.sptk.many b6\n" \ ".endp xen_" #name "\n") #define DEFINE_VOID_FUNC0(name, code) \ extern void \ xen_ ## name (void); \ __DEFINE_FUNC(name, code) #define DEFINE_VOID_FUNC1(name, code) \ extern void \ xen_ ## name (unsigned long arg); \ __DEFINE_FUNC(name, code) #define DEFINE_VOID_FUNC1_VOID(name, code) \ extern void \ xen_ ## name (void *arg); \ __DEFINE_FUNC(name, code) #define DEFINE_VOID_FUNC2(name, code) \ extern void \ xen_ ## name (unsigned long arg0, \ unsigned long arg1); \ __DEFINE_FUNC(name, code) #define DEFINE_FUNC0(name, code) \ extern unsigned long \ xen_ ## name (void); \ __DEFINE_FUNC(name, code) #define DEFINE_FUNC1(name, type, code) \ extern unsigned long \ xen_ ## name (type arg); \ __DEFINE_FUNC(name, code) #define XEN_PSR_I_ADDR_ADDR (XSI_BASE + XSI_PSR_I_ADDR_OFS) /* * static void xen_set_itm_with_offset(unsigned long val) * xen_set_itm(val - XEN_MAPPEDREGS->itc_offset); */ /* 2 bundles */ DEFINE_VOID_FUNC1(set_itm_with_offset, "mov r2 = " __stringify(XSI_BASE) " + " __stringify(XSI_ITC_OFFSET_OFS) "\n" ";;\n" "ld8 r3 = [r2]\n" ";;\n" "sub r8 = r8, r3\n" "break " __stringify(HYPERPRIVOP_SET_ITM) "\n"); /* * static unsigned long xen_get_itm_with_offset(void) * return ia64_native_getreg(_IA64_REG_CR_ITM) + XEN_MAPPEDREGS->itc_offset; */ /* 2 bundles */ DEFINE_FUNC0(get_itm_with_offset, "mov r2 = " __stringify(XSI_BASE) " + " __stringify(XSI_ITC_OFFSET_OFS) "\n" ";;\n" "ld8 r3 = [r2]\n" "mov r8 = cr.itm\n" ";;\n" "add r8 = r8, r2\n"); /* * static void xen_set_itc(unsigned long val) * unsigned long mitc; * * WARN_ON(!irqs_disabled()); * mitc = ia64_native_getreg(_IA64_REG_AR_ITC); * XEN_MAPPEDREGS->itc_offset = val - mitc; * XEN_MAPPEDREGS->itc_last = val; */ /* 2 bundles */ DEFINE_VOID_FUNC1(set_itc, "mov r2 = " __stringify(XSI_BASE) " + " __stringify(XSI_ITC_LAST_OFS) "\n" "mov r3 = ar.itc\n" ";;\n" "sub r3 = r8, r3\n" "st8 [r2] = r8, " __stringify(XSI_ITC_LAST_OFS) " - " __stringify(XSI_ITC_OFFSET_OFS) "\n" ";;\n" "st8 [r2] = r3\n"); /* * static unsigned long xen_get_itc(void) * unsigned long res; * unsigned long itc_offset; * unsigned long itc_last; * unsigned long ret_itc_last; * * itc_offset = XEN_MAPPEDREGS->itc_offset; * do { * itc_last = XEN_MAPPEDREGS->itc_last; * res = ia64_native_getreg(_IA64_REG_AR_ITC); * res += itc_offset; * if (itc_last >= res) * res = itc_last + 1; * ret_itc_last = cmpxchg(&XEN_MAPPEDREGS->itc_last, * itc_last, res); * } while (unlikely(ret_itc_last != itc_last)); * return res; */ /* 5 bundles */ DEFINE_FUNC0(get_itc, "mov r2 = " __stringify(XSI_BASE) " + " __stringify(XSI_ITC_OFFSET_OFS) "\n" ";;\n" "ld8 r9 = [r2], " __stringify(XSI_ITC_LAST_OFS) " - " __stringify(XSI_ITC_OFFSET_OFS) "\n" /* r9 = itc_offset */ /* r2 = XSI_ITC_OFFSET */ "888:\n" "mov r8 = ar.itc\n" /* res = ar.itc */ ";;\n" "ld8 r3 = [r2]\n" /* r3 = itc_last */ "add r8 = r8, r9\n" /* res = ar.itc + itc_offset */ ";;\n" "cmp.gtu p6, p0 = r3, r8\n" ";;\n" "(p6) add r8 = 1, r3\n" /* if (itc_last > res) itc_last + 1 */ ";;\n" "mov ar.ccv = r8\n" ";;\n" "cmpxchg8.acq r10 = [r2], r8, ar.ccv\n" ";;\n" "cmp.ne p6, p0 = r10, r3\n" "(p6) hint @pause\n" "(p6) br.cond.spnt 888b\n"); DEFINE_VOID_FUNC1_VOID(fc, "break " __stringify(HYPERPRIVOP_FC) "\n"); /* * psr_i_addr_addr = XEN_PSR_I_ADDR_ADDR * masked_addr = *psr_i_addr_addr * pending_intr_addr = masked_addr - 1 * if (val & IA64_PSR_I) { * masked = *masked_addr * *masked_addr = 0:xen_set_virtual_psr_i(1) * compiler barrier * if (masked) { * uint8_t pending = *pending_intr_addr; * if (pending) * XEN_HYPER_SSM_I * } * } else { * *masked_addr = 1:xen_set_virtual_psr_i(0) * } */ /* 6 bundles */ DEFINE_VOID_FUNC1(intrin_local_irq_restore, /* r8 = input value: 0 or IA64_PSR_I * p6 = (flags & IA64_PSR_I) * = if clause * p7 = !(flags & IA64_PSR_I) * = else clause */ "cmp.ne p6, p7 = r8, r0\n" "mov r9 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" ";;\n" /* r9 = XEN_PSR_I_ADDR */ "ld8 r9 = [r9]\n" ";;\n" /* r10 = masked previous value */ "(p6) ld1.acq r10 = [r9]\n" ";;\n" /* p8 = !masked interrupt masked previously? */ "(p6) cmp.ne.unc p8, p0 = r10, r0\n" /* p7 = else clause */ "(p7) mov r11 = 1\n" ";;\n" /* masked = 1 */ "(p7) st1.rel [r9] = r11\n" /* p6 = if clause */ /* masked = 0 * r9 = masked_addr - 1 * = pending_intr_addr */ "(p8) st1.rel [r9] = r0, -1\n" ";;\n" /* r8 = pending_intr */ "(p8) ld1.acq r11 = [r9]\n" ";;\n" /* p9 = interrupt pending? */ "(p8) cmp.ne.unc p9, p10 = r11, r0\n" ";;\n" "(p10) mf\n" /* issue hypercall to trigger interrupt */ "(p9) break " __stringify(HYPERPRIVOP_SSM_I) "\n"); DEFINE_VOID_FUNC2(ptcga, "break " __stringify(HYPERPRIVOP_PTC_GA) "\n"); DEFINE_VOID_FUNC2(set_rr, "break " __stringify(HYPERPRIVOP_SET_RR) "\n"); /* * tmp = XEN_MAPPEDREGS->interrupt_mask_addr = XEN_PSR_I_ADDR_ADDR; * tmp = *tmp * tmp = *tmp; * psr_i = tmp? 0: IA64_PSR_I; */ /* 4 bundles */ DEFINE_FUNC0(get_psr_i, "mov r9 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" ";;\n" "ld8 r9 = [r9]\n" /* r9 = XEN_PSR_I_ADDR */ "mov r8 = 0\n" /* psr_i = 0 */ ";;\n" "ld1.acq r9 = [r9]\n" /* r9 = XEN_PSR_I */ ";;\n" "cmp.eq.unc p6, p0 = r9, r0\n" /* p6 = (XEN_PSR_I != 0) */ ";;\n" "(p6) mov r8 = " __stringify(1 << IA64_PSR_I_BIT) "\n"); DEFINE_FUNC1(thash, unsigned long, "break " __stringify(HYPERPRIVOP_THASH) "\n"); DEFINE_FUNC1(get_cpuid, int, "break " __stringify(HYPERPRIVOP_GET_CPUID) "\n"); DEFINE_FUNC1(get_pmd, int, "break " __stringify(HYPERPRIVOP_GET_PMD) "\n"); DEFINE_FUNC1(get_rr, unsigned long, "break " __stringify(HYPERPRIVOP_GET_RR) "\n"); /* * void xen_privop_ssm_i(void) * * int masked = !xen_get_virtual_psr_i(); * // masked = *(*XEN_MAPPEDREGS->interrupt_mask_addr) * xen_set_virtual_psr_i(1) * // *(*XEN_MAPPEDREGS->interrupt_mask_addr) = 0 * // compiler barrier * if (masked) { * uint8_t* pend_int_addr = * (uint8_t*)(*XEN_MAPPEDREGS->interrupt_mask_addr) - 1; * uint8_t pending = *pend_int_addr; * if (pending) * XEN_HYPER_SSM_I * } */ /* 4 bundles */ DEFINE_VOID_FUNC0(ssm_i, "mov r8 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" ";;\n" "ld8 r8 = [r8]\n" /* r8 = XEN_PSR_I_ADDR */ ";;\n" "ld1.acq r9 = [r8]\n" /* r9 = XEN_PSR_I */ ";;\n" "st1.rel [r8] = r0, -1\n" /* psr_i = 0. enable interrupt * r8 = XEN_PSR_I_ADDR - 1 * = pend_int_addr */ "cmp.eq.unc p0, p6 = r9, r0\n"/* p6 = !XEN_PSR_I * previously interrupt * masked? */ ";;\n" "(p6) ld1.acq r8 = [r8]\n" /* r8 = xen_pend_int */ ";;\n" "(p6) cmp.eq.unc p6, p7 = r8, r0\n" /*interrupt pending?*/ ";;\n" /* issue hypercall to get interrupt */ "(p7) break " __stringify(HYPERPRIVOP_SSM_I) "\n" ";;\n"); /* * psr_i_addr_addr = XEN_MAPPEDREGS->interrupt_mask_addr * = XEN_PSR_I_ADDR_ADDR; * psr_i_addr = *psr_i_addr_addr; * *psr_i_addr = 1; */ /* 2 bundles */ DEFINE_VOID_FUNC0(rsm_i, "mov r8 = " __stringify(XEN_PSR_I_ADDR_ADDR) "\n" /* r8 = XEN_PSR_I_ADDR */ "mov r9 = 1\n" ";;\n" "ld8 r8 = [r8]\n" /* r8 = XEN_PSR_I */ ";;\n" "st1.rel [r8] = r9\n"); /* XEN_PSR_I = 1 */ extern void xen_set_rr0_to_rr4(unsigned long val0, unsigned long val1, unsigned long val2, unsigned long val3, unsigned long val4); __DEFINE_FUNC(set_rr0_to_rr4, "break " __stringify(HYPERPRIVOP_SET_RR0_TO_RR4) "\n"); extern unsigned long xen_getreg(int regnum); #define __DEFINE_GET_REG(id, privop) \ "mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \ ";;\n" \ "cmp.eq p6, p0 = r2, r8\n" \ ";;\n" \ "(p6) break " __stringify(HYPERPRIVOP_GET_ ## privop) "\n" \ "(p6) br.cond.sptk.many b6\n" \ ";;\n" __DEFINE_FUNC(getreg, __DEFINE_GET_REG(PSR, PSR) /* get_itc */ "mov r2 = " __stringify(_IA64_REG_AR_ITC) "\n" ";;\n" "cmp.eq p6, p0 = r2, r8\n" ";;\n" "(p6) br.cond.spnt xen_get_itc\n" ";;\n" /* get itm */ "mov r2 = " __stringify(_IA64_REG_CR_ITM) "\n" ";;\n" "cmp.eq p6, p0 = r2, r8\n" ";;\n" "(p6) br.cond.spnt xen_get_itm_with_offset\n" ";;\n" __DEFINE_GET_REG(CR_IVR, IVR) __DEFINE_GET_REG(CR_TPR, TPR) /* fall back */ "movl r2 = ia64_native_getreg_func\n" ";;\n" "mov b7 = r2\n" ";;\n" "br.cond.sptk.many b7\n"); extern void xen_setreg(int regnum, unsigned long val); #define __DEFINE_SET_REG(id, privop) \ "mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \ ";;\n" \ "cmp.eq p6, p0 = r2, r9\n" \ ";;\n" \ "(p6) break " __stringify(HYPERPRIVOP_ ## privop) "\n" \ "(p6) br.cond.sptk.many b6\n" \ ";;\n" __DEFINE_FUNC(setreg, /* kr0 .. kr 7*/ /* * if (_IA64_REG_AR_KR0 <= regnum && * regnum <= _IA64_REG_AR_KR7) { * register __index asm ("r8") = regnum - _IA64_REG_AR_KR0 * register __val asm ("r9") = val * "break HYPERPRIVOP_SET_KR" * } */ "mov r17 = r9\n" "mov r2 = " __stringify(_IA64_REG_AR_KR0) "\n" ";;\n" "cmp.ge p6, p0 = r9, r2\n" "sub r17 = r17, r2\n" ";;\n" "(p6) cmp.ge.unc p7, p0 = " __stringify(_IA64_REG_AR_KR7) " - " __stringify(_IA64_REG_AR_KR0) ", r17\n" ";;\n" "(p7) mov r9 = r8\n" ";;\n" "(p7) mov r8 = r17\n" "(p7) break " __stringify(HYPERPRIVOP_SET_KR) "\n" /* set itm */ "mov r2 = " __stringify(_IA64_REG_CR_ITM) "\n" ";;\n" "cmp.eq p6, p0 = r2, r8\n" ";;\n" "(p6) br.cond.spnt xen_set_itm_with_offset\n" /* set itc */ "mov r2 = " __stringify(_IA64_REG_AR_ITC) "\n" ";;\n" "cmp.eq p6, p0 = r2, r8\n" ";;\n" "(p6) br.cond.spnt xen_set_itc\n" __DEFINE_SET_REG(CR_TPR, SET_TPR) __DEFINE_SET_REG(CR_EOI, EOI) /* fall back */ "movl r2 = ia64_native_setreg_func\n" ";;\n" "mov b7 = r2\n" ";;\n" "br.cond.sptk.many b7\n"); #endif static const struct pv_cpu_ops xen_cpu_ops __initconst = { .fc = xen_fc, .thash = xen_thash, .get_cpuid = xen_get_cpuid, .get_pmd = xen_get_pmd, .getreg = xen_getreg, .setreg = xen_setreg, .ptcga = xen_ptcga, .get_rr = xen_get_rr, .set_rr = xen_set_rr, .set_rr0_to_rr4 = xen_set_rr0_to_rr4, .ssm_i = xen_ssm_i, .rsm_i = xen_rsm_i, .get_psr_i = xen_get_psr_i, .intrin_local_irq_restore = xen_intrin_local_irq_restore, }; /****************************************************************************** * replacement of hand written assembly codes. */ extern char xen_switch_to; extern char xen_leave_syscall; extern char xen_work_processed_syscall; extern char xen_leave_kernel; const struct pv_cpu_asm_switch xen_cpu_asm_switch = { .switch_to = (unsigned long)&xen_switch_to, .leave_syscall = (unsigned long)&xen_leave_syscall, .work_processed_syscall = (unsigned long)&xen_work_processed_syscall, .leave_kernel = (unsigned long)&xen_leave_kernel, }; /*************************************************************************** * pv_iosapic_ops * iosapic read/write hooks. */ static void xen_pcat_compat_init(void) { /* nothing */ } static struct irq_chip* xen_iosapic_get_irq_chip(unsigned long trigger) { return NULL; } static unsigned int xen_iosapic_read(char __iomem *iosapic, unsigned int reg) { struct physdev_apic apic_op; int ret; apic_op.apic_physbase = (unsigned long)iosapic - __IA64_UNCACHED_OFFSET; apic_op.reg = reg; ret = HYPERVISOR_physdev_op(PHYSDEVOP_apic_read, &apic_op); if (ret) return ret; return apic_op.value; } static void xen_iosapic_write(char __iomem *iosapic, unsigned int reg, u32 val) { struct physdev_apic apic_op; apic_op.apic_physbase = (unsigned long)iosapic - __IA64_UNCACHED_OFFSET; apic_op.reg = reg; apic_op.value = val; HYPERVISOR_physdev_op(PHYSDEVOP_apic_write, &apic_op); } static struct pv_iosapic_ops xen_iosapic_ops __initdata = { .pcat_compat_init = xen_pcat_compat_init, .__get_irq_chip = xen_iosapic_get_irq_chip, .__read = xen_iosapic_read, .__write = xen_iosapic_write, }; /*************************************************************************** * pv_ops initialization */ void __init xen_setup_pv_ops(void) { xen_info_init(); pv_info = xen_info; pv_init_ops = xen_init_ops; pv_fsys_data = xen_fsys_data; pv_patchdata = xen_patchdata; pv_cpu_ops = xen_cpu_ops; pv_iosapic_ops = xen_iosapic_ops; pv_irq_ops = xen_irq_ops; pv_time_ops = xen_time_ops; paravirt_cpu_asm_init(&xen_cpu_asm_switch); } #ifdef ASM_SUPPORTED /*************************************************************************** * binary pacthing * pv_init_ops.patch_bundle */ #define DEFINE_FUNC_GETREG(name, privop) \ DEFINE_FUNC0(get_ ## name, \ "break "__stringify(HYPERPRIVOP_GET_ ## privop) "\n") DEFINE_FUNC_GETREG(psr, PSR); DEFINE_FUNC_GETREG(eflag, EFLAG); DEFINE_FUNC_GETREG(ivr, IVR); DEFINE_FUNC_GETREG(tpr, TPR); #define DEFINE_FUNC_SET_KR(n) \ DEFINE_VOID_FUNC0(set_kr ## n, \ ";;\n" \ "mov r9 = r8\n" \ "mov r8 = " #n "\n" \ "break " __stringify(HYPERPRIVOP_SET_KR) "\n") DEFINE_FUNC_SET_KR(0); DEFINE_FUNC_SET_KR(1); DEFINE_FUNC_SET_KR(2); DEFINE_FUNC_SET_KR(3); DEFINE_FUNC_SET_KR(4); DEFINE_FUNC_SET_KR(5); DEFINE_FUNC_SET_KR(6); DEFINE_FUNC_SET_KR(7); #define __DEFINE_FUNC_SETREG(name, privop) \ DEFINE_VOID_FUNC0(name, \ "break "__stringify(HYPERPRIVOP_ ## privop) "\n") #define DEFINE_FUNC_SETREG(name, privop) \ __DEFINE_FUNC_SETREG(set_ ## name, SET_ ## privop) DEFINE_FUNC_SETREG(eflag, EFLAG); DEFINE_FUNC_SETREG(tpr, TPR); __DEFINE_FUNC_SETREG(eoi, EOI); extern const char xen_check_events[]; extern const char __xen_intrin_local_irq_restore_direct_start[]; extern const char __xen_intrin_local_irq_restore_direct_end[]; extern const unsigned long __xen_intrin_local_irq_restore_direct_reloc; asm ( ".align 32\n" ".proc xen_check_events\n" "xen_check_events:\n" /* masked = 0 * r9 = masked_addr - 1 * = pending_intr_addr */ "st1.rel [r9] = r0, -1\n" ";;\n" /* r8 = pending_intr */ "ld1.acq r11 = [r9]\n" ";;\n" /* p9 = interrupt pending? */ "cmp.ne p9, p10 = r11, r0\n" ";;\n" "(p10) mf\n" /* issue hypercall to trigger interrupt */ "(p9) break " __stringify(HYPERPRIVOP_SSM_I) "\n" "br.cond.sptk.many b6\n" ".endp xen_check_events\n" "\n" ".align 32\n" ".proc __xen_intrin_local_irq_restore_direct\n" "__xen_intrin_local_irq_restore_direct:\n" "__xen_intrin_local_irq_restore_direct_start:\n" "1:\n" "{\n" "cmp.ne p6, p7 = r8, r0\n" "mov r17 = ip\n" /* get ip to calc return address */ "mov r9 = "__stringify(XEN_PSR_I_ADDR_ADDR) "\n" ";;\n" "}\n" "{\n" /* r9 = XEN_PSR_I_ADDR */ "ld8 r9 = [r9]\n" ";;\n" /* r10 = masked previous value */ "(p6) ld1.acq r10 = [r9]\n" "adds r17 = 1f - 1b, r17\n" /* calculate return address */ ";;\n" "}\n" "{\n" /* p8 = !masked interrupt masked previously? */ "(p6) cmp.ne.unc p8, p0 = r10, r0\n" "\n" /* p7 = else clause */ "(p7) mov r11 = 1\n" ";;\n" "(p8) mov b6 = r17\n" /* set return address */ "}\n" "{\n" /* masked = 1 */ "(p7) st1.rel [r9] = r11\n" "\n" "[99:]\n" "(p8) brl.cond.dptk.few xen_check_events\n" "}\n" /* pv calling stub is 5 bundles. fill nop to adjust return address */ "{\n" "nop 0\n" "nop 0\n" "nop 0\n" "}\n" "1:\n" "__xen_intrin_local_irq_restore_direct_end:\n" ".endp __xen_intrin_local_irq_restore_direct\n" "\n" ".align 8\n" "__xen_intrin_local_irq_restore_direct_reloc:\n" "data8 99b\n" ); static struct paravirt_patch_bundle_elem xen_patch_bundle_elems[] __initdata_or_module = { #define XEN_PATCH_BUNDLE_ELEM(name, type) \ { \ (void*)xen_ ## name ## _direct_start, \ (void*)xen_ ## name ## _direct_end, \ PARAVIRT_PATCH_TYPE_ ## type, \ } XEN_PATCH_BUNDLE_ELEM(fc, FC), XEN_PATCH_BUNDLE_ELEM(thash, THASH), XEN_PATCH_BUNDLE_ELEM(get_cpuid, GET_CPUID), XEN_PATCH_BUNDLE_ELEM(get_pmd, GET_PMD), XEN_PATCH_BUNDLE_ELEM(ptcga, PTCGA), XEN_PATCH_BUNDLE_ELEM(get_rr, GET_RR), XEN_PATCH_BUNDLE_ELEM(set_rr, SET_RR), XEN_PATCH_BUNDLE_ELEM(set_rr0_to_rr4, SET_RR0_TO_RR4), XEN_PATCH_BUNDLE_ELEM(ssm_i, SSM_I), XEN_PATCH_BUNDLE_ELEM(rsm_i, RSM_I), XEN_PATCH_BUNDLE_ELEM(get_psr_i, GET_PSR_I), { (void*)__xen_intrin_local_irq_restore_direct_start, (void*)__xen_intrin_local_irq_restore_direct_end, PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE, }, #define XEN_PATCH_BUNDLE_ELEM_GETREG(name, reg) \ { \ xen_get_ ## name ## _direct_start, \ xen_get_ ## name ## _direct_end, \ PARAVIRT_PATCH_TYPE_GETREG + _IA64_REG_ ## reg, \ } XEN_PATCH_BUNDLE_ELEM_GETREG(psr, PSR), XEN_PATCH_BUNDLE_ELEM_GETREG(eflag, AR_EFLAG), XEN_PATCH_BUNDLE_ELEM_GETREG(ivr, CR_IVR), XEN_PATCH_BUNDLE_ELEM_GETREG(tpr, CR_TPR), XEN_PATCH_BUNDLE_ELEM_GETREG(itc, AR_ITC), XEN_PATCH_BUNDLE_ELEM_GETREG(itm_with_offset, CR_ITM), #define __XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \ { \ xen_ ## name ## _direct_start, \ xen_ ## name ## _direct_end, \ PARAVIRT_PATCH_TYPE_SETREG + _IA64_REG_ ## reg, \ } #define XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \ __XEN_PATCH_BUNDLE_ELEM_SETREG(set_ ## name, reg) XEN_PATCH_BUNDLE_ELEM_SETREG(kr0, AR_KR0), XEN_PATCH_BUNDLE_ELEM_SETREG(kr1, AR_KR1), XEN_PATCH_BUNDLE_ELEM_SETREG(kr2, AR_KR2), XEN_PATCH_BUNDLE_ELEM_SETREG(kr3, AR_KR3), XEN_PATCH_BUNDLE_ELEM_SETREG(kr4, AR_KR4), XEN_PATCH_BUNDLE_ELEM_SETREG(kr5, AR_KR5), XEN_PATCH_BUNDLE_ELEM_SETREG(kr6, AR_KR6), XEN_PATCH_BUNDLE_ELEM_SETREG(kr7, AR_KR7), XEN_PATCH_BUNDLE_ELEM_SETREG(eflag, AR_EFLAG), XEN_PATCH_BUNDLE_ELEM_SETREG(tpr, CR_TPR), __XEN_PATCH_BUNDLE_ELEM_SETREG(eoi, CR_EOI), XEN_PATCH_BUNDLE_ELEM_SETREG(itc, AR_ITC), XEN_PATCH_BUNDLE_ELEM_SETREG(itm_with_offset, CR_ITM), }; static unsigned long __init_or_module xen_patch_bundle(void *sbundle, void *ebundle, unsigned long type) { const unsigned long nelems = sizeof(xen_patch_bundle_elems) / sizeof(xen_patch_bundle_elems[0]); unsigned long used; const struct paravirt_patch_bundle_elem *found; used = __paravirt_patch_apply_bundle(sbundle, ebundle, type, xen_patch_bundle_elems, nelems, &found); if (found == NULL) /* fallback */ return ia64_native_patch_bundle(sbundle, ebundle, type); if (used == 0) return used; /* relocation */ switch (type) { case PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE: { unsigned long reloc = __xen_intrin_local_irq_restore_direct_reloc; unsigned long reloc_offset = reloc - (unsigned long) __xen_intrin_local_irq_restore_direct_start; unsigned long tag = (unsigned long)sbundle + reloc_offset; paravirt_patch_reloc_brl(tag, xen_check_events); break; } default: /* nothing */ break; } return used; } #endif /* ASM_SUPPOTED */ const struct paravirt_patch_branch_target xen_branch_target[] __initconst = { #define PARAVIRT_BR_TARGET(name, type) \ { \ &xen_ ## name, \ PARAVIRT_PATCH_TYPE_BR_ ## type, \ } PARAVIRT_BR_TARGET(switch_to, SWITCH_TO), PARAVIRT_BR_TARGET(leave_syscall, LEAVE_SYSCALL), PARAVIRT_BR_TARGET(work_processed_syscall, WORK_PROCESSED_SYSCALL), PARAVIRT_BR_TARGET(leave_kernel, LEAVE_KERNEL), }; static void __init xen_patch_branch(unsigned long tag, unsigned long type) { __paravirt_patch_apply_branch(tag, type, xen_branch_target, ARRAY_SIZE(xen_branch_target)); }