05f0ecbda5
This patch fixes the following compilation error caused by recursive inclusion of kernel.h which defines BUILD_BUG_ON(). In this case, the case it catches will be caught by the case CONFIG_PARAVIRT=n, so removing it would not hurt compile time check very much. So fix the breakage by removing it. CC arch/ia64/kernel/asm-offsets.s In file included from include/linux/bitops.h:17, from include/linux/kernel.h:15, from include/linux/sched.h:52, from arch/ia64/kernel/asm-offsets.c:9: arch/ia64/include/asm/bitops.h: In function 'set_bit': arch/ia64/include/asm/bitops.h:47: error: implicit declaration of function 'BUILD_BUG_ON' Signed-off-by: Isaku Yamahata <yamahata@valinux.co.jp> Signed-off-by: Tony Luck <tony.luck@intel.com>
479 lines
14 KiB
C
479 lines
14 KiB
C
/******************************************************************************
|
|
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
|
|
* VA Linux Systems Japan K.K.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*
|
|
*/
|
|
|
|
#ifndef _ASM_IA64_PARAVIRT_PRIVOP_H
|
|
#define _ASM_IA64_PARAVIRT_PRIVOP_H
|
|
|
|
#ifdef CONFIG_PARAVIRT
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <linux/types.h>
|
|
#include <asm/kregs.h> /* for IA64_PSR_I */
|
|
|
|
/******************************************************************************
|
|
* replacement of intrinsics operations.
|
|
*/
|
|
|
|
struct pv_cpu_ops {
|
|
void (*fc)(void *addr);
|
|
unsigned long (*thash)(unsigned long addr);
|
|
unsigned long (*get_cpuid)(int index);
|
|
unsigned long (*get_pmd)(int index);
|
|
unsigned long (*getreg)(int reg);
|
|
void (*setreg)(int reg, unsigned long val);
|
|
void (*ptcga)(unsigned long addr, unsigned long size);
|
|
unsigned long (*get_rr)(unsigned long index);
|
|
void (*set_rr)(unsigned long index, unsigned long val);
|
|
void (*set_rr0_to_rr4)(unsigned long val0, unsigned long val1,
|
|
unsigned long val2, unsigned long val3,
|
|
unsigned long val4);
|
|
void (*ssm_i)(void);
|
|
void (*rsm_i)(void);
|
|
unsigned long (*get_psr_i)(void);
|
|
void (*intrin_local_irq_restore)(unsigned long flags);
|
|
};
|
|
|
|
extern struct pv_cpu_ops pv_cpu_ops;
|
|
|
|
extern void ia64_native_setreg_func(int regnum, unsigned long val);
|
|
extern unsigned long ia64_native_getreg_func(int regnum);
|
|
|
|
/************************************************/
|
|
/* Instructions paravirtualized for performance */
|
|
/************************************************/
|
|
|
|
#ifndef ASM_SUPPORTED
|
|
#define paravirt_ssm_i() pv_cpu_ops.ssm_i()
|
|
#define paravirt_rsm_i() pv_cpu_ops.rsm_i()
|
|
#define __paravirt_getreg() pv_cpu_ops.getreg()
|
|
#endif
|
|
|
|
/* mask for ia64_native_ssm/rsm() must be constant.("i" constraing).
|
|
* static inline function doesn't satisfy it. */
|
|
#define paravirt_ssm(mask) \
|
|
do { \
|
|
if ((mask) == IA64_PSR_I) \
|
|
paravirt_ssm_i(); \
|
|
else \
|
|
ia64_native_ssm(mask); \
|
|
} while (0)
|
|
|
|
#define paravirt_rsm(mask) \
|
|
do { \
|
|
if ((mask) == IA64_PSR_I) \
|
|
paravirt_rsm_i(); \
|
|
else \
|
|
ia64_native_rsm(mask); \
|
|
} while (0)
|
|
|
|
/* returned ip value should be the one in the caller,
|
|
* not in __paravirt_getreg() */
|
|
#define paravirt_getreg(reg) \
|
|
({ \
|
|
unsigned long res; \
|
|
if ((reg) == _IA64_REG_IP) \
|
|
res = ia64_native_getreg(_IA64_REG_IP); \
|
|
else \
|
|
res = __paravirt_getreg(reg); \
|
|
res; \
|
|
})
|
|
|
|
/******************************************************************************
|
|
* replacement of hand written assembly codes.
|
|
*/
|
|
struct pv_cpu_asm_switch {
|
|
unsigned long switch_to;
|
|
unsigned long leave_syscall;
|
|
unsigned long work_processed_syscall;
|
|
unsigned long leave_kernel;
|
|
};
|
|
void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#define IA64_PARAVIRT_ASM_FUNC(name) paravirt_ ## name
|
|
|
|
#else
|
|
|
|
/* fallback for native case */
|
|
#define IA64_PARAVIRT_ASM_FUNC(name) ia64_native_ ## name
|
|
|
|
#endif /* CONFIG_PARAVIRT */
|
|
|
|
#if defined(CONFIG_PARAVIRT) && defined(ASM_SUPPORTED)
|
|
#define paravirt_dv_serialize_data() ia64_dv_serialize_data()
|
|
#else
|
|
#define paravirt_dv_serialize_data() /* nothing */
|
|
#endif
|
|
|
|
/* these routines utilize privilege-sensitive or performance-sensitive
|
|
* privileged instructions so the code must be replaced with
|
|
* paravirtualized versions */
|
|
#define ia64_switch_to IA64_PARAVIRT_ASM_FUNC(switch_to)
|
|
#define ia64_leave_syscall IA64_PARAVIRT_ASM_FUNC(leave_syscall)
|
|
#define ia64_work_processed_syscall \
|
|
IA64_PARAVIRT_ASM_FUNC(work_processed_syscall)
|
|
#define ia64_leave_kernel IA64_PARAVIRT_ASM_FUNC(leave_kernel)
|
|
|
|
|
|
#if defined(CONFIG_PARAVIRT)
|
|
/******************************************************************************
|
|
* binary patching infrastructure
|
|
*/
|
|
#define PARAVIRT_PATCH_TYPE_FC 1
|
|
#define PARAVIRT_PATCH_TYPE_THASH 2
|
|
#define PARAVIRT_PATCH_TYPE_GET_CPUID 3
|
|
#define PARAVIRT_PATCH_TYPE_GET_PMD 4
|
|
#define PARAVIRT_PATCH_TYPE_PTCGA 5
|
|
#define PARAVIRT_PATCH_TYPE_GET_RR 6
|
|
#define PARAVIRT_PATCH_TYPE_SET_RR 7
|
|
#define PARAVIRT_PATCH_TYPE_SET_RR0_TO_RR4 8
|
|
#define PARAVIRT_PATCH_TYPE_SSM_I 9
|
|
#define PARAVIRT_PATCH_TYPE_RSM_I 10
|
|
#define PARAVIRT_PATCH_TYPE_GET_PSR_I 11
|
|
#define PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE 12
|
|
|
|
/* PARAVIRT_PATY_TYPE_[GS]ETREG + _IA64_REG_xxx */
|
|
#define PARAVIRT_PATCH_TYPE_GETREG 0x10000000
|
|
#define PARAVIRT_PATCH_TYPE_SETREG 0x20000000
|
|
|
|
/*
|
|
* struct task_struct* (*ia64_switch_to)(void* next_task);
|
|
* void *ia64_leave_syscall;
|
|
* void *ia64_work_processed_syscall
|
|
* void *ia64_leave_kernel;
|
|
*/
|
|
|
|
#define PARAVIRT_PATCH_TYPE_BR_START 0x30000000
|
|
#define PARAVIRT_PATCH_TYPE_BR_SWITCH_TO \
|
|
(PARAVIRT_PATCH_TYPE_BR_START + 0)
|
|
#define PARAVIRT_PATCH_TYPE_BR_LEAVE_SYSCALL \
|
|
(PARAVIRT_PATCH_TYPE_BR_START + 1)
|
|
#define PARAVIRT_PATCH_TYPE_BR_WORK_PROCESSED_SYSCALL \
|
|
(PARAVIRT_PATCH_TYPE_BR_START + 2)
|
|
#define PARAVIRT_PATCH_TYPE_BR_LEAVE_KERNEL \
|
|
(PARAVIRT_PATCH_TYPE_BR_START + 3)
|
|
|
|
#ifdef ASM_SUPPORTED
|
|
#include <asm/paravirt_patch.h>
|
|
|
|
/*
|
|
* pv_cpu_ops calling stub.
|
|
* normal function call convension can't be written by gcc
|
|
* inline assembly.
|
|
*
|
|
* from the caller's point of view,
|
|
* the following registers will be clobbered.
|
|
* r2, r3
|
|
* r8-r15
|
|
* r16, r17
|
|
* b6, b7
|
|
* p6-p15
|
|
* ar.ccv
|
|
*
|
|
* from the callee's point of view ,
|
|
* the following registers can be used.
|
|
* r2, r3: scratch
|
|
* r8: scratch, input argument0 and return value
|
|
* r0-r15: scratch, input argument1-5
|
|
* b6: return pointer
|
|
* b7: scratch
|
|
* p6-p15: scratch
|
|
* ar.ccv: scratch
|
|
*
|
|
* other registers must not be changed. especially
|
|
* b0: rp: preserved. gcc ignores b0 in clobbered register.
|
|
* r16: saved gp
|
|
*/
|
|
/* 5 bundles */
|
|
#define __PARAVIRT_BR \
|
|
";;\n" \
|
|
"{ .mlx\n" \
|
|
"nop 0\n" \
|
|
"movl r2 = %[op_addr]\n"/* get function pointer address */ \
|
|
";;\n" \
|
|
"}\n" \
|
|
"1:\n" \
|
|
"{ .mii\n" \
|
|
"ld8 r2 = [r2]\n" /* load function descriptor address */ \
|
|
"mov r17 = ip\n" /* get ip to calc return address */ \
|
|
"mov r16 = gp\n" /* save gp */ \
|
|
";;\n" \
|
|
"}\n" \
|
|
"{ .mii\n" \
|
|
"ld8 r3 = [r2], 8\n" /* load entry address */ \
|
|
"adds r17 = 1f - 1b, r17\n" /* calculate return address */ \
|
|
";;\n" \
|
|
"mov b7 = r3\n" /* set entry address */ \
|
|
"}\n" \
|
|
"{ .mib\n" \
|
|
"ld8 gp = [r2]\n" /* load gp value */ \
|
|
"mov b6 = r17\n" /* set return address */ \
|
|
"br.cond.sptk.few b7\n" /* intrinsics are very short isns */ \
|
|
"}\n" \
|
|
"1:\n" \
|
|
"{ .mii\n" \
|
|
"mov gp = r16\n" /* restore gp value */ \
|
|
"nop 0\n" \
|
|
"nop 0\n" \
|
|
";;\n" \
|
|
"}\n"
|
|
|
|
#define PARAVIRT_OP(op) \
|
|
[op_addr] "i"(&pv_cpu_ops.op)
|
|
|
|
#define PARAVIRT_TYPE(type) \
|
|
PARAVIRT_PATCH_TYPE_ ## type
|
|
|
|
#define PARAVIRT_REG_CLOBBERS0 \
|
|
"r2", "r3", /*"r8",*/ "r9", "r10", "r11", "r14", \
|
|
"r15", "r16", "r17"
|
|
|
|
#define PARAVIRT_REG_CLOBBERS1 \
|
|
"r2","r3", /*"r8",*/ "r9", "r10", "r11", "r14", \
|
|
"r15", "r16", "r17"
|
|
|
|
#define PARAVIRT_REG_CLOBBERS2 \
|
|
"r2", "r3", /*"r8", "r9",*/ "r10", "r11", "r14", \
|
|
"r15", "r16", "r17"
|
|
|
|
#define PARAVIRT_REG_CLOBBERS5 \
|
|
"r2", "r3", /*"r8", "r9", "r10", "r11", "r14",*/ \
|
|
"r15", "r16", "r17"
|
|
|
|
#define PARAVIRT_BR_CLOBBERS \
|
|
"b6", "b7"
|
|
|
|
#define PARAVIRT_PR_CLOBBERS \
|
|
"p6", "p7", "p8", "p9", "p10", "p11", "p12", "p13", "p14", "p15"
|
|
|
|
#define PARAVIRT_AR_CLOBBERS \
|
|
"ar.ccv"
|
|
|
|
#define PARAVIRT_CLOBBERS0 \
|
|
PARAVIRT_REG_CLOBBERS0, \
|
|
PARAVIRT_BR_CLOBBERS, \
|
|
PARAVIRT_PR_CLOBBERS, \
|
|
PARAVIRT_AR_CLOBBERS, \
|
|
"memory"
|
|
|
|
#define PARAVIRT_CLOBBERS1 \
|
|
PARAVIRT_REG_CLOBBERS1, \
|
|
PARAVIRT_BR_CLOBBERS, \
|
|
PARAVIRT_PR_CLOBBERS, \
|
|
PARAVIRT_AR_CLOBBERS, \
|
|
"memory"
|
|
|
|
#define PARAVIRT_CLOBBERS2 \
|
|
PARAVIRT_REG_CLOBBERS2, \
|
|
PARAVIRT_BR_CLOBBERS, \
|
|
PARAVIRT_PR_CLOBBERS, \
|
|
PARAVIRT_AR_CLOBBERS, \
|
|
"memory"
|
|
|
|
#define PARAVIRT_CLOBBERS5 \
|
|
PARAVIRT_REG_CLOBBERS5, \
|
|
PARAVIRT_BR_CLOBBERS, \
|
|
PARAVIRT_PR_CLOBBERS, \
|
|
PARAVIRT_AR_CLOBBERS, \
|
|
"memory"
|
|
|
|
#define PARAVIRT_BR0(op, type) \
|
|
register unsigned long ia64_clobber asm ("r8"); \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_clobber) \
|
|
: PARAVIRT_OP(op) \
|
|
: PARAVIRT_CLOBBERS0)
|
|
|
|
#define PARAVIRT_BR0_RET(op, type) \
|
|
register unsigned long ia64_intri_res asm ("r8"); \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_intri_res) \
|
|
: PARAVIRT_OP(op) \
|
|
: PARAVIRT_CLOBBERS0)
|
|
|
|
#define PARAVIRT_BR1(op, type, arg1) \
|
|
register unsigned long __##arg1 asm ("r8") = arg1; \
|
|
register unsigned long ia64_clobber asm ("r8"); \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_clobber) \
|
|
: PARAVIRT_OP(op), "0"(__##arg1) \
|
|
: PARAVIRT_CLOBBERS1)
|
|
|
|
#define PARAVIRT_BR1_RET(op, type, arg1) \
|
|
register unsigned long ia64_intri_res asm ("r8"); \
|
|
register unsigned long __##arg1 asm ("r8") = arg1; \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_intri_res) \
|
|
: PARAVIRT_OP(op), "0"(__##arg1) \
|
|
: PARAVIRT_CLOBBERS1)
|
|
|
|
#define PARAVIRT_BR1_VOID(op, type, arg1) \
|
|
register void *__##arg1 asm ("r8") = arg1; \
|
|
register unsigned long ia64_clobber asm ("r8"); \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_clobber) \
|
|
: PARAVIRT_OP(op), "0"(__##arg1) \
|
|
: PARAVIRT_CLOBBERS1)
|
|
|
|
#define PARAVIRT_BR2(op, type, arg1, arg2) \
|
|
register unsigned long __##arg1 asm ("r8") = arg1; \
|
|
register unsigned long __##arg2 asm ("r9") = arg2; \
|
|
register unsigned long ia64_clobber1 asm ("r8"); \
|
|
register unsigned long ia64_clobber2 asm ("r9"); \
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(type)) \
|
|
: "=r"(ia64_clobber1), "=r"(ia64_clobber2) \
|
|
: PARAVIRT_OP(op), "0"(__##arg1), "1"(__##arg2) \
|
|
: PARAVIRT_CLOBBERS2)
|
|
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP0(op, type) \
|
|
static inline void \
|
|
paravirt_ ## op (void) \
|
|
{ \
|
|
PARAVIRT_BR0(op, type); \
|
|
}
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP0_RET(op, type) \
|
|
static inline unsigned long \
|
|
paravirt_ ## op (void) \
|
|
{ \
|
|
PARAVIRT_BR0_RET(op, type); \
|
|
return ia64_intri_res; \
|
|
}
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP1_VOID(op, type) \
|
|
static inline void \
|
|
paravirt_ ## op (void *arg1) \
|
|
{ \
|
|
PARAVIRT_BR1_VOID(op, type, arg1); \
|
|
}
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP1(op, type) \
|
|
static inline void \
|
|
paravirt_ ## op (unsigned long arg1) \
|
|
{ \
|
|
PARAVIRT_BR1(op, type, arg1); \
|
|
}
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP1_RET(op, type) \
|
|
static inline unsigned long \
|
|
paravirt_ ## op (unsigned long arg1) \
|
|
{ \
|
|
PARAVIRT_BR1_RET(op, type, arg1); \
|
|
return ia64_intri_res; \
|
|
}
|
|
|
|
#define PARAVIRT_DEFINE_CPU_OP2(op, type) \
|
|
static inline void \
|
|
paravirt_ ## op (unsigned long arg1, \
|
|
unsigned long arg2) \
|
|
{ \
|
|
PARAVIRT_BR2(op, type, arg1, arg2); \
|
|
}
|
|
|
|
|
|
PARAVIRT_DEFINE_CPU_OP1_VOID(fc, FC);
|
|
PARAVIRT_DEFINE_CPU_OP1_RET(thash, THASH)
|
|
PARAVIRT_DEFINE_CPU_OP1_RET(get_cpuid, GET_CPUID)
|
|
PARAVIRT_DEFINE_CPU_OP1_RET(get_pmd, GET_PMD)
|
|
PARAVIRT_DEFINE_CPU_OP2(ptcga, PTCGA)
|
|
PARAVIRT_DEFINE_CPU_OP1_RET(get_rr, GET_RR)
|
|
PARAVIRT_DEFINE_CPU_OP2(set_rr, SET_RR)
|
|
PARAVIRT_DEFINE_CPU_OP0(ssm_i, SSM_I)
|
|
PARAVIRT_DEFINE_CPU_OP0(rsm_i, RSM_I)
|
|
PARAVIRT_DEFINE_CPU_OP0_RET(get_psr_i, GET_PSR_I)
|
|
PARAVIRT_DEFINE_CPU_OP1(intrin_local_irq_restore, INTRIN_LOCAL_IRQ_RESTORE)
|
|
|
|
static inline void
|
|
paravirt_set_rr0_to_rr4(unsigned long val0, unsigned long val1,
|
|
unsigned long val2, unsigned long val3,
|
|
unsigned long val4)
|
|
{
|
|
register unsigned long __val0 asm ("r8") = val0;
|
|
register unsigned long __val1 asm ("r9") = val1;
|
|
register unsigned long __val2 asm ("r10") = val2;
|
|
register unsigned long __val3 asm ("r11") = val3;
|
|
register unsigned long __val4 asm ("r14") = val4;
|
|
|
|
register unsigned long ia64_clobber0 asm ("r8");
|
|
register unsigned long ia64_clobber1 asm ("r9");
|
|
register unsigned long ia64_clobber2 asm ("r10");
|
|
register unsigned long ia64_clobber3 asm ("r11");
|
|
register unsigned long ia64_clobber4 asm ("r14");
|
|
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,
|
|
PARAVIRT_TYPE(SET_RR0_TO_RR4))
|
|
: "=r"(ia64_clobber0),
|
|
"=r"(ia64_clobber1),
|
|
"=r"(ia64_clobber2),
|
|
"=r"(ia64_clobber3),
|
|
"=r"(ia64_clobber4)
|
|
: PARAVIRT_OP(set_rr0_to_rr4),
|
|
"0"(__val0), "1"(__val1), "2"(__val2),
|
|
"3"(__val3), "4"(__val4)
|
|
: PARAVIRT_CLOBBERS5);
|
|
}
|
|
|
|
/* unsigned long paravirt_getreg(int reg) */
|
|
#define __paravirt_getreg(reg) \
|
|
({ \
|
|
register unsigned long ia64_intri_res asm ("r8"); \
|
|
register unsigned long __reg asm ("r8") = (reg); \
|
|
\
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(GETREG) \
|
|
+ (reg)) \
|
|
: "=r"(ia64_intri_res) \
|
|
: PARAVIRT_OP(getreg), "0"(__reg) \
|
|
: PARAVIRT_CLOBBERS1); \
|
|
\
|
|
ia64_intri_res; \
|
|
})
|
|
|
|
/* void paravirt_setreg(int reg, unsigned long val) */
|
|
#define paravirt_setreg(reg, val) \
|
|
do { \
|
|
register unsigned long __val asm ("r8") = val; \
|
|
register unsigned long __reg asm ("r9") = reg; \
|
|
register unsigned long ia64_clobber1 asm ("r8"); \
|
|
register unsigned long ia64_clobber2 asm ("r9"); \
|
|
\
|
|
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
|
|
PARAVIRT_TYPE(SETREG) \
|
|
+ (reg)) \
|
|
: "=r"(ia64_clobber1), \
|
|
"=r"(ia64_clobber2) \
|
|
: PARAVIRT_OP(setreg), \
|
|
"1"(__reg), "0"(__val) \
|
|
: PARAVIRT_CLOBBERS2); \
|
|
} while (0)
|
|
|
|
#endif /* ASM_SUPPORTED */
|
|
#endif /* CONFIG_PARAVIRT && ASM_SUPPOTED */
|
|
|
|
#endif /* _ASM_IA64_PARAVIRT_PRIVOP_H */
|