x86: unify load_segment macro
This patch unifies the load_segment() macro, making them equal in both x86_64 and i386 architectures. The common version goes to system.h, and the old are deleted. Signed-off-by: Glauber de Oliveira Costa <gcosta@redhat.com> Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
This commit is contained in:
committed by
Ingo Molnar
parent
d89542229b
commit
a6b4655258
@@ -38,6 +38,27 @@ __asm__ __volatile__ ("movw %%dx,%1\n\t" \
|
||||
#define set_base(ldt, base) _set_base(((char *)&(ldt)) , (base))
|
||||
#define set_limit(ldt, limit) _set_limit(((char *)&(ldt)) , ((limit)-1))
|
||||
|
||||
/*
|
||||
* Load a segment. Fall back on loading the zero
|
||||
* segment if something goes wrong..
|
||||
*/
|
||||
#define loadsegment(seg, value) \
|
||||
asm volatile("\n" \
|
||||
"1:\t" \
|
||||
"movl %k0,%%" #seg "\n" \
|
||||
"2:\n" \
|
||||
".section .fixup,\"ax\"\n" \
|
||||
"3:\t" \
|
||||
"movl %k1, %%" #seg "\n\t" \
|
||||
"jmp 2b\n" \
|
||||
".previous\n" \
|
||||
".section __ex_table,\"a\"\n\t" \
|
||||
_ASM_ALIGN "\n\t" \
|
||||
_ASM_PTR " 1b,3b\n" \
|
||||
".previous" \
|
||||
: :"r" (value), "r" (0))
|
||||
|
||||
|
||||
/*
|
||||
* Save a segment register away
|
||||
*/
|
||||
|
Reference in New Issue
Block a user