Commit 8d51b032 authored by Anton Blanchard's avatar Anton Blanchard Committed by Linus Torvalds

[PATCH] ppc64: cleanup asm/processor.h

Wrap a lot more of processor.h in #ifdef __KERNEL__ / __ASSEMBLY__.  Remove
now unised EXC_FRAME_SIZE.
Signed-off-by: default avatarAnton Blanchard <anton@samba.org>
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent d19b41b9
...@@ -20,12 +20,6 @@ ...@@ -20,12 +20,6 @@
#include <asm/ptrace.h> #include <asm/ptrace.h>
#include <asm/types.h> #include <asm/types.h>
/*
* Default implementation of macro that returns current
* instruction pointer ("program counter").
*/
#define current_text_addr() ({ __label__ _l; _l: &&_l;})
/* Machine State Register (MSR) Fields */ /* Machine State Register (MSR) Fields */
#define MSR_SF_LG 63 /* Enable 64 bit mode */ #define MSR_SF_LG 63 /* Enable 64 bit mode */
#define MSR_ISF_LG 61 /* Interrupt 64b mode valid on 630 */ #define MSR_ISF_LG 61 /* Interrupt 64b mode valid on 630 */
...@@ -410,6 +404,12 @@ ...@@ -410,6 +404,12 @@
#define XGLUE(a,b) a##b #define XGLUE(a,b) a##b
#define GLUE(a,b) XGLUE(a,b) #define GLUE(a,b) XGLUE(a,b)
/* iSeries CTRL register (for runlatch) */
#define CTRLT 0x098
#define CTRLF 0x088
#define RUNLATCH 0x0001
#ifdef __ASSEMBLY__ #ifdef __ASSEMBLY__
#define _GLOBAL(name) \ #define _GLOBAL(name) \
...@@ -438,8 +438,13 @@ name: \ ...@@ -438,8 +438,13 @@ name: \
.type GLUE(.,name),@function; \ .type GLUE(.,name),@function; \
GLUE(.,name): GLUE(.,name):
#endif /* __ASSEMBLY__ */ #else /* __ASSEMBLY__ */
/*
* Default implementation of macro that returns current
* instruction pointer ("program counter").
*/
#define current_text_addr() ({ __label__ _l; _l: &&_l;})
/* Macros for setting and retrieving special purpose registers */ /* Macros for setting and retrieving special purpose registers */
...@@ -461,20 +466,9 @@ GLUE(.,name): ...@@ -461,20 +466,9 @@ GLUE(.,name):
#define mttbl(v) asm volatile("mttbl %0":: "r"(v)) #define mttbl(v) asm volatile("mttbl %0":: "r"(v))
#define mttbu(v) asm volatile("mttbu %0":: "r"(v)) #define mttbu(v) asm volatile("mttbu %0":: "r"(v))
/* iSeries CTRL register (for runlatch) */
#define CTRLT 0x098
#define CTRLF 0x088
#define RUNLATCH 0x0001
/* Size of an exception stack frame contained in the paca. */
#define EXC_FRAME_SIZE 64
#define mfasr() ({unsigned long rval; \ #define mfasr() ({unsigned long rval; \
asm volatile("mfasr %0" : "=r" (rval)); rval;}) asm volatile("mfasr %0" : "=r" (rval)); rval;})
#ifndef __ASSEMBLY__
static inline void set_tb(unsigned int upper, unsigned int lower) static inline void set_tb(unsigned int upper, unsigned int lower)
{ {
mttbl(0); mttbl(0);
...@@ -485,6 +479,8 @@ static inline void set_tb(unsigned int upper, unsigned int lower) ...@@ -485,6 +479,8 @@ static inline void set_tb(unsigned int upper, unsigned int lower)
#define __get_SP() ({unsigned long sp; \ #define __get_SP() ({unsigned long sp; \
asm volatile("mr %0,1": "=r" (sp)); sp;}) asm volatile("mr %0,1": "=r" (sp)); sp;})
#ifdef __KERNEL__
extern int have_of; extern int have_of;
struct task_struct; struct task_struct;
...@@ -507,8 +503,6 @@ extern long kernel_thread(int (*fn)(void *), void *arg, unsigned long flags); ...@@ -507,8 +503,6 @@ extern long kernel_thread(int (*fn)(void *), void *arg, unsigned long flags);
extern struct task_struct *last_task_used_math; extern struct task_struct *last_task_used_math;
extern struct task_struct *last_task_used_altivec; extern struct task_struct *last_task_used_altivec;
#ifdef __KERNEL__
/* 64-bit user address space is 41-bits (2TBs user VM) */ /* 64-bit user address space is 41-bits (2TBs user VM) */
#define TASK_SIZE_USER64 (0x0000020000000000UL) #define TASK_SIZE_USER64 (0x0000020000000000UL)
...@@ -520,8 +514,6 @@ extern struct task_struct *last_task_used_altivec; ...@@ -520,8 +514,6 @@ extern struct task_struct *last_task_used_altivec;
#define TASK_SIZE (test_thread_flag(TIF_32BIT) ? \ #define TASK_SIZE (test_thread_flag(TIF_32BIT) ? \
TASK_SIZE_USER32 : TASK_SIZE_USER64) TASK_SIZE_USER32 : TASK_SIZE_USER64)
#endif /* __KERNEL__ */
/* This decides where the kernel will search for a free chunk of vm /* This decides where the kernel will search for a free chunk of vm
* space during mmap's. * space during mmap's.
...@@ -627,10 +619,12 @@ static inline void prefetchw(const void *x) ...@@ -627,10 +619,12 @@ static inline void prefetchw(const void *x)
#define spin_lock_prefetch(x) prefetchw(x) #define spin_lock_prefetch(x) prefetchw(x)
#endif /* ASSEMBLY */
#define HAVE_ARCH_PICK_MMAP_LAYOUT #define HAVE_ARCH_PICK_MMAP_LAYOUT
#endif /* __KERNEL__ */
#endif /* __ASSEMBLY__ */
/* /*
* Number of entries in the SLB. If this ever changes we should handle * Number of entries in the SLB. If this ever changes we should handle
* it with a use a cpu feature fixup. * it with a use a cpu feature fixup.
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment