Commit c9f61748 authored by Paul Mackerras's avatar Paul Mackerras

[PATCH] PPC64 Segment table code cleanup - kill bitfields

Remove the overly verbose and hard to follow use of bitfields in the
PPC64 segment table code, replacing it with explicit bitmask
operations.
Signed-off-by: default avatarDavid Gibson <dwg@au1.ibm.com>
Signed-off-by: default avatarPaul Mackerras <paulus@samba.org>
parent 87218e11
...@@ -61,33 +61,32 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid) ...@@ -61,33 +61,32 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid)
{ {
unsigned long entry, group, old_esid, castout_entry, i; unsigned long entry, group, old_esid, castout_entry, i;
unsigned int global_entry; unsigned int global_entry;
STE *ste, *castout_ste; struct stab_entry *ste, *castout_ste;
unsigned long kernel_segment = (REGION_ID(esid << SID_SHIFT) != unsigned long kernel_segment = (REGION_ID(esid << SID_SHIFT) !=
USER_REGION_ID); USER_REGION_ID);
unsigned long esid_data;
/* Search the primary group first. */ /* Search the primary group first. */
global_entry = (esid & 0x1f) << 3; global_entry = (esid & 0x1f) << 3;
ste = (STE *)(stab | ((esid & 0x1f) << 7)); ste = (struct stab_entry *)(stab | ((esid & 0x1f) << 7));
/* Find an empty entry, if one exists. */ /* Find an empty entry, if one exists. */
for (group = 0; group < 2; group++) { for (group = 0; group < 2; group++) {
for (entry = 0; entry < 8; entry++, ste++) { for (entry = 0; entry < 8; entry++, ste++) {
if (!(ste->dw0.dw0.v)) { if (!(ste->esid_data & STE_ESID_V)) {
ste->dw0.dword0 = 0; ste->vsid_data = vsid << STE_VSID_SHIFT;
ste->dw1.dword1 = 0;
ste->dw1.dw1.vsid = vsid;
ste->dw0.dw0.esid = esid;
ste->dw0.dw0.kp = 1;
if (!kernel_segment)
ste->dw0.dw0.ks = 1;
asm volatile("eieio":::"memory"); asm volatile("eieio":::"memory");
ste->dw0.dw0.v = 1; esid_data = esid << SID_SHIFT;
esid_data |= STE_ESID_KP | STE_ESID_V;
if (! kernel_segment)
esid_data |= STE_ESID_KS;
ste->esid_data = esid_data;
return (global_entry | entry); return (global_entry | entry);
} }
} }
/* Now search the secondary group. */ /* Now search the secondary group. */
global_entry = ((~esid) & 0x1f) << 3; global_entry = ((~esid) & 0x1f) << 3;
ste = (STE *)(stab | (((~esid) & 0x1f) << 7)); ste = (struct stab_entry *)(stab | (((~esid) & 0x1f) << 7));
} }
/* /*
...@@ -98,16 +97,16 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid) ...@@ -98,16 +97,16 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid)
for (i = 0; i < 16; i++) { for (i = 0; i < 16; i++) {
if (castout_entry < 8) { if (castout_entry < 8) {
global_entry = (esid & 0x1f) << 3; global_entry = (esid & 0x1f) << 3;
ste = (STE *)(stab | ((esid & 0x1f) << 7)); ste = (struct stab_entry *)(stab | ((esid & 0x1f) << 7));
castout_ste = ste + castout_entry; castout_ste = ste + castout_entry;
} else { } else {
global_entry = ((~esid) & 0x1f) << 3; global_entry = ((~esid) & 0x1f) << 3;
ste = (STE *)(stab | (((~esid) & 0x1f) << 7)); ste = (struct stab_entry *)(stab | (((~esid) & 0x1f) << 7));
castout_ste = ste + (castout_entry - 8); castout_ste = ste + (castout_entry - 8);
} }
/* Dont cast out the first kernel segment */ /* Dont cast out the first kernel segment */
if (castout_ste->dw0.dw0.esid != GET_ESID(KERNELBASE)) if ((castout_ste->esid_data & ESID_MASK) != KERNELBASE)
break; break;
castout_entry = (castout_entry + 1) & 0xf; castout_entry = (castout_entry + 1) & 0xf;
...@@ -120,19 +119,21 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid) ...@@ -120,19 +119,21 @@ static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid)
/* Force previous translations to complete. DRENG */ /* Force previous translations to complete. DRENG */
asm volatile("isync" : : : "memory"); asm volatile("isync" : : : "memory");
castout_ste->dw0.dw0.v = 0; old_esid = castout_ste->esid_data >> SID_SHIFT;
castout_ste->esid_data = 0; /* Invalidate old entry */
asm volatile("sync" : : : "memory"); /* Order update */ asm volatile("sync" : : : "memory"); /* Order update */
castout_ste->dw0.dword0 = 0; castout_ste->vsid_data = vsid << STE_VSID_SHIFT;
castout_ste->dw1.dword1 = 0;
castout_ste->dw1.dw1.vsid = vsid;
old_esid = castout_ste->dw0.dw0.esid;
castout_ste->dw0.dw0.esid = esid;
castout_ste->dw0.dw0.kp = 1;
if (!kernel_segment)
castout_ste->dw0.dw0.ks = 1;
asm volatile("eieio" : : : "memory"); /* Order update */ asm volatile("eieio" : : : "memory"); /* Order update */
castout_ste->dw0.dw0.v = 1;
esid_data = esid << SID_SHIFT;
esid_data |= STE_ESID_KP | STE_ESID_V;
if (!kernel_segment)
esid_data |= STE_ESID_KS;
castout_ste->esid_data = esid_data;
asm volatile("slbie %0" : : "r" (old_esid << SID_SHIFT)); asm volatile("slbie %0" : : "r" (old_esid << SID_SHIFT));
/* Ensure completion of slbie */ /* Ensure completion of slbie */
asm volatile("sync" : : : "memory"); asm volatile("sync" : : : "memory");
...@@ -240,8 +241,8 @@ static void preload_stab(struct task_struct *tsk, struct mm_struct *mm) ...@@ -240,8 +241,8 @@ static void preload_stab(struct task_struct *tsk, struct mm_struct *mm)
/* Flush all user entries from the segment table of the current processor. */ /* Flush all user entries from the segment table of the current processor. */
void flush_stab(struct task_struct *tsk, struct mm_struct *mm) void flush_stab(struct task_struct *tsk, struct mm_struct *mm)
{ {
STE *stab = (STE *) get_paca()->stab_addr; struct stab_entry *stab = (struct stab_entry *) get_paca()->stab_addr;
STE *ste; struct stab_entry *ste;
unsigned long offset = __get_cpu_var(stab_cache_ptr); unsigned long offset = __get_cpu_var(stab_cache_ptr);
/* Force previous translations to complete. DRENG */ /* Force previous translations to complete. DRENG */
...@@ -252,7 +253,7 @@ void flush_stab(struct task_struct *tsk, struct mm_struct *mm) ...@@ -252,7 +253,7 @@ void flush_stab(struct task_struct *tsk, struct mm_struct *mm)
for (i = 0; i < offset; i++) { for (i = 0; i < offset; i++) {
ste = stab + __get_cpu_var(stab_cache[i]); ste = stab + __get_cpu_var(stab_cache[i]);
ste->dw0.dw0.v = 0; ste->esid_data = 0; /* invalidate entry */
} }
} else { } else {
unsigned long entry; unsigned long entry;
...@@ -263,12 +264,12 @@ void flush_stab(struct task_struct *tsk, struct mm_struct *mm) ...@@ -263,12 +264,12 @@ void flush_stab(struct task_struct *tsk, struct mm_struct *mm)
/* Never flush the first entry. */ /* Never flush the first entry. */
ste += 1; ste += 1;
for (entry = 1; for (entry = 1;
entry < (PAGE_SIZE / sizeof(STE)); entry < (PAGE_SIZE / sizeof(struct stab_entry));
entry++, ste++) { entry++, ste++) {
unsigned long ea; unsigned long ea;
ea = ste->dw0.dw0.esid << SID_SHIFT; ea = ste->esid_data & ESID_MASK;
if (ea < KERNELBASE) { if (ea < KERNELBASE) {
ste->dw0.dw0.v = 0; ste->esid_data = 0;
} }
} }
} }
......
...@@ -37,33 +37,17 @@ typedef struct { ...@@ -37,33 +37,17 @@ typedef struct {
mm_context_t ctx = { .id = REGION_ID(ea), KERNEL_LOW_HPAGES}; \ mm_context_t ctx = { .id = REGION_ID(ea), KERNEL_LOW_HPAGES}; \
ctx; }) ctx; })
typedef struct { #define STE_ESID_V 0x80
unsigned long esid: 36; /* Effective segment ID */ #define STE_ESID_KS 0x20
unsigned long resv0:20; /* Reserved */ #define STE_ESID_KP 0x10
unsigned long v: 1; /* Entry valid (v=1) or invalid */ #define STE_ESID_N 0x08
unsigned long resv1: 1; /* Reserved */
unsigned long ks: 1; /* Supervisor (privileged) state storage key */
unsigned long kp: 1; /* Problem state storage key */
unsigned long n: 1; /* No-execute if n=1 */
unsigned long resv2: 3; /* padding to a 64b boundary */
} ste_dword0;
typedef struct { #define STE_VSID_SHIFT 12
unsigned long vsid: 52; /* Virtual segment ID */
unsigned long resv0:12; /* Padding to a 64b boundary */
} ste_dword1;
typedef struct _STE { struct stab_entry {
union { unsigned long esid_data;
unsigned long dword0; unsigned long vsid_data;
ste_dword0 dw0; };
} dw0;
union {
unsigned long dword1;
ste_dword1 dw1;
} dw1;
} STE;
/* Hardware Page Table Entry */ /* Hardware Page Table Entry */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment