Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
0983f050
Commit
0983f050
authored
Jul 26, 2005
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge master.kernel.org:/home/rmk/linux-2.6-arm-smp
parents
db776a14
186efd52
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
130 additions
and
58 deletions
+130
-58
arch/arm/mm/mm-armv.c
arch/arm/mm/mm-armv.c
+3
-0
include/asm-arm/bitops.h
include/asm-arm/bitops.h
+2
-2
include/asm-arm/locks.h
include/asm-arm/locks.h
+24
-12
include/asm-arm/spinlock.h
include/asm-arm/spinlock.h
+38
-15
include/asm-arm/system.h
include/asm-arm/system.h
+63
-29
No files found.
arch/arm/mm/mm-armv.c
View file @
0983f050
...
@@ -425,6 +425,9 @@ static void __init build_mem_type_table(void)
...
@@ -425,6 +425,9 @@ static void __init build_mem_type_table(void)
mem_types
[
MT_ROM
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_ROM
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_MINICLEAN
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_MINICLEAN
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_CACHECLEAN
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_CACHECLEAN
].
prot_sect
|=
PMD_SECT_APX
|
PMD_SECT_AP_WRITE
;
mem_types
[
MT_DEVICE
].
prot_pte
|=
L_PTE_BUFFERABLE
;
mem_types
[
MT_DEVICE
].
prot_sect
|=
PMD_SECT_BUFFERED
;
}
}
cp
=
&
cache_policies
[
cachepolicy
];
cp
=
&
cache_policies
[
cachepolicy
];
...
...
include/asm-arm/bitops.h
View file @
0983f050
...
@@ -21,8 +21,8 @@
...
@@ -21,8 +21,8 @@
#include <asm/system.h>
#include <asm/system.h>
#define smp_mb__before_clear_bit()
do { } while (0
)
#define smp_mb__before_clear_bit()
mb(
)
#define smp_mb__after_clear_bit()
do { } while (0
)
#define smp_mb__after_clear_bit()
mb(
)
/*
/*
* These functions are the basis of our bit ops.
* These functions are the basis of our bit ops.
...
...
include/asm-arm/locks.h
View file @
0983f050
...
@@ -28,7 +28,8 @@
...
@@ -28,7 +28,8 @@
" blmi " #fail \
" blmi " #fail \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
})
})
#define __down_op_ret(ptr,fail) \
#define __down_op_ret(ptr,fail) \
...
@@ -48,12 +49,14 @@
...
@@ -48,12 +49,14 @@
" mov %0, ip" \
" mov %0, ip" \
: "=&r" (ret) \
: "=&r" (ret) \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
ret; \
ret; \
})
})
#define __up_op(ptr,wake) \
#define __up_op(ptr,wake) \
({ \
({ \
smp_mb(); \
__asm__ __volatile__( \
__asm__ __volatile__( \
"@ up_op\n" \
"@ up_op\n" \
"1: ldrex lr, [%0]\n" \
"1: ldrex lr, [%0]\n" \
...
@@ -66,7 +69,7 @@
...
@@ -66,7 +69,7 @@
" blle " #wake \
" blle " #wake \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc"
, "memory");
\
: "ip", "lr", "cc"
);
\
})
})
/*
/*
...
@@ -92,11 +95,13 @@
...
@@ -92,11 +95,13 @@
" blne " #fail \
" blne " #fail \
: \
: \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
})
})
#define __up_op_write(ptr,wake) \
#define __up_op_write(ptr,wake) \
({ \
({ \
smp_mb(); \
__asm__ __volatile__( \
__asm__ __volatile__( \
"@ up_op_read\n" \
"@ up_op_read\n" \
"1: ldrex lr, [%0]\n" \
"1: ldrex lr, [%0]\n" \
...
@@ -108,7 +113,7 @@
...
@@ -108,7 +113,7 @@
" blcs " #wake \
" blcs " #wake \
: \
: \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "ip", "lr", "cc"
, "memory");
\
: "ip", "lr", "cc"
);
\
})
})
#define __down_op_read(ptr,fail) \
#define __down_op_read(ptr,fail) \
...
@@ -116,6 +121,7 @@
...
@@ -116,6 +121,7 @@
#define __up_op_read(ptr,wake) \
#define __up_op_read(ptr,wake) \
({ \
({ \
smp_mb(); \
__asm__ __volatile__( \
__asm__ __volatile__( \
"@ up_op_read\n" \
"@ up_op_read\n" \
"1: ldrex lr, [%0]\n" \
"1: ldrex lr, [%0]\n" \
...
@@ -128,7 +134,7 @@
...
@@ -128,7 +134,7 @@
" bleq " #wake \
" bleq " #wake \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc"
, "memory");
\
: "ip", "lr", "cc"
);
\
})
})
#else
#else
...
@@ -148,7 +154,8 @@
...
@@ -148,7 +154,8 @@
" blmi " #fail \
" blmi " #fail \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
})
})
#define __down_op_ret(ptr,fail) \
#define __down_op_ret(ptr,fail) \
...
@@ -169,12 +176,14 @@
...
@@ -169,12 +176,14 @@
" mov %0, ip" \
" mov %0, ip" \
: "=&r" (ret) \
: "=&r" (ret) \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
ret; \
ret; \
})
})
#define __up_op(ptr,wake) \
#define __up_op(ptr,wake) \
({ \
({ \
smp_mb(); \
__asm__ __volatile__( \
__asm__ __volatile__( \
"@ up_op\n" \
"@ up_op\n" \
" mrs ip, cpsr\n" \
" mrs ip, cpsr\n" \
...
@@ -188,7 +197,7 @@
...
@@ -188,7 +197,7 @@
" blle " #wake \
" blle " #wake \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc"
, "memory");
\
: "ip", "lr", "cc"
);
\
})
})
/*
/*
...
@@ -215,7 +224,8 @@
...
@@ -215,7 +224,8 @@
" blne " #fail \
" blne " #fail \
: \
: \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
})
})
#define __up_op_write(ptr,wake) \
#define __up_op_write(ptr,wake) \
...
@@ -233,7 +243,8 @@
...
@@ -233,7 +243,8 @@
" blcs " #wake \
" blcs " #wake \
: \
: \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "r" (ptr), "I" (RW_LOCK_BIAS) \
: "ip", "lr", "cc", "memory"); \
: "ip", "lr", "cc"); \
smp_mb(); \
})
})
#define __down_op_read(ptr,fail) \
#define __down_op_read(ptr,fail) \
...
@@ -241,6 +252,7 @@
...
@@ -241,6 +252,7 @@
#define __up_op_read(ptr,wake) \
#define __up_op_read(ptr,wake) \
({ \
({ \
smp_mb(); \
__asm__ __volatile__( \
__asm__ __volatile__( \
"@ up_op_read\n" \
"@ up_op_read\n" \
" mrs ip, cpsr\n" \
" mrs ip, cpsr\n" \
...
@@ -254,7 +266,7 @@
...
@@ -254,7 +266,7 @@
" bleq " #wake \
" bleq " #wake \
: \
: \
: "r" (ptr), "I" (1) \
: "r" (ptr), "I" (1) \
: "ip", "lr", "cc"
, "memory");
\
: "ip", "lr", "cc"
);
\
})
})
#endif
#endif
...
...
include/asm-arm/spinlock.h
View file @
0983f050
...
@@ -8,9 +8,10 @@
...
@@ -8,9 +8,10 @@
/*
/*
* ARMv6 Spin-locking.
* ARMv6 Spin-locking.
*
*
* We (exclusively) read the old value, and decrement it. If it
* We exclusively read the old value. If it is zero, we may have
* hits zero, we may have won the lock, so we try (exclusively)
* won the lock, so we try exclusively storing it. A memory barrier
* storing it.
* is required after we get a lock, and before we release it, because
* V6 CPUs are assumed to have weakly ordered memory.
*
*
* Unlocked value: 0
* Unlocked value: 0
* Locked value: 1
* Locked value: 1
...
@@ -41,7 +42,9 @@ static inline void _raw_spin_lock(spinlock_t *lock)
...
@@ -41,7 +42,9 @@ static inline void _raw_spin_lock(spinlock_t *lock)
" bne 1b"
" bne 1b"
:
"=&r"
(
tmp
)
:
"=&r"
(
tmp
)
:
"r"
(
&
lock
->
lock
),
"r"
(
1
)
:
"r"
(
&
lock
->
lock
),
"r"
(
1
)
:
"cc"
,
"memory"
);
:
"cc"
);
smp_mb
();
}
}
static
inline
int
_raw_spin_trylock
(
spinlock_t
*
lock
)
static
inline
int
_raw_spin_trylock
(
spinlock_t
*
lock
)
...
@@ -54,18 +57,25 @@ static inline int _raw_spin_trylock(spinlock_t *lock)
...
@@ -54,18 +57,25 @@ static inline int _raw_spin_trylock(spinlock_t *lock)
" strexeq %0, %2, [%1]"
" strexeq %0, %2, [%1]"
:
"=&r"
(
tmp
)
:
"=&r"
(
tmp
)
:
"r"
(
&
lock
->
lock
),
"r"
(
1
)
:
"r"
(
&
lock
->
lock
),
"r"
(
1
)
:
"cc"
,
"memory"
);
:
"cc"
);
return
tmp
==
0
;
if
(
tmp
==
0
)
{
smp_mb
();
return
1
;
}
else
{
return
0
;
}
}
}
static
inline
void
_raw_spin_unlock
(
spinlock_t
*
lock
)
static
inline
void
_raw_spin_unlock
(
spinlock_t
*
lock
)
{
{
smp_mb
();
__asm__
__volatile__
(
__asm__
__volatile__
(
" str %1, [%0]"
" str %1, [%0]"
:
:
:
"r"
(
&
lock
->
lock
),
"r"
(
0
)
:
"r"
(
&
lock
->
lock
),
"r"
(
0
)
:
"cc"
,
"memory"
);
:
"cc"
);
}
}
/*
/*
...
@@ -98,7 +108,9 @@ static inline void _raw_write_lock(rwlock_t *rw)
...
@@ -98,7 +108,9 @@ static inline void _raw_write_lock(rwlock_t *rw)
" bne 1b"
" bne 1b"
:
"=&r"
(
tmp
)
:
"=&r"
(
tmp
)
:
"r"
(
&
rw
->
lock
),
"r"
(
0x80000000
)
:
"r"
(
&
rw
->
lock
),
"r"
(
0x80000000
)
:
"cc"
,
"memory"
);
:
"cc"
);
smp_mb
();
}
}
static
inline
int
_raw_write_trylock
(
rwlock_t
*
rw
)
static
inline
int
_raw_write_trylock
(
rwlock_t
*
rw
)
...
@@ -111,18 +123,25 @@ static inline int _raw_write_trylock(rwlock_t *rw)
...
@@ -111,18 +123,25 @@ static inline int _raw_write_trylock(rwlock_t *rw)
" strexeq %0, %2, [%1]"
" strexeq %0, %2, [%1]"
:
"=&r"
(
tmp
)
:
"=&r"
(
tmp
)
:
"r"
(
&
rw
->
lock
),
"r"
(
0x80000000
)
:
"r"
(
&
rw
->
lock
),
"r"
(
0x80000000
)
:
"cc"
,
"memory"
);
:
"cc"
);
return
tmp
==
0
;
if
(
tmp
==
0
)
{
smp_mb
();
return
1
;
}
else
{
return
0
;
}
}
}
static
inline
void
_raw_write_unlock
(
rwlock_t
*
rw
)
static
inline
void
_raw_write_unlock
(
rwlock_t
*
rw
)
{
{
smp_mb
();
__asm__
__volatile__
(
__asm__
__volatile__
(
"str %1, [%0]"
"str %1, [%0]"
:
:
:
"r"
(
&
rw
->
lock
),
"r"
(
0
)
:
"r"
(
&
rw
->
lock
),
"r"
(
0
)
:
"cc"
,
"memory"
);
:
"cc"
);
}
}
/*
/*
...
@@ -149,13 +168,17 @@ static inline void _raw_read_lock(rwlock_t *rw)
...
@@ -149,13 +168,17 @@ static inline void _raw_read_lock(rwlock_t *rw)
" bmi 1b"
" bmi 1b"
:
"=&r"
(
tmp
),
"=&r"
(
tmp2
)
:
"=&r"
(
tmp
),
"=&r"
(
tmp2
)
:
"r"
(
&
rw
->
lock
)
:
"r"
(
&
rw
->
lock
)
:
"cc"
,
"memory"
);
:
"cc"
);
smp_mb
();
}
}
static
inline
void
_raw_read_unlock
(
rwlock_t
*
rw
)
static
inline
void
_raw_read_unlock
(
rwlock_t
*
rw
)
{
{
unsigned
long
tmp
,
tmp2
;
unsigned
long
tmp
,
tmp2
;
smp_mb
();
__asm__
__volatile__
(
__asm__
__volatile__
(
"1: ldrex %0, [%2]
\n
"
"1: ldrex %0, [%2]
\n
"
" sub %0, %0, #1
\n
"
" sub %0, %0, #1
\n
"
...
@@ -164,7 +187,7 @@ static inline void _raw_read_unlock(rwlock_t *rw)
...
@@ -164,7 +187,7 @@ static inline void _raw_read_unlock(rwlock_t *rw)
" bne 1b"
" bne 1b"
:
"=&r"
(
tmp
),
"=&r"
(
tmp2
)
:
"=&r"
(
tmp
),
"=&r"
(
tmp2
)
:
"r"
(
&
rw
->
lock
)
:
"r"
(
&
rw
->
lock
)
:
"cc"
,
"memory"
);
:
"cc"
);
}
}
#define _raw_read_trylock(lock) generic_raw_read_trylock(lock)
#define _raw_read_trylock(lock) generic_raw_read_trylock(lock)
...
...
include/asm-arm/system.h
View file @
0983f050
...
@@ -139,7 +139,12 @@ extern unsigned int user_debug;
...
@@ -139,7 +139,12 @@ extern unsigned int user_debug;
#define vectors_high() (0)
#define vectors_high() (0)
#endif
#endif
#if __LINUX_ARM_ARCH__ >= 6
#define mb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" \
: : "r" (0) : "memory")
#else
#define mb() __asm__ __volatile__ ("" : : : "memory")
#define mb() __asm__ __volatile__ ("" : : : "memory")
#endif
#define rmb() mb()
#define rmb() mb()
#define wmb() mb()
#define wmb() mb()
#define read_barrier_depends() do { } while(0)
#define read_barrier_depends() do { } while(0)
...
@@ -323,12 +328,8 @@ do { \
...
@@ -323,12 +328,8 @@ do { \
* NOTE that this solution won't work on an SMP system, so explcitly
* NOTE that this solution won't work on an SMP system, so explcitly
* forbid it here.
* forbid it here.
*/
*/
#ifdef CONFIG_SMP
#error SMP is not supported on SA1100/SA110
#else
#define swp_is_buggy
#define swp_is_buggy
#endif
#endif
#endif
static
inline
unsigned
long
__xchg
(
unsigned
long
x
,
volatile
void
*
ptr
,
int
size
)
static
inline
unsigned
long
__xchg
(
unsigned
long
x
,
volatile
void
*
ptr
,
int
size
)
{
{
...
@@ -337,35 +338,68 @@ static inline unsigned long __xchg(unsigned long x, volatile void *ptr, int size
...
@@ -337,35 +338,68 @@ static inline unsigned long __xchg(unsigned long x, volatile void *ptr, int size
#ifdef swp_is_buggy
#ifdef swp_is_buggy
unsigned
long
flags
;
unsigned
long
flags
;
#endif
#endif
#if __LINUX_ARM_ARCH__ >= 6
unsigned
int
tmp
;
#endif
switch
(
size
)
{
switch
(
size
)
{
#ifdef swp_is_buggy
#if __LINUX_ARM_ARCH__ >= 6
case
1
:
case
1
:
local_irq_save
(
flags
);
asm
volatile
(
"@ __xchg1
\n
"
ret
=
*
(
volatile
unsigned
char
*
)
ptr
;
"1: ldrexb %0, [%3]
\n
"
*
(
volatile
unsigned
char
*
)
ptr
=
x
;
" strexb %1, %2, [%3]
\n
"
local_irq_restore
(
flags
);
" teq %1, #0
\n
"
break
;
" bne 1b"
:
"=&r"
(
ret
),
"=&r"
(
tmp
)
case
4
:
:
"r"
(
x
),
"r"
(
ptr
)
local_irq_save
(
flags
);
:
"memory"
,
"cc"
);
ret
=
*
(
volatile
unsigned
long
*
)
ptr
;
break
;
*
(
volatile
unsigned
long
*
)
ptr
=
x
;
case
4
:
local_irq_restore
(
flags
);
asm
volatile
(
"@ __xchg4
\n
"
break
;
"1: ldrex %0, [%3]
\n
"
" strex %1, %2, [%3]
\n
"
" teq %1, #0
\n
"
" bne 1b"
:
"=&r"
(
ret
),
"=&r"
(
tmp
)
:
"r"
(
x
),
"r"
(
ptr
)
:
"memory"
,
"cc"
);
break
;
#elif defined(swp_is_buggy)
#ifdef CONFIG_SMP
#error SMP is not supported on this platform
#endif
case
1
:
local_irq_save
(
flags
);
ret
=
*
(
volatile
unsigned
char
*
)
ptr
;
*
(
volatile
unsigned
char
*
)
ptr
=
x
;
local_irq_restore
(
flags
);
break
;
case
4
:
local_irq_save
(
flags
);
ret
=
*
(
volatile
unsigned
long
*
)
ptr
;
*
(
volatile
unsigned
long
*
)
ptr
=
x
;
local_irq_restore
(
flags
);
break
;
#else
#else
case
1
:
__asm__
__volatile__
(
"swpb %0, %1, [%2]"
case
1
:
:
"=&r"
(
ret
)
asm
volatile
(
"@ __xchg1
\n
"
:
"r"
(
x
),
"r"
(
ptr
)
" swpb %0, %1, [%2]"
:
"memory"
,
"cc"
);
:
"=&r"
(
ret
)
break
;
:
"r"
(
x
),
"r"
(
ptr
)
case
4
:
__asm__
__volatile__
(
"swp %0, %1, [%2]"
:
"memory"
,
"cc"
);
:
"=&r"
(
ret
)
break
;
:
"r"
(
x
),
"r"
(
ptr
)
case
4
:
:
"memory"
,
"cc"
);
asm
volatile
(
"@ __xchg4
\n
"
break
;
" swp %0, %1, [%2]"
:
"=&r"
(
ret
)
:
"r"
(
x
),
"r"
(
ptr
)
:
"memory"
,
"cc"
);
break
;
#endif
#endif
default:
__bad_xchg
(
ptr
,
size
),
ret
=
0
;
default:
__bad_xchg
(
ptr
,
size
),
ret
=
0
;
break
;
}
}
return
ret
;
return
ret
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment