Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
ab7c828b
Commit
ab7c828b
authored
Jun 08, 2003
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge
bk://are.twiddle.net/axp-2.5
into home.transmeta.com:/home/torvalds/v2.5/linux
parents
6c94e37b
4964368c
Changes
11
Show whitespace changes
Inline
Side-by-side
Showing
11 changed files
with
87 additions
and
97 deletions
+87
-97
arch/alpha/kernel/asm-offsets.c
arch/alpha/kernel/asm-offsets.c
+1
-0
arch/alpha/kernel/head.S
arch/alpha/kernel/head.S
+2
-1
arch/alpha/kernel/systbls.S
arch/alpha/kernel/systbls.S
+1
-0
arch/alpha/kernel/traps.c
arch/alpha/kernel/traps.c
+40
-40
arch/alpha/lib/csum_partial_copy.c
arch/alpha/lib/csum_partial_copy.c
+2
-2
arch/alpha/lib/memmove.S
arch/alpha/lib/memmove.S
+9
-0
include/asm-alpha/ptrace.h
include/asm-alpha/ptrace.h
+2
-9
include/asm-alpha/string.h
include/asm-alpha/string.h
+1
-0
include/asm-alpha/uaccess.h
include/asm-alpha/uaccess.h
+26
-43
include/asm-alpha/unaligned.h
include/asm-alpha/unaligned.h
+1
-1
include/asm-alpha/unistd.h
include/asm-alpha/unistd.h
+2
-1
No files found.
arch/alpha/kernel/asm-offsets.c
View file @
ab7c828b
...
@@ -31,6 +31,7 @@ void foo(void)
...
@@ -31,6 +31,7 @@ void foo(void)
DEFINE
(
TASK_TGID
,
offsetof
(
struct
task_struct
,
tgid
));
DEFINE
(
TASK_TGID
,
offsetof
(
struct
task_struct
,
tgid
));
BLANK
();
BLANK
();
DEFINE
(
SIZEOF_PT_REGS
,
sizeof
(
struct
pt_regs
));
DEFINE
(
PT_PTRACED
,
PT_PTRACED
);
DEFINE
(
PT_PTRACED
,
PT_PTRACED
);
DEFINE
(
CLONE_VM
,
CLONE_VM
);
DEFINE
(
CLONE_VM
,
CLONE_VM
);
DEFINE
(
CLONE_UNTRACED
,
CLONE_UNTRACED
);
DEFINE
(
CLONE_UNTRACED
,
CLONE_UNTRACED
);
...
...
arch/alpha/kernel/head.S
View file @
ab7c828b
...
@@ -9,6 +9,7 @@
...
@@ -9,6 +9,7 @@
#include <linux/config.h>
#include <linux/config.h>
#include <asm/system.h>
#include <asm/system.h>
#include <asm/asm_offsets.h>
.
globl
swapper_pg_dir
.
globl
swapper_pg_dir
.
globl
_stext
.
globl
_stext
...
@@ -25,7 +26,7 @@ __start:
...
@@ -25,7 +26,7 @@ __start:
/
*
We
need
to
get
current_task_info
loaded
up
...
*/
/
*
We
need
to
get
current_task_info
loaded
up
...
*/
lda
$
8
,
init_thread_union
lda
$
8
,
init_thread_union
/
*
...
and
find
our
stack
...
*/
/
*
...
and
find
our
stack
...
*/
lda
$
30
,
0x4000
(
$
8
)
lda
$
30
,
0x4000
-
SIZEOF_PT_REGS
(
$
8
)
/
*
...
and
then
we
can
start
the
kernel
.
*/
/
*
...
and
then
we
can
start
the
kernel
.
*/
jsr
$
26
,
start_kernel
jsr
$
26
,
start_kernel
call_pal
PAL_halt
call_pal
PAL_halt
...
...
arch/alpha/kernel/systbls.S
View file @
ab7c828b
...
@@ -442,6 +442,7 @@ sys_call_table:
...
@@ -442,6 +442,7 @@ sys_call_table:
.
quad
sys_clock_gettime
/*
420
*/
.
quad
sys_clock_gettime
/*
420
*/
.
quad
sys_clock_getres
.
quad
sys_clock_getres
.
quad
sys_clock_nanosleep
.
quad
sys_clock_nanosleep
.
quad
sys_semtimedop
.
size
sys_call_table
,
.
-
sys_call_table
.
size
sys_call_table
,
.
-
sys_call_table
.
type
sys_call_table
,
@
object
.
type
sys_call_table
,
@
object
...
...
arch/alpha/kernel/traps.c
View file @
ab7c828b
...
@@ -485,9 +485,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -485,9 +485,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
" extwh %2,%3,%2
\n
"
" extwh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -505,9 +505,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -505,9 +505,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
" extlh %2,%3,%2
\n
"
" extlh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -525,9 +525,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -525,9 +525,9 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
" extqh %2,%3,%2
\n
"
" extqh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -554,13 +554,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -554,13 +554,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
@@ -584,13 +584,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -584,13 +584,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
@@ -614,13 +614,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
...
@@ -614,13 +614,13 @@ do_entUna(void * va, unsigned long opcode, unsigned long reg,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n\t
"
".section __ex_table,
\"
a
\"\n\t
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
@@ -845,9 +845,9 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -845,9 +845,9 @@ do_entUnaUser(void * va, unsigned long opcode,
" extwh %2,%3,%2
\n
"
" extwh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -865,9 +865,9 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -865,9 +865,9 @@ do_entUnaUser(void * va, unsigned long opcode,
" extlh %2,%3,%2
\n
"
" extlh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -885,9 +885,9 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -885,9 +885,9 @@ do_entUnaUser(void * va, unsigned long opcode,
" extqh %2,%3,%2
\n
"
" extqh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -905,9 +905,9 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -905,9 +905,9 @@ do_entUnaUser(void * va, unsigned long opcode,
" extlh %2,%3,%2
\n
"
" extlh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -925,9 +925,9 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -925,9 +925,9 @@ do_entUnaUser(void * va, unsigned long opcode,
" extqh %2,%3,%2
\n
"
" extqh %2,%3,%2
\n
"
"3:
\n
"
"3:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %1,3b-1b(%0)
\n
"
" lda %1,3b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %2,3b-2b(%0)
\n
"
" lda %2,3b-2b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
)
...
@@ -954,13 +954,13 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -954,13 +954,13 @@ do_entUnaUser(void * va, unsigned long opcode,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
@@ -988,13 +988,13 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -988,13 +988,13 @@ do_entUnaUser(void * va, unsigned long opcode,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n
"
".section __ex_table,
\"
a
\"\n
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
@@ -1022,13 +1022,13 @@ do_entUnaUser(void * va, unsigned long opcode,
...
@@ -1022,13 +1022,13 @@ do_entUnaUser(void * va, unsigned long opcode,
"4: stq_u %1,0(%5)
\n
"
"4: stq_u %1,0(%5)
\n
"
"5:
\n
"
"5:
\n
"
".section __ex_table,
\"
a
\"\n\t
"
".section __ex_table,
\"
a
\"\n\t
"
" .
gprel32 1b
\n
"
" .
long 1b - .
\n
"
" lda %2,5b-1b(%0)
\n
"
" lda %2,5b-1b(%0)
\n
"
" .
gprel32 2b
\n
"
" .
long 2b - .
\n
"
" lda %1,5b-2b(%0)
\n
"
" lda %1,5b-2b(%0)
\n
"
" .
gprel32 3b
\n
"
" .
long 3b - .
\n
"
" lda $31,5b-3b(%0)
\n
"
" lda $31,5b-3b(%0)
\n
"
" .
gprel32 4b
\n
"
" .
long 4b - .
\n
"
" lda $31,5b-4b(%0)
\n
"
" lda $31,5b-4b(%0)
\n
"
".previous"
".previous"
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
:
"=r"
(
error
),
"=&r"
(
tmp1
),
"=&r"
(
tmp2
),
...
...
arch/alpha/lib/csum_partial_copy.c
View file @
ab7c828b
...
@@ -46,7 +46,7 @@ __asm__ __volatile__("insqh %1,%2,%0":"=r" (z):"r" (x),"r" (y))
...
@@ -46,7 +46,7 @@ __asm__ __volatile__("insqh %1,%2,%0":"=r" (z):"r" (x),"r" (y))
"1: ldq_u %0,%2\n" \
"1: ldq_u %0,%2\n" \
"2:\n" \
"2:\n" \
".section __ex_table,\"a\"\n" \
".section __ex_table,\"a\"\n" \
" .
gprel32 1b
\n" \
" .
long 1b - .
\n" \
" lda %0,2b-1b(%1)\n" \
" lda %0,2b-1b(%1)\n" \
".previous" \
".previous" \
: "=r"(x), "=r"(__guu_err) \
: "=r"(x), "=r"(__guu_err) \
...
@@ -61,7 +61,7 @@ __asm__ __volatile__("insqh %1,%2,%0":"=r" (z):"r" (x),"r" (y))
...
@@ -61,7 +61,7 @@ __asm__ __volatile__("insqh %1,%2,%0":"=r" (z):"r" (x),"r" (y))
"1: stq_u %2,%1\n" \
"1: stq_u %2,%1\n" \
"2:\n" \
"2:\n" \
".section __ex_table,\"a\"\n" \
".section __ex_table,\"a\"\n" \
" .
gprel32 1b
" \
" .
long 1b - .
" \
" lda $31,2b-1b(%0)\n" \
" lda $31,2b-1b(%0)\n" \
".previous" \
".previous" \
: "=r"(__puu_err) \
: "=r"(__puu_err) \
...
...
arch/alpha/lib/memmove.S
View file @
ab7c828b
...
@@ -11,6 +11,15 @@
...
@@ -11,6 +11,15 @@
.
set
noreorder
.
set
noreorder
.
text
.
text
.
align
4
.
globl
bcopy
.
ent
bcopy
bcopy
:
mov
$
16
,
$
0
mov
$
17
,
$
16
mov
$
0
,
$
17
.
end
bcopy
.
align
4
.
align
4
.
globl
memmove
.
globl
memmove
.
ent
memmove
.
ent
memmove
...
...
include/asm-alpha/ptrace.h
View file @
ab7c828b
...
@@ -71,15 +71,8 @@ struct switch_stack {
...
@@ -71,15 +71,8 @@ struct switch_stack {
#define instruction_pointer(regs) ((regs)->pc)
#define instruction_pointer(regs) ((regs)->pc)
extern
void
show_regs
(
struct
pt_regs
*
);
extern
void
show_regs
(
struct
pt_regs
*
);
/*
#define alpha_task_regs(task) \
* TODO: if kernel-only threads do not have a dummy pt_regs structure at the
((struct pt_regs *) ((long) (task)->thread_info + 2*PAGE_SIZE) - 1)
* top of the stack, this would cause kernel stack corruption. Either check
* first that we're not dealing with a kernel thread or change the kernel
* stacks to allocate a dummy pt_regs structure.
*/
#define alpha_task_regs(task) ((struct pt_regs *) \
((long) task->thread_info + PAGE_SIZE) - 1)
#define force_successful_syscall_return() (alpha_task_regs(current)->r0 = 0)
#define force_successful_syscall_return() (alpha_task_regs(current)->r0 = 0)
...
...
include/asm-alpha/string.h
View file @
ab7c828b
...
@@ -13,6 +13,7 @@
...
@@ -13,6 +13,7 @@
#define __HAVE_ARCH_MEMCPY
#define __HAVE_ARCH_MEMCPY
extern
void
*
memcpy
(
void
*
,
const
void
*
,
size_t
);
extern
void
*
memcpy
(
void
*
,
const
void
*
,
size_t
);
#define __HAVE_ARCH_MEMMOVE
#define __HAVE_ARCH_MEMMOVE
#define __HAVE_ARCH_BCOPY
extern
void
*
memmove
(
void
*
,
const
void
*
,
size_t
);
extern
void
*
memmove
(
void
*
,
const
void
*
,
size_t
);
/* For backward compatibility with modules. Unused otherwise. */
/* For backward compatibility with modules. Unused otherwise. */
...
...
include/asm-alpha/uaccess.h
View file @
ab7c828b
...
@@ -340,25 +340,31 @@ __asm__ __volatile__("1: stb %r2,%1\n" \
...
@@ -340,25 +340,31 @@ __asm__ __volatile__("1: stb %r2,%1\n" \
* Complex access routines
* Complex access routines
*/
*/
/* This little bit of silliness is to get the GP loaded for a function
that ordinarily wouldn't. Otherwise we could have it done by the macro
directly, which can be optimized the linker. */
#ifdef MODULE
#define __module_address(sym) "r"(sym),
#define __module_call(ra, arg, sym) "jsr $" #ra ",(%" #arg ")," #sym
#else
#define __module_address(sym)
#define __module_call(ra, arg, sym) "bsr $" #ra "," #sym " !samegp"
#endif
extern
void
__copy_user
(
void
);
extern
void
__copy_user
(
void
);
extern
inline
long
extern
inline
long
__copy_tofrom_user_nocheck
(
void
*
to
,
const
void
*
from
,
long
len
)
__copy_tofrom_user_nocheck
(
void
*
to
,
const
void
*
from
,
long
len
)
{
{
/* This little bit of silliness is to get the GP loaded for
a function that ordinarily wouldn't. Otherwise we could
have it done by the macro directly, which can be optimized
the linker. */
register
void
*
pv
__asm__
(
"$27"
)
=
__copy_user
;
register
void
*
__cu_to
__asm__
(
"$6"
)
=
to
;
register
void
*
__cu_to
__asm__
(
"$6"
)
=
to
;
register
const
void
*
__cu_from
__asm__
(
"$7"
)
=
from
;
register
const
void
*
__cu_from
__asm__
(
"$7"
)
=
from
;
register
long
__cu_len
__asm__
(
"$0"
)
=
len
;
register
long
__cu_len
__asm__
(
"$0"
)
=
len
;
__asm__
__volatile__
(
__asm__
__volatile__
(
"jsr $28,(%3),__copy_user"
__module_call
(
28
,
3
,
__copy_user
)
:
"=r"
(
__cu_len
),
"=r"
(
__cu_from
),
"=r"
(
__cu_to
),
"=r"
(
pv
)
:
"=r"
(
__cu_len
),
"=r"
(
__cu_from
),
"=r"
(
__cu_to
)
:
"0"
(
__cu_len
),
"1"
(
__cu_from
),
"2"
(
__cu_to
),
"3"
(
pv
)
:
__module_address
(
__copy_user
)
"0"
(
__cu_len
),
"1"
(
__cu_from
),
"2"
(
__cu_to
)
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
return
__cu_len
;
return
__cu_len
;
...
@@ -367,20 +373,8 @@ __copy_tofrom_user_nocheck(void *to, const void *from, long len)
...
@@ -367,20 +373,8 @@ __copy_tofrom_user_nocheck(void *to, const void *from, long len)
extern
inline
long
extern
inline
long
__copy_tofrom_user
(
void
*
to
,
const
void
*
from
,
long
len
,
const
void
*
validate
)
__copy_tofrom_user
(
void
*
to
,
const
void
*
from
,
long
len
,
const
void
*
validate
)
{
{
if
(
__access_ok
((
long
)
validate
,
len
,
get_fs
()))
{
if
(
__access_ok
((
long
)
validate
,
len
,
get_fs
()))
register
void
*
pv
__asm__
(
"$27"
)
=
__copy_user
;
len
=
__copy_tofrom_user_nocheck
(
to
,
from
,
len
);
register
void
*
__cu_to
__asm__
(
"$6"
)
=
to
;
register
const
void
*
__cu_from
__asm__
(
"$7"
)
=
from
;
register
long
__cu_len
__asm__
(
"$0"
)
=
len
;
__asm__
__volatile__
(
"jsr $28,(%3),__copy_user"
:
"=r"
(
__cu_len
),
"=r"
(
__cu_from
),
"=r"
(
__cu_to
),
"=r"
(
pv
)
:
"0"
(
__cu_len
),
"1"
(
__cu_from
),
"2"
(
__cu_to
),
"3"
(
pv
)
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
len
=
__cu_len
;
}
return
len
;
return
len
;
}
}
...
@@ -404,18 +398,13 @@ extern void __do_clear_user(void);
...
@@ -404,18 +398,13 @@ extern void __do_clear_user(void);
extern
inline
long
extern
inline
long
__clear_user
(
void
*
to
,
long
len
)
__clear_user
(
void
*
to
,
long
len
)
{
{
/* This little bit of silliness is to get the GP loaded for
a function that ordinarily wouldn't. Otherwise we could
have it done by the macro directly, which can be optimized
the linker. */
register
void
*
pv
__asm__
(
"$27"
)
=
__do_clear_user
;
register
void
*
__cl_to
__asm__
(
"$6"
)
=
to
;
register
void
*
__cl_to
__asm__
(
"$6"
)
=
to
;
register
long
__cl_len
__asm__
(
"$0"
)
=
len
;
register
long
__cl_len
__asm__
(
"$0"
)
=
len
;
__asm__
__volatile__
(
__asm__
__volatile__
(
"jsr $28,(%2),__do_clear_user"
__module_call
(
28
,
2
,
__do_clear_user
)
:
"=r"
(
__cl_len
),
"=r"
(
__cl_to
),
"=r"
(
pv
)
:
"=r"
(
__cl_len
),
"=r"
(
__cl_to
)
:
"0"
(
__cl_len
),
"1"
(
__cl_to
),
"2"
(
pv
)
:
__module_address
(
__do_clear_user
)
"0"
(
__cl_len
),
"1"
(
__cl_to
)
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
return
__cl_len
;
return
__cl_len
;
}
}
...
@@ -423,20 +412,14 @@ __clear_user(void *to, long len)
...
@@ -423,20 +412,14 @@ __clear_user(void *to, long len)
extern
inline
long
extern
inline
long
clear_user
(
void
*
to
,
long
len
)
clear_user
(
void
*
to
,
long
len
)
{
{
if
(
__access_ok
((
long
)
to
,
len
,
get_fs
()))
{
if
(
__access_ok
((
long
)
to
,
len
,
get_fs
()))
register
void
*
pv
__asm__
(
"$27"
)
=
__do_clear_user
;
len
=
__clear_user
(
to
,
len
);
register
void
*
__cl_to
__asm__
(
"$6"
)
=
to
;
register
long
__cl_len
__asm__
(
"$0"
)
=
len
;
__asm__
__volatile__
(
"jsr $28,(%2),__do_clear_user"
:
"=r"
(
__cl_len
),
"=r"
(
__cl_to
),
"=r"
(
pv
)
:
"0"
(
__cl_len
),
"1"
(
__cl_to
),
"2"
(
pv
)
:
"$1"
,
"$2"
,
"$3"
,
"$4"
,
"$5"
,
"$28"
,
"memory"
);
len
=
__cl_len
;
}
return
len
;
return
len
;
}
}
#undef __module_address
#undef __module_call
/* Returns: -EFAULT if exception before terminator, N if the entire
/* Returns: -EFAULT if exception before terminator, N if the entire
buffer filled, else strlen. */
buffer filled, else strlen. */
...
...
include/asm-alpha/unaligned.h
View file @
ab7c828b
...
@@ -14,7 +14,7 @@
...
@@ -14,7 +14,7 @@
* the get/put functions are indeed always optimized,
* the get/put functions are indeed always optimized,
* and that we use the correct sizes.
* and that we use the correct sizes.
*/
*/
extern
void
bad_unaligned_access_length
(
void
);
extern
void
bad_unaligned_access_length
(
void
)
__attribute__
((
noreturn
))
;
/*
/*
* EGCS 1.1 knows about arbitrary unaligned loads. Define some
* EGCS 1.1 knows about arbitrary unaligned loads. Define some
...
...
include/asm-alpha/unistd.h
View file @
ab7c828b
...
@@ -358,7 +358,8 @@
...
@@ -358,7 +358,8 @@
#define __NR_clock_gettime 420
#define __NR_clock_gettime 420
#define __NR_clock_getres 421
#define __NR_clock_getres 421
#define __NR_clock_nanosleep 422
#define __NR_clock_nanosleep 422
#define NR_SYSCALLS 423
#define __NR_semtimedop 423
#define NR_SYSCALLS 424
#if defined(__GNUC__)
#if defined(__GNUC__)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment