Commit bcfc6e63 authored by Kai Backman's avatar Kai Backman

fixed register usage and removed some dead code.

R=rsc
APPROVED=rsc
DELTA=17  (0 added, 12 deleted, 5 changed)
OCL=34659
CL=34665
parent ff196fbd
...@@ -31,31 +31,19 @@ arg=0 ...@@ -31,31 +31,19 @@ arg=0
/* replaced use of R10 by R11 because the former can be the data segment base register */ /* replaced use of R10 by R11 because the former can be the data segment base register */
TEXT _mulv(SB), $0 TEXT _mulv(SB), $0
MOVW 8(FP), R9 /* l0 */ MOVW 8(FP), R2 /* l0 */
MOVW 4(FP), R11 /* h0 */ MOVW 4(FP), R3 /* h0 */
MOVW 16(FP), R4 /* l1 */ MOVW 16(FP), R4 /* l1 */
MOVW 12(FP), R5 /* h1 */ MOVW 12(FP), R5 /* h1 */
UMULL(4, 9, 7, 6, 0) UMULL(4, 2, 7, 6, 0)
MUL(11, 4, 8, 0) MUL(3, 4, 8, 0)
ADD R8, R7 ADD R8, R7
MUL(9, 5, 8, 0) MUL(2, 5, 8, 0)
ADD R8, R7 ADD R8, R7
MOVW R6, 4(R(arg)) MOVW R6, 4(R(arg))
MOVW R7, 0(R(arg)) MOVW R7, 0(R(arg))
RET RET
/* multiply, add, and right-shift, yielding a 32-bit result, while
using 64-bit accuracy for the multiply -- for fast fixed-point math */
TEXT _mularsv(SB), $0
MOVW 4(FP), R11 /* m1 */
MOVW 8(FP), R8 /* a */
MOVW 12(FP), R4 /* rs */
MOVW $0, R9
UMLAL(0, 11, 9, 8, 0)
MOVW R8>>R4, R8
RSB $32, R4, R4
ORR R9<<R4, R8, R0
RET
Q = 0 Q = 0
N = 1 N = 1
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment