From: Kai Backman Date: Tue, 15 Sep 2009 23:15:17 +0000 (-0700) Subject: fixed register usage and removed some dead code. X-Git-Tag: weekly.2009-11-06~570 X-Git-Url: http://www.git.cypherpunks.su/?a=commitdiff_plain;h=bcfc6e631fbc0f06f2d975301ce71554c197970a;p=gostls13.git fixed register usage and removed some dead code. R=rsc APPROVED=rsc DELTA=17 (0 added, 12 deleted, 5 changed) OCL=34659 CL=34665 --- diff --git a/src/pkg/runtime/arm/vlop.s b/src/pkg/runtime/arm/vlop.s index 010e62ae65..56771ec840 100644 --- a/src/pkg/runtime/arm/vlop.s +++ b/src/pkg/runtime/arm/vlop.s @@ -31,31 +31,19 @@ arg=0 /* replaced use of R10 by R11 because the former can be the data segment base register */ TEXT _mulv(SB), $0 - MOVW 8(FP), R9 /* l0 */ - MOVW 4(FP), R11 /* h0 */ + MOVW 8(FP), R2 /* l0 */ + MOVW 4(FP), R3 /* h0 */ MOVW 16(FP), R4 /* l1 */ MOVW 12(FP), R5 /* h1 */ - UMULL(4, 9, 7, 6, 0) - MUL(11, 4, 8, 0) + UMULL(4, 2, 7, 6, 0) + MUL(3, 4, 8, 0) ADD R8, R7 - MUL(9, 5, 8, 0) + MUL(2, 5, 8, 0) ADD R8, R7 MOVW R6, 4(R(arg)) MOVW R7, 0(R(arg)) RET -/* multiply, add, and right-shift, yielding a 32-bit result, while - using 64-bit accuracy for the multiply -- for fast fixed-point math */ -TEXT _mularsv(SB), $0 - MOVW 4(FP), R11 /* m1 */ - MOVW 8(FP), R8 /* a */ - MOVW 12(FP), R4 /* rs */ - MOVW $0, R9 - UMLAL(0, 11, 9, 8, 0) - MOVW R8>>R4, R8 - RSB $32, R4, R4 - ORR R9<