summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorkettenis <kettenis@openbsd.org>2011-11-27 21:02:38 +0000
committerkettenis <kettenis@openbsd.org>2011-11-27 21:02:38 +0000
commit4a61e410cc0141a83350e9116c4fd8164ba0a7ea (patch)
tree7ad46a702db89d7a62a312151308420a47b1877d
parentSwitch to proper PA-RISC 2.0 mnemonics. Works around a bug in the assembler (diff)
downloadwireguard-openbsd-4a61e410cc0141a83350e9116c4fd8164ba0a7ea.tar.xz
wireguard-openbsd-4a61e410cc0141a83350e9116c4fd8164ba0a7ea.zip
Regen.
-rw-r--r--sys/lib/libkern/arch/hppa64/bcopy.S68
-rw-r--r--sys/lib/libkern/arch/hppa64/spcopy.S68
2 files changed, 68 insertions, 68 deletions
diff --git a/sys/lib/libkern/arch/hppa64/bcopy.S b/sys/lib/libkern/arch/hppa64/bcopy.S
index 78e548192b6..f59e0571e38 100644
--- a/sys/lib/libkern/arch/hppa64/bcopy.S
+++ b/sys/lib/libkern/arch/hppa64/bcopy.S
@@ -48,63 +48,63 @@ ALTENTRY(bcopy)
addi -16, %arg2, %arg2
$bcopy_f.loop16a
ldw 32(%sr0, %arg0), %r0
- ldws,ma 4(%sr0, %arg0), %r22
- ldws,ma 4(%sr0, %arg0), %r21
- ldws,ma 4(%sr0, %arg0), %r20
- ldws,ma 4(%sr0, %arg0), %r19
- stbys,b,m %r22, 4(%sr0, %arg1)
- stws,ma %r21, 4(%sr0, %arg1)
- stws,ma %r20, 4(%sr0, %arg1)
+ ldw,ma 4(%sr0, %arg0), %r22
+ ldw,ma 4(%sr0, %arg0), %r21
+ ldw,ma 4(%sr0, %arg0), %r20
+ ldw,ma 4(%sr0, %arg0), %r19
+ stby,b,m %r22, 4(%sr0, %arg1)
+ stw,ma %r21, 4(%sr0, %arg1)
+ stw,ma %r20, 4(%sr0, %arg1)
addib,*>= -16, %arg2, $bcopy_f.loop16a
- stws,ma %r19, 4(%sr0, %arg1)
+ stw,ma %r19, 4(%sr0, %arg1)
addib,*<,n 12, %arg2, $bcopy_f.cleanup
$bcopy_f.word
- ldws,ma 4(%sr0, %arg0), %r22
+ ldw,ma 4(%sr0, %arg0), %r22
addib,*>= -4, %arg2, $bcopy_f.word
- stws,ma %r22, 4(%sr0, %arg1)
+ stw,ma %r22, 4(%sr0, %arg1)
$bcopy_f.cleanup
addib,*=,n 4, %arg2, $bcopy_f.done
- ldws 0(%sr0, %arg0), %r22
+ ldw 0(%sr0, %arg0), %r22
add %arg1, %arg2, %arg1
b $bcopy_f.done
- stbys,e %r22, 0(%sr0, %arg1)
+ stby,e %r22, 0(%sr0, %arg1)
$bcopy_f.unaligned
sub,*>= %r19, %r20, %r21
- ldwm 4(%sr0, %arg0), %ret1
+ ldw,ma 4(%sr0, %arg0), %ret1
depd,z %r21, 60, 61, %r22
mtsar %r22
addi -16, %arg2, %arg2
$bcopy_f.loop16u
ldw 32(%sr0, %arg0), %r0
- ldws,ma 4(%sr0, %arg0), %r22
- ldws,ma 4(%sr0, %arg0), %r21
- ldws,ma 4(%sr0, %arg0), %r20
- ldws,ma 4(%sr0, %arg0), %r19
- vshd %ret1, %r22, %r31
- stbys,b,m %r31, 4(%sr0, %arg1)
- vshd %r22, %r21, %r31
- stws,ma %r31, 4(%sr0, %arg1)
- vshd %r21, %r20, %r31
- stws,ma %r31, 4(%sr0, %arg1)
- vshd %r20, %r19, %r31
- stws,ma %r31, 4(%sr0, %arg1)
+ ldw,ma 4(%sr0, %arg0), %r22
+ ldw,ma 4(%sr0, %arg0), %r21
+ ldw,ma 4(%sr0, %arg0), %r20
+ ldw,ma 4(%sr0, %arg0), %r19
+ shrpw %ret1, %r22, %sar, %r31
+ stby,b,m %r31, 4(%sr0, %arg1)
+ shrpw %r22, %r21, %sar, %r31
+ stw,ma %r31, 4(%sr0, %arg1)
+ shrpw %r21, %r20, %sar, %r31
+ stw,ma %r31, 4(%sr0, %arg1)
+ shrpw %r20, %r19, %sar, %r31
+ stw,ma %r31, 4(%sr0, %arg1)
addib,*>= -16, %arg2, $bcopy_f.loop16u
copy %r19, %ret1
addib,*<,n 12, %arg2, $bcopy_f.cleanup_un
$bcopy_f.word_un
- ldws,ma 4(%sr0, %arg0), %r22
- vshd %ret1, %r22, %r21
+ ldw,ma 4(%sr0, %arg0), %r22
+ shrpw %ret1, %r22, %sar, %r21
addib,*< -4, %arg2, $bcopy_f.cleanup1_un
- stws,ma %r21, 4(%sr0, %arg1)
- ldws,ma 4(%sr0, %arg0), %ret1
- vshd %r22, %ret1, %r21
+ stw,ma %r21, 4(%sr0, %arg1)
+ ldw,ma 4(%sr0, %arg0), %ret1
+ shrpw %r22, %ret1, %sar, %r21
addib,*>= -4, %arg2, $bcopy_f.word_un
- stws,ma %r21, 4(%sr0, %arg1)
+ stw,ma %r21, 4(%sr0, %arg1)
$bcopy_f.cleanup_un
addib,*<=,n 4, %arg2, $bcopy_f.done
@@ -112,10 +112,10 @@ $bcopy_f.cleanup_un
add %arg1, %arg2, %arg1
extrd,u %r19, 60, 2, %r19
sub,*<= %arg2, %r19, %r0
- ldws,ma 4(%sr0, %arg0), %r22
- vshd %ret1, %r22, %r21
+ ldw,ma 4(%sr0, %arg0), %r22
+ shrpw %ret1, %r22, %sar, %r21
b $bcopy_f.done
- stbys,e %r21, 0(%sr0, %arg1)
+ stby,e %r21, 0(%sr0, %arg1)
$bcopy_f.cleanup1_un
b $bcopy_f.cleanup_un
diff --git a/sys/lib/libkern/arch/hppa64/spcopy.S b/sys/lib/libkern/arch/hppa64/spcopy.S
index 73e7bc37a24..9b3ecb9acbc 100644
--- a/sys/lib/libkern/arch/hppa64/spcopy.S
+++ b/sys/lib/libkern/arch/hppa64/spcopy.S
@@ -71,63 +71,63 @@ LEAF_ENTRY(spcopy)
addi -16, %ret0, %ret0
$spcopy.loop16a
ldw 32(%sr1, %arg1), %r0
- ldws,ma 4(%sr1, %arg1), %r22
- ldws,ma 4(%sr1, %arg1), %r21
- ldws,ma 4(%sr1, %arg1), %r20
- ldws,ma 4(%sr1, %arg1), %r19
- stbys,b,m %r22, 4(%sr2, %arg3)
- stws,ma %r21, 4(%sr2, %arg3)
- stws,ma %r20, 4(%sr2, %arg3)
+ ldw,ma 4(%sr1, %arg1), %r22
+ ldw,ma 4(%sr1, %arg1), %r21
+ ldw,ma 4(%sr1, %arg1), %r20
+ ldw,ma 4(%sr1, %arg1), %r19
+ stby,b,m %r22, 4(%sr2, %arg3)
+ stw,ma %r21, 4(%sr2, %arg3)
+ stw,ma %r20, 4(%sr2, %arg3)
addib,*>= -16, %ret0, $spcopy.loop16a
- stws,ma %r19, 4(%sr2, %arg3)
+ stw,ma %r19, 4(%sr2, %arg3)
addib,*<,n 12, %ret0, $spcopy.cleanup
$spcopy.word
- ldws,ma 4(%sr1, %arg1), %r22
+ ldw,ma 4(%sr1, %arg1), %r22
addib,*>= -4, %ret0, $spcopy.word
- stws,ma %r22, 4(%sr2, %arg3)
+ stw,ma %r22, 4(%sr2, %arg3)
$spcopy.cleanup
addib,*=,n 4, %ret0, $spcopy.done
- ldws 0(%sr1, %arg1), %r22
+ ldw 0(%sr1, %arg1), %r22
add %arg3, %ret0, %arg3
b $spcopy.done
- stbys,e %r22, 0(%sr2, %arg3)
+ stby,e %r22, 0(%sr2, %arg3)
$spcopy.unaligned
sub,*>= %r19, %r20, %r21
- ldwm 4(%sr1, %arg1), %ret1
+ ldw,ma 4(%sr1, %arg1), %ret1
depd,z %r21, 60, 61, %r22
mtsar %r22
addi -16, %ret0, %ret0
$spcopy.loop16u
ldw 32(%sr1, %arg1), %r0
- ldws,ma 4(%sr1, %arg1), %r22
- ldws,ma 4(%sr1, %arg1), %r21
- ldws,ma 4(%sr1, %arg1), %r20
- ldws,ma 4(%sr1, %arg1), %r19
- vshd %ret1, %r22, %r31
- stbys,b,m %r31, 4(%sr2, %arg3)
- vshd %r22, %r21, %r31
- stws,ma %r31, 4(%sr2, %arg3)
- vshd %r21, %r20, %r31
- stws,ma %r31, 4(%sr2, %arg3)
- vshd %r20, %r19, %r31
- stws,ma %r31, 4(%sr2, %arg3)
+ ldw,ma 4(%sr1, %arg1), %r22
+ ldw,ma 4(%sr1, %arg1), %r21
+ ldw,ma 4(%sr1, %arg1), %r20
+ ldw,ma 4(%sr1, %arg1), %r19
+ shrpw %ret1, %r22, %sar, %r31
+ stby,b,m %r31, 4(%sr2, %arg3)
+ shrpw %r22, %r21, %sar, %r31
+ stw,ma %r31, 4(%sr2, %arg3)
+ shrpw %r21, %r20, %sar, %r31
+ stw,ma %r31, 4(%sr2, %arg3)
+ shrpw %r20, %r19, %sar, %r31
+ stw,ma %r31, 4(%sr2, %arg3)
addib,*>= -16, %ret0, $spcopy.loop16u
copy %r19, %ret1
addib,*<,n 12, %ret0, $spcopy.cleanup_un
$spcopy.word_un
- ldws,ma 4(%sr1, %arg1), %r22
- vshd %ret1, %r22, %r21
+ ldw,ma 4(%sr1, %arg1), %r22
+ shrpw %ret1, %r22, %sar, %r21
addib,*< -4, %ret0, $spcopy.cleanup1_un
- stws,ma %r21, 4(%sr2, %arg3)
- ldws,ma 4(%sr1, %arg1), %ret1
- vshd %r22, %ret1, %r21
+ stw,ma %r21, 4(%sr2, %arg3)
+ ldw,ma 4(%sr1, %arg1), %ret1
+ shrpw %r22, %ret1, %sar, %r21
addib,*>= -4, %ret0, $spcopy.word_un
- stws,ma %r21, 4(%sr2, %arg3)
+ stw,ma %r21, 4(%sr2, %arg3)
$spcopy.cleanup_un
addib,*<=,n 4, %ret0, $spcopy.done
@@ -135,10 +135,10 @@ $spcopy.cleanup_un
add %arg3, %ret0, %arg3
extrd,u %r19, 60, 2, %r19
sub,*<= %ret0, %r19, %r0
- ldws,ma 4(%sr1, %arg1), %r22
- vshd %ret1, %r22, %r21
+ ldw,ma 4(%sr1, %arg1), %r22
+ shrpw %ret1, %r22, %sar, %r21
b $spcopy.done
- stbys,e %r21, 0(%sr2, %arg3)
+ stby,e %r21, 0(%sr2, %arg3)
$spcopy.cleanup1_un
b $spcopy.cleanup_un