886 lines
17 KiB
ArmAsm
886 lines
17 KiB
ArmAsm
# Copyright (c) 2014, Google Inc.
|
|
#
|
|
# Permission to use, copy, modify, and/or distribute this software for any
|
|
# purpose with or without fee is hereby granted, provided that the above
|
|
# copyright notice and this permission notice appear in all copies.
|
|
#
|
|
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
|
|
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
|
|
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
|
|
# SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
|
|
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
|
|
# OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
|
|
# CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
|
|
|
# This file contains a pre-compiled version of chacha_vec.c for ARM. This is
|
|
# needed to support switching on NEON code at runtime. If the whole of OpenSSL
|
|
# were to be compiled with the needed flags to build chacha_vec.c, then it
|
|
# wouldn't be possible to run on non-NEON systems.
|
|
#
|
|
# This file was generated by:
|
|
#
|
|
# /opt/gcc-linaro-arm-linux-gnueabihf-4.7-2012.10-20121022_linux/bin/arm-linux-gnueabihf-gcc -O3 -mcpu=cortex-a8 -mfpu=neon -S chacha_vec.c -I ../../include -fpic -o chacha_vec_arm.S
|
|
|
|
.syntax unified
|
|
.cpu cortex-a8
|
|
.eabi_attribute 27, 3
|
|
|
|
# EABI attribute 28 sets whether VFP register arguments were used to build this
|
|
# file. If object files are inconsistent on this point, the linker will refuse
|
|
# to link them. Thus we report whatever the compiler expects since we don't use
|
|
# VFP arguments.
|
|
|
|
#if defined(__ARM_PCS_VFP)
|
|
.eabi_attribute 28, 1
|
|
#else
|
|
.eabi_attribute 28, 0
|
|
#endif
|
|
|
|
.fpu neon
|
|
.eabi_attribute 20, 1
|
|
.eabi_attribute 21, 1
|
|
.eabi_attribute 23, 3
|
|
.eabi_attribute 24, 1
|
|
.eabi_attribute 25, 1
|
|
.eabi_attribute 26, 2
|
|
.eabi_attribute 30, 2
|
|
.eabi_attribute 34, 1
|
|
.eabi_attribute 18, 4
|
|
.thumb
|
|
.file "chacha_vec.c"
|
|
.text
|
|
.align 2
|
|
.global CRYPTO_chacha_20_neon
|
|
.thumb
|
|
.thumb_func
|
|
.type CRYPTO_chacha_20_neon, %function
|
|
CRYPTO_chacha_20_neon:
|
|
@ args = 8, pretend = 0, frame = 304
|
|
@ frame_needed = 1, uses_anonymous_args = 0
|
|
@ link register save eliminated.
|
|
push {r4, r5, r6, r7, r8, r9, sl, fp}
|
|
fstmfdd sp!, {d8, d9, d10, d11, d12, d13, d14, d15}
|
|
sub sp, sp, #304
|
|
add r7, sp, #0
|
|
movw ip, #43691
|
|
movt ip, 43690
|
|
str r2, [r7, #196]
|
|
sub sp, sp, #96
|
|
ldr r4, [r7, #196]
|
|
ldr r6, [r7, #400]
|
|
ldr r2, .L38+16
|
|
umull r4, ip, ip, r4
|
|
ldr r6, [r6, #0]
|
|
ldr r8, [r7, #400]
|
|
.LPIC24:
|
|
add r2, pc
|
|
add r4, sp, #15
|
|
str r3, [r7, #244]
|
|
str r6, [r7, #176]
|
|
bic r4, r4, #15
|
|
str r0, [r7, #188]
|
|
str r4, [r7, #200]
|
|
lsrs ip, ip, #7
|
|
str r1, [r7, #184]
|
|
ldmia r2, {r0, r1, r2, r3}
|
|
ldr r4, [r8, #4]
|
|
ldr r5, [r7, #244]
|
|
vld1.64 {d24-d25}, [r5:64]
|
|
vldr d26, [r5, #16]
|
|
vldr d27, [r5, #24]
|
|
ldr r9, [r7, #200]
|
|
ldr r8, [r7, #404]
|
|
ldr r5, [r7, #176]
|
|
add r6, r9, #64
|
|
str r4, [r7, #300]
|
|
mov r4, #0
|
|
str r8, [r7, #288]
|
|
str r5, [r7, #296]
|
|
str r4, [r7, #292]
|
|
stmia r6, {r0, r1, r2, r3}
|
|
vldr d22, [r9, #64]
|
|
vldr d23, [r9, #72]
|
|
vldr d20, [r7, #288]
|
|
vldr d21, [r7, #296]
|
|
str ip, [r7, #192]
|
|
beq .L20
|
|
lsl r6, ip, #1
|
|
ldr r1, [r9, #68]
|
|
add r3, r6, ip
|
|
str r6, [r7, #180]
|
|
ldr r2, [r9, #72]
|
|
add r8, r8, #2
|
|
ldr r5, [r9, #76]
|
|
vldr d18, .L38
|
|
vldr d19, .L38+8
|
|
str r4, [r7, #240]
|
|
ldr r6, [r7, #184]
|
|
ldr r4, [r7, #188]
|
|
str r0, [r7, #224]
|
|
str r1, [r7, #220]
|
|
str r8, [r7, #208]
|
|
str r2, [r7, #216]
|
|
str r3, [r7, #204]
|
|
str r5, [r7, #212]
|
|
str r6, [r7, #252]
|
|
str r4, [r7, #248]
|
|
.L4:
|
|
ldr r2, [r7, #244]
|
|
add r9, r7, #216
|
|
ldr r3, [r7, #244]
|
|
vadd.i32 q8, q10, q9
|
|
ldr r6, [r7, #208]
|
|
vmov q15, q13 @ v4si
|
|
ldr r5, [r7, #240]
|
|
vmov q3, q12 @ v4si
|
|
ldr r4, [r7, #244]
|
|
vmov q2, q11 @ v4si
|
|
adds r5, r5, r6
|
|
ldr r2, [r2, #8]
|
|
ldr r6, [r7, #400]
|
|
vmov q5, q10 @ v4si
|
|
ldr r3, [r3, #12]
|
|
vmov q1, q13 @ v4si
|
|
ldr r0, [r7, #244]
|
|
vmov q0, q12 @ v4si
|
|
ldr r1, [r7, #244]
|
|
vmov q4, q11 @ v4si
|
|
ldmia r9, {r9, sl, fp}
|
|
str r5, [r7, #228]
|
|
ldr r5, [r4, #24]
|
|
ldr r0, [r0, #0]
|
|
ldr r1, [r1, #4]
|
|
str r2, [r7, #264]
|
|
str r3, [r7, #236]
|
|
ldr r2, [r6, #4]
|
|
ldr r3, [r4, #28]
|
|
str r5, [r7, #280]
|
|
ldr r5, [r6, #0]
|
|
movs r6, #0
|
|
ldr ip, [r7, #228]
|
|
ldr r8, [r7, #212]
|
|
str r0, [r7, #232]
|
|
str r1, [r7, #268]
|
|
ldr r0, [r4, #16]
|
|
ldr r1, [r4, #20]
|
|
movs r4, #10
|
|
str r2, [r7, #24]
|
|
str r3, [r7, #284]
|
|
str r4, [r7, #256]
|
|
ldr r2, [r7, #264]
|
|
str r9, [r7, #276]
|
|
mov r9, r6
|
|
ldr r6, [r7, #280]
|
|
str r8, [r7, #260]
|
|
mov r8, sl
|
|
str r1, [r7, #272]
|
|
mov sl, ip
|
|
str r6, [r7, #264]
|
|
mov r6, r5
|
|
ldr r3, [r7, #236]
|
|
mov r5, r0
|
|
ldr ip, [r7, #24]
|
|
ldr r1, [r7, #268]
|
|
ldr r0, [r7, #232]
|
|
b .L39
|
|
.L40:
|
|
.align 3
|
|
.L38:
|
|
.word 1
|
|
.word 0
|
|
.word 0
|
|
.word 0
|
|
.word .LANCHOR0-(.LPIC24+4)
|
|
.L39:
|
|
.L3:
|
|
vadd.i32 q4, q4, q0
|
|
add r8, r8, r1
|
|
vadd.i32 q2, q2, q3
|
|
str r8, [r7, #268]
|
|
veor q5, q5, q4
|
|
ldr r8, [r7, #276]
|
|
veor q8, q8, q2
|
|
add fp, fp, r0
|
|
str fp, [r7, #280]
|
|
add r8, r8, r2
|
|
vrev32.16 q5, q5
|
|
str r8, [r7, #276]
|
|
vrev32.16 q8, q8
|
|
vadd.i32 q1, q1, q5
|
|
vadd.i32 q15, q15, q8
|
|
ldr r8, [r7, #280]
|
|
veor q0, q1, q0
|
|
ldr r4, [r7, #260]
|
|
veor q3, q15, q3
|
|
eor sl, sl, r8
|
|
ldr r8, [r7, #276]
|
|
add fp, r4, r3
|
|
vshl.i32 q7, q0, #12
|
|
ldr r4, [r7, #268]
|
|
vshl.i32 q6, q3, #12
|
|
eor r6, r6, r8
|
|
eor r9, r9, r4
|
|
ldr r4, [r7, #272]
|
|
vsri.32 q7, q0, #20
|
|
ror r8, r6, #16
|
|
ldr r6, [r7, #264]
|
|
eor ip, ip, fp
|
|
vsri.32 q6, q3, #20
|
|
ror sl, sl, #16
|
|
ror r9, r9, #16
|
|
add r5, r5, sl
|
|
vadd.i32 q4, q4, q7
|
|
str r5, [r7, #236]
|
|
vadd.i32 q2, q2, q6
|
|
add r5, r4, r9
|
|
add r4, r6, r8
|
|
ldr r6, [r7, #284]
|
|
ror ip, ip, #16
|
|
veor q5, q4, q5
|
|
veor q8, q2, q8
|
|
add r6, r6, ip
|
|
str r6, [r7, #264]
|
|
eors r1, r1, r5
|
|
ldr r6, [r7, #236]
|
|
vshl.i32 q3, q5, #8
|
|
vshl.i32 q14, q8, #8
|
|
eors r2, r2, r4
|
|
eors r0, r0, r6
|
|
ldr r6, [r7, #264]
|
|
vsri.32 q3, q5, #24
|
|
ror r1, r1, #20
|
|
eors r3, r3, r6
|
|
ldr r6, [r7, #280]
|
|
ror r0, r0, #20
|
|
vsri.32 q14, q8, #24
|
|
adds r6, r0, r6
|
|
str r6, [r7, #284]
|
|
ldr r6, [r7, #268]
|
|
vadd.i32 q1, q1, q3
|
|
vadd.i32 q15, q15, q14
|
|
ror r2, r2, #20
|
|
adds r6, r1, r6
|
|
str r6, [r7, #260]
|
|
ldr r6, [r7, #276]
|
|
veor q6, q15, q6
|
|
veor q7, q1, q7
|
|
ror r3, r3, #20
|
|
adds r6, r2, r6
|
|
str r6, [r7, #280]
|
|
ldr r6, [r7, #284]
|
|
vshl.i32 q0, q6, #7
|
|
vshl.i32 q5, q7, #7
|
|
add fp, r3, fp
|
|
eor sl, r6, sl
|
|
ldr r6, [r7, #260]
|
|
eor ip, fp, ip
|
|
vsri.32 q0, q6, #25
|
|
eor r9, r6, r9
|
|
ldr r6, [r7, #280]
|
|
ror sl, sl, #24
|
|
vsri.32 q5, q7, #25
|
|
eor r8, r6, r8
|
|
ldr r6, [r7, #236]
|
|
ror r9, r9, #24
|
|
ror ip, ip, #24
|
|
add r6, sl, r6
|
|
str r6, [r7, #276]
|
|
ldr r6, [r7, #264]
|
|
add r5, r9, r5
|
|
str r5, [r7, #272]
|
|
vext.32 q5, q5, q5, #1
|
|
add r5, ip, r6
|
|
ldr r6, [r7, #276]
|
|
vext.32 q0, q0, q0, #1
|
|
vadd.i32 q4, q4, q5
|
|
eors r0, r0, r6
|
|
ldr r6, [r7, #272]
|
|
vadd.i32 q2, q2, q0
|
|
vext.32 q3, q3, q3, #3
|
|
ror r8, r8, #24
|
|
eors r1, r1, r6
|
|
vext.32 q14, q14, q14, #3
|
|
add r4, r8, r4
|
|
ldr r6, [r7, #284]
|
|
veor q3, q4, q3
|
|
veor q14, q2, q14
|
|
eors r2, r2, r4
|
|
ror r1, r1, #25
|
|
vext.32 q1, q1, q1, #2
|
|
adds r6, r1, r6
|
|
str r6, [r7, #284]
|
|
vext.32 q15, q15, q15, #2
|
|
ldr r6, [r7, #260]
|
|
eors r3, r3, r5
|
|
ror r2, r2, #25
|
|
vrev32.16 q8, q14
|
|
adds r6, r2, r6
|
|
vrev32.16 q3, q3
|
|
str r6, [r7, #268]
|
|
vadd.i32 q1, q1, q3
|
|
ldr r6, [r7, #280]
|
|
vadd.i32 q15, q15, q8
|
|
ror r3, r3, #25
|
|
veor q5, q1, q5
|
|
adds r6, r3, r6
|
|
veor q0, q15, q0
|
|
str r6, [r7, #264]
|
|
ldr r6, [r7, #268]
|
|
ror r0, r0, #25
|
|
add fp, r0, fp
|
|
vshl.i32 q6, q5, #12
|
|
eor sl, r6, sl
|
|
ldr r6, [r7, #284]
|
|
vshl.i32 q14, q0, #12
|
|
eor r8, fp, r8
|
|
eor ip, r6, ip
|
|
ldr r6, [r7, #264]
|
|
vsri.32 q6, q5, #20
|
|
ror sl, sl, #16
|
|
eor r9, r6, r9
|
|
ror r6, r8, #16
|
|
vsri.32 q14, q0, #20
|
|
ldr r8, [r7, #272]
|
|
ror ip, ip, #16
|
|
add r5, sl, r5
|
|
add r8, r6, r8
|
|
add r4, ip, r4
|
|
str r4, [r7, #236]
|
|
eor r0, r8, r0
|
|
str r5, [r7, #280]
|
|
vadd.i32 q4, q4, q6
|
|
ldr r5, [r7, #236]
|
|
vadd.i32 q2, q2, q14
|
|
ldr r4, [r7, #276]
|
|
ror r0, r0, #20
|
|
veor q3, q4, q3
|
|
eors r1, r1, r5
|
|
veor q0, q2, q8
|
|
str r8, [r7, #272]
|
|
str r0, [r7, #24]
|
|
add fp, r0, fp
|
|
ldr r8, [r7, #280]
|
|
ror r9, r9, #16
|
|
ldr r0, [r7, #284]
|
|
add r4, r9, r4
|
|
str fp, [r7, #260]
|
|
ror r1, r1, #20
|
|
add fp, r1, r0
|
|
eor r2, r8, r2
|
|
ldr r0, [r7, #260]
|
|
eors r3, r3, r4
|
|
vshl.i32 q5, q3, #8
|
|
str r4, [r7, #232]
|
|
vshl.i32 q8, q0, #8
|
|
ldr r4, [r7, #268]
|
|
ldr r5, [r7, #264]
|
|
ror r2, r2, #20
|
|
ror r3, r3, #20
|
|
eors r6, r6, r0
|
|
adds r5, r3, r5
|
|
add r8, r2, r4
|
|
vsri.32 q5, q3, #24
|
|
ldr r4, [r7, #272]
|
|
eor r9, r5, r9
|
|
eor ip, fp, ip
|
|
vsri.32 q8, q0, #24
|
|
eor sl, r8, sl
|
|
ror r6, r6, #24
|
|
ldr r0, [r7, #280]
|
|
str r5, [r7, #276]
|
|
adds r4, r6, r4
|
|
ldr r5, [r7, #236]
|
|
vadd.i32 q1, q1, q5
|
|
str r4, [r7, #272]
|
|
vadd.i32 q15, q15, q8
|
|
ldr r4, [r7, #232]
|
|
ror ip, ip, #24
|
|
ror sl, sl, #24
|
|
ror r9, r9, #24
|
|
add r5, ip, r5
|
|
add r0, sl, r0
|
|
str r5, [r7, #264]
|
|
add r5, r9, r4
|
|
str r0, [r7, #284]
|
|
veor q6, q1, q6
|
|
ldr r4, [r7, #24]
|
|
veor q14, q15, q14
|
|
ldr r0, [r7, #272]
|
|
eors r3, r3, r5
|
|
vshl.i32 q0, q6, #7
|
|
vext.32 q1, q1, q1, #2
|
|
eors r0, r0, r4
|
|
ldr r4, [r7, #284]
|
|
str r0, [r7, #280]
|
|
vshl.i32 q3, q14, #7
|
|
eors r2, r2, r4
|
|
ldr r4, [r7, #280]
|
|
ldr r0, [r7, #264]
|
|
vsri.32 q0, q6, #25
|
|
ror r2, r2, #25
|
|
ror r3, r3, #25
|
|
eors r1, r1, r0
|
|
vsri.32 q3, q14, #25
|
|
ror r0, r4, #25
|
|
ldr r4, [r7, #256]
|
|
ror r1, r1, #25
|
|
vext.32 q5, q5, q5, #1
|
|
subs r4, r4, #1
|
|
str r4, [r7, #256]
|
|
vext.32 q15, q15, q15, #2
|
|
vext.32 q8, q8, q8, #1
|
|
vext.32 q0, q0, q0, #3
|
|
vext.32 q3, q3, q3, #3
|
|
bne .L3
|
|
ldr r4, [r7, #264]
|
|
vadd.i32 q14, q10, q9
|
|
str r2, [r7, #264]
|
|
vadd.i32 q10, q10, q5
|
|
ldr r2, [r7, #252]
|
|
vld1.64 {d12-d13}, [r2:64]
|
|
ldr r2, [r7, #220]
|
|
vadd.i32 q4, q11, q4
|
|
str ip, [r7, #24]
|
|
mov ip, sl
|
|
mov sl, r8
|
|
ldr r8, [r7, #260]
|
|
add sl, sl, r2
|
|
ldr r2, [r7, #212]
|
|
str r4, [r7, #280]
|
|
vadd.i32 q0, q12, q0
|
|
ldr r4, [r7, #224]
|
|
add r8, r8, r2
|
|
ldr r2, [r7, #240]
|
|
vadd.i32 q1, q13, q1
|
|
str r0, [r7, #232]
|
|
add fp, fp, r4
|
|
mov r0, r5
|
|
ldr r4, [r7, #216]
|
|
mov r5, r6
|
|
mov r6, r9
|
|
ldr r9, [r7, #276]
|
|
adds r2, r2, #3
|
|
str r2, [r7, #240]
|
|
vadd.i32 q2, q11, q2
|
|
ldr r2, [r7, #252]
|
|
add r9, r9, r4
|
|
vadd.i32 q3, q12, q3
|
|
ldr r4, [r7, #228]
|
|
vadd.i32 q15, q13, q15
|
|
str r1, [r7, #268]
|
|
vadd.i32 q8, q14, q8
|
|
str r3, [r7, #236]
|
|
veor q4, q4, q6
|
|
ldr r3, [r7, #284]
|
|
ldr r1, [r7, #272]
|
|
add ip, r4, ip
|
|
ldr r4, [r7, #248]
|
|
vst1.64 {d8-d9}, [r4:64]
|
|
vldr d8, [r2, #16]
|
|
vldr d9, [r2, #24]
|
|
veor q0, q0, q4
|
|
vstr d0, [r4, #16]
|
|
vstr d1, [r4, #24]
|
|
vldr d0, [r2, #32]
|
|
vldr d1, [r2, #40]
|
|
veor q1, q1, q0
|
|
vstr d2, [r4, #32]
|
|
vstr d3, [r4, #40]
|
|
vldr d2, [r2, #48]
|
|
vldr d3, [r2, #56]
|
|
veor q10, q10, q1
|
|
vstr d20, [r4, #48]
|
|
vstr d21, [r4, #56]
|
|
vldr d8, [r2, #64]
|
|
vldr d9, [r2, #72]
|
|
veor q2, q2, q4
|
|
vstr d4, [r4, #64]
|
|
vstr d5, [r4, #72]
|
|
vldr d10, [r2, #80]
|
|
vldr d11, [r2, #88]
|
|
veor q3, q3, q5
|
|
vstr d6, [r4, #80]
|
|
vstr d7, [r4, #88]
|
|
vldr d12, [r2, #96]
|
|
vldr d13, [r2, #104]
|
|
veor q15, q15, q6
|
|
vstr d30, [r4, #96]
|
|
vstr d31, [r4, #104]
|
|
vldr d20, [r2, #112]
|
|
vldr d21, [r2, #120]
|
|
veor q8, q8, q10
|
|
vstr d16, [r4, #112]
|
|
vstr d17, [r4, #120]
|
|
ldr r4, [r2, #128]
|
|
ldr r2, [r7, #248]
|
|
vadd.i32 q10, q14, q9
|
|
eor r4, fp, r4
|
|
vadd.i32 q10, q10, q9
|
|
str r4, [r2, #128]
|
|
ldr r4, [r7, #252]
|
|
ldr r2, [r4, #132]
|
|
eor r2, sl, r2
|
|
ldr sl, [r7, #248]
|
|
str r2, [sl, #132]
|
|
ldr r2, [r4, #136]
|
|
eor r2, r9, r2
|
|
str r2, [sl, #136]
|
|
ldr r2, [r4, #140]
|
|
eor r2, r8, r2
|
|
str r2, [sl, #140]
|
|
ldr r2, [r7, #244]
|
|
ldr r4, [r4, #144]
|
|
ldr r2, [r2, #0]
|
|
str r4, [r7, #44]
|
|
ldr r4, [r7, #232]
|
|
add r8, r4, r2
|
|
ldr r2, [r7, #44]
|
|
ldr r4, [r7, #244]
|
|
eor r8, r8, r2
|
|
ldr r2, [r7, #252]
|
|
str r8, [sl, #144]
|
|
ldr r4, [r4, #4]
|
|
ldr r2, [r2, #148]
|
|
str r2, [r7, #40]
|
|
ldr r2, [r7, #268]
|
|
add r8, r2, r4
|
|
ldr r4, [r7, #40]
|
|
ldr r2, [r7, #244]
|
|
eor r8, r8, r4
|
|
ldr r4, [r7, #252]
|
|
str r8, [sl, #148]
|
|
ldr r2, [r2, #8]
|
|
ldr r4, [r4, #152]
|
|
str r4, [r7, #36]
|
|
ldr r4, [r7, #264]
|
|
add r8, r4, r2
|
|
ldr r2, [r7, #36]
|
|
eor r8, r8, r2
|
|
str r8, [sl, #152]
|
|
ldr r2, [r7, #252]
|
|
ldr r4, [r7, #244]
|
|
ldr r2, [r2, #156]
|
|
ldr r4, [r4, #12]
|
|
str r2, [r7, #32]
|
|
ldr r2, [r7, #236]
|
|
add r8, r2, r4
|
|
ldr r4, [r7, #32]
|
|
ldr r2, [r7, #252]
|
|
eor r8, r8, r4
|
|
str r8, [sl, #156]
|
|
ldr r8, [r7, #244]
|
|
ldr r2, [r2, #160]
|
|
ldr r4, [r8, #16]
|
|
adds r0, r0, r4
|
|
ldr r4, [r7, #252]
|
|
eors r0, r0, r2
|
|
str r0, [sl, #160]
|
|
ldr r0, [r8, #20]
|
|
ldr r2, [r4, #164]
|
|
adds r1, r1, r0
|
|
ldr r0, [r7, #280]
|
|
eors r1, r1, r2
|
|
str r1, [sl, #164]
|
|
ldr r2, [r8, #24]
|
|
ldr r1, [r4, #168]
|
|
adds r2, r0, r2
|
|
eors r2, r2, r1
|
|
str r2, [sl, #168]
|
|
ldr r1, [r8, #28]
|
|
ldr r2, [r4, #172]
|
|
adds r3, r3, r1
|
|
eors r3, r3, r2
|
|
str r3, [sl, #172]
|
|
ldr r3, [r4, #176]
|
|
eor r3, ip, r3
|
|
str r3, [sl, #176]
|
|
ldr r3, [r4, #180]
|
|
ldr r4, [r7, #400]
|
|
eors r6, r6, r3
|
|
str r6, [sl, #180]
|
|
ldr r6, [r7, #252]
|
|
ldr r2, [r4, #0]
|
|
ldr r3, [r6, #184]
|
|
adds r5, r5, r2
|
|
eors r5, r5, r3
|
|
str r5, [sl, #184]
|
|
ldr r2, [r6, #188]
|
|
adds r6, r6, #192
|
|
ldr r3, [r4, #4]
|
|
str r6, [r7, #252]
|
|
ldr r0, [r7, #24]
|
|
ldr r1, [r7, #240]
|
|
adds r4, r0, r3
|
|
eors r4, r4, r2
|
|
ldr r2, [r7, #204]
|
|
str r4, [sl, #188]
|
|
add sl, sl, #192
|
|
cmp r1, r2
|
|
str sl, [r7, #248]
|
|
bne .L4
|
|
ldr r4, [r7, #192]
|
|
ldr r3, [r7, #180]
|
|
ldr r6, [r7, #188]
|
|
adds r5, r3, r4
|
|
ldr r8, [r7, #184]
|
|
lsls r5, r5, #6
|
|
adds r4, r6, r5
|
|
add r5, r8, r5
|
|
.L2:
|
|
ldr r9, [r7, #196]
|
|
movw r3, #43691
|
|
movt r3, 43690
|
|
ldr sl, [r7, #196]
|
|
umull r9, r3, r3, r9
|
|
lsrs r3, r3, #7
|
|
add r3, r3, r3, lsl #1
|
|
sub r3, sl, r3, lsl #6
|
|
lsrs r6, r3, #6
|
|
beq .L5
|
|
add r1, r5, #16
|
|
add r2, r4, #16
|
|
mov r0, r6
|
|
vldr d30, .L41
|
|
vldr d31, .L41+8
|
|
.L6:
|
|
vmov q8, q10 @ v4si
|
|
movs r3, #10
|
|
vmov q1, q13 @ v4si
|
|
vmov q14, q12 @ v4si
|
|
vmov q3, q11 @ v4si
|
|
.L7:
|
|
vadd.i32 q3, q3, q14
|
|
subs r3, r3, #1
|
|
veor q2, q8, q3
|
|
vrev32.16 q2, q2
|
|
vadd.i32 q8, q1, q2
|
|
veor q9, q8, q14
|
|
vshl.i32 q14, q9, #12
|
|
vsri.32 q14, q9, #20
|
|
vadd.i32 q3, q3, q14
|
|
veor q2, q3, q2
|
|
vshl.i32 q9, q2, #8
|
|
vsri.32 q9, q2, #24
|
|
vadd.i32 q8, q8, q9
|
|
vext.32 q9, q9, q9, #3
|
|
veor q14, q8, q14
|
|
vext.32 q1, q8, q8, #2
|
|
vshl.i32 q8, q14, #7
|
|
vsri.32 q8, q14, #25
|
|
vext.32 q8, q8, q8, #1
|
|
vadd.i32 q3, q3, q8
|
|
veor q2, q3, q9
|
|
vrev32.16 q2, q2
|
|
vadd.i32 q9, q1, q2
|
|
veor q8, q9, q8
|
|
vshl.i32 q14, q8, #12
|
|
vsri.32 q14, q8, #20
|
|
vadd.i32 q3, q3, q14
|
|
veor q2, q3, q2
|
|
vshl.i32 q8, q2, #8
|
|
vsri.32 q8, q2, #24
|
|
vadd.i32 q9, q9, q8
|
|
vext.32 q8, q8, q8, #1
|
|
veor q14, q9, q14
|
|
vext.32 q1, q9, q9, #2
|
|
vshl.i32 q9, q14, #7
|
|
vsri.32 q9, q14, #25
|
|
vext.32 q14, q9, q9, #3
|
|
bne .L7
|
|
vadd.i32 q8, q10, q8
|
|
subs r0, r0, #1
|
|
vadd.i32 q3, q11, q3
|
|
vldr d0, [r1, #-16]
|
|
vldr d1, [r1, #-8]
|
|
vadd.i32 q14, q12, q14
|
|
vadd.i32 q1, q13, q1
|
|
veor q3, q3, q0
|
|
vstr d6, [r2, #-16]
|
|
vstr d7, [r2, #-8]
|
|
vadd.i32 q10, q10, q15
|
|
vld1.64 {d8-d9}, [r1:64]
|
|
veor q14, q14, q4
|
|
vst1.64 {d28-d29}, [r2:64]
|
|
vldr d10, [r1, #16]
|
|
vldr d11, [r1, #24]
|
|
veor q1, q1, q5
|
|
vstr d2, [r2, #16]
|
|
vstr d3, [r2, #24]
|
|
vldr d18, [r1, #32]
|
|
vldr d19, [r1, #40]
|
|
add r1, r1, #64
|
|
veor q8, q8, q9
|
|
vstr d16, [r2, #32]
|
|
vstr d17, [r2, #40]
|
|
add r2, r2, #64
|
|
bne .L6
|
|
lsls r6, r6, #6
|
|
adds r4, r4, r6
|
|
adds r5, r5, r6
|
|
.L5:
|
|
ldr r6, [r7, #196]
|
|
ands ip, r6, #63
|
|
beq .L1
|
|
vmov q8, q10 @ v4si
|
|
movs r3, #10
|
|
vmov q14, q13 @ v4si
|
|
vmov q9, q12 @ v4si
|
|
vmov q15, q11 @ v4si
|
|
.L10:
|
|
vadd.i32 q15, q15, q9
|
|
subs r3, r3, #1
|
|
veor q8, q8, q15
|
|
vrev32.16 q8, q8
|
|
vadd.i32 q3, q14, q8
|
|
veor q9, q3, q9
|
|
vshl.i32 q14, q9, #12
|
|
vsri.32 q14, q9, #20
|
|
vadd.i32 q15, q15, q14
|
|
veor q9, q15, q8
|
|
vshl.i32 q8, q9, #8
|
|
vsri.32 q8, q9, #24
|
|
vadd.i32 q9, q3, q8
|
|
vext.32 q8, q8, q8, #3
|
|
veor q2, q9, q14
|
|
vext.32 q14, q9, q9, #2
|
|
vshl.i32 q9, q2, #7
|
|
vsri.32 q9, q2, #25
|
|
vext.32 q9, q9, q9, #1
|
|
vadd.i32 q15, q15, q9
|
|
veor q3, q15, q8
|
|
vrev32.16 q3, q3
|
|
vadd.i32 q14, q14, q3
|
|
veor q8, q14, q9
|
|
vshl.i32 q9, q8, #12
|
|
vsri.32 q9, q8, #20
|
|
vadd.i32 q15, q15, q9
|
|
veor q3, q15, q3
|
|
vshl.i32 q8, q3, #8
|
|
vsri.32 q8, q3, #24
|
|
vadd.i32 q14, q14, q8
|
|
vext.32 q8, q8, q8, #1
|
|
veor q3, q14, q9
|
|
vext.32 q14, q14, q14, #2
|
|
vshl.i32 q9, q3, #7
|
|
vsri.32 q9, q3, #25
|
|
vext.32 q9, q9, q9, #3
|
|
bne .L10
|
|
cmp ip, #15
|
|
vadd.i32 q11, q11, q15
|
|
bhi .L37
|
|
ldr r9, [r7, #200]
|
|
vst1.64 {d22-d23}, [r9:128]
|
|
.L14:
|
|
ldr sl, [r7, #196]
|
|
and r3, sl, #48
|
|
cmp ip, r3
|
|
bls .L1
|
|
adds r0, r5, r3
|
|
adds r1, r4, r3
|
|
add r2, r0, #16
|
|
add r6, r1, #16
|
|
cmp r1, r2
|
|
it cc
|
|
cmpcc r0, r6
|
|
rsb r9, r3, ip
|
|
ite cc
|
|
movcc r2, #0
|
|
movcs r2, #1
|
|
cmp r9, #15
|
|
ite ls
|
|
movls r2, #0
|
|
andhi r2, r2, #1
|
|
lsr r8, r9, #4
|
|
eor r2, r2, #1
|
|
cmp r8, #0
|
|
it eq
|
|
orreq r2, r2, #1
|
|
lsl sl, r8, #4
|
|
cbnz r2, .L35
|
|
ldr fp, [r7, #200]
|
|
add r6, fp, r3
|
|
.L17:
|
|
vld1.8 {q8}, [r0]!
|
|
adds r2, r2, #1
|
|
cmp r8, r2
|
|
vld1.8 {q9}, [r6]!
|
|
veor q8, q9, q8
|
|
vst1.8 {q8}, [r1]!
|
|
bhi .L17
|
|
cmp r9, sl
|
|
add r3, r3, sl
|
|
beq .L1
|
|
.L35:
|
|
ldr r0, [r7, #200]
|
|
.L25:
|
|
ldrb r2, [r5, r3] @ zero_extendqisi2
|
|
ldrb r1, [r3, r0] @ zero_extendqisi2
|
|
eors r2, r2, r1
|
|
strb r2, [r4, r3]
|
|
adds r3, r3, #1
|
|
cmp ip, r3
|
|
bhi .L25
|
|
.L1:
|
|
add r7, r7, #304
|
|
mov sp, r7
|
|
fldmfdd sp!, {d8, d9, d10, d11, d12, d13, d14, d15}
|
|
pop {r4, r5, r6, r7, r8, r9, sl, fp}
|
|
bx lr
|
|
.L37:
|
|
cmp ip, #31
|
|
vld1.64 {d0-d1}, [r5:64]
|
|
vadd.i32 q9, q12, q9
|
|
veor q11, q11, q0
|
|
vst1.64 {d22-d23}, [r4:64]
|
|
bls .L12
|
|
cmp ip, #47
|
|
vldr d2, [r5, #16]
|
|
vldr d3, [r5, #24]
|
|
vadd.i32 q13, q13, q14
|
|
veor q9, q9, q1
|
|
vstr d18, [r4, #16]
|
|
vstr d19, [r4, #24]
|
|
bls .L13
|
|
vadd.i32 q8, q8, q10
|
|
vldr d0, [r5, #32]
|
|
vldr d1, [r5, #40]
|
|
ldr r6, [r7, #200]
|
|
vstr d16, [r6, #48]
|
|
vstr d17, [r6, #56]
|
|
veor q8, q13, q0
|
|
vstr d16, [r4, #32]
|
|
vstr d17, [r4, #40]
|
|
b .L14
|
|
.L12:
|
|
ldr r8, [r7, #200]
|
|
vstr d18, [r8, #16]
|
|
vstr d19, [r8, #24]
|
|
b .L14
|
|
.L20:
|
|
ldr r5, [r7, #184]
|
|
ldr r4, [r7, #188]
|
|
b .L2
|
|
.L13:
|
|
ldr r6, [r7, #200]
|
|
vstr d26, [r6, #32]
|
|
vstr d27, [r6, #40]
|
|
b .L14
|
|
.L42:
|
|
.align 3
|
|
.L41:
|
|
.word 1
|
|
.word 0
|
|
.word 0
|
|
.word 0
|
|
.size CRYPTO_chacha_20_neon, .-CRYPTO_chacha_20_neon
|
|
.section .rodata
|
|
.align 3
|
|
.LANCHOR0 = . + 0
|
|
.LC0:
|
|
.word 1634760805
|
|
.word 857760878
|
|
.word 2036477234
|
|
.word 1797285236
|
|
.ident "GCC: (crosstool-NG linaro-1.13.1-4.7-2012.10-20121022 - Linaro GCC 2012.10) 4.7.3 20121001 (prerelease)"
|
|
.section .note.GNU-stack,"",%progbits
|