Update patch files and gyp from r235188.
git-svn-id: http://src.chromium.org/svn/trunk/deps/third_party/openssl@235201 4ff67af0-8c30-449e-8e8b-ad334ec8d88c
diff --git a/openssl.gypi b/openssl.gypi
index 3d914c3..6280456 100644
--- a/openssl.gypi
+++ b/openssl.gypi
@@ -609,7 +609,6 @@
'openssl/crypto/des/des_enc.c',
'openssl/crypto/des/fcrypt_b.c',
'openssl/crypto/mem_clr.c',
- 'openssl/crypto/poly1305/poly1305.c',
],
'openssl_x86_sources': [
'openssl/crypto/aes/asm/aes-586.S',
@@ -625,7 +624,6 @@
'openssl/crypto/des/asm/des-586.S',
'openssl/crypto/md5/asm/md5-586.S',
'openssl/crypto/modes/asm/ghash-x86.S',
- 'openssl/crypto/poly1305/poly1305_vec.c',
'openssl/crypto/sha/asm/sha1-586.S',
'openssl/crypto/sha/asm/sha256-586.S',
'openssl/crypto/sha/asm/sha512-586.S',
diff --git a/openssl/Crypto-config.mk b/openssl/Crypto-config.mk
index 35f8365..aab86d3 100644
--- a/openssl/Crypto-config.mk
+++ b/openssl/Crypto-config.mk
@@ -598,7 +598,6 @@
crypto/des/asm/des-586.S \
crypto/md5/asm/md5-586.S \
crypto/modes/asm/ghash-x86.S \
- crypto/poly1305/poly1305_vec.c \
crypto/sha/asm/sha1-586.S \
crypto/sha/asm/sha256-586.S \
crypto/sha/asm/sha512-586.S \
@@ -613,7 +612,6 @@
crypto/des/des_enc.c \
crypto/des/fcrypt_b.c \
crypto/mem_clr.c \
- crypto/poly1305/poly1305.c \
x86_64_c_flags := \
-DAES_ASM \
diff --git a/patches.chromium/0007-chacha.patch b/patches.chromium/0007-chacha.patch
index dbfa73d..b42492f 100644
--- a/patches.chromium/0007-chacha.patch
+++ b/patches.chromium/0007-chacha.patch
@@ -44,7 +44,7 @@
crypto/sha/asm/sha1-armv4-large.S \
crypto/sha/asm/sha256-armv4.S \
crypto/sha/asm/sha512-armv4.S \
-@@ -821,10 +828,12 @@
+@@ -821,10 +828,11 @@
crypto/bn/asm/co-586.S \
crypto/bn/asm/x86-gf2m.S \
crypto/bn/asm/x86-mont.S \
@@ -53,11 +53,10 @@
crypto/des/asm/des-586.S \
crypto/md5/asm/md5-586.S \
crypto/modes/asm/ghash-x86.S \
-+crypto/poly1305/poly1305_vec.c \
crypto/sha/asm/sha1-586.S \
crypto/sha/asm/sha256-586.S \
crypto/sha/asm/sha512-586.S \
-@@ -836,9 +845,11 @@
+@@ -836,9 +845,10 @@
crypto/aes/aes_cbc.c \
crypto/bf/bf_enc.c \
crypto/bn/bn_asm.c \
@@ -65,7 +64,6 @@
crypto/des/des_enc.c \
crypto/des/fcrypt_b.c \
crypto/mem_clr.c \
-+crypto/poly1305/poly1305.c \
"
OPENSSL_CRYPTO_SOURCES_x86_64="\
@@ -1620,7 +1618,7 @@
diff -burN android-openssl-lhash2/patches/chacha20poly1305.patch android-openssl/patches/chacha20poly1305.patch
--- android-openssl-lhash2/patches/chacha20poly1305.patch 1969-12-31 19:00:00.000000000 -0500
+++ android-openssl/patches/chacha20poly1305.patch 2013-11-05 15:15:28.454480948 -0500
-@@ -0,0 +1,5726 @@
+@@ -0,0 +1,5725 @@
+From 2688f00904e4ffd647afcff69bb8fe6df8c5902b Mon Sep 17 00:00:00 2001
+From: Adam Langley <agl@chromium.org>
+Date: Mon, 9 Sep 2013 12:13:24 -0400
@@ -2491,7 +2489,7 @@
++ op[9] = REVW_BE(REVW_BE(ip[9]) ^ (x9 + kp[5]));
++ op[10] = REVW_BE(REVW_BE(ip[10]) ^ (x10 + kp[6]));
++ op[11] = REVW_BE(REVW_BE(ip[11]) ^ (x11 + kp[7]));
-++ op[12] = REVW_BE(REVW_BE(ip[12]) ^ (x12 + BPI*iters+(BPI-1)));
+++ op[12] = REVW_BE(REVW_BE(ip[12]) ^ (x12 + counter+BPI*iters+(BPI-1)));
++ op[13] = REVW_BE(REVW_BE(ip[13]) ^ (x13));
++ op[14] = REVW_BE(REVW_BE(ip[14]) ^ (x14 + np[0]));
++ op[15] = REVW_BE(REVW_BE(ip[15]) ^ (x15 + np[1]));
@@ -2848,7 +2846,7 @@
+index 0000000..1c0c0fb
+--- /dev/null
++++ b/crypto/evp/e_chacha20poly1305.c
-+@@ -0,0 +1,261 @@
++@@ -0,0 +1,267 @@
++/* ====================================================================
++ * Copyright (c) 2013 The OpenSSL Project. All rights reserved.
++ *
@@ -2971,6 +2969,12 @@
++ CRYPTO_poly1305_update(poly1305, length_bytes, sizeof(length_bytes));
++}
++
+++#if __arm__
+++#define ALIGNED __attribute__((aligned(16)))
+++#else
+++#define ALIGNED
+++#endif
+++
++static ssize_t aead_chacha20_poly1305_seal(const EVP_AEAD_CTX *ctx,
++ unsigned char *out, size_t max_out_len,
++ const unsigned char *nonce, size_t nonce_len,
@@ -2978,7 +2982,7 @@
++ const unsigned char *ad, size_t ad_len)
++ {
++ const struct aead_chacha20_poly1305_ctx *c20_ctx = ctx->aead_state;
-++ unsigned char poly1305_key[32];
+++ unsigned char poly1305_key[32] ALIGNED;
++ poly1305_state poly1305;
++ const uint64_t in_len_64 = in_len;
++
@@ -3035,7 +3039,7 @@
++ {
++ const struct aead_chacha20_poly1305_ctx *c20_ctx = ctx->aead_state;
++ unsigned char mac[POLY1305_TAG_LEN];
-++ unsigned char poly1305_key[32];
+++ unsigned char poly1305_key[32] ALIGNED;
++ size_t out_len;
++ poly1305_state poly1305;
++ const uint64_t in_len_64 = in_len;
@@ -3663,7 +3667,7 @@
+index 0000000..adcef35
+--- /dev/null
++++ b/crypto/poly1305/poly1305_arm.c
-+@@ -0,0 +1,335 @@
++@@ -0,0 +1,327 @@
++/* ====================================================================
++ * Copyright (c) 2011-2013 The OpenSSL Project. All rights reserved.
++ *
@@ -3901,8 +3905,6 @@
++ fe1305x2 *const c = h + 1;
++ fe1305x2 *const precomp = c + 1;
++ unsigned int i;
-++ unsigned char data[sizeof(fe1305x2) + 16];
-++ fe1305x2 *const r2r = (fe1305x2 *) (data + (15 & (-(int) data)));
++
++ if (st->buf_used)
++ {
@@ -3915,20 +3917,12 @@
++ in_len -= todo;
++ in += todo;
++
-++ if (st->buf_used == sizeof(st->buf))
+++ if (st->buf_used == sizeof(st->buf) && in_len)
++ {
+++ addmulmod(h,h,precomp,&zero);
++ fe1305x2_frombytearray(c, st->buf, sizeof(st->buf));
-++ r2r->v[0] = precomp->v[0];
-++ r2r->v[2] = precomp->v[2];
-++ r2r->v[4] = precomp->v[4];
-++ r2r->v[6] = precomp->v[6];
-++ r2r->v[8] = precomp->v[8];
-++ r2r->v[1] = r->v[1];
-++ r2r->v[3] = r->v[3];
-++ r2r->v[5] = r->v[5];
-++ r2r->v[7] = r->v[7];
-++ r2r->v[9] = r->v[9];
-++ addmulmod(h,h,r2r,c);
+++ for (i = 0; i < 10; i++)
+++ h->v[i] += c->v[i];
++ st->buf_used = 0;
++ }
++ }
@@ -3936,7 +3930,7 @@
++ while (in_len > 32)
++ {
++ unsigned int tlen = 1048576;
-++ if (in_len < 1048576)
+++ if (in_len < tlen)
++ tlen = in_len;
++ tlen -= blocks(h, precomp, in, tlen);
++ in_len -= tlen;
@@ -3959,6 +3953,8 @@
++ fe1305x2 *const c = h + 1;
++ fe1305x2 *const precomp = c + 1;
++
+++ addmulmod(h,h,precomp,&zero);
+++
++ if (st->buf_used > 16)
++ {
++ fe1305x2_frombytearray(c, st->buf, st->buf_used);
@@ -7351,7 +7347,7 @@
diff -burN android-openssl-lhash2/patches/neon_runtime.patch android-openssl/patches/neon_runtime.patch
--- android-openssl-lhash2/patches/neon_runtime.patch 1969-12-31 19:00:00.000000000 -0500
+++ android-openssl/patches/neon_runtime.patch 2013-11-05 16:51:28.668287449 -0500
-@@ -0,0 +1,1118 @@
+@@ -0,0 +1,1123 @@
+From aea47606333cfd3e7a09cab3e42e488c79a416af Mon Sep 17 00:00:00 2001
+From: Adam Langley <agl@chromium.org>
+Date: Tue, 5 Nov 2013 13:10:11 -0500
@@ -7415,12 +7411,14 @@
+ /* chacha_core performs |num_rounds| rounds of ChaCha20 on the input words in
+ * |input| and writes the 64 output bytes to |output|. */
+ static void chacha_core(unsigned char output[64], const uint32_t input[16],
-+@@ -124,6 +134,14 @@ void CRYPTO_chacha_20(unsigned char *out,
++@@ -124,6 +134,16 @@ void CRYPTO_chacha_20(unsigned char *out,
+ unsigned char buf[64];
+ size_t todo, i;
+
++#if __arm__
-++ if (CRYPTO_is_NEON_capable())
+++ if (CRYPTO_is_NEON_capable() &&
+++ ((intptr_t)in & 15) == 0 &&
+++ ((intptr_t)out & 15) == 0)
++ {
++ CRYPTO_chacha_20_neon(out, in, in_len, key, nonce, counter);
++ return;
@@ -7454,7 +7452,7 @@
+index 0000000..24a5050
+--- /dev/null
++++ b/crypto/chacha/chacha_vec_arm.S
-+@@ -0,0 +1,860 @@
++@@ -0,0 +1,863 @@
++# This file contains a pre-compiled version of chacha_vec.c for ARM. This is
++# needed to support switching on NEON code at runtime. If the whole of OpenSSL
++# were to be compiled with the needed flags to build chacha_vec.c, then it
@@ -7462,7 +7460,7 @@
++#
++# This file was generated by:
++#
-++# /opt/gcc-linaro-arm-linux-gnueabihf-4.7-2012.10-20121022_linux/bin/arm-linux-gnueabihf-gcc -O3 -mcpu=cortex-a8 -mfpu=neon -S chacha_vec.c -I ../../include -fpic
+++# /opt/gcc-linaro-arm-linux-gnueabihf-4.7-2012.10-20121022_linux/bin/arm-linux-gnueabihf-gcc -O3 -mcpu=cortex-a8 -mfpu=neon -S chacha_vec.c -I ../../include -fpic -o chacha_vec_arm.S
++#
++# And then EABI attribute 28 was set to zero to allow linking with soft-float
++# code.
@@ -7490,130 +7488,133 @@
++ .thumb_func
++ .type CRYPTO_chacha_20_neon, %function
++CRYPTO_chacha_20_neon:
-++ @ args = 8, pretend = 0, frame = 296
+++ @ args = 8, pretend = 0, frame = 304
++ @ frame_needed = 1, uses_anonymous_args = 0
++ @ link register save eliminated.
++ push {r4, r5, r6, r7, r8, r9, sl, fp}
++ fstmfdd sp!, {d8, d9, d10, d11, d12, d13, d14, d15}
-++ sub sp, sp, #296
+++ sub sp, sp, #304
++ add r7, sp, #0
++ movw ip, #43691
++ movt ip, 43690
-++ str r2, [r7, #192]
+++ str r2, [r7, #196]
++ sub sp, sp, #96
-++ ldr r4, [r7, #192]
-++ ldr r6, [r7, #392]
+++ ldr r4, [r7, #196]
+++ ldr r6, [r7, #400]
++ ldr r2, .L38+16
++ umull r4, ip, ip, r4
++ ldr r6, [r6, #0]
-++ ldr r8, [r7, #392]
+++ ldr r8, [r7, #400]
++.LPIC24:
++ add r2, pc
++ add r4, sp, #15
-++ str r3, [r7, #236]
-++ str r6, [r7, #172]
+++ str r3, [r7, #244]
+++ str r6, [r7, #176]
++ bic r4, r4, #15
-++ str r0, [r7, #184]
-++ str r4, [r7, #196]
+++ str r0, [r7, #188]
+++ str r4, [r7, #200]
++ lsrs ip, ip, #7
-++ str r1, [r7, #180]
+++ str r1, [r7, #184]
++ ldmia r2, {r0, r1, r2, r3}
++ ldr r4, [r8, #4]
-++ ldr r5, [r7, #236]
+++ ldr r5, [r7, #244]
++ vld1.64 {d24-d25}, [r5:64]
++ vldr d26, [r5, #16]
++ vldr d27, [r5, #24]
-++ ldr fp, [r7, #196]
-++ ldr r8, [r7, #396]
-++ ldr r5, [r7, #172]
-++ add r6, fp, #64
-++ str r4, [r7, #292]
+++ ldr r9, [r7, #200]
+++ ldr r8, [r7, #404]
+++ ldr r5, [r7, #176]
+++ add r6, r9, #64
+++ str r4, [r7, #300]
++ mov r4, #0
-++ str r8, [r7, #280]
-++ str r5, [r7, #288]
-++ str r4, [r7, #284]
+++ str r8, [r7, #288]
+++ str r5, [r7, #296]
+++ str r4, [r7, #292]
++ stmia r6, {r0, r1, r2, r3}
-++ vldr d22, [fp, #64]
-++ vldr d23, [fp, #72]
-++ vldr d20, [r7, #280]
-++ vldr d21, [r7, #288]
-++ str ip, [r7, #188]
+++ vldr d22, [r9, #64]
+++ vldr d23, [r9, #72]
+++ vldr d20, [r7, #288]
+++ vldr d21, [r7, #296]
+++ str ip, [r7, #192]
++ beq .L20
++ lsl r6, ip, #1
-++ ldr r1, [fp, #68]
+++ ldr r1, [r9, #68]
++ add r3, r6, ip
-++ str r6, [r7, #176]
-++ ldr r2, [fp, #72]
+++ str r6, [r7, #180]
+++ ldr r2, [r9, #72]
++ add r8, r8, #2
-++ ldr r5, [fp, #76]
+++ ldr r5, [r9, #76]
++ vldr d18, .L38
++ vldr d19, .L38+8
-++ str r4, [r7, #232]
-++ ldr r6, [r7, #180]
-++ ldr r4, [r7, #184]
-++ str r0, [r7, #220]
-++ str r1, [r7, #216]
-++ str r8, [r7, #200]
-++ str r2, [r7, #212]
-++ str r3, [r7, #204]
-++ str r5, [r7, #208]
-++ str r6, [r7, #244]
++ str r4, [r7, #240]
+++ ldr r6, [r7, #184]
+++ ldr r4, [r7, #188]
+++ str r0, [r7, #224]
+++ str r1, [r7, #220]
+++ str r8, [r7, #208]
+++ str r2, [r7, #216]
+++ str r3, [r7, #204]
+++ str r5, [r7, #212]
+++ str r6, [r7, #252]
+++ str r4, [r7, #248]
++.L4:
-++ ldr r6, [r7, #236]
+++ ldr r2, [r7, #244]
+++ add r9, r7, #216
+++ ldr r3, [r7, #244]
++ vadd.i32 q8, q10, q9
-++ ldr r5, [r7, #236]
+++ ldr r6, [r7, #208]
++ vmov q15, q13 @ v4si
-++ ldr r8, [r7, #232]
+++ ldr r5, [r7, #240]
++ vmov q3, q12 @ v4si
-++ ldr r6, [r6, #4]
+++ ldr r4, [r7, #244]
++ vmov q2, q11 @ v4si
-++ ldr fp, [r7, #200]
+++ adds r5, r5, r6
+++ ldr r2, [r2, #8]
+++ ldr r6, [r7, #400]
++ vmov q5, q10 @ v4si
-++ ldr r4, [r7, #236]
+++ ldr r3, [r3, #12]
++ vmov q1, q13 @ v4si
-++ add ip, r8, fp
-++ ldr r5, [r5, #0]
-++ ldr r0, [r7, #236]
-++ add r8, r7, #208
-++ ldr r1, [r7, #236]
+++ ldr r0, [r7, #244]
++ vmov q0, q12 @ v4si
-++ str r6, [r7, #260]
+++ ldr r1, [r7, #244]
++ vmov q4, q11 @ v4si
-++ ldr r6, [r7, #392]
-++ ldmia r8, {r8, r9, sl, fp}
-++ ldr r0, [r0, #8]
-++ ldr r1, [r1, #12]
-++ str r5, [r7, #224]
+++ ldmia r9, {r9, sl, fp}
+++ str r5, [r7, #228]
++ ldr r5, [r4, #24]
-++ ldr r3, [r4, #28]
+++ ldr r0, [r0, #0]
+++ ldr r1, [r1, #4]
+++ str r2, [r7, #264]
+++ str r3, [r7, #236]
++ ldr r2, [r6, #4]
-++ str r0, [r7, #256]
-++ str r1, [r7, #228]
-++ str r5, [r7, #272]
+++ ldr r3, [r4, #28]
+++ str r5, [r7, #280]
++ ldr r5, [r6, #0]
++ movs r6, #0
+++ ldr ip, [r7, #228]
+++ ldr r8, [r7, #212]
+++ str r0, [r7, #232]
+++ str r1, [r7, #268]
++ ldr r0, [r4, #16]
++ ldr r1, [r4, #20]
++ movs r4, #10
-++ str r2, [r7, #20]
-++ str r3, [r7, #276]
-++ str r9, [r7, #268]
+++ str r2, [r7, #24]
+++ str r3, [r7, #284]
+++ str r4, [r7, #256]
+++ ldr r2, [r7, #264]
+++ str r9, [r7, #276]
++ mov r9, r6
-++ str r4, [r7, #248]
-++ ldr r2, [r7, #256]
-++ ldr r3, [r7, #228]
-++ str r8, [r7, #252]
+++ ldr r6, [r7, #280]
+++ str r8, [r7, #260]
++ mov r8, sl
-++ ldr r6, [r7, #272]
+++ str r1, [r7, #272]
++ mov sl, ip
-++ str r1, [r7, #264]
-++ ldr ip, [r7, #20]
-++ str r6, [r7, #256]
+++ str r6, [r7, #264]
++ mov r6, r5
-++ ldr r1, [r7, #260]
+++ ldr r3, [r7, #236]
++ mov r5, r0
-++ ldr r0, [r7, #224]
+++ ldr ip, [r7, #24]
+++ ldr r1, [r7, #268]
+++ ldr r0, [r7, #232]
++ b .L39
++.L40:
++ .align 3
@@ -7628,445 +7629,445 @@
++ vadd.i32 q4, q4, q0
++ add r8, r8, r1
++ vadd.i32 q2, q2, q3
-++ str r8, [r7, #260]
+++ str r8, [r7, #268]
++ veor q5, q5, q4
-++ ldr r8, [r7, #268]
+++ ldr r8, [r7, #276]
++ veor q8, q8, q2
++ add fp, fp, r0
-++ str fp, [r7, #272]
+++ str fp, [r7, #280]
++ add r8, r8, r2
++ vrev32.16 q5, q5
-++ str r8, [r7, #268]
+++ str r8, [r7, #276]
++ vrev32.16 q8, q8
++ vadd.i32 q1, q1, q5
++ vadd.i32 q15, q15, q8
-++ ldr r8, [r7, #272]
+++ ldr r8, [r7, #280]
++ veor q0, q1, q0
-++ ldr r4, [r7, #252]
+++ ldr r4, [r7, #260]
++ veor q3, q15, q3
++ eor sl, sl, r8
-++ ldr r8, [r7, #268]
+++ ldr r8, [r7, #276]
++ add fp, r4, r3
++ vshl.i32 q7, q0, #12
-++ ldr r4, [r7, #260]
+++ ldr r4, [r7, #268]
++ vshl.i32 q6, q3, #12
++ eor r6, r6, r8
++ eor r9, r9, r4
-++ ldr r4, [r7, #264]
+++ ldr r4, [r7, #272]
++ vsri.32 q7, q0, #20
++ ror r8, r6, #16
-++ ldr r6, [r7, #256]
+++ ldr r6, [r7, #264]
++ eor ip, ip, fp
++ vsri.32 q6, q3, #20
++ ror sl, sl, #16
++ ror r9, r9, #16
++ add r5, r5, sl
++ vadd.i32 q4, q4, q7
-++ str r5, [r7, #228]
+++ str r5, [r7, #236]
++ vadd.i32 q2, q2, q6
++ add r5, r4, r9
++ add r4, r6, r8
-++ ldr r6, [r7, #276]
+++ ldr r6, [r7, #284]
++ ror ip, ip, #16
++ veor q5, q4, q5
++ veor q8, q2, q8
++ add r6, r6, ip
-++ str r6, [r7, #256]
+++ str r6, [r7, #264]
++ eors r1, r1, r5
-++ ldr r6, [r7, #228]
+++ ldr r6, [r7, #236]
++ vshl.i32 q3, q5, #8
++ vshl.i32 q14, q8, #8
++ eors r2, r2, r4
++ eors r0, r0, r6
-++ ldr r6, [r7, #256]
+++ ldr r6, [r7, #264]
++ vsri.32 q3, q5, #24
++ ror r1, r1, #20
++ eors r3, r3, r6
-++ ldr r6, [r7, #272]
+++ ldr r6, [r7, #280]
++ ror r0, r0, #20
++ vsri.32 q14, q8, #24
++ adds r6, r0, r6
-++ str r6, [r7, #276]
-++ ldr r6, [r7, #260]
+++ str r6, [r7, #284]
+++ ldr r6, [r7, #268]
++ vadd.i32 q1, q1, q3
++ vadd.i32 q15, q15, q14
++ ror r2, r2, #20
++ adds r6, r1, r6
-++ str r6, [r7, #252]
-++ ldr r6, [r7, #268]
+++ str r6, [r7, #260]
+++ ldr r6, [r7, #276]
++ veor q6, q15, q6
++ veor q7, q1, q7
++ ror r3, r3, #20
++ adds r6, r2, r6
-++ str r6, [r7, #272]
-++ ldr r6, [r7, #276]
+++ str r6, [r7, #280]
+++ ldr r6, [r7, #284]
++ vshl.i32 q0, q6, #7
++ vshl.i32 q5, q7, #7
++ add fp, r3, fp
++ eor sl, r6, sl
-++ ldr r6, [r7, #252]
+++ ldr r6, [r7, #260]
++ eor ip, fp, ip
++ vsri.32 q0, q6, #25
++ eor r9, r6, r9
-++ ldr r6, [r7, #272]
+++ ldr r6, [r7, #280]
++ ror sl, sl, #24
++ vsri.32 q5, q7, #25
++ eor r8, r6, r8
-++ ldr r6, [r7, #228]
+++ ldr r6, [r7, #236]
++ ror r9, r9, #24
++ ror ip, ip, #24
++ add r6, sl, r6
-++ str r6, [r7, #268]
-++ ldr r6, [r7, #256]
+++ str r6, [r7, #276]
+++ ldr r6, [r7, #264]
++ add r5, r9, r5
-++ str r5, [r7, #264]
+++ str r5, [r7, #272]
++ vext.32 q5, q5, q5, #1
++ add r5, ip, r6
-++ ldr r6, [r7, #268]
+++ ldr r6, [r7, #276]
++ vext.32 q0, q0, q0, #1
++ vadd.i32 q4, q4, q5
++ eors r0, r0, r6
-++ ldr r6, [r7, #264]
+++ ldr r6, [r7, #272]
++ vadd.i32 q2, q2, q0
++ vext.32 q3, q3, q3, #3
++ ror r8, r8, #24
++ eors r1, r1, r6
++ vext.32 q14, q14, q14, #3
++ add r4, r8, r4
-++ ldr r6, [r7, #276]
+++ ldr r6, [r7, #284]
++ veor q3, q4, q3
++ veor q14, q2, q14
++ eors r2, r2, r4
++ ror r1, r1, #25
++ vext.32 q1, q1, q1, #2
++ adds r6, r1, r6
-++ str r6, [r7, #276]
+++ str r6, [r7, #284]
++ vext.32 q15, q15, q15, #2
-++ ldr r6, [r7, #252]
+++ ldr r6, [r7, #260]
++ eors r3, r3, r5
++ ror r2, r2, #25
++ vrev32.16 q8, q14
++ adds r6, r2, r6
++ vrev32.16 q3, q3
-++ str r6, [r7, #260]
+++ str r6, [r7, #268]
++ vadd.i32 q1, q1, q3
-++ ldr r6, [r7, #272]
+++ ldr r6, [r7, #280]
++ vadd.i32 q15, q15, q8
++ ror r3, r3, #25
++ veor q5, q1, q5
++ adds r6, r3, r6
++ veor q0, q15, q0
-++ str r6, [r7, #256]
-++ ldr r6, [r7, #260]
+++ str r6, [r7, #264]
+++ ldr r6, [r7, #268]
++ ror r0, r0, #25
++ add fp, r0, fp
++ vshl.i32 q6, q5, #12
++ eor sl, r6, sl
-++ ldr r6, [r7, #276]
+++ ldr r6, [r7, #284]
++ vshl.i32 q14, q0, #12
++ eor r8, fp, r8
++ eor ip, r6, ip
-++ ldr r6, [r7, #256]
+++ ldr r6, [r7, #264]
++ vsri.32 q6, q5, #20
++ ror sl, sl, #16
++ eor r9, r6, r9
++ ror r6, r8, #16
++ vsri.32 q14, q0, #20
-++ ldr r8, [r7, #264]
+++ ldr r8, [r7, #272]
++ ror ip, ip, #16
++ add r5, sl, r5
++ add r8, r6, r8
++ add r4, ip, r4
-++ str r4, [r7, #228]
+++ str r4, [r7, #236]
++ eor r0, r8, r0
-++ str r5, [r7, #272]
+++ str r5, [r7, #280]
++ vadd.i32 q4, q4, q6
-++ ldr r5, [r7, #228]
+++ ldr r5, [r7, #236]
++ vadd.i32 q2, q2, q14
-++ ldr r4, [r7, #268]
+++ ldr r4, [r7, #276]
++ ror r0, r0, #20
++ veor q3, q4, q3
++ eors r1, r1, r5
++ veor q0, q2, q8
-++ str r8, [r7, #264]
-++ str r0, [r7, #20]
+++ str r8, [r7, #272]
+++ str r0, [r7, #24]
++ add fp, r0, fp
-++ ldr r8, [r7, #272]
+++ ldr r8, [r7, #280]
++ ror r9, r9, #16
-++ ldr r0, [r7, #276]
+++ ldr r0, [r7, #284]
++ add r4, r9, r4
-++ str fp, [r7, #252]
+++ str fp, [r7, #260]
++ ror r1, r1, #20
++ add fp, r1, r0
++ eor r2, r8, r2
-++ ldr r0, [r7, #252]
+++ ldr r0, [r7, #260]
++ eors r3, r3, r4
++ vshl.i32 q5, q3, #8
-++ str r4, [r7, #224]
+++ str r4, [r7, #232]
++ vshl.i32 q8, q0, #8
-++ ldr r4, [r7, #260]
-++ ldr r5, [r7, #256]
+++ ldr r4, [r7, #268]
+++ ldr r5, [r7, #264]
++ ror r2, r2, #20
++ ror r3, r3, #20
++ eors r6, r6, r0
++ adds r5, r3, r5
++ add r8, r2, r4
++ vsri.32 q5, q3, #24
-++ ldr r4, [r7, #264]
+++ ldr r4, [r7, #272]
++ eor r9, r5, r9
++ eor ip, fp, ip
++ vsri.32 q8, q0, #24
++ eor sl, r8, sl
++ ror r6, r6, #24
-++ ldr r0, [r7, #272]
-++ str r5, [r7, #268]
+++ ldr r0, [r7, #280]
+++ str r5, [r7, #276]
++ adds r4, r6, r4
-++ ldr r5, [r7, #228]
+++ ldr r5, [r7, #236]
++ vadd.i32 q1, q1, q5
-++ str r4, [r7, #264]
+++ str r4, [r7, #272]
++ vadd.i32 q15, q15, q8
-++ ldr r4, [r7, #224]
+++ ldr r4, [r7, #232]
++ ror ip, ip, #24
++ ror sl, sl, #24
++ ror r9, r9, #24
++ add r5, ip, r5
++ add r0, sl, r0
-++ str r5, [r7, #256]
+++ str r5, [r7, #264]
++ add r5, r9, r4
-++ str r0, [r7, #276]
+++ str r0, [r7, #284]
++ veor q6, q1, q6
-++ ldr r4, [r7, #20]
+++ ldr r4, [r7, #24]
++ veor q14, q15, q14
-++ ldr r0, [r7, #264]
+++ ldr r0, [r7, #272]
++ eors r3, r3, r5
++ vshl.i32 q0, q6, #7
++ vext.32 q1, q1, q1, #2
++ eors r0, r0, r4
-++ ldr r4, [r7, #276]
-++ str r0, [r7, #272]
+++ ldr r4, [r7, #284]
+++ str r0, [r7, #280]
++ vshl.i32 q3, q14, #7
++ eors r2, r2, r4
-++ ldr r4, [r7, #272]
-++ ldr r0, [r7, #256]
+++ ldr r4, [r7, #280]
+++ ldr r0, [r7, #264]
++ vsri.32 q0, q6, #25
++ ror r2, r2, #25
++ ror r3, r3, #25
++ eors r1, r1, r0
++ vsri.32 q3, q14, #25
++ ror r0, r4, #25
-++ ldr r4, [r7, #248]
+++ ldr r4, [r7, #256]
++ ror r1, r1, #25
++ vext.32 q5, q5, q5, #1
++ subs r4, r4, #1
-++ str r4, [r7, #248]
+++ str r4, [r7, #256]
++ vext.32 q15, q15, q15, #2
++ vext.32 q8, q8, q8, #1
++ vext.32 q0, q0, q0, #3
++ vext.32 q3, q3, q3, #3
++ bne .L3
-++ ldr r4, [r7, #256]
-++ vadd.i32 q4, q11, q4
-++ str r2, [r7, #256]
+++ ldr r4, [r7, #264]
++ vadd.i32 q14, q10, q9
-++ ldr r2, [r7, #244]
-++ vld1.64 {d12-d13}, [r2:64]
-++ str r4, [r7, #272]
-++ veor q4, q4, q6
-++ ldr r4, [r7, #220]
+++ str r2, [r7, #264]
++ vadd.i32 q10, q10, q5
-++ ldr r2, [r7, #216]
-++ vadd.i32 q0, q12, q0
-++ add fp, fp, r4
-++ str ip, [r7, #20]
-++ ldr r4, [r7, #212]
+++ ldr r2, [r7, #252]
+++ vld1.64 {d12-d13}, [r2:64]
+++ ldr r2, [r7, #220]
+++ vadd.i32 q4, q11, q4
+++ str ip, [r7, #24]
++ mov ip, sl
-++ str r0, [r7, #224]
++ mov sl, r8
-++ mov r0, r5
-++ ldr r8, [r7, #252]
-++ mov r5, r6
+++ ldr r8, [r7, #260]
++ add sl, sl, r2
-++ mov r6, r9
-++ ldr r2, [r7, #208]
-++ ldr r9, [r7, #268]
-++ vadd.i32 q1, q13, q1
-++ vadd.i32 q2, q11, q2
-++ str r1, [r7, #260]
-++ add r9, r9, r4
-++ add r4, r8, r2
-++ ldr r8, [r7, #232]
-++ vadd.i32 q3, q12, q3
-++ vadd.i32 q15, q13, q15
-++ str r3, [r7, #228]
-++ add r2, r8, #2
-++ vadd.i32 q8, q14, q8
-++ add ip, r2, ip
+++ ldr r2, [r7, #212]
+++ str r4, [r7, #280]
+++ vadd.i32 q0, q12, q0
+++ ldr r4, [r7, #224]
+++ add r8, r8, r2
++ ldr r2, [r7, #240]
-++ vst1.64 {d8-d9}, [r2:64]
-++ ldr r2, [r7, #244]
-++ ldr r3, [r7, #276]
+++ vadd.i32 q1, q13, q1
+++ str r0, [r7, #232]
+++ add fp, fp, r4
+++ mov r0, r5
+++ ldr r4, [r7, #216]
+++ mov r5, r6
+++ mov r6, r9
+++ ldr r9, [r7, #276]
+++ adds r2, r2, #3
+++ str r2, [r7, #240]
+++ vadd.i32 q2, q11, q2
+++ ldr r2, [r7, #252]
+++ add r9, r9, r4
+++ vadd.i32 q3, q12, q3
+++ ldr r4, [r7, #228]
+++ vadd.i32 q15, q13, q15
+++ str r1, [r7, #268]
+++ vadd.i32 q8, q14, q8
+++ str r3, [r7, #236]
+++ veor q4, q4, q6
+++ ldr r3, [r7, #284]
+++ ldr r1, [r7, #272]
+++ add ip, r4, ip
+++ ldr r4, [r7, #248]
+++ vst1.64 {d8-d9}, [r4:64]
++ vldr d8, [r2, #16]
++ vldr d9, [r2, #24]
-++ ldr r1, [r7, #264]
++ veor q0, q0, q4
-++ add r8, r8, #3
-++ str r8, [r7, #232]
-++ ldr r8, [r7, #240]
-++ vstr d0, [r8, #16]
-++ vstr d1, [r8, #24]
+++ vstr d0, [r4, #16]
+++ vstr d1, [r4, #24]
++ vldr d0, [r2, #32]
++ vldr d1, [r2, #40]
++ veor q1, q1, q0
-++ vstr d2, [r8, #32]
-++ vstr d3, [r8, #40]
+++ vstr d2, [r4, #32]
+++ vstr d3, [r4, #40]
++ vldr d2, [r2, #48]
++ vldr d3, [r2, #56]
++ veor q10, q10, q1
-++ vstr d20, [r8, #48]
-++ vstr d21, [r8, #56]
+++ vstr d20, [r4, #48]
+++ vstr d21, [r4, #56]
++ vldr d8, [r2, #64]
++ vldr d9, [r2, #72]
++ veor q2, q2, q4
-++ vstr d4, [r8, #64]
-++ vstr d5, [r8, #72]
+++ vstr d4, [r4, #64]
+++ vstr d5, [r4, #72]
++ vldr d10, [r2, #80]
++ vldr d11, [r2, #88]
++ veor q3, q3, q5
-++ vstr d6, [r8, #80]
-++ vstr d7, [r8, #88]
+++ vstr d6, [r4, #80]
+++ vstr d7, [r4, #88]
++ vldr d12, [r2, #96]
++ vldr d13, [r2, #104]
++ veor q15, q15, q6
-++ vstr d30, [r8, #96]
-++ vstr d31, [r8, #104]
+++ vstr d30, [r4, #96]
+++ vstr d31, [r4, #104]
++ vldr d20, [r2, #112]
++ vldr d21, [r2, #120]
++ veor q8, q8, q10
-++ vstr d16, [r8, #112]
-++ vstr d17, [r8, #120]
-++ mov r8, r2
-++ ldr r2, [r2, #128]
+++ vstr d16, [r4, #112]
+++ vstr d17, [r4, #120]
+++ ldr r4, [r2, #128]
+++ ldr r2, [r7, #248]
++ vadd.i32 q10, q14, q9
-++ eor r2, fp, r2
-++ ldr fp, [r7, #240]
+++ eor r4, fp, r4
++ vadd.i32 q10, q10, q9
-++ str r2, [fp, #128]
-++ ldr r2, [r8, #132]
+++ str r4, [r2, #128]
+++ ldr r4, [r7, #252]
+++ ldr r2, [r4, #132]
++ eor r2, sl, r2
-++ str r2, [fp, #132]
-++ ldr r2, [r8, #136]
+++ ldr sl, [r7, #248]
+++ str r2, [sl, #132]
+++ ldr r2, [r4, #136]
++ eor r2, r9, r2
-++ str r2, [fp, #136]
-++ ldr r2, [r8, #140]
-++ eors r2, r2, r4
-++ str r2, [fp, #140]
-++ ldr r2, [r7, #236]
-++ ldr r4, [r8, #144]
-++ ldr r2, [r2, #0]
-++ str r4, [r7, #168]
-++ ldr r4, [r7, #224]
-++ add r8, r4, r2
-++ ldr r2, [r7, #168]
-++ ldr r4, [r7, #236]
-++ eor r8, r8, r2
+++ str r2, [sl, #136]
+++ ldr r2, [r4, #140]
+++ eor r2, r8, r2
+++ str r2, [sl, #140]
++ ldr r2, [r7, #244]
-++ str r8, [fp, #144]
+++ ldr r4, [r4, #144]
+++ ldr r2, [r2, #0]
+++ str r4, [r7, #44]
+++ ldr r4, [r7, #232]
+++ add r8, r4, r2
+++ ldr r2, [r7, #44]
+++ ldr r4, [r7, #244]
+++ eor r8, r8, r2
+++ ldr r2, [r7, #252]
+++ str r8, [sl, #144]
++ ldr r4, [r4, #4]
++ ldr r2, [r2, #148]
-++ str r2, [r7, #36]
-++ ldr r2, [r7, #260]
+++ str r2, [r7, #40]
+++ ldr r2, [r7, #268]
++ add r8, r2, r4
-++ ldr r4, [r7, #36]
-++ ldr r2, [r7, #236]
+++ ldr r4, [r7, #40]
+++ ldr r2, [r7, #244]
++ eor r8, r8, r4
-++ ldr r4, [r7, #244]
-++ str r8, [fp, #148]
+++ ldr r4, [r7, #252]
+++ str r8, [sl, #148]
++ ldr r2, [r2, #8]
++ ldr r4, [r4, #152]
-++ str r4, [r7, #32]
-++ ldr r4, [r7, #256]
+++ str r4, [r7, #36]
+++ ldr r4, [r7, #264]
++ add r8, r4, r2
-++ ldr r2, [r7, #32]
+++ ldr r2, [r7, #36]
++ eor r8, r8, r2
-++ str r8, [fp, #152]
-++ ldr r2, [r7, #244]
-++ ldr r4, [r7, #236]
+++ str r8, [sl, #152]
+++ ldr r2, [r7, #252]
+++ ldr r4, [r7, #244]
++ ldr r2, [r2, #156]
++ ldr r4, [r4, #12]
-++ str r2, [r7, #28]
-++ ldr r2, [r7, #228]
+++ str r2, [r7, #32]
+++ ldr r2, [r7, #236]
++ add r8, r2, r4
-++ ldr r4, [r7, #28]
-++ ldr r2, [r7, #244]
+++ ldr r4, [r7, #32]
+++ ldr r2, [r7, #252]
++ eor r8, r8, r4
-++ str r8, [fp, #156]
-++ ldr r8, [r7, #236]
+++ str r8, [sl, #156]
+++ ldr r8, [r7, #244]
++ ldr r2, [r2, #160]
++ ldr r4, [r8, #16]
++ adds r0, r0, r4
-++ ldr r4, [r7, #244]
+++ ldr r4, [r7, #252]
++ eors r0, r0, r2
-++ str r0, [fp, #160]
+++ str r0, [sl, #160]
++ ldr r0, [r8, #20]
++ ldr r2, [r4, #164]
++ adds r1, r1, r0
-++ ldr r0, [r7, #272]
+++ ldr r0, [r7, #280]
++ eors r1, r1, r2
-++ str r1, [fp, #164]
+++ str r1, [sl, #164]
++ ldr r2, [r8, #24]
++ ldr r1, [r4, #168]
++ adds r2, r0, r2
++ eors r2, r2, r1
-++ str r2, [fp, #168]
+++ str r2, [sl, #168]
++ ldr r1, [r8, #28]
++ ldr r2, [r4, #172]
++ adds r3, r3, r1
++ eors r3, r3, r2
-++ str r3, [fp, #172]
+++ str r3, [sl, #172]
++ ldr r3, [r4, #176]
++ eor r3, ip, r3
-++ str r3, [fp, #176]
+++ str r3, [sl, #176]
++ ldr r3, [r4, #180]
-++ ldr r4, [r7, #392]
+++ ldr r4, [r7, #400]
++ eors r6, r6, r3
-++ str r6, [fp, #180]
-++ ldr r6, [r7, #244]
+++ str r6, [sl, #180]
+++ ldr r6, [r7, #252]
++ ldr r2, [r4, #0]
++ ldr r3, [r6, #184]
++ adds r5, r5, r2
++ eors r5, r5, r3
-++ str r5, [fp, #184]
+++ str r5, [sl, #184]
++ ldr r2, [r6, #188]
++ adds r6, r6, #192
++ ldr r3, [r4, #4]
-++ str r6, [r7, #244]
-++ ldr r0, [r7, #20]
-++ ldr r1, [r7, #232]
+++ str r6, [r7, #252]
+++ ldr r0, [r7, #24]
+++ ldr r1, [r7, #240]
++ adds r4, r0, r3
++ eors r4, r4, r2
++ ldr r2, [r7, #204]
-++ str r4, [fp, #188]
-++ add fp, fp, #192
+++ str r4, [sl, #188]
+++ add sl, sl, #192
++ cmp r1, r2
-++ str fp, [r7, #240]
+++ str sl, [r7, #248]
++ bne .L4
-++ ldr r4, [r7, #188]
-++ ldr r3, [r7, #176]
-++ ldr r6, [r7, #184]
+++ ldr r4, [r7, #192]
+++ ldr r3, [r7, #180]
+++ ldr r6, [r7, #188]
++ adds r5, r3, r4
-++ ldr r8, [r7, #180]
+++ ldr r8, [r7, #184]
++ lsls r5, r5, #6
++ adds r4, r6, r5
++ add r5, r8, r5
++.L2:
-++ ldr fp, [r7, #192]
+++ ldr r9, [r7, #196]
++ movw r3, #43691
++ movt r3, 43690
-++ ldr r6, [r7, #192]
-++ umull fp, r3, r3, fp
+++ ldr sl, [r7, #196]
+++ umull r9, r3, r3, r9
++ lsrs r3, r3, #7
++ add r3, r3, r3, lsl #1
-++ sub r3, r6, r3, lsl #6
+++ sub r3, sl, r3, lsl #6
++ lsrs r6, r3, #6
++ beq .L5
++ add r1, r5, #16
@@ -8150,7 +8151,7 @@
++ adds r4, r4, r6
++ adds r5, r5, r6
++.L5:
-++ ldr r6, [r7, #192]
+++ ldr r6, [r7, #196]
++ ands ip, r6, #63
++ beq .L1
++ vmov q8, q10 @ v4si
@@ -8200,11 +8201,11 @@
++ cmp ip, #15
++ vadd.i32 q11, q11, q15
++ bhi .L37
-++ ldr fp, [r7, #196]
-++ vst1.64 {d22-d23}, [fp:128]
+++ ldr r9, [r7, #200]
+++ vst1.64 {d22-d23}, [r9:128]
++.L14:
-++ ldr r6, [r7, #192]
-++ and r3, r6, #48
+++ ldr sl, [r7, #196]
+++ and r3, sl, #48
++ cmp ip, r3
++ bls .L1
++ adds r0, r5, r3
@@ -8229,7 +8230,7 @@
++ orreq r2, r2, #1
++ lsl sl, r8, #4
++ cbnz r2, .L35
-++ ldr fp, [r7, #196]
+++ ldr fp, [r7, #200]
++ add r6, fp, r3
++.L17:
++ vld1.8 {q8}, [r0]!
@@ -8243,7 +8244,7 @@
++ add r3, r3, sl
++ beq .L1
++.L35:
-++ ldr r0, [r7, #196]
+++ ldr r0, [r7, #200]
++.L25:
++ ldrb r2, [r5, r3] @ zero_extendqisi2
++ ldrb r1, [r3, r0] @ zero_extendqisi2
@@ -8253,7 +8254,7 @@
++ cmp ip, r3
++ bhi .L25
++.L1:
-++ add r7, r7, #296
+++ add r7, r7, #304
++ mov sp, r7
++ fldmfdd sp!, {d8, d9, d10, d11, d12, d13, d14, d15}
++ pop {r4, r5, r6, r7, r8, r9, sl, fp}
@@ -8276,7 +8277,7 @@
++ vadd.i32 q8, q8, q10
++ vldr d0, [r5, #32]
++ vldr d1, [r5, #40]
-++ ldr r6, [r7, #196]
+++ ldr r6, [r7, #200]
++ vstr d16, [r6, #48]
++ vstr d17, [r6, #56]
++ veor q8, q13, q0
@@ -8284,16 +8285,16 @@
++ vstr d17, [r4, #40]
++ b .L14
++.L12:
-++ ldr r8, [r7, #196]
+++ ldr r8, [r7, #200]
++ vstr d18, [r8, #16]
++ vstr d19, [r8, #24]
++ b .L14
++.L20:
-++ ldr r5, [r7, #180]
-++ ldr r4, [r7, #184]
+++ ldr r5, [r7, #184]
+++ ldr r4, [r7, #188]
++ b .L2
++.L13:
-++ ldr r6, [r7, #196]
+++ ldr r6, [r7, #200]
++ vstr d26, [r6, #32]
++ vstr d27, [r6, #40]
++ b .L14