2014-06-20 20:00:00 +01:00
|
|
|
/* ====================================================================
|
|
|
|
* Copyright (c) 2001-2011 The OpenSSL Project. All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
*
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
*
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in
|
|
|
|
* the documentation and/or other materials provided with the
|
|
|
|
* distribution.
|
|
|
|
*
|
|
|
|
* 3. All advertising materials mentioning features or use of this
|
|
|
|
* software must display the following acknowledgment:
|
|
|
|
* "This product includes software developed by the OpenSSL Project
|
|
|
|
* for use in the OpenSSL Toolkit. (http://www.openssl.org/)"
|
|
|
|
*
|
|
|
|
* 4. The names "OpenSSL Toolkit" and "OpenSSL Project" must not be used to
|
|
|
|
* endorse or promote products derived from this software without
|
|
|
|
* prior written permission. For written permission, please contact
|
|
|
|
* openssl-core@openssl.org.
|
|
|
|
*
|
|
|
|
* 5. Products derived from this software may not be called "OpenSSL"
|
|
|
|
* nor may "OpenSSL" appear in their names without prior written
|
|
|
|
* permission of the OpenSSL Project.
|
|
|
|
*
|
|
|
|
* 6. Redistributions of any form whatsoever must retain the following
|
|
|
|
* acknowledgment:
|
|
|
|
* "This product includes software developed by the OpenSSL Project
|
|
|
|
* for use in the OpenSSL Toolkit (http://www.openssl.org/)"
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE OpenSSL PROJECT ``AS IS'' AND ANY
|
|
|
|
* EXPRESSED OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE OpenSSL PROJECT OR
|
|
|
|
* ITS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
|
|
|
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
|
|
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
|
|
|
|
* OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
* ==================================================================== */
|
|
|
|
|
2019-01-09 03:35:56 +00:00
|
|
|
#include <assert.h>
|
2015-01-31 01:08:37 +00:00
|
|
|
#include <string.h>
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
#include <openssl/aead.h>
|
2014-06-20 20:00:00 +01:00
|
|
|
#include <openssl/aes.h>
|
|
|
|
#include <openssl/cipher.h>
|
|
|
|
#include <openssl/cpu.h>
|
|
|
|
#include <openssl/err.h>
|
|
|
|
#include <openssl/mem.h>
|
2016-03-25 22:07:11 +00:00
|
|
|
#include <openssl/nid.h>
|
2014-06-20 20:00:00 +01:00
|
|
|
#include <openssl/rand.h>
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
#include "internal.h"
|
2017-05-03 21:23:37 +01:00
|
|
|
#include "../../internal.h"
|
|
|
|
#include "../aes/internal.h"
|
|
|
|
#include "../modes/internal.h"
|
2017-05-20 02:11:20 +01:00
|
|
|
#include "../delocate.h"
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
|
2018-05-06 07:05:19 +01:00
|
|
|
OPENSSL_MSVC_PRAGMA(warning(push))
|
2017-08-18 19:06:02 +01:00
|
|
|
OPENSSL_MSVC_PRAGMA(warning(disable: 4702)) // Unreachable code.
|
Fix build when using Visual Studio 2015 Update 1.
Many of the compatibility issues are described at
https://msdn.microsoft.com/en-us/library/mt612856.aspx. The macros
that suppressed warnings on a per-function basis no longer work in
Update 1, so replace them with #pragmas. Update 1 warns when |size_t|
arguments to |printf| are casted, so stop doing that casting.
Unfortunately, this requires an ugly hack to continue working in
MSVC 2013 as MSVC 2013 doesn't support "%zu". Finally, Update 1 has new
warnings, some of which need to be suppressed.
---
Updated by davidben to give up on suppressing warnings in crypto/x509 and
crypto/x509v3 as those directories aren't changed much from upstream. In each
of these cases, upstream opted just blindly initialize the variable, so do the
same. Also switch C4265 to level 4, per Microsoft's recommendation and work
around a bug in limits.h that happens to get fixed by Google include order
style.
(limits.h is sensitive to whether corecrt.h, pulled in by stddef.h and some
other headers, is included before it. The reason it affected just one file is
we often put the file's header first, which means base.h is pulling in
stddef.h. Relying on this is ugly, but it's no worse than what everything else
is doing and this doesn't seem worth making something as tame as limits.h so
messy to use.)
Change-Id: I02d1f935356899f424d3525d03eca401bfa3e6cd
Reviewed-on: https://boringssl-review.googlesource.com/7480
Reviewed-by: David Benjamin <davidben@google.com>
2016-01-18 08:21:42 +00:00
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
typedef struct {
|
|
|
|
union {
|
|
|
|
double align;
|
|
|
|
AES_KEY ks;
|
|
|
|
} ks;
|
|
|
|
block128_f block;
|
|
|
|
union {
|
|
|
|
cbc128_f cbc;
|
|
|
|
ctr128_f ctr;
|
|
|
|
} stream;
|
|
|
|
} EVP_AES_KEY;
|
|
|
|
|
|
|
|
typedef struct {
|
2019-01-09 03:35:56 +00:00
|
|
|
GCM128_CONTEXT gcm;
|
2014-06-20 20:00:00 +01:00
|
|
|
union {
|
|
|
|
double align;
|
|
|
|
AES_KEY ks;
|
2017-08-18 19:06:02 +01:00
|
|
|
} ks; // AES key schedule to use
|
|
|
|
int key_set; // Set if key initialised
|
|
|
|
int iv_set; // Set if an iv is set
|
|
|
|
uint8_t *iv; // Temporary IV store
|
|
|
|
int ivlen; // IV length
|
2014-06-20 20:00:00 +01:00
|
|
|
int taglen;
|
2017-08-18 19:06:02 +01:00
|
|
|
int iv_gen; // It is OK to generate IVs
|
2014-06-20 20:00:00 +01:00
|
|
|
ctr128_f ctr;
|
|
|
|
} EVP_AES_GCM_CTX;
|
|
|
|
|
|
|
|
static int aes_init_key(EVP_CIPHER_CTX *ctx, const uint8_t *key,
|
Fix build when using Visual Studio 2015 Update 1.
Many of the compatibility issues are described at
https://msdn.microsoft.com/en-us/library/mt612856.aspx. The macros
that suppressed warnings on a per-function basis no longer work in
Update 1, so replace them with #pragmas. Update 1 warns when |size_t|
arguments to |printf| are casted, so stop doing that casting.
Unfortunately, this requires an ugly hack to continue working in
MSVC 2013 as MSVC 2013 doesn't support "%zu". Finally, Update 1 has new
warnings, some of which need to be suppressed.
---
Updated by davidben to give up on suppressing warnings in crypto/x509 and
crypto/x509v3 as those directories aren't changed much from upstream. In each
of these cases, upstream opted just blindly initialize the variable, so do the
same. Also switch C4265 to level 4, per Microsoft's recommendation and work
around a bug in limits.h that happens to get fixed by Google include order
style.
(limits.h is sensitive to whether corecrt.h, pulled in by stddef.h and some
other headers, is included before it. The reason it affected just one file is
we often put the file's header first, which means base.h is pulling in
stddef.h. Relying on this is ugly, but it's no worse than what everything else
is doing and this doesn't seem worth making something as tame as limits.h so
messy to use.)
Change-Id: I02d1f935356899f424d3525d03eca401bfa3e6cd
Reviewed-on: https://boringssl-review.googlesource.com/7480
Reviewed-by: David Benjamin <davidben@google.com>
2016-01-18 08:21:42 +00:00
|
|
|
const uint8_t *iv, int enc) {
|
2014-06-20 20:00:00 +01:00
|
|
|
int ret, mode;
|
|
|
|
EVP_AES_KEY *dat = (EVP_AES_KEY *)ctx->cipher_data;
|
|
|
|
|
|
|
|
mode = ctx->cipher->flags & EVP_CIPH_MODE_MASK;
|
|
|
|
if ((mode == EVP_CIPH_ECB_MODE || mode == EVP_CIPH_CBC_MODE) && !enc) {
|
2015-01-09 23:44:37 +00:00
|
|
|
if (hwaes_capable()) {
|
2016-09-23 20:47:24 +01:00
|
|
|
ret = aes_hw_set_decrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->block = aes_hw_decrypt;
|
2015-01-09 23:44:37 +00:00
|
|
|
dat->stream.cbc = NULL;
|
|
|
|
if (mode == EVP_CIPH_CBC_MODE) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->stream.cbc = aes_hw_cbc_encrypt;
|
2015-01-09 23:44:37 +00:00
|
|
|
}
|
|
|
|
} else if (bsaes_capable() && mode == EVP_CIPH_CBC_MODE) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
ret = aes_nohw_set_decrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
|
|
|
dat->block = aes_nohw_decrypt;
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->stream.cbc = bsaes_cbc_encrypt;
|
2014-06-20 20:00:00 +01:00
|
|
|
} else if (vpaes_capable()) {
|
|
|
|
ret = vpaes_set_decrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->block = vpaes_decrypt;
|
|
|
|
dat->stream.cbc = mode == EVP_CIPH_CBC_MODE ? vpaes_cbc_encrypt : NULL;
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
ret = aes_nohw_set_decrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
|
|
|
dat->block = aes_nohw_decrypt;
|
|
|
|
dat->stream.cbc = NULL;
|
|
|
|
#if defined(AES_NOHW_CBC)
|
|
|
|
if (mode == EVP_CIPH_CBC_MODE) {
|
|
|
|
dat->stream.cbc = aes_nohw_cbc_encrypt;
|
|
|
|
}
|
|
|
|
#endif
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2015-01-09 23:44:37 +00:00
|
|
|
} else if (hwaes_capable()) {
|
2016-09-23 20:47:24 +01:00
|
|
|
ret = aes_hw_set_encrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->block = aes_hw_encrypt;
|
2015-01-09 23:44:37 +00:00
|
|
|
dat->stream.cbc = NULL;
|
|
|
|
if (mode == EVP_CIPH_CBC_MODE) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->stream.cbc = aes_hw_cbc_encrypt;
|
2015-01-09 23:44:37 +00:00
|
|
|
} else if (mode == EVP_CIPH_CTR_MODE) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->stream.ctr = aes_hw_ctr32_encrypt_blocks;
|
2015-01-09 23:44:37 +00:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
} else if (bsaes_capable() && mode == EVP_CIPH_CTR_MODE) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
ret = aes_nohw_set_encrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
|
|
|
dat->block = aes_nohw_encrypt;
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->stream.ctr = bsaes_ctr32_encrypt_blocks;
|
2014-06-20 20:00:00 +01:00
|
|
|
} else if (vpaes_capable()) {
|
|
|
|
ret = vpaes_set_encrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
dat->block = vpaes_encrypt;
|
|
|
|
dat->stream.cbc = mode == EVP_CIPH_CBC_MODE ? vpaes_cbc_encrypt : NULL;
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
ret = aes_nohw_set_encrypt_key(key, ctx->key_len * 8, &dat->ks.ks);
|
|
|
|
dat->block = aes_nohw_encrypt;
|
|
|
|
dat->stream.cbc = NULL;
|
|
|
|
#if defined(AES_NOHW_CBC)
|
|
|
|
if (mode == EVP_CIPH_CBC_MODE) {
|
|
|
|
dat->stream.cbc = aes_nohw_cbc_encrypt;
|
|
|
|
}
|
|
|
|
#endif
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if (ret < 0) {
|
2015-06-29 05:28:17 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_AES_KEY_SETUP_FAILED);
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-02-24 20:44:40 +00:00
|
|
|
static int aes_cbc_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out, const uint8_t *in,
|
|
|
|
size_t len) {
|
2014-06-20 20:00:00 +01:00
|
|
|
EVP_AES_KEY *dat = (EVP_AES_KEY *)ctx->cipher_data;
|
|
|
|
|
|
|
|
if (dat->stream.cbc) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
(*dat->stream.cbc)(in, out, len, &dat->ks.ks, ctx->iv, ctx->encrypt);
|
2014-06-20 20:00:00 +01:00
|
|
|
} else if (ctx->encrypt) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_cbc128_encrypt(in, out, len, &dat->ks.ks, ctx->iv, dat->block);
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_cbc128_decrypt(in, out, len, &dat->ks.ks, ctx->iv, dat->block);
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-02-24 20:44:40 +00:00
|
|
|
static int aes_ecb_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out, const uint8_t *in,
|
|
|
|
size_t len) {
|
2014-06-20 20:00:00 +01:00
|
|
|
size_t bl = ctx->cipher->block_size;
|
|
|
|
EVP_AES_KEY *dat = (EVP_AES_KEY *)ctx->cipher_data;
|
|
|
|
|
|
|
|
if (len < bl) {
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2016-09-05 17:47:25 +01:00
|
|
|
len -= bl;
|
|
|
|
for (size_t i = 0; i <= len; i += bl) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
(*dat->block)(in + i, out + i, &dat->ks.ks);
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-02-24 20:44:40 +00:00
|
|
|
static int aes_ctr_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out, const uint8_t *in,
|
|
|
|
size_t len) {
|
2014-06-20 20:00:00 +01:00
|
|
|
EVP_AES_KEY *dat = (EVP_AES_KEY *)ctx->cipher_data;
|
|
|
|
|
|
|
|
if (dat->stream.ctr) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_ctr128_encrypt_ctr32(in, out, len, &dat->ks.ks, ctx->iv, ctx->buf,
|
2016-04-16 20:20:07 +01:00
|
|
|
&ctx->num, dat->stream.ctr);
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_ctr128_encrypt(in, out, len, &dat->ks.ks, ctx->iv, ctx->buf,
|
|
|
|
&ctx->num, dat->block);
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-02-24 20:44:40 +00:00
|
|
|
static int aes_ofb_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out, const uint8_t *in,
|
|
|
|
size_t len) {
|
|
|
|
EVP_AES_KEY *dat = (EVP_AES_KEY *)ctx->cipher_data;
|
|
|
|
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_ofb128_encrypt(in, out, len, &dat->ks.ks, ctx->iv, &ctx->num,
|
|
|
|
dat->block);
|
2015-02-24 20:44:40 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-09-17 22:44:22 +01:00
|
|
|
ctr128_f aes_ctr_set_key(AES_KEY *aes_key, GCM128_KEY *gcm_key,
|
2017-04-07 19:46:13 +01:00
|
|
|
block128_f *out_block, const uint8_t *key,
|
2019-01-12 15:29:10 +00:00
|
|
|
size_t key_bytes, int large_inputs) {
|
2015-01-09 23:44:37 +00:00
|
|
|
if (hwaes_capable()) {
|
2017-04-07 19:46:13 +01:00
|
|
|
aes_hw_set_encrypt_key(key, key_bytes * 8, aes_key);
|
2018-09-17 22:44:22 +01:00
|
|
|
if (gcm_key != NULL) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_gcm128_init_key(gcm_key, aes_key, aes_hw_encrypt, 1);
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
|
|
|
if (out_block) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
*out_block = aes_hw_encrypt;
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
return aes_hw_ctr32_encrypt_blocks;
|
2015-01-09 23:44:37 +00:00
|
|
|
}
|
|
|
|
|
2019-01-12 15:29:10 +00:00
|
|
|
const int bsaes_ok = bsaes_capable();
|
|
|
|
const int vpaes_ok = vpaes_capable();
|
|
|
|
if (bsaes_ok && (large_inputs || !vpaes_ok)) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
aes_nohw_set_encrypt_key(key, key_bytes * 8, aes_key);
|
2018-09-17 22:44:22 +01:00
|
|
|
if (gcm_key != NULL) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
CRYPTO_gcm128_init_key(gcm_key, aes_key, aes_nohw_encrypt, 0);
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
|
|
|
if (out_block) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
*out_block = aes_nohw_encrypt;
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
return bsaes_ctr32_encrypt_blocks;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
2019-01-12 15:29:10 +00:00
|
|
|
if (vpaes_ok) {
|
2017-04-07 19:46:13 +01:00
|
|
|
vpaes_set_encrypt_key(key, key_bytes * 8, aes_key);
|
2015-03-13 18:49:22 +00:00
|
|
|
if (out_block) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
*out_block = vpaes_encrypt;
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
2018-09-17 22:44:22 +01:00
|
|
|
if (gcm_key != NULL) {
|
Fix undefined block128_f, etc., casts.
This one is a little thorny. All the various block cipher modes
functions and callbacks take a void *key. This allows them to be used
with multiple kinds of block ciphers.
However, the implementations of those callbacks are the normal typed
functions, like AES_encrypt. Those take AES_KEY *key. While, at the ABI
level, this is perfectly fine, C considers this undefined behavior.
If we wish to preserve this genericness, we could either instantiate
multiple versions of these mode functions or create wrappers of
AES_encrypt, etc., that take void *key.
The former means more code and is tedious without C++ templates (maybe
someday...). The latter would not be difficult for a compiler to
optimize out. C mistakenly allowed comparing function pointers for
equality, which means a compiler cannot replace pointers to wrapper
functions with the real thing. (That said, the performance-sensitive
bits already act in chunks, e.g. ctr128_f, so the function call overhead
shouldn't matter.)
But our only 128-bit block cipher is AES anyway, so I just switched
things to use AES_KEY throughout. AES is doing fine, and hopefully we
would have the sense not to pair a hypothetical future block cipher with
so many modes!
Change-Id: Ied3e843f0e3042a439f09e655b29847ade9d4c7d
Reviewed-on: https://boringssl-review.googlesource.com/32107
Reviewed-by: Adam Langley <agl@google.com>
2018-09-23 02:37:01 +01:00
|
|
|
CRYPTO_gcm128_init_key(gcm_key, aes_key, vpaes_encrypt, 0);
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
aes_nohw_set_encrypt_key(key, key_bytes * 8, aes_key);
|
2018-09-17 22:44:22 +01:00
|
|
|
if (gcm_key != NULL) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
CRYPTO_gcm128_init_key(gcm_key, aes_key, aes_nohw_encrypt, 0);
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
|
|
|
if (out_block) {
|
Avoid double-dispatch with AES_* vs aes_nohw_*.
In particular, consistently pair bsaes with aes_nohw.
Ideally the aes_nohw_* calls in bsaes-*.pl would be patched out and
bsaes grows its own constant-time key setup
(https://crbug.com/boringssl/256), but I'll sort that out separately. In
the meantime, avoid going through AES_* which now dispatch. This avoids
several nuisances:
1. If we were to add, say, a vpaes-armv7.pl the ABI tests would break.
Fundamentally, we cannot assume that an AES_KEY has one and only one
representation and must keep everything matching up.
2. AES_* functions should enable vpaes. This makes AES_* faster and
constant-time for vector-capable CPUs
(https://crbug.com/boringssl/263), relevant for QUIC packet number
encryption, allowing us to add vpaes-armv8.pl
(https://crbug.com/boringssl/246) without carrying a (likely) mostly
unused AES implementation.
3. It's silly to double-dispatch when the EVP layer has already
dispatched.
4. We should avoid asm calling into C. Otherwise, we need to test asm
for ABI compliance as both caller and callee. Currently we only test
it for callee compliance. When asm calls into asm, it *should* comply
with the ABI as caller too, but mistakes don't matter as long as the
called function triggers it. If the function is asm, this is fixed.
If it is C, we must care about arbitrary C compiler output.
Bug: 263
Change-Id: Ic85af5c765fd57cbffeaf301c3872bad6c5bbf78
Reviewed-on: https://boringssl-review.googlesource.com/c/34874
Commit-Queue: Adam Langley <agl@google.com>
Reviewed-by: Adam Langley <agl@google.com>
2019-02-10 04:05:43 +00:00
|
|
|
*out_block = aes_nohw_encrypt;
|
2015-03-13 18:49:22 +00:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2019-01-09 03:35:56 +00:00
|
|
|
#if defined(OPENSSL_32_BIT)
|
|
|
|
#define EVP_AES_GCM_CTX_PADDING (4+8)
|
|
|
|
#else
|
|
|
|
#define EVP_AES_GCM_CTX_PADDING 8
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static EVP_AES_GCM_CTX *aes_gcm_from_cipher_ctx(EVP_CIPHER_CTX *ctx) {
|
|
|
|
#if defined(__GNUC__) || defined(__clang__)
|
|
|
|
OPENSSL_STATIC_ASSERT(
|
|
|
|
alignof(EVP_AES_GCM_CTX) <= 16,
|
|
|
|
"EVP_AES_GCM_CTX needs more alignment than this function provides");
|
|
|
|
#endif
|
|
|
|
|
|
|
|
// |malloc| guarantees up to 4-byte alignment on 32-bit and 8-byte alignment
|
|
|
|
// on 64-bit systems, so we need to adjust to reach 16-byte alignment.
|
|
|
|
assert(ctx->cipher->ctx_size ==
|
|
|
|
sizeof(EVP_AES_GCM_CTX) + EVP_AES_GCM_CTX_PADDING);
|
|
|
|
|
|
|
|
char *ptr = ctx->cipher_data;
|
|
|
|
#if defined(OPENSSL_32_BIT)
|
|
|
|
assert((uintptr_t)ptr % 4 == 0);
|
|
|
|
ptr += (uintptr_t)ptr & 4;
|
|
|
|
#endif
|
|
|
|
assert((uintptr_t)ptr % 8 == 0);
|
|
|
|
ptr += (uintptr_t)ptr & 8;
|
|
|
|
return (EVP_AES_GCM_CTX *)ptr;
|
|
|
|
}
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
static int aes_gcm_init_key(EVP_CIPHER_CTX *ctx, const uint8_t *key,
|
|
|
|
const uint8_t *iv, int enc) {
|
2019-01-09 03:35:56 +00:00
|
|
|
EVP_AES_GCM_CTX *gctx = aes_gcm_from_cipher_ctx(ctx);
|
2014-06-20 20:00:00 +01:00
|
|
|
if (!iv && !key) {
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
if (key) {
|
2018-09-17 22:44:22 +01:00
|
|
|
OPENSSL_memset(&gctx->gcm, 0, sizeof(gctx->gcm));
|
|
|
|
gctx->ctr = aes_ctr_set_key(&gctx->ks.ks, &gctx->gcm.gcm_key, NULL, key,
|
2019-01-12 15:29:10 +00:00
|
|
|
ctx->key_len, 1 /* large inputs */);
|
2017-08-18 19:06:02 +01:00
|
|
|
// If we have an iv can set it directly, otherwise use saved IV.
|
2014-06-20 20:00:00 +01:00
|
|
|
if (iv == NULL && gctx->iv_set) {
|
|
|
|
iv = gctx->iv;
|
|
|
|
}
|
|
|
|
if (iv) {
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gctx->gcm, &gctx->ks.ks, iv, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->iv_set = 1;
|
|
|
|
}
|
|
|
|
gctx->key_set = 1;
|
|
|
|
} else {
|
2017-08-18 19:06:02 +01:00
|
|
|
// If key set use IV, otherwise copy
|
2014-06-20 20:00:00 +01:00
|
|
|
if (gctx->key_set) {
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gctx->gcm, &gctx->ks.ks, iv, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(gctx->iv, iv, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
gctx->iv_set = 1;
|
|
|
|
gctx->iv_gen = 0;
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2015-02-23 07:15:50 +00:00
|
|
|
static void aes_gcm_cleanup(EVP_CIPHER_CTX *c) {
|
2019-01-09 03:35:56 +00:00
|
|
|
EVP_AES_GCM_CTX *gctx = aes_gcm_from_cipher_ctx(c);
|
2014-06-20 20:00:00 +01:00
|
|
|
OPENSSL_cleanse(&gctx->gcm, sizeof(gctx->gcm));
|
|
|
|
if (gctx->iv != c->iv) {
|
|
|
|
OPENSSL_free(gctx->iv);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-18 19:06:02 +01:00
|
|
|
// increment counter (64-bit int) by 1
|
2014-06-20 20:00:00 +01:00
|
|
|
static void ctr64_inc(uint8_t *counter) {
|
|
|
|
int n = 8;
|
|
|
|
uint8_t c;
|
|
|
|
|
|
|
|
do {
|
|
|
|
--n;
|
|
|
|
c = counter[n];
|
|
|
|
++c;
|
|
|
|
counter[n] = c;
|
|
|
|
if (c) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
} while (n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int aes_gcm_ctrl(EVP_CIPHER_CTX *c, int type, int arg, void *ptr) {
|
2019-01-09 03:35:56 +00:00
|
|
|
EVP_AES_GCM_CTX *gctx = aes_gcm_from_cipher_ctx(c);
|
2014-06-20 20:00:00 +01:00
|
|
|
switch (type) {
|
|
|
|
case EVP_CTRL_INIT:
|
|
|
|
gctx->key_set = 0;
|
|
|
|
gctx->iv_set = 0;
|
|
|
|
gctx->ivlen = c->cipher->iv_len;
|
|
|
|
gctx->iv = c->iv;
|
|
|
|
gctx->taglen = -1;
|
|
|
|
gctx->iv_gen = 0;
|
|
|
|
return 1;
|
|
|
|
|
2018-08-09 17:56:34 +01:00
|
|
|
case EVP_CTRL_AEAD_SET_IVLEN:
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg <= 0) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-08-18 19:06:02 +01:00
|
|
|
// Allocate memory for IV if needed
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg > EVP_MAX_IV_LENGTH && arg > gctx->ivlen) {
|
|
|
|
if (gctx->iv != c->iv) {
|
|
|
|
OPENSSL_free(gctx->iv);
|
|
|
|
}
|
|
|
|
gctx->iv = OPENSSL_malloc(arg);
|
|
|
|
if (!gctx->iv) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
gctx->ivlen = arg;
|
|
|
|
return 1;
|
|
|
|
|
2018-08-09 17:56:34 +01:00
|
|
|
case EVP_CTRL_AEAD_SET_TAG:
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg <= 0 || arg > 16 || c->encrypt) {
|
|
|
|
return 0;
|
|
|
|
}
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(c->buf, ptr, arg);
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->taglen = arg;
|
|
|
|
return 1;
|
|
|
|
|
2018-08-09 17:56:34 +01:00
|
|
|
case EVP_CTRL_AEAD_GET_TAG:
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg <= 0 || arg > 16 || !c->encrypt || gctx->taglen < 0) {
|
|
|
|
return 0;
|
|
|
|
}
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(ptr, c->buf, arg);
|
2014-06-20 20:00:00 +01:00
|
|
|
return 1;
|
|
|
|
|
2018-08-09 17:56:34 +01:00
|
|
|
case EVP_CTRL_AEAD_SET_IV_FIXED:
|
2017-08-18 19:06:02 +01:00
|
|
|
// Special case: -1 length restores whole IV
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg == -1) {
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(gctx->iv, ptr, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->iv_gen = 1;
|
|
|
|
return 1;
|
|
|
|
}
|
2017-08-18 19:06:02 +01:00
|
|
|
// Fixed field must be at least 4 bytes and invocation field
|
|
|
|
// at least 8.
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg < 4 || (gctx->ivlen - arg) < 8) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (arg) {
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(gctx->iv, ptr, arg);
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-12-16 12:48:10 +00:00
|
|
|
if (c->encrypt && !RAND_bytes(gctx->iv + arg, gctx->ivlen - arg)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
gctx->iv_gen = 1;
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
case EVP_CTRL_GCM_IV_GEN:
|
|
|
|
if (gctx->iv_gen == 0 || gctx->key_set == 0) {
|
|
|
|
return 0;
|
|
|
|
}
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gctx->gcm, &gctx->ks.ks, gctx->iv, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
if (arg <= 0 || arg > gctx->ivlen) {
|
|
|
|
arg = gctx->ivlen;
|
|
|
|
}
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(ptr, gctx->iv + gctx->ivlen - arg, arg);
|
2017-08-18 19:06:02 +01:00
|
|
|
// Invocation field will be at least 8 bytes in size and
|
|
|
|
// so no need to check wrap around or increment more than
|
|
|
|
// last 8 bytes.
|
2014-06-20 20:00:00 +01:00
|
|
|
ctr64_inc(gctx->iv + gctx->ivlen - 8);
|
|
|
|
gctx->iv_set = 1;
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
case EVP_CTRL_GCM_SET_IV_INV:
|
|
|
|
if (gctx->iv_gen == 0 || gctx->key_set == 0 || c->encrypt) {
|
|
|
|
return 0;
|
|
|
|
}
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(gctx->iv + gctx->ivlen - arg, ptr, arg);
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gctx->gcm, &gctx->ks.ks, gctx->iv, gctx->ivlen);
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->iv_set = 1;
|
|
|
|
return 1;
|
|
|
|
|
2014-07-25 01:42:11 +01:00
|
|
|
case EVP_CTRL_COPY: {
|
|
|
|
EVP_CIPHER_CTX *out = ptr;
|
2019-01-09 03:35:56 +00:00
|
|
|
EVP_AES_GCM_CTX *gctx_out = aes_gcm_from_cipher_ctx(out);
|
2014-07-25 01:42:11 +01:00
|
|
|
if (gctx->iv == c->iv) {
|
|
|
|
gctx_out->iv = out->iv;
|
|
|
|
} else {
|
|
|
|
gctx_out->iv = OPENSSL_malloc(gctx->ivlen);
|
|
|
|
if (!gctx_out->iv) {
|
|
|
|
return 0;
|
|
|
|
}
|
2016-12-13 06:07:13 +00:00
|
|
|
OPENSSL_memcpy(gctx_out->iv, gctx->iv, gctx->ivlen);
|
2014-07-25 01:42:11 +01:00
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
default:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int aes_gcm_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out, const uint8_t *in,
|
|
|
|
size_t len) {
|
2019-01-09 03:35:56 +00:00
|
|
|
EVP_AES_GCM_CTX *gctx = aes_gcm_from_cipher_ctx(ctx);
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-08-18 19:06:02 +01:00
|
|
|
// If not set up, return error
|
2014-06-20 20:00:00 +01:00
|
|
|
if (!gctx->key_set) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
if (!gctx->iv_set) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
if (in) {
|
|
|
|
if (out == NULL) {
|
2014-06-20 20:00:00 +01:00
|
|
|
if (!CRYPTO_gcm128_aad(&gctx->gcm, in, len)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
} else if (ctx->encrypt) {
|
|
|
|
if (gctx->ctr) {
|
2016-02-18 04:59:19 +00:00
|
|
|
if (!CRYPTO_gcm128_encrypt_ctr32(&gctx->gcm, &gctx->ks.ks, in, out, len,
|
|
|
|
gctx->ctr)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
2016-02-18 04:59:19 +00:00
|
|
|
if (!CRYPTO_gcm128_encrypt(&gctx->gcm, &gctx->ks.ks, in, out, len)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (gctx->ctr) {
|
2016-02-18 04:59:19 +00:00
|
|
|
if (!CRYPTO_gcm128_decrypt_ctr32(&gctx->gcm, &gctx->ks.ks, in, out, len,
|
|
|
|
gctx->ctr)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
} else {
|
2016-02-18 04:59:19 +00:00
|
|
|
if (!CRYPTO_gcm128_decrypt(&gctx->gcm, &gctx->ks.ks, in, out, len)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return len;
|
|
|
|
} else {
|
|
|
|
if (!ctx->encrypt) {
|
2014-06-20 20:00:00 +01:00
|
|
|
if (gctx->taglen < 0 ||
|
2015-08-14 21:05:21 +01:00
|
|
|
!CRYPTO_gcm128_finish(&gctx->gcm, ctx->buf, gctx->taglen)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return -1;
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->iv_set = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
CRYPTO_gcm128_tag(&gctx->gcm, ctx->buf, 16);
|
|
|
|
gctx->taglen = 16;
|
2017-08-18 19:06:02 +01:00
|
|
|
// Don't reuse the IV
|
2014-06-20 20:00:00 +01:00
|
|
|
gctx->iv_set = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_128_cbc_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_cbc;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CBC_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_cbc_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_128_ctr_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_ctr;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CTR_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ctr_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_128_ecb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ecb_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_128_ofb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_ofb128;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_OFB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ofb_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_128_gcm_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_gcm;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->iv_len = 12;
|
2019-01-09 03:35:56 +00:00
|
|
|
out->ctx_size = sizeof(EVP_AES_GCM_CTX) + EVP_AES_GCM_CTX_PADDING;
|
2017-05-03 21:23:37 +01:00
|
|
|
out->flags = EVP_CIPH_GCM_MODE | EVP_CIPH_CUSTOM_IV |
|
|
|
|
EVP_CIPH_FLAG_CUSTOM_CIPHER | EVP_CIPH_ALWAYS_CALL_INIT |
|
|
|
|
EVP_CIPH_CTRL_INIT | EVP_CIPH_FLAG_AEAD_CIPHER;
|
|
|
|
out->init = aes_gcm_init_key;
|
|
|
|
out->cipher = aes_gcm_cipher;
|
|
|
|
out->cleanup = aes_gcm_cleanup;
|
|
|
|
out->ctrl = aes_gcm_ctrl;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_192_cbc_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_cbc;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CBC_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_cbc_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_192_ctr_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_ctr;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CTR_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ctr_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_192_ecb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ecb_cipher;
|
|
|
|
}
|
|
|
|
|
2018-06-16 01:14:16 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_192_ofb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_ofb128;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_OFB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ofb_cipher;
|
|
|
|
}
|
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_192_gcm_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_gcm;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->iv_len = 12;
|
2019-01-09 03:35:56 +00:00
|
|
|
out->ctx_size = sizeof(EVP_AES_GCM_CTX) + EVP_AES_GCM_CTX_PADDING;
|
2017-05-03 21:23:37 +01:00
|
|
|
out->flags = EVP_CIPH_GCM_MODE | EVP_CIPH_CUSTOM_IV |
|
|
|
|
EVP_CIPH_FLAG_CUSTOM_CIPHER | EVP_CIPH_ALWAYS_CALL_INIT |
|
|
|
|
EVP_CIPH_CTRL_INIT | EVP_CIPH_FLAG_AEAD_CIPHER;
|
|
|
|
out->init = aes_gcm_init_key;
|
|
|
|
out->cipher = aes_gcm_cipher;
|
|
|
|
out->cleanup = aes_gcm_cleanup;
|
|
|
|
out->ctrl = aes_gcm_ctrl;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_256_cbc_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_cbc;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CBC_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_cbc_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_256_ctr_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_ctr;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_CTR_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ctr_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_256_ecb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ecb_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_256_ofb_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_ofb128;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->iv_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_OFB_MODE;
|
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_ofb_cipher;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_256_gcm_generic) {
|
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_gcm;
|
|
|
|
out->block_size = 1;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->iv_len = 12;
|
2019-01-09 03:35:56 +00:00
|
|
|
out->ctx_size = sizeof(EVP_AES_GCM_CTX) + EVP_AES_GCM_CTX_PADDING;
|
2017-05-03 21:23:37 +01:00
|
|
|
out->flags = EVP_CIPH_GCM_MODE | EVP_CIPH_CUSTOM_IV |
|
|
|
|
EVP_CIPH_FLAG_CUSTOM_CIPHER | EVP_CIPH_ALWAYS_CALL_INIT |
|
|
|
|
EVP_CIPH_CTRL_INIT | EVP_CIPH_FLAG_AEAD_CIPHER;
|
|
|
|
out->init = aes_gcm_init_key;
|
|
|
|
out->cipher = aes_gcm_cipher;
|
|
|
|
out->cleanup = aes_gcm_cleanup;
|
|
|
|
out->ctrl = aes_gcm_ctrl;
|
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
#if defined(HWAES_ECB)
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
static int aes_hw_ecb_cipher(EVP_CIPHER_CTX *ctx, uint8_t *out,
|
|
|
|
const uint8_t *in, size_t len) {
|
2014-06-20 20:00:00 +01:00
|
|
|
size_t bl = ctx->cipher->block_size;
|
|
|
|
|
|
|
|
if (len < bl) {
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
aes_hw_ecb_encrypt(in, out, len, ctx->cipher_data, ctx->encrypt);
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_hw_128_ecb) {
|
2017-05-03 21:23:37 +01:00
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_128_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 16;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
2018-05-14 22:28:36 +01:00
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_hw_ecb_cipher;
|
2018-05-12 16:04:57 +01:00
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_hw_192_ecb) {
|
2017-05-03 21:23:37 +01:00
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_192_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 24;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
2018-05-14 22:28:36 +01:00
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_hw_ecb_cipher;
|
2018-05-12 16:04:57 +01:00
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
DEFINE_LOCAL_DATA(EVP_CIPHER, aes_hw_256_ecb) {
|
2017-05-03 21:23:37 +01:00
|
|
|
memset(out, 0, sizeof(EVP_CIPHER));
|
|
|
|
|
|
|
|
out->nid = NID_aes_256_ecb;
|
|
|
|
out->block_size = 16;
|
|
|
|
out->key_len = 32;
|
|
|
|
out->ctx_size = sizeof(EVP_AES_KEY);
|
|
|
|
out->flags = EVP_CIPH_ECB_MODE;
|
2018-05-14 22:28:36 +01:00
|
|
|
out->init = aes_init_key;
|
|
|
|
out->cipher = aes_hw_ecb_cipher;
|
2018-05-12 16:04:57 +01:00
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
#define EVP_ECB_CIPHER_FUNCTION(keybits) \
|
|
|
|
const EVP_CIPHER *EVP_aes_##keybits##_ecb(void) { \
|
|
|
|
if (hwaes_capable()) { \
|
|
|
|
return aes_hw_##keybits##_ecb(); \
|
|
|
|
} \
|
|
|
|
return aes_##keybits##_ecb_generic(); \
|
|
|
|
}
|
2018-05-11 18:38:51 +01:00
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
#else
|
2018-05-12 16:04:57 +01:00
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
#define EVP_ECB_CIPHER_FUNCTION(keybits) \
|
|
|
|
const EVP_CIPHER *EVP_aes_##keybits##_ecb(void) { \
|
|
|
|
return aes_##keybits##_ecb_generic(); \
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
#endif // HWAES_ECB
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
#define EVP_CIPHER_FUNCTION(keybits, mode) \
|
|
|
|
const EVP_CIPHER *EVP_aes_##keybits##_##mode(void) { \
|
2017-05-03 21:23:37 +01:00
|
|
|
return aes_##keybits##_##mode##_generic(); \
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
EVP_CIPHER_FUNCTION(128, cbc)
|
|
|
|
EVP_CIPHER_FUNCTION(128, ctr)
|
2015-02-24 20:44:40 +00:00
|
|
|
EVP_CIPHER_FUNCTION(128, ofb)
|
2014-06-20 20:00:00 +01:00
|
|
|
EVP_CIPHER_FUNCTION(128, gcm)
|
|
|
|
|
2015-04-02 21:09:01 +01:00
|
|
|
EVP_CIPHER_FUNCTION(192, cbc)
|
|
|
|
EVP_CIPHER_FUNCTION(192, ctr)
|
2018-06-16 01:14:16 +01:00
|
|
|
EVP_CIPHER_FUNCTION(192, ofb)
|
2015-04-02 21:09:01 +01:00
|
|
|
EVP_CIPHER_FUNCTION(192, gcm)
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
EVP_CIPHER_FUNCTION(256, cbc)
|
|
|
|
EVP_CIPHER_FUNCTION(256, ctr)
|
2015-02-24 20:44:40 +00:00
|
|
|
EVP_CIPHER_FUNCTION(256, ofb)
|
2014-06-20 20:00:00 +01:00
|
|
|
EVP_CIPHER_FUNCTION(256, gcm)
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2018-05-14 22:28:36 +01:00
|
|
|
EVP_ECB_CIPHER_FUNCTION(128)
|
|
|
|
EVP_ECB_CIPHER_FUNCTION(192)
|
|
|
|
EVP_ECB_CIPHER_FUNCTION(256)
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
#define EVP_AEAD_AES_GCM_TAG_LEN 16
|
|
|
|
|
|
|
|
struct aead_aes_gcm_ctx {
|
|
|
|
union {
|
|
|
|
double align;
|
|
|
|
AES_KEY ks;
|
|
|
|
} ks;
|
2018-09-17 22:44:22 +01:00
|
|
|
GCM128_KEY gcm_key;
|
2014-06-20 20:00:00 +01:00
|
|
|
ctr128_f ctr;
|
|
|
|
};
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
static int aead_aes_gcm_init_impl(struct aead_aes_gcm_ctx *gcm_ctx,
|
2017-06-27 23:49:40 +01:00
|
|
|
size_t *out_tag_len, const uint8_t *key,
|
|
|
|
size_t key_len, size_t tag_len) {
|
2014-06-20 20:00:00 +01:00
|
|
|
const size_t key_bits = key_len * 8;
|
|
|
|
|
|
|
|
if (key_bits != 128 && key_bits != 256) {
|
2015-06-29 05:28:17 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_BAD_KEY_LENGTH);
|
2017-08-18 19:06:02 +01:00
|
|
|
return 0; // EVP_AEAD_CTX_init should catch this.
|
2014-06-20 20:00:00 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if (tag_len == EVP_AEAD_DEFAULT_TAG_LENGTH) {
|
|
|
|
tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tag_len > EVP_AEAD_AES_GCM_TAG_LEN) {
|
2015-06-29 05:28:17 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_TAG_TOO_LARGE);
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-01-12 15:29:10 +00:00
|
|
|
gcm_ctx->ctr = aes_ctr_set_key(&gcm_ctx->ks.ks, &gcm_ctx->gcm_key, NULL, key,
|
|
|
|
key_len, 1 /* large inputs */);
|
2017-06-27 23:49:40 +01:00
|
|
|
*out_tag_len = tag_len;
|
2017-05-24 21:54:35 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(sizeof(((EVP_AEAD_CTX *)NULL)->state) >=
|
|
|
|
sizeof(struct aead_aes_gcm_ctx),
|
|
|
|
"AEAD state is too small");
|
2018-10-16 20:11:51 +01:00
|
|
|
#if defined(__GNUC__) || defined(__clang__)
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(alignof(union evp_aead_ctx_st_state) >=
|
|
|
|
alignof(struct aead_aes_gcm_ctx),
|
|
|
|
"AEAD state has insufficient alignment");
|
2018-10-16 20:11:51 +01:00
|
|
|
#endif
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
static int aead_aes_gcm_init(EVP_AEAD_CTX *ctx, const uint8_t *key,
|
2017-06-27 23:49:40 +01:00
|
|
|
size_t key_len, size_t requested_tag_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_ctx *gcm_ctx = (struct aead_aes_gcm_ctx *) &ctx->state;
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
size_t actual_tag_len;
|
|
|
|
if (!aead_aes_gcm_init_impl(gcm_ctx, &actual_tag_len, key, key_len,
|
|
|
|
requested_tag_len)) {
|
2017-05-24 21:54:35 +01:00
|
|
|
return 0;
|
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
ctx->tag_len = actual_tag_len;
|
2014-06-20 20:00:00 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-10-16 20:11:51 +01:00
|
|
|
static void aead_aes_gcm_cleanup(EVP_AEAD_CTX *ctx) {}
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-06-06 20:29:48 +01:00
|
|
|
static int aead_aes_gcm_seal_scatter(const EVP_AEAD_CTX *ctx, uint8_t *out,
|
|
|
|
uint8_t *out_tag, size_t *out_tag_len,
|
|
|
|
size_t max_out_tag_len,
|
|
|
|
const uint8_t *nonce, size_t nonce_len,
|
|
|
|
const uint8_t *in, size_t in_len,
|
2017-06-23 22:49:22 +01:00
|
|
|
const uint8_t *extra_in,
|
|
|
|
size_t extra_in_len,
|
2017-06-06 20:29:48 +01:00
|
|
|
const uint8_t *ad, size_t ad_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_ctx *gcm_ctx = (struct aead_aes_gcm_ctx *) &ctx->state;
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-06-26 18:16:50 +01:00
|
|
|
if (extra_in_len + ctx->tag_len < ctx->tag_len) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_TOO_LARGE);
|
|
|
|
return 0;
|
|
|
|
}
|
2017-07-22 00:27:54 +01:00
|
|
|
if (max_out_tag_len < extra_in_len + ctx->tag_len) {
|
2017-06-26 18:16:50 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_BUFFER_TOO_SMALL);
|
|
|
|
return 0;
|
|
|
|
}
|
2017-05-23 17:40:58 +01:00
|
|
|
if (nonce_len == 0) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_INVALID_NONCE_SIZE);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-09-27 07:12:01 +01:00
|
|
|
const AES_KEY *key = &gcm_ctx->ks.ks;
|
|
|
|
|
2018-09-17 22:44:22 +01:00
|
|
|
GCM128_CONTEXT gcm;
|
|
|
|
OPENSSL_memset(&gcm, 0, sizeof(gcm));
|
|
|
|
OPENSSL_memcpy(&gcm.gcm_key, &gcm_ctx->gcm_key, sizeof(gcm.gcm_key));
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gcm, key, nonce, nonce_len);
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
if (ad_len > 0 && !CRYPTO_gcm128_aad(&gcm, ad, ad_len)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gcm_ctx->ctr) {
|
2015-09-27 07:12:01 +01:00
|
|
|
if (!CRYPTO_gcm128_encrypt_ctr32(&gcm, key, in, out, in_len,
|
|
|
|
gcm_ctx->ctr)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
} else {
|
2015-09-27 07:12:01 +01:00
|
|
|
if (!CRYPTO_gcm128_encrypt(&gcm, key, in, out, in_len)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-06-26 18:16:50 +01:00
|
|
|
if (extra_in_len) {
|
|
|
|
if (gcm_ctx->ctr) {
|
|
|
|
if (!CRYPTO_gcm128_encrypt_ctr32(&gcm, key, extra_in, out_tag,
|
|
|
|
extra_in_len, gcm_ctx->ctr)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (!CRYPTO_gcm128_encrypt(&gcm, key, extra_in, out_tag, extra_in_len)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
CRYPTO_gcm128_tag(&gcm, out_tag + extra_in_len, ctx->tag_len);
|
|
|
|
*out_tag_len = ctx->tag_len + extra_in_len;
|
|
|
|
|
2014-06-20 20:00:00 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2017-06-06 20:29:48 +01:00
|
|
|
static int aead_aes_gcm_open_gather(const EVP_AEAD_CTX *ctx, uint8_t *out,
|
|
|
|
const uint8_t *nonce, size_t nonce_len,
|
|
|
|
const uint8_t *in, size_t in_len,
|
|
|
|
const uint8_t *in_tag, size_t in_tag_len,
|
|
|
|
const uint8_t *ad, size_t ad_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_ctx *gcm_ctx = (struct aead_aes_gcm_ctx *) &ctx->state;
|
2014-06-20 20:00:00 +01:00
|
|
|
uint8_t tag[EVP_AEAD_AES_GCM_TAG_LEN];
|
|
|
|
|
2017-05-23 17:40:58 +01:00
|
|
|
if (nonce_len == 0) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_INVALID_NONCE_SIZE);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
if (in_tag_len != ctx->tag_len) {
|
2015-06-29 05:28:17 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_BAD_DECRYPT);
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-09-27 07:12:01 +01:00
|
|
|
const AES_KEY *key = &gcm_ctx->ks.ks;
|
|
|
|
|
2018-09-17 22:44:22 +01:00
|
|
|
GCM128_CONTEXT gcm;
|
|
|
|
OPENSSL_memset(&gcm, 0, sizeof(gcm));
|
|
|
|
OPENSSL_memcpy(&gcm.gcm_key, &gcm_ctx->gcm_key, sizeof(gcm.gcm_key));
|
2015-09-27 07:12:01 +01:00
|
|
|
CRYPTO_gcm128_setiv(&gcm, key, nonce, nonce_len);
|
2014-06-20 20:00:00 +01:00
|
|
|
|
|
|
|
if (!CRYPTO_gcm128_aad(&gcm, ad, ad_len)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (gcm_ctx->ctr) {
|
2017-06-06 20:29:48 +01:00
|
|
|
if (!CRYPTO_gcm128_decrypt_ctr32(&gcm, key, in, out, in_len,
|
|
|
|
gcm_ctx->ctr)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
} else {
|
2017-06-06 20:29:48 +01:00
|
|
|
if (!CRYPTO_gcm128_decrypt(&gcm, key, in, out, in_len)) {
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
CRYPTO_gcm128_tag(&gcm, tag, ctx->tag_len);
|
|
|
|
if (CRYPTO_memcmp(tag, in_tag, ctx->tag_len) != 0) {
|
2015-06-29 05:28:17 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_BAD_DECRYPT);
|
2014-06-20 20:00:00 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_128_gcm) {
|
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 16;
|
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-06-26 18:16:50 +01:00
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
out->init = aead_aes_gcm_init;
|
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
2017-06-06 20:29:48 +01:00
|
|
|
out->seal_scatter = aead_aes_gcm_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
2017-05-03 21:23:37 +01:00
|
|
|
}
|
2014-06-20 20:00:00 +01:00
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_256_gcm) {
|
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 32;
|
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-06-26 18:16:50 +01:00
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
2017-05-03 21:23:37 +01:00
|
|
|
out->init = aead_aes_gcm_init;
|
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
2017-06-06 20:29:48 +01:00
|
|
|
out->seal_scatter = aead_aes_gcm_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
2017-05-03 21:23:37 +01:00
|
|
|
}
|
2014-07-25 23:40:44 +01:00
|
|
|
|
2018-06-18 22:04:41 +01:00
|
|
|
struct aead_aes_gcm_tls12_ctx {
|
|
|
|
struct aead_aes_gcm_ctx gcm_ctx;
|
|
|
|
uint64_t min_next_nonce;
|
|
|
|
};
|
|
|
|
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(sizeof(((EVP_AEAD_CTX *)NULL)->state) >=
|
|
|
|
sizeof(struct aead_aes_gcm_tls12_ctx),
|
|
|
|
"AEAD state is too small");
|
2018-10-16 20:11:51 +01:00
|
|
|
#if defined(__GNUC__) || defined(__clang__)
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(alignof(union evp_aead_ctx_st_state) >=
|
|
|
|
alignof(struct aead_aes_gcm_tls12_ctx),
|
|
|
|
"AEAD state has insufficient alignment");
|
2018-10-16 20:11:51 +01:00
|
|
|
#endif
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
static int aead_aes_gcm_tls12_init(EVP_AEAD_CTX *ctx, const uint8_t *key,
|
2017-06-27 23:49:40 +01:00
|
|
|
size_t key_len, size_t requested_tag_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_tls12_ctx *gcm_ctx =
|
|
|
|
(struct aead_aes_gcm_tls12_ctx *) &ctx->state;
|
2017-04-19 21:45:09 +01:00
|
|
|
|
2018-01-25 23:43:48 +00:00
|
|
|
gcm_ctx->min_next_nonce = 0;
|
2017-05-24 21:54:35 +01:00
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
size_t actual_tag_len;
|
|
|
|
if (!aead_aes_gcm_init_impl(&gcm_ctx->gcm_ctx, &actual_tag_len, key, key_len,
|
|
|
|
requested_tag_len)) {
|
2017-04-19 21:45:09 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-27 23:49:40 +01:00
|
|
|
ctx->tag_len = actual_tag_len;
|
2017-05-24 21:54:35 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2017-06-06 20:29:48 +01:00
|
|
|
static int aead_aes_gcm_tls12_seal_scatter(
|
|
|
|
const EVP_AEAD_CTX *ctx, uint8_t *out, uint8_t *out_tag,
|
|
|
|
size_t *out_tag_len, size_t max_out_tag_len, const uint8_t *nonce,
|
2017-06-23 22:49:22 +01:00
|
|
|
size_t nonce_len, const uint8_t *in, size_t in_len, const uint8_t *extra_in,
|
|
|
|
size_t extra_in_len, const uint8_t *ad, size_t ad_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_tls12_ctx *gcm_ctx =
|
|
|
|
(struct aead_aes_gcm_tls12_ctx *) &ctx->state;
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
if (nonce_len != 12) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_UNSUPPORTED_NONCE_SIZE);
|
2017-04-19 21:45:09 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-01-25 23:43:48 +00:00
|
|
|
// The given nonces must be strictly monotonically increasing.
|
|
|
|
uint64_t given_counter;
|
|
|
|
OPENSSL_memcpy(&given_counter, nonce + nonce_len - sizeof(given_counter),
|
|
|
|
sizeof(given_counter));
|
|
|
|
given_counter = CRYPTO_bswap8(given_counter);
|
|
|
|
if (given_counter == UINT64_MAX ||
|
|
|
|
given_counter < gcm_ctx->min_next_nonce) {
|
2017-05-24 21:54:35 +01:00
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_INVALID_NONCE);
|
2017-04-19 21:45:09 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-01-25 23:43:48 +00:00
|
|
|
gcm_ctx->min_next_nonce = given_counter + 1;
|
2017-04-19 21:45:09 +01:00
|
|
|
|
2017-06-06 20:29:48 +01:00
|
|
|
return aead_aes_gcm_seal_scatter(ctx, out, out_tag, out_tag_len,
|
|
|
|
max_out_tag_len, nonce, nonce_len, in,
|
2017-06-23 22:49:22 +01:00
|
|
|
in_len, extra_in, extra_in_len, ad, ad_len);
|
2017-04-19 21:45:09 +01:00
|
|
|
}
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_128_gcm_tls12) {
|
2017-05-03 21:23:37 +01:00
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 16;
|
2017-05-24 21:54:35 +01:00
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-05-03 21:23:37 +01:00
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-06-26 18:16:50 +01:00
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
out->init = aead_aes_gcm_tls12_init;
|
2018-06-18 22:04:41 +01:00
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
2017-06-06 20:29:48 +01:00
|
|
|
out->seal_scatter = aead_aes_gcm_tls12_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
2017-04-19 21:45:09 +01:00
|
|
|
}
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_256_gcm_tls12) {
|
2017-05-03 21:23:37 +01:00
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 32;
|
2017-05-24 21:54:35 +01:00
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-05-03 21:23:37 +01:00
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
2017-06-26 18:16:50 +01:00
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
2017-05-24 21:54:35 +01:00
|
|
|
out->init = aead_aes_gcm_tls12_init;
|
2018-06-18 22:04:41 +01:00
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
2017-06-06 20:29:48 +01:00
|
|
|
out->seal_scatter = aead_aes_gcm_tls12_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
2017-04-19 21:45:09 +01:00
|
|
|
}
|
2017-05-03 21:23:37 +01:00
|
|
|
|
2018-06-18 22:04:41 +01:00
|
|
|
struct aead_aes_gcm_tls13_ctx {
|
|
|
|
struct aead_aes_gcm_ctx gcm_ctx;
|
|
|
|
uint64_t min_next_nonce;
|
|
|
|
uint64_t mask;
|
|
|
|
uint8_t first;
|
|
|
|
};
|
|
|
|
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(sizeof(((EVP_AEAD_CTX *)NULL)->state) >=
|
|
|
|
sizeof(struct aead_aes_gcm_tls13_ctx),
|
|
|
|
"AEAD state is too small");
|
2018-10-16 20:11:51 +01:00
|
|
|
#if defined(__GNUC__) || defined(__clang__)
|
2018-10-24 23:08:00 +01:00
|
|
|
OPENSSL_STATIC_ASSERT(alignof(union evp_aead_ctx_st_state) >=
|
|
|
|
alignof(struct aead_aes_gcm_tls13_ctx),
|
|
|
|
"AEAD state has insufficient alignment");
|
2018-10-16 20:11:51 +01:00
|
|
|
#endif
|
|
|
|
|
2018-06-18 22:04:41 +01:00
|
|
|
static int aead_aes_gcm_tls13_init(EVP_AEAD_CTX *ctx, const uint8_t *key,
|
|
|
|
size_t key_len, size_t requested_tag_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_tls13_ctx *gcm_ctx =
|
|
|
|
(struct aead_aes_gcm_tls13_ctx *) &ctx->state;
|
2018-06-18 22:04:41 +01:00
|
|
|
|
|
|
|
gcm_ctx->min_next_nonce = 0;
|
|
|
|
gcm_ctx->first = 1;
|
|
|
|
|
|
|
|
size_t actual_tag_len;
|
|
|
|
if (!aead_aes_gcm_init_impl(&gcm_ctx->gcm_ctx, &actual_tag_len, key, key_len,
|
|
|
|
requested_tag_len)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx->tag_len = actual_tag_len;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int aead_aes_gcm_tls13_seal_scatter(
|
|
|
|
const EVP_AEAD_CTX *ctx, uint8_t *out, uint8_t *out_tag,
|
|
|
|
size_t *out_tag_len, size_t max_out_tag_len, const uint8_t *nonce,
|
|
|
|
size_t nonce_len, const uint8_t *in, size_t in_len, const uint8_t *extra_in,
|
|
|
|
size_t extra_in_len, const uint8_t *ad, size_t ad_len) {
|
2018-10-16 20:11:51 +01:00
|
|
|
struct aead_aes_gcm_tls13_ctx *gcm_ctx =
|
|
|
|
(struct aead_aes_gcm_tls13_ctx *) &ctx->state;
|
|
|
|
|
2018-06-18 22:04:41 +01:00
|
|
|
if (nonce_len != 12) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_UNSUPPORTED_NONCE_SIZE);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
// The given nonces must be strictly monotonically increasing. See
|
2018-08-15 04:26:39 +01:00
|
|
|
// https://tools.ietf.org/html/rfc8446#section-5.3 for details of the TLS 1.3
|
|
|
|
// nonce construction.
|
2018-06-18 22:04:41 +01:00
|
|
|
uint64_t given_counter;
|
|
|
|
OPENSSL_memcpy(&given_counter, nonce + nonce_len - sizeof(given_counter),
|
|
|
|
sizeof(given_counter));
|
|
|
|
given_counter = CRYPTO_bswap8(given_counter);
|
|
|
|
|
|
|
|
if (gcm_ctx->first) {
|
|
|
|
// In the first call the sequence number will be zero and therefore the
|
|
|
|
// given nonce will be 0 ^ mask = mask.
|
|
|
|
gcm_ctx->mask = given_counter;
|
|
|
|
gcm_ctx->first = 0;
|
|
|
|
}
|
|
|
|
given_counter ^= gcm_ctx->mask;
|
|
|
|
|
|
|
|
if (given_counter == UINT64_MAX ||
|
|
|
|
given_counter < gcm_ctx->min_next_nonce) {
|
|
|
|
OPENSSL_PUT_ERROR(CIPHER, CIPHER_R_INVALID_NONCE);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
gcm_ctx->min_next_nonce = given_counter + 1;
|
|
|
|
|
|
|
|
return aead_aes_gcm_seal_scatter(ctx, out, out_tag, out_tag_len,
|
|
|
|
max_out_tag_len, nonce, nonce_len, in,
|
|
|
|
in_len, extra_in, extra_in_len, ad, ad_len);
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_128_gcm_tls13) {
|
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 16;
|
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
|
|
|
out->init = aead_aes_gcm_tls13_init;
|
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
|
|
|
out->seal_scatter = aead_aes_gcm_tls13_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEFINE_METHOD_FUNCTION(EVP_AEAD, EVP_aead_aes_256_gcm_tls13) {
|
|
|
|
memset(out, 0, sizeof(EVP_AEAD));
|
|
|
|
|
|
|
|
out->key_len = 32;
|
|
|
|
out->nonce_len = 12;
|
|
|
|
out->overhead = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->max_tag_len = EVP_AEAD_AES_GCM_TAG_LEN;
|
|
|
|
out->seal_scatter_supports_extra_in = 1;
|
|
|
|
|
|
|
|
out->init = aead_aes_gcm_tls13_init;
|
|
|
|
out->cleanup = aead_aes_gcm_cleanup;
|
|
|
|
out->seal_scatter = aead_aes_gcm_tls13_seal_scatter;
|
|
|
|
out->open_gather = aead_aes_gcm_open_gather;
|
|
|
|
}
|
|
|
|
|
2014-08-20 19:19:54 +01:00
|
|
|
int EVP_has_aes_hardware(void) {
|
|
|
|
#if defined(OPENSSL_X86) || defined(OPENSSL_X86_64)
|
2018-05-14 22:28:36 +01:00
|
|
|
return hwaes_capable() && crypto_gcm_clmul_enabled();
|
2015-01-09 23:44:37 +00:00
|
|
|
#elif defined(OPENSSL_ARM) || defined(OPENSSL_AARCH64)
|
2015-10-16 23:46:46 +01:00
|
|
|
return hwaes_capable() && CRYPTO_is_ARMv8_PMULL_capable();
|
2014-08-20 19:19:54 +01:00
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
}
|
2018-05-06 07:05:19 +01:00
|
|
|
|
|
|
|
OPENSSL_MSVC_PRAGMA(warning(pop))
|