|
@ -20,23 +20,25 @@ |
|
|
#include "cc20.h" |
|
|
#include "cc20.h" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#if defined (HAVE_OPENSSL_1_1) // openSSL 1.1 ---------------------------------------------
|
|
|
#if defined (HAVE_OPENSSL_1_1) // openSSL 1.1 ---------------------------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/* get any erorr message out of openssl
|
|
|
// get any erorr message out of openssl
|
|
|
taken from https://en.wikibooks.org/wiki/OpenSSL/Error_handling */
|
|
|
// taken from https://en.wikibooks.org/wiki/OpenSSL/Error_handling
|
|
|
static char *openssl_err_as_string (void) { |
|
|
static char *openssl_err_as_string (void) { |
|
|
BIO *bio = BIO_new (BIO_s_mem ()); |
|
|
|
|
|
ERR_print_errors (bio); |
|
|
|
|
|
char *buf = NULL; |
|
|
|
|
|
size_t len = BIO_get_mem_data (bio, &buf); |
|
|
|
|
|
char *ret = (char *) calloc (1, 1 + len); |
|
|
|
|
|
|
|
|
|
|
|
if(ret) |
|
|
BIO *bio = BIO_new(BIO_s_mem()); |
|
|
memcpy (ret, buf, len); |
|
|
ERR_print_errors(bio); |
|
|
|
|
|
char *buf = NULL; |
|
|
|
|
|
size_t len = BIO_get_mem_data(bio, &buf); |
|
|
|
|
|
char *ret = (char *)calloc(1, 1 + len); |
|
|
|
|
|
|
|
|
BIO_free (bio); |
|
|
if(ret) |
|
|
return ret; |
|
|
memcpy(ret, buf, len); |
|
|
|
|
|
|
|
|
|
|
|
BIO_free(bio); |
|
|
|
|
|
|
|
|
|
|
|
return ret; |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
@ -47,35 +49,35 @@ int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
int evp_len; |
|
|
int evp_len; |
|
|
int evp_ciphertext_len; |
|
|
int evp_ciphertext_len; |
|
|
|
|
|
|
|
|
if(1 == EVP_EncryptInit_ex(ctx->ctx, ctx->cipher, NULL, ctx->key, iv)) { |
|
|
if(1 == EVP_EncryptInit_ex(ctx->ctx, ctx->cipher, NULL, ctx->key, iv)) { |
|
|
if(1 == EVP_CIPHER_CTX_set_padding(ctx->ctx, 0)) { |
|
|
if(1 == EVP_CIPHER_CTX_set_padding(ctx->ctx, 0)) { |
|
|
if(1 == EVP_EncryptUpdate(ctx->ctx, out, &evp_len, in, in_len)) { |
|
|
if(1 == EVP_EncryptUpdate(ctx->ctx, out, &evp_len, in, in_len)) { |
|
|
evp_ciphertext_len = evp_len; |
|
|
evp_ciphertext_len = evp_len; |
|
|
if(1 == EVP_EncryptFinal_ex(ctx->ctx, out + evp_len, &evp_len)) { |
|
|
if(1 == EVP_EncryptFinal_ex(ctx->ctx, out + evp_len, &evp_len)) { |
|
|
evp_ciphertext_len += evp_len; |
|
|
evp_ciphertext_len += evp_len; |
|
|
if(evp_ciphertext_len != in_len) |
|
|
if(evp_ciphertext_len != in_len) |
|
|
traceEvent(TRACE_ERROR, "cc20_crypt openssl encryption: encrypted %u bytes where %u were expected", |
|
|
traceEvent(TRACE_ERROR, "cc20_crypt openssl encryption: encrypted %u bytes where %u were expected", |
|
|
evp_ciphertext_len, in_len); |
|
|
evp_ciphertext_len, in_len); |
|
|
|
|
|
} else |
|
|
|
|
|
traceEvent(TRACE_ERROR, "cc20_crypt openssl final encryption: %s", |
|
|
|
|
|
openssl_err_as_string()); |
|
|
|
|
|
} else |
|
|
|
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl encrpytion: %s", |
|
|
|
|
|
openssl_err_as_string()); |
|
|
} else |
|
|
} else |
|
|
traceEvent(TRACE_ERROR, "cc20_crypt openssl final encryption: %s", |
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl padding setup: %s", |
|
|
openssl_err_as_string()); |
|
|
openssl_err_as_string()); |
|
|
} else |
|
|
|
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl encrpytion: %s", |
|
|
|
|
|
openssl_err_as_string()); |
|
|
|
|
|
} else |
|
|
} else |
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl padding setup: %s", |
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl init: %s", |
|
|
openssl_err_as_string()); |
|
|
openssl_err_as_string()); |
|
|
} else |
|
|
|
|
|
traceEvent(TRACE_ERROR, "cc20_encrypt openssl init: %s", |
|
|
|
|
|
openssl_err_as_string()); |
|
|
|
|
|
|
|
|
|
|
|
EVP_CIPHER_CTX_reset(ctx->ctx); |
|
|
EVP_CIPHER_CTX_reset(ctx->ctx); |
|
|
|
|
|
|
|
|
return 0; |
|
|
return 0; |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#elif defined (__SSE2__) // SSE ----------------------------------------------------------
|
|
|
#elif defined (__SSE2__) // SSE2 ---------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// taken (and heavily modified and enhanced) from
|
|
|
// taken (and heavily modified and enhanced) from
|
|
@ -93,152 +95,154 @@ int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
#define TWO _mm_setr_epi32(2, 0, 0, 0) |
|
|
#define TWO _mm_setr_epi32(2, 0, 0, 0) |
|
|
|
|
|
|
|
|
#if defined (__SSSE3__) // --- SSSE3
|
|
|
#if defined (__SSSE3__) // --- SSSE3
|
|
|
|
|
|
|
|
|
#define L8 _mm_set_epi32(0x0e0d0c0fL, 0x0a09080bL, 0x06050407L, 0x02010003L) |
|
|
#define L8 _mm_set_epi32(0x0e0d0c0fL, 0x0a09080bL, 0x06050407L, 0x02010003L) |
|
|
#define L16 _mm_set_epi32(0x0d0c0f0eL, 0x09080b0aL, 0x05040706L, 0x01000302L) |
|
|
#define L16 _mm_set_epi32(0x0d0c0f0eL, 0x09080b0aL, 0x05040706L, 0x01000302L) |
|
|
#define ROL8(X) ( _mm_shuffle_epi8(X, L8)) /* SSSE 3 */ |
|
|
#define ROL8(X) ( _mm_shuffle_epi8(X, L8)) /* SSSE 3 */ |
|
|
#define ROL16(X) ( _mm_shuffle_epi8(X, L16)) /* SSSE 3 */ |
|
|
#define ROL16(X) ( _mm_shuffle_epi8(X, L16)) /* SSSE 3 */ |
|
|
|
|
|
|
|
|
#else // --- regular SSE2 ----------
|
|
|
#else // --- regular SSE2 ----------
|
|
|
|
|
|
|
|
|
#define ROL8(X) ROL(X,8) |
|
|
#define ROL8(X) ROL(X,8) |
|
|
#define ROL16(X) ROL(X,16) |
|
|
#define ROL16(X) ROL(X,16) |
|
|
|
|
|
|
|
|
#endif // --------------------------
|
|
|
#endif // --------------------------
|
|
|
|
|
|
|
|
|
#define CC20_PERMUTE_ROWS(A,B,C,D) \ |
|
|
|
|
|
B = _mm_shuffle_epi32(B, _MM_SHUFFLE(0, 3, 2, 1)); \ |
|
|
|
|
|
C = _mm_shuffle_epi32(C, _MM_SHUFFLE(1, 0, 3, 2)); \ |
|
|
|
|
|
D = _mm_shuffle_epi32(D, _MM_SHUFFLE(2, 1, 0, 3)) |
|
|
|
|
|
|
|
|
|
|
|
#define CC20_PERMUTE_ROWS_INV(A,B,C,D) \ |
|
|
|
|
|
B = _mm_shuffle_epi32(B, _MM_SHUFFLE(2, 1, 0, 3)); \ |
|
|
|
|
|
C = _mm_shuffle_epi32(C, _MM_SHUFFLE(1, 0, 3, 2)); \ |
|
|
|
|
|
D = _mm_shuffle_epi32(D, _MM_SHUFFLE(0, 3, 2, 1)) |
|
|
|
|
|
|
|
|
|
|
|
#define CC20_ODD_ROUND(A,B,C,D) \ |
|
|
|
|
|
/* odd round */ \ |
|
|
|
|
|
A = ADD(A, B); D = ROL16(XOR(D, A)); \ |
|
|
|
|
|
C = ADD(C, D); B = ROL(XOR(B, C), 12); \ |
|
|
|
|
|
A = ADD(A, B); D = ROL8(XOR(D, A)); \ |
|
|
|
|
|
C = ADD(C, D); B = ROL(XOR(B, C), 7) |
|
|
|
|
|
|
|
|
|
|
|
#define CC20_EVEN_ROUND(A,B,C,D) \ |
|
|
|
|
|
CC20_PERMUTE_ROWS (A, B, C, D); \ |
|
|
|
|
|
CC20_ODD_ROUND (A, B, C, D); \ |
|
|
|
|
|
CC20_PERMUTE_ROWS_INV(A, B, C, D) |
|
|
|
|
|
|
|
|
|
|
|
#define CC20_DOUBLE_ROUND(A,B,C,D) \ |
|
|
|
|
|
CC20_ODD_ROUND (A, B, C, D); \ |
|
|
|
|
|
CC20_EVEN_ROUND(A, B, C, D) |
|
|
|
|
|
|
|
|
|
|
|
#define STOREXOR(O,I,X) \ |
|
|
|
|
|
_mm_storeu_si128 ((__m128i*)O, \ |
|
|
|
|
|
_mm_xor_si128 (_mm_loadu_si128((__m128i*)I), X)); \ |
|
|
|
|
|
I += 16; O += 16 \ |
|
|
|
|
|
|
|
|
|
|
|
int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
#define CC20_PERMUTE_ROWS(A,B,C,D) \ |
|
|
const unsigned char *iv, cc20_context_t *ctx) { |
|
|
B = _mm_shuffle_epi32(B, _MM_SHUFFLE(0, 3, 2, 1)); \ |
|
|
|
|
|
C = _mm_shuffle_epi32(C, _MM_SHUFFLE(1, 0, 3, 2)); \ |
|
|
|
|
|
D = _mm_shuffle_epi32(D, _MM_SHUFFLE(2, 1, 0, 3)) |
|
|
|
|
|
|
|
|
__m128i a, b, c, d, k0, k1, k2, k3, k4, k5, k6, k7; |
|
|
#define CC20_PERMUTE_ROWS_INV(A,B,C,D) \ |
|
|
|
|
|
B = _mm_shuffle_epi32(B, _MM_SHUFFLE(2, 1, 0, 3)); \ |
|
|
|
|
|
C = _mm_shuffle_epi32(C, _MM_SHUFFLE(1, 0, 3, 2)); \ |
|
|
|
|
|
D = _mm_shuffle_epi32(D, _MM_SHUFFLE(0, 3, 2, 1)) |
|
|
|
|
|
|
|
|
uint8_t *keystream8 = (uint8_t*)ctx->keystream32; |
|
|
#define CC20_ODD_ROUND(A,B,C,D) \ |
|
|
|
|
|
/* odd round */ \ |
|
|
|
|
|
A = ADD(A, B); D = ROL16(XOR(D, A)); \ |
|
|
|
|
|
C = ADD(C, D); B = ROL(XOR(B, C), 12); \ |
|
|
|
|
|
A = ADD(A, B); D = ROL8(XOR(D, A)); \ |
|
|
|
|
|
C = ADD(C, D); B = ROL(XOR(B, C), 7) |
|
|
|
|
|
|
|
|
const uint8_t *magic_constant = (uint8_t*)"expand 32-byte k"; |
|
|
#define CC20_EVEN_ROUND(A,B,C,D) \ |
|
|
|
|
|
CC20_PERMUTE_ROWS (A, B, C, D); \ |
|
|
|
|
|
CC20_ODD_ROUND (A, B, C, D); \ |
|
|
|
|
|
CC20_PERMUTE_ROWS_INV(A, B, C, D) |
|
|
|
|
|
|
|
|
a = _mm_loadu_si128 ((__m128i*)magic_constant); |
|
|
#define CC20_DOUBLE_ROUND(A,B,C,D) \ |
|
|
b = _mm_loadu_si128 ((__m128i*)(ctx->key)); |
|
|
CC20_ODD_ROUND (A, B, C, D); \ |
|
|
c = _mm_loadu_si128 ( (__m128i*)((ctx->key)+16)); |
|
|
CC20_EVEN_ROUND(A, B, C, D) |
|
|
d = _mm_loadu_si128 ((__m128i*)iv); |
|
|
|
|
|
|
|
|
|
|
|
while (in_len >= 128) { |
|
|
#define STOREXOR(O,I,X) \ |
|
|
|
|
|
_mm_storeu_si128((__m128i*)O, \ |
|
|
|
|
|
_mm_xor_si128(_mm_loadu_si128((__m128i*)I), X)); \ |
|
|
|
|
|
I += 16; O += 16 \ |
|
|
|
|
|
|
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
|
|
|
k4 = a; k5 = b; k6 = c; k7 = ADD(d, ONE); |
|
|
|
|
|
|
|
|
|
|
|
// 10 double rounds -- in parallel to make better use of all 8 SSE registers
|
|
|
int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
const unsigned char *iv, cc20_context_t *ctx) { |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
|
|
|
|
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
__m128i a, b, c, d, k0, k1, k2, k3, k4, k5, k6, k7; |
|
|
k4 = ADD(k4, a); k5 = ADD(k5, b); k6 = ADD(k6, c); k7 = ADD(k7, d); k7 = ADD(k7, ONE); |
|
|
|
|
|
|
|
|
|
|
|
STOREXOR(out, in, k0); STOREXOR(out, in, k1); STOREXOR(out, in, k2); STOREXOR(out, in, k3); |
|
|
uint8_t *keystream8 = (uint8_t*)ctx->keystream32; |
|
|
STOREXOR(out, in, k4); STOREXOR(out, in, k5); STOREXOR(out, in, k6); STOREXOR(out, in, k7); |
|
|
|
|
|
|
|
|
|
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
const uint8_t *magic_constant = (uint8_t*)"expand 32-byte k"; |
|
|
d = ADD(d, TWO); |
|
|
|
|
|
|
|
|
|
|
|
in_len -= 128; |
|
|
a = _mm_loadu_si128((__m128i*)magic_constant); |
|
|
} |
|
|
b = _mm_loadu_si128((__m128i*)(ctx->key)); |
|
|
|
|
|
c = _mm_loadu_si128( (__m128i*)((ctx->key)+16)); |
|
|
|
|
|
d = _mm_loadu_si128((__m128i*)iv); |
|
|
|
|
|
|
|
|
if (in_len >= 64) { |
|
|
while(in_len >= 128) { |
|
|
|
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
|
|
|
k4 = a; k5 = b; k6 = c; k7 = ADD(d, ONE); |
|
|
|
|
|
|
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
// 10 double rounds -- two in parallel to make better use of all 8 SSE registers
|
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); CC20_DOUBLE_ROUND(k4, k5, k6, k7); |
|
|
|
|
|
|
|
|
// 10 double rounds
|
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
k4 = ADD(k4, a); k5 = ADD(k5, b); k6 = ADD(k6, c); k7 = ADD(k7, d); k7 = ADD(k7, ONE); |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
|
|
|
|
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
STOREXOR(out, in, k0); STOREXOR(out, in, k1); STOREXOR(out, in, k2); STOREXOR(out, in, k3); |
|
|
|
|
|
STOREXOR(out, in, k4); STOREXOR(out, in, k5); STOREXOR(out, in, k6); STOREXOR(out, in, k7); |
|
|
|
|
|
|
|
|
STOREXOR(out, in, k0); STOREXOR(out, in, k1); STOREXOR(out, in, k2); STOREXOR(out, in, k3); |
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
|
|
|
d = ADD(d, TWO); |
|
|
|
|
|
|
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
in_len -= 128; |
|
|
d = ADD(d, ONE); |
|
|
} |
|
|
|
|
|
|
|
|
in_len -= 64; |
|
|
if(in_len >= 64) { |
|
|
} |
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
|
|
|
|
|
|
if (in_len) { |
|
|
// 10 double rounds
|
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
|
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
|
|
|
|
|
|
// 10 double rounds
|
|
|
STOREXOR(out, in, k0); STOREXOR(out, in, k1); STOREXOR(out, in, k2); STOREXOR(out, in, k3); |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
d = ADD(d, ONE); |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
in_len -= 64; |
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
|
|
|
|
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
|
|
|
|
|
|
|
|
|
_mm_storeu_si128 ((__m128i*)&(ctx->keystream32[ 0]), k0); |
|
|
|
|
|
_mm_storeu_si128 ((__m128i*)&(ctx->keystream32[ 4]), k1); |
|
|
|
|
|
_mm_storeu_si128 ((__m128i*)&(ctx->keystream32[ 8]), k2); |
|
|
|
|
|
_mm_storeu_si128 ((__m128i*)&(ctx->keystream32[12]), k3); |
|
|
|
|
|
|
|
|
|
|
|
// keep in mind that out and in got increased inside the last loop
|
|
|
|
|
|
// and point to current position now
|
|
|
|
|
|
while(in_len > 0) { |
|
|
|
|
|
in_len--; |
|
|
|
|
|
out[in_len] = in[in_len] ^ keystream8[in_len]; |
|
|
|
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
} |
|
|
if(in_len) { |
|
|
|
|
|
k0 = a; k1 = b; k2 = c; k3 = d; |
|
|
|
|
|
|
|
|
|
|
|
// 10 double rounds
|
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
CC20_DOUBLE_ROUND(k0, k1, k2, k3); |
|
|
|
|
|
|
|
|
|
|
|
k0 = ADD(k0, a); k1 = ADD(k1, b); k2 = ADD(k2, c); k3 = ADD(k3, d); |
|
|
|
|
|
|
|
|
|
|
|
_mm_storeu_si128((__m128i*)&(ctx->keystream32[ 0]), k0); |
|
|
|
|
|
_mm_storeu_si128((__m128i*)&(ctx->keystream32[ 4]), k1); |
|
|
|
|
|
_mm_storeu_si128((__m128i*)&(ctx->keystream32[ 8]), k2); |
|
|
|
|
|
_mm_storeu_si128((__m128i*)&(ctx->keystream32[12]), k3); |
|
|
|
|
|
|
|
|
|
|
|
// keep in mind that out and in got increased inside the last loop
|
|
|
|
|
|
// and point to current position now
|
|
|
|
|
|
while(in_len > 0) { |
|
|
|
|
|
in_len--; |
|
|
|
|
|
out[in_len] = in[in_len] ^ keystream8[in_len]; |
|
|
|
|
|
} |
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
return(0); |
|
|
return(0); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#else // plain C --------------------------------------------------------------------------
|
|
|
#else // plain C --------------------------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// taken (and modified) from https://github.com/Ginurx/chacha20-c (public domain)
|
|
|
// taken (and modified) from https://github.com/Ginurx/chacha20-c (public domain)
|
|
@ -246,170 +250,172 @@ int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
|
|
|
|
|
|
static void cc20_init_block(cc20_context_t *ctx, const uint8_t nonce[]) { |
|
|
static void cc20_init_block(cc20_context_t *ctx, const uint8_t nonce[]) { |
|
|
|
|
|
|
|
|
const uint8_t *magic_constant = (uint8_t*)"expand 32-byte k"; |
|
|
const uint8_t *magic_constant = (uint8_t*)"expand 32-byte k"; |
|
|
|
|
|
|
|
|
memcpy(&(ctx->state[ 0]), magic_constant, 16); |
|
|
memcpy(&(ctx->state[ 0]), magic_constant, 16); |
|
|
memcpy(&(ctx->state[ 4]), ctx->key, CC20_KEY_BYTES); |
|
|
memcpy(&(ctx->state[ 4]), ctx->key, CC20_KEY_BYTES); |
|
|
memcpy(&(ctx->state[12]), nonce, CC20_IV_SIZE); |
|
|
memcpy(&(ctx->state[12]), nonce, CC20_IV_SIZE); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#define ROL32(x,r) (((x)<<(r))|((x)>>(32-(r)))) |
|
|
#define ROL32(x,r) (((x)<<(r))|((x)>>(32-(r)))) |
|
|
#define CC20_QUARTERROUND(x, a, b, c, d) \ |
|
|
|
|
|
x[a] += x[b]; x[d] = ROL32(x[d] ^ x[a], 16); \ |
|
|
#define CC20_QUARTERROUND(x, a, b, c, d) \ |
|
|
x[c] += x[d]; x[b] = ROL32(x[b] ^ x[c], 12); \ |
|
|
x[a] += x[b]; x[d] = ROL32(x[d] ^ x[a], 16); \ |
|
|
x[a] += x[b]; x[d] = ROL32(x[d] ^ x[a], 8); \ |
|
|
x[c] += x[d]; x[b] = ROL32(x[b] ^ x[c], 12); \ |
|
|
x[c] += x[d]; x[b] = ROL32(x[b] ^ x[c], 7) |
|
|
x[a] += x[b]; x[d] = ROL32(x[d] ^ x[a], 8); \ |
|
|
#define CC20_DOUBLE_ROUND(s) \ |
|
|
x[c] += x[d]; x[b] = ROL32(x[b] ^ x[c], 7) |
|
|
/* odd round */ \ |
|
|
|
|
|
CC20_QUARTERROUND(s, 0, 4, 8, 12); \ |
|
|
#define CC20_DOUBLE_ROUND(s) \ |
|
|
CC20_QUARTERROUND(s, 1, 5, 9, 13); \ |
|
|
/* odd round */ \ |
|
|
CC20_QUARTERROUND(s, 2, 6, 10, 14); \ |
|
|
CC20_QUARTERROUND(s, 0, 4, 8, 12); \ |
|
|
CC20_QUARTERROUND(s, 3, 7, 11, 15); \ |
|
|
CC20_QUARTERROUND(s, 1, 5, 9, 13); \ |
|
|
/* even round */ \ |
|
|
CC20_QUARTERROUND(s, 2, 6, 10, 14); \ |
|
|
CC20_QUARTERROUND(s, 0, 5, 10, 15); \ |
|
|
CC20_QUARTERROUND(s, 3, 7, 11, 15); \ |
|
|
CC20_QUARTERROUND(s, 1, 6, 11, 12); \ |
|
|
/* even round */ \ |
|
|
CC20_QUARTERROUND(s, 2, 7, 8, 13); \ |
|
|
CC20_QUARTERROUND(s, 0, 5, 10, 15); \ |
|
|
CC20_QUARTERROUND(s, 3, 4, 9, 14) |
|
|
CC20_QUARTERROUND(s, 1, 6, 11, 12); \ |
|
|
|
|
|
CC20_QUARTERROUND(s, 2, 7, 8, 13); \ |
|
|
|
|
|
CC20_QUARTERROUND(s, 3, 4, 9, 14) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static void cc20_block_next(cc20_context_t *ctx) { |
|
|
static void cc20_block_next(cc20_context_t *ctx) { |
|
|
|
|
|
|
|
|
uint32_t *counter = ctx->state + 12; |
|
|
uint32_t *counter = ctx->state + 12; |
|
|
|
|
|
|
|
|
ctx->keystream32[ 0] = ctx->state[ 0]; |
|
|
ctx->keystream32[ 0] = ctx->state[ 0]; |
|
|
ctx->keystream32[ 1] = ctx->state[ 1]; |
|
|
ctx->keystream32[ 1] = ctx->state[ 1]; |
|
|
ctx->keystream32[ 2] = ctx->state[ 2]; |
|
|
ctx->keystream32[ 2] = ctx->state[ 2]; |
|
|
ctx->keystream32[ 3] = ctx->state[ 3]; |
|
|
ctx->keystream32[ 3] = ctx->state[ 3]; |
|
|
ctx->keystream32[ 4] = ctx->state[ 4]; |
|
|
ctx->keystream32[ 4] = ctx->state[ 4]; |
|
|
ctx->keystream32[ 5] = ctx->state[ 5]; |
|
|
ctx->keystream32[ 5] = ctx->state[ 5]; |
|
|
ctx->keystream32[ 6] = ctx->state[ 6]; |
|
|
ctx->keystream32[ 6] = ctx->state[ 6]; |
|
|
ctx->keystream32[ 7] = ctx->state[ 7]; |
|
|
ctx->keystream32[ 7] = ctx->state[ 7]; |
|
|
ctx->keystream32[ 8] = ctx->state[ 8]; |
|
|
ctx->keystream32[ 8] = ctx->state[ 8]; |
|
|
ctx->keystream32[ 9] = ctx->state[ 9]; |
|
|
ctx->keystream32[ 9] = ctx->state[ 9]; |
|
|
ctx->keystream32[10] = ctx->state[10]; |
|
|
ctx->keystream32[10] = ctx->state[10]; |
|
|
ctx->keystream32[11] = ctx->state[11]; |
|
|
ctx->keystream32[11] = ctx->state[11]; |
|
|
ctx->keystream32[12] = ctx->state[12]; |
|
|
ctx->keystream32[12] = ctx->state[12]; |
|
|
ctx->keystream32[13] = ctx->state[13]; |
|
|
ctx->keystream32[13] = ctx->state[13]; |
|
|
ctx->keystream32[14] = ctx->state[14]; |
|
|
ctx->keystream32[14] = ctx->state[14]; |
|
|
ctx->keystream32[15] = ctx->state[15]; |
|
|
ctx->keystream32[15] = ctx->state[15]; |
|
|
|
|
|
|
|
|
// 10 double rounds
|
|
|
// 10 double rounds
|
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
CC20_DOUBLE_ROUND(ctx->keystream32); |
|
|
|
|
|
|
|
|
ctx->keystream32[ 0] += ctx->state[ 0]; |
|
|
ctx->keystream32[ 0] += ctx->state[ 0]; |
|
|
ctx->keystream32[ 1] += ctx->state[ 1]; |
|
|
ctx->keystream32[ 1] += ctx->state[ 1]; |
|
|
ctx->keystream32[ 2] += ctx->state[ 2]; |
|
|
ctx->keystream32[ 2] += ctx->state[ 2]; |
|
|
ctx->keystream32[ 3] += ctx->state[ 3]; |
|
|
ctx->keystream32[ 3] += ctx->state[ 3]; |
|
|
ctx->keystream32[ 4] += ctx->state[ 4]; |
|
|
ctx->keystream32[ 4] += ctx->state[ 4]; |
|
|
ctx->keystream32[ 5] += ctx->state[ 5]; |
|
|
ctx->keystream32[ 5] += ctx->state[ 5]; |
|
|
ctx->keystream32[ 6] += ctx->state[ 6]; |
|
|
ctx->keystream32[ 6] += ctx->state[ 6]; |
|
|
ctx->keystream32[ 7] += ctx->state[ 7]; |
|
|
ctx->keystream32[ 7] += ctx->state[ 7]; |
|
|
ctx->keystream32[ 8] += ctx->state[ 8]; |
|
|
ctx->keystream32[ 8] += ctx->state[ 8]; |
|
|
ctx->keystream32[ 9] += ctx->state[ 9]; |
|
|
ctx->keystream32[ 9] += ctx->state[ 9]; |
|
|
ctx->keystream32[10] += ctx->state[10]; |
|
|
ctx->keystream32[10] += ctx->state[10]; |
|
|
ctx->keystream32[11] += ctx->state[11]; |
|
|
ctx->keystream32[11] += ctx->state[11]; |
|
|
ctx->keystream32[12] += ctx->state[12]; |
|
|
ctx->keystream32[12] += ctx->state[12]; |
|
|
ctx->keystream32[13] += ctx->state[13]; |
|
|
ctx->keystream32[13] += ctx->state[13]; |
|
|
ctx->keystream32[14] += ctx->state[14]; |
|
|
ctx->keystream32[14] += ctx->state[14]; |
|
|
ctx->keystream32[15] += ctx->state[15]; |
|
|
ctx->keystream32[15] += ctx->state[15]; |
|
|
|
|
|
|
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
// increment counter, make sure it is and stays little endian in memory
|
|
|
*counter = htole32(le32toh(*counter)+1); |
|
|
*counter = htole32(le32toh(*counter)+1); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static void cc20_init_context(cc20_context_t *ctx, const uint8_t *nonce) { |
|
|
static void cc20_init_context(cc20_context_t *ctx, const uint8_t *nonce) { |
|
|
|
|
|
|
|
|
cc20_init_block(ctx, nonce); |
|
|
cc20_init_block(ctx, nonce); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
int cc20_crypt (unsigned char *out, const unsigned char *in, size_t in_len, |
|
|
const unsigned char *iv, cc20_context_t *ctx) { |
|
|
const unsigned char *iv, cc20_context_t *ctx) { |
|
|
|
|
|
|
|
|
uint8_t *keystream8 = (uint8_t*)ctx->keystream32; |
|
|
uint8_t *keystream8 = (uint8_t*)ctx->keystream32; |
|
|
uint32_t * in_p = (uint32_t*)in; |
|
|
uint32_t * in_p = (uint32_t*)in; |
|
|
uint32_t * out_p = (uint32_t*)out; |
|
|
uint32_t * out_p = (uint32_t*)out; |
|
|
size_t tmp_len = in_len; |
|
|
size_t tmp_len = in_len; |
|
|
|
|
|
|
|
|
cc20_init_context(ctx, iv); |
|
|
cc20_init_context(ctx, iv); |
|
|
|
|
|
|
|
|
while(in_len >= 64) { |
|
|
while(in_len >= 64) { |
|
|
|
|
|
cc20_block_next(ctx); |
|
|
cc20_block_next(ctx); |
|
|
|
|
|
|
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 0]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 0]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 1]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 1]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 2]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 2]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 3]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 3]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 4]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 4]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 5]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 5]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 6]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 6]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 7]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 7]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 8]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 8]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 9]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[ 9]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[10]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[10]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[11]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[11]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[12]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[12]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[13]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[13]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[14]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[14]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[15]; in_p++; out_p++; |
|
|
*(uint32_t*)out_p = *(uint32_t*)in_p ^ ctx->keystream32[15]; in_p++; out_p++; |
|
|
|
|
|
in_len -= 64; |
|
|
in_len -= 64; |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
if(in_len > 0) { |
|
|
|
|
|
|
|
|
|
|
|
cc20_block_next(ctx); |
|
|
|
|
|
|
|
|
|
|
|
tmp_len -= in_len; |
|
|
|
|
|
while(in_len > 0) { |
|
|
|
|
|
out[tmp_len] = in[tmp_len] ^ keystream8[tmp_len%64]; |
|
|
|
|
|
tmp_len++; |
|
|
|
|
|
in_len--; |
|
|
|
|
|
} |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
return(0); |
|
|
if(in_len > 0) { |
|
|
|
|
|
cc20_block_next(ctx); |
|
|
|
|
|
|
|
|
|
|
|
tmp_len -= in_len; |
|
|
|
|
|
while(in_len > 0) { |
|
|
|
|
|
out[tmp_len] = in[tmp_len] ^ keystream8[tmp_len%64]; |
|
|
|
|
|
tmp_len++; |
|
|
|
|
|
in_len--; |
|
|
|
|
|
} |
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
return(0); |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#endif // openSSL 1.1, plain C ------------------------------------------------------------
|
|
|
#endif // openSSL 1.1, plain C ------------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
int cc20_init (const unsigned char *key, cc20_context_t **ctx) { |
|
|
int cc20_init (const unsigned char *key, cc20_context_t **ctx) { |
|
|
|
|
|
|
|
|
// allocate context...
|
|
|
// allocate context...
|
|
|
*ctx = (cc20_context_t*) calloc(1, sizeof(cc20_context_t)); |
|
|
*ctx = (cc20_context_t*)calloc(1, sizeof(cc20_context_t)); |
|
|
if (!(*ctx)) |
|
|
if(!(*ctx)) |
|
|
return -1; |
|
|
return -1; |
|
|
#if defined (HAVE_OPENSSL_1_1) |
|
|
#if defined (HAVE_OPENSSL_1_1) |
|
|
if(!((*ctx)->ctx = EVP_CIPHER_CTX_new())) { |
|
|
if(!((*ctx)->ctx = EVP_CIPHER_CTX_new())) { |
|
|
traceEvent(TRACE_ERROR, "cc20_init openssl's evp_* encryption context creation failed: %s", |
|
|
traceEvent(TRACE_ERROR, "cc20_init openssl's evp_* encryption context creation failed: %s", |
|
|
openssl_err_as_string()); |
|
|
openssl_err_as_string()); |
|
|
return -1; |
|
|
return -1; |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
(*ctx)->cipher = EVP_chacha20(); |
|
|
(*ctx)->cipher = EVP_chacha20(); |
|
|
#endif |
|
|
#endif |
|
|
memcpy((*ctx)->key, key, CC20_KEY_BYTES); |
|
|
memcpy((*ctx)->key, key, CC20_KEY_BYTES); |
|
|
|
|
|
|
|
|
return 0; |
|
|
return 0; |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
int cc20_deinit (cc20_context_t *ctx) { |
|
|
int cc20_deinit (cc20_context_t *ctx) { |
|
|
|
|
|
|
|
|
#if defined (HAVE_OPENSSL_1_1) |
|
|
#if defined (HAVE_OPENSSL_1_1) |
|
|
if (ctx->ctx) EVP_CIPHER_CTX_free(ctx->ctx); |
|
|
if(ctx->ctx) EVP_CIPHER_CTX_free(ctx->ctx); |
|
|
#endif |
|
|
#endif |
|
|
return 0; |
|
|
return 0; |
|
|
} |
|
|
} |
|
|