diff --git a/crypto/evp/e_aes_cbc_hmac_sha1.c b/crypto/evp/e_aes_cbc_hmac_sha1.c index 6ece66f27a..b2d797dba2 100644 --- a/crypto/evp/e_aes_cbc_hmac_sha1.c +++ b/crypto/evp/e_aes_cbc_hmac_sha1.c @@ -210,6 +210,8 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA1 *key, u8 *IVs; #if defined(BSWAP8) u64 seqnum; +#else + unsigned int carry,j; #endif if (RAND_bytes((IVs=blocks[0].c),16*x4)<=0) /* ask for IVs in bulk */ @@ -257,13 +259,9 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA1 *key, #if defined(BSWAP8) blocks[i].q[0] = BSWAP8(seqnum+i); #else - blocks[i].c[7] += ((u8*)key->md.data)[7]+i; - if (blocks[i].c[7] < i) { - int j; - - for (j=6;j>=0;j--) { - if (blocks[i].c[j]=((u8*)key->md.data)[j]+1) break; - } + for (carry=i,j=8;j--;) { + blocks[i].c[j] = ((u8*)key->md.data)[j]+carry; + carry = (blocks[i].c[j]-carry)>>(sizeof(carry)*8-1); } #endif blocks[i].c[8] = ((u8*)key->md.data)[8]; @@ -331,10 +329,10 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA1 *key, len += 64+13; /* 64 is HMAC header */ len *= 8; /* convert to bits */ if (off<(64-8)) { - blocks[i].d[15] = BSWAP4(len); + PUTU32(blocks[i].c+60,len); edges[i].blocks = 1; } else { - blocks[i].d[31] = BSWAP4(len); + PUTU32(blocks[i].c+124,len); edges[i].blocks = 2; } edges[i].ptr = blocks[i].c; @@ -345,13 +343,13 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA1 *key, memset(blocks,0,sizeof(blocks)); for (i=0;iA[i]); ctx->A[i] = key->tail.h0; - blocks[i].d[1] = BSWAP4(ctx->B[i]); ctx->B[i] = key->tail.h1; - blocks[i].d[2] = BSWAP4(ctx->C[i]); ctx->C[i] = key->tail.h2; - blocks[i].d[3] = BSWAP4(ctx->D[i]); ctx->D[i] = key->tail.h3; - blocks[i].d[4] = BSWAP4(ctx->E[i]); ctx->E[i] = key->tail.h4; + PUTU32(blocks[i].c+0,ctx->A[i]); ctx->A[i] = key->tail.h0; + PUTU32(blocks[i].c+4,ctx->B[i]); ctx->B[i] = key->tail.h1; + PUTU32(blocks[i].c+8,ctx->C[i]); ctx->C[i] = key->tail.h2; + PUTU32(blocks[i].c+12,ctx->D[i]); ctx->D[i] = key->tail.h3; + PUTU32(blocks[i].c+16,ctx->E[i]); ctx->E[i] = key->tail.h4; blocks[i].c[20] = 0x80; - blocks[i].d[15] = BSWAP4((64+20)*8); + PUTU32(blocks[i].c+60,(64+20)*8); edges[i].ptr = blocks[i].c; edges[i].blocks = 1; } @@ -369,11 +367,11 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA1 *key, out += 5+16+len; /* write MAC */ - ((u32 *)out)[0] = BSWAP4(ctx->A[i]); - ((u32 *)out)[1] = BSWAP4(ctx->B[i]); - ((u32 *)out)[2] = BSWAP4(ctx->C[i]); - ((u32 *)out)[3] = BSWAP4(ctx->D[i]); - ((u32 *)out)[4] = BSWAP4(ctx->E[i]); + PUTU32(out+0,ctx->A[i]); + PUTU32(out+4,ctx->B[i]); + PUTU32(out+8,ctx->C[i]); + PUTU32(out+12,ctx->D[i]); + PUTU32(out+16,ctx->E[i]); out += 20; len += 20; @@ -822,7 +820,7 @@ static int aesni_cbc_hmac_sha1_ctrl(EVP_CIPHER_CTX *ctx, int type, int arg, void unsigned int n4x=1, x4; unsigned int frag, last, packlen, inp_len; - if (arginp[11]<<8|param->inp[12]; diff --git a/crypto/evp/e_aes_cbc_hmac_sha256.c b/crypto/evp/e_aes_cbc_hmac_sha256.c index df031cc1c0..63c6acda97 100644 --- a/crypto/evp/e_aes_cbc_hmac_sha256.c +++ b/crypto/evp/e_aes_cbc_hmac_sha256.c @@ -206,6 +206,8 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA256 *key, u8 *IVs; #if defined(BSWAP8) u64 seqnum; +#else + unsigned int carry,j; #endif if (RAND_bytes((IVs=blocks[0].c),16*x4)<=0) /* ask for IVs in bulk */ @@ -256,13 +258,9 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA256 *key, #if defined(BSWAP8) blocks[i].q[0] = BSWAP8(seqnum+i); #else - blocks[i].c[7] += ((u8*)key->md.data)[7]+i; - if (blocks[i].c[7] < i) { - int j; - - for (j=6;j>=0;j--) { - if (blocks[i].c[j]=((u8*)key->md.data)[j]+1) break; - } + for (carry=i,j=8;j--;) { + blocks[i].c[j] = ((u8*)key->md.data)[j]+carry; + carry = (blocks[i].c[j]-carry)>>(sizeof(carry)*8-1); } #endif blocks[i].c[8] = ((u8*)key->md.data)[8]; @@ -330,10 +328,10 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA256 *key, len += 64+13; /* 64 is HMAC header */ len *= 8; /* convert to bits */ if (off<(64-8)) { - blocks[i].d[15] = BSWAP4(len); + PUTU32(blocks[i].c+60,len); edges[i].blocks = 1; } else { - blocks[i].d[31] = BSWAP4(len); + PUTU32(blocks[i].c+124,len); edges[i].blocks = 2; } edges[i].ptr = blocks[i].c; @@ -344,16 +342,16 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA256 *key, memset(blocks,0,sizeof(blocks)); for (i=0;iA[i]); ctx->A[i] = key->tail.h[0]; - blocks[i].d[1] = BSWAP4(ctx->B[i]); ctx->B[i] = key->tail.h[1]; - blocks[i].d[2] = BSWAP4(ctx->C[i]); ctx->C[i] = key->tail.h[2]; - blocks[i].d[3] = BSWAP4(ctx->D[i]); ctx->D[i] = key->tail.h[3]; - blocks[i].d[4] = BSWAP4(ctx->E[i]); ctx->E[i] = key->tail.h[4]; - blocks[i].d[5] = BSWAP4(ctx->F[i]); ctx->F[i] = key->tail.h[5]; - blocks[i].d[6] = BSWAP4(ctx->G[i]); ctx->G[i] = key->tail.h[6]; - blocks[i].d[7] = BSWAP4(ctx->H[i]); ctx->H[i] = key->tail.h[7]; + PUTU32(blocks[i].c+0,ctx->A[i]); ctx->A[i] = key->tail.h[0]; + PUTU32(blocks[i].c+4,ctx->B[i]); ctx->B[i] = key->tail.h[1]; + PUTU32(blocks[i].c+8,ctx->C[i]); ctx->C[i] = key->tail.h[2]; + PUTU32(blocks[i].c+12,ctx->D[i]); ctx->D[i] = key->tail.h[3]; + PUTU32(blocks[i].c+16,ctx->E[i]); ctx->E[i] = key->tail.h[4]; + PUTU32(blocks[i].c+20,ctx->F[i]); ctx->F[i] = key->tail.h[5]; + PUTU32(blocks[i].c+24,ctx->G[i]); ctx->G[i] = key->tail.h[6]; + PUTU32(blocks[i].c+28,ctx->H[i]); ctx->H[i] = key->tail.h[7]; blocks[i].c[32] = 0x80; - blocks[i].d[15] = BSWAP4((64+32)*8); + PUTU32(blocks[i].c+60,(64+32)*8); edges[i].ptr = blocks[i].c; edges[i].blocks = 1; } @@ -371,14 +369,14 @@ static size_t tls1_1_multi_block_encrypt(EVP_AES_HMAC_SHA256 *key, out += 5+16+len; /* write MAC */ - ((u32 *)out)[0] = BSWAP4(ctx->A[i]); - ((u32 *)out)[1] = BSWAP4(ctx->B[i]); - ((u32 *)out)[2] = BSWAP4(ctx->C[i]); - ((u32 *)out)[3] = BSWAP4(ctx->D[i]); - ((u32 *)out)[4] = BSWAP4(ctx->E[i]); - ((u32 *)out)[5] = BSWAP4(ctx->F[i]); - ((u32 *)out)[6] = BSWAP4(ctx->G[i]); - ((u32 *)out)[7] = BSWAP4(ctx->H[i]); + PUTU32(out+0,ctx->A[i]); + PUTU32(out+4,ctx->B[i]); + PUTU32(out+8,ctx->C[i]); + PUTU32(out+12,ctx->D[i]); + PUTU32(out+16,ctx->E[i]); + PUTU32(out+20,ctx->F[i]); + PUTU32(out+24,ctx->G[i]); + PUTU32(out+28,ctx->H[i]); out += 32; len += 32;