tmp3
unsigned char tmp2[EVP_MAX_BLOCK_LENGTH], tmp3[EVP_MAX_BLOCK_LENGTH];
memcpy(tmp3, p + blocksize, len);
memcpy(tmp3 + len, tmp2 + len, blocksize - len); /* xor 0 */
p[i + blocksize] = tmp2[i] ^ tmp3[i];
EVP_Cipher(c, p, tmp3, blocksize);
nn k, r, s, tmp, tmp2, tmp3;
k.magic = r.magic = s.magic = tmp.magic = tmp2.magic = tmp3.magic = WORD(0);
ret = nn_mod_mul(&tmp3, &r, &b, q); EG(ret, err); /* rb */
ret = nn_mod_mul(&tmp3, &tmp3, x, q); EG(ret, err); /* (rb)x mod q */
ret = nn_mod_sub(&tmp2, &k, &tmp3, q); EG(ret, err); /* tmp2 = (kb - (rb)x) mod q */
ret = nn_mod_mul(&tmp3, &r, x, q); EG(ret, err); /* rx mod q */
ret = nn_mod_sub(&tmp2, &k, &tmp3, q); EG(ret, err); /* tmp2 = (k - rx) mod q */
nn_uninit(&tmp3);
unsigned char tmp3[AES_BLOCK_SIZE];
memcpy(tmp3, tmp2, AES_BLOCK_SIZE);
iv = tmp3;
memcpy(tmp3, tmp2, AES_BLOCK_SIZE);
iv = tmp3;
ALIGN32 BN_ULONG tmp3[P256_LIMBS] = { 0 };
ecp_sm2p256_mul(tmp3, tmp2, tmp0);
ecp_sm2p256_sub(R->X, R->X, tmp3);
ecp_sm2p256_mul(tmp3, tmp3, P->Y);
ecp_sm2p256_sub(R->Y, tmp2, tmp3);
BIGNUM *tmp = NULL, *tmp2 = NULL, *tmp3 = NULL, *k = NULL, *K = NULL;
if ((tmp = BN_new()) == NULL || (tmp2 = BN_new()) == NULL || (tmp3 = BN_new()) == NULL || (xtmp = BN_new()) == NULL)
if (!BN_mul(tmp3, u, xtmp, bn_ctx))
if (!BN_add(tmp2, a, tmp3))
BN_clear_free(tmp3);
__m128i tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8, tmp9;
tmp3 = _mm_clmulepi64_si128(a, b, 0x00);
tmp3 = _mm_xor_si128(tmp3, tmp5);
tmp7 = _mm_srli_epi32(tmp3, 31);
tmp3 = _mm_slli_epi32(tmp3, 1);
tmp3 = _mm_or_si128(tmp3, tmp7);
tmp7 = _mm_slli_epi32(tmp3, 31);
tmp8 = _mm_slli_epi32(tmp3, 30);
tmp9 = _mm_slli_epi32(tmp3, 25);
tmp3 = _mm_xor_si128(tmp3, tmp7);
tmp2 = _mm_srli_epi32(tmp3, 1);
tmp4 = _mm_srli_epi32(tmp3, 2);
tmp5 = _mm_srli_epi32(tmp3, 7);
tmp3 = _mm_xor_si128(tmp3, tmp2);
tmp6 = _mm_xor_si128(tmp6, tmp3);
__m128i tmp0, tmp1, tmp2, tmp3;
tmp3 = _mm_shuffle_epi32(H4, 78);
tmp3 = _mm_xor_si128(tmp3, H4);
tmp3 = _mm_clmulepi64_si128(tmp3, tmp7, 0x00);
tmp0 = _mm_xor_si128(tmp3, tmp0);
tmp3 = lo;
tmp7 = _mm_srli_epi32(tmp3, 31);
tmp3 = _mm_slli_epi32(tmp3, 1);
tmp3 = _mm_or_si128(tmp3, tmp7);
tmp7 = _mm_slli_epi32(tmp3, 31);
tmp8 = _mm_slli_epi32(tmp3, 30);
tmp9 = _mm_slli_epi32(tmp3, 25);
tmp3 = _mm_xor_si128(tmp3, tmp7);
tmp2 = _mm_srli_epi32(tmp3, 1);
tmp4 = _mm_srli_epi32(tmp3, 2);
tmp5 = _mm_srli_epi32(tmp3, 7);
tmp3 = _mm_xor_si128(tmp3, tmp2);
tmp6 = _mm_xor_si128(tmp6, tmp3);
__m128i tmp1, tmp2, tmp3, tmp4;
tmp3 = _mm_loadu_si128(&((const __m128i *)addt)[i*4+2]);
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
tmp3 =_mm_xor_si128(tmp3, KEY[0]);
tmp3 = _mm_aesenc_si128(tmp3, KEY[j]);
tmp3 =_mm_aesenclast_si128(tmp3, KEY[nr]);
tmp3 = _mm_xor_si128(tmp3,
_mm_storeu_si128(&((__m128i*)out)[i*8+2], tmp3);
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
__m128i tmp1, tmp2, tmp3, tmp4;
tmp3 = _mm_loadu_si128(&((const __m128i *)addt)[i*4+2]);
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
tmp3 = _mm_loadu_si128(&((const __m128i *)in)[i*4+2]);
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
tmp3 =_mm_xor_si128(tmp3, KEY[0]);
tmp3 = _mm_aesenc_si128(tmp3, KEY[j]);
tmp3 =_mm_aesenclast_si128(tmp3, KEY[nr]);
tmp3 = _mm_xor_si128(tmp3,
_mm_storeu_si128(&((__m128i*)out)[i*8+2], tmp3);
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
__m128i tmp1, tmp2, tmp3, tmp4;
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
aesni_enc8(rounds - 1, key_schedule, tmp1, tmp2, tmp3, tmp4,
char tmp2[max_string + 1], *tmp3;
tmp3 = malloc(len * 4 + 1);
if (strvisx(tmp3, tmp2, len,
fprintf(fp, "\"%s\"%s", tmp3,
free(tmp3);
char tmp2[max_string + 1], *tmp3;
tmp3 = malloc(len * 4 + 1);
if (strvisx(tmp3, tmp2, len,
fprintf(fp, "\"%s\"%s", tmp3, truncated ?
free(tmp3);