Symbol: tmp3
crypto/heimdal/lib/krb5/crypto-evp.c
147
unsigned char tmp2[EVP_MAX_BLOCK_LENGTH], tmp3[EVP_MAX_BLOCK_LENGTH];
crypto/heimdal/lib/krb5/crypto-evp.c
169
memcpy(tmp3, p + blocksize, len);
crypto/heimdal/lib/krb5/crypto-evp.c
170
memcpy(tmp3 + len, tmp2 + len, blocksize - len); /* xor 0 */
crypto/heimdal/lib/krb5/crypto-evp.c
173
p[i + blocksize] = tmp2[i] ^ tmp3[i];
crypto/heimdal/lib/krb5/crypto-evp.c
176
EVP_Cipher(c, p, tmp3, blocksize);
crypto/libecc/src/sig/sm2.c
320
nn k, r, s, tmp, tmp2, tmp3;
crypto/libecc/src/sig/sm2.c
327
k.magic = r.magic = s.magic = tmp.magic = tmp2.magic = tmp3.magic = WORD(0);
crypto/libecc/src/sig/sm2.c
428
ret = nn_mod_mul(&tmp3, &r, &b, q); EG(ret, err); /* rb */
crypto/libecc/src/sig/sm2.c
430
ret = nn_mod_mul(&tmp3, &tmp3, x, q); EG(ret, err); /* (rb)x mod q */
crypto/libecc/src/sig/sm2.c
431
ret = nn_mod_sub(&tmp2, &k, &tmp3, q); EG(ret, err); /* tmp2 = (kb - (rb)x) mod q */
crypto/libecc/src/sig/sm2.c
441
ret = nn_mod_mul(&tmp3, &r, x, q); EG(ret, err); /* rx mod q */
crypto/libecc/src/sig/sm2.c
442
ret = nn_mod_sub(&tmp2, &k, &tmp3, q); EG(ret, err); /* tmp2 = (k - rx) mod q */
crypto/libecc/src/sig/sm2.c
464
nn_uninit(&tmp3);
crypto/openssl/crypto/aes/aes_ige.c
189
unsigned char tmp3[AES_BLOCK_SIZE];
crypto/openssl/crypto/aes/aes_ige.c
269
memcpy(tmp3, tmp2, AES_BLOCK_SIZE);
crypto/openssl/crypto/aes/aes_ige.c
270
iv = tmp3;
crypto/openssl/crypto/aes/aes_ige.c
287
memcpy(tmp3, tmp2, AES_BLOCK_SIZE);
crypto/openssl/crypto/aes/aes_ige.c
288
iv = tmp3;
crypto/openssl/crypto/ec/ecp_sm2p256.c
221
ALIGN32 BN_ULONG tmp3[P256_LIMBS] = { 0 };
crypto/openssl/crypto/ec/ecp_sm2p256.c
264
ecp_sm2p256_mul(tmp3, tmp2, tmp0);
crypto/openssl/crypto/ec/ecp_sm2p256.c
269
ecp_sm2p256_sub(R->X, R->X, tmp3);
crypto/openssl/crypto/ec/ecp_sm2p256.c
272
ecp_sm2p256_mul(tmp3, tmp3, P->Y);
crypto/openssl/crypto/ec/ecp_sm2p256.c
273
ecp_sm2p256_sub(R->Y, tmp2, tmp3);
crypto/openssl/crypto/srp/srp_lib.c
214
BIGNUM *tmp = NULL, *tmp2 = NULL, *tmp3 = NULL, *k = NULL, *K = NULL;
crypto/openssl/crypto/srp/srp_lib.c
222
if ((tmp = BN_new()) == NULL || (tmp2 = BN_new()) == NULL || (tmp3 = BN_new()) == NULL || (xtmp = BN_new()) == NULL)
crypto/openssl/crypto/srp/srp_lib.c
235
if (!BN_mul(tmp3, u, xtmp, bn_ctx))
crypto/openssl/crypto/srp/srp_lib.c
237
if (!BN_add(tmp2, a, tmp3))
crypto/openssl/crypto/srp/srp_lib.c
250
BN_clear_free(tmp3);
sys/crypto/aesni/aesni_ghash.c
110
__m128i tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8, tmp9;
sys/crypto/aesni/aesni_ghash.c
112
tmp3 = _mm_clmulepi64_si128(a, b, 0x00);
sys/crypto/aesni/aesni_ghash.c
120
tmp3 = _mm_xor_si128(tmp3, tmp5);
sys/crypto/aesni/aesni_ghash.c
123
tmp7 = _mm_srli_epi32(tmp3, 31);
sys/crypto/aesni/aesni_ghash.c
125
tmp3 = _mm_slli_epi32(tmp3, 1);
sys/crypto/aesni/aesni_ghash.c
131
tmp3 = _mm_or_si128(tmp3, tmp7);
sys/crypto/aesni/aesni_ghash.c
135
tmp7 = _mm_slli_epi32(tmp3, 31);
sys/crypto/aesni/aesni_ghash.c
136
tmp8 = _mm_slli_epi32(tmp3, 30);
sys/crypto/aesni/aesni_ghash.c
137
tmp9 = _mm_slli_epi32(tmp3, 25);
sys/crypto/aesni/aesni_ghash.c
143
tmp3 = _mm_xor_si128(tmp3, tmp7);
sys/crypto/aesni/aesni_ghash.c
145
tmp2 = _mm_srli_epi32(tmp3, 1);
sys/crypto/aesni/aesni_ghash.c
146
tmp4 = _mm_srli_epi32(tmp3, 2);
sys/crypto/aesni/aesni_ghash.c
147
tmp5 = _mm_srli_epi32(tmp3, 7);
sys/crypto/aesni/aesni_ghash.c
151
tmp3 = _mm_xor_si128(tmp3, tmp2);
sys/crypto/aesni/aesni_ghash.c
152
tmp6 = _mm_xor_si128(tmp6, tmp3);
sys/crypto/aesni/aesni_ghash.c
167
__m128i tmp0, tmp1, tmp2, tmp3;
sys/crypto/aesni/aesni_ghash.c
201
tmp3 = _mm_shuffle_epi32(H4, 78);
sys/crypto/aesni/aesni_ghash.c
203
tmp3 = _mm_xor_si128(tmp3, H4);
sys/crypto/aesni/aesni_ghash.c
209
tmp3 = _mm_clmulepi64_si128(tmp3, tmp7, 0x00);
sys/crypto/aesni/aesni_ghash.c
215
tmp0 = _mm_xor_si128(tmp3, tmp0);
sys/crypto/aesni/aesni_ghash.c
223
tmp3 = lo;
sys/crypto/aesni/aesni_ghash.c
226
tmp7 = _mm_srli_epi32(tmp3, 31);
sys/crypto/aesni/aesni_ghash.c
228
tmp3 = _mm_slli_epi32(tmp3, 1);
sys/crypto/aesni/aesni_ghash.c
234
tmp3 = _mm_or_si128(tmp3, tmp7);
sys/crypto/aesni/aesni_ghash.c
238
tmp7 = _mm_slli_epi32(tmp3, 31);
sys/crypto/aesni/aesni_ghash.c
239
tmp8 = _mm_slli_epi32(tmp3, 30);
sys/crypto/aesni/aesni_ghash.c
240
tmp9 = _mm_slli_epi32(tmp3, 25);
sys/crypto/aesni/aesni_ghash.c
246
tmp3 = _mm_xor_si128(tmp3, tmp7);
sys/crypto/aesni/aesni_ghash.c
248
tmp2 = _mm_srli_epi32(tmp3, 1);
sys/crypto/aesni/aesni_ghash.c
249
tmp4 = _mm_srli_epi32(tmp3, 2);
sys/crypto/aesni/aesni_ghash.c
250
tmp5 = _mm_srli_epi32(tmp3, 7);
sys/crypto/aesni/aesni_ghash.c
254
tmp3 = _mm_xor_si128(tmp3, tmp2);
sys/crypto/aesni/aesni_ghash.c
255
tmp6 = _mm_xor_si128(tmp6, tmp3);
sys/crypto/aesni/aesni_ghash.c
275
__m128i tmp1, tmp2, tmp3, tmp4;
sys/crypto/aesni/aesni_ghash.c
352
tmp3 = _mm_loadu_si128(&((const __m128i *)addt)[i*4+2]);
sys/crypto/aesni/aesni_ghash.c
357
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
sys/crypto/aesni/aesni_ghash.c
361
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
sys/crypto/aesni/aesni_ghash.c
392
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
sys/crypto/aesni/aesni_ghash.c
410
tmp3 =_mm_xor_si128(tmp3, KEY[0]);
sys/crypto/aesni/aesni_ghash.c
420
tmp3 = _mm_aesenc_si128(tmp3, KEY[j]);
sys/crypto/aesni/aesni_ghash.c
429
tmp3 =_mm_aesenclast_si128(tmp3, KEY[nr]);
sys/crypto/aesni/aesni_ghash.c
440
tmp3 = _mm_xor_si128(tmp3,
sys/crypto/aesni/aesni_ghash.c
455
_mm_storeu_si128(&((__m128i*)out)[i*8+2], tmp3);
sys/crypto/aesni/aesni_ghash.c
464
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
sys/crypto/aesni/aesni_ghash.c
473
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
sys/crypto/aesni/aesni_ghash.c
536
__m128i tmp1, tmp2, tmp3, tmp4;
sys/crypto/aesni/aesni_ghash.c
613
tmp3 = _mm_loadu_si128(&((const __m128i *)addt)[i*4+2]);
sys/crypto/aesni/aesni_ghash.c
618
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
sys/crypto/aesni/aesni_ghash.c
623
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
sys/crypto/aesni/aesni_ghash.c
645
tmp3 = _mm_loadu_si128(&((const __m128i *)in)[i*4+2]);
sys/crypto/aesni/aesni_ghash.c
650
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
sys/crypto/aesni/aesni_ghash.c
655
reduce4(H, H2, H3, H4, tmp4, tmp3, tmp2, tmp1, &X);
sys/crypto/aesni/aesni_ghash.c
697
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
sys/crypto/aesni/aesni_ghash.c
715
tmp3 =_mm_xor_si128(tmp3, KEY[0]);
sys/crypto/aesni/aesni_ghash.c
725
tmp3 = _mm_aesenc_si128(tmp3, KEY[j]);
sys/crypto/aesni/aesni_ghash.c
734
tmp3 =_mm_aesenclast_si128(tmp3, KEY[nr]);
sys/crypto/aesni/aesni_ghash.c
745
tmp3 = _mm_xor_si128(tmp3,
sys/crypto/aesni/aesni_ghash.c
760
_mm_storeu_si128(&((__m128i*)out)[i*8+2], tmp3);
sys/crypto/aesni/aesni_ghash.c
769
tmp3 = _mm_shuffle_epi8(tmp3, BSWAP_MASK);
sys/crypto/aesni/aesni_wrap.c
207
__m128i tmp1, tmp2, tmp3, tmp4;
sys/crypto/aesni/aesni_wrap.c
229
tmp3 = _mm_shuffle_epi8(ctr3, BSWAP_EPI64);
sys/crypto/aesni/aesni_wrap.c
244
aesni_enc8(rounds - 1, key_schedule, tmp1, tmp2, tmp3, tmp4,
usr.bin/truss/syscalls.c
1164
char tmp2[max_string + 1], *tmp3;
usr.bin/truss/syscalls.c
1196
tmp3 = malloc(len * 4 + 1);
usr.bin/truss/syscalls.c
1198
if (strvisx(tmp3, tmp2, len,
usr.bin/truss/syscalls.c
1205
fprintf(fp, "\"%s\"%s", tmp3,
usr.bin/truss/syscalls.c
1207
free(tmp3);
usr.bin/truss/syscalls.c
1705
char tmp2[max_string + 1], *tmp3;
usr.bin/truss/syscalls.c
1724
tmp3 = malloc(len * 4 + 1);
usr.bin/truss/syscalls.c
1726
if (strvisx(tmp3, tmp2, len,
usr.bin/truss/syscalls.c
1732
fprintf(fp, "\"%s\"%s", tmp3, truncated ?
usr.bin/truss/syscalls.c
1734
free(tmp3);