p4
BN_ULONG p4[P256_LIMBS];
ecp_nistz256_mul_mont(p4, res, p2); /* f*p */
ecp_nistz256_sqr_mont(res, p4);
ecp_nistz256_mul_mont(p8, res, p4); /* ff*p */
ecp_nistz256_mul_mont(res, res, p4);
if (!TEST_ptr(obj->p4 = OPENSSL_strdup(p4_init)))
OPENSSL_free(obj->p4);
if (!TEST_ptr(obj->p4 = OPENSSL_strndup(params->data,
obj->p4_l = strlen(obj->p4);
params->return_size = strlen(obj->p4);
strcpy(params->data, obj->p4);
OPENSSL_free(obj->p4);
obj->p4 = NULL;
if (!TEST_true(OSSL_PARAM_get_utf8_string(p, &obj->p4, 0)))
&& !TEST_true(OSSL_PARAM_set_utf8_string(p, obj->p4)))
|| !TEST_str_eq(sneakpeek->p4, app_p4) /* app value set */
char *p4;
OPENSSL_free(obj->p4);
obj->p4 = NULL;
static const unsigned char p4[] = { 0x0F, 0xFF, 0xFF, 0xFF, 0xFF };
&& TEST_ptr(bn_p4 = bn_load_new(p4, sizeof(p4)))
c = x - z * (p1 + z * (p2 + z * (p3 + z * (p4 +
p4 = -1.6534060280704225e-06, /* 0xbebbbd53, 0x273e8fb7 */
c = x - z * (p1 + z * (p2 + z * (p3 + z * (p4 +
struct in_addr mask4, *p4;
p4 = (struct in_addr *)prefix;
p4->s_addr &= mask4.s_addr;
#define SVM_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->sc->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define VMX_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->vmx->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define VMM_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define LINUX_CTR6(f, m, p1, p2, p3, p4, p5, p6) do { \
p1, p2, p3, p4, p5, p6); \
#define LINUX_CTR4(f, m, p1, p2, p3, p4) LINUX_CTR6(f, m, p1, p2, p3, p4, 0, 0)
#define LINUX_CTR5(f, m, p1, p2, p3, p4, p5) LINUX_CTR6(f, m, p1, p2, p3, p4, p5, 0)
#define LINUX_CTR4(f, m, p1, p2, p3, p4)
#define LINUX_CTR5(f, m, p1, p2, p3, p4, p5)
#define LINUX_CTR6(f, m, p1, p2, p3, p4, p5, p6)
#define Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
X##p4 += X##p5; X##p5 = RotL_64(X##p5,ROT##_2); X##p5 ^= X##p4; \
#define R512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) /* unrolled */ \
Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
#define R512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
#define Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rNum) \
X##p4 += X##p5; X##p5 = RotL_64(X##p5,ROT##_2); X##p5 ^= X##p4; \
#define R1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
#define R1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
data.data = __DECONST(void *, p4);
static const uint8_t p4[4] = {
#define VCPU_CTR4(vm, vcpuid, format, p1, p2, p3, p4) \
(p1), (p2), (p3), (p4))
#define VM_CTR4(vm, format, p1, p2, p3, p4) \
CTR5(KTR_VMM, "vm %s: " format, vm_name((vm)), (p1), (p2), (p3), (p4))
struct sctp_ipv4addr_param *p4, p4_buf;
p4 = (struct sctp_ipv4addr_param *)phdr;
sin4.sin_addr.s_addr = p4->addr;
struct sctp_ipv4addr_param ip4_param, *p4;
p4 = (struct sctp_ipv4addr_param *)phdr;
memcpy(&sin4.sin_addr, &p4->addr, sizeof(p4->addr));
struct sctp_ipv4addr_param *p4, p4_buf;
p4 = (struct sctp_ipv4addr_param *)sctp_get_next_param(m,
if (p4 == NULL) {
memcpy(&sin->sin_addr, &p4->addr, sizeof(struct in_addr));
struct sctp_ipv4addr_param *p4, p4_buf;
p4 = (struct sctp_ipv4addr_param *)phdr;
sin.sin_addr.s_addr = p4->addr;
int lv1_net_control(uint64_t bus, uint64_t dev, uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4, uint64_t *v1, uint64_t *v2);
int lv1_gpu_context_attribute(uint64_t chandle, uint64_t op, uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4);
int lv1_gpu_attribute(uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4, uint64_t p5);
#define SBI_CALL4(e, f, p1, p2, p3, p4) SBI_CALL5(e, f, p1, p2, p3, p4, 0)
#define SBI_CALL5(e, f, p1, p2, p3, p4, p5) sbi_call(e, f, p1, p2, p3, p4, p5)
#define TR4(d, p1, p2, p3, p4) CTR4(KTR_GEN, d, p1, p2, p3, p4)
#define TR5(d, p1, p2, p3, p4, p5) CTR5(KTR_GEN, d, p1, p2, p3, p4, p5)
#define TR6(d, p1, p2, p3, p4, p5, p6) CTR6(KTR_GEN, d, p1, p2, p3, p4, p5, p6)
#define ITR4(d, p1, p2, p3, p4) CTR4(KTR_INIT, d, p1, p2, p3, p4)
#define ITR5(d, p1, p2, p3, p4, p5) CTR5(KTR_INIT, d, p1, p2, p3, p4, p5)
#define ITR6(d, p1, p2, p3, p4, p5, p6) CTR6(KTR_INIT, d, p1, p2, p3, p4, p5, p6)
#define ITR4(d, p1, p2, p3, p4)
#define ITR5(d, p1, p2, p3, p4, p5)
#define ITR6(d, p1, p2, p3, p4, p5, p6)
#define CTR6(m, format, p1, p2, p3, p4, p5, p6) do { \
(u_long)(p4), (u_long)(p5), (u_long)(p6)); \
#define CTR4(m, format, p1, p2, p3, p4) CTR6(m, format, p1, p2, p3, p4, 0, 0)
#define CTR5(m, format, p1, p2, p3, p4, p5) CTR6(m, format, p1, p2, p3, p4, p5, 0)
#define CTR4(m, d, p1, p2, p3, p4) (void)0
#define CTR5(m, d, p1, p2, p3, p4, p5) (void)0
#define CTR6(m, d, p1, p2, p3, p4, p5, p6) (void)0
#define PMCDBG4(M, N, L, F, p1, p2, p3, p4) do { \
CTR4(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4);\
#define PMCDBG5(M, N, L, F, p1, p2, p3, p4, p5) do { \
CTR5(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4, \
#define PMCDBG6(M, N, L, F, p1, p2, p3, p4, p5, p6) do { \
CTR6(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4, \
#define PMCDBG4(M, N, L, F, p1, p2, p3, p4)
#define PMCDBG5(M, N, L, F, p1, p2, p3, p4, p5)
#define PMCDBG6(M, N, L, F, p1, p2, p3, p4, p5, p6)