p3
unsigned char *p1 = os1.data, *p2 = os2.data, *p3 = res->keyvalue.data;
p3[i] = p1[i] ^ p2[i];
char *p1, *p2, *p3;
p3 = estrdup(p);
if (strcmp(p3, test_cc_name) != 0)
free(p3);
ASSERT_INT_EQ(sshbuf_froms(p2, &p3), 0);
ASSERT_PTR_NE(p3, NULL);
ASSERT_PTR_NE(sshbuf_ptr(p3), NULL);
ASSERT_SIZE_T_EQ(sshbuf_len(p3), sizeof(test_buf) - 1);
ASSERT_MEM_EQ(sshbuf_ptr(p3), test_buf, sizeof(test_buf) - 1);
sshbuf_free(p3);
ASSERT_INT_EQ(sshbuf_froms(p2, &p3), SSH_ERR_MESSAGE_INCOMPLETE);
ASSERT_PTR_EQ(p3, NULL);
struct sshbuf *p1, *p2, *p3;
if (!TEST_true(BN_hex2bn(&obj->p3, p3_init)))
BN_free(obj->p3);
if (!TEST_ptr(obj->p3 = BN_native2bn(params->data,
params->return_size = BN_num_bytes(obj->p3);
BN_bn2nativepad(obj->p3, params->data, params->return_size);
&& !TEST_true(OSSL_PARAM_get_BN(p, &obj->p3)))
&& !TEST_true(OSSL_PARAM_set_BN(p, obj->p3)))
|| !TEST_BN_eq(sneakpeek->p3, app_p3) /* app value set */
BIGNUM *p3;
BN_free(obj->p3);
obj->p3 = NULL;
static const unsigned char p3[] = { 0x0B, 0x50, 0x4F, 0x33, 0x40 };
&& TEST_ptr(bn_p3 = bn_load_new(p3, sizeof(p3)))
static const unsigned char p3[] = { 0x0F, 0x50, 0x00, 0x03, 0x75 };
&& TEST_ptr(bn_p3 = bn_load_new(p3, sizeof(p3)))
c = x - z * (p1 + z * (p2 + z * (p3 + z * (p4 +
p3 = 6.6137564717940088e-05, /* 0x3f11566a, 0xb5c2ba0d */
long double nadj,p,p1,p2,p3,q,r,t,w,y,z;
c = x - z * (p1 + z * (p2 + z * (p3 + z * (p4 +
double nadj,p,p1,p2,p3,q,r,t,w,y,z;
p3 = t2+w*(t5+w*(t8+w*(t11+w*t14)));
p = z*p1-(tt-w*(p2+y*p3));
value = get_variable(s, &p3, line);
if (p3 != NULL) {
if (c == '{' && *p3 == '}')
p3++;
len += strlen(p3);
if (p3 != NULL)
strcat(p1, p3);
s = p1 + len - strlen(p3) - 1;
char c, *s, *p1, *p2, *p3, *newstring, *value;
p3 = NULL;
#define SVM_CTR3(vcpu, format, p1, p2, p3) \
VCPU_CTR3((vcpu)->sc->vm, (vcpu)->vcpuid, format, p1, p2, p3)
#define SVM_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->sc->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define VMX_CTR3(vcpu, format, p1, p2, p3) \
VCPU_CTR3((vcpu)->vmx->vm, (vcpu)->vcpuid, format, p1, p2, p3)
#define VMX_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->vmx->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define VLAPIC_CTR3(vlapic, format, p1, p2, p3) \
VCPU_CTR3((vlapic)->vm, (vlapic)->vcpuid, format, p1, p2, p3)
#define VMM_CTR3(vcpu, format, p1, p2, p3) \
VCPU_CTR3((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2, p3)
#define VMM_CTR4(vcpu, format, p1, p2, p3, p4) \
VCPU_CTR4((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4)
#define LINUX_CTR6(f, m, p1, p2, p3, p4, p5, p6) do { \
p1, p2, p3, p4, p5, p6); \
#define LINUX_CTR3(f, m, p1, p2, p3) LINUX_CTR6(f, m, p1, p2, p3, 0, 0, 0)
#define LINUX_CTR4(f, m, p1, p2, p3, p4) LINUX_CTR6(f, m, p1, p2, p3, p4, 0, 0)
#define LINUX_CTR5(f, m, p1, p2, p3, p4, p5) LINUX_CTR6(f, m, p1, p2, p3, p4, p5, 0)
#define LINUX_CTR3(f, m, p1, p2, p3)
#define LINUX_CTR4(f, m, p1, p2, p3, p4)
#define LINUX_CTR5(f, m, p1, p2, p3, p4, p5)
#define LINUX_CTR6(f, m, p1, p2, p3, p4, p5, p6)
#define Round256(p0,p1,p2,p3,ROT,rNum) \
X##p2 += X##p3; X##p3 = RotL_64(X##p3,ROT##_1); X##p3 ^= X##p2; \
#define R256(p0,p1,p2,p3,ROT,rNum) /* fully unrolled */ \
Round256(p0,p1,p2,p3,ROT,rNum) \
#define R256(p0,p1,p2,p3,ROT,rNum) \
Round256(p0,p1,p2,p3,ROT,rNum) \
#define Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
X##p2 += X##p3; X##p3 = RotL_64(X##p3,ROT##_1); X##p3 ^= X##p2; \
#define R512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) /* unrolled */ \
Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
#define R512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
Round512(p0,p1,p2,p3,p4,p5,p6,p7,ROT,rNum) \
#define Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rNum) \
X##p2 += X##p3; X##p3 = RotL_64(X##p3,ROT##_1); X##p3 ^= X##p2; \
#define R1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
#define R1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
Round1024(p0,p1,p2,p3,p4,p5,p6,p7,p8,p9,pA,pB,pC,pD,pE,pF,ROT,rn) \
#define hpt_rebuild_mirror(p1, p2, p3) hpt_rebuild_data_block(p1, p2, p3)
#define hpt_rebuild_mirror(p1, p2, p3) hpt_rebuild_data_block(p1, p2, p3)
#define hpt_rebuild_mirror(p1, p2, p3) hpt_rebuild_data_block(p1, p2, p3)
#define hpt_rebuild_mirror(p1, p2, p3) hpt_rebuild_data_block(p1, p2, p3)
isp_gen_role_str(char *buf, size_t len, uint16_t p3)
if (p3 & PRLI_WD3_ENHANCED_DISCOVERY) {
if (p3 & PRLI_WD3_REC_SUPPORT) {
if (p3 & PRLI_WD3_TASK_RETRY_IDENTIFICATION_REQUESTED) {
if (p3 & PRLI_WD3_RETRY) {
if (p3 & PRLI_WD3_CONFIRMED_COMPLETION_ALLOWED) {
if (p3 & PRLI_WD3_DATA_OVERLAY_ALLOWED) {
if (p3 & PRLI_WD3_INITIATOR_FUNCTION) {
if (p3 & PRLI_WD3_TARGET_FUNCTION) {
if (p3 & PRLI_WD3_READ_FCP_XFER_RDY_DISABLED) {
if (p3 & PRLI_WD3_WRITE_FCP_XFER_RDY_DISABLED) {
#define VCPU_CTR3(vm, vcpuid, format, p1, p2, p3) \
CTR5(KTR_VMM, "vm %s[%d]: " format, vm_name((vm)), (vcpuid), (p1), (p2), (p3))
#define VCPU_CTR4(vm, vcpuid, format, p1, p2, p3, p4) \
(p1), (p2), (p3), (p4))
#define VM_CTR3(vm, format, p1, p2, p3) \
CTR4(KTR_VMM, "vm %s: " format, vm_name((vm)), (p1), (p2), (p3))
#define VM_CTR4(vm, format, p1, p2, p3, p4) \
CTR5(KTR_VMM, "vm %s: " format, vm_name((vm)), (p1), (p2), (p3), (p4))
int lv1_net_control(uint64_t bus, uint64_t dev, uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4, uint64_t *v1, uint64_t *v2);
int lv1_gpu_context_attribute(uint64_t chandle, uint64_t op, uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4);
int lv1_gpu_attribute(uint64_t p1, uint64_t p2, uint64_t p3, uint64_t p4, uint64_t p5);
#define SBI_CALL3(e, f, p1, p2, p3) SBI_CALL5(e, f, p1, p2, p3, 0, 0)
#define SBI_CALL4(e, f, p1, p2, p3, p4) SBI_CALL5(e, f, p1, p2, p3, p4, 0)
#define SBI_CALL5(e, f, p1, p2, p3, p4, p5) sbi_call(e, f, p1, p2, p3, p4, p5)
#define TR3(d, p1, p2, p3) CTR3(KTR_GEN, d, p1, p2, p3)
#define TR4(d, p1, p2, p3, p4) CTR4(KTR_GEN, d, p1, p2, p3, p4)
#define TR5(d, p1, p2, p3, p4, p5) CTR5(KTR_GEN, d, p1, p2, p3, p4, p5)
#define TR6(d, p1, p2, p3, p4, p5, p6) CTR6(KTR_GEN, d, p1, p2, p3, p4, p5, p6)
#define ITR3(d, p1, p2, p3) CTR3(KTR_INIT, d, p1, p2, p3)
#define ITR4(d, p1, p2, p3, p4) CTR4(KTR_INIT, d, p1, p2, p3, p4)
#define ITR5(d, p1, p2, p3, p4, p5) CTR5(KTR_INIT, d, p1, p2, p3, p4, p5)
#define ITR6(d, p1, p2, p3, p4, p5, p6) CTR6(KTR_INIT, d, p1, p2, p3, p4, p5, p6)
#define ITR3(d, p1, p2, p3)
#define ITR4(d, p1, p2, p3, p4)
#define ITR5(d, p1, p2, p3, p4, p5)
#define ITR6(d, p1, p2, p3, p4, p5, p6)
#define CTR6(m, format, p1, p2, p3, p4, p5, p6) do { \
(u_long)(p1), (u_long)(p2), (u_long)(p3), \
#define CTR3(m, format, p1, p2, p3) CTR6(m, format, p1, p2, p3, 0, 0, 0)
#define CTR4(m, format, p1, p2, p3, p4) CTR6(m, format, p1, p2, p3, p4, 0, 0)
#define CTR5(m, format, p1, p2, p3, p4, p5) CTR6(m, format, p1, p2, p3, p4, p5, 0)
#define CTR3(m, d, p1, p2, p3) (void)0
#define CTR4(m, d, p1, p2, p3, p4) (void)0
#define CTR5(m, d, p1, p2, p3, p4, p5) (void)0
#define CTR6(m, d, p1, p2, p3, p4, p5, p6) (void)0
#define PMCDBG3(M, N, L, F, p1, p2, p3) do { \
CTR3(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3); \
#define PMCDBG4(M, N, L, F, p1, p2, p3, p4) do { \
CTR4(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4);\
#define PMCDBG5(M, N, L, F, p1, p2, p3, p4, p5) do { \
CTR5(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4, \
#define PMCDBG6(M, N, L, F, p1, p2, p3, p4, p5, p6) do { \
CTR6(KTR_PMC, #M ":" #N ":" #L ": " F, p1, p2, p3, p4, \
#define PMCDBG3(M, N, L, F, p1, p2, p3)
#define PMCDBG4(M, N, L, F, p1, p2, p3, p4)
#define PMCDBG5(M, N, L, F, p1, p2, p3, p4, p5)
#define PMCDBG6(M, N, L, F, p1, p2, p3, p4, p5, p6)
char *p1, *p2, *p3;
p3 = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
ATF_REQUIRE(p3 != MAP_FAILED);
ATF_REQUIRE(p3[0] == 0);
munmap(p3, pagesize);
char *p1, *p2, *p3;
p3 = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_SHARED, fd,
ATF_REQUIRE(p3 != MAP_FAILED);
ATF_REQUIRE(p3[0] == 0);
munmap(p3, pagesize);
p3(tcps_sc_added, "\t{:entries-added/%ju} "
p3(tcps_hc_added, "\t{:entries-added/%ju} "