pbits
int pbits;
(pbits = BN_num_bits(dh_p)) <= 0 ||
need > INT_MAX / 2 || 2 * need > pbits)
if (!DH_set_length(dh, MINIMUM(need * 2, pbits - 1)))
int ossl_dh_generate_ffc_parameters(DH *dh, int type, int pbits, int qbits,
pbits, qbits, &res, cb);
pbits, qbits, &res, cb);
int ossl_dh_get_named_group_uid_from_size(int pbits)
switch (pbits) {
int ossl_dsa_generate_ffc_parameters(DSA *dsa, int type, int pbits, int qbits,
pbits, qbits, &res, cb);
pbits, qbits, &res, cb);
int EVP_PKEY_CTX_set_dh_paramgen_prime_len(EVP_PKEY_CTX *ctx, int pbits)
size_t bits = pbits;
unsigned int pbits = 2048;
params[1] = OSSL_PARAM_construct_uint(OSSL_PKEY_PARAM_FFC_PBITS, &pbits);
int ossl_dh_generate_ffc_parameters(DH *dh, int type, int pbits, int qbits,
int ossl_dh_get_named_group_uid_from_size(int pbits);
int ossl_dsa_generate_ffc_parameters(DSA *dsa, int type, int pbits, int qbits,
int EVP_PKEY_CTX_set_dh_paramgen_prime_len(EVP_PKEY_CTX *ctx, int pbits);
gctx->pbits = 2048;
&& !OSSL_PARAM_get_size_t(p, &gctx->pbits))
size_t pbits;
gctx->group_nid = ossl_dh_get_named_group_uid_from_size(gctx->pbits);
ret = DH_generate_parameters_ex(dh, gctx->pbits,
gctx->pbits, gctx->qbits,
gctx->pbits = 2048;
&& !OSSL_PARAM_get_size_t(p, &gctx->pbits))
size_t pbits;
gctx->gen_type = (gctx->pbits >= 2048 ? DSA_PARAMGEN_TYPE_FIPS_186_4 : DSA_PARAMGEN_TYPE_FIPS_186_2);
gctx->pbits, gctx->qbits,
static size_t pbits = 1024; /* With 160 Q bits, we MUST use 1024 P bits */
OSSL_PARAM_size_t("pbits", &pbits),
static size_t pbits = 1024; /* With 160 Q bits, we MUST use 1024 P bits */
OSSL_PARAM_size_t("pbits", &pbits),
unsigned int pbits = 512; /* minimum allowed for speed */
params[0] = OSSL_PARAM_construct_uint(OSSL_PKEY_PARAM_FFC_PBITS, &pbits);
unsigned int pbits = 512; /* minimum allowed for speed */
*p++ = OSSL_PARAM_construct_uint(OSSL_PKEY_PARAM_FFC_PBITS, &pbits);
*pbits = bits;
inet_cidr_pton_ipv6(const char *src, u_char *dst, int *pbits) {
*pbits = bits;
inet_cidr_pton_ipv4(const char *src, u_char *dst, int *pbits, int ipv6) {
pbits(cg_clustersfree(&acg), acg.cg_nclusterblks);
pbits(cg_inosused(&acg), afs.fs_ipg);
pbits(cg_blksfree(&acg), afs.fs_fpg);
static void pbits(void *, int);
pt_entry_t obits, pbits;
obits = pbits = *pdpe;
MPASS((pbits & (PG_MANAGED | PG_G)) == 0);
pbits &= ~(PG_RW | PG_M);
pbits |= pg_nx;
if (pbits != obits) {
if (!atomic_cmpset_long(pdpe, obits, pbits))
pbits = ptpaddr;
pbits &= ~(PG_RW | PG_M);
pbits |= pg_nx;
if (ptpaddr == pbits || !pmap_demote_pde(pmap, pde,
obits = pbits = *pte;
if ((pbits & PG_V) == 0)
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
m = PHYS_TO_VM_PAGE(pbits & PG_FRAME);
pbits &= ~(PG_RW | PG_M);
pbits |= pg_nx;
if (pbits != obits) {
if (!atomic_cmpset_long(pte, obits, pbits))
pt_entry_t obits, pbits;
obits = pbits = *pte;
if ((pbits & PG_V) == 0)
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
m = PHYS_TO_VM_PAGE(pbits & PG_FRAME);
pbits &= ~(PG_RW | PG_M);
pbits |= pg_nx;
if (pbits != obits) {
if (!atomic_cmpset_64(pte, obits, pbits))
pbits))
pt_entry_t obits, pbits;
obits = pbits = be64toh(*pte);
if ((pbits & PG_V) == 0)
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
m = PHYS_TO_VM_PAGE(pbits & PG_FRAME);
pbits &= ~(PG_RW | PG_M);
pbits |= RPTE_EAA_R;
pbits |= PG_X;
if (pbits != obits) {
if (!atomic_cmpset_long(pte, htobe64(obits), htobe64(pbits)))
sva, obits, pbits);
print_flags_generic(int flags, const struct bits *pbits, const char *format,
xo_emit(format, fmt_flags(pbits, flags));
for (p = pbits; p->b_mask; p++)
void print_flags_generic(int flags, const struct bits *pbits,