Symbol: pd
crypto/heimdal/kdc/hprop.c
118
ret = krb5_write_message(context, &pd->sock, &data);
crypto/heimdal/kdc/hprop.c
120
ret = krb5_write_priv_message(context, pd->auth_context,
crypto/heimdal/kdc/hprop.c
121
&pd->sock, &data);
crypto/heimdal/kdc/hprop.c
229
struct prop_data *pd)
crypto/heimdal/kdc/hprop.c
235
ret = mit_prop_dump(pd, database_name);
crypto/heimdal/kdc/hprop.c
240
ret = hdb_foreach(context, db, HDB_F_DECRYPT, v5_prop, pd);
crypto/heimdal/kdc/hprop.c
255
struct prop_data pd;
crypto/heimdal/kdc/hprop.c
258
pd.context = context;
crypto/heimdal/kdc/hprop.c
259
pd.auth_context = NULL;
crypto/heimdal/kdc/hprop.c
260
pd.sock = STDOUT_FILENO;
crypto/heimdal/kdc/hprop.c
262
ret = iterate (context, database_name, db, type, &pd);
crypto/heimdal/kdc/hprop.c
266
ret = krb5_write_message (context, &pd.sock, &data);
crypto/heimdal/kdc/hprop.c
286
struct prop_data pd;
crypto/heimdal/kdc/hprop.c
348
pd.context = context;
crypto/heimdal/kdc/hprop.c
349
pd.auth_context = auth_context;
crypto/heimdal/kdc/hprop.c
350
pd.sock = fd;
crypto/heimdal/kdc/hprop.c
352
ret = iterate (context, database_name, db, type, &pd);
crypto/heimdal/kdc/hprop.c
93
struct prop_data *pd = appdata;
crypto/heimdal/kdc/mit_dump.c
143
struct prop_data *pd = arg;
crypto/heimdal/kdc/mit_dump.c
187
ret = _hdb_mit_dump2mitdb_entry(pd->context, line, sp);
crypto/heimdal/kdc/mit_dump.c
191
ret = _hdb_mdb_value2entry(pd->context, &kdb_ent, 0, &ent.entry);
crypto/heimdal/kdc/mit_dump.c
194
ret = v5_prop(pd->context, NULL, &ent, arg);
crypto/heimdal/kdc/mit_dump.c
195
hdb_free_entry(pd->context, &ent);
crypto/heimdal/lib/hx509/hxtool.c
214
struct pem_data pd;
crypto/heimdal/lib/hx509/hxtool.c
217
pd.os = &co;
crypto/heimdal/lib/hx509/hxtool.c
218
pd.detached_data = 0;
crypto/heimdal/lib/hx509/hxtool.c
224
ret = hx509_pem_read(context, f, pem_reader, &pd);
crypto/heimdal/lib/hx509/hxtool.c
229
if (pd.detached_data && opt->signed_content_string == NULL) {
crypto/heimdal/lib/roken/qsort.c
119
char *pa, *pb, *pc, *pd, *pl, *pm, *pn;
crypto/heimdal/lib/roken/qsort.c
149
pc = pd = (char *)a + (n - 1) * es;
crypto/heimdal/lib/roken/qsort.c
162
swap(pc, pd);
crypto/heimdal/lib/roken/qsort.c
163
pd -= es;
crypto/heimdal/lib/roken/qsort.c
186
r = min(pd - pc, pn - pd - es);
crypto/heimdal/lib/roken/qsort.c
194
if ((r = pd - pc) > es) {
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
119
krb5_data pd;
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
128
pd.length = strlen (p);
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
129
pd.data = (char *) p;
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
141
r = krb5int_des_string_to_key (0, &pd, &sd, 0, &key);
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
166
krb5_data pd, sd;
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
172
pd.length = strlen (p);
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
173
pd.data = (char *) p;
crypto/krb5/src/lib/crypto/crypto_tests/vectors.c
186
r = krb5_c_string_to_key (0, enctype, &pd, &sd, &key);
crypto/krb5/src/lib/rpc/auth_gss.c
252
authgss_get_private_data(AUTH *auth, struct authgss_private_data *pd)
crypto/krb5/src/lib/rpc/auth_gss.c
258
if (!auth || !pd)
crypto/krb5/src/lib/rpc/auth_gss.c
266
pd->pd_ctx = gd->ctx;
crypto/krb5/src/lib/rpc/auth_gss.c
267
pd->pd_ctx_hndl = gd->gc.gc_ctx;
crypto/krb5/src/lib/rpc/auth_gss.c
268
pd->pd_seq_win = gd->win;
crypto/openssl/crypto/bio/bss_conn.c
598
BIO_POLL_DESCRIPTOR *pd = ptr;
crypto/openssl/crypto/bio/bss_conn.c
604
pd->type = BIO_POLL_DESCRIPTOR_TYPE_SOCK_FD;
crypto/openssl/crypto/bio/bss_conn.c
605
pd->value.fd = b->num;
crypto/openssl/crypto/bio/bss_dgram.c
1034
BIO_POLL_DESCRIPTOR *pd = ptr;
crypto/openssl/crypto/bio/bss_dgram.c
1036
pd->type = BIO_POLL_DESCRIPTOR_TYPE_SOCK_FD;
crypto/openssl/crypto/bio/bss_dgram.c
1037
pd->value.fd = b->num;
crypto/openssl/crypto/bio/bss_sock.c
209
BIO_POLL_DESCRIPTOR *pd = ptr;
crypto/openssl/crypto/bio/bss_sock.c
216
pd->type = BIO_POLL_DESCRIPTOR_TYPE_SOCK_FD;
crypto/openssl/crypto/bio/bss_sock.c
217
pd->value.fd = b->num;
crypto/openssl/crypto/evp/evp_enc.c
1418
unsigned int pd = pad;
crypto/openssl/crypto/evp/evp_enc.c
1427
params[0] = OSSL_PARAM_construct_uint(OSSL_CIPHER_PARAM_PADDING, &pd);
crypto/openssl/crypto/param_build.c
252
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
286
pd = param_push(bld, key, sz, sz, type, secure);
crypto/openssl/crypto/param_build.c
287
if (pd == NULL)
crypto/openssl/crypto/param_build.c
289
pd->bn = bn;
crypto/openssl/crypto/param_build.c
325
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
336
pd = param_push(bld, key, bsize, bsize + 1, OSSL_PARAM_UTF8_STRING, secure);
crypto/openssl/crypto/param_build.c
337
if (pd == NULL)
crypto/openssl/crypto/param_build.c
339
pd->string = buf;
crypto/openssl/crypto/param_build.c
346
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
355
pd = param_push(bld, key, bsize, sizeof(buf), OSSL_PARAM_UTF8_PTR, 0);
crypto/openssl/crypto/param_build.c
356
if (pd == NULL)
crypto/openssl/crypto/param_build.c
358
pd->string = buf;
crypto/openssl/crypto/param_build.c
365
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
374
pd = param_push(bld, key, bsize, bsize, OSSL_PARAM_OCTET_STRING, secure);
crypto/openssl/crypto/param_build.c
375
if (pd == NULL)
crypto/openssl/crypto/param_build.c
377
pd->string = buf;
crypto/openssl/crypto/param_build.c
384
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
391
pd = param_push(bld, key, bsize, sizeof(buf), OSSL_PARAM_OCTET_PTR, 0);
crypto/openssl/crypto/param_build.c
392
if (pd == NULL)
crypto/openssl/crypto/param_build.c
394
pd->string = buf;
crypto/openssl/crypto/param_build.c
403
OSSL_PARAM_BLD_DEF *pd;
crypto/openssl/crypto/param_build.c
407
pd = sk_OSSL_PARAM_BLD_DEF_value(bld->params, i);
crypto/openssl/crypto/param_build.c
408
param[i].key = pd->key;
crypto/openssl/crypto/param_build.c
409
param[i].data_type = pd->type;
crypto/openssl/crypto/param_build.c
410
param[i].data_size = pd->size;
crypto/openssl/crypto/param_build.c
413
if (pd->secure) {
crypto/openssl/crypto/param_build.c
415
secure += pd->alloc_blocks;
crypto/openssl/crypto/param_build.c
418
blk += pd->alloc_blocks;
crypto/openssl/crypto/param_build.c
421
if (pd->bn != NULL) {
crypto/openssl/crypto/param_build.c
423
if (pd->type == OSSL_PARAM_UNSIGNED_INTEGER)
crypto/openssl/crypto/param_build.c
424
BN_bn2nativepad(pd->bn, (unsigned char *)p, pd->size);
crypto/openssl/crypto/param_build.c
426
BN_signed_bn2native(pd->bn, (unsigned char *)p, pd->size);
crypto/openssl/crypto/param_build.c
427
} else if (pd->type == OSSL_PARAM_OCTET_PTR
crypto/openssl/crypto/param_build.c
428
|| pd->type == OSSL_PARAM_UTF8_PTR) {
crypto/openssl/crypto/param_build.c
430
*(const void **)p = pd->string;
crypto/openssl/crypto/param_build.c
431
} else if (pd->type == OSSL_PARAM_OCTET_STRING
crypto/openssl/crypto/param_build.c
432
|| pd->type == OSSL_PARAM_UTF8_STRING) {
crypto/openssl/crypto/param_build.c
433
if (pd->string != NULL)
crypto/openssl/crypto/param_build.c
434
memcpy(p, pd->string, pd->size);
crypto/openssl/crypto/param_build.c
436
memset(p, 0, pd->size);
crypto/openssl/crypto/param_build.c
437
if (pd->type == OSSL_PARAM_UTF8_STRING)
crypto/openssl/crypto/param_build.c
438
((char *)p)[pd->size] = '\0';
crypto/openssl/crypto/param_build.c
441
if (pd->size > sizeof(pd->num))
crypto/openssl/crypto/param_build.c
442
memset(p, 0, pd->size);
crypto/openssl/crypto/param_build.c
443
else if (pd->size > 0)
crypto/openssl/crypto/param_build.c
444
memcpy(p, &pd->num, pd->size);
crypto/openssl/crypto/param_build.c
55
OSSL_PARAM_BLD_DEF *pd = OPENSSL_zalloc(sizeof(*pd));
crypto/openssl/crypto/param_build.c
57
if (pd == NULL)
crypto/openssl/crypto/param_build.c
59
pd->key = key;
crypto/openssl/crypto/param_build.c
60
pd->type = type;
crypto/openssl/crypto/param_build.c
61
pd->size = size;
crypto/openssl/crypto/param_build.c
62
pd->alloc_blocks = ossl_param_bytes_to_blocks(alloc);
crypto/openssl/crypto/param_build.c
63
if ((pd->secure = secure) != 0)
crypto/openssl/crypto/param_build.c
64
bld->secure_blocks += pd->alloc_blocks;
crypto/openssl/crypto/param_build.c
66
bld->total_blocks += pd->alloc_blocks;
crypto/openssl/crypto/param_build.c
67
if (sk_OSSL_PARAM_BLD_DEF_push(bld->params, pd) <= 0) {
crypto/openssl/crypto/param_build.c
68
OPENSSL_free(pd);
crypto/openssl/crypto/param_build.c
69
pd = NULL;
crypto/openssl/crypto/param_build.c
71
return pd;
crypto/openssl/crypto/param_build.c
77
OSSL_PARAM_BLD_DEF *pd = param_push(bld, key, size, size, type, 0);
crypto/openssl/crypto/param_build.c
79
if (pd == NULL) {
crypto/openssl/crypto/param_build.c
83
if (size > sizeof(pd->num)) {
crypto/openssl/crypto/param_build.c
87
memcpy(&pd->num, num, size);
crypto/openssl/crypto/property/property_parse.c
300
static void pd_free(OSSL_PROPERTY_DEFINITION *pd)
crypto/openssl/crypto/property/property_parse.c
302
OPENSSL_free(pd);
crypto/openssl/providers/common/include/prov/provider_util.h
66
const EVP_MD *ossl_prov_digest_fetch(PROV_DIGEST *pd, OSSL_LIB_CTX *libctx,
crypto/openssl/providers/common/include/prov/provider_util.h
75
int ossl_prov_digest_load_from_params(PROV_DIGEST *pd,
crypto/openssl/providers/common/include/prov/provider_util.h
80
void ossl_prov_digest_reset(PROV_DIGEST *pd);
crypto/openssl/providers/common/include/prov/provider_util.h
86
const EVP_MD *ossl_prov_digest_md(const PROV_DIGEST *pd);
crypto/openssl/providers/common/include/prov/provider_util.h
87
ENGINE *ossl_prov_digest_engine(const PROV_DIGEST *pd);
crypto/openssl/providers/common/include/prov/provider_util.h
90
void ossl_prov_digest_set_md(PROV_DIGEST *pd, EVP_MD *md);
crypto/openssl/providers/common/provider_util.c
140
void ossl_prov_digest_reset(PROV_DIGEST *pd)
crypto/openssl/providers/common/provider_util.c
142
EVP_MD_free(pd->alloc_md);
crypto/openssl/providers/common/provider_util.c
143
pd->alloc_md = NULL;
crypto/openssl/providers/common/provider_util.c
144
pd->md = NULL;
crypto/openssl/providers/common/provider_util.c
146
ENGINE_finish(pd->engine);
crypto/openssl/providers/common/provider_util.c
148
pd->engine = NULL;
crypto/openssl/providers/common/provider_util.c
167
const EVP_MD *ossl_prov_digest_fetch(PROV_DIGEST *pd, OSSL_LIB_CTX *libctx,
crypto/openssl/providers/common/provider_util.c
170
EVP_MD_free(pd->alloc_md);
crypto/openssl/providers/common/provider_util.c
171
pd->md = pd->alloc_md = EVP_MD_fetch(libctx, mdname, propquery);
crypto/openssl/providers/common/provider_util.c
173
return pd->md;
crypto/openssl/providers/common/provider_util.c
176
int ossl_prov_digest_load_from_params(PROV_DIGEST *pd,
crypto/openssl/providers/common/provider_util.c
186
if (!load_common(params, &propquery, &pd->engine))
crypto/openssl/providers/common/provider_util.c
196
ossl_prov_digest_fetch(pd, ctx, p->data, propquery);
crypto/openssl/providers/common/provider_util.c
198
if (pd->md == NULL) {
crypto/openssl/providers/common/provider_util.c
204
pd->md = md;
crypto/openssl/providers/common/provider_util.c
207
if (pd->md != NULL)
crypto/openssl/providers/common/provider_util.c
211
return pd->md != NULL;
crypto/openssl/providers/common/provider_util.c
214
void ossl_prov_digest_set_md(PROV_DIGEST *pd, EVP_MD *md)
crypto/openssl/providers/common/provider_util.c
216
ossl_prov_digest_reset(pd);
crypto/openssl/providers/common/provider_util.c
217
pd->md = pd->alloc_md = md;
crypto/openssl/providers/common/provider_util.c
220
const EVP_MD *ossl_prov_digest_md(const PROV_DIGEST *pd)
crypto/openssl/providers/common/provider_util.c
222
return pd->md;
crypto/openssl/providers/common/provider_util.c
225
ENGINE *ossl_prov_digest_engine(const PROV_DIGEST *pd)
crypto/openssl/providers/common/provider_util.c
227
return pd->engine;
crypto/openssl/test/rsa_test.c
492
BIGNUM *pn = NULL, *pe = NULL, *pd = NULL;
crypto/openssl/test/rsa_test.c
544
pd = BN_bin2bn(d, sizeof(d), NULL);
crypto/openssl/test/rsa_test.c
547
|| (priv && pd == NULL)
crypto/openssl/test/rsa_test.c
548
|| !RSA_set0_key(rsa, pn, pe, pd))) {
crypto/openssl/test/rsa_test.c
551
BN_free(pd);
lib/libc/net/getproto.c
124
struct protodata *pd;
lib/libc/net/getproto.c
127
if ((pd = __protodata_init()) == NULL)
lib/libc/net/getproto.c
129
if (getprotobynumber_r(proto, &pd->proto, pd->data, sizeof(pd->data),
lib/libc/net/getprotoent.c
541
struct protodata *pd;
lib/libc/net/getprotoent.c
544
if ((pd = __protodata_init()) == NULL)
lib/libc/net/getprotoent.c
546
if (getprotoent_r(&pd->proto, pd->data, sizeof(pd->data), &rval) != 0)
lib/libc/net/getprotoname.c
138
struct protodata *pd;
lib/libc/net/getprotoname.c
141
if ((pd = __protodata_init()) == NULL)
lib/libc/net/getprotoname.c
143
if (getprotobyname_r(name, &pd->proto, pd->data, sizeof(pd->data),
lib/libc/stdlib/qsort.c
106
char *pa, *pb, *pc, *pd, *pl, *pm, *pn;
lib/libc/stdlib/qsort.c
138
pc = pd = (char *)a + (n - 1) * es;
lib/libc/stdlib/qsort.c
149
swapfunc(pc, pd, es);
lib/libc/stdlib/qsort.c
150
pd -= es;
lib/libc/stdlib/qsort.c
169
d1 = MIN(pd - pc, pn - pd - (ssize_t)es);
lib/libc/stdlib/qsort.c
173
d2 = pd - pc;
lib/libcasper/libcasper/libcasper.c
69
cap_add_pd(cap_channel_t *chan, int pd)
lib/libcasper/libcasper/libcasper.c
72
if (!fd_is_valid(pd))
lib/libcasper/libcasper/libcasper.c
74
chan->cch_pd = pd;
lib/libkvm/kvm_arm.c
196
arm_pd_entry_t pd;
lib/libkvm/kvm_arm.c
203
pd = _kvm32toh(kd, vm->l1pt[ARM_L1_IDX(va)]);
lib/libkvm/kvm_arm.c
204
if (!l1pte_valid(pd))
lib/libkvm/kvm_arm.c
206
if (l1pte_section_p(pd)) {
lib/libkvm/kvm_arm.c
208
*pa = (pd & ARM_L1_S_ADDR_MASK) + (va & ARM_L1_S_OFFSET);
lib/libkvm/kvm_arm.c
211
pte_pa = (pd & ARM_L1_C_ADDR_MASK) + l2pte_index(va) * sizeof(pte);
lib/libpam/modules/pam_exec/pam_exec.c
171
int chin[2], chout[2], cherr[2], pd;
lib/libpam/modules/pam_exec/pam_exec.c
178
pd = -1;
lib/libpam/modules/pam_exec/pam_exec.c
312
if ((pid = pdfork(&pd, 0)) == 0) {
lib/libpam/modules/pam_exec/pam_exec.c
345
pfd[0].fd = pd;
lib/libpam/modules/pam_exec/pam_exec.c
442
if (pd >= 0)
lib/libpam/modules/pam_exec/pam_exec.c
443
close(pd);
lib/libpmc/libpmc.c
1761
pmc_name_of_disposition(enum pmc_disp pd)
lib/libpmc/libpmc.c
1763
if ((int) pd >= PMC_DISP_FIRST &&
lib/libpmc/libpmc.c
1764
pd <= PMC_DISP_LAST)
lib/libpmc/libpmc.c
1765
return (pmc_disposition_names[pd]);
lib/libpmc/pmu-events/jevents.c
355
struct perf_entry_data *pd = data;
lib/libpmc/pmu-events/jevents.c
356
FILE *outfp = pd->outfp;
lib/libpmc/pmu-events/jevents.c
357
char *topic_local = pd->topic;
lib/libprocstat/libprocstat.c
2079
struct pwddesc pd;
lib/libprocstat/libprocstat.c
2085
if (!kvm_read_all(kd, (unsigned long)kp->ki_pd, &pd, sizeof(pd))) {
lib/libprocstat/libprocstat.c
2090
*maskp = pd.pd_cmask;
lib/libusb/libusb10_desc.c
720
struct libusb_platform_descriptor **pd)
lib/libusb/libusb10_desc.c
727
pd == NULL)
lib/libusb/libusb10_desc.c
755
*pd = desc;
lib/libusb/libusb20_desc.c
322
libusb20_me_encode(void *ptr, uint16_t len, const void *pd)
lib/libusb/libusb20_desc.c
337
pf = (*((struct libusb20_me_format *const *)pd))->format;
lib/libusb/libusb20_desc.c
360
LIBUSB20_ADD_BYTES(pd, pd_offset));
lib/libusb/libusb20_desc.c
379
LIBUSB20_ADD_BYTES(pd, pd_offset));
lib/libusb/libusb20_desc.c
398
LIBUSB20_ADD_BYTES(pd, pd_offset));
lib/libusb/libusb20_desc.c
420
LIBUSB20_ADD_BYTES(pd, pd_offset));
lib/libusb/libusb20_desc.c
444
ps = LIBUSB20_ADD_BYTES(pd, pd_offset);
lib/libusb/libusb20_desc.c
462
src_len = libusb20_me_get_1(pd, 0);
lib/libusb/libusb20_desc.c
466
src_len = libusb20_me_get_2(pd, 1);
lib/libusb/libusb20_desc.c
554
libusb20_me_decode(const void *ptr, uint16_t len, void *pd)
lib/libusb/libusb20_desc.c
569
pf = (*((struct libusb20_me_format **)pd))->format;
lib/libusb/libusb20_desc.c
596
*((uint8_t *)LIBUSB20_ADD_BYTES(pd,
lib/libusb/libusb20_desc.c
616
*((uint16_t *)LIBUSB20_ADD_BYTES(pd,
lib/libusb/libusb20_desc.c
639
*((uint32_t *)LIBUSB20_ADD_BYTES(pd,
lib/libusb/libusb20_desc.c
666
*((uint64_t *)LIBUSB20_ADD_BYTES(pd,
lib/libusb/libusb20_desc.c
679
ps = LIBUSB20_ADD_BYTES(pd, pd_offset);
lib/libusb/libusb20_desc.h
595
uint16_t libusb20_me_encode(void *ptr, uint16_t len, const void *pd);
lib/libusb/libusb20_desc.h
596
uint16_t libusb20_me_decode(const void *ptr, uint16_t len, void *pd);
libexec/tftpd/tests/functional.c
337
int pd[2];
libexec/tftpd/tests/functional.c
339
ATF_REQUIRE_EQ(0, pipe2(pd, O_CLOEXEC));
libexec/tftpd/tests/functional.c
400
ATF_REQUIRE_INTEQ(0, close(pd[1]));
libexec/tftpd/tests/functional.c
402
ATF_REQUIRE_INTEQ(0, read(pd[0], &pd[1], sizeof(pd[1])));
libexec/tftpd/tests/functional.c
403
ATF_REQUIRE_INTEQ(0, close(pd[0]));
sbin/ipfw/ipfw2.c
563
pr_u64(struct buf_pr *b, void *pd, int width)
sbin/ipfw/ipfw2.c
573
bcopy (pd, &u, sizeof(u));
sbin/ipfw/ipfw2.h
348
int pr_u64(struct buf_pr *bp, void *pd, int width);
stand/common/gfx_fb.c
1387
EFI_GRAPHICS_OUTPUT_BLT_PIXEL *ps, *pd;
stand/common/gfx_fb.c
1389
struct paletteentry *ps, *pd;
stand/common/gfx_fb.c
1395
pd = dst;
stand/common/gfx_fb.c
1402
pd[i].Red = alpha_blend(ps[i].Red, pd[i].Red, a);
stand/common/gfx_fb.c
1403
pd[i].Green = alpha_blend(ps[i].Green, pd[i].Green, a);
stand/common/gfx_fb.c
1404
pd[i].Blue = alpha_blend(ps[i].Blue, pd[i].Blue, a);
stand/common/gfx_fb.c
1405
pd[i].Reserved = a;
stand/efi/libefi/efipart.c
1049
pdinfo_t *pd;
stand/efi/libefi/efipart.c
1055
pd = efiblk_get_pdinfo((struct devdesc *)dev);
stand/efi/libefi/efipart.c
1056
if (pd == NULL)
stand/efi/libefi/efipart.c
1059
if (pd->pd_blkio->Media->RemovableMedia &&
stand/efi/libefi/efipart.c
1060
!pd->pd_blkio->Media->MediaPresent)
stand/efi/libefi/efipart.c
1065
bcd.dv_cache = pd->pd_bcache;
stand/efi/libefi/efipart.c
1070
offset = dev->d_offset * pd->pd_blkio->Media->BlockSize;
stand/efi/libefi/efipart.c
1083
pdinfo_t *pd;
stand/efi/libefi/efipart.c
1096
pd = efiblk_get_pdinfo((struct devdesc *)dev);
stand/efi/libefi/efipart.c
1097
if (pd == NULL)
stand/efi/libefi/efipart.c
1100
blkio = pd->pd_blkio;
stand/efi/libefi/efipart.c
138
pdinfo_t *pd = NULL;
stand/efi/libefi/efipart.c
142
return (pd);
stand/efi/libefi/efipart.c
144
STAILQ_FOREACH(pd, pdi, pd_link) {
stand/efi/libefi/efipart.c
145
if (pd->pd_unit == dev->d_unit)
stand/efi/libefi/efipart.c
146
return (pd);
stand/efi/libefi/efipart.c
148
return (pd);
stand/efi/libefi/efipart.c
165
same_handle(pdinfo_t *pd, EFI_HANDLE h)
stand/efi/libefi/efipart.c
168
return (pd->pd_handle == h || pd->pd_alias == h);
stand/efi/libefi/efipart.c
205
pdinfo_t *pd;
stand/efi/libefi/efipart.c
208
STAILQ_FOREACH(pd, pdi, pd_link) {
stand/efi/libefi/efipart.c
217
pdinfo_t *pd;
stand/efi/libefi/efipart.c
226
STAILQ_FOREACH(pd, pdi, pd_link) {
stand/efi/libefi/efipart.c
228
if (efi_devpath_match(pd->pd_devpath, parent))
stand/efi/libefi/efipart.c
232
return (pd);
stand/efi/libefi/efipart.c
356
pdinfo_t *pd;
stand/efi/libefi/efipart.c
400
if ((pd = calloc(1, sizeof(*pd))) == NULL) {
stand/efi/libefi/efipart.c
405
STAILQ_INIT(&pd->pd_part);
stand/efi/libefi/efipart.c
407
pd->pd_handle = hin[i];
stand/efi/libefi/efipart.c
408
pd->pd_devpath = devpath;
stand/efi/libefi/efipart.c
409
pd->pd_blkio = blkio;
stand/efi/libefi/efipart.c
410
STAILQ_INSERT_TAIL(&pdinfo, pd, pd_link);
stand/efi/libefi/efipart.c
416
STAILQ_FOREACH(pd, &pdinfo, pd_link) {
stand/efi/libefi/efipart.c
417
pd->pd_parent = efipart_find_parent(&pdinfo, pd->pd_devpath);
stand/efi/libefi/efipart.c
429
pdinfo_t *pd;
stand/efi/libefi/efipart.c
431
STAILQ_FOREACH(pd, plist, pd_link) {
stand/efi/libefi/efipart.c
432
if (pd_test(pd, data))
stand/efi/libefi/efipart.c
436
return (pd);
stand/efi/libefi/efipart.c
510
pdinfo_t *parent, *pd, *last;
stand/efi/libefi/efipart.c
519
STAILQ_FOREACH(pd, &pdinfo, pd_link) {
stand/efi/libefi/efipart.c
520
if (efi_devpath_match(pd->pd_devpath, cd->pd_devpath)) {
stand/efi/libefi/efipart.c
525
if (pd == NULL) {
stand/efi/libefi/efipart.c
581
efipart_testchild(pdinfo_t *dev, pdinfo_t *pd)
stand/efi/libefi/efipart.c
587
if (efi_devpath_match(dev->pd_parent->pd_devpath, pd->pd_devpath)) {
stand/efi/libefi/efipart.c
714
pdinfo_t *parent, *pd, *last;
stand/efi/libefi/efipart.c
724
STAILQ_FOREACH(pd, &pdinfo, pd_link) {
stand/efi/libefi/efipart.c
725
if (efi_devpath_match(pd->pd_devpath, hd->pd_devpath)) {
stand/efi/libefi/efipart.c
730
if (pd == NULL) {
stand/efi/libefi/efipart.c
793
pdinfo_t *pd;
stand/efi/libefi/efipart.c
805
STAILQ_FOREACH(pd, pdlist, pd_link) {
stand/efi/libefi/efipart.c
806
h = pd->pd_handle;
stand/efi/libefi/efipart.c
817
" %s%d", dev->dv_name, pd->pd_unit);
stand/efi/libefi/efipart.c
839
pd->pd_blkio = blkio;
stand/efi/libefi/efipart.c
841
pd_dev.dd.d_unit = pd->pd_unit;
stand/efi/libefi/efipart.c
887
pdinfo_t *pd;
stand/efi/libefi/efipart.c
897
pd = efiblk_get_pdinfo((struct devdesc *)dev);
stand/efi/libefi/efipart.c
898
if (pd == NULL)
stand/efi/libefi/efipart.c
901
if (pd->pd_blkio == NULL) {
stand/efi/libefi/efipart.c
902
status = OpenProtocolByHandle(pd->pd_handle, &blkio_guid,
stand/efi/libefi/efipart.c
903
(void **)&pd->pd_blkio);
stand/efi/libefi/efipart.c
908
blkio = pd->pd_blkio;
stand/efi/libefi/efipart.c
912
pd->pd_open++;
stand/efi/libefi/efipart.c
913
if (pd->pd_bcache == NULL)
stand/efi/libefi/efipart.c
914
pd->pd_bcache = bcache_allocate();
stand/efi/libefi/efipart.c
923
pd->pd_open--;
stand/efi/libefi/efipart.c
924
if (pd->pd_open == 0) {
stand/efi/libefi/efipart.c
925
pd->pd_blkio = NULL;
stand/efi/libefi/efipart.c
926
bcache_free(pd->pd_bcache);
stand/efi/libefi/efipart.c
927
pd->pd_bcache = NULL;
stand/efi/libefi/efipart.c
939
pdinfo_t *pd;
stand/efi/libefi/efipart.c
945
pd = efiblk_get_pdinfo((struct devdesc *)dev);
stand/efi/libefi/efipart.c
946
if (pd == NULL)
stand/efi/libefi/efipart.c
949
pd->pd_open--;
stand/efi/libefi/efipart.c
950
if (pd->pd_open == 0) {
stand/efi/libefi/efipart.c
951
pd->pd_blkio = NULL;
stand/efi/libefi/efipart.c
953
bcache_free(pd->pd_bcache);
stand/efi/libefi/efipart.c
954
pd->pd_bcache = NULL;
stand/efi/libefi/efipart.c
966
pdinfo_t *pd;
stand/efi/libefi/efipart.c
973
pd = efiblk_get_pdinfo((struct devdesc *)dev);
stand/efi/libefi/efipart.c
974
if (pd == NULL)
stand/efi/libefi/efipart.c
985
*(u_int *)data = pd->pd_blkio->Media->BlockSize;
stand/efi/libefi/efipart.c
988
*(uint64_t *)data = pd->pd_blkio->Media->BlockSize *
stand/efi/libefi/efipart.c
989
(pd->pd_blkio->Media->LastBlock + 1);
stand/efi/libefi/efizfs.c
112
STAILQ_FOREACH(pd, &hd->pd_part, pd_link) {
stand/efi/libefi/efizfs.c
114
efipart_hddev.dv_name, hd->pd_unit, pd->pd_unit);
stand/efi/libefi/efizfs.c
116
insert_zfs(pd->pd_handle, guid,
stand/efi/libefi/efizfs.c
117
pd->pd_handle == boot_img->DeviceHandle);
stand/efi/libefi/efizfs.c
98
pdinfo_t *hd, *pd = NULL;
stand/efi/loader/main.c
1965
pdinfo_t *hd, *pd;
stand/efi/loader/main.c
1986
STAILQ_FOREACH(pd, &hd->pd_part, pd_link) {
stand/efi/loader/main.c
1990
if (pd->pd_unit == (uint32_t)d_dev->d_slice) {
stand/efi/loader/main.c
1992
pd->pd_handle;
stand/efi/loader/main.c
898
int br, db, io, rs, rw, xo, pv, pd;
stand/efi/loader/main.c
919
pd = spcr->PciDeviceId;
stand/efi/loader/main.c
920
if (pv == 0xffff && pd == 0xffff) {
stand/efi/loader/main.c
959
} else if (pv != 0xffff && pd != 0xffff) {
stand/efi/loader/main.c
961
db, dt, pv, pd, pa, br, xo);
stand/ficl/ficl.h
1075
int isAFiclWord(FICL_DICT *pd, FICL_WORD *pFW);
stand/ficl/tools.c
112
int isAFiclWord(FICL_DICT *pd, FICL_WORD *pFW)
stand/ficl/tools.c
115
if (!dictIncludes(pd, pFW))
stand/ficl/tools.c
118
if (!dictIncludes(pd, pFW->name))
stand/ficl/tools.c
121
if ((pFW->link != NULL) && !dictIncludes(pd, pFW->link))
stand/ficl/tools.c
156
FICL_DICT *pd = vmGetDict(pVM);
stand/ficl/tools.c
159
if (!dictIncludes(pd, (void *)cp))
stand/ficl/tools.c
165
if (isAFiclWord(pd, pFW))
stand/ficl/tools.c
190
FICL_DICT *pd = vmGetDict(pVM);
stand/ficl/tools.c
205
if (isAFiclWord(pd, pFW))
stand/ficl/tools.c
214
if (isAFiclWord(pd, c.p))
stand/ficl/words.c
1771
FICL_DICT *pd = vmGetDict(pVM);
stand/ficl/words.c
1776
if ((pStep != NULL) && isAFiclWord(pd, pStep))
sys/amd64/amd64/minidump_machdep.c
167
uint64_t *pml4, *pdp, *pd, *pt, pa;
sys/amd64/amd64/minidump_machdep.c
219
pd = (uint64_t *)PHYS_TO_DMAP(pdpe & PG_FRAME);
sys/amd64/amd64/minidump_machdep.c
221
pde = atomic_load_64(&pd[pmap_pde_index(va)]);
sys/amd64/amd64/minidump_machdep.c
363
pd = (uint64_t *)PHYS_TO_DMAP(pa);
sys/amd64/amd64/minidump_machdep.c
364
error = blk_write(di, (char *)pd, 0, PAGE_SIZE);
sys/amd64/amd64/pmap.c
10718
pd_entry_t *pde, pd;
sys/amd64/amd64/pmap.c
10735
pd = *pde;
sys/amd64/amd64/pmap.c
10736
KASSERT((pd & X86_PG_V) != 0,
sys/amd64/amd64/pmap.c
10737
("invalid pd va %#lx pde %#lx pd %#lx", va, (u_long)pde, pd));
sys/amd64/amd64/pmap.c
10738
if ((pd & X86_PG_PS) != 0)
sys/amd64/amd64/pmap.c
10739
return ((pd & PG_PS_FRAME) | (va & PDRMASK));
sys/amd64/amd64/pmap.c
10852
pd_entry_t *pde, pd;
sys/amd64/amd64/pmap.c
10889
pd = *pde;
sys/amd64/amd64/pmap.c
10890
KASSERT((pd & X86_PG_V) != 0,
sys/amd64/amd64/pmap.c
10892
(u_long)pde, pd));
sys/amd64/amd64/pmap.c
10893
if ((pd & X86_PG_PS) != 0) {
sys/amd64/amd64/pmap.c
10896
(u_long)pde, pd));
sys/amd64/amd64/pmap.c
10900
pd, len));
sys/amd64/amd64/pmap.c
12012
pd_entry_t *pd, pde;
sys/amd64/amd64/pmap.c
12116
pd = (pd_entry_t *)PHYS_TO_DMAP(pa);
sys/amd64/amd64/pmap.c
12119
pde = pd[k];
sys/amd64/amd64/pmap.c
12260
pd_entry_t *pd;
sys/amd64/amd64/pmap.c
12283
pd = (pd_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(pg2));
sys/amd64/amd64/pmap.c
12285
if ((pd[i2] & PG_V) == 0)
sys/amd64/amd64/pmap.c
12287
pg1 = PHYS_TO_VM_PAGE(pd[i2] & PG_FRAME);
sys/amd64/amd64/pmap.c
12289
ptpages_show_complain(1, i2, pd[i2]);
sys/amd64/amd64/pmap.c
4084
pd_entry_t *pd;
sys/amd64/amd64/pmap.c
4116
pd = pmap_pde(pmap, va);
sys/amd64/amd64/pmap.c
4117
*pd = 0;
sys/amd64/amd64/pmap.c
4585
pd_entry_t *pd;
sys/amd64/amd64/pmap.c
4685
pd = (pd_entry_t *)PHYS_TO_DMAP(*pdp & PG_FRAME);
sys/amd64/amd64/pmap.c
4688
pd = &pd[pmap_pde_index(va)];
sys/amd64/amd64/pmap.c
4689
KASSERT((*pd & PG_V) == 0,
sys/amd64/amd64/pmap.c
4690
("pmap %p va %#lx pd %#lx", pmap, va, *pd));
sys/amd64/amd64/pmap.c
4691
*pd = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4767
pd_entry_t *pd, PG_V;
sys/amd64/amd64/pmap.c
4780
pd = pmap_pde(pmap, va);
sys/amd64/amd64/pmap.c
4786
if (pd != NULL && (*pd & (PG_PS | PG_V)) == (PG_PS | PG_V)) {
sys/amd64/amd64/pmap.c
4787
if (!pmap_demote_pde_locked(pmap, pd, va, lockp)) {
sys/amd64/amd64/pmap.c
4792
pd = NULL;
sys/amd64/amd64/pmap.c
4800
if (pd != NULL && (*pd & PG_V) != 0) {
sys/amd64/amd64/pmap.c
4801
m = PHYS_TO_VM_PAGE(*pd & PG_FRAME);
sys/arm/freescale/imx/imx6_sdma.h
147
uint32_t pd;
sys/cam/ctl/ctl.c
9761
struct scsi_vpd_port_designation *pd;
sys/cam/ctl/ctl.c
9809
pd = &sp->design[0];
sys/cam/ctl/ctl.c
9818
scsi_ulto2b(port->targ_port, pd->relative_port_id);
sys/cam/ctl/ctl.c
9821
memcpy(pd->initiator_transportid,
sys/cam/ctl/ctl.c
9825
scsi_ulto2b(iid_len, pd->initiator_transportid_length);
sys/cam/ctl/ctl.c
9827
(&pd->initiator_transportid[iid_len]);
sys/cam/ctl/ctl.c
9835
pd = (struct scsi_vpd_port_designation *)
sys/cddl/dev/kinst/aarch64/kinst_isa.c
295
dtrace_kinst_probedesc_t *pd;
sys/cddl/dev/kinst/aarch64/kinst_isa.c
301
pd = opaque;
sys/cddl/dev/kinst/aarch64/kinst_isa.c
306
if (strcmp(func, pd->kpd_func) != 0)
sys/cddl/dev/kinst/aarch64/kinst_isa.c
386
if (pd->kpd_off != -1 && off != pd->kpd_off)
sys/cddl/dev/kinst/amd64/kinst_isa.c
500
dtrace_kinst_probedesc_t *pd;
sys/cddl/dev/kinst/amd64/kinst_isa.c
506
pd = opaque;
sys/cddl/dev/kinst/amd64/kinst_isa.c
510
if (strcmp(func, pd->kpd_func) != 0)
sys/cddl/dev/kinst/amd64/kinst_isa.c
543
if (pd->kpd_off != -1 && off != pd->kpd_off) {
sys/cddl/dev/kinst/amd64/kinst_isa.c
560
if (pd->kpd_off != -1)
sys/cddl/dev/kinst/kinst.c
184
dtrace_kinst_probedesc_t *pd;
sys/cddl/dev/kinst/kinst.c
186
pd = arg;
sys/cddl/dev/kinst/kinst.c
187
if (pd->kpd_mod[0] != '\0' && strcmp(pd->kpd_mod, lf->filename) != 0)
sys/cddl/dev/kinst/kinst.c
201
dtrace_kinst_probedesc_t *pd;
sys/cddl/dev/kinst/kinst.c
206
pd = (dtrace_kinst_probedesc_t *)addr;
sys/cddl/dev/kinst/kinst.c
207
pd->kpd_func[sizeof(pd->kpd_func) - 1] = '\0';
sys/cddl/dev/kinst/kinst.c
208
pd->kpd_mod[sizeof(pd->kpd_mod) - 1] = '\0';
sys/cddl/dev/kinst/kinst.c
211
error = linker_file_foreach(kinst_linker_file_cb, pd);
sys/cddl/dev/kinst/riscv/kinst_isa.c
446
dtrace_kinst_probedesc_t *pd;
sys/cddl/dev/kinst/riscv/kinst_isa.c
453
pd = opaque;
sys/cddl/dev/kinst/riscv/kinst_isa.c
458
if (strcmp(func, pd->kpd_func) != 0)
sys/cddl/dev/kinst/riscv/kinst_isa.c
511
if (pd->kpd_off != -1 && off != pd->kpd_off)
sys/dev/ath/if_ath_pci.c
172
const struct pci_device_table *pd;
sys/dev/ath/if_ath_pci.c
177
pd = PCI_MATCH(dev, ath_pci_id_table);
sys/dev/ath/if_ath_pci.c
178
if (pd)
sys/dev/ath/if_ath_pci.c
179
sc->sc_pci_devinfo = pd->driver_data;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1050
struct ib_pd *ib_pd = ib_ah->pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1052
struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
1053
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1066
ah->qplib_ah.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1210
struct bnxt_re_pd *pd,
sys/dev/bnxt/bnxt_re/ib_verbs.c
1222
context = pd->ibpd.uobject->context;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1265
struct bnxt_re_pd *pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1268
struct ib_pd *ib_pd = ib_srq->pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1273
pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
1274
rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1290
context = pd->ibpd.uobject->context;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1307
srq->qplib_srq.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1327
rc = bnxt_re_init_user_srq(rdev, pd, srq, udata);
sys/dev/bnxt/bnxt_re/ib_verbs.c
1696
struct bnxt_re_pd *pd, struct bnxt_re_qp *qp,
sys/dev/bnxt/bnxt_re/ib_verbs.c
1710
context = pd->ibpd.uobject->context;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1794
static struct bnxt_re_ah *bnxt_re_create_shadow_qp_ah(struct bnxt_re_pd *pd,
sys/dev/bnxt/bnxt_re/ib_verbs.c
1798
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1810
ah->qplib_ah.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1850
struct bnxt_re_pd *pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1869
ib_pd = gsi_qp->ib_qp.pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1870
pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
1882
sah = bnxt_re_create_shadow_qp_ah(pd, &rdev->qplib_res,
sys/dev/bnxt/bnxt_re/ib_verbs.c
1894
static struct bnxt_re_qp *bnxt_re_create_shadow_qp(struct bnxt_re_pd *pd,
sys/dev/bnxt/bnxt_re/ib_verbs.c
1898
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
1912
qp->qplib_qp.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
2160
static int bnxt_re_init_qp_attr(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd,
sys/dev/bnxt/bnxt_re/ib_verbs.c
2177
context = pd->ibpd.uobject->context;
sys/dev/bnxt/bnxt_re/ib_verbs.c
2183
qplqp->pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
2247
rc = bnxt_re_init_user_qp(rdev, pd, qp, udata);
sys/dev/bnxt/bnxt_re/ib_verbs.c
2253
struct bnxt_re_pd *pd)
sys/dev/bnxt/bnxt_re/ib_verbs.c
2269
sqp = bnxt_re_create_shadow_qp(pd, &rdev->qplib_res, &qp->qplib_qp);
sys/dev/bnxt/bnxt_re/ib_verbs.c
2280
sah = bnxt_re_create_shadow_qp_ah(pd, &rdev->qplib_res,
sys/dev/bnxt/bnxt_re/ib_verbs.c
2312
static int bnxt_re_create_gsi_qp(struct bnxt_re_qp *qp, struct bnxt_re_pd *pd)
sys/dev/bnxt/bnxt_re/ib_verbs.c
2339
rc = bnxt_re_create_shadow_gsi(qp, pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
2387
struct bnxt_re_pd *pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
2395
pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
2396
rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
2423
rc = bnxt_re_init_qp_attr(qp, pd, qp_init_attr, udata);
sys/dev/bnxt/bnxt_re/ib_verbs.c
2429
rc = bnxt_re_create_gsi_qp(qp, pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
4858
struct bnxt_re_pd *pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
4864
pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
4865
rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
4874
mr->qplib_mr.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
4969
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
4970
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
4990
mr->qplib_mr.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
5036
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
5037
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
5049
mw->qplib_mw.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
5136
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
5137
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
5166
mr->qplib_mr.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
5246
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
5294
mr->qplib_mr.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
588
static void bnxt_re_legacy_create_fence_wqe(struct bnxt_re_pd *pd)
sys/dev/bnxt/bnxt_re/ib_verbs.c
590
struct bnxt_re_legacy_fence_data *fence = &pd->fence;
sys/dev/bnxt/bnxt_re/ib_verbs.c
593
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
620
struct ib_pd *ib_pd = qp->ib_qp.pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
621
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
622
struct bnxt_re_legacy_fence_data *fence = &pd->fence;
sys/dev/bnxt/bnxt_re/ib_verbs.c
636
wqe.bind.r_key, qp->qplib_qp.id, pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
647
static int bnxt_re_legacy_create_fence_mr(struct bnxt_re_pd *pd)
sys/dev/bnxt/bnxt_re/ib_verbs.c
650
struct bnxt_re_legacy_fence_data *fence = &pd->fence;
sys/dev/bnxt/bnxt_re/ib_verbs.c
651
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
686
mr->qplib_mr.pd = &pd->qplib_pd;
sys/dev/bnxt/bnxt_re/ib_verbs.c
722
ib_mw = bnxt_re_alloc_mw(&pd->ibpd, IB_MW_TYPE_1, NULL);
sys/dev/bnxt/bnxt_re/ib_verbs.c
726
"Failed to create fence-MW for PD: %p\n", pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
732
bnxt_re_legacy_create_fence_wqe(pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
754
static void bnxt_re_legacy_destroy_fence_mr(struct bnxt_re_pd *pd)
sys/dev/bnxt/bnxt_re/ib_verbs.c
756
struct bnxt_re_legacy_fence_data *fence = &pd->fence;
sys/dev/bnxt/bnxt_re/ib_verbs.c
757
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
818
struct bnxt_re_pd *pd = to_bnxt_re(ib_pd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
819
struct bnxt_re_dev *rdev = pd->rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
822
bnxt_re_legacy_destroy_fence_mr(pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
826
&pd->qplib_pd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
846
struct bnxt_re_pd *pd = container_of(ibpd, struct bnxt_re_pd, ibpd);
sys/dev/bnxt/bnxt_re/ib_verbs.c
848
pd->rdev = rdev;
sys/dev/bnxt/bnxt_re/ib_verbs.c
849
if (bnxt_qplib_alloc_pd(&rdev->qplib_res, &pd->qplib_pd)) {
sys/dev/bnxt/bnxt_re/ib_verbs.c
865
resp.pdid = pd->qplib_pd.id;
sys/dev/bnxt/bnxt_re/ib_verbs.c
888
if (bnxt_re_legacy_create_fence_mr(pd))
sys/dev/bnxt/bnxt_re/ib_verbs.c
900
&pd->qplib_pd);
sys/dev/bnxt/bnxt_re/ib_verbs.h
348
struct ib_mr *bnxt_re_get_dma_mr(struct ib_pd *pd, int mr_access_flags);
sys/dev/bnxt/bnxt_re/ib_verbs.h
357
struct ib_mr *bnxt_re_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/bnxt/bnxt_re/ib_verbs.h
362
u64 virt_addr, int mr_access_flags, struct ib_pd *pd,
sys/dev/bnxt/bnxt_re/qplib_fp.c
1230
req.pd_id = cpu_to_le32(qp->pd->id);
sys/dev/bnxt/bnxt_re/qplib_fp.c
675
req.pd_id = cpu_to_le32(srq->pd->id);
sys/dev/bnxt/bnxt_re/qplib_fp.c
986
req.pd_id = cpu_to_le32(qp->pd->id);
sys/dev/bnxt/bnxt_re/qplib_fp.h
329
struct bnxt_qplib_pd *pd;
sys/dev/bnxt/bnxt_re/qplib_fp.h
77
struct bnxt_qplib_pd *pd;
sys/dev/bnxt/bnxt_re/qplib_res.c
772
int bnxt_qplib_alloc_pd(struct bnxt_qplib_res *res, struct bnxt_qplib_pd *pd)
sys/dev/bnxt/bnxt_re/qplib_res.c
787
pd->id = bit_num;
sys/dev/bnxt/bnxt_re/qplib_res.c
795
struct bnxt_qplib_pd *pd)
sys/dev/bnxt/bnxt_re/qplib_res.c
798
if (test_and_set_bit(pd->id, pdt->tbl)) {
sys/dev/bnxt/bnxt_re/qplib_res.c
800
pd->id);
sys/dev/bnxt/bnxt_re/qplib_res.c
805
pd->id = pdt->max - 1;
sys/dev/bnxt/bnxt_re/qplib_res.h
475
struct bnxt_qplib_pd *pd);
sys/dev/bnxt/bnxt_re/qplib_res.h
478
struct bnxt_qplib_pd *pd);
sys/dev/bnxt/bnxt_re/qplib_sp.c
487
req.pd_id = cpu_to_le32(ah->pd->id);
sys/dev/bnxt/bnxt_re/qplib_sp.c
580
req.pd_id = cpu_to_le32(mrw->pd->id);
sys/dev/bnxt/bnxt_re/qplib_sp.c
694
req.key = cpu_to_le32(mr->pd->id);
sys/dev/bnxt/bnxt_re/qplib_sp.h
111
struct bnxt_qplib_pd *pd;
sys/dev/bnxt/bnxt_re/qplib_sp.h
96
struct bnxt_qplib_pd *pd;
sys/dev/bxe/bxe_elink.c
6589
uint16_t pd, gp_status4;
sys/dev/bxe/bxe_elink.c
6602
&pd);
sys/dev/bxe/bxe_elink.c
6603
if (pd & (1<<15))
sys/dev/bxe/bxe_elink.c
6607
pd, vars->link_status);
sys/dev/cxgbe/iw_cxgbe/device.c
129
rdev->stats.pd.total = T4_MAX_NUM_PD;
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
132
struct c4iw_stat pd;
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
461
u32 pd;
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
932
struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
937
struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
939
struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, u64
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
941
struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc);
sys/dev/cxgbe/iw_cxgbe/iw_cxgbe.h
951
struct ib_qp *c4iw_create_qp(struct ib_pd *pd,
sys/dev/cxgbe/iw_cxgbe/mem.c
345
struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc)
sys/dev/cxgbe/iw_cxgbe/mem.c
353
CTR2(KTR_IW_CXGBE, "%s ib_pd %p", __func__, pd);
sys/dev/cxgbe/iw_cxgbe/mem.c
354
php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/mem.c
389
struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/cxgbe/iw_cxgbe/mem.c
401
CTR2(KTR_IW_CXGBE, "%s ib_pd %p", __func__, pd);
sys/dev/cxgbe/iw_cxgbe/mem.c
409
php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/mem.c
421
mhp->umem = ib_umem_get(pd->uobject->context, start, length, acc, 0);
sys/dev/cxgbe/iw_cxgbe/mem.c
492
struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/cxgbe/iw_cxgbe/mem.c
505
php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/mem.c
548
struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd,
sys/dev/cxgbe/iw_cxgbe/mem.c
560
php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/provider.c
210
c4iw_deallocate_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/cxgbe/iw_cxgbe/provider.c
212
struct c4iw_pd *php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/provider.c
215
CTR3(KTR_IW_CXGBE, "%s: pd %p, pdid 0x%x", __func__, pd, php->pdid);
sys/dev/cxgbe/iw_cxgbe/provider.c
219
rhp->rdev.stats.pd.cur--;
sys/dev/cxgbe/iw_cxgbe/provider.c
224
c4iw_allocate_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/cxgbe/iw_cxgbe/provider.c
226
struct c4iw_pd *php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/provider.c
227
struct ib_device *ibdev = pd->device;
sys/dev/cxgbe/iw_cxgbe/provider.c
232
pd, udata);
sys/dev/cxgbe/iw_cxgbe/provider.c
249
rhp->rdev.stats.pd.cur++;
sys/dev/cxgbe/iw_cxgbe/provider.c
250
if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max)
sys/dev/cxgbe/iw_cxgbe/provider.c
251
rhp->rdev.stats.pd.max = rhp->rdev.stats.pd.cur;
sys/dev/cxgbe/iw_cxgbe/qp.c
1401
wqe->u.init.pdid = cpu_to_be32(qhp->attr.pd);
sys/dev/cxgbe/iw_cxgbe/qp.c
1714
c4iw_create_qp(struct ib_pd *pd, struct ib_qp_init_attr *attrs,
sys/dev/cxgbe/iw_cxgbe/qp.c
1729
CTR2(KTR_IW_CXGBE, "%s ib_pd %p", __func__, pd);
sys/dev/cxgbe/iw_cxgbe/qp.c
1734
php = to_c4iw_pd(pd);
sys/dev/cxgbe/iw_cxgbe/qp.c
1756
ucontext = pd->uobject ? to_c4iw_ucontext(pd->uobject->context) : NULL;
sys/dev/cxgbe/iw_cxgbe/qp.c
1789
qhp->attr.pd = php->pdid;
sys/dev/hwpmc/hwpmc_amd.c
187
const struct amd_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
198
pd = &amd_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_amd.c
202
pd->pm_descr.pd_class);
sys/dev/hwpmc/hwpmc_amd.c
204
tmp = rdmsr(pd->pm_perfctr); /* RDMSR serializes */
sys/dev/hwpmc/hwpmc_amd.c
233
const struct amd_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
241
pd = &amd_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_amd.c
251
wrmsr(pd->pm_perfctr, v);
sys/dev/hwpmc/hwpmc_amd.c
331
const struct pmc_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
339
pd = &amd_pmcdesc[ri].pm_descr;
sys/dev/hwpmc/hwpmc_amd.c
342
if (pd->pd_class != a->pm_class)
sys/dev/hwpmc/hwpmc_amd.c
351
((pd->pd_caps & PMC_CAP_PRECISE) == 0))
sys/dev/hwpmc/hwpmc_amd.c
446
const struct amd_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
454
pd = &amd_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_amd.c
463
KASSERT(AMD_PMC_IS_STOPPED(pd->pm_evsel),
sys/dev/hwpmc/hwpmc_amd.c
465
ri, cpu, pd->pm_descr.pd_name));
sys/dev/hwpmc/hwpmc_amd.c
472
wrmsr(pd->pm_evsel, config);
sys/dev/hwpmc/hwpmc_amd.c
482
const struct amd_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
491
pd = &amd_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_amd.c
493
KASSERT(!AMD_PMC_IS_STOPPED(pd->pm_evsel),
sys/dev/hwpmc/hwpmc_amd.c
495
__LINE__, ri, cpu, pd->pm_descr.pd_name));
sys/dev/hwpmc/hwpmc_amd.c
501
wrmsr(pd->pm_evsel, config);
sys/dev/hwpmc/hwpmc_amd.c
511
if (rdmsr(pd->pm_perfctr) & (1 << (pd->pm_descr.pd_width - 1)))
sys/dev/hwpmc/hwpmc_amd.c
636
const struct amd_descr *pd;
sys/dev/hwpmc/hwpmc_amd.c
645
pd = &amd_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_amd.c
647
strlcpy(pi->pm_name, pd->pm_descr.pd_name, sizeof(pi->pm_name));
sys/dev/hwpmc/hwpmc_amd.c
648
pi->pm_class = pd->pm_descr.pd_class;
sys/dev/hwpmc/hwpmc_cmn600.c
317
const struct pmc_descr *pd;
sys/dev/hwpmc/hwpmc_cmn600.c
334
pd = &desc->pd_descr;
sys/dev/hwpmc/hwpmc_cmn600.c
339
if (pd->pd_class != a->pm_class)
sys/dev/hwpmc/hwpmc_cmn600.c
563
struct pmc_descr *pd;
sys/dev/hwpmc/hwpmc_cmn600.c
572
pd = &cmn600desc(ri)->pd_descr;
sys/dev/hwpmc/hwpmc_cmn600.c
574
strlcpy(pi->pm_name, pd->pd_name, sizeof(pi->pm_name));
sys/dev/hwpmc/hwpmc_cmn600.c
575
pi->pm_class = pd->pd_class;
sys/dev/hwpmc/hwpmc_dmc620.c
288
const struct pmc_descr *pd;
sys/dev/hwpmc/hwpmc_dmc620.c
300
pd = &dmc620desc(class, cpu, ri)->pd_descr;
sys/dev/hwpmc/hwpmc_dmc620.c
306
if (pd->pd_class != a->pm_class)
sys/dev/hwpmc/hwpmc_dmc620.c
428
struct pmc_descr *pd;
sys/dev/hwpmc/hwpmc_dmc620.c
437
pd = &dmc620desc(class, cpu, ri)->pd_descr;
sys/dev/hwpmc/hwpmc_dmc620.c
439
strlcpy(pi->pm_name, pd->pd_name, sizeof(pi->pm_name));
sys/dev/hwpmc/hwpmc_dmc620.c
440
pi->pm_class = pd->pd_class;
sys/dev/hwpmc/hwpmc_soft.c
159
const struct soft_descr *pd;
sys/dev/hwpmc/hwpmc_soft.c
168
pd = &soft_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_soft.c
170
strlcpy(pi->pm_name, pd->pm_descr.pd_name, sizeof(pi->pm_name));
sys/dev/hwpmc/hwpmc_soft.c
171
pi->pm_class = pd->pm_descr.pd_class;
sys/dev/hwpmc/hwpmc_tsc.c
114
const struct tsc_descr *pd;
sys/dev/hwpmc/hwpmc_tsc.c
122
pd = &tsc_pmcdesc[ri];
sys/dev/hwpmc/hwpmc_tsc.c
124
strlcpy(pi->pm_name, pd->pm_descr.pd_name, sizeof(pi->pm_name));
sys/dev/hwpmc/hwpmc_tsc.c
125
pi->pm_class = pd->pm_descr.pd_class;
sys/dev/irdma/fbsd_kcompat.h
216
struct ib_mr *irdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 len,
sys/dev/irdma/fbsd_kcompat.h
219
struct ib_mr *irdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/irdma/fbsd_kcompat.h
226
struct ib_mw *irdma_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/irdma/fbsd_kcompat.h
256
int irdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
sys/dev/irdma/irdma_cm.c
3778
ibmr->pd = &iwpd->ibpd;
sys/dev/irdma/irdma_ctrl.c
2250
cqp = qp->pd->dev->cqp;
sys/dev/irdma/irdma_ctrl.c
2309
cqp = qp->pd->dev->cqp;
sys/dev/irdma/irdma_ctrl.c
250
irdma_sc_pd_init(struct irdma_sc_dev *dev, struct irdma_sc_pd *pd, u32 pd_id,
sys/dev/irdma/irdma_ctrl.c
253
pd->pd_id = pd_id;
sys/dev/irdma/irdma_ctrl.c
254
pd->abi_ver = abi_ver;
sys/dev/irdma/irdma_ctrl.c
255
pd->dev = dev;
sys/dev/irdma/irdma_ctrl.c
471
info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags ||
sys/dev/irdma/irdma_ctrl.c
473
info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags)
sys/dev/irdma/irdma_ctrl.c
476
qp->dev = info->pd->dev;
sys/dev/irdma/irdma_ctrl.c
485
qp->pd = info->pd;
sys/dev/irdma/irdma_ctrl.c
487
info->qp_uk_init_info.wqe_alloc_db = qp->pd->dev->wqe_alloc_db;
sys/dev/irdma/irdma_ctrl.c
493
pble_obj_cnt = info->pd->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
sys/dev/irdma/irdma_kcompat.c
107
irdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/irdma/irdma_kcompat.c
110
struct irdma_device *iwdev = to_iwdev(pd->device);
sys/dev/irdma/irdma_kcompat.c
1270
irdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 len,
sys/dev/irdma/irdma_kcompat.c
1275
struct irdma_device *iwdev = to_iwdev(pd->device);
sys/dev/irdma/irdma_kcompat.c
1287
region = ib_umem_get(pd->uobject->context, start, len, access, 0);
sys/dev/irdma/irdma_kcompat.c
1301
iwmr = irdma_alloc_iwmr(region, pd, virt, req.reg_type);
sys/dev/irdma/irdma_kcompat.c
131
iwmr->ibmr.pd = pd;
sys/dev/irdma/irdma_kcompat.c
132
iwmr->ibmr.device = pd->device;
sys/dev/irdma/irdma_kcompat.c
1369
iwmr->ibmr.pd = new_pd;
sys/dev/irdma/irdma_kcompat.c
284
irdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/irdma/irdma_kcompat.c
286
struct irdma_pd *iwpd = to_iwpd(pd);
sys/dev/irdma/irdma_kcompat.c
287
struct irdma_device *iwdev = to_iwdev(pd->device);
sys/dev/irdma/irdma_kcompat.c
351
struct irdma_pd *pd, u8 user_pri)
sys/dev/irdma/irdma_kcompat.c
359
spin_lock_irqsave(&pd->udqp_list_lock, flags);
sys/dev/irdma/irdma_kcompat.c
360
list_for_each_safe(list_node, tmp_node, &pd->udqp_list) {
sys/dev/irdma/irdma_kcompat.c
377
spin_unlock_irqrestore(&pd->udqp_list_lock, flags);
sys/dev/irdma/irdma_kcompat.c
388
spin_unlock_irqrestore(&pd->udqp_list_lock, flags);
sys/dev/irdma/irdma_kcompat.c
431
struct irdma_pd *pd,
sys/dev/irdma/irdma_kcompat.c
456
irdma_find_qp_update_qs(iwdev->rf, pd, vlan_prio);
sys/dev/irdma/irdma_kcompat.c
533
struct irdma_pd *pd = to_iwpd(ib_ah->pd);
sys/dev/irdma/irdma_kcompat.c
535
struct irdma_device *iwdev = to_iwdev(ib_ah->pd->device);
sys/dev/irdma/irdma_kcompat.c
557
ah->pd = pd;
sys/dev/irdma/irdma_kcompat.c
585
ah_info->pd_idx = pd->sc_pd.pd_id;
sys/dev/irdma/irdma_kcompat.c
599
err = irdma_create_ah_vlan_tag(iwdev, pd, ah_info, &sgid_attr, dmac);
sys/dev/irdma/irdma_kcompat.c
797
init_info.pd = &iwpd->sc_pd;
sys/dev/irdma/irdma_puda.c
718
qp->pd = &rsrc->sc_pd;
sys/dev/irdma/irdma_puda.c
742
ukqp->wqe_alloc_db = qp->pd->dev->wqe_alloc_db;
sys/dev/irdma/irdma_type.h
1084
struct irdma_sc_pd *pd;
sys/dev/irdma/irdma_type.h
1223
void irdma_sc_pd_init(struct irdma_sc_dev *dev, struct irdma_sc_pd *pd, u32 pd_id,
sys/dev/irdma/irdma_type.h
493
struct irdma_sc_pd *pd;
sys/dev/irdma/irdma_utils.c
2664
iwmr.ibmr.pd = &iwpd->ibpd;
sys/dev/irdma/irdma_verbs.c
1872
struct irdma_pd *iwpd = to_iwpd(iwmr->ibmr.pd);
sys/dev/irdma/irdma_verbs.c
1911
struct ib_pd *pd = iwmr->ibmr.pd;
sys/dev/irdma/irdma_verbs.c
1912
struct irdma_pd *iwpd = to_iwpd(pd);
sys/dev/irdma/irdma_verbs.c
1927
info->all_memory = (pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) ? true : false;
sys/dev/irdma/irdma_verbs.c
2002
struct ib_pd *pd = iwmr->ibmr.pd;
sys/dev/irdma/irdma_verbs.c
2003
struct irdma_pd *iwpd = to_iwpd(pd);
sys/dev/irdma/irdma_verbs.c
2019
if ((pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) || iwmr->dma_mr)
sys/dev/irdma/irdma_verbs.c
2064
struct ib_pd *pd, u64 virt,
sys/dev/irdma/irdma_verbs.c
2077
iwmr->ibmr.pd = pd;
sys/dev/irdma/irdma_verbs.c
2078
iwmr->ibmr.device = pd->device;
sys/dev/irdma/irdma_verbs.c
2236
struct irdma_pd *iwpd = to_iwpd(ib_mr->pd);
sys/dev/irdma/irdma_verbs.c
2290
struct ib_pd *pd = iwmr->ibmr.pd;
sys/dev/irdma/irdma_verbs.c
2294
region = ib_umem_get(pd->uobject->context, start, len, iwmr->access, 0);
sys/dev/irdma/irdma_verbs.c
2305
iwmr->ibmr.pd = pd;
sys/dev/irdma/irdma_verbs.c
2334
irdma_reg_phys_mr(struct ib_pd *pd, u64 addr, u64 size, int access,
sys/dev/irdma/irdma_verbs.c
2337
struct irdma_device *iwdev = to_iwdev(pd->device);
sys/dev/irdma/irdma_verbs.c
2347
iwmr->ibmr.pd = pd;
sys/dev/irdma/irdma_verbs.c
2348
iwmr->ibmr.device = pd->device;
sys/dev/irdma/irdma_verbs.c
2388
irdma_get_dma_mr(struct ib_pd *pd, int acc)
sys/dev/irdma/irdma_verbs.c
2392
return irdma_reg_phys_mr(pd, 0, 0, acc, &kva, true);
sys/dev/irdma/irdma_verbs.c
820
struct irdma_pd *iwpd = to_iwpd(ibqp->pd);
sys/dev/irdma/irdma_verbs.h
386
struct ib_pd *pd, u64 virt,
sys/dev/irdma/irdma_verbs.h
88
struct irdma_pd *pd;
sys/dev/iser/icl_iser.h
351
struct ib_pd *pd;
sys/dev/iser/iser_verbs.c
222
device->pd = ib_alloc_pd(device->ib_device, IB_PD_UNSAFE_GLOBAL_RKEY);
sys/dev/iser/iser_verbs.c
223
if (IS_ERR(device->pd))
sys/dev/iser/iser_verbs.c
255
device->mr = device->pd->__internal_mr;
sys/dev/iser/iser_verbs.c
278
ib_dealloc_pd(device->pd);
sys/dev/iser/iser_verbs.c
303
(void)ib_dealloc_pd(device->pd);
sys/dev/iser/iser_verbs.c
309
device->pd = NULL;
sys/dev/iser/iser_verbs.c
314
struct ib_pd *pd,
sys/dev/iser/iser_verbs.c
319
res->mr = ib_alloc_mr(pd, IB_MR_TYPE_MEM_REG, ISCSI_ISER_SG_TABLESIZE + 1);
sys/dev/iser/iser_verbs.c
337
iser_create_fastreg_desc(struct ib_device *ib_device, struct ib_pd *pd)
sys/dev/iser/iser_verbs.c
343
ret = iser_alloc_reg_res(ib_device, pd, &desc->rsc);
sys/dev/iser/iser_verbs.c
370
desc = iser_create_fastreg_desc(device->ib_device, device->pd);
sys/dev/iser/iser_verbs.c
467
ret = rdma_create_qp(ib_conn->cma_id, device->pd, &init_attr);
sys/dev/mfi/mfi.c
1601
detail->args.pd.device_id) {
sys/dev/mfi/mfi.c
1625
sc->mfi_cam_rescan_cb(sc, detail->args.pd.device_id);
sys/dev/mfi/mfireg.h
1007
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1010
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1016
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1021
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1026
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1031
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
1308
uint8_t pd;
sys/dev/mfi/mfireg.h
1311
} pd[MFI_MAX_ROW_SIZE];
sys/dev/mfi/mfireg.h
958
struct mfi_evt_pd pd;
sys/dev/mfi/mfireg.h
988
struct mfi_evt_pd pd;
sys/dev/mlx4/device.h
1099
int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access,
sys/dev/mlx4/device.h
1103
int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type,
sys/dev/mlx4/device.h
1398
int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages,
sys/dev/mlx4/device.h
686
u32 pd;
sys/dev/mlx4/device.h
698
u32 pd;
sys/dev/mlx4/mlx4_core/mlx4.h
376
__be32 pd;
sys/dev/mlx4/mlx4_core/mlx4_mr.c
1049
int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
1070
err = mlx4_mr_alloc(dev, pd, 0, 0, access, max_pages,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
376
mmr->pd = be32_to_cpu((*mpt_entry)->pd_flags) & MLX4_MPT_PD_MASK;
sys/dev/mlx4/mlx4_core/mlx4_mr.c
422
static int mlx4_mr_alloc_reserved(struct mlx4_dev *dev, u32 mridx, u32 pd,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
428
mr->pd = pd;
sys/dev/mlx4/mlx4_core/mlx4_mr.c
532
int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
542
err = mlx4_mr_alloc_reserved(dev, index, pd, iova, size,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
653
mpt_entry->pd_flags = cpu_to_be32(mr->pd | MLX4_MPT_PD_FLAG_EN_INV);
sys/dev/mlx4/mlx4_core/mlx4_mr.c
820
int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type,
sys/dev/mlx4/mlx4_core/mlx4_mr.c
836
mw->pd = pd;
sys/dev/mlx4/mlx4_core/mlx4_mr.c
865
mpt_entry->pd_flags = cpu_to_be32(mw->pd);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
2714
u32 pd;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
2729
pd = mr_get_pd(inbox->buf);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
2730
pd_slave = (pd >> 17) & 0x7f;
sys/dev/mlx4/mlx4_core/mlx4_srq.c
198
srq_context->pd = cpu_to_be32(pdn);
sys/dev/mlx4/mlx4_en/mlx4_en_resources.c
50
context->pd = cpu_to_be32(mdev->priv_pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
101
struct ib_pd *pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
439
struct ib_pd *pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
718
struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc);
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
721
struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
725
struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
728
struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
759
struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
783
struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int mr_access_flags,
sys/dev/mlx4/mlx4_ib/mlx4_ib.h
882
int mr_access_flags, struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
106
ah->av.eth.port_pd = cpu_to_be32(to_mpd(pd)->pdn | (ah_attr->port_num << 24));
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
135
if (rdma_port_get_link_layer(ib_ah->pd->device, ah_attr->port_num) == IB_LINK_LAYER_ETHERNET) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
168
if (rdma_port_get_link_layer(ah->pd->device, ah_attr->port_num) == IB_LINK_LAYER_ETHERNET)
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
47
struct ib_pd *pd = ib_ah->pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
51
ah->av.ib.port_pd = cpu_to_be32(to_mpd(pd)->pdn | (ah_attr->port_num << 24));
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
76
struct ib_pd *pd = ib_ah->pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib_ah.c
94
ret = ib_get_cached_gid(pd->device, ah_attr->port_num,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1287
sg_list.lkey = ctx->pd->local_dma_lkey;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1359
ah = rdma_zalloc_drv_obj(sqp_ctx->pd->device, ib_ah);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1363
ah->device = sqp_ctx->pd->device;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1364
ah->pd = sqp_ctx->pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1400
list.lkey = sqp_ctx->pd->local_dma_lkey;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1792
tun_qp->qp = ib_create_qp(ctx->pd, &qp_init_attr.init_attr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1982
ctx->pd = ib_alloc_pd(ctx->ib_dev, 0);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1983
if (IS_ERR(ctx->pd)) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
1984
ret = PTR_ERR(ctx->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
201
new_ah = ib_create_ah(dev->send_agent[port_num - 1][0]->qp->pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2032
ib_dealloc_pd(ctx->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2033
ctx->pd = NULL;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2067
ib_dealloc_pd(ctx->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2068
ctx->pd = NULL;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2199
ib_dealloc_pd(sqp_ctx->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
2200
sqp_ctx->pd = NULL;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
543
ah = ib_create_ah(tun_ctx->pd, &attr, 0);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mad.c
613
list.lkey = tun_ctx->pd->local_dma_lkey;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1141
struct mlx4_ib_pd *pd = to_mpd(ibpd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1145
err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1149
if (udata && ib_copy_to_udata(udata, &pd->pdn, sizeof(__u32))) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1150
mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1156
static void mlx4_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1158
mlx4_pd_free(to_mdev(pd->device)->dev, to_mpd(pd)->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1179
xrcd->pd = ib_alloc_pd(ibdev, 0);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1180
if (IS_ERR(xrcd->pd)) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1181
err = PTR_ERR(xrcd->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1195
ib_dealloc_pd(xrcd->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
1206
ib_dealloc_pd(to_mxrcd(xrcd)->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
137
struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
141
struct mlx4_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
153
mr->umem = ib_umem_get(pd->uobject->context, start, length,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
163
err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, virt_addr, length,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
181
(void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
194
int mr_access_flags, struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
214
to_mpd(pd)->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
343
struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
346
struct mlx4_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
354
err = mlx4_mw_alloc(dev->dev, to_mpd(pd)->pdn,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
386
struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
389
struct mlx4_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
401
err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, 0, 0, 0,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
406
err = mlx4_alloc_priv_pages(pd->device, mr, max_num_sg);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
421
mr->ibmr.device = pd->device;
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
430
struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int acc,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
433
struct mlx4_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
441
err = mlx4_fmr_alloc(dev->dev, to_mpd(pd)->pdn, convert_access(acc),
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
447
err = mlx4_fmr_enable(to_mdev(pd->device)->dev, &fmr->mfmr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
456
(void) mlx4_mr_free(to_mdev(pd->device)->dev, &fmr->mfmr.mr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
61
struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc)
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
70
err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0,
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
75
err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_mr.c
85
(void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1001
return to_mpd(to_mxrcd(qp->ibqp.xrcd)->pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1003
return to_mpd(qp->ibqp.pd);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1133
static struct ib_qp *_mlx4_ib_create_qp(struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1181
pd = to_mxrcd(init_attr->xrcd)->pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1186
if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC))
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1201
err = create_qp_common(to_mdev(pd->device), pd, init_attr,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1222
int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev, 1, 1, &sqpn, 0);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1227
sqpn = get_sqp_num(to_mdev(pd->device), init_attr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1230
err = create_qp_common(to_mdev(pd->device), pd, init_attr, udata,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1249
struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1252
struct ib_device *device = pd ? pd->device : init_attr->xrcd->device;
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1256
ibqp = _mlx4_ib_create_qp(pd, init_attr, udata);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1267
sqp->roce_v2_gsi = ib_create_qp(pd, init_attr);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1632
struct mlx4_ib_pd *pd;
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1844
pd = get_pd(qp);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
1846
context->pd = cpu_to_be32(pd->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
260
av->port_pd = cpu_to_be32((qp->port << 24) | to_mpd(qp->ibqp.pd)->pdn);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
642
static int create_qp_common(struct mlx4_ib_dev *dev, struct ib_pd *pd,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
729
err = set_rq_size(dev, &init_attr->cap, !!pd->uobject, qp_has_rq(init_attr), qp);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
733
if (pd->uobject) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
747
qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr,
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
764
err = mlx4_ib_db_map_user(to_mucontext(pd->uobject->context),
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
841
if (alloc_proxy_bufs(pd->device, qp)) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
915
free_proxy_bufs(pd->device, qp);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
917
if (pd->uobject) {
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
919
mlx4_ib_db_unmap_user(to_mucontext(pd->uobject->context), &qp->db);
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
929
if (pd->uobject)
sys/dev/mlx4/mlx4_ib/mlx4_ib_qp.c
935
if (!pd->uobject && qp_has_rq(init_attr))
sys/dev/mlx4/mlx4_ib/mlx4_ib_srq.c
185
err = mlx4_srq_alloc(dev->dev, to_mpd(ib_srq->pd)->pdn, cqn, xrcdn,
sys/dev/mlx4/qp.h
198
__be32 pd;
sys/dev/mlx5/driver.h
400
u32 pd;
sys/dev/mlx5/driver.h
428
u32 pd;
sys/dev/mlx5/mlx5_core/mlx5_crypto.c
52
MLX5_SET(create_encryption_key_in, in, encryption_key_object.pd, pdn);
sys/dev/mlx5/mlx5_core/mlx5_crypto.c
61
MLX5_SET(create_encryption_key_in, in, encryption_key_object.pd, pdn);
sys/dev/mlx5/mlx5_core/mlx5_mr.c
197
MLX5_SET(create_psv_in, in, pd, pdn);
sys/dev/mlx5/mlx5_core/mlx5_mr.c
95
mkey->pd = MLX5_GET(mkc, mkc, pd);
sys/dev/mlx5/mlx5_core/mlx5_pd.c
47
*pdn = MLX5_GET(alloc_pd_out, out, pd);
sys/dev/mlx5/mlx5_core/mlx5_pd.c
59
MLX5_SET(dealloc_pd_in, in, pd, pdn);
sys/dev/mlx5/mlx5_core/mlx5_srq.c
109
in->pd = MLX5_GET(srqc, srqc, pd);
sys/dev/mlx5/mlx5_core/mlx5_srq.c
69
MLX5_SET(wq, wq, pd, in->pd);
sys/dev/mlx5/mlx5_core/mlx5_srq.c
81
MLX5_SET(srqc, srqc, pd, in->pd);
sys/dev/mlx5/mlx5_core/mlx5_srq.c
96
in->pd = MLX5_GET(wq, wq, pd);
sys/dev/mlx5/mlx5_core/mlx5_tls.c
45
MLX5_SET(tisc, tisc, pd, pdn);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
342
MLX5_SET(sw_tls_cntx, ptag->crypto_params, progress.pd, ptag->tisn);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls_rx.c
571
MLX5_SET(sw_tls_rx_cntx, ptag->crypto_params, progress.pd, ptag->tirn);
sys/dev/mlx5/mlx5_en/mlx5_en_main.c
1527
MLX5_SET(wq, rqc_wq, pd, priv->pdn);
sys/dev/mlx5/mlx5_en/mlx5_en_main.c
2355
MLX5_SET(wq, wq, pd, priv->pdn);
sys/dev/mlx5/mlx5_en/mlx5_en_main.c
2369
MLX5_SET(wq, wq, pd, priv->pdn);
sys/dev/mlx5/mlx5_en/mlx5_en_main.c
4006
MLX5_SET(mkc, mkc, pd, pdn);
sys/dev/mlx5/mlx5_en/mlx5_en_rl.c
63
MLX5_SET(wq, wq, pd, rl->priv->pdn);
sys/dev/mlx5/mlx5_fpga/mlx5fpga_conn.c
237
MLX5_SET(mkc, mkc, pd, pdn);
sys/dev/mlx5/mlx5_fpga/mlx5fpga_conn.c
581
MLX5_SET(qpc, qpc, pd, fdev->conn_res.pdn);
sys/dev/mlx5/mlx5_fpga/mlx5fpga_conn.c
690
MLX5_SET(qpc, qpc, pd, conn->fdev->conn_res.pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
1002
struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
1059
struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
448
struct ib_pd *pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
574
struct ib_pd *pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
916
struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
939
struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc);
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
940
struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
943
struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
950
struct ib_pd *pd, struct ib_udata *udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib.h
952
struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
111
struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
114
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
118
const int num_pkeys = pd->device->attrs.max_pkeys;
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
167
gsi->cq = ib_alloc_cq(pd->device, gsi, init_attr->cap.max_send_wr, 0,
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
183
gsi->rx_qp = ib_create_qp(pd, &hw_init_attr);
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
249
struct ib_pd *pd = gsi->rx_qp->pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_gsi.c
265
return ib_create_qp(pd, &init_attr);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1741
struct mlx5_ib_pd *pd = to_mpd(ibpd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1749
err = mlx5_core_alloc_pd(to_mdev(ibdev)->mdev, &pd->pdn, uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1753
pd->uid = uid;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1755
resp.pdn = pd->pdn;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1757
mlx5_core_dealloc_pd(to_mdev(ibdev)->mdev, pd->pdn, uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1765
static void mlx5_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1767
struct mlx5_ib_dev *mdev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
1768
struct mlx5_ib_pd *mpd = to_mpd(pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2788
if (dev->umrc.pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2789
ib_dealloc_pd(dev->umrc.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2800
struct ib_pd *pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2812
pd = ib_alloc_pd(&dev->ib_dev, 0);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2813
if (IS_ERR(pd)) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2815
ret = PTR_ERR(pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2833
qp = mlx5_ib_create_qp(pd, init_attr, NULL);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2873
dev->umrc.pd = pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2896
ib_dealloc_pd(pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2897
dev->umrc.pd = NULL;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
2980
devr->s0->pd = devr->p0;
sys/dev/mlx5/mlx5_ib/mlx5_ib_main.c
3004
devr->s1->pd = devr->p0;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1057
static int rereg_umr(struct ib_pd *pd, struct mlx5_ib_mr *mr, u64 virt_addr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1061
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1089
prep_umr_wqe_common(pd, &umrwr, &sg, dma, npages, mr->mmkey.key,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1093
umrwr.pd = pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1131
struct ib_pd *pd = (flags & IB_MR_REREG_PD) ? new_pd : ib_mr->pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1153
mr->umem = mr_umem_get(pd, addr, len, access_flags, &npages,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1178
mr = reg_create(ib_mr, pd, addr, len, mr->umem, ncont,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1189
err = rereg_umr(pd, mr, addr, len, npages, page_shift,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1198
ib_mr->pd = pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1199
mr->mmkey.pd = to_mpd(pd)->pdn;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1346
struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1350
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1372
MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1377
err = mlx5_alloc_priv_descs(pd->device, mr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1387
err = mlx5_alloc_priv_descs(pd->device, mr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1405
err = mlx5_core_create_psv(dev->mdev, to_mpd(pd)->pdn,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1459
struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1462
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1500
MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
1594
u32 lkey = mr->ibmr.pd->local_dma_lkey;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
477
struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc)
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
479
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
507
MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
577
static void prep_umr_wqe_common(struct ib_pd *pd, struct mlx5_umr_wr *umrwr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
581
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
585
sg->lkey = dev->umrc.pd->local_dma_lkey;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
601
static void prep_umr_reg_wqe(struct ib_pd *pd, struct mlx5_umr_wr *umrwr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
606
prep_umr_wqe_common(pd, umrwr, sg, dma, n, key, page_shift);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
613
umrwr->pd = pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
624
static struct ib_umem *mr_umem_get(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
628
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
629
struct ib_umem *umem = ib_umem_get(pd->uobject->context, start, length,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
665
static struct mlx5_ib_mr *reg_umr(struct ib_pd *pd, struct ib_umem *umem,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
669
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
706
prep_umr_reg_wqe(pd, &umrwr, &sg, dma, npages, mr->mmkey.key,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
724
mr->mmkey.pd = to_mpd(pd)->pdn;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
828
sg.lkey = dev->umrc.pd->local_dma_lkey;
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
870
static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
875
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
913
MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
957
struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
961
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
972
umem = mr_umem_get(pd, start, length, access_flags, &npages,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
979
mr = reg_umr(pd, umem, virt_addr, length, ncont, page_shift,
sys/dev/mlx5/mlx5_ib/mlx5_ib_mr.c
993
mr = reg_create(NULL, pd, virt_addr, length, umem, ncont,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1063
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1068
MLX5_SET(create_tis_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1074
struct mlx5_ib_sq *sq, struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1076
mlx5_core_destroy_tis(dev->mdev, sq->tisn, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1081
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1099
err = mlx5_ib_umem_get(dev, pd, ubuffer->buf_addr, ubuffer->buf_size,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1112
MLX5_SET(create_sq_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1124
MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd));
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1175
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1196
MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1214
MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd));
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1239
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1251
MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1266
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1268
mlx5_core_destroy_tir(dev->mdev, rq->tirn, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1273
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1278
struct ib_uobject *uobj = pd->uobject;
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1285
err = create_raw_packet_qp_tis(dev, sq, tdn, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1289
err = create_raw_packet_qp_sq(dev, sq, in, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1299
err = create_raw_packet_qp_rq(dev, rq, in, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1304
err = create_raw_packet_qp_tir(dev, rq, tdn, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1321
destroy_raw_packet_qp_tis(dev, sq, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1334
destroy_raw_packet_qp_tir(dev, rq, qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1340
destroy_raw_packet_qp_tis(dev, sq, qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1359
to_mpd(qp->ibqp.pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1363
struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1367
struct ib_uobject *uobj = pd->uobject;
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1431
MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1586
static int create_qp_common(struct mlx5_ib_dev *dev, struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1619
err = create_rss_raw_qp_tir(dev, qp, pd, init_attr, udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1671
if (pd && pd->uobject) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1677
err = get_qp_user_index(to_mucontext(pd->uobject->context),
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1690
qp, (pd && pd->uobject) ? &ucmd : NULL);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1696
if (pd) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1697
if (pd->uobject) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1716
err = create_user_qp(dev, pd, qp, udata, init_attr, &in,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1746
MLX5_SET(qpc, qpc, pd, to_mpd(pd ? pd : devr->p0)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1844
err = create_raw_packet_qp(dev, qp, in, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1880
destroy_qp_user(dev, pd, qp, base, udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
1948
return to_mpd(qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2094
struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2103
if (pd) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2104
dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2107
if (!pd->uobject) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2110
} else if (!to_mucontext(pd->uobject->context)->cqe_version) {
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2151
err = create_qp_common(dev, pd, init_attr, udata, qp);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2175
return mlx5_ib_gsi_create_qp(pd, init_attr);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2272
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2285
MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2299
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2312
MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2393
ah->sl & 0xf, qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2543
const struct mlx5_modify_raw_qp_param *raw_qp_param, struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2557
MLX5_SET(modify_rq_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2585
struct ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2598
MLX5_SET(modify_sq_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2657
qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2666
qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2671
return modify_raw_packet_qp_sq(dev->mdev, sq, sq_state, qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2723
struct mlx5_ib_pd *pd;
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2810
pd = get_pd(qp);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
2814
context->flags_pd = cpu_to_be32(pd ? pd->pdn : to_mpd(dev->devr.p0)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3355
if (umrwr->pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3356
seg->flags_pd = cpu_to_be32(to_mpd(umrwr->pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3367
struct mlx5_ib_pd *pd)
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3373
dseg->lkey = cpu_to_be32(pd->ibpd.local_dma_lkey);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3768
struct mlx5_ib_pd *pd = to_mpd(qp->ibqp.pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
3788
set_reg_data_seg(*seg, mr, pd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4690
static int create_rq(struct mlx5_ib_rwq *rwq, struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4701
dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4708
MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4721
MLX5_SET(wq, wq, pd, to_mpd(pd)->pdn);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4753
static int prepare_user_rq(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4758
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4797
err = create_user_rq(dev, pd, rwq, &ucmd);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4808
struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4825
dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4831
err = prepare_user_rq(pd, init_attr, udata, rwq);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4834
err = create_rq(rwq, pd, init_attr);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4861
destroy_user_rq(pd, rwq, udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4873
destroy_user_rq(wq->pd, rwq, udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
4928
rwq_ind_tbl->uid = to_mpd(init_attr->ind_tbl[0]->pd)->uid;
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
5012
MLX5_SET(modify_rq_in, in, uid, to_mpd(wq->pd)->uid);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
650
struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
658
*umem = ib_umem_get(pd->uobject->context, addr, size, 0, 0);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
684
static void destroy_user_rq(struct ib_pd *pd, struct mlx5_ib_rwq *rwq,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
698
static int create_user_rq(struct mlx5_ib_dev *dev, struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
712
context = to_mucontext(pd->uobject->context);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
713
rwq->umem = ib_umem_get(pd->uobject->context, ucmd->buf_addr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
760
static int create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
788
context = to_mucontext(pd->uobject->context);
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
829
err = mlx5_ib_umem_get(dev, pd, ubuffer->buf_addr,
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
848
attr->qp_type != IB_QPT_XRC_INI) ? to_mpd(pd)->uid : 0;
sys/dev/mlx5/mlx5_ib/mlx5_ib_qp.c
897
static void destroy_qp_user(struct mlx5_ib_dev *dev, struct ib_pd *pd, struct mlx5_ib_qp *qp,
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
140
in->uid = (in->type != IB_SRQT_XRC) ? to_mpd(pd)->uid : 0;
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
218
static void destroy_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq,
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
283
err = create_srq_user(ib_srq->pd, srq, &in, udata, buf_size);
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
308
in.pd = to_mpd(ib_srq->pd)->pdn;
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
338
destroy_srq_user(ib_srq->pd, srq, udata);
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
69
static int create_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq,
sys/dev/mlx5/mlx5_ib/mlx5_ib_srq.c
73
struct mlx5_ib_dev *dev = to_mdev(pd->device);
sys/dev/mlx5/mlx5_ifc.h
10833
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
1790
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
2380
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
2647
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
2705
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
2801
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
3135
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
3417
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
6893
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
7744
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
8146
u8 pd[0x18];
sys/dev/mlx5/mlx5_ifc.h
8645
u8 pd[0x18];
sys/dev/mlx5/mlx5_lib/mlx5_aso.c
277
MLX5_SET(wq, wq, pd, pdn);
sys/dev/mlx5/srq.h
47
u32 pd;
sys/dev/mlxfw/mlxfw_mfa2.c
290
const struct mlxfw_mfa2_tlv_package_descriptor *pd;
sys/dev/mlxfw/mlxfw_mfa2.c
320
pd = mlxfw_mfa2_tlv_package_descriptor_get(mfa2_file, multi_child);
sys/dev/mlxfw/mlxfw_mfa2.c
321
if (!pd) {
sys/dev/mlxfw/mlxfw_mfa2.c
332
mfa2_file->dev_count = be16_to_cpu(pd->num_devices);
sys/dev/mlxfw/mlxfw_mfa2.c
336
mfa2_file->component_count = be16_to_cpu(pd->num_components);
sys/dev/mlxfw/mlxfw_mfa2.c
337
mfa2_file->cb = (const u8 *) fw->data + NLA_ALIGN(be32_to_cpu(pd->cb_offset));
sys/dev/mlxfw/mlxfw_mfa2.c
342
mfa2_file->cb_archive_size = be32_to_cpu(pd->cb_archive_size);
sys/dev/mrsas/mrsas.h
2851
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2881
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2900
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2903
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2909
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2914
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2919
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
2924
struct mrsas_evtarg_pd pd;
sys/dev/mrsas/mrsas.h
808
u_int16_t pd[MAX_RAIDMAP_ROW_SIZE];
sys/dev/mrsas/mrsas_fp.c
1654
u_int32_t pd, arRef, r1_alt_pd;
sys/dev/mrsas/mrsas_fp.c
1695
pd = MR_ArPdGet(arRef, physArm, map); /* Get the Pd. */
sys/dev/mrsas/mrsas_fp.c
1697
if (pd != MR_PD_INVALID) {
sys/dev/mrsas/mrsas_fp.c
1699
*pDevHandle = MR_PdDevHandleGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
1700
*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
1716
pd = MR_ArPdGet(arRef, physArm, map);
sys/dev/mrsas/mrsas_fp.c
1717
if (pd != MR_PD_INVALID) {
sys/dev/mrsas/mrsas_fp.c
1719
*pDevHandle = MR_PdDevHandleGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
1720
*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
209
MR_PdDevHandleGet(u_int32_t pd, MR_DRV_RAID_MAP_ALL * map)
sys/dev/mrsas/mrsas_fp.c
211
return map->raidMap.devHndlInfo[pd].curDevHdl;
sys/dev/mrsas/mrsas_fp.c
214
static u_int8_t MR_PdInterfaceTypeGet(u_int32_t pd, MR_DRV_RAID_MAP_ALL *map)
sys/dev/mrsas/mrsas_fp.c
216
return map->raidMap.devHndlInfo[pd].interfaceType;
sys/dev/mrsas/mrsas_fp.c
222
return le16toh(map->raidMap.arMapInfo[ar].pd[arm]);
sys/dev/mrsas/mrsas_fp.c
89
static u_int16_t MR_PdDevHandleGet(u_int32_t pd, MR_DRV_RAID_MAP_ALL * map);
sys/dev/mrsas/mrsas_fp.c
922
u_int32_t pd, arRef, r1_alt_pd;
sys/dev/mrsas/mrsas_fp.c
949
pd = MR_ArPdGet(arRef, physArm, map);
sys/dev/mrsas/mrsas_fp.c
951
if (pd != MR_PD_INVALID) {
sys/dev/mrsas/mrsas_fp.c
952
*pDevHandle = MR_PdDevHandleGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
953
*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
968
pd = MR_ArPdGet(arRef, physArm, map);
sys/dev/mrsas/mrsas_fp.c
969
if (pd != MR_PD_INVALID) {
sys/dev/mrsas/mrsas_fp.c
970
*pDevHandle = MR_PdDevHandleGet(pd, map);
sys/dev/mrsas/mrsas_fp.c
971
*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
sys/dev/mthca/mthca_allocator.c
195
union mthca_buf *buf, int *is_direct, struct mthca_pd *pd,
sys/dev/mthca/mthca_allocator.c
260
err = mthca_mr_alloc_phys(dev, pd->pd_num,
sys/dev/mthca/mthca_av.c
154
struct mthca_pd *pd,
sys/dev/mthca/mthca_av.c
170
} else if (!atomic_read(&pd->sqp_count) &&
sys/dev/mthca/mthca_av.c
197
ah->key = pd->ntmr.ibmr.lkey;
sys/dev/mthca/mthca_av.c
201
av->port_pd = cpu_to_be32(pd->pd_num | (ah_attr->port_num << 24));
sys/dev/mthca/mthca_cq.c
70
__be32 pd;
sys/dev/mthca/mthca_cq.c
838
cq_context->pd = cpu_to_be32(pdn);
sys/dev/mthca/mthca_dev.h
427
union mthca_buf *buf, int *is_direct, struct mthca_pd *pd,
sys/dev/mthca/mthca_dev.h
464
int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd);
sys/dev/mthca/mthca_dev.h
465
void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd);
sys/dev/mthca/mthca_dev.h
473
int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift,
sys/dev/mthca/mthca_dev.h
475
int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_dev.h
477
int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_dev.h
483
int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_dev.h
514
int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd,
sys/dev/mthca/mthca_dev.h
547
struct mthca_pd *pd,
sys/dev/mthca/mthca_dev.h
556
struct mthca_pd *pd,
sys/dev/mthca/mthca_dev.h
567
struct mthca_pd *pd,
sys/dev/mthca/mthca_mad.c
89
new_ah = ib_create_ah(dev->send_agent[port_num - 1][0]->qp->pd,
sys/dev/mthca/mthca_mr.c
429
int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift,
sys/dev/mthca/mthca_mr.c
468
mpt_entry->pd = cpu_to_be32(pd);
sys/dev/mthca/mthca_mr.c
512
int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_mr.c
516
return mthca_mr_alloc(dev, pd, 12, 0, ~0ULL, access, mr);
sys/dev/mthca/mthca_mr.c
519
int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_mr.c
536
err = mthca_mr_alloc(dev, pd, buffer_size_shift, iova,
sys/dev/mthca/mthca_mr.c
54
__be32 pd;
sys/dev/mthca/mthca_mr.c
567
int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd,
sys/dev/mthca/mthca_mr.c
638
mpt_entry->pd = cpu_to_be32(pd);
sys/dev/mthca/mthca_pd.c
39
int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd)
sys/dev/mthca/mthca_pd.c
43
pd->privileged = privileged;
sys/dev/mthca/mthca_pd.c
45
atomic_set(&pd->sqp_count, 0);
sys/dev/mthca/mthca_pd.c
46
pd->pd_num = mthca_alloc(&dev->pd_table.alloc);
sys/dev/mthca/mthca_pd.c
47
if (pd->pd_num == -1)
sys/dev/mthca/mthca_pd.c
51
err = mthca_mr_alloc_notrans(dev, pd->pd_num,
sys/dev/mthca/mthca_pd.c
54
&pd->ntmr);
sys/dev/mthca/mthca_pd.c
56
mthca_free(&dev->pd_table.alloc, pd->pd_num);
sys/dev/mthca/mthca_pd.c
62
void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd)
sys/dev/mthca/mthca_pd.c
64
if (pd->privileged)
sys/dev/mthca/mthca_pd.c
65
mthca_free_mr(dev, &pd->ntmr);
sys/dev/mthca/mthca_pd.c
66
mthca_free(&dev->pd_table.alloc, pd->pd_num);
sys/dev/mthca/mthca_provider.c
367
struct mthca_pd *pd = to_mpd(ibpd);
sys/dev/mthca/mthca_provider.c
370
err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd);
sys/dev/mthca/mthca_provider.c
375
if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
sys/dev/mthca/mthca_provider.c
376
mthca_pd_free(to_mdev(ibdev), pd);
sys/dev/mthca/mthca_provider.c
384
static void mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
sys/dev/mthca/mthca_provider.c
386
mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
sys/dev/mthca/mthca_provider.c
396
return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd),
sys/dev/mthca/mthca_provider.c
433
err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd),
sys/dev/mthca/mthca_provider.c
467
static struct ib_qp *mthca_create_qp(struct ib_pd *pd,
sys/dev/mthca/mthca_provider.c
495
err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
sys/dev/mthca/mthca_provider.c
503
err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
sys/dev/mthca/mthca_provider.c
507
mthca_unmap_user_db(to_mdev(pd->device),
sys/dev/mthca/mthca_provider.c
520
err = mthca_alloc_qp(to_mdev(pd->device), to_mpd(pd),
sys/dev/mthca/mthca_provider.c
527
mthca_unmap_user_db(to_mdev(pd->device),
sys/dev/mthca/mthca_provider.c
531
mthca_unmap_user_db(to_mdev(pd->device),
sys/dev/mthca/mthca_provider.c
549
err = mthca_alloc_sqp(to_mdev(pd->device), to_mpd(pd),
sys/dev/mthca/mthca_provider.c
826
static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
sys/dev/mthca/mthca_provider.c
835
err = mthca_mr_alloc_notrans(to_mdev(pd->device),
sys/dev/mthca/mthca_provider.c
836
to_mpd(pd)->pd_num,
sys/dev/mthca/mthca_provider.c
849
static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
sys/dev/mthca/mthca_provider.c
852
struct mthca_dev *dev = to_mdev(pd->device);
sys/dev/mthca/mthca_provider.c
863
if (!to_mucontext(pd->uobject->context)->reg_mr_warned) {
sys/dev/mthca/mthca_provider.c
868
++to_mucontext(pd->uobject->context)->reg_mr_warned;
sys/dev/mthca/mthca_provider.c
877
mr->umem = ib_umem_get(pd->uobject->context, start, length, acc,
sys/dev/mthca/mthca_provider.c
930
err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, shift, virt, length,
sys/dev/mthca/mthca_provider.c
960
static struct ib_fmr *mthca_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
sys/dev/mthca/mthca_provider.c
971
err = mthca_fmr_alloc(to_mdev(pd->device), to_mpd(pd)->pd_num,
sys/dev/mthca/mthca_qp.c
1056
&qp->queue, &qp->is_direct, pd, 0, &qp->mr);
sys/dev/mthca/mthca_qp.c
1146
struct mthca_pd *pd,
sys/dev/mthca/mthca_qp.c
1174
ret = mthca_alloc_wqe_buf(dev, pd, qp, udata);
sys/dev/mthca/mthca_qp.c
1180
mthca_adjust_qp_caps(dev, pd, qp);
sys/dev/mthca/mthca_qp.c
1235
struct mthca_pd *pd, struct mthca_qp *qp)
sys/dev/mthca/mthca_qp.c
1244
cap->max_inline_data > mthca_max_inline_data(pd, max_data_size))
sys/dev/mthca/mthca_qp.c
1274
struct mthca_pd *pd,
sys/dev/mthca/mthca_qp.c
1292
err = mthca_set_qp_size(dev, cap, pd, qp);
sys/dev/mthca/mthca_qp.c
1303
err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
sys/dev/mthca/mthca_qp.c
134
__be32 pd;
sys/dev/mthca/mthca_qp.c
1349
struct mthca_pd *pd,
sys/dev/mthca/mthca_qp.c
1363
err = mthca_set_qp_size(dev, cap, pd, &sqp->qp);
sys/dev/mthca/mthca_qp.c
1387
err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
sys/dev/mthca/mthca_qp.c
1392
atomic_inc(&pd->sqp_count);
sys/dev/mthca/mthca_qp.c
1474
atomic_dec(&(to_mpd(qp->ibqp.pd)->sqp_count));
sys/dev/mthca/mthca_qp.c
1544
data->lkey = cpu_to_be32(to_mpd(sqp->qp.ibqp.pd)->ntmr.ibmr.lkey);
sys/dev/mthca/mthca_qp.c
697
qp_context->pd = cpu_to_be32(to_mpd(ibqp->pd)->pd_num);
sys/dev/mthca/mthca_qp.c
943
static inline int mthca_max_inline_data(struct mthca_pd *pd, int max_data_size)
sys/dev/mthca/mthca_qp.c
946
return pd->ibpd.uobject ? max_data_size - MTHCA_INLINE_HEADER_SIZE : 0;
sys/dev/mthca/mthca_qp.c
950
struct mthca_pd *pd,
sys/dev/mthca/mthca_qp.c
957
qp->max_inline_data = mthca_max_inline_data(pd, max_data_size);
sys/dev/mthca/mthca_qp.c
975
struct mthca_pd *pd,
sys/dev/mthca/mthca_srq.c
109
context->state_pd = cpu_to_be32(pd->pd_num);
sys/dev/mthca/mthca_srq.c
119
struct mthca_pd *pd,
sys/dev/mthca/mthca_srq.c
138
context->eq_pd = cpu_to_be32(MTHCA_EQ_ASYNC << 24 | pd->pd_num);
sys/dev/mthca/mthca_srq.c
148
static int mthca_alloc_srq_buf(struct mthca_dev *dev, struct mthca_pd *pd,
sys/dev/mthca/mthca_srq.c
165
&srq->queue, &srq->is_direct, pd, 1, &srq->mr);
sys/dev/mthca/mthca_srq.c
200
int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd,
sys/dev/mthca/mthca_srq.c
256
err = mthca_alloc_srq_buf(dev, pd, srq, udata);
sys/dev/mthca/mthca_srq.c
266
mthca_arbel_init_srq_context(dev, pd, srq, mailbox->buf, udata);
sys/dev/mthca/mthca_srq.c
268
mthca_tavor_init_srq_context(dev, pd, srq, mailbox->buf, udata);
sys/dev/mthca/mthca_srq.c
98
struct mthca_pd *pd,
sys/dev/oce/oce_if.c
1032
struct oce_packet_desc *pd;
sys/dev/oce/oce_if.c
1091
pd = &wq->pckts[wq->pkt_desc_head];
sys/dev/oce/oce_if.c
1095
pd->map,
sys/dev/oce/oce_if.c
1096
m, segs, &pd->nsegs, BUS_DMA_NOWAIT);
sys/dev/oce/oce_if.c
1098
num_wqes = pd->nsegs + 1;
sys/dev/oce/oce_if.c
1105
bus_dmamap_unload(wq->tag, pd->map);
sys/dev/oce/oce_if.c
1111
bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_PREWRITE);
sys/dev/oce/oce_if.c
1112
pd->mbuf = m;
sys/dev/oce/oce_if.c
1151
for (i = 0; i < pd->nsegs; i++) {
sys/dev/oce/oce_if.c
1159
pd->wqe_idx = wq->ring->pidx;
sys/dev/oce/oce_if.c
1163
if (num_wqes > (pd->nsegs + 1)) {
sys/dev/oce/oce_if.c
1171
pd->wqe_idx = wq->ring->pidx;
sys/dev/oce/oce_if.c
1174
pd->nsegs++;
sys/dev/oce/oce_if.c
1231
struct oce_packet_desc *pd;
sys/dev/oce/oce_if.c
1235
pd = &wq->pckts[wq->pkt_desc_tail];
sys/dev/oce/oce_if.c
1238
atomic_subtract_int(&wq->ring->num_used, pd->nsegs + 1);
sys/dev/oce/oce_if.c
1239
bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
sys/dev/oce/oce_if.c
1240
bus_dmamap_unload(wq->tag, pd->map);
sys/dev/oce/oce_if.c
1242
m = pd->mbuf;
sys/dev/oce/oce_if.c
1244
pd->mbuf = NULL;
sys/dev/oce/oce_if.c
1514
struct oce_packet_desc *pd;
sys/dev/oce/oce_if.c
1523
pd = &rq->pckts[rq->ring->cidx];
sys/dev/oce/oce_if.c
1525
bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
sys/dev/oce/oce_if.c
1526
bus_dmamap_unload(rq->tag, pd->map);
sys/dev/oce/oce_if.c
1531
pd->mbuf->m_len = frag_len;
sys/dev/oce/oce_if.c
1535
pd->mbuf->m_flags &= ~M_PKTHDR;
sys/dev/oce/oce_if.c
1536
tail->m_next = pd->mbuf;
sys/dev/oce/oce_if.c
1539
tail = pd->mbuf;
sys/dev/oce/oce_if.c
1542
pd->mbuf->m_pkthdr.len = len;
sys/dev/oce/oce_if.c
1544
pd->mbuf->m_nextpkt = NULL;
sys/dev/oce/oce_if.c
1545
pd->mbuf->m_pkthdr.csum_flags = 0;
sys/dev/oce/oce_if.c
1549
pd->mbuf->m_pkthdr.csum_flags |=
sys/dev/oce/oce_if.c
1553
pd->mbuf->m_pkthdr.csum_flags |=
sys/dev/oce/oce_if.c
1557
pd->mbuf->m_pkthdr.csum_data = 0xffff;
sys/dev/oce/oce_if.c
1560
pd->mbuf->m_pkthdr.csum_flags |=
sys/dev/oce/oce_if.c
1564
*m = tail = pd->mbuf;
sys/dev/oce/oce_if.c
1566
pd->mbuf = NULL;
sys/dev/oce/oce_if.c
1756
struct oce_packet_desc *pd;
sys/dev/oce/oce_if.c
1765
pd = &rq->pckts[rq->ring->cidx];
sys/dev/oce/oce_if.c
1766
bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
sys/dev/oce/oce_if.c
1767
bus_dmamap_unload(rq->tag, pd->map);
sys/dev/oce/oce_if.c
1768
if (pd->mbuf != NULL) {
sys/dev/oce/oce_if.c
1769
m_freem(pd->mbuf);
sys/dev/oce/oce_if.c
1770
pd->mbuf = NULL;
sys/dev/oce/oce_if.c
1866
struct oce_packet_desc *pd;
sys/dev/oce/oce_if.c
1876
pd = &rq->pckts[rq->ring->pidx];
sys/dev/oce/oce_if.c
1877
pd->mbuf = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, oce_rq_buf_size);
sys/dev/oce/oce_if.c
1878
if (pd->mbuf == NULL) {
sys/dev/oce/oce_if.c
1882
pd->mbuf->m_nextpkt = NULL;
sys/dev/oce/oce_if.c
1884
pd->mbuf->m_len = pd->mbuf->m_pkthdr.len = rq->cfg.frag_size;
sys/dev/oce/oce_if.c
1887
pd->map,
sys/dev/oce/oce_if.c
1888
pd->mbuf,
sys/dev/oce/oce_if.c
1891
m_free(pd->mbuf);
sys/dev/oce/oce_if.c
1901
bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_PREREAD);
sys/dev/oce/oce_queue.c
1091
struct oce_packet_desc *pd;
sys/dev/oce/oce_queue.c
1094
pd = &rq->pckts[rq->ring->cidx];
sys/dev/oce/oce_queue.c
1095
bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
sys/dev/oce/oce_queue.c
1096
bus_dmamap_unload(rq->tag, pd->map);
sys/dev/oce/oce_queue.c
1097
if (pd->mbuf != NULL) {
sys/dev/oce/oce_queue.c
1098
m_freem(pd->mbuf);
sys/dev/oce/oce_queue.c
1099
pd->mbuf = NULL;
sys/dev/ocs_fc/ocs_scsi.c
2142
uint8_t *pd = fcprsp->data;
sys/dev/ocs_fc/ocs_scsi.c
2160
rsp.response_data = pd;
sys/dev/ocs_fc/ocs_scsi.c
2162
pd += rsp.response_data_length;
sys/dev/ocs_fc/ocs_scsi.c
2166
rsp.sense_data = pd;
sys/dev/ocs_fc/ocs_scsi.c
2168
pd += sns_len;
sys/dev/pbio/pbio.c
111
struct portdata pd[PBIO_NPORTS];/* Per port data */
sys/dev/pbio/pbio.c
236
(void)make_dev_s(&args, &sc->pd[i].port, "pbio%d%s", unit,
sys/dev/pbio/pbio.c
255
scp->pd[port].diff = *(int *)data;
sys/dev/pbio/pbio.c
258
scp->pd[port].ipace = *(int *)data;
sys/dev/pbio/pbio.c
261
scp->pd[port].opace = *(int *)data;
sys/dev/pbio/pbio.c
264
*(int *)data = scp->pd[port].diff;
sys/dev/pbio/pbio.c
267
*(int *)data = scp->pd[port].ipace;
sys/dev/pbio/pbio.c
270
*(int *)data = scp->pd[port].opace;
sys/dev/pbio/pbio.c
340
if (scp->pd[port].diff) {
sys/dev/pbio/pbio.c
341
if (*val != scp->pd[port].oldval) {
sys/dev/pbio/pbio.c
342
scp->pd[port].oldval = *val;
sys/dev/pbio/pbio.c
345
err = pause_sig("pbiopl", max(1, scp->pd[port].ipace));
sys/dev/pbio/pbio.c
367
if ((err = uiomove(scp->pd[port].buff, toread, uio)) != 0)
sys/dev/pbio/pbio.c
372
scp->pd[port].buff[i] = val;
sys/dev/pbio/pbio.c
373
if (!scp->pd[port].diff && scp->pd[port].ipace)
sys/dev/pbio/pbio.c
374
pause_sig("pbioip", scp->pd[port].ipace);
sys/dev/pbio/pbio.c
395
if ((ret = uiomove(scp->pd[port].buff, towrite, uio)) != 0)
sys/dev/pbio/pbio.c
398
val = scp->pd[port].buff[i];
sys/dev/pbio/pbio.c
419
if (scp->pd[port].opace)
sys/dev/pbio/pbio.c
420
pause_sig("pbioop", scp->pd[port].opace);
sys/dev/qlnx/qlnxe/ecore_hsi_iwarp.h
912
__le16 pd;
sys/dev/qlnx/qlnxe/ecore_hsi_rdma.h
674
__le16 pd;
sys/dev/qlnx/qlnxe/ecore_hsi_roce.h
158
__le16 pd;
sys/dev/qlnx/qlnxe/ecore_hsi_roce.h
218
__le16 pd;
sys/dev/qlnx/qlnxe/ecore_iwarp.c
324
p_ramrod->pd = qp->pd;
sys/dev/qlnx/qlnxe/ecore_rdma.c
1185
u16 *pd)
sys/dev/qlnx/qlnxe/ecore_rdma.c
1202
*pd = (u16)returned_id;
sys/dev/qlnx/qlnxe/ecore_rdma.c
1209
u16 pd)
sys/dev/qlnx/qlnxe/ecore_rdma.c
1213
DP_VERBOSE(p_hwfn, ECORE_MSG_RDMA, "pd = %08x\n", pd);
sys/dev/qlnx/qlnxe/ecore_rdma.c
1217
ecore_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->pd_map, pd);
sys/dev/qlnx/qlnxe/ecore_rdma.c
1581
qp->pd = in_params->pd;
sys/dev/qlnx/qlnxe/ecore_rdma.c
1890
p_ramrod->pd = OSAL_CPU_TO_LE16(params->pd);
sys/dev/qlnx/qlnxe/ecore_rdma.h
167
u16 pd; /* Protection domain */
sys/dev/qlnx/qlnxe/ecore_rdma_api.h
376
u16 pd;
sys/dev/qlnx/qlnxe/ecore_rdma_api.h
472
u16 pd;
sys/dev/qlnx/qlnxe/ecore_rdma_api.h
676
u16 *pd);
sys/dev/qlnx/qlnxe/ecore_rdma_api.h
718
u16 pd);
sys/dev/qlnx/qlnxe/ecore_roce.c
487
p_ramrod->pd = OSAL_CPU_TO_LE16(qp->pd);
sys/dev/qlnx/qlnxe/ecore_roce.c
645
p_ramrod->pd = OSAL_CPU_TO_LE16(qp->pd);
sys/dev/qlnx/qlnxe/ecore_roce_api.h
356
u16 pd;
sys/dev/qlnx/qlnxe/ecore_roce_api.h
500
u16 pd;
sys/dev/qlnx/qlnxe/ecore_roce_api.h
599
u16 *pd);
sys/dev/qlnx/qlnxe/ecore_roce_api.h
640
u16 pd);
sys/dev/qlnx/qlnxr/qlnxr_def.h
426
struct qlnxr_pd *pd;
sys/dev/qlnx/qlnxr/qlnxr_def.h
640
struct qlnxr_pd *pd;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1087
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1116
mr->hw_mr.pd = pd->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1457
struct qlnxr_pd *pd;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1464
pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1468
pd->pd_id, start, len, usr_addr, acc);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
1512
mr->hw_mr.pd = pd->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2203
struct qlnxr_pd *pd,
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2215
qp->pd = pd;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2238
pd->pd_id, qp->qp_type, qp->max_inline_data,
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
225
in_params.pd_id = get_qlnxr_pd(ibsrq->pd)->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2425
struct qlnxr_pd *pd,
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2443
params->pd = pd->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2444
params->dpi = pd->uctx ? pd->uctx->dpi : dev->dpi;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2573
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2616
qlnxr_init_common_qp_in_params(dev, pd, qp, attrs, false, &in_params);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
277
if (ibsrq->pd->uobject && ibsrq->pd->uobject->context)
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2851
in_params->pd,
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
2947
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
3009
in_params.pd = pd->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
3010
in_params.dpi = pd->uctx ? pd->uctx->dpi : dev->dpi;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
3055
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
3073
attrs->event_handler, pd,
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
3086
qlnxr_set_common_qp_params(dev, qp, pd, attrs);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
5359
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
5369
ibpd, pd, pd->pd_id, max_page_list_len);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
5396
mr->hw_mr.pd = pd->pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
705
struct qlnxr_pd *pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
728
pd->pd_id = pd_id;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
731
rc = ib_copy_to_udata(udata, &pd->pd_id, sizeof(pd->pd_id));
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
738
pd->uctx = rdma_udata_to_drv_context(
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
740
pd->uctx->pd = pd;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
745
pd, pd_id, dev->pd_count);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
757
struct qlnxr_pd *pd;
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
761
pd = get_qlnxr_pd(ibpd);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
767
if (pd == NULL) {
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
770
ecore_rdma_free_pd(dev->rdma_ctx, pd->pd_id);
sys/dev/qlnx/qlnxr/qlnxr_verbs.c
773
pd, pd->pd_id, dev->pd_count);
sys/dev/qlnx/qlnxr/qlnxr_verbs.h
162
extern struct ib_mr *qlnxr_alloc_mr(struct ib_pd *pd,
sys/dev/qlnx/qlnxr/qlnxr_verbs.h
80
extern void qlnxr_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
sys/dev/safe/safe.c
1000
pd->pd_size = re->re_src_segs[i].ds_len;
sys/dev/safe/safe.c
1001
pd->pd_flags = SAFE_PD_READY;
sys/dev/safe/safe.c
1186
pd = sc->sc_dpfree;
sys/dev/safe/safe.c
1187
KASSERT((pd->pd_flags&3) == 0 ||
sys/dev/safe/safe.c
1188
(pd->pd_flags&3) == SAFE_PD_DONE,
sys/dev/safe/safe.c
1190
pd->pd_flags));
sys/dev/safe/safe.c
1193
pd->pd_addr = re->re_dst_segs[i].ds_addr;
sys/dev/safe/safe.c
1194
pd->pd_flags = SAFE_PD_READY;
sys/dev/safe/safe.c
776
struct safe_pdesc *pd;
sys/dev/safe/safe.c
991
pd = sc->sc_spfree;
sys/dev/safe/safe.c
995
KASSERT((pd->pd_flags&3) == 0 ||
sys/dev/safe/safe.c
996
(pd->pd_flags&3) == SAFE_PD_DONE,
sys/dev/safe/safe.c
998
pd->pd_flags));
sys/dev/safe/safe.c
999
pd->pd_addr = re->re_src_segs[i].ds_addr;
sys/dev/usb/controller/usb_controller.c
652
struct usb_fs_privdata *pd;
sys/dev/usb/controller/usb_controller.c
656
while ((pd = SLIST_FIRST(&bus->pd_cleanup_list)) != NULL) {
sys/dev/usb/controller/usb_controller.c
657
SLIST_REMOVE(&bus->pd_cleanup_list, pd, usb_fs_privdata, pd_next);
sys/dev/usb/controller/usb_controller.c
660
usb_destroy_dev_sync(pd);
sys/dev/usb/usb_dev.c
168
usb_loc_fill(struct usb_fs_privdata* pd, struct usb_cdev_privdata *cpd)
sys/dev/usb/usb_dev.c
170
cpd->bus_index = pd->bus_index;
sys/dev/usb/usb_dev.c
171
cpd->dev_index = pd->dev_index;
sys/dev/usb/usb_dev.c
172
cpd->ep_addr = pd->ep_addr;
sys/dev/usb/usb_dev.c
173
cpd->fifo_index = pd->fifo_index;
sys/dev/usb/usb_dev.c
871
struct usb_fs_privdata* pd = (struct usb_fs_privdata*)dev->si_drv1;
sys/dev/usb/usb_dev.c
879
if (((fflags & FREAD) && !(pd->mode & FREAD)) ||
sys/dev/usb/usb_dev.c
880
((fflags & FWRITE) && !(pd->mode & FWRITE))) {
sys/dev/usb/usb_dev.c
887
usb_loc_fill(pd, cpd);
sys/dev/usb/usb_device.c
2123
struct usb_fs_privdata* pd;
sys/dev/usb/usb_device.c
2128
pd = malloc(sizeof(struct usb_fs_privdata), M_USBDEV,
sys/dev/usb/usb_device.c
2130
pd->bus_index = device_get_unit(udev->bus->bdev);
sys/dev/usb/usb_device.c
2131
pd->dev_index = udev->device_index;
sys/dev/usb/usb_device.c
2132
pd->ep_addr = ep;
sys/dev/usb/usb_device.c
2133
pd->fifo_index = fi;
sys/dev/usb/usb_device.c
2134
pd->mode = rwmode;
sys/dev/usb/usb_device.c
2140
pd->bus_index, pd->dev_index, pd->ep_addr);
sys/dev/usb/usb_device.c
2149
args.mda_si_drv1 = pd;
sys/dev/usb/usb_device.c
2151
if (make_dev_s(&args, &pd->cdev, "%s", devname) != 0) {
sys/dev/usb/usb_device.c
2153
free(pd, M_USBDEV);
sys/dev/usb/usb_device.c
2156
return (pd);
sys/dev/usb/usb_device.c
2160
usb_destroy_dev_sync(struct usb_fs_privdata *pd)
sys/dev/usb/usb_device.c
2163
pd->bus_index, pd->dev_index);
sys/dev/usb/usb_device.c
2169
destroy_dev(pd->cdev);
sys/dev/usb/usb_device.c
2171
free(pd, M_USBDEV);
sys/dev/usb/usb_device.c
2175
usb_destroy_dev(struct usb_fs_privdata *pd)
sys/dev/usb/usb_device.c
2179
if (pd == NULL)
sys/dev/usb/usb_device.c
2183
bus = devclass_get_softc(usb_devclass_ptr, pd->bus_index);
sys/dev/usb/usb_device.c
2187
usb_destroy_dev_sync(pd);
sys/dev/usb/usb_device.c
2192
delist_dev(pd->cdev);
sys/dev/usb/usb_device.c
2195
SLIST_INSERT_HEAD(&bus->pd_cleanup_list, pd, pd_next);
sys/dev/usb/usb_device.c
2208
struct usb_fs_privdata* pd;
sys/dev/usb/usb_device.c
2255
pd = usb_make_dev(udev, NULL, ep, 0,
sys/dev/usb/usb_device.c
2258
if (pd != NULL)
sys/dev/usb/usb_device.c
2259
SLIST_INSERT_HEAD(&udev->pd_list, pd, pd_next);
sys/dev/usb/usb_device.c
2266
struct usb_fs_privdata* pd;
sys/dev/usb/usb_device.c
2270
while ((pd = SLIST_FIRST(&udev->pd_list)) != NULL) {
sys/dev/usb/usb_device.c
2271
KASSERT(pd->cdev->si_drv1 == pd, ("privdata corrupt"));
sys/dev/usb/usb_device.c
2273
SLIST_REMOVE(&udev->pd_list, pd, usb_fs_privdata, pd_next);
sys/dev/usb/usb_device.c
2275
usb_destroy_dev(pd);
sys/dev/vnic/thunder_mdio.c
367
struct phy_desc *pd = NULL;
sys/dev/vnic/thunder_mdio.c
370
TAILQ_FOREACH(pd, &sc->phy_desc_head, phy_desc_list) {
sys/dev/vnic/thunder_mdio.c
371
if (pd->lmacid == lmacid)
sys/dev/vnic/thunder_mdio.c
375
return (pd);
sys/dev/vnic/thunder_mdio.c
383
struct phy_desc *pd;
sys/dev/vnic/thunder_mdio.c
388
pd = get_phy_desc(sc, lmacid);
sys/dev/vnic/thunder_mdio.c
389
if (pd == NULL) {
sys/dev/vnic/thunder_mdio.c
396
mii_sc = device_get_softc(pd->miibus);
sys/dev/vnic/thunder_mdio.c
443
struct phy_desc *pd;
sys/dev/vnic/thunder_mdio.c
449
pd = get_phy_desc(sc, lmacid);
sys/dev/vnic/thunder_mdio.c
451
if (pd == NULL) {
sys/dev/vnic/thunder_mdio.c
452
pd = malloc(sizeof(*pd), M_THUNDER_MDIO, (M_NOWAIT | M_ZERO));
sys/dev/vnic/thunder_mdio.c
453
if (pd == NULL)
sys/dev/vnic/thunder_mdio.c
455
pd->ifp = if_alloc(IFT_ETHER);
sys/dev/vnic/thunder_mdio.c
456
pd->lmacid = lmacid;
sys/dev/vnic/thunder_mdio.c
459
err = mii_attach(dev, &pd->miibus, pd->ifp,
sys/dev/vnic/thunder_mdio.c
465
if_free(pd->ifp);
sys/dev/vnic/thunder_mdio.c
466
free(pd, M_THUNDER_MDIO);
sys/dev/vnic/thunder_mdio.c
471
TAILQ_INSERT_TAIL(&sc->phy_desc_head, pd, phy_desc_list);
sys/dev/vnic/thunder_mdio.c
481
struct phy_desc *pd;
sys/dev/vnic/thunder_mdio.c
486
pd = get_phy_desc(sc, lmacid);
sys/dev/vnic/thunder_mdio.c
487
if (pd == NULL) {
sys/dev/vnic/thunder_mdio.c
493
TAILQ_REMOVE(&sc->phy_desc_head, pd, phy_desc_list);
sys/dev/vnic/thunder_mdio.c
498
if_free(pd->ifp);
sys/dev/vnic/thunder_mdio.c
500
free(pd, M_THUNDER_MDIO);
sys/fs/pseudofs/pseudofs_vnops.c
372
struct pfs_node *pd = pvd->pvd_pn;
sys/fs/pseudofs/pseudofs_vnops.c
384
pfs_lock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
386
if (vp->v_type == VDIR && pd->pn_type == pfstype_root) {
sys/fs/pseudofs/pseudofs_vnops.c
389
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
391
} else if (vp->v_type == VDIR && pd->pn_type == pfstype_procdir) {
sys/fs/pseudofs/pseudofs_vnops.c
400
len = strlen(pd->pn_name);
sys/fs/pseudofs/pseudofs_vnops.c
406
bcopy(pd->pn_name, buf + i, len);
sys/fs/pseudofs/pseudofs_vnops.c
409
pn = pd->pn_parent;
sys/fs/pseudofs/pseudofs_vnops.c
410
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
437
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
451
struct pfs_node *pd = pvd->pvd_pn;
sys/fs/pseudofs/pseudofs_vnops.c
459
pfs_assert_not_owned(pd);
sys/fs/pseudofs/pseudofs_vnops.c
463
KASSERT_PN_IS_DIR(pd);
sys/fs/pseudofs/pseudofs_vnops.c
479
if (!pfs_visible(curthread, pd, pvd->pvd_pid, NULL))
sys/fs/pseudofs/pseudofs_vnops.c
486
pn = pd;
sys/fs/pseudofs/pseudofs_vnops.c
496
if (pd->pn_type == pfstype_root)
sys/fs/pseudofs/pseudofs_vnops.c
513
KASSERT(pd->pn_parent != NULL,
sys/fs/pseudofs/pseudofs_vnops.c
524
if (pd->pn_type == pfstype_procdir)
sys/fs/pseudofs/pseudofs_vnops.c
526
pfs_lock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
527
pn = pd->pn_parent;
sys/fs/pseudofs/pseudofs_vnops.c
528
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
532
pfs_lock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
535
for (pn = pd->pn_nodes; pn != NULL; pn = pn->pn_next)
sys/fs/pseudofs/pseudofs_vnops.c
540
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
551
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
556
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
561
pfs_assert_not_owned(pd);
sys/fs/pseudofs/pseudofs_vnops.c
786
pfs_iterate(struct thread *td, struct proc *proc, struct pfs_node *pd,
sys/fs/pseudofs/pseudofs_vnops.c
792
pfs_assert_owned(pd);
sys/fs/pseudofs/pseudofs_vnops.c
796
*pn = pd->pn_nodes;
sys/fs/pseudofs/pseudofs_vnops.c
846
struct pfs_node *pd = pvd->pvd_pn;
sys/fs/pseudofs/pseudofs_vnops.c
858
KASSERT(pd->pn_info == vn->v_mount->mnt_data,
sys/fs/pseudofs/pseudofs_vnops.c
860
PFS_TRACE(("%s pid %lu", pd->pn_name, (unsigned long)pid));
sys/fs/pseudofs/pseudofs_vnops.c
861
pfs_assert_not_owned(pd);
sys/fs/pseudofs/pseudofs_vnops.c
868
KASSERT_PN_IS_DIR(pd);
sys/fs/pseudofs/pseudofs_vnops.c
889
pfs_lock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
898
if (!pfs_visible_proc(curthread, pd, proc)) {
sys/fs/pseudofs/pseudofs_vnops.c
901
pfs_unlock(pd);
sys/fs/pseudofs/pseudofs_vnops.c
909
if (pfs_iterate(curthread, proc, pd, &pn, &p) == -1) {
sys/fs/pseudofs/pseudofs_vnops.c
961
pfs_unlock(pd);
sys/fs/udf/ecma167-udf.h
375
struct part_desc pd;
sys/fs/udf/udf_vfsops.c
302
struct part_desc *pd;
sys/fs/udf/udf_vfsops.c
417
pd = (struct part_desc *)bp->b_data;
sys/fs/udf/udf_vfsops.c
418
if (!udf_checktag(&pd->tag, TAGID_PARTITION)) {
sys/fs/udf/udf_vfsops.c
420
part_num = le16toh(pd->part_num);
sys/fs/udf/udf_vfsops.c
421
udfmp->part_len = le32toh(pd->part_len);
sys/fs/udf/udf_vfsops.c
422
udfmp->part_start = le32toh(pd->start_loc);
sys/geom/raid/md_ddf.c
1406
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1410
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
1411
meta = &pd->pd_meta;
sys/geom/raid/md_ddf.c
1451
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1458
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
1461
for (i = 0; i < pd->pd_subdisks; ) {
sys/geom/raid/md_ddf.c
1463
pd->pd_meta[i]->volume_id);
sys/geom/raid/md_ddf.c
1468
free(pd->pd_meta[i], M_MD_DDF);
sys/geom/raid/md_ddf.c
1469
for (j = i; j < pd->pd_subdisks - 1; j++)
sys/geom/raid/md_ddf.c
1470
pd->pd_meta[j] = pd->pd_meta[j + 1];
sys/geom/raid/md_ddf.c
1471
pd->pd_meta[DDF_MAX_SUBDISKS - 1] = NULL;
sys/geom/raid/md_ddf.c
1472
pd->pd_subdisks--;
sys/geom/raid/md_ddf.c
1473
pd->pd_updated = 1;
sys/geom/raid/md_ddf.c
1477
if (pd->pd_subdisks == 0) {
sys/geom/raid/md_ddf.c
1613
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1628
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
1629
pdmeta = &pd->pd_meta;
sys/geom/raid/md_ddf.c
1630
reference = GET32(&pd->pd_meta, pdd->PD_Reference);
sys/geom/raid/md_ddf.c
1663
if (ddf_meta_count_vdc(&pd->pd_meta, NULL) >=
sys/geom/raid/md_ddf.c
1664
GET16(&pd->pd_meta, hdr->Max_Partitions)) {
sys/geom/raid/md_ddf.c
1669
ddf_meta_unused_range(&pd->pd_meta, &eoff, &esize);
sys/geom/raid/md_ddf.c
1675
eoff *= pd->pd_meta.sectorsize;
sys/geom/raid/md_ddf.c
1676
esize *= pd->pd_meta.sectorsize;
sys/geom/raid/md_ddf.c
1712
sa = ddf_meta_find_sa(&pd->pd_meta, 0);
sys/geom/raid/md_ddf.c
1714
if ((GET8D(&pd->pd_meta, sa->Spare_Type) &
sys/geom/raid/md_ddf.c
1716
SET32D(&pd->pd_meta, sa->Signature, 0xffffffff);
sys/geom/raid/md_ddf.c
1718
SET8D(&pd->pd_meta, sa->Spare_Type,
sys/geom/raid/md_ddf.c
1719
GET8D(&pd->pd_meta, sa->Spare_Type) |
sys/geom/raid/md_ddf.c
1796
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1836
pd = disk->d_md_data;
sys/geom/raid/md_ddf.c
1837
if (ddf_meta_count_vdc(&pd->pd_meta, NULL) <
sys/geom/raid/md_ddf.c
1838
GET16(&pd->pd_meta, hdr->Max_Partitions)) {
sys/geom/raid/md_ddf.c
1860
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1915
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
1916
if (ddf_meta_find_vdc(&pd->pd_meta, vmeta->vdc->VD_GUID) != NULL)
sys/geom/raid/md_ddf.c
1954
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
1969
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
1970
pdmeta = &pd->pd_meta;
sys/geom/raid/md_ddf.c
2117
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
2181
pd = malloc(sizeof(*pd), M_MD_DDF, M_WAITOK | M_ZERO);
sys/geom/raid/md_ddf.c
2182
pd->pd_meta = meta;
sys/geom/raid/md_ddf.c
2184
disk->d_md_data = (void *)pd;
sys/geom/raid/md_ddf.c
2251
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
2332
pd = disk->d_md_data;
sys/geom/raid/md_ddf.c
2333
if (ddf_meta_count_vdc(&pd->pd_meta, NULL) >=
sys/geom/raid/md_ddf.c
2334
GET16(&pd->pd_meta, hdr->Max_Partitions)) {
sys/geom/raid/md_ddf.c
2343
ddf_meta_unused_range(&pd->pd_meta,
sys/geom/raid/md_ddf.c
2361
pd = malloc(sizeof(*pd), M_MD_DDF, M_WAITOK | M_ZERO);
sys/geom/raid/md_ddf.c
2363
disk->d_md_data = (void *)pd;
sys/geom/raid/md_ddf.c
2369
ddf_meta_copy(&mdi->mdio_meta, &pd->pd_meta);
sys/geom/raid/md_ddf.c
2371
ddf_meta_update(&mdi->mdio_meta, &pd->pd_meta);
sys/geom/raid/md_ddf.c
2377
size = MIN(size, GET64(&pd->pd_meta,
sys/geom/raid/md_ddf.c
2684
pd = malloc(sizeof(*pd), M_MD_DDF, M_WAITOK | M_ZERO);
sys/geom/raid/md_ddf.c
2688
disk->d_md_data = (void *)pd;
sys/geom/raid/md_ddf.c
2696
sa = ddf_meta_find_sa(&pd->pd_meta, 1);
sys/geom/raid/md_ddf.c
2698
SET32D(&pd->pd_meta, sa->Signature,
sys/geom/raid/md_ddf.c
2700
SET8D(&pd->pd_meta, sa->Spare_Type, 0);
sys/geom/raid/md_ddf.c
2701
SET16D(&pd->pd_meta, sa->Populated_SAEs, 0);
sys/geom/raid/md_ddf.c
2702
SET16D(&pd->pd_meta, sa->MAX_SAE_Supported,
sys/geom/raid/md_ddf.c
2703
(GET16(&pd->pd_meta, hdr->Configuration_Record_Length) *
sys/geom/raid/md_ddf.c
2704
pd->pd_meta.sectorsize -
sys/geom/raid/md_ddf.c
2709
ddf_meta_copy(&mdi->mdio_meta, &pd->pd_meta);
sys/geom/raid/md_ddf.c
2711
ddf_meta_update(&mdi->mdio_meta, &pd->pd_meta);
sys/geom/raid/md_ddf.c
2728
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
2832
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
2847
GET32(&pd->pd_meta, pdd->PD_Reference));
sys/geom/raid/md_ddf.c
2884
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
2886
GET32(&pd->pd_meta, pdd->PD_Reference));
sys/geom/raid/md_ddf.c
2896
sa = ddf_meta_find_sa(&pd->pd_meta, 0);
sys/geom/raid/md_ddf.c
2898
(GET8D(&pd->pd_meta, sa->Spare_Type) &
sys/geom/raid/md_ddf.c
2931
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
2936
memcpy(pd->pd_meta.pdr, gmeta->pdr,
sys/geom/raid/md_ddf.c
2937
GET32(&pd->pd_meta, hdr->pdr_length) *
sys/geom/raid/md_ddf.c
2938
pd->pd_meta.sectorsize);
sys/geom/raid/md_ddf.c
2940
SET16(&pd->pd_meta, vdr->Populated_VDEs, 0);
sys/geom/raid/md_ddf.c
2945
i = ddf_meta_find_vd(&pd->pd_meta,
sys/geom/raid/md_ddf.c
2948
i = ddf_meta_find_vd(&pd->pd_meta, NULL);
sys/geom/raid/md_ddf.c
2950
memcpy(&pd->pd_meta.vdr->entry[i],
sys/geom/raid/md_ddf.c
2957
j = GETCRNUM(&pd->pd_meta);
sys/geom/raid/md_ddf.c
2959
vdc = GETVDCPTR(&pd->pd_meta, i);
sys/geom/raid/md_ddf.c
2960
if (GET32D(&pd->pd_meta, vdc->Signature) !=
sys/geom/raid/md_ddf.c
2963
SET32D(&pd->pd_meta, vdc->Signature, 0xffffffff);
sys/geom/raid/md_ddf.c
2972
vdc = ddf_meta_find_vdc(&pd->pd_meta,
sys/geom/raid/md_ddf.c
2975
vdc = ddf_meta_find_vdc(&pd->pd_meta, NULL);
sys/geom/raid/md_ddf.c
2980
GET16(&pd->pd_meta,
sys/geom/raid/md_ddf.c
2982
pd->pd_meta.sectorsize);
sys/geom/raid/md_ddf.c
2987
g_raid_md_ddf_print(&pd->pd_meta);
sys/geom/raid/md_ddf.c
2988
ddf_meta_write(disk->d_consumer, &pd->pd_meta);
sys/geom/raid/md_ddf.c
2998
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
3003
pd = (struct g_raid_md_ddf_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_ddf.c
3015
i = ddf_meta_find_pd(&pd->pd_meta, NULL, GET32(&pd->pd_meta, pdd->PD_Reference));
sys/geom/raid/md_ddf.c
3016
SET16(&pd->pd_meta, pdr->entry[i].PD_State, DDF_PDE_FAILED | DDF_PDE_PFA);
sys/geom/raid/md_ddf.c
3018
ddf_meta_write(tdisk->d_consumer, &pd->pd_meta);
sys/geom/raid/md_ddf.c
3040
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
3042
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
3043
ddf_meta_free(&pd->pd_meta);
sys/geom/raid/md_ddf.c
3044
free(pd, M_MD_DDF);
sys/geom/raid/md_ddf.c
568
struct g_raid_md_ddf_perdisk *pd;
sys/geom/raid/md_ddf.c
581
pd = (struct g_raid_md_ddf_perdisk *)disk->d_md_data;
sys/geom/raid/md_ddf.c
582
meta = &pd->pd_meta;
sys/geom/raid/md_intel.c
1010
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_intel.c
1019
memcpy(&meta->disk[disk_pos], &pd->pd_disk_meta,
sys/geom/raid/md_intel.c
1025
!(pd->pd_disk_meta.flags & INTEL_F_SPARE))
sys/geom/raid/md_intel.c
1052
!(pd->pd_disk_meta.flags & INTEL_F_SPARE)) {
sys/geom/raid/md_intel.c
1252
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
1313
pd = malloc(sizeof(*pd), M_MD_INTEL, M_WAITOK | M_ZERO);
sys/geom/raid/md_intel.c
1314
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_intel.c
1315
pd->pd_disk_meta = meta->disk[disk_pos];
sys/geom/raid/md_intel.c
1317
disk->d_md_data = (void *)pd;
sys/geom/raid/md_intel.c
1370
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
1375
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
1376
pdmeta = pd->pd_meta;
sys/geom/raid/md_intel.c
1478
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
1596
pd = malloc(sizeof(*pd), M_MD_INTEL, M_WAITOK | M_ZERO);
sys/geom/raid/md_intel.c
1597
pd->pd_meta = meta;
sys/geom/raid/md_intel.c
1598
pd->pd_disk_pos = -1;
sys/geom/raid/md_intel.c
1600
memcpy(&pd->pd_disk_meta.serial[0], serial, INTEL_SERIAL_LEN);
sys/geom/raid/md_intel.c
1601
intel_set_disk_sectors(&pd->pd_disk_meta,
sys/geom/raid/md_intel.c
1603
pd->pd_disk_meta.id = 0;
sys/geom/raid/md_intel.c
1604
pd->pd_disk_meta.flags = INTEL_F_SPARE;
sys/geom/raid/md_intel.c
1606
pd->pd_disk_meta = meta->disk[disk_pos];
sys/geom/raid/md_intel.c
1609
disk->d_md_data = (void *)pd;
sys/geom/raid/md_intel.c
1635
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
1648
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
1652
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_intel.c
1694
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
1768
pd = malloc(sizeof(*pd), M_MD_INTEL, M_WAITOK | M_ZERO);
sys/geom/raid/md_intel.c
1769
pd->pd_disk_pos = i;
sys/geom/raid/md_intel.c
1771
disk->d_md_data = (void *)pd;
sys/geom/raid/md_intel.c
1774
strcpy(&pd->pd_disk_meta.serial[0], "NONE");
sys/geom/raid/md_intel.c
1775
pd->pd_disk_meta.id = 0xffffffff;
sys/geom/raid/md_intel.c
1776
pd->pd_disk_meta.flags = INTEL_F_ASSIGNED;
sys/geom/raid/md_intel.c
1783
&pd->pd_disk_meta.serial[0], INTEL_SERIAL_LEN);
sys/geom/raid/md_intel.c
1794
intel_set_disk_sectors(&pd->pd_disk_meta,
sys/geom/raid/md_intel.c
1800
pd->pd_disk_meta.id = 0;
sys/geom/raid/md_intel.c
1801
pd->pd_disk_meta.flags = INTEL_F_ASSIGNED | INTEL_F_ONLINE;
sys/geom/raid/md_intel.c
1886
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
1887
sd = &vol->v_subdisks[pd->pd_disk_pos];
sys/geom/raid/md_intel.c
1972
pd = (struct g_raid_md_intel_perdisk *)
sys/geom/raid/md_intel.c
1975
intel_get_disk_sectors(&pd->pd_disk_meta);
sys/geom/raid/md_intel.c
2227
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
2233
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_intel.c
2302
pd = malloc(sizeof(*pd), M_MD_INTEL, M_WAITOK | M_ZERO);
sys/geom/raid/md_intel.c
2303
pd->pd_disk_pos = -1;
sys/geom/raid/md_intel.c
2307
disk->d_md_data = (void *)pd;
sys/geom/raid/md_intel.c
2312
memcpy(&pd->pd_disk_meta.serial[0], &serial[0],
sys/geom/raid/md_intel.c
2314
intel_set_disk_sectors(&pd->pd_disk_meta,
sys/geom/raid/md_intel.c
2316
pd->pd_disk_meta.id = 0;
sys/geom/raid/md_intel.c
2317
pd->pd_disk_meta.flags = INTEL_F_SPARE;
sys/geom/raid/md_intel.c
2322
intel_meta_write_spare(cp, &pd->pd_disk_meta);
sys/geom/raid/md_intel.c
2351
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
2371
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
2372
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_intel.c
2376
pd->pd_disk_meta.flags =
sys/geom/raid/md_intel.c
2379
pd->pd_disk_meta.flags = INTEL_F_FAILED |
sys/geom/raid/md_intel.c
2382
pd->pd_disk_meta.flags = INTEL_F_FAILED |
sys/geom/raid/md_intel.c
2385
if (!(pd->pd_disk_meta.flags & INTEL_F_DISABLED))
sys/geom/raid/md_intel.c
2386
pd->pd_disk_meta.flags = INTEL_F_ASSIGNED;
sys/geom/raid/md_intel.c
2387
if (pd->pd_disk_meta.id != 0xffffffff) {
sys/geom/raid/md_intel.c
2388
pd->pd_disk_meta.id = 0xffffffff;
sys/geom/raid/md_intel.c
2389
len = strlen(pd->pd_disk_meta.serial);
sys/geom/raid/md_intel.c
2391
strcpy(pd->pd_disk_meta.serial + len, ":0");
sys/geom/raid/md_intel.c
2407
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
2408
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_intel.c
2410
meta->disk[pd->pd_disk_pos] = pd->pd_disk_meta;
sys/geom/raid/md_intel.c
2411
if (pd->pd_disk_meta.sectors_hi != 0)
sys/geom/raid/md_intel.c
2562
pd = (struct g_raid_md_intel_perdisk *)
sys/geom/raid/md_intel.c
2564
mmap0->disk_idx[sdi] = pd->pd_disk_pos;
sys/geom/raid/md_intel.c
2566
mmap1->disk_idx[sdi] = pd->pd_disk_pos;
sys/geom/raid/md_intel.c
2602
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
2605
if (pd->pd_meta != NULL) {
sys/geom/raid/md_intel.c
2606
free(pd->pd_meta, M_MD_INTEL);
sys/geom/raid/md_intel.c
2607
pd->pd_meta = NULL;
sys/geom/raid/md_intel.c
2609
pd->pd_meta = intel_meta_copy(meta);
sys/geom/raid/md_intel.c
2621
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
2626
pd = (struct g_raid_md_intel_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_intel.c
2629
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_intel.c
2636
mdi->mdio_meta->disk[pd->pd_disk_pos].flags = INTEL_F_FAILED;
sys/geom/raid/md_intel.c
2637
pd->pd_disk_meta.flags = INTEL_F_FAILED;
sys/geom/raid/md_intel.c
2667
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
2669
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
2670
if (pd->pd_meta != NULL) {
sys/geom/raid/md_intel.c
2671
free(pd->pd_meta, M_MD_INTEL);
sys/geom/raid/md_intel.c
2672
pd->pd_meta = NULL;
sys/geom/raid/md_intel.c
2674
free(pd, M_MD_INTEL);
sys/geom/raid/md_intel.c
842
struct g_raid_md_intel_perdisk *pd;
sys/geom/raid/md_intel.c
845
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
846
if (pd->pd_disk_pos == id)
sys/geom/raid/md_intel.c
914
struct g_raid_md_intel_perdisk *pd, *oldpd;
sys/geom/raid/md_intel.c
924
pd = (struct g_raid_md_intel_perdisk *)disk->d_md_data;
sys/geom/raid/md_intel.c
928
disk_pos = intel_meta_find_disk(meta, pd->pd_disk_meta.serial);
sys/geom/raid/md_intel.c
932
if ((pd->pd_disk_meta.flags & INTEL_F_FAILED) &&
sys/geom/raid/md_intel.c
933
!(pd->pd_disk_meta.flags & INTEL_F_DISABLED)) {
sys/geom/raid/md_intel.c
951
intel_get_disk_sectors(&pd->pd_disk_meta);
sys/geom/raid/md_intel.c
974
if (pd->pd_disk_meta.flags & INTEL_F_SPARE) {
sys/geom/raid/md_jmicron.c
1027
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
1100
pd = malloc(sizeof(*pd), M_MD_JMICRON, M_WAITOK | M_ZERO);
sys/geom/raid/md_jmicron.c
1101
pd->pd_disk_pos = i;
sys/geom/raid/md_jmicron.c
1102
pd->pd_disk_id = arc4random() & JMICRON_DISK_MASK;
sys/geom/raid/md_jmicron.c
1104
disk->d_md_data = (void *)pd;
sys/geom/raid/md_jmicron.c
1113
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_jmicron.c
1207
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
1208
sd = &vol->v_subdisks[pd->pd_disk_pos];
sys/geom/raid/md_jmicron.c
1288
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
1294
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_jmicron.c
1350
pd = malloc(sizeof(*pd), M_MD_JMICRON, M_WAITOK | M_ZERO);
sys/geom/raid/md_jmicron.c
1351
pd->pd_disk_pos = -3;
sys/geom/raid/md_jmicron.c
1352
pd->pd_disk_id = arc4random() & JMICRON_DISK_MASK;
sys/geom/raid/md_jmicron.c
1353
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_jmicron.c
1357
disk->d_md_data = (void *)pd;
sys/geom/raid/md_jmicron.c
1393
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
1429
pd = (struct g_raid_md_jmicron_perdisk *)
sys/geom/raid/md_jmicron.c
1431
meta->disks[i] = pd->pd_disk_id;
sys/geom/raid/md_jmicron.c
1442
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
1445
meta->spare[spares] = pd->pd_disk_id;
sys/geom/raid/md_jmicron.c
1455
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
1459
if (pd->pd_meta != NULL) {
sys/geom/raid/md_jmicron.c
1460
free(pd->pd_meta, M_MD_JMICRON);
sys/geom/raid/md_jmicron.c
1461
pd->pd_meta = NULL;
sys/geom/raid/md_jmicron.c
1463
pd->pd_meta = jmicron_meta_copy(meta);
sys/geom/raid/md_jmicron.c
1464
pd->pd_meta->disk_id = pd->pd_disk_id;
sys/geom/raid/md_jmicron.c
1466
pd->pd_meta->offset =
sys/geom/raid/md_jmicron.c
1468
pd->pd_meta->disk_sectors_high =
sys/geom/raid/md_jmicron.c
1470
pd->pd_meta->disk_sectors_low =
sys/geom/raid/md_jmicron.c
1473
pd->pd_meta->flags &= ~JMICRON_F_BADSEC;
sys/geom/raid/md_jmicron.c
1475
pd->pd_meta->flags |= JMICRON_F_UNSYNC;
sys/geom/raid/md_jmicron.c
1479
g_raid_md_jmicron_print(pd->pd_meta);
sys/geom/raid/md_jmicron.c
1480
jmicron_meta_write(disk->d_consumer, pd->pd_meta);
sys/geom/raid/md_jmicron.c
1490
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
1494
pd = (struct g_raid_md_jmicron_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_jmicron.c
1497
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_jmicron.c
1528
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
1530
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
1531
if (pd->pd_meta != NULL) {
sys/geom/raid/md_jmicron.c
1532
free(pd->pd_meta, M_MD_JMICRON);
sys/geom/raid/md_jmicron.c
1533
pd->pd_meta = NULL;
sys/geom/raid/md_jmicron.c
1535
free(pd, M_MD_JMICRON);
sys/geom/raid/md_jmicron.c
358
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
361
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
362
if (pd->pd_disk_pos == id)
sys/geom/raid/md_jmicron.c
427
struct g_raid_md_jmicron_perdisk *pd, *oldpd;
sys/geom/raid/md_jmicron.c
435
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
439
if (pd->pd_meta != NULL)
sys/geom/raid/md_jmicron.c
440
disk_pos = jmicron_meta_find_disk(meta, pd->pd_disk_id);
sys/geom/raid/md_jmicron.c
459
pd->pd_disk_size) {
sys/geom/raid/md_jmicron.c
462
pd->pd_disk_size,
sys/geom/raid/md_jmicron.c
477
if (disk_pos == -3 || pd->pd_disk_pos == -3) {
sys/geom/raid/md_jmicron.c
513
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_jmicron.c
515
meta->disks[disk_pos] = pd->pd_disk_id;
sys/geom/raid/md_jmicron.c
534
(off_t)pd->pd_meta->offset * 16 * 512; //ZZZ
sys/geom/raid/md_jmicron.c
536
(((off_t)pd->pd_meta->disk_sectors_high << 16) +
sys/geom/raid/md_jmicron.c
537
pd->pd_meta->disk_sectors_low) * 512;
sys/geom/raid/md_jmicron.c
545
(pd->pd_meta->flags & JMICRON_F_BADSEC) == 0) {
sys/geom/raid/md_jmicron.c
549
} else if (pd->pd_meta->flags & JMICRON_F_UNSYNC) {
sys/geom/raid/md_jmicron.c
646
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
700
pd = malloc(sizeof(*pd), M_MD_JMICRON, M_WAITOK | M_ZERO);
sys/geom/raid/md_jmicron.c
701
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_jmicron.c
702
pd->pd_disk_id = meta->disks[disk_pos];
sys/geom/raid/md_jmicron.c
704
disk->d_md_data = (void *)pd;
sys/geom/raid/md_jmicron.c
743
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
748
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
749
pdmeta = pd->pd_meta;
sys/geom/raid/md_jmicron.c
827
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
934
pd = malloc(sizeof(*pd), M_MD_JMICRON, M_WAITOK | M_ZERO);
sys/geom/raid/md_jmicron.c
935
pd->pd_meta = meta;
sys/geom/raid/md_jmicron.c
937
pd->pd_disk_pos = -3;
sys/geom/raid/md_jmicron.c
938
pd->pd_disk_id = arc4random() & JMICRON_DISK_MASK;
sys/geom/raid/md_jmicron.c
940
pd->pd_disk_pos = -1;
sys/geom/raid/md_jmicron.c
941
pd->pd_disk_id = meta->disk_id;
sys/geom/raid/md_jmicron.c
943
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_jmicron.c
945
disk->d_md_data = (void *)pd;
sys/geom/raid/md_jmicron.c
969
struct g_raid_md_jmicron_perdisk *pd;
sys/geom/raid/md_jmicron.c
982
pd = (struct g_raid_md_jmicron_perdisk *)disk->d_md_data;
sys/geom/raid/md_jmicron.c
986
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_nvidia.c
1004
if (pd->pd_disk_pos >= 0)
sys/geom/raid/md_nvidia.c
1031
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
1104
pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO);
sys/geom/raid/md_nvidia.c
1105
pd->pd_disk_pos = i;
sys/geom/raid/md_nvidia.c
1107
disk->d_md_data = (void *)pd;
sys/geom/raid/md_nvidia.c
1116
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_nvidia.c
1212
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
1213
sd = &vol->v_subdisks[pd->pd_disk_pos];
sys/geom/raid/md_nvidia.c
1293
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
1299
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_nvidia.c
1355
pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO);
sys/geom/raid/md_nvidia.c
1356
pd->pd_disk_pos = -3;
sys/geom/raid/md_nvidia.c
1357
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_nvidia.c
1361
disk->d_md_data = (void *)pd;
sys/geom/raid/md_nvidia.c
1397
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
1475
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
1479
if (pd->pd_meta != NULL) {
sys/geom/raid/md_nvidia.c
1480
free(pd->pd_meta, M_MD_NVIDIA);
sys/geom/raid/md_nvidia.c
1481
pd->pd_meta = NULL;
sys/geom/raid/md_nvidia.c
1483
pd->pd_meta = nvidia_meta_copy(meta);
sys/geom/raid/md_nvidia.c
1486
pd->pd_meta->disk_number =
sys/geom/raid/md_nvidia.c
1489
pd->pd_meta->disk_status = 0x100;
sys/geom/raid/md_nvidia.c
1490
pd->pd_meta->rebuild_lba =
sys/geom/raid/md_nvidia.c
1495
pd->pd_meta->disk_number = meta->total_disks + spares++;
sys/geom/raid/md_nvidia.c
1498
g_raid_md_nvidia_print(pd->pd_meta);
sys/geom/raid/md_nvidia.c
1499
nvidia_meta_write(disk->d_consumer, pd->pd_meta);
sys/geom/raid/md_nvidia.c
1509
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
1513
pd = (struct g_raid_md_nvidia_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_nvidia.c
1516
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_nvidia.c
1548
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
1550
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
1551
if (pd->pd_meta != NULL) {
sys/geom/raid/md_nvidia.c
1552
free(pd->pd_meta, M_MD_NVIDIA);
sys/geom/raid/md_nvidia.c
1553
pd->pd_meta = NULL;
sys/geom/raid/md_nvidia.c
1555
free(pd, M_MD_NVIDIA);
sys/geom/raid/md_nvidia.c
368
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
371
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
372
if (pd->pd_disk_pos == id)
sys/geom/raid/md_nvidia.c
434
struct g_raid_md_nvidia_perdisk *pd, *oldpd;
sys/geom/raid/md_nvidia.c
442
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
446
if (pd->pd_meta != NULL) {
sys/geom/raid/md_nvidia.c
447
disk_pos = pd->pd_meta->disk_number;
sys/geom/raid/md_nvidia.c
470
pd->pd_disk_size) {
sys/geom/raid/md_nvidia.c
473
pd->pd_disk_size,
sys/geom/raid/md_nvidia.c
517
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_nvidia.c
541
sd->sd_size = pd->pd_disk_size - 0x800 * 512;
sys/geom/raid/md_nvidia.c
548
(pd->pd_meta->disk_status & 0x100)) {
sys/geom/raid/md_nvidia.c
552
sd->sd_rebuild_pos = (off_t)pd->pd_meta->rebuild_lba /
sys/geom/raid/md_nvidia.c
553
meta->array_width * pd->pd_meta->sector_size;
sys/geom/raid/md_nvidia.c
558
sd->sd_rebuild_pos = (off_t)pd->pd_meta->rebuild_lba /
sys/geom/raid/md_nvidia.c
559
meta->array_width * pd->pd_meta->sector_size;
sys/geom/raid/md_nvidia.c
653
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
710
pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO);
sys/geom/raid/md_nvidia.c
711
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_nvidia.c
713
disk->d_md_data = (void *)pd;
sys/geom/raid/md_nvidia.c
752
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
757
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
758
pdmeta = pd->pd_meta;
sys/geom/raid/md_nvidia.c
832
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
931
pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO);
sys/geom/raid/md_nvidia.c
932
pd->pd_meta = meta;
sys/geom/raid/md_nvidia.c
934
pd->pd_disk_pos = -3;
sys/geom/raid/md_nvidia.c
936
pd->pd_disk_pos = -1;
sys/geom/raid/md_nvidia.c
938
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_nvidia.c
940
disk->d_md_data = (void *)pd;
sys/geom/raid/md_nvidia.c
964
struct g_raid_md_nvidia_perdisk *pd;
sys/geom/raid/md_nvidia.c
980
pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data;
sys/geom/raid/md_nvidia.c
984
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_promise.c
1003
for (i = 0; i < pd->pd_subdisks; i++) {
sys/geom/raid/md_promise.c
1004
pdmeta = pd->pd_meta[i];
sys/geom/raid/md_promise.c
1042
for (i = 0; i < pd->pd_subdisks; i++) {
sys/geom/raid/md_promise.c
1043
pdmeta = pd->pd_meta[i];
sys/geom/raid/md_promise.c
1103
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
1192
pd = malloc(sizeof(*pd), M_MD_PROMISE, M_WAITOK | M_ZERO);
sys/geom/raid/md_promise.c
1193
pd->pd_subdisks = subdisks;
sys/geom/raid/md_promise.c
1195
pd->pd_meta[i] = metaarr[i];
sys/geom/raid/md_promise.c
1197
disk->d_md_data = (void *)pd;
sys/geom/raid/md_promise.c
1264
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
1343
pd = disk->d_md_data;
sys/geom/raid/md_promise.c
1344
if (pd->pd_subdisks >= PROMISE_MAX_SUBDISKS) {
sys/geom/raid/md_promise.c
1347
diskname, pd->pd_subdisks);
sys/geom/raid/md_promise.c
1353
promise_meta_unused_range(pd->pd_meta,
sys/geom/raid/md_promise.c
1354
pd->pd_subdisks,
sys/geom/raid/md_promise.c
1372
pd = malloc(sizeof(*pd), M_MD_PROMISE, M_WAITOK | M_ZERO);
sys/geom/raid/md_promise.c
1374
disk->d_md_data = (void *)pd;
sys/geom/raid/md_promise.c
1673
pd = malloc(sizeof(*pd), M_MD_PROMISE, M_WAITOK | M_ZERO);
sys/geom/raid/md_promise.c
1677
disk->d_md_data = (void *)pd;
sys/geom/raid/md_promise.c
1700
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
1853
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
1854
for (j = 0; j < pd->pd_subdisks; j++) {
sys/geom/raid/md_promise.c
1855
if (pd->pd_meta[j]->volume_id == meta->volume_id)
sys/geom/raid/md_promise.c
1858
if (j == pd->pd_subdisks)
sys/geom/raid/md_promise.c
1859
pd->pd_subdisks++;
sys/geom/raid/md_promise.c
1860
if (pd->pd_meta[j] != NULL)
sys/geom/raid/md_promise.c
1861
free(pd->pd_meta[j], M_MD_PROMISE);
sys/geom/raid/md_promise.c
1862
pd->pd_meta[j] = promise_meta_copy(meta);
sys/geom/raid/md_promise.c
1863
pd->pd_meta[j]->disk = meta->disks[pos];
sys/geom/raid/md_promise.c
1864
pd->pd_meta[j]->disk.number = pos;
sys/geom/raid/md_promise.c
1865
pd->pd_meta[j]->disk_offset_high =
sys/geom/raid/md_promise.c
1867
pd->pd_meta[j]->disk_offset = sd->sd_offset / 512;
sys/geom/raid/md_promise.c
1868
pd->pd_meta[j]->disk_sectors_high =
sys/geom/raid/md_promise.c
1870
pd->pd_meta[j]->disk_sectors = sd->sd_size / 512;
sys/geom/raid/md_promise.c
1872
pd->pd_meta[j]->disk_rebuild_high =
sys/geom/raid/md_promise.c
1874
pd->pd_meta[j]->disk_rebuild =
sys/geom/raid/md_promise.c
1877
pd->pd_meta[j]->disk_rebuild_high = 0;
sys/geom/raid/md_promise.c
1878
pd->pd_meta[j]->disk_rebuild = 0;
sys/geom/raid/md_promise.c
1880
pd->pd_meta[j]->disk_rebuild_high = UINT32_MAX;
sys/geom/raid/md_promise.c
1881
pd->pd_meta[j]->disk_rebuild = UINT32_MAX;
sys/geom/raid/md_promise.c
1883
pd->pd_updated = 1;
sys/geom/raid/md_promise.c
1888
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
1891
if (!pd->pd_updated)
sys/geom/raid/md_promise.c
1895
for (i = 0; i < pd->pd_subdisks; i++)
sys/geom/raid/md_promise.c
1896
g_raid_md_promise_print(pd->pd_meta[i]);
sys/geom/raid/md_promise.c
1898
pd->pd_meta, pd->pd_subdisks);
sys/geom/raid/md_promise.c
1899
pd->pd_updated = 0;
sys/geom/raid/md_promise.c
1910
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
1915
pd = (struct g_raid_md_promise_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_promise.c
1925
if (pd->pd_subdisks > 0 && tdisk->d_consumer != NULL)
sys/geom/raid/md_promise.c
1928
for (i = 0; i < pd->pd_subdisks; i++) {
sys/geom/raid/md_promise.c
1929
pd->pd_meta[i]->disk.flags |=
sys/geom/raid/md_promise.c
1931
pos = pd->pd_meta[i]->disk.number;
sys/geom/raid/md_promise.c
1933
pd->pd_meta[i]->disks[pos].flags |=
sys/geom/raid/md_promise.c
1936
g_raid_md_promise_print(pd->pd_meta[i]);
sys/geom/raid/md_promise.c
1940
pd->pd_meta, pd->pd_subdisks);
sys/geom/raid/md_promise.c
1962
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
1965
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
1966
for (i = 0; i < pd->pd_subdisks; i++) {
sys/geom/raid/md_promise.c
1967
if (pd->pd_meta[i] != NULL) {
sys/geom/raid/md_promise.c
1968
free(pd->pd_meta[i], M_MD_PROMISE);
sys/geom/raid/md_promise.c
1969
pd->pd_meta[i] = NULL;
sys/geom/raid/md_promise.c
1972
free(pd, M_MD_PROMISE);
sys/geom/raid/md_promise.c
579
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
586
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
589
for (i = 0; i < pd->pd_subdisks; ) {
sys/geom/raid/md_promise.c
591
pd->pd_meta[i]->volume_id);
sys/geom/raid/md_promise.c
596
free(pd->pd_meta[i], M_MD_PROMISE);
sys/geom/raid/md_promise.c
597
for (j = i; j < pd->pd_subdisks - 1; j++)
sys/geom/raid/md_promise.c
598
pd->pd_meta[j] = pd->pd_meta[j + 1];
sys/geom/raid/md_promise.c
599
pd->pd_meta[pd->pd_subdisks - 1] = NULL;
sys/geom/raid/md_promise.c
600
pd->pd_subdisks--;
sys/geom/raid/md_promise.c
601
pd->pd_updated = 1;
sys/geom/raid/md_promise.c
605
if (pd->pd_subdisks == 0) {
sys/geom/raid/md_promise.c
669
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
676
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
683
md_disk_pos = promise_meta_find_disk(meta, pd->pd_meta[sdn]->disk.id);
sys/geom/raid/md_promise.c
695
pd->pd_meta[sdn]->disk.flags & PROMISE_F_DOWN) {
sys/geom/raid/md_promise.c
701
free(pd->pd_meta[sdn], M_MD_PROMISE);
sys/geom/raid/md_promise.c
702
for (i = sdn; i < pd->pd_subdisks - 1; i++)
sys/geom/raid/md_promise.c
703
pd->pd_meta[i] = pd->pd_meta[i + 1];
sys/geom/raid/md_promise.c
704
pd->pd_meta[pd->pd_subdisks - 1] = NULL;
sys/geom/raid/md_promise.c
705
pd->pd_subdisks--;
sys/geom/raid/md_promise.c
714
promise_meta_unused_range(pd->pd_meta, pd->pd_subdisks,
sys/geom/raid/md_promise.c
749
if (pd->pd_subdisks == 0) {
sys/geom/raid/md_promise.c
783
sd->sd_offset = (((off_t)pd->pd_meta[sdn]->disk_offset_high
sys/geom/raid/md_promise.c
784
<< 32) + pd->pd_meta[sdn]->disk_offset) * 512;
sys/geom/raid/md_promise.c
785
sd->sd_size = (((off_t)pd->pd_meta[sdn]->disk_sectors_high
sys/geom/raid/md_promise.c
786
<< 32) + pd->pd_meta[sdn]->disk_sectors) * 512;
sys/geom/raid/md_promise.c
801
if (pd->pd_meta[sdn]->generation != meta->generation)
sys/geom/raid/md_promise.c
805
(((off_t)pd->pd_meta[sdn]->disk_rebuild_high << 32) +
sys/geom/raid/md_promise.c
806
pd->pd_meta[sdn]->disk_rebuild) * 512;
sys/geom/raid/md_promise.c
812
} else if (pd->pd_meta[sdn]->generation != meta->generation ||
sys/geom/raid/md_promise.c
835
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
875
pd = disk->d_md_data;
sys/geom/raid/md_promise.c
876
if (pd->pd_subdisks < PROMISE_MAX_SUBDISKS) {
sys/geom/raid/md_promise.c
899
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
946
pd = disk->d_md_data;
sys/geom/raid/md_promise.c
947
for (i = 0; i < pd->pd_subdisks; i++) {
sys/geom/raid/md_promise.c
948
if (pd->pd_meta[i]->volume_id == meta->volume_id)
sys/geom/raid/md_promise.c
987
struct g_raid_md_promise_perdisk *pd;
sys/geom/raid/md_promise.c
995
pd = (struct g_raid_md_promise_perdisk *)disk->d_md_data;
sys/geom/raid/md_promise.c
997
if (pd->pd_subdisks == 0) {
sys/geom/raid/md_sii.c
1031
pd = malloc(sizeof(*pd), M_MD_SII, M_WAITOK | M_ZERO);
sys/geom/raid/md_sii.c
1032
pd->pd_meta = meta;
sys/geom/raid/md_sii.c
1034
pd->pd_disk_pos = -3;
sys/geom/raid/md_sii.c
1036
pd->pd_disk_pos = -1;
sys/geom/raid/md_sii.c
1038
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_sii.c
1040
disk->d_md_data = (void *)pd;
sys/geom/raid/md_sii.c
1064
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
1077
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
1081
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_sii.c
1122
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
1195
pd = malloc(sizeof(*pd), M_MD_SII, M_WAITOK | M_ZERO);
sys/geom/raid/md_sii.c
1196
pd->pd_disk_pos = i;
sys/geom/raid/md_sii.c
1198
disk->d_md_data = (void *)pd;
sys/geom/raid/md_sii.c
1207
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_sii.c
1301
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
1302
sd = &vol->v_subdisks[pd->pd_disk_pos];
sys/geom/raid/md_sii.c
1382
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
1388
if (pd->pd_disk_pos >= 0) {
sys/geom/raid/md_sii.c
1444
pd = malloc(sizeof(*pd), M_MD_SII, M_WAITOK | M_ZERO);
sys/geom/raid/md_sii.c
1445
pd->pd_disk_pos = -3;
sys/geom/raid/md_sii.c
1446
pd->pd_disk_size = pp->mediasize;
sys/geom/raid/md_sii.c
1450
disk->d_md_data = (void *)pd;
sys/geom/raid/md_sii.c
1488
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
1552
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
1555
if (pd->pd_meta != NULL) {
sys/geom/raid/md_sii.c
1556
free(pd->pd_meta, M_MD_SII);
sys/geom/raid/md_sii.c
1557
pd->pd_meta = NULL;
sys/geom/raid/md_sii.c
1559
pd->pd_meta = sii_meta_copy(meta);
sys/geom/raid/md_sii.c
1562
pd->pd_meta->disk_status = SII_S_DROPPED;
sys/geom/raid/md_sii.c
1564
pd->pd_meta->disk_status = SII_S_REBUILD;
sys/geom/raid/md_sii.c
1565
pd->pd_meta->rebuild_lba =
sys/geom/raid/md_sii.c
1568
pd->pd_meta->disk_status = SII_S_CURRENT;
sys/geom/raid/md_sii.c
1570
pd->pd_meta->disk_number = sd->sd_pos;
sys/geom/raid/md_sii.c
1571
pd->pd_meta->raid0_ident = 0xff;
sys/geom/raid/md_sii.c
1572
pd->pd_meta->raid1_ident = 0;
sys/geom/raid/md_sii.c
1574
pd->pd_meta->disk_number = sd->sd_pos / meta->raid1_disks;
sys/geom/raid/md_sii.c
1575
pd->pd_meta->raid0_ident = sd->sd_pos % meta->raid1_disks;
sys/geom/raid/md_sii.c
1576
pd->pd_meta->raid1_ident = sd->sd_pos / meta->raid1_disks;
sys/geom/raid/md_sii.c
1578
pd->pd_meta->disk_number = sd->sd_pos;
sys/geom/raid/md_sii.c
1579
pd->pd_meta->raid0_ident = 0;
sys/geom/raid/md_sii.c
1580
pd->pd_meta->raid1_ident = 0xff;
sys/geom/raid/md_sii.c
1585
g_raid_md_sii_print(pd->pd_meta);
sys/geom/raid/md_sii.c
1586
sii_meta_write(disk->d_consumer, pd->pd_meta);
sys/geom/raid/md_sii.c
1596
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
1600
pd = (struct g_raid_md_sii_perdisk *)tdisk->d_md_data;
sys/geom/raid/md_sii.c
1603
if (pd->pd_disk_pos < 0)
sys/geom/raid/md_sii.c
1611
if (pd->pd_meta) {
sys/geom/raid/md_sii.c
1612
pd->pd_meta->disk_status = SII_S_REMOVED;
sys/geom/raid/md_sii.c
1613
sii_meta_write(tdisk->d_consumer, pd->pd_meta);
sys/geom/raid/md_sii.c
1643
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
1645
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
1646
if (pd->pd_meta != NULL) {
sys/geom/raid/md_sii.c
1647
free(pd->pd_meta, M_MD_SII);
sys/geom/raid/md_sii.c
1648
pd->pd_meta = NULL;
sys/geom/raid/md_sii.c
1650
free(pd, M_MD_SII);
sys/geom/raid/md_sii.c
415
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
418
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
419
if (pd->pd_disk_pos == id)
sys/geom/raid/md_sii.c
482
struct g_raid_md_sii_perdisk *pd, *oldpd;
sys/geom/raid/md_sii.c
490
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
494
if (pd->pd_meta != NULL)
sys/geom/raid/md_sii.c
495
disk_pos = sii_meta_disk_pos(meta, pd->pd_meta);
sys/geom/raid/md_sii.c
514
pd->pd_disk_size) {
sys/geom/raid/md_sii.c
517
pd->pd_disk_size,
sys/geom/raid/md_sii.c
532
if (disk_pos == -3 || pd->pd_disk_pos == -3) {
sys/geom/raid/md_sii.c
568
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_sii.c
581
else if (pd->pd_meta->disk_status == SII_S_CURRENT ||
sys/geom/raid/md_sii.c
582
pd->pd_meta->disk_status == SII_S_REBUILD)
sys/geom/raid/md_sii.c
592
sd->sd_size = pd->pd_disk_size - 0x800 * 512;
sys/geom/raid/md_sii.c
598
} else if (pd->pd_meta->disk_status == SII_S_REBUILD) {
sys/geom/raid/md_sii.c
602
if (pd->pd_meta->generation == meta->generation)
sys/geom/raid/md_sii.c
603
sd->sd_rebuild_pos = pd->pd_meta->rebuild_lba * 512;
sys/geom/raid/md_sii.c
606
} else if (pd->pd_meta->disk_status == SII_S_CURRENT) {
sys/geom/raid/md_sii.c
607
if (pd->pd_meta->raid_status == SII_S_ONLINE ||
sys/geom/raid/md_sii.c
608
pd->pd_meta->generation != meta->generation) {
sys/geom/raid/md_sii.c
709
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
766
pd = malloc(sizeof(*pd), M_MD_SII, M_WAITOK | M_ZERO);
sys/geom/raid/md_sii.c
767
pd->pd_disk_pos = disk_pos;
sys/geom/raid/md_sii.c
769
disk->d_md_data = (void *)pd;
sys/geom/raid/md_sii.c
786
pd = disk->d_md_data;
sys/geom/raid/md_sii.c
787
if (pd->pd_meta == NULL)
sys/geom/raid/md_sii.c
791
pd->pd_meta->generation;
sys/geom/raid/md_sii.c
823
struct g_raid_md_sii_perdisk *pd;
sys/geom/raid/md_sii.c
828
pd = (struct g_raid_md_sii_perdisk *)disk->d_md_data;
sys/geom/raid/md_sii.c
829
pdmeta = pd->pd_meta;
sys/geom/raid/md_sii.c
914
struct g_raid_md_sii_perdisk *pd;
sys/i386/i386/bios.c
571
struct pnp_sysdev *pd;
sys/i386/i386/bios.c
612
pd = &pda->node;
sys/i386/i386/bios.c
615
bzero(pd, bigdev);
sys/i386/i386/bios.c
629
if (pd->size < sizeof(struct pnp_sysdev)) {
sys/i386/i386/bios.c
639
if (!strcmp(pnp_eisaformat(pd->devid), "PNP0000")) /* ISA PIC */
sys/i386/i386/bios.c
641
if (!strcmp(pnp_eisaformat(pd->devid), "PNP0003")) /* APIC */
sys/i386/i386/bios.c
646
isa_set_vendorid(dev, pd->devid);
sys/i386/i386/bios.c
647
isa_set_logicalid(dev, pd->devid);
sys/i386/i386/bios.c
661
(!(pd->attrib & PNPATTR_NOCONFIG) &&
sys/i386/i386/bios.c
662
PNPATTR_CONFIG(pd->attrib) != PNPATTR_CONFIG_STATIC)
sys/i386/i386/bios.c
664
isa_set_pnpbios_handle(dev, pd->handle);
sys/i386/i386/bios.c
666
pnp_parse_resources(dev, &pd->devdata[0],
sys/i386/i386/bios.c
667
pd->size - sizeof(struct pnp_sysdev), 0);
sys/i386/i386/bios.c
669
device_set_desc_copy(dev, pnp_eisaformat(pd->devid));
sys/i386/i386/bios.c
672
devid = &pd->devid;
sys/i386/i386/bios.c
676
left = pd->size - sizeof(struct pnp_sysdev);
sys/i386/i386/bios.c
679
tag = pd->devdata[idx++];
sys/i386/i386/bios.c
684
compid = (u_int32_t *)(pd->devdata + idx);
sys/i386/i386/bios.c
686
printf("pnpbios: node %d compat ID 0x%08x\n", pd->handle, *compid);
sys/i386/i386/bios.c
697
idx += *(u_int16_t *)(pd->devdata + idx) + 2;
sys/i386/i386/bios.c
701
pd->handle, pnp_eisaformat(*devid), *devid);
sys/i386/i386/minidump_machdep_base.c
158
pd_entry_t *pd, pde;
sys/i386/i386/minidump_machdep_base.c
181
pd = IdlePTD; /* always mapped! */
sys/i386/i386/minidump_machdep_base.c
182
pde = pte_load(&pd[va >> PDRSHIFT]);
sys/i386/i386/minidump_machdep_base.c
281
pd = IdlePTD; /* always mapped! */
sys/i386/i386/minidump_machdep_base.c
282
pde = pte_load(&pd[va >> PDRSHIFT]);
sys/i386/pci/pci_pir.c
675
struct pci_dev_lookup *pd;
sys/i386/pci/pci_pir.c
677
pd = (struct pci_dev_lookup *)arg;
sys/i386/pci/pci_pir.c
678
if (intpin->link != pd->link || pd->bus != -1)
sys/i386/pci/pci_pir.c
680
pd->bus = entry->pe_bus;
sys/i386/pci/pci_pir.c
681
pd->device = entry->pe_device;
sys/i386/pci/pci_pir.c
682
pd->pin = intpin - entry->pe_intpin;
sys/i386/pci/pci_pir.c
688
struct pci_dev_lookup pd;
sys/i386/pci/pci_pir.c
701
pd.bus = -1;
sys/i386/pci/pci_pir.c
702
pd.link = pci_link->pl_id;
sys/i386/pci/pci_pir.c
703
pci_pir_walk_table(pir_resume_find_device, &pd);
sys/i386/pci/pci_pir.c
704
KASSERT(pd.bus != -1,
sys/i386/pci/pci_pir.c
710
pd.bus, pd.device, pd.pin + 'A',
sys/i386/pci/pci_pir.c
712
error = pci_pir_biosroute(pd.bus, pd.device, 0, pd.pin,
sys/kern/kern_exit.c
1512
struct procdesc *pd;
sys/kern/kern_exit.c
1529
pd = fp->f_data;
sys/kern/kern_exit.c
1533
KASSERT((pd->pd_flags & PDF_CLOSED) == 0,
sys/kern/kern_exit.c
1535
p, pd, pd->pd_flags));
sys/kern/kern_exit.c
1538
p = pd->pd_proc;
sys/kern/kern_fork.c
420
struct pwddesc *pd;
sys/kern/kern_fork.c
466
pd = pdinit(p1->p_pd, false);
sys/kern/kern_fork.c
471
pd = pdshare(p1->p_pd);
sys/kern/kern_fork.c
473
pd = pdcopy(p1->p_pd);
sys/kern/kern_fork.c
478
pd = pdcopy(p1->p_pd);
sys/kern/kern_fork.c
480
pd = pdshare(p1->p_pd);
sys/kern/kern_fork.c
584
p2->p_pd = pd;
sys/kern/subr_devmap.c
107
const struct devmap_entry *pd;
sys/kern/subr_devmap.c
114
for (pd = devmap_table; pd != NULL && pd->pd_size != 0; ++pd) {
sys/kern/subr_devmap.c
115
if (lowaddr > pd->pd_va)
sys/kern/subr_devmap.c
116
lowaddr = pd->pd_va;
sys/kern/subr_devmap.c
178
const struct devmap_entry *pd;
sys/kern/subr_devmap.c
189
for (pd = devmap_table; pd->pd_size != 0; ++pd) {
sys/kern/subr_devmap.c
190
pmap_preboot_map_attr(pd->pd_pa, pd->pd_va, pd->pd_size,
sys/kern/subr_devmap.c
202
const struct devmap_entry *pd;
sys/kern/subr_devmap.c
207
for (pd = devmap_table; pd->pd_size != 0; ++pd) {
sys/kern/subr_devmap.c
208
if (pa >= pd->pd_pa && pa + size <= pd->pd_pa + pd->pd_size)
sys/kern/subr_devmap.c
209
return ((void *)(pd->pd_va + (pa - pd->pd_pa)));
sys/kern/subr_devmap.c
222
const struct devmap_entry *pd;
sys/kern/subr_devmap.c
229
for (pd = devmap_table; pd->pd_size != 0; ++pd) {
sys/kern/subr_devmap.c
230
if (va >= pd->pd_va && va + size <= pd->pd_va + pd->pd_size)
sys/kern/subr_devmap.c
231
return ((vm_paddr_t)(pd->pd_pa + (va - pd->pd_va)));
sys/kern/subr_devmap.c
74
const struct devmap_entry *pd;
sys/kern/subr_devmap.c
82
for (pd = devmap_table; pd->pd_size != 0; ++pd) {
sys/kern/subr_devmap.c
84
(uintmax_t)pd->pd_pa,
sys/kern/subr_devmap.c
85
(uintmax_t)(pd->pd_pa + pd->pd_size - 1),
sys/kern/subr_devmap.c
86
(uintmax_t)pd->pd_va);
sys/kern/sys_procdesc.c
125
struct procdesc *pd;
sys/kern/sys_procdesc.c
136
pd = fp->f_data;
sys/kern/sys_procdesc.c
138
if (pd->pd_proc != NULL) {
sys/kern/sys_procdesc.c
139
*p = pd->pd_proc;
sys/kern/sys_procdesc.c
156
struct procdesc *pd;
sys/kern/sys_procdesc.c
161
pd = fp_procdesc->f_data;
sys/kern/sys_procdesc.c
162
return (pd->pd_pid);
sys/kern/sys_procdesc.c
213
struct procdesc *pd;
sys/kern/sys_procdesc.c
215
pd = malloc(sizeof(*pd), M_PROCDESC, M_WAITOK | M_ZERO);
sys/kern/sys_procdesc.c
216
pd->pd_proc = p;
sys/kern/sys_procdesc.c
217
pd->pd_pid = p->p_pid;
sys/kern/sys_procdesc.c
218
p->p_procdesc = pd;
sys/kern/sys_procdesc.c
219
pd->pd_flags = 0;
sys/kern/sys_procdesc.c
221
pd->pd_flags |= PDF_DAEMON;
sys/kern/sys_procdesc.c
222
PROCDESC_LOCK_INIT(pd);
sys/kern/sys_procdesc.c
223
knlist_init_mtx(&pd->pd_selinfo.si_note, &pd->pd_lock);
sys/kern/sys_procdesc.c
229
refcount_init(&pd->pd_refcount, 2);
sys/kern/sys_procdesc.c
259
procdesc_free(struct procdesc *pd)
sys/kern/sys_procdesc.c
268
if (refcount_release(&pd->pd_refcount)) {
sys/kern/sys_procdesc.c
269
KASSERT(pd->pd_proc == NULL,
sys/kern/sys_procdesc.c
271
KASSERT((pd->pd_flags & PDF_CLOSED),
sys/kern/sys_procdesc.c
274
if (pd->pd_pid != -1)
sys/kern/sys_procdesc.c
275
proc_id_clear(PROC_ID_PID, pd->pd_pid);
sys/kern/sys_procdesc.c
277
knlist_destroy(&pd->pd_selinfo.si_note);
sys/kern/sys_procdesc.c
278
PROCDESC_LOCK_DESTROY(pd);
sys/kern/sys_procdesc.c
279
free(pd, M_PROCDESC);
sys/kern/sys_procdesc.c
291
struct procdesc *pd;
sys/kern/sys_procdesc.c
297
pd = p->p_procdesc;
sys/kern/sys_procdesc.c
299
PROCDESC_LOCK(pd);
sys/kern/sys_procdesc.c
300
KASSERT((pd->pd_flags & PDF_CLOSED) == 0 || p->p_pptr == p->p_reaper,
sys/kern/sys_procdesc.c
303
pd->pd_flags |= PDF_EXITED;
sys/kern/sys_procdesc.c
304
pd->pd_xstat = KW_EXITCODE(p->p_xexit, p->p_xsig);
sys/kern/sys_procdesc.c
312
if (pd->pd_flags & PDF_CLOSED) {
sys/kern/sys_procdesc.c
313
PROCDESC_UNLOCK(pd);
sys/kern/sys_procdesc.c
314
pd->pd_proc = NULL;
sys/kern/sys_procdesc.c
316
procdesc_free(pd);
sys/kern/sys_procdesc.c
319
if (pd->pd_flags & PDF_SELECTED) {
sys/kern/sys_procdesc.c
320
pd->pd_flags &= ~PDF_SELECTED;
sys/kern/sys_procdesc.c
321
selwakeup(&pd->pd_selinfo);
sys/kern/sys_procdesc.c
323
KNOTE_LOCKED(&pd->pd_selinfo.si_note, NOTE_EXIT);
sys/kern/sys_procdesc.c
324
PROCDESC_UNLOCK(pd);
sys/kern/sys_procdesc.c
338
struct procdesc *pd;
sys/kern/sys_procdesc.c
343
pd = p->p_procdesc;
sys/kern/sys_procdesc.c
344
pd->pd_proc = NULL;
sys/kern/sys_procdesc.c
346
procdesc_free(pd);
sys/kern/sys_procdesc.c
358
struct procdesc *pd;
sys/kern/sys_procdesc.c
363
pd = fp->f_data;
sys/kern/sys_procdesc.c
368
PROCDESC_LOCK(pd);
sys/kern/sys_procdesc.c
369
pd->pd_flags |= PDF_CLOSED;
sys/kern/sys_procdesc.c
370
PROCDESC_UNLOCK(pd);
sys/kern/sys_procdesc.c
371
p = pd->pd_proc;
sys/kern/sys_procdesc.c
397
pd->pd_proc = NULL;
sys/kern/sys_procdesc.c
399
pd->pd_pid = -1;
sys/kern/sys_procdesc.c
400
procdesc_free(pd);
sys/kern/sys_procdesc.c
415
if ((pd->pd_flags & PDF_DAEMON) == 0)
sys/kern/sys_procdesc.c
425
procdesc_free(pd);
sys/kern/sys_procdesc.c
433
struct procdesc *pd;
sys/kern/sys_procdesc.c
437
pd = fp->f_data;
sys/kern/sys_procdesc.c
438
PROCDESC_LOCK(pd);
sys/kern/sys_procdesc.c
439
if (pd->pd_flags & PDF_EXITED)
sys/kern/sys_procdesc.c
442
selrecord(td, &pd->pd_selinfo);
sys/kern/sys_procdesc.c
443
pd->pd_flags |= PDF_SELECTED;
sys/kern/sys_procdesc.c
445
PROCDESC_UNLOCK(pd);
sys/kern/sys_procdesc.c
452
struct procdesc *pd;
sys/kern/sys_procdesc.c
454
pd = kn->kn_fp->f_data;
sys/kern/sys_procdesc.c
455
knlist_remove(&pd->pd_selinfo.si_note, kn, 0);
sys/kern/sys_procdesc.c
461
struct procdesc *pd;
sys/kern/sys_procdesc.c
464
pd = kn->kn_fp->f_data;
sys/kern/sys_procdesc.c
470
event = pd->pd_flags & PDF_EXITED ? NOTE_EXIT : 0;
sys/kern/sys_procdesc.c
484
kn->kn_data = pd->pd_xstat;
sys/kern/sys_procdesc.c
503
struct procdesc *pd;
sys/kern/sys_procdesc.c
505
pd = fp->f_data;
sys/kern/sys_procdesc.c
510
knlist_add(&pd->pd_selinfo.si_note, kn, 0);
sys/kern/sys_procdesc.c
520
struct procdesc *pd;
sys/kern/sys_procdesc.c
529
pd = fp->f_data;
sys/kern/sys_procdesc.c
531
if (pd->pd_proc != NULL) {
sys/kern/sys_procdesc.c
532
PROC_LOCK(pd->pd_proc);
sys/kern/sys_procdesc.c
533
AUDIT_ARG_PROCESS(pd->pd_proc);
sys/kern/sys_procdesc.c
536
pstart = pd->pd_proc->p_stats->p_start;
sys/kern/sys_procdesc.c
543
if (pd->pd_proc->p_state != PRS_ZOMBIE)
sys/kern/sys_procdesc.c
547
sb->st_uid = pd->pd_proc->p_ucred->cr_ruid;
sys/kern/sys_procdesc.c
548
sb->st_gid = pd->pd_proc->p_ucred->cr_rgid;
sys/kern/sys_procdesc.c
549
PROC_UNLOCK(pd->pd_proc);
sys/kern/uipc_mqueue.c
1002
pn = mqfs_create_file(pd, cnp->cn_nameptr, cnp->cn_namelen,
sys/kern/uipc_mqueue.c
1384
struct mqfs_node *pd;
sys/kern/uipc_mqueue.c
1395
pd = VTON(vp);
sys/kern/uipc_mqueue.c
1416
LIST_FOREACH(pn, &pd->mn_children, mn_sibling) {
sys/kern/uipc_mqueue.c
1488
struct mqfs_node *pd = VTON(ap->a_dvp);
sys/kern/uipc_mqueue.c
1492
if (pd->mn_type != mqfstype_root && pd->mn_type != mqfstype_dir)
sys/kern/uipc_mqueue.c
1495
pn = mqfs_create_dir(pd, cnp->cn_nameptr, cnp->cn_namelen,
sys/kern/uipc_mqueue.c
826
mqfs_search(struct mqfs_node *pd, const char *name, int len, struct ucred *cred)
sys/kern/uipc_mqueue.c
831
sx_assert(&pd->mn_info->mi_lock, SX_LOCKED);
sys/kern/uipc_mqueue.c
833
LIST_FOREACH(pn, &pd->mn_children, mn_sibling) {
sys/kern/uipc_mqueue.c
851
struct mqfs_node *pd;
sys/kern/uipc_mqueue.c
867
pd = VTON(dvp);
sys/kern/uipc_mqueue.c
869
mqfs = pd->mn_info;
sys/kern/uipc_mqueue.c
887
pn = pd;
sys/kern/uipc_mqueue.c
900
KASSERT(pd->mn_parent, ("non-root directory has no parent"));
sys/kern/uipc_mqueue.c
901
pn = pd->mn_parent;
sys/kern/uipc_mqueue.c
909
pn = mqfs_search(pd, pname, namelen, cnp->cn_cred);
sys/kern/uipc_mqueue.c
990
struct mqfs_node *pd;
sys/kern/uipc_mqueue.c
995
pd = VTON(ap->a_dvp);
sys/kern/uipc_mqueue.c
996
if (pd->mn_type != mqfstype_root && pd->mn_type != mqfstype_dir)
sys/libkern/qsort.c
114
char *pa, *pb, *pc, *pd, *pl, *pm, *pn;
sys/libkern/qsort.c
145
pc = pd = (char *)a + (n - 1) * es;
sys/libkern/qsort.c
156
swap(pc, pd);
sys/libkern/qsort.c
157
pd -= es;
sys/libkern/qsort.c
171
d1 = MIN(pd - pc, pn - pd - es);
sys/libkern/qsort.c
175
d2 = pd - pc;
sys/net/pfvar.h
1434
struct pf_pdesc *pd;
sys/net/pfvar.h
2019
struct pf_pdesc pd;
sys/net/pfvar.h
3166
int pf_normalize_mss(struct pf_pdesc *pd);
sys/netinet/sctp_cc_functions.c
1248
SDT_PROBE5(sctp, cwnd, net, pd,
sys/netinet/sctp_kdtrace.c
127
SDT_PROBE_DEFINE5(sctp, cwnd, net, pd,
sys/netinet/sctp_kdtrace.h
72
SDT_PROBE_DECLARE(sctp, cwnd, net, pd);
sys/netipsec/key.c
574
#define KEY_SETSECSPIDX(_dir, s, d, ps, pd, ulp, idx) \
sys/netipsec/key.c
579
(idx)->prefd = (pd); \
sys/netpfil/pf/if_pflog.c
246
struct pf_kruleset *ruleset, struct pf_pdesc *pd, int lookupsafe,
sys/netpfil/pf/if_pflog.c
252
if (rm == NULL || pd == NULL)
sys/netpfil/pf/if_pflog.c
266
hdr.af = pd->af;
sys/netpfil/pf/if_pflog.c
269
memcpy(hdr.ifname, pd->kif->pfik_name, sizeof(hdr.ifname));
sys/netpfil/pf/if_pflog.c
287
if (trigger->log & PF_LOG_USER && !pd->lookup.done && lookupsafe)
sys/netpfil/pf/if_pflog.c
288
pd->lookup.done = pf_socket_lookup(pd);
sys/netpfil/pf/if_pflog.c
289
if (trigger->log & PF_LOG_USER && pd->lookup.done > 0)
sys/netpfil/pf/if_pflog.c
290
hdr.uid = pd->lookup.uid;
sys/netpfil/pf/if_pflog.c
296
hdr.dir = pd->dir;
sys/netpfil/pf/if_pflog.c
299
if (pd->af == AF_INET && pd->dir == PF_OUT) {
sys/netpfil/pf/if_pflog.c
302
ip = mtod(pd->m, struct ip *);
sys/netpfil/pf/if_pflog.c
304
ip->ip_sum = in_cksum(pd->m, ip->ip_hl << 2);
sys/netpfil/pf/if_pflog.c
309
if_inc_counter(ifn, IFCOUNTER_OBYTES, pd->m->m_pkthdr.len);
sys/netpfil/pf/if_pflog.c
310
bpf_mtap2(ifn->if_bpf, &hdr, PFLOG_HDRLEN, pd->m);
sys/netpfil/pf/if_pfsync.c
2220
struct pfsync_deferral *pd;
sys/netpfil/pf/if_pfsync.c
2245
pd = malloc(sizeof(*pd), M_PFSYNC, M_NOWAIT);
sys/netpfil/pf/if_pfsync.c
2246
if (pd == NULL) {
sys/netpfil/pf/if_pfsync.c
2255
pd->pd_sc = sc;
sys/netpfil/pf/if_pfsync.c
2256
pd->pd_st = st;
sys/netpfil/pf/if_pfsync.c
2258
pd->pd_m = m;
sys/netpfil/pf/if_pfsync.c
2260
TAILQ_INSERT_TAIL(&b->b_deferrals, pd, pd_entry);
sys/netpfil/pf/if_pfsync.c
2261
callout_init_mtx(&pd->pd_tmo, &b->b_mtx, CALLOUT_RETURNUNLOCKED);
sys/netpfil/pf/if_pfsync.c
2262
callout_reset(&pd->pd_tmo, (V_pfsync_defer_timeout * hz) / 1000,
sys/netpfil/pf/if_pfsync.c
2263
pfsync_defer_tmo, pd);
sys/netpfil/pf/if_pfsync.c
2272
pfsync_undefer(struct pfsync_deferral *pd, int drop)
sys/netpfil/pf/if_pfsync.c
2274
struct pfsync_softc *sc = pd->pd_sc;
sys/netpfil/pf/if_pfsync.c
2275
struct mbuf *m = pd->pd_m;
sys/netpfil/pf/if_pfsync.c
2276
struct pf_kstate *st = pd->pd_st;
sys/netpfil/pf/if_pfsync.c
2281
TAILQ_REMOVE(&b->b_deferrals, pd, pd_entry);
sys/netpfil/pf/if_pfsync.c
2283
pd->pd_st->state_flags &= ~PFSTATE_ACK; /* XXX: locking! */
sys/netpfil/pf/if_pfsync.c
2284
free(pd, M_PFSYNC);
sys/netpfil/pf/if_pfsync.c
2299
struct pfsync_deferral *pd = arg;
sys/netpfil/pf/if_pfsync.c
2300
struct pfsync_softc *sc = pd->pd_sc;
sys/netpfil/pf/if_pfsync.c
2301
struct mbuf *m = pd->pd_m;
sys/netpfil/pf/if_pfsync.c
2302
struct pf_kstate *st = pd->pd_st;
sys/netpfil/pf/if_pfsync.c
2311
TAILQ_REMOVE(&b->b_deferrals, pd, pd_entry);
sys/netpfil/pf/if_pfsync.c
2313
pd->pd_st->state_flags &= ~PFSTATE_ACK; /* XXX: locking! */
sys/netpfil/pf/if_pfsync.c
2315
free(pd, M_PFSYNC);
sys/netpfil/pf/if_pfsync.c
2338
struct pfsync_deferral *pd;
sys/netpfil/pf/if_pfsync.c
2343
TAILQ_FOREACH(pd, &b->b_deferrals, pd_entry) {
sys/netpfil/pf/if_pfsync.c
2344
if (pd->pd_st == st) {
sys/netpfil/pf/if_pfsync.c
2345
if (callout_stop(&pd->pd_tmo) > 0)
sys/netpfil/pf/if_pfsync.c
2346
pfsync_undefer(pd, drop);
sys/netpfil/pf/if_pfsync.c
465
struct pfsync_deferral *pd =
sys/netpfil/pf/if_pfsync.c
468
ret = callout_stop(&pd->pd_tmo);
sys/netpfil/pf/if_pfsync.c
470
pfsync_undefer(pd, 1);
sys/netpfil/pf/if_pfsync.c
473
callout_drain(&pd->pd_tmo);
sys/netpfil/pf/pf.c
10002
tmp = pd->act.dnrpipe;
sys/netpfil/pf/pf.c
10003
pd->act.dnrpipe = pd->act.dnpipe;
sys/netpfil/pf/pf.c
10004
pd->act.dnpipe = tmp;
sys/netpfil/pf/pf.c
10021
error = pf_dummynet_route(pd, s, r, ifp,
sys/netpfil/pf/pf.c
10035
if (pd->act.rt != PF_DUPTO) {
sys/netpfil/pf/pf.c
10037
MPASS(m0 == pd->m);
sys/netpfil/pf/pf.c
10038
PACKET_UNDO_NAT(pd,
sys/netpfil/pf/pf.c
10044
ifp->if_mtu, pd->af, r, pd->act.rtableid);
sys/netpfil/pf/pf.c
10065
pd->pf_mtag = pf_find_mtag(md);
sys/netpfil/pf/pf.c
10066
error = pf_dummynet_route(pd, s, r, ifp,
sys/netpfil/pf/pf.c
10082
if (pd->act.rt != PF_DUPTO)
sys/netpfil/pf/pf.c
10083
pd->m = NULL;
sys/netpfil/pf/pf.c
10100
struct pf_kstate *s, struct pf_pdesc *pd, struct inpcb *inp)
sys/netpfil/pf/pf.c
10111
KASSERT(pd->m && r && oifp, ("%s: invalid parameters", __func__));
sys/netpfil/pf/pf.c
10113
SDT_PROBE4(pf, ip6, route_to, entry, pd->m, pd, s, oifp);
sys/netpfil/pf/pf.c
10121
KASSERT(pd->dir == PF_IN || pd->dir == PF_OUT ||
sys/netpfil/pf/pf.c
10125
if ((pd->pf_mtag == NULL &&
sys/netpfil/pf/pf.c
10126
((pd->pf_mtag = pf_get_mtag(pd->m)) == NULL)) ||
sys/netpfil/pf/pf.c
10127
pd->pf_mtag->routed++ > 3) {
sys/netpfil/pf/pf.c
10128
m0 = pd->m;
sys/netpfil/pf/pf.c
10129
pd->m = NULL;
sys/netpfil/pf/pf.c
10135
if (pd->act.rt_kif != NULL)
sys/netpfil/pf/pf.c
10136
ifp = pd->act.rt_kif->pfik_ifp;
sys/netpfil/pf/pf.c
10138
if (pd->act.rt == PF_DUPTO) {
sys/netpfil/pf/pf.c
10139
if ((pd->pf_mtag->flags & PF_MTAG_FLAG_DUPLICATED)) {
sys/netpfil/pf/pf.c
10147
m0 = pd->m;
sys/netpfil/pf/pf.c
10148
pd->m = NULL;
sys/netpfil/pf/pf.c
10154
pd->pf_mtag->flags |= PF_MTAG_FLAG_DUPLICATED;
sys/netpfil/pf/pf.c
10155
if (((m0 = m_dup(pd->m, M_NOWAIT)) == NULL)) {
sys/netpfil/pf/pf.c
10162
if ((pd->act.rt == PF_REPLYTO) == (r_dir == pd->dir)) {
sys/netpfil/pf/pf.c
10163
if (pd->af == pd->naf) {
sys/netpfil/pf/pf.c
10164
pf_dummynet(pd, s, r, &pd->m);
sys/netpfil/pf/pf.c
10179
if (pd->act.rt_kif && pd->act.rt_kif->pfik_ifp &&
sys/netpfil/pf/pf.c
10180
pd->af != pd->naf) {
sys/netpfil/pf/pf.c
10181
if (pd->act.rt == PF_ROUTETO && r->naf != AF_INET6) {
sys/netpfil/pf/pf.c
10185
if (pd->act.rt == PF_REPLYTO && r->naf != AF_INET) {
sys/netpfil/pf/pf.c
10190
m0 = pd->m;
sys/netpfil/pf/pf.c
10198
pf_addrcpy((struct pf_addr *)&dst.sin6_addr, &pd->act.rt_addr,
sys/netpfil/pf/pf.c
10201
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
10203
if (r->rt != PF_DUPTO && pd->naf == pd->af)
sys/netpfil/pf/pf.c
10205
ICMP6_TIME_EXCEED_TRANSIT, 0, pd->af, r,
sys/netpfil/pf/pf.c
10206
pd->act.rtableid);
sys/netpfil/pf/pf.c
10214
if (ifp == NULL && (pd->af != pd->naf)) {
sys/netpfil/pf/pf.c
10231
if (pd->af != pd->naf) {
sys/netpfil/pf/pf.c
10232
struct udphdr *uh = &pd->hdr.udp;
sys/netpfil/pf/pf.c
10234
if (pd->proto == IPPROTO_UDP && uh->uh_sum == 0) {
sys/netpfil/pf/pf.c
10237
m_copyback(m0, pd->off, sizeof(*uh), pd->hdr.any);
sys/netpfil/pf/pf.c
10242
m0 = pd->m;
sys/netpfil/pf/pf.c
10243
pd->m = NULL;
sys/netpfil/pf/pf.c
10256
MPASS(r->rt == PF_REPLYTO || (pd->af != pd->naf && s->direction == PF_IN));
sys/netpfil/pf/pf.c
10258
if (pd->act.rt == PF_REPLYTO) {
sys/netpfil/pf/pf.c
10263
if (r->rt == PF_DUPTO || (pd->af != pd->naf && s->direction == PF_IN))
sys/netpfil/pf/pf.c
10266
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
10272
&out_kif->pfik_bytes[pd->naf == AF_INET6][1]
sys/netpfil/pf/pf.c
10273
[action != PF_PASS && action != PF_AFRT], pd->tot_len);
sys/netpfil/pf/pf.c
10275
&out_kif->pfik_packets[pd->naf == AF_INET6][1]
sys/netpfil/pf/pf.c
10280
ifp, &m0, inp, &pd->act) != PF_PASS) {
sys/netpfil/pf/pf.c
10311
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
10317
pd->dir = PF_OUT;
sys/netpfil/pf/pf.c
10328
tmp = pd->act.dnrpipe;
sys/netpfil/pf/pf.c
10329
pd->act.dnrpipe = pd->act.dnpipe;
sys/netpfil/pf/pf.c
10330
pd->act.dnpipe = tmp;
sys/netpfil/pf/pf.c
10349
pf_dummynet_route(pd, s, r, ifp, sintosa(&dst), &md);
sys/netpfil/pf/pf.c
10358
if (pd->act.rt != PF_DUPTO) {
sys/netpfil/pf/pf.c
10360
MPASS(m0 == pd->m);
sys/netpfil/pf/pf.c
10361
PACKET_UNDO_NAT(pd,
sys/netpfil/pf/pf.c
10368
ifp->if_mtu, pd->af, r, pd->act.rtableid);
sys/netpfil/pf/pf.c
10377
if (pd->act.rt != PF_DUPTO)
sys/netpfil/pf/pf.c
10378
pd->m = NULL;
sys/netpfil/pf/pf.c
10514
pf_pdesc_to_dnflow(const struct pf_pdesc *pd, const struct pf_krule *r,
sys/netpfil/pf/pf.c
10518
sa_family_t af = pd->naf;
sys/netpfil/pf/pf.c
10525
dndir = pd->dir;
sys/netpfil/pf/pf.c
10528
if (pd->pf_mtag->flags & PF_MTAG_FLAG_DUMMYNETED)
sys/netpfil/pf/pf.c
10533
if (pd->dport != NULL)
sys/netpfil/pf/pf.c
10534
dnflow->f_id.dst_port = ntohs(*pd->dport);
sys/netpfil/pf/pf.c
10535
if (pd->sport != NULL)
sys/netpfil/pf/pf.c
10536
dnflow->f_id.src_port = ntohs(*pd->sport);
sys/netpfil/pf/pf.c
10538
if (pd->dir == PF_IN)
sys/netpfil/pf/pf.c
10543
if (pd->dir != dndir && pd->act.dnrpipe) {
sys/netpfil/pf/pf.c
10544
dnflow->rule.info = pd->act.dnrpipe;
sys/netpfil/pf/pf.c
10546
else if (pd->dir == dndir && pd->act.dnpipe) {
sys/netpfil/pf/pf.c
10547
dnflow->rule.info = pd->act.dnpipe;
sys/netpfil/pf/pf.c
10554
if (r->free_flags & PFRULE_DN_IS_PIPE || pd->act.flags & PFSTATE_DN_IS_PIPE)
sys/netpfil/pf/pf.c
10557
dnflow->f_id.proto = pd->proto;
sys/netpfil/pf/pf.c
10567
s->key[PF_SK_STACK]->addr[pd->sidx].v4.s_addr);
sys/netpfil/pf/pf.c
10569
s->key[PF_SK_STACK]->addr[pd->didx].v4.s_addr);
sys/netpfil/pf/pf.c
10571
dnflow->f_id.src_ip = ntohl(pd->src->v4.s_addr);
sys/netpfil/pf/pf.c
10572
dnflow->f_id.dst_ip = ntohl(pd->dst->v4.s_addr);
sys/netpfil/pf/pf.c
10580
s->key[PF_SK_STACK]->addr[pd->sidx].v6;
sys/netpfil/pf/pf.c
10582
s->key[PF_SK_STACK]->addr[pd->didx].v6;
sys/netpfil/pf/pf.c
10584
dnflow->f_id.src_ip6 = pd->src->v6;
sys/netpfil/pf/pf.c
10585
dnflow->f_id.dst_ip6 = pd->dst->v6;
sys/netpfil/pf/pf.c
10596
if (pd->naf == AF_INET6)
sys/netpfil/pf/pf.c
10655
pf_dummynet(struct pf_pdesc *pd, struct pf_kstate *s,
sys/netpfil/pf/pf.c
10658
return (pf_dummynet_route(pd, s, r, NULL, NULL, m0));
sys/netpfil/pf/pf.c
10662
pf_dummynet_route(struct pf_pdesc *pd, struct pf_kstate *s,
sys/netpfil/pf/pf.c
10670
if (pd->act.dnpipe == 0 && pd->act.dnrpipe == 0)
sys/netpfil/pf/pf.c
10679
if (pd->pf_mtag == NULL &&
sys/netpfil/pf/pf.c
10680
((pd->pf_mtag = pf_get_mtag(*m0)) == NULL)) {
sys/netpfil/pf/pf.c
10687
pd->pf_mtag->flags |= PF_MTAG_FLAG_ROUTE_TO;
sys/netpfil/pf/pf.c
10689
pd->pf_mtag->if_index = ifp->if_index;
sys/netpfil/pf/pf.c
10690
pd->pf_mtag->if_idxgen = ifp->if_idxgen;
sys/netpfil/pf/pf.c
10696
memcpy(&pd->pf_mtag->dst, sa,
sys/netpfil/pf/pf.c
10700
memcpy(&pd->pf_mtag->dst, sa,
sys/netpfil/pf/pf.c
10710
(pd->af == AF_INET && IN_LOOPBACK(ntohl(pd->dst->v4.s_addr))) ||
sys/netpfil/pf/pf.c
10712
(pd->af == AF_INET6 && IN6_IS_ADDR_LOOPBACK(&pd->dst->v6)))) {
sys/netpfil/pf/pf.c
10721
if (pf_pdesc_to_dnflow(pd, r, s, &dnflow)) {
sys/netpfil/pf/pf.c
10722
pd->pf_mtag->flags |= PF_MTAG_FLAG_DUMMYNET;
sys/netpfil/pf/pf.c
10723
pd->pf_mtag->flags |= PF_MTAG_FLAG_DUMMYNETED;
sys/netpfil/pf/pf.c
10726
pd->pf_mtag->flags &= ~PF_MTAG_FLAG_ROUTE_TO;
sys/netpfil/pf/pf.c
10727
pf_dummynet_flag_remove(*m0, pd->pf_mtag);
sys/netpfil/pf/pf.c
10735
pf_walk_option(struct pf_pdesc *pd, struct ip *h, int off, int end,
sys/netpfil/pf/pf.c
10741
if (pd->m->m_pkthdr.len < end) {
sys/netpfil/pf/pf.c
10748
m_copydata(pd->m, off, end - off, opts);
sys/netpfil/pf/pf.c
10778
pd->badopts |= PF_OPT_ROUTER_ALERT;
sys/netpfil/pf/pf.c
10781
pd->badopts |= PF_OPT_OTHER;
sys/netpfil/pf/pf.c
10791
pf_walk_header(struct pf_pdesc *pd, struct ip *h, u_short *reason)
sys/netpfil/pf/pf.c
10803
if (pf_walk_option(pd, h, pd->off + sizeof(struct ip),
sys/netpfil/pf/pf.c
10804
pd->off + hlen, reason) != PF_PASS)
sys/netpfil/pf/pf.c
10807
if (pd->badopts == 0)
sys/netpfil/pf/pf.c
10808
pd->badopts |= PF_OPT_OTHER;
sys/netpfil/pf/pf.c
10810
end = pd->off + ntohs(h->ip_len);
sys/netpfil/pf/pf.c
10811
pd->off += hlen;
sys/netpfil/pf/pf.c
10812
pd->proto = h->ip_p;
sys/netpfil/pf/pf.c
10814
if (pd->proto == IPPROTO_IGMP) {
sys/netpfil/pf/pf.c
10825
pd->badopts &= ~PF_OPT_ROUTER_ALERT;
sys/netpfil/pf/pf.c
10831
switch (pd->proto) {
sys/netpfil/pf/pf.c
10835
end < pd->off + sizeof(ext))
sys/netpfil/pf/pf.c
10837
if (!pf_pull_hdr(pd->m, pd->off, &ext, sizeof(ext),
sys/netpfil/pf/pf.c
10842
pd->off += (ext.ah_len + 2) * 4;
sys/netpfil/pf/pf.c
10843
pd->proto = ext.ah_nxt;
sys/netpfil/pf/pf.c
10856
pf_walk_option6(struct pf_pdesc *pd, struct ip6_hdr *h, int off, int end,
sys/netpfil/pf/pf.c
10863
if (!pf_pull_hdr(pd->m, off, &opt.ip6o_type,
sys/netpfil/pf/pf.c
10872
if (!pf_pull_hdr(pd->m, off, &opt, sizeof(opt),
sys/netpfil/pf/pf.c
10886
pd->badopts |= PF_OPT_JUMBO;
sys/netpfil/pf/pf.c
10887
if (pd->jumbolen != 0) {
sys/netpfil/pf/pf.c
10897
if (!pf_pull_hdr(pd->m, off, &jumbo, sizeof(jumbo),
sys/netpfil/pf/pf.c
10902
memcpy(&pd->jumbolen, jumbo.ip6oj_jumbo_len,
sys/netpfil/pf/pf.c
10903
sizeof(pd->jumbolen));
sys/netpfil/pf/pf.c
10904
pd->jumbolen = ntohl(pd->jumbolen);
sys/netpfil/pf/pf.c
10905
if (pd->jumbolen < IPV6_MAXPACKET) {
sys/netpfil/pf/pf.c
10912
pd->badopts |= PF_OPT_ROUTER_ALERT;
sys/netpfil/pf/pf.c
10915
pd->badopts |= PF_OPT_OTHER;
sys/netpfil/pf/pf.c
10925
pf_walk_header6(struct pf_pdesc *pd, struct ip6_hdr *h, u_short *reason)
sys/netpfil/pf/pf.c
10934
pd->off += sizeof(struct ip6_hdr);
sys/netpfil/pf/pf.c
10935
end = pd->off + ntohs(h->ip6_plen);
sys/netpfil/pf/pf.c
10936
pd->fragoff = pd->extoff = pd->jumbolen = 0;
sys/netpfil/pf/pf.c
10937
pd->proto = h->ip6_nxt;
sys/netpfil/pf/pf.c
10939
switch (pd->proto) {
sys/netpfil/pf/pf.c
10942
pd->badopts |= PF_OPT_OTHER;
sys/netpfil/pf/pf.c
10945
if (!pf_pull_hdr(pd->m, pd->off, &ext, sizeof(ext),
sys/netpfil/pf/pf.c
10950
if (pf_walk_option6(pd, h, pd->off + sizeof(ext),
sys/netpfil/pf/pf.c
10951
pd->off + (ext.ip6e_len + 1) * 8,
sys/netpfil/pf/pf.c
10955
if (pd->badopts == 0)
sys/netpfil/pf/pf.c
10956
pd->badopts |= PF_OPT_OTHER;
sys/netpfil/pf/pf.c
10959
switch (pd->proto) {
sys/netpfil/pf/pf.c
10967
if (pd->jumbolen != 0) {
sys/netpfil/pf/pf.c
10972
if (!pf_pull_hdr(pd->m, pd->off, &frag, sizeof(frag),
sys/netpfil/pf/pf.c
10979
pd->fragoff = pd->off;
sys/netpfil/pf/pf.c
10984
pd->fragoff = pd->off;
sys/netpfil/pf/pf.c
10985
pd->off += sizeof(frag);
sys/netpfil/pf/pf.c
10986
pd->proto = frag.ip6f_nxt;
sys/netpfil/pf/pf.c
10995
if (pd->fragoff != 0 && end < pd->off + sizeof(rthdr)) {
sys/netpfil/pf/pf.c
10996
pd->off = pd->fragoff;
sys/netpfil/pf/pf.c
10997
pd->proto = IPPROTO_FRAGMENT;
sys/netpfil/pf/pf.c
11000
if (!pf_pull_hdr(pd->m, pd->off, &rthdr, sizeof(rthdr),
sys/netpfil/pf/pf.c
11013
if (pd->proto == IPPROTO_HOPOPTS && hdr_cnt > 0) {
sys/netpfil/pf/pf.c
11021
if (!pf_pull_hdr(pd->m, pd->off, &ext, sizeof(ext),
sys/netpfil/pf/pf.c
11027
if (pd->fragoff != 0 && end < pd->off + sizeof(ext)) {
sys/netpfil/pf/pf.c
11028
pd->off = pd->fragoff;
sys/netpfil/pf/pf.c
11029
pd->proto = IPPROTO_FRAGMENT;
sys/netpfil/pf/pf.c
11033
if (pd->fragoff == 0)
sys/netpfil/pf/pf.c
11034
pd->extoff = pd->off;
sys/netpfil/pf/pf.c
11035
if (pd->proto == IPPROTO_HOPOPTS && pd->fragoff == 0 &&
sys/netpfil/pf/pf.c
11036
ntohs(h->ip6_plen) == 0 && pd->jumbolen != 0) {
sys/netpfil/pf/pf.c
11041
if (pd->proto == IPPROTO_AH)
sys/netpfil/pf/pf.c
11042
pd->off += (ext.ip6e_len + 2) * 4;
sys/netpfil/pf/pf.c
11044
pd->off += (ext.ip6e_len + 1) * 8;
sys/netpfil/pf/pf.c
11045
pd->proto = ext.ip6e_nxt;
sys/netpfil/pf/pf.c
11049
if (pd->fragoff != 0 && end < pd->off + sizeof(icmp6)) {
sys/netpfil/pf/pf.c
11050
pd->off = pd->fragoff;
sys/netpfil/pf/pf.c
11051
pd->proto = IPPROTO_FRAGMENT;
sys/netpfil/pf/pf.c
11054
if (!pf_pull_hdr(pd->m, pd->off, &icmp6, sizeof(icmp6),
sys/netpfil/pf/pf.c
11079
pd->badopts &= ~PF_OPT_ROUTER_ALERT;
sys/netpfil/pf/pf.c
11087
if (pd->fragoff != 0 && end < pd->off +
sys/netpfil/pf/pf.c
11088
(pd->proto == IPPROTO_TCP ? sizeof(struct tcphdr) :
sys/netpfil/pf/pf.c
11089
pd->proto == IPPROTO_UDP ? sizeof(struct udphdr) :
sys/netpfil/pf/pf.c
11090
pd->proto == IPPROTO_SCTP ? sizeof(struct sctphdr) :
sys/netpfil/pf/pf.c
11092
pd->off = pd->fragoff;
sys/netpfil/pf/pf.c
11093
pd->proto = IPPROTO_FRAGMENT;
sys/netpfil/pf/pf.c
11107
pf_init_pdesc(struct pf_pdesc *pd, struct mbuf *m)
sys/netpfil/pf/pf.c
11109
memset(pd, 0, sizeof(*pd));
sys/netpfil/pf/pf.c
11110
pd->pf_mtag = pf_find_mtag(m);
sys/netpfil/pf/pf.c
11111
pd->m = m;
sys/netpfil/pf/pf.c
11115
pf_setup_pdesc(sa_family_t af, int dir, struct pf_pdesc *pd, struct mbuf **m0,
sys/netpfil/pf/pf.c
11119
pd->dir = dir;
sys/netpfil/pf/pf.c
11120
pd->kif = kif;
sys/netpfil/pf/pf.c
11121
pd->m = *m0;
sys/netpfil/pf/pf.c
11122
pd->sidx = (dir == PF_IN) ? 0 : 1;
sys/netpfil/pf/pf.c
11123
pd->didx = (dir == PF_IN) ? 1 : 0;
sys/netpfil/pf/pf.c
11124
pd->af = pd->naf = af;
sys/netpfil/pf/pf.c
11128
TAILQ_INIT(&pd->sctp_multihome_jobs);
sys/netpfil/pf/pf.c
11130
memcpy(&pd->act, default_actions, sizeof(pd->act));
sys/netpfil/pf/pf.c
11132
if (pd->pf_mtag && pd->pf_mtag->dnpipe) {
sys/netpfil/pf/pf.c
11133
pd->act.dnpipe = pd->pf_mtag->dnpipe;
sys/netpfil/pf/pf.c
11134
pd->act.flags = pd->pf_mtag->dnflags;
sys/netpfil/pf/pf.c
11143
(pd->m = *m0 = m_pullup(*m0, sizeof(struct ip))) == NULL) {
sys/netpfil/pf/pf.c
11152
h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf.c
11153
if (pd->m->m_pkthdr.len < ntohs(h->ip_len)) {
sys/netpfil/pf/pf.c
11159
if (pf_normalize_ip(reason, pd) != PF_PASS) {
sys/netpfil/pf/pf.c
11161
*m0 = pd->m;
sys/netpfil/pf/pf.c
11165
*m0 = pd->m;
sys/netpfil/pf/pf.c
11166
h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf.c
11168
if (pf_walk_header(pd, h, reason) != PF_PASS) {
sys/netpfil/pf/pf.c
11173
pd->src = (struct pf_addr *)&h->ip_src;
sys/netpfil/pf/pf.c
11174
pd->dst = (struct pf_addr *)&h->ip_dst;
sys/netpfil/pf/pf.c
11175
pf_addrcpy(&pd->osrc, pd->src, af);
sys/netpfil/pf/pf.c
11176
pf_addrcpy(&pd->odst, pd->dst, af);
sys/netpfil/pf/pf.c
11177
pd->ip_sum = &h->ip_sum;
sys/netpfil/pf/pf.c
11178
pd->tos = h->ip_tos & ~IPTOS_ECN_MASK;
sys/netpfil/pf/pf.c
11179
pd->ttl = h->ip_ttl;
sys/netpfil/pf/pf.c
11180
pd->tot_len = ntohs(h->ip_len);
sys/netpfil/pf/pf.c
11181
pd->act.rtableid = -1;
sys/netpfil/pf/pf.c
11182
pd->df = h->ip_off & htons(IP_DF);
sys/netpfil/pf/pf.c
11183
pd->virtual_proto = (h->ip_off & htons(IP_MF | IP_OFFMASK)) ?
sys/netpfil/pf/pf.c
11184
PF_VPROTO_FRAGMENT : pd->proto;
sys/netpfil/pf/pf.c
11194
(pd->m = *m0 = m_pullup(*m0, sizeof(struct ip6_hdr))) == NULL) {
sys/netpfil/pf/pf.c
11203
h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf.c
11204
if (pd->m->m_pkthdr.len <
sys/netpfil/pf/pf.c
11220
if (pf_walk_header6(pd, h, reason) != PF_PASS) {
sys/netpfil/pf/pf.c
11225
h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf.c
11226
pd->src = (struct pf_addr *)&h->ip6_src;
sys/netpfil/pf/pf.c
11227
pd->dst = (struct pf_addr *)&h->ip6_dst;
sys/netpfil/pf/pf.c
11228
pf_addrcpy(&pd->osrc, pd->src, af);
sys/netpfil/pf/pf.c
11229
pf_addrcpy(&pd->odst, pd->dst, af);
sys/netpfil/pf/pf.c
11230
pd->ip_sum = NULL;
sys/netpfil/pf/pf.c
11231
pd->tos = IPV6_DSCP(h);
sys/netpfil/pf/pf.c
11232
pd->ttl = h->ip6_hlim;
sys/netpfil/pf/pf.c
11233
pd->tot_len = ntohs(h->ip6_plen) + sizeof(struct ip6_hdr);
sys/netpfil/pf/pf.c
11234
pd->act.rtableid = -1;
sys/netpfil/pf/pf.c
11236
pd->virtual_proto = (pd->fragoff != 0) ?
sys/netpfil/pf/pf.c
11237
PF_VPROTO_FRAGMENT : pd->proto;
sys/netpfil/pf/pf.c
11240
if (pf_normalize_ip6(pd->fragoff, reason, pd) !=
sys/netpfil/pf/pf.c
11242
*m0 = pd->m;
sys/netpfil/pf/pf.c
11246
*m0 = pd->m;
sys/netpfil/pf/pf.c
11247
if (pd->m == NULL) {
sys/netpfil/pf/pf.c
11254
h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf.c
11255
pd->src = (struct pf_addr *)&h->ip6_src;
sys/netpfil/pf/pf.c
11256
pd->dst = (struct pf_addr *)&h->ip6_dst;
sys/netpfil/pf/pf.c
11258
pd->off = 0;
sys/netpfil/pf/pf.c
11260
if (pf_walk_header6(pd, h, reason) != PF_PASS) {
sys/netpfil/pf/pf.c
11265
if (m_tag_find(pd->m, PACKET_TAG_PF_REASSEMBLED, NULL) != NULL) {
sys/netpfil/pf/pf.c
11270
pd->virtual_proto = pd->proto;
sys/netpfil/pf/pf.c
11271
MPASS(pd->fragoff == 0);
sys/netpfil/pf/pf.c
11274
if (pd->fragoff != 0)
sys/netpfil/pf/pf.c
11275
pd->virtual_proto = PF_VPROTO_FRAGMENT;
sys/netpfil/pf/pf.c
11284
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
11286
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
11288
if (!pf_pull_hdr(pd->m, pd->off, th, sizeof(*th),
sys/netpfil/pf/pf.c
11294
pd->hdrlen = sizeof(*th);
sys/netpfil/pf/pf.c
11295
pd->p_len = pd->tot_len - pd->off - (th->th_off << 2);
sys/netpfil/pf/pf.c
11296
pd->sport = &th->th_sport;
sys/netpfil/pf/pf.c
11297
pd->dport = &th->th_dport;
sys/netpfil/pf/pf.c
11298
pd->pcksum = &th->th_sum;
sys/netpfil/pf/pf.c
11302
struct udphdr *uh = &pd->hdr.udp;
sys/netpfil/pf/pf.c
11304
if (!pf_pull_hdr(pd->m, pd->off, uh, sizeof(*uh),
sys/netpfil/pf/pf.c
11310
pd->hdrlen = sizeof(*uh);
sys/netpfil/pf/pf.c
11312
ntohs(uh->uh_ulen) > pd->m->m_pkthdr.len - pd->off ||
sys/netpfil/pf/pf.c
11318
pd->sport = &uh->uh_sport;
sys/netpfil/pf/pf.c
11319
pd->dport = &uh->uh_dport;
sys/netpfil/pf/pf.c
11320
pd->pcksum = &uh->uh_sum;
sys/netpfil/pf/pf.c
11324
if (!pf_pull_hdr(pd->m, pd->off, &pd->hdr.sctp, sizeof(pd->hdr.sctp),
sys/netpfil/pf/pf.c
11330
pd->hdrlen = sizeof(pd->hdr.sctp);
sys/netpfil/pf/pf.c
11331
pd->p_len = pd->tot_len - pd->off;
sys/netpfil/pf/pf.c
11333
pd->sport = &pd->hdr.sctp.src_port;
sys/netpfil/pf/pf.c
11334
pd->dport = &pd->hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
11335
if (pd->hdr.sctp.src_port == 0 || pd->hdr.sctp.dest_port == 0) {
sys/netpfil/pf/pf.c
11348
pd->pcksum = &pd->sctp_dummy_sum;
sys/netpfil/pf/pf.c
11350
if (pf_scan_sctp(pd) != PF_PASS) {
sys/netpfil/pf/pf.c
11358
if (!pf_pull_hdr(pd->m, pd->off, &pd->hdr.icmp, ICMP_MINLEN,
sys/netpfil/pf/pf.c
11364
pd->pcksum = &pd->hdr.icmp.icmp_cksum;
sys/netpfil/pf/pf.c
11365
pd->hdrlen = ICMP_MINLEN;
sys/netpfil/pf/pf.c
11372
if (!pf_pull_hdr(pd->m, pd->off, &pd->hdr.icmp6, icmp_hlen,
sys/netpfil/pf/pf.c
11379
switch (pd->hdr.icmp6.icmp6_type) {
sys/netpfil/pf/pf.c
11391
if (pd->ttl != 255) {
sys/netpfil/pf/pf.c
11398
!pf_pull_hdr(pd->m, pd->off, &pd->hdr.icmp6, icmp_hlen,
sys/netpfil/pf/pf.c
11404
pd->hdrlen = icmp_hlen;
sys/netpfil/pf/pf.c
11405
pd->pcksum = &pd->hdr.icmp6.icmp6_cksum;
sys/netpfil/pf/pf.c
11414
pd->pcksum = &pd->sctp_dummy_sum;
sys/netpfil/pf/pf.c
11418
if (pd->sport)
sys/netpfil/pf/pf.c
11419
pd->osport = pd->nsport = *pd->sport;
sys/netpfil/pf/pf.c
11420
if (pd->dport)
sys/netpfil/pf/pf.c
11421
pd->odport = pd->ndport = *pd->dport;
sys/netpfil/pf/pf.c
11423
MPASS(pd->pcksum != NULL);
sys/netpfil/pf/pf.c
11429
pf_rule_counters_inc(struct pf_pdesc *pd, struct pf_krule *r, int dir_out,
sys/netpfil/pf/pf.c
11434
pf_counter_u64_add_protected(&(r->bytes[dir_out]), pd->tot_len);
sys/netpfil/pf/pf.c
11439
pd->tot_len, dir_out, op_pass, r->src.neg);
sys/netpfil/pf/pf.c
11442
pd->tot_len, dir_out, op_pass, r->dst.neg);
sys/netpfil/pf/pf.c
11446
pf_counters_inc(int action, struct pf_pdesc *pd, struct pf_kstate *s,
sys/netpfil/pf/pf.c
11452
struct pf_addr *src_host = pd->src;
sys/netpfil/pf/pf.c
11453
struct pf_addr *dst_host = pd->dst;
sys/netpfil/pf/pf.c
11455
int dir_out = (pd->dir == PF_OUT);
sys/netpfil/pf/pf.c
11459
sa_family_t af = pd->af;
sys/netpfil/pf/pf.c
11470
af = pd->naf;
sys/netpfil/pf/pf.c
11474
&pd->kif->pfik_bytes[af == AF_INET6][dir_out][!op_pass],
sys/netpfil/pf/pf.c
11475
pd->tot_len);
sys/netpfil/pf/pf.c
11477
&pd->kif->pfik_packets[af == AF_INET6][dir_out][!op_pass],
sys/netpfil/pf/pf.c
11501
dir_out = (pd->naf == s->rule->naf);
sys/netpfil/pf/pf.c
11504
s_dir_rev = (pd->naf == s->rule->af);
sys/netpfil/pf/pf.c
11506
dir_out = (pd->dir == PF_OUT);
sys/netpfil/pf/pf.c
11509
s_dir_rev = (pd->dir != s->direction);
sys/netpfil/pf/pf.c
11519
s->bytes[s_dir_rev] += pd->tot_len;
sys/netpfil/pf/pf.c
11533
pd->tot_len);
sys/netpfil/pf/pf.c
11548
pf_rule_counters_inc(pd, s->nat_rule, dir_out,
sys/netpfil/pf/pf.c
11560
pf_rule_counters_inc(pd, ri->r, dir_out, op_r_pass, af,
sys/netpfil/pf/pf.c
11576
pf_rule_counters_inc(pd, a, dir_out, op_r_pass, af,
sys/netpfil/pf/pf.c
11581
pf_rule_counters_inc(pd, r, dir_out, op_r_pass, af,
sys/netpfil/pf/pf.c
11589
pf_log_matches(struct pf_pdesc *pd, struct pf_krule *rm,
sys/netpfil/pf/pf.c
11602
ruleset, pd, 1, ri->r);
sys/netpfil/pf/pf.c
11618
struct pf_pdesc pd;
sys/netpfil/pf/pf.c
11653
pf_init_pdesc(&pd, *m0);
sys/netpfil/pf/pf.c
11656
if (pd.pf_mtag != NULL && (pd.pf_mtag->flags & PF_MTAG_FLAG_ROUTE_TO)) {
sys/netpfil/pf/pf.c
11657
pd.pf_mtag->flags &= ~PF_MTAG_FLAG_ROUTE_TO;
sys/netpfil/pf/pf.c
11659
ifp = ifnet_byindexgen(pd.pf_mtag->if_index,
sys/netpfil/pf/pf.c
11660
pd.pf_mtag->if_idxgen);
sys/netpfil/pf/pf.c
11666
(ifp->if_output)(ifp, *m0, sintosa(&pd.pf_mtag->dst), NULL);
sys/netpfil/pf/pf.c
11671
if (ip_dn_io_ptr != NULL && pd.pf_mtag != NULL &&
sys/netpfil/pf/pf.c
11672
pd.pf_mtag->flags & PF_MTAG_FLAG_DUMMYNET) {
sys/netpfil/pf/pf.c
11679
pf_dummynet_flag_remove(pd.m, pd.pf_mtag);
sys/netpfil/pf/pf.c
11686
if (pf_setup_pdesc(af, dir, &pd, m0, &action, &reason,
sys/netpfil/pf/pf.c
11689
pd.act.log |= PF_LOG_FORCE;
sys/netpfil/pf/pf.c
11695
pd.df && (*m0)->m_pkthdr.len > ifp->if_mtu) {
sys/netpfil/pf/pf.c
11719
((mtag = m_tag_locate(pd.m, MTAG_PF_DIVERT, 0, NULL)) != NULL)) {
sys/netpfil/pf/pf.c
11723
if (pd.pf_mtag == NULL &&
sys/netpfil/pf/pf.c
11724
((pd.pf_mtag = pf_get_mtag(pd.m)) == NULL)) {
sys/netpfil/pf/pf.c
11728
pd.pf_mtag->flags |= PF_MTAG_FLAG_PACKET_LOOPED;
sys/netpfil/pf/pf.c
11730
if (pd.pf_mtag && pd.pf_mtag->flags & PF_MTAG_FLAG_FASTFWD_OURS_PRESENT) {
sys/netpfil/pf/pf.c
11731
pd.m->m_flags |= M_FASTFWD_OURS;
sys/netpfil/pf/pf.c
11732
pd.pf_mtag->flags &= ~PF_MTAG_FLAG_FASTFWD_OURS_PRESENT;
sys/netpfil/pf/pf.c
11734
m_tag_delete(pd.m, mtag);
sys/netpfil/pf/pf.c
11736
mtag = m_tag_locate(pd.m, MTAG_IPFW_RULE, 0, NULL);
sys/netpfil/pf/pf.c
11738
m_tag_delete(pd.m, mtag);
sys/netpfil/pf/pf.c
11741
switch (pd.virtual_proto) {
sys/netpfil/pf/pf.c
11750
action = pf_test_rule(&r, &s, &pd, &a,
sys/netpfil/pf/pf.c
11758
if ((tcp_get_flags(&pd.hdr.tcp) & (TH_SYN|TH_ACK|TH_RST)) == TH_SYN &&
sys/netpfil/pf/pf.c
11759
pd.dir == PF_IN && pf_synflood_check(&pd)) {
sys/netpfil/pf/pf.c
11760
pf_syncookie_send(&pd, &reason);
sys/netpfil/pf/pf.c
11765
if ((tcp_get_flags(&pd.hdr.tcp) & TH_ACK) && pd.p_len == 0)
sys/netpfil/pf/pf.c
11767
action = pf_normalize_tcp(&pd);
sys/netpfil/pf/pf.c
11770
action = pf_test_state(&s, &pd, &reason);
sys/netpfil/pf/pf.c
11781
if ((tcp_get_flags(&pd.hdr.tcp) & (TH_SYN|TH_ACK|TH_RST)) ==
sys/netpfil/pf/pf.c
11782
TH_ACK && pf_syncookie_validate(&pd) &&
sys/netpfil/pf/pf.c
11783
pd.dir == PF_IN) {
sys/netpfil/pf/pf.c
11786
msyn = pf_syncookie_recreate_syn(&pd, &reason);
sys/netpfil/pf/pf.c
11793
&pd.act);
sys/netpfil/pf/pf.c
11798
action = pf_test_state(&s, &pd, &reason);
sys/netpfil/pf/pf.c
11804
s->src.seqhi = ntohl(pd.hdr.tcp.th_ack) - 1;
sys/netpfil/pf/pf.c
11805
s->src.seqlo = ntohl(pd.hdr.tcp.th_seq) - 1;
sys/netpfil/pf/pf.c
11807
action = pf_synproxy(&pd, s, &reason);
sys/netpfil/pf/pf.c
11810
action = pf_test_rule(&r, &s, &pd,
sys/netpfil/pf/pf.c
11818
action = pf_normalize_sctp(&pd);
sys/netpfil/pf/pf.c
11824
action = pf_test_state(&s, &pd, &reason);
sys/netpfil/pf/pf.c
11834
&pd, &a, &ruleset, &reason, inp, &match_rules);
sys/netpfil/pf/pf.c
11840
if (pd.virtual_proto == IPPROTO_ICMP && af != AF_INET) {
sys/netpfil/pf/pf.c
11847
if (pd.virtual_proto == IPPROTO_ICMPV6 && af != AF_INET6) {
sys/netpfil/pf/pf.c
11854
action = pf_test_state_icmp(&s, &pd, &reason);
sys/netpfil/pf/pf.c
11863
action = pf_test_rule(&r, &s, &pd,
sys/netpfil/pf/pf.c
11874
if (pd.m == NULL) {
sys/netpfil/pf/pf.c
11880
memcpy(&pd.act, &s->act, sizeof(s->act));
sys/netpfil/pf/pf.c
11882
if (action == PF_PASS && pd.badopts != 0 && !pd.act.allow_opts) {
sys/netpfil/pf/pf.c
11885
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
11890
if (pd.act.max_pkt_size && pd.act.max_pkt_size &&
sys/netpfil/pf/pf.c
11891
pd.tot_len > pd.act.max_pkt_size) {
sys/netpfil/pf/pf.c
11894
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
11900
uint8_t log = pd.act.log;
sys/netpfil/pf/pf.c
11901
memcpy(&pd.act, &s->act, sizeof(struct pf_rule_actions));
sys/netpfil/pf/pf.c
11902
pd.act.log |= log;
sys/netpfil/pf/pf.c
11908
if (tag > 0 && pf_tag_packet(&pd, tag)) {
sys/netpfil/pf/pf.c
11913
pf_scrub(&pd);
sys/netpfil/pf/pf.c
11914
if (pd.proto == IPPROTO_TCP && pd.act.max_mss)
sys/netpfil/pf/pf.c
11915
pf_normalize_mss(&pd);
sys/netpfil/pf/pf.c
11917
if (pd.act.rtableid >= 0)
sys/netpfil/pf/pf.c
11918
M_SETFIB(pd.m, pd.act.rtableid);
sys/netpfil/pf/pf.c
11920
if (pd.act.flags & PFSTATE_SETPRIO) {
sys/netpfil/pf/pf.c
11921
if (pd.tos & IPTOS_LOWDELAY)
sys/netpfil/pf/pf.c
11923
if (vlan_set_pcp(pd.m, pd.act.set_prio[use_2nd_queue])) {
sys/netpfil/pf/pf.c
11926
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
11933
if (action == PF_PASS && pd.act.qid) {
sys/netpfil/pf/pf.c
11934
if (pd.pf_mtag == NULL &&
sys/netpfil/pf/pf.c
11935
((pd.pf_mtag = pf_get_mtag(pd.m)) == NULL)) {
sys/netpfil/pf/pf.c
11940
pd.pf_mtag->qid_hash = pf_state_hash(s);
sys/netpfil/pf/pf.c
11941
if (use_2nd_queue || (pd.tos & IPTOS_LOWDELAY))
sys/netpfil/pf/pf.c
11942
pd.pf_mtag->qid = pd.act.pqid;
sys/netpfil/pf/pf.c
11944
pd.pf_mtag->qid = pd.act.qid;
sys/netpfil/pf/pf.c
11946
pd.pf_mtag->hdr = mtod(pd.m, void *);
sys/netpfil/pf/pf.c
11956
if (dir == PF_IN && action == PF_PASS && (pd.proto == IPPROTO_TCP ||
sys/netpfil/pf/pf.c
11957
pd.proto == IPPROTO_UDP) && s != NULL && s->nat_rule != NULL &&
sys/netpfil/pf/pf.c
11960
pf_is_loopback(af, pd.dst))
sys/netpfil/pf/pf.c
11961
pd.m->m_flags |= M_SKIP_FIREWALL;
sys/netpfil/pf/pf.c
11963
if (action == PF_PASS && r->divert.port && !PACKET_LOOPED(&pd)) {
sys/netpfil/pf/pf.c
11973
pf_counters_inc(action, &pd, s, r, a, &match_rules);
sys/netpfil/pf/pf.c
11978
m_tag_prepend(pd.m, mtag);
sys/netpfil/pf/pf.c
11979
if (pd.m->m_flags & M_FASTFWD_OURS) {
sys/netpfil/pf/pf.c
11980
if (pd.pf_mtag == NULL &&
sys/netpfil/pf/pf.c
11981
((pd.pf_mtag = pf_get_mtag(pd.m)) == NULL)) {
sys/netpfil/pf/pf.c
11984
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
11988
pd.pf_mtag->flags |=
sys/netpfil/pf/pf.c
11990
pd.m->m_flags &= ~M_FASTFWD_OURS;
sys/netpfil/pf/pf.c
12000
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
12006
pd.act.log = PF_LOG_FORCE;
sys/netpfil/pf/pf.c
12013
if (pd.pf_mtag)
sys/netpfil/pf/pf.c
12014
pd.pf_mtag->flags &= ~PF_MTAG_FLAG_PACKET_LOOPED;
sys/netpfil/pf/pf.c
12016
if (pd.act.log) {
sys/netpfil/pf/pf.c
12025
if (pd.act.log & PF_LOG_FORCE || lr->log & PF_LOG_ALL)
sys/netpfil/pf/pf.c
12027
ruleset, &pd, (s == NULL), NULL);
sys/netpfil/pf/pf.c
12032
reason, ri->r, a, ruleset, &pd, 0, NULL);
sys/netpfil/pf/pf.c
12036
pf_counters_inc(action, &pd, s, r, a, &match_rules);
sys/netpfil/pf/pf.c
12050
if (pf_translate_af(&pd, r)) {
sys/netpfil/pf/pf.c
12051
*m0 = pd.m;
sys/netpfil/pf/pf.c
12056
if (pd.naf == AF_INET) {
sys/netpfil/pf/pf.c
12057
action = pf_route(r, kif->pfik_ifp, s, &pd,
sys/netpfil/pf/pf.c
12062
if (pd.naf == AF_INET6) {
sys/netpfil/pf/pf.c
12063
action = pf_route6(r, kif->pfik_ifp, s, &pd,
sys/netpfil/pf/pf.c
12067
*m0 = pd.m;
sys/netpfil/pf/pf.c
12071
if (pd.act.rt) {
sys/netpfil/pf/pf.c
12076
action = pf_route(r, kif->pfik_ifp, s, &pd,
sys/netpfil/pf/pf.c
12083
action = pf_route6(r, kif->pfik_ifp, s, &pd,
sys/netpfil/pf/pf.c
12088
*m0 = pd.m;
sys/netpfil/pf/pf.c
12091
if (pf_dummynet(&pd, s, r, m0) != 0) {
sys/netpfil/pf/pf.c
12116
(mtag = m_tag_find(pd.m, PACKET_TAG_PF_REASSEMBLED, NULL)) != NULL)
sys/netpfil/pf/pf.c
12120
pf_sctp_multihome_delayed(&pd, kif, s, action);
sys/netpfil/pf/pf.c
1951
pf_state_key_addr_setup(struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
1954
struct pf_addr *saddr = pd->src;
sys/netpfil/pf/pf.c
1955
struct pf_addr *daddr = pd->dst;
sys/netpfil/pf/pf.c
1960
if (pd->af == AF_INET || pd->proto != IPPROTO_ICMPV6)
sys/netpfil/pf/pf.c
1963
switch (pd->hdr.icmp6.icmp6_type) {
sys/netpfil/pf/pf.c
1967
if (!pf_pull_hdr(pd->m, pd->off, &nd, sizeof(nd), NULL,
sys/netpfil/pf/pf.c
1968
pd->af))
sys/netpfil/pf/pf.c
1976
if (!pf_pull_hdr(pd->m, pd->off, &nd, sizeof(nd), NULL,
sys/netpfil/pf/pf.c
1977
pd->af))
sys/netpfil/pf/pf.c
1981
if (IN6_IS_ADDR_MULTICAST(&pd->dst->v6)) {
sys/netpfil/pf/pf.c
1982
key->addr[pd->didx].addr32[0] = 0;
sys/netpfil/pf/pf.c
1983
key->addr[pd->didx].addr32[1] = 0;
sys/netpfil/pf/pf.c
1984
key->addr[pd->didx].addr32[2] = 0;
sys/netpfil/pf/pf.c
1985
key->addr[pd->didx].addr32[3] = 0;
sys/netpfil/pf/pf.c
1991
key->addr[pd->sidx].addr32[0] = IPV6_ADDR_INT32_MLL;
sys/netpfil/pf/pf.c
1992
key->addr[pd->sidx].addr32[1] = 0;
sys/netpfil/pf/pf.c
1993
key->addr[pd->sidx].addr32[2] = 0;
sys/netpfil/pf/pf.c
1994
key->addr[pd->sidx].addr32[3] = IPV6_ADDR_INT32_ONE;
sys/netpfil/pf/pf.c
2001
pf_addrcpy(&key->addr[pd->sidx], saddr, pd->af);
sys/netpfil/pf/pf.c
2003
pf_addrcpy(&key->addr[pd->didx], daddr, pd->af);
sys/netpfil/pf/pf.c
2009
pf_state_key_setup(struct pf_pdesc *pd, u_int16_t sport, u_int16_t dport,
sys/netpfil/pf/pf.c
2016
if (pf_state_key_addr_setup(pd, (struct pf_state_key_cmp *)*sk,
sys/netpfil/pf/pf.c
2023
(*sk)->port[pd->sidx] = sport;
sys/netpfil/pf/pf.c
2024
(*sk)->port[pd->didx] = dport;
sys/netpfil/pf/pf.c
2025
(*sk)->proto = pd->proto;
sys/netpfil/pf/pf.c
2026
(*sk)->af = pd->af;
sys/netpfil/pf/pf.c
2035
if (pd->af != pd->naf) {
sys/netpfil/pf/pf.c
2036
(*sk)->port[pd->sidx] = pd->osport;
sys/netpfil/pf/pf.c
2037
(*sk)->port[pd->didx] = pd->odport;
sys/netpfil/pf/pf.c
2039
(*nk)->af = pd->naf;
sys/netpfil/pf/pf.c
2047
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
2048
pf_addrcpy(&(*nk)->addr[pd->didx], &pd->nsaddr,
sys/netpfil/pf/pf.c
2049
pd->naf);
sys/netpfil/pf/pf.c
2050
pf_addrcpy(&(*nk)->addr[pd->sidx], &pd->ndaddr,
sys/netpfil/pf/pf.c
2051
pd->naf);
sys/netpfil/pf/pf.c
2052
(*nk)->port[pd->didx] = pd->nsport;
sys/netpfil/pf/pf.c
2053
(*nk)->port[pd->sidx] = pd->ndport;
sys/netpfil/pf/pf.c
2055
pf_addrcpy(&(*nk)->addr[pd->sidx], &pd->nsaddr,
sys/netpfil/pf/pf.c
2056
pd->naf);
sys/netpfil/pf/pf.c
2057
pf_addrcpy(&(*nk)->addr[pd->didx], &pd->ndaddr,
sys/netpfil/pf/pf.c
2058
pd->naf);
sys/netpfil/pf/pf.c
2059
(*nk)->port[pd->sidx] = pd->nsport;
sys/netpfil/pf/pf.c
2060
(*nk)->port[pd->didx] = pd->ndport;
sys/netpfil/pf/pf.c
2063
switch (pd->proto) {
sys/netpfil/pf/pf.c
2071
(*nk)->proto = pd->proto;
sys/netpfil/pf/pf.c
2181
pf_find_state(struct pf_pdesc *pd, const struct pf_state_key_cmp *key,
sys/netpfil/pf/pf.c
2204
idx = (pd->dir == PF_IN ? PF_SK_WIRE : PF_SK_STACK);
sys/netpfil/pf/pf.c
2208
if (s->kif == V_pfi_all || s->kif == pd->kif ||
sys/netpfil/pf/pf.c
2209
s->orig_kif == pd->kif) {
sys/netpfil/pf/pf.c
2219
SDT_PROBE5(pf, ip, state, lookup, pd->kif,
sys/netpfil/pf/pf.c
2220
key, (pd->dir), pd, *state);
sys/netpfil/pf/pf.c
2233
if (s->kif == V_pfi_all || s->kif == pd->kif ||
sys/netpfil/pf/pf.c
2234
s->orig_kif == pd->kif) {
sys/netpfil/pf/pf.c
2244
SDT_PROBE5(pf, ip, state, lookup, pd->kif,
sys/netpfil/pf/pf.c
2245
key, (pd->dir), pd, NULL);
sys/netpfil/pf/pf.c
2255
SDT_PROBE5(pf, ip, state, lookup, pd->kif, key, (pd->dir), pd, *state);
sys/netpfil/pf/pf.c
2263
if ((s)->rule->pktrate.limit && pd->dir == (s)->direction) {
sys/netpfil/pf/pf.c
2269
if (PACKET_LOOPED(pd)) {
sys/netpfil/pf/pf.c
2561
pf_icmp_mapping(struct pf_pdesc *pd, u_int8_t type,
sys/netpfil/pf/pf.c
2572
switch (pd->af) {
sys/netpfil/pf/pf.c
2581
*virtual_id = pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
2589
*virtual_id = pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
2597
*virtual_id = pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
2605
*virtual_id = pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
2665
*virtual_id = pd->hdr.icmp6.icmp6_id;
sys/netpfil/pf/pf.c
2727
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
3645
pf_patch_8(struct pf_pdesc *pd, u_int8_t *f, u_int8_t v, bool hi)
sys/netpfil/pf/pf.c
3655
if (! (pd->m->m_pkthdr.csum_flags & (CSUM_DELAY_DATA |
sys/netpfil/pf/pf.c
3657
*pd->pcksum = pf_cksum_fixup(*pd->pcksum, old, new,
sys/netpfil/pf/pf.c
3658
pd->proto == IPPROTO_UDP);
sys/netpfil/pf/pf.c
3667
pf_patch_16(struct pf_pdesc *pd, void *f, u_int16_t v, bool hi)
sys/netpfil/pf/pf.c
3673
rewrite += pf_patch_8(pd, fb++, *vb++, hi);
sys/netpfil/pf/pf.c
3674
rewrite += pf_patch_8(pd, fb++, *vb++, !hi);
sys/netpfil/pf/pf.c
3680
pf_patch_32(struct pf_pdesc *pd, void *f, u_int32_t v, bool hi)
sys/netpfil/pf/pf.c
3686
rewrite += pf_patch_8(pd, fb++, *vb++, hi);
sys/netpfil/pf/pf.c
3687
rewrite += pf_patch_8(pd, fb++, *vb++, !hi);
sys/netpfil/pf/pf.c
3688
rewrite += pf_patch_8(pd, fb++, *vb++, hi);
sys/netpfil/pf/pf.c
3689
rewrite += pf_patch_8(pd, fb++, *vb++, !hi);
sys/netpfil/pf/pf.c
3705
pf_change_ap(struct pf_pdesc *pd, struct pf_addr *a, u_int16_t *p,
sys/netpfil/pf/pf.c
3710
uint8_t u = pd->virtual_proto == IPPROTO_UDP;
sys/netpfil/pf/pf.c
3712
MPASS(pd->pcksum != NULL);
sys/netpfil/pf/pf.c
3713
if (pd->af == AF_INET) {
sys/netpfil/pf/pf.c
3714
MPASS(pd->ip_sum);
sys/netpfil/pf/pf.c
3717
pf_addrcpy(&ao, a, pd->af);
sys/netpfil/pf/pf.c
3718
if (pd->af == pd->naf)
sys/netpfil/pf/pf.c
3719
pf_addrcpy(a, an, pd->af);
sys/netpfil/pf/pf.c
3721
if (pd->m->m_pkthdr.csum_flags & (CSUM_DELAY_DATA | CSUM_DELAY_DATA_IPV6))
sys/netpfil/pf/pf.c
3722
*pd->pcksum = ~*pd->pcksum;
sys/netpfil/pf/pf.c
3729
switch (pd->af) {
sys/netpfil/pf/pf.c
3732
switch (pd->naf) {
sys/netpfil/pf/pf.c
3734
*pd->ip_sum = pf_cksum_fixup(pf_cksum_fixup(*pd->ip_sum,
sys/netpfil/pf/pf.c
3739
*pd->pcksum = pf_cksum_fixup(pf_cksum_fixup(*pd->pcksum,
sys/netpfil/pf/pf.c
3743
*pd->pcksum = pf_proto_cksum_fixup(pd->m, *pd->pcksum, po, pn, u);
sys/netpfil/pf/pf.c
3747
*pd->pcksum = pf_cksum_fixup(pf_cksum_fixup(pf_cksum_fixup(
sys/netpfil/pf/pf.c
3749
pf_cksum_fixup(pf_cksum_fixup(pf_cksum_fixup(*pd->pcksum,
sys/netpfil/pf/pf.c
3762
unhandled_af(pd->naf);
sys/netpfil/pf/pf.c
3768
switch (pd->naf) {
sys/netpfil/pf/pf.c
3771
*pd->pcksum = pf_cksum_fixup(pf_cksum_fixup(pf_cksum_fixup(
sys/netpfil/pf/pf.c
3773
pf_cksum_fixup(pf_cksum_fixup(pf_cksum_fixup(*pd->pcksum,
sys/netpfil/pf/pf.c
3786
*pd->pcksum = pf_cksum_fixup(pf_cksum_fixup(pf_cksum_fixup(
sys/netpfil/pf/pf.c
3788
pf_cksum_fixup(pf_cksum_fixup(*pd->pcksum,
sys/netpfil/pf/pf.c
3798
*pd->pcksum = pf_proto_cksum_fixup(pd->m, *pd->pcksum, po, pn, u);
sys/netpfil/pf/pf.c
3801
unhandled_af(pd->naf);
sys/netpfil/pf/pf.c
3806
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
3809
if (pd->m->m_pkthdr.csum_flags & (CSUM_DELAY_DATA |
sys/netpfil/pf/pf.c
3811
*pd->pcksum = ~*pd->pcksum;
sys/netpfil/pf/pf.c
3812
if (! *pd->pcksum)
sys/netpfil/pf/pf.c
3813
*pd->pcksum = 0xffff;
sys/netpfil/pf/pf.c
3954
pf_translate_af(struct pf_pdesc *pd, struct pf_krule *r)
sys/netpfil/pf/pf.c
3965
if (pd->ttl == 1) {
sys/netpfil/pf/pf.c
3967
if (pd->af == AF_INET) {
sys/netpfil/pf/pf.c
3968
pf_send_icmp(pd->m, ICMP_TIMXCEED,
sys/netpfil/pf/pf.c
3969
ICMP_TIMXCEED_INTRANS, 0, pd->af, r,
sys/netpfil/pf/pf.c
3970
pd->act.rtableid);
sys/netpfil/pf/pf.c
3972
pf_send_icmp(pd->m, ICMP6_TIME_EXCEEDED,
sys/netpfil/pf/pf.c
3973
ICMP6_TIME_EXCEED_TRANSIT, 0, pd->af, r,
sys/netpfil/pf/pf.c
3974
pd->act.rtableid);
sys/netpfil/pf/pf.c
3980
hlen = pd->naf == AF_INET ? sizeof(*ip4) : sizeof(*ip6);
sys/netpfil/pf/pf.c
3983
m_adj(pd->m, pd->off);
sys/netpfil/pf/pf.c
3986
M_PREPEND(pd->m, hlen, M_NOWAIT);
sys/netpfil/pf/pf.c
3987
if (pd->m == NULL)
sys/netpfil/pf/pf.c
3990
switch (pd->naf) {
sys/netpfil/pf/pf.c
3992
ip4 = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf.c
3996
ip4->ip_tos = pd->tos;
sys/netpfil/pf/pf.c
3997
ip4->ip_len = htons(hlen + (pd->tot_len - pd->off));
sys/netpfil/pf/pf.c
3999
ip4->ip_ttl = pd->ttl;
sys/netpfil/pf/pf.c
4000
ip4->ip_p = pd->proto;
sys/netpfil/pf/pf.c
4001
ip4->ip_src = pd->nsaddr.v4;
sys/netpfil/pf/pf.c
4002
ip4->ip_dst = pd->ndaddr.v4;
sys/netpfil/pf/pf.c
4003
pd->src = (struct pf_addr *)&ip4->ip_src;
sys/netpfil/pf/pf.c
4004
pd->dst = (struct pf_addr *)&ip4->ip_dst;
sys/netpfil/pf/pf.c
4005
pd->off = sizeof(struct ip);
sys/netpfil/pf/pf.c
4006
if (pd->m->m_pkthdr.csum_flags & CSUM_TCP_IPV6) {
sys/netpfil/pf/pf.c
4007
pd->m->m_pkthdr.csum_flags &= ~CSUM_TCP_IPV6;
sys/netpfil/pf/pf.c
4008
pd->m->m_pkthdr.csum_flags |= CSUM_TCP;
sys/netpfil/pf/pf.c
4010
if (pd->m->m_pkthdr.csum_flags & CSUM_UDP_IPV6) {
sys/netpfil/pf/pf.c
4011
pd->m->m_pkthdr.csum_flags &= ~CSUM_UDP_IPV6;
sys/netpfil/pf/pf.c
4012
pd->m->m_pkthdr.csum_flags |= CSUM_UDP;
sys/netpfil/pf/pf.c
4014
if (pd->m->m_pkthdr.csum_flags & CSUM_SCTP_IPV6) {
sys/netpfil/pf/pf.c
4015
pd->m->m_pkthdr.csum_flags &= ~CSUM_SCTP_IPV6;
sys/netpfil/pf/pf.c
4016
pd->m->m_pkthdr.csum_flags |= CSUM_SCTP;
sys/netpfil/pf/pf.c
4020
ip6 = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf.c
4023
ip6->ip6_flow |= htonl((u_int32_t)pd->tos << 20);
sys/netpfil/pf/pf.c
4024
ip6->ip6_plen = htons(pd->tot_len - pd->off);
sys/netpfil/pf/pf.c
4025
ip6->ip6_nxt = pd->proto;
sys/netpfil/pf/pf.c
4026
if (!pd->ttl || pd->ttl > IPV6_DEFHLIM)
sys/netpfil/pf/pf.c
4029
ip6->ip6_hlim = pd->ttl;
sys/netpfil/pf/pf.c
4030
ip6->ip6_src = pd->nsaddr.v6;
sys/netpfil/pf/pf.c
4031
ip6->ip6_dst = pd->ndaddr.v6;
sys/netpfil/pf/pf.c
4032
pd->src = (struct pf_addr *)&ip6->ip6_src;
sys/netpfil/pf/pf.c
4033
pd->dst = (struct pf_addr *)&ip6->ip6_dst;
sys/netpfil/pf/pf.c
4034
pd->off = sizeof(struct ip6_hdr);
sys/netpfil/pf/pf.c
4035
if (pd->m->m_pkthdr.csum_flags & CSUM_TCP) {
sys/netpfil/pf/pf.c
4036
pd->m->m_pkthdr.csum_flags &= ~CSUM_TCP;
sys/netpfil/pf/pf.c
4037
pd->m->m_pkthdr.csum_flags |= CSUM_TCP_IPV6;
sys/netpfil/pf/pf.c
4039
if (pd->m->m_pkthdr.csum_flags & CSUM_UDP) {
sys/netpfil/pf/pf.c
4040
pd->m->m_pkthdr.csum_flags &= ~CSUM_UDP;
sys/netpfil/pf/pf.c
4041
pd->m->m_pkthdr.csum_flags |= CSUM_UDP_IPV6;
sys/netpfil/pf/pf.c
4043
if (pd->m->m_pkthdr.csum_flags & CSUM_SCTP) {
sys/netpfil/pf/pf.c
4044
pd->m->m_pkthdr.csum_flags &= ~CSUM_SCTP;
sys/netpfil/pf/pf.c
4045
pd->m->m_pkthdr.csum_flags |= CSUM_SCTP_IPV6;
sys/netpfil/pf/pf.c
4053
mtag = m_tag_find(pd->m, PACKET_TAG_PF_REASSEMBLED, NULL);
sys/netpfil/pf/pf.c
4066
if (pd->proto == IPPROTO_ICMP || pd->proto == IPPROTO_ICMPV6) {
sys/netpfil/pf/pf.c
4068
if ((mp = m_pulldown(pd->m, hlen, sizeof(*icmp), &off)) ==
sys/netpfil/pf/pf.c
4070
pd->m = NULL;
sys/netpfil/pf/pf.c
4075
icmp->icmp6_cksum = pd->naf == AF_INET ?
sys/netpfil/pf/pf.c
4076
in4_cksum(pd->m, 0, hlen, ntohs(ip4->ip_len) - hlen) :
sys/netpfil/pf/pf.c
4077
in6_cksum(pd->m, IPPROTO_ICMPV6, hlen,
sys/netpfil/pf/pf.c
4086
pf_change_icmp_af(struct mbuf *m, int off, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
4158
pd->tot_len += hlen - olen;
sys/netpfil/pf/pf.c
4405
pf_modulate_sack(struct pf_pdesc *pd, struct tcphdr *th,
sys/netpfil/pf/pf.c
4413
olen = (pd->hdr.tcp.th_off << 2) - sizeof(struct tcphdr);
sys/netpfil/pf/pf.c
4414
optsoff = pd->off + sizeof(struct tcphdr);
sys/netpfil/pf/pf.c
4417
!pf_pull_hdr(pd->m, optsoff, opts, olen, NULL, pd->af))
sys/netpfil/pf/pf.c
4429
pf_patch_32(pd, &sack.start,
sys/netpfil/pf/pf.c
4432
pf_patch_32(pd, &sack.end,
sys/netpfil/pf/pf.c
4442
m_copyback(pd->m, optsoff, olen, (caddr_t)opts);
sys/netpfil/pf/pf.c
4602
pf_send_sctp_abort(sa_family_t af, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
4617
MPASS(af == pd->af);
sys/netpfil/pf/pf.c
4642
h->ip_src = pd->dst->v4;
sys/netpfil/pf/pf.c
4643
h->ip_dst = pd->src->v4;
sys/netpfil/pf/pf.c
4659
memcpy(&h6->ip6_src, &pd->dst->v6, sizeof(struct in6_addr));
sys/netpfil/pf/pf.c
4660
memcpy(&h6->ip6_dst, &pd->src->v6, sizeof(struct in6_addr));
sys/netpfil/pf/pf.c
4672
hdr->src_port = pd->hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
4673
hdr->dest_port = pd->hdr.sctp.src_port;
sys/netpfil/pf/pf.c
4674
hdr->v_tag = pd->sctp_initiate_tag;
sys/netpfil/pf/pf.c
4761
pf_undo_nat(struct pf_krule *nr, struct pf_pdesc *pd, uint16_t bip_sum)
sys/netpfil/pf/pf.c
4765
pf_addrcpy(pd->src, &pd->osrc, pd->af);
sys/netpfil/pf/pf.c
4766
pf_addrcpy(pd->dst, &pd->odst, pd->af);
sys/netpfil/pf/pf.c
4767
if (pd->sport)
sys/netpfil/pf/pf.c
4768
*pd->sport = pd->osport;
sys/netpfil/pf/pf.c
4769
if (pd->dport)
sys/netpfil/pf/pf.c
4770
*pd->dport = pd->odport;
sys/netpfil/pf/pf.c
4771
if (pd->ip_sum)
sys/netpfil/pf/pf.c
4772
*pd->ip_sum = bip_sum;
sys/netpfil/pf/pf.c
4773
m_copyback(pd->m, pd->off, pd->hdrlen, pd->hdr.any);
sys/netpfil/pf/pf.c
4778
pf_return(struct pf_krule *r, struct pf_krule *nr, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
4782
pf_undo_nat(nr, pd, bip_sum);
sys/netpfil/pf/pf.c
4784
if (pd->proto == IPPROTO_TCP &&
sys/netpfil/pf/pf.c
4788
u_int32_t ack = ntohl(th->th_seq) + pd->p_len;
sys/netpfil/pf/pf.c
4790
if (pf_check_proto_cksum(pd->m, pd->off, pd->tot_len - pd->off,
sys/netpfil/pf/pf.c
4791
IPPROTO_TCP, pd->af))
sys/netpfil/pf/pf.c
4798
pf_send_tcp(r, pd->af, pd->dst,
sys/netpfil/pf/pf.c
4799
pd->src, th->th_dport, th->th_sport,
sys/netpfil/pf/pf.c
4804
} else if (pd->proto == IPPROTO_SCTP &&
sys/netpfil/pf/pf.c
4806
pf_send_sctp_abort(pd->af, pd, r->return_ttl, rtableid);
sys/netpfil/pf/pf.c
4807
} else if (pd->proto != IPPROTO_ICMP && pd->af == AF_INET &&
sys/netpfil/pf/pf.c
4809
pf_send_icmp(pd->m, r->return_icmp >> 8,
sys/netpfil/pf/pf.c
4810
r->return_icmp & 255, 0, pd->af, r, rtableid);
sys/netpfil/pf/pf.c
4811
else if (pd->proto != IPPROTO_ICMPV6 && pd->af == AF_INET6 &&
sys/netpfil/pf/pf.c
4813
pf_send_icmp(pd->m, r->return_icmp6 >> 8,
sys/netpfil/pf/pf.c
4814
r->return_icmp6 & 255, 0, pd->af, r, rtableid);
sys/netpfil/pf/pf.c
4852
pf_send_challenge_ack(struct pf_pdesc *pd, struct pf_kstate *s,
sys/netpfil/pf/pf.c
4865
pf_send_tcp(s->rule, pd->af, pd->dst, pd->src,
sys/netpfil/pf/pf.c
4866
pd->hdr.tcp.th_dport, pd->hdr.tcp.th_sport, dst->seqlo,
sys/netpfil/pf/pf.c
5086
pf_tag_packet(struct pf_pdesc *pd, int tag)
sys/netpfil/pf/pf.c
5091
if (pd->pf_mtag == NULL && ((pd->pf_mtag = pf_get_mtag(pd->m)) == NULL))
sys/netpfil/pf/pf.c
5094
pd->pf_mtag->tag = tag;
sys/netpfil/pf/pf.c
5369
pf_socket_lookup(struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
5376
pd->lookup.uid = -1;
sys/netpfil/pf/pf.c
5377
pd->lookup.gid = -1;
sys/netpfil/pf/pf.c
5379
switch (pd->proto) {
sys/netpfil/pf/pf.c
5381
sport = pd->hdr.tcp.th_sport;
sys/netpfil/pf/pf.c
5382
dport = pd->hdr.tcp.th_dport;
sys/netpfil/pf/pf.c
5386
sport = pd->hdr.udp.uh_sport;
sys/netpfil/pf/pf.c
5387
dport = pd->hdr.udp.uh_dport;
sys/netpfil/pf/pf.c
5393
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
5394
saddr = pd->src;
sys/netpfil/pf/pf.c
5395
daddr = pd->dst;
sys/netpfil/pf/pf.c
5402
saddr = pd->dst;
sys/netpfil/pf/pf.c
5403
daddr = pd->src;
sys/netpfil/pf/pf.c
5405
switch (pd->af) {
sys/netpfil/pf/pf.c
5409
dport, INPLOOKUP_RLOCKPCB, NULL, pd->m);
sys/netpfil/pf/pf.c
5413
INPLOOKUP_RLOCKPCB, NULL, pd->m);
sys/netpfil/pf/pf.c
5422
dport, INPLOOKUP_RLOCKPCB, NULL, pd->m);
sys/netpfil/pf/pf.c
5426
INPLOOKUP_RLOCKPCB, NULL, pd->m);
sys/netpfil/pf/pf.c
5433
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
5436
pd->lookup.uid = inp->inp_cred->cr_uid;
sys/netpfil/pf/pf.c
5437
pd->lookup.gid = inp->inp_cred->cr_gid;
sys/netpfil/pf/pf.c
5479
pf_get_wscale(struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
5485
olen = (pd->hdr.tcp.th_off << 2) - sizeof(struct tcphdr);
sys/netpfil/pf/pf.c
5486
if (olen < TCPOLEN_WINDOW || !pf_pull_hdr(pd->m,
sys/netpfil/pf/pf.c
5487
pd->off + sizeof(struct tcphdr), opts, olen, NULL, pd->af))
sys/netpfil/pf/pf.c
5504
pf_get_mss(struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
5510
olen = (pd->hdr.tcp.th_off << 2) - sizeof(struct tcphdr);
sys/netpfil/pf/pf.c
5511
if (olen < TCPOLEN_MAXSEG || !pf_pull_hdr(pd->m,
sys/netpfil/pf/pf.c
5512
pd->off + sizeof(struct tcphdr), opts, olen, NULL, pd->af))
sys/netpfil/pf/pf.c
5566
pf_tcp_iss(struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
5584
SHA512_Update(&ctx, &pd->hdr.tcp.th_sport, sizeof(u_short));
sys/netpfil/pf/pf.c
5585
SHA512_Update(&ctx, &pd->hdr.tcp.th_dport, sizeof(u_short));
sys/netpfil/pf/pf.c
5586
switch (pd->af) {
sys/netpfil/pf/pf.c
5588
SHA512_Update(&ctx, &pd->src->v6, sizeof(struct in6_addr));
sys/netpfil/pf/pf.c
5589
SHA512_Update(&ctx, &pd->dst->v6, sizeof(struct in6_addr));
sys/netpfil/pf/pf.c
5592
SHA512_Update(&ctx, &pd->src->v4, sizeof(struct in_addr));
sys/netpfil/pf/pf.c
5593
SHA512_Update(&ctx, &pd->dst->v4, sizeof(struct in_addr));
sys/netpfil/pf/pf.c
5944
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf.c
5958
pd->naf = r->naf;
sys/netpfil/pf/pf.c
5959
if (pf_get_transaddr_af(ctx->nr, pd) == -1) {
sys/netpfil/pf/pf.c
5997
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf.c
6007
if (ctx->pd->related_rule) {
sys/netpfil/pf/pf.c
6008
*ctx->rm = ctx->pd->related_rule;
sys/netpfil/pf/pf.c
6015
PF_TEST_ATTRIB(pfi_kkif_match(r->kif, pd->kif) == r->ifnot,
sys/netpfil/pf/pf.c
6017
PF_TEST_ATTRIB(r->direction && r->direction != pd->dir,
sys/netpfil/pf/pf.c
6019
PF_TEST_ATTRIB(r->af && r->af != pd->af,
sys/netpfil/pf/pf.c
6021
PF_TEST_ATTRIB(r->proto && r->proto != pd->proto,
sys/netpfil/pf/pf.c
6023
PF_TEST_ATTRIB(PF_MISMATCHAW(&r->src.addr, &pd->nsaddr, pd->naf,
sys/netpfil/pf/pf.c
6024
r->src.neg, pd->kif, M_GETFIB(pd->m)),
sys/netpfil/pf/pf.c
6026
PF_TEST_ATTRIB(PF_MISMATCHAW(&r->dst.addr, &pd->ndaddr, pd->af,
sys/netpfil/pf/pf.c
6027
r->dst.neg, NULL, M_GETFIB(pd->m)),
sys/netpfil/pf/pf.c
6029
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
6034
PF_TEST_ATTRIB((pd->proto == IPPROTO_TCP && r->flagset),
sys/netpfil/pf/pf.c
6053
r->src.port[0], r->src.port[1], pd->nsport),
sys/netpfil/pf/pf.c
6057
r->dst.port[0], r->dst.port[1], pd->ndport),
sys/netpfil/pf/pf.c
6060
PF_TEST_ATTRIB(r->uid.op && (pd->lookup.done || (pd->lookup.done =
sys/netpfil/pf/pf.c
6061
pf_socket_lookup(pd), 1)) &&
sys/netpfil/pf/pf.c
6063
pd->lookup.uid),
sys/netpfil/pf/pf.c
6066
PF_TEST_ATTRIB(r->gid.op && (pd->lookup.done || (pd->lookup.done =
sys/netpfil/pf/pf.c
6067
pf_socket_lookup(pd), 1)) &&
sys/netpfil/pf/pf.c
6069
pd->lookup.gid),
sys/netpfil/pf/pf.c
6086
PF_TEST_ATTRIB(r->tos && !(r->tos == pd->tos),
sys/netpfil/pf/pf.c
6089
!pf_match_ieee8021q_pcp(r->prio, pd->m),
sys/netpfil/pf/pf.c
6094
PF_TEST_ATTRIB(r->match_tag && !pf_match_tag(pd->m, r,
sys/netpfil/pf/pf.c
6095
&ctx->tag, pd->pf_mtag ? pd->pf_mtag->tag : 0),
sys/netpfil/pf/pf.c
6097
PF_TEST_ATTRIB((r->rcv_kif && pf_match_rcvif(pd->m, r) ==
sys/netpfil/pf/pf.c
6101
pd->virtual_proto != PF_VPROTO_FRAGMENT),
sys/netpfil/pf/pf.c
6104
(pd->virtual_proto != IPPROTO_TCP || !pf_osfp_match(
sys/netpfil/pf/pf.c
6105
pf_osfp_fingerprint(pd, ctx->th),
sys/netpfil/pf/pf.c
6180
pf_source_key(srlim, &key, ctx->pd->af,
sys/netpfil/pf/pf.c
6181
ctx->pd->src);
sys/netpfil/pf/pf.c
6310
pf_rule_to_actions(r, &pd->act);
sys/netpfil/pf/pf.c
6313
ctx->a, ruleset, pd, 1, NULL);
sys/netpfil/pf/pf.c
6339
if (pd->act.log & PF_LOG_MATCHES)
sys/netpfil/pf/pf.c
6340
pf_log_matches(pd, r, ctx->a, ruleset, ctx->match_rules);
sys/netpfil/pf/pf.c
6372
struct pf_pdesc *pd, struct pf_krule **am,
sys/netpfil/pf/pf.c
6388
ctx.pd = pd;
sys/netpfil/pf/pf.c
6392
ctx.th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
6396
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
6397
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
6401
pd->lookup.uid = inp->inp_cred->cr_uid;
sys/netpfil/pf/pf.c
6402
pd->lookup.gid = inp->inp_cred->cr_gid;
sys/netpfil/pf/pf.c
6403
pd->lookup.done = 1;
sys/netpfil/pf/pf.c
6406
if (pd->ip_sum)
sys/netpfil/pf/pf.c
6407
bip_sum = *pd->ip_sum;
sys/netpfil/pf/pf.c
6409
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
6412
pd->nsport = ctx.th->th_sport;
sys/netpfil/pf/pf.c
6413
pd->ndport = ctx.th->th_dport;
sys/netpfil/pf/pf.c
6416
bproto_sum = pd->hdr.udp.uh_sum;
sys/netpfil/pf/pf.c
6417
pd->nsport = pd->hdr.udp.uh_sport;
sys/netpfil/pf/pf.c
6418
pd->ndport = pd->hdr.udp.uh_dport;
sys/netpfil/pf/pf.c
6421
pd->nsport = pd->hdr.sctp.src_port;
sys/netpfil/pf/pf.c
6422
pd->ndport = pd->hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
6426
MPASS(pd->af == AF_INET);
sys/netpfil/pf/pf.c
6427
ctx.icmptype = pd->hdr.icmp.icmp_type;
sys/netpfil/pf/pf.c
6428
ctx.icmpcode = pd->hdr.icmp.icmp_code;
sys/netpfil/pf/pf.c
6429
ctx.state_icmp = pf_icmp_mapping(pd, ctx.icmptype,
sys/netpfil/pf/pf.c
6432
pd->nsport = ctx.virtual_id;
sys/netpfil/pf/pf.c
6433
pd->ndport = ctx.virtual_type;
sys/netpfil/pf/pf.c
6435
pd->nsport = ctx.virtual_type;
sys/netpfil/pf/pf.c
6436
pd->ndport = ctx.virtual_id;
sys/netpfil/pf/pf.c
6442
MPASS(pd->af == AF_INET6);
sys/netpfil/pf/pf.c
6443
ctx.icmptype = pd->hdr.icmp6.icmp6_type;
sys/netpfil/pf/pf.c
6444
ctx.icmpcode = pd->hdr.icmp6.icmp6_code;
sys/netpfil/pf/pf.c
6445
ctx.state_icmp = pf_icmp_mapping(pd, ctx.icmptype,
sys/netpfil/pf/pf.c
6448
pd->nsport = ctx.virtual_id;
sys/netpfil/pf/pf.c
6449
pd->ndport = ctx.virtual_type;
sys/netpfil/pf/pf.c
6451
pd->nsport = ctx.virtual_type;
sys/netpfil/pf/pf.c
6452
pd->ndport = ctx.virtual_id;
sys/netpfil/pf/pf.c
6458
pd->nsport = pd->ndport = 0;
sys/netpfil/pf/pf.c
6461
pd->osport = pd->nsport;
sys/netpfil/pf/pf.c
6462
pd->odport = pd->ndport;
sys/netpfil/pf/pf.c
6479
ruleset, pd, 1, NULL);
sys/netpfil/pf/pf.c
6504
pf_rule_to_actions(r, &pd->act);
sys/netpfil/pf/pf.c
6523
m_copyback(pd->m, pd->off, pd->hdrlen, pd->hdr.any);
sys/netpfil/pf/pf.c
6524
PFLOG_PACKET(r->action, ctx.reason, r, ctx.a, ruleset, pd, 1, NULL);
sys/netpfil/pf/pf.c
6526
if (pd->act.log & PF_LOG_MATCHES)
sys/netpfil/pf/pf.c
6527
pf_log_matches(pd, r, ctx.a, ruleset, ctx.match_rules);
sys/netpfil/pf/pf.c
6528
if (pd->virtual_proto != PF_VPROTO_FRAGMENT &&
sys/netpfil/pf/pf.c
6533
pf_return(r, ctx.nr, pd, ctx.th, bproto_sum,
sys/netpfil/pf/pf.c
6540
if (ctx.tag > 0 && pf_tag_packet(pd, ctx.tag)) {
sys/netpfil/pf/pf.c
6544
if (pd->act.rtableid >= 0)
sys/netpfil/pf/pf.c
6545
M_SETFIB(pd->m, pd->act.rtableid);
sys/netpfil/pf/pf.c
6554
pd->act.rt = r->rt;
sys/netpfil/pf/pf.c
6556
pd->act.rt_af = pd->af;
sys/netpfil/pf/pf.c
6558
pd->act.rt_af = pd->naf;
sys/netpfil/pf/pf.c
6559
if ((transerror = pf_map_addr_sn(pd->af, r, pd->src,
sys/netpfil/pf/pf.c
6560
&pd->act.rt_addr, &pd->act.rt_af, &pd->act.rt_kif, NULL,
sys/netpfil/pf/pf.c
6567
if (pd->virtual_proto != PF_VPROTO_FRAGMENT &&
sys/netpfil/pf/pf.c
6569
(pd->flags & PFDESC_TCP_NORM)))) {
sys/netpfil/pf/pf.c
6578
pd->act.log |= PF_LOG_FORCE;
sys/netpfil/pf/pf.c
6581
pf_return(r, ctx.nr, pd, ctx.th,
sys/netpfil/pf/pf.c
6583
pd->act.rtableid);
sys/netpfil/pf/pf.c
6588
if (pd->proto == IPPROTO_TCP &&
sys/netpfil/pf/pf.c
6589
r->keep_state == PF_STATE_SYNPROXY && pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
6590
action = pf_synproxy_ack(r, pd, sm, &ctx.act);
sys/netpfil/pf/pf.c
6595
nat64 = pd->af != pd->naf;
sys/netpfil/pf/pf.c
6600
ctx.sk = (*sm)->key[pd->dir == PF_IN ? PF_SK_STACK : PF_SK_WIRE];
sys/netpfil/pf/pf.c
6602
ctx.nk = (*sm)->key[pd->dir == PF_IN ? PF_SK_WIRE : PF_SK_STACK];
sys/netpfil/pf/pf.c
6604
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
6605
ret = pf_translate(pd, &ctx.sk->addr[pd->didx],
sys/netpfil/pf/pf.c
6606
ctx.sk->port[pd->didx], &ctx.sk->addr[pd->sidx],
sys/netpfil/pf/pf.c
6607
ctx.sk->port[pd->sidx], ctx.virtual_type,
sys/netpfil/pf/pf.c
6610
ret = pf_translate(pd, &ctx.sk->addr[pd->sidx],
sys/netpfil/pf/pf.c
6611
ctx.sk->port[pd->sidx], &ctx.sk->addr[pd->didx],
sys/netpfil/pf/pf.c
6612
ctx.sk->port[pd->didx], ctx.virtual_type,
sys/netpfil/pf/pf.c
6633
m_copyback(pd->m, pd->off, pd->hdrlen, pd->hdr.any);
sys/netpfil/pf/pf.c
6636
pd->dir == PF_OUT &&
sys/netpfil/pf/pf.c
6637
V_pfsync_defer_ptr != NULL && V_pfsync_defer_ptr(*sm, pd->m)) {
sys/netpfil/pf/pf.c
6664
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf.c
6677
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
6690
(sn_reason = pf_insert_src_node(sns, snhs, r, pd->src, pd->af,
sys/netpfil/pf/pf.c
6691
NULL, NULL, pd->af, PF_SN_LIMIT)) != 0) {
sys/netpfil/pf/pf.c
6698
(sn_reason = pf_insert_src_node(sns, snhs, r, pd->src,
sys/netpfil/pf/pf.c
6699
pd->af, &pd->act.rt_addr, pd->act.rt_kif, pd->act.rt_af,
sys/netpfil/pf/pf.c
6716
ctx->sk ? &(ctx->sk->addr[pd->sidx]) : pd->src, pd->af,
sys/netpfil/pf/pf.c
6717
ctx->nk ? &(ctx->nk->addr[1]) : &(pd->nsaddr), NULL,
sys/netpfil/pf/pf.c
6718
pd->naf, PF_SN_NAT)) != 0 ) {
sys/netpfil/pf/pf.c
6733
memcpy(&s->act, &pd->act, sizeof(struct pf_rule_actions));
sys/netpfil/pf/pf.c
6735
if (pd->act.allow_opts)
sys/netpfil/pf/pf.c
6739
if (pd->flags & PFDESC_TCP_NORM) /* Set by old-style scrub rules */
sys/netpfil/pf/pf.c
6745
s->act.log = pd->act.log & PF_LOG_ALL;
sys/netpfil/pf/pf.c
6747
s->state_flags |= pd->act.flags; /* Only needed for pfsync and state export */
sys/netpfil/pf/pf.c
6751
switch (pd->proto) {
sys/netpfil/pf/pf.c
6754
s->src.seqhi = s->src.seqlo + pd->p_len + 1;
sys/netpfil/pf/pf.c
6758
if ((s->src.seqdiff = pf_tcp_iss(pd) - s->src.seqlo) ==
sys/netpfil/pf/pf.c
6761
pf_change_proto_a(pd->m, &th->th_seq, &th->th_sum,
sys/netpfil/pf/pf.c
6768
s->src.wscale = pf_get_wscale(pd);
sys/netpfil/pf/pf.c
6811
if (pd->proto == IPPROTO_TCP) {
sys/netpfil/pf/pf.c
6813
pf_normalize_tcp_init(pd, th, &s->src)) {
sys/netpfil/pf/pf.c
6818
pf_normalize_tcp_stateful(pd, &ctx->reason, th, s,
sys/netpfil/pf/pf.c
6826
} else if (pd->proto == IPPROTO_SCTP) {
sys/netpfil/pf/pf.c
6827
if (pf_normalize_sctp_init(pd, &s->src, &s->dst))
sys/netpfil/pf/pf.c
6829
if (! (pd->sctp_flags & (PFDESC_SCTP_INIT | PFDESC_SCTP_ADD_IP)))
sys/netpfil/pf/pf.c
6832
s->direction = pd->dir;
sys/netpfil/pf/pf.c
6838
MPASS(pd->sport == NULL || (pd->osport == *pd->sport));
sys/netpfil/pf/pf.c
6839
MPASS(pd->dport == NULL || (pd->odport == *pd->dport));
sys/netpfil/pf/pf.c
6840
if (pf_state_key_setup(pd, pd->nsport, pd->ndport,
sys/netpfil/pf/pf.c
6889
pf_source_key(srlim, sr, ctx->pd->af, ctx->pd->src);
sys/netpfil/pf/pf.c
6941
if (pf_state_insert(BOUND_IFACE(s, pd), pd->kif,
sys/netpfil/pf/pf.c
6942
(pd->dir == PF_IN) ? ctx->sk : ctx->nk,
sys/netpfil/pf/pf.c
6943
(pd->dir == PF_IN) ? ctx->nk : ctx->sk, s)) {
sys/netpfil/pf/pf.c
6964
if (pd->proto == IPPROTO_TCP && (tcp_get_flags(th) & (TH_SYN|TH_ACK)) ==
sys/netpfil/pf/pf.c
6965
TH_SYN && r->keep_state == PF_STATE_SYNPROXY && pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
6967
pf_undo_nat(ctx->nr, pd, bip_sum);
sys/netpfil/pf/pf.c
6970
int rtid = M_GETFIB(pd->m);
sys/netpfil/pf/pf.c
6971
mss = pf_get_mss(pd);
sys/netpfil/pf/pf.c
6972
mss = pf_calc_mss(pd->src, pd->af, rtid, mss);
sys/netpfil/pf/pf.c
6973
mss = pf_calc_mss(pd->dst, pd->af, rtid, mss);
sys/netpfil/pf/pf.c
6975
pf_send_tcp(r, pd->af, pd->dst, pd->src, th->th_dport,
sys/netpfil/pf/pf.c
6978
pd->act.rtableid, &ctx->reason);
sys/netpfil/pf/pf.c
702
if ((pd->dir) == PF_OUT) \
sys/netpfil/pf/pf.c
7053
pf_translate(struct pf_pdesc *pd, struct pf_addr *saddr, u_int16_t sport,
sys/netpfil/pf/pf.c
7068
int afto = pd->af != pd->naf;
sys/netpfil/pf/pf.c
7072
switch (pd->proto) {
sys/netpfil/pf/pf.c
7076
if (afto || *pd->sport != sport) {
sys/netpfil/pf/pf.c
7077
pf_change_ap(pd, pd->src, pd->sport,
sys/netpfil/pf/pf.c
7081
if (afto || *pd->dport != dport) {
sys/netpfil/pf/pf.c
7082
pf_change_ap(pd, pd->dst, pd->dport,
sys/netpfil/pf/pf.c
709
#define PACKET_LOOPED(pd) ((pd)->pf_mtag && \
sys/netpfil/pf/pf.c
7091
if (pd->af != AF_INET)
sys/netpfil/pf/pf.c
7095
if (pf_translate_icmp_af(AF_INET6, &pd->hdr.icmp))
sys/netpfil/pf/pf.c
7097
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
710
(pd)->pf_mtag->flags & PF_MTAG_FLAG_PACKET_LOOPED)
sys/netpfil/pf/pf.c
7103
if (icmpid != pd->hdr.icmp.icmp_id) {
sys/netpfil/pf/pf.c
7104
pd->hdr.icmp.icmp_cksum = pf_cksum_fixup(
sys/netpfil/pf/pf.c
7105
pd->hdr.icmp.icmp_cksum,
sys/netpfil/pf/pf.c
7106
pd->hdr.icmp.icmp_id, icmpid, 0);
sys/netpfil/pf/pf.c
7107
pd->hdr.icmp.icmp_id = icmpid;
sys/netpfil/pf/pf.c
7118
if (pd->af != AF_INET6)
sys/netpfil/pf/pf.c
7123
if (pf_translate_icmp_af(AF_INET, &pd->hdr.icmp6))
sys/netpfil/pf/pf.c
7125
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
713
BOUND_IFACE(struct pf_kstate *st, struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
7141
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf.c
7143
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7149
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
715
struct pfi_kkif *k = pd->kif;
sys/netpfil/pf/pf.c
7151
if (PF_ANEQ(&pd->nsaddr, &nk->addr[pd->sidx], pd->af) ||
sys/netpfil/pf/pf.c
7152
nk->port[pd->sidx] != pd->nsport) {
sys/netpfil/pf/pf.c
7153
pf_change_ap(pd, pd->src, &th->th_sport,
sys/netpfil/pf/pf.c
7154
&nk->addr[pd->sidx], nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
7155
pd->sport = &th->th_sport;
sys/netpfil/pf/pf.c
7156
pd->nsport = th->th_sport;
sys/netpfil/pf/pf.c
7157
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
7160
if (PF_ANEQ(&pd->ndaddr, &nk->addr[pd->didx], pd->af) ||
sys/netpfil/pf/pf.c
7161
nk->port[pd->didx] != pd->ndport) {
sys/netpfil/pf/pf.c
7162
pf_change_ap(pd, pd->dst, &th->th_dport,
sys/netpfil/pf/pf.c
7163
&nk->addr[pd->didx], nk->port[pd->didx]);
sys/netpfil/pf/pf.c
7164
pd->dport = &th->th_dport;
sys/netpfil/pf/pf.c
7165
pd->ndport = th->th_dport;
sys/netpfil/pf/pf.c
7166
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7171
if (PF_ANEQ(&pd->nsaddr, &nk->addr[pd->sidx], pd->af) ||
sys/netpfil/pf/pf.c
7172
nk->port[pd->sidx] != pd->nsport) {
sys/netpfil/pf/pf.c
7173
pf_change_ap(pd, pd->src,
sys/netpfil/pf/pf.c
7174
&pd->hdr.udp.uh_sport,
sys/netpfil/pf/pf.c
7175
&nk->addr[pd->sidx],
sys/netpfil/pf/pf.c
7176
nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
7177
pd->sport = &pd->hdr.udp.uh_sport;
sys/netpfil/pf/pf.c
7178
pd->nsport = pd->hdr.udp.uh_sport;
sys/netpfil/pf/pf.c
7179
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
7182
if (PF_ANEQ(&pd->ndaddr, &nk->addr[pd->didx], pd->af) ||
sys/netpfil/pf/pf.c
7183
nk->port[pd->didx] != pd->ndport) {
sys/netpfil/pf/pf.c
7184
pf_change_ap(pd, pd->dst,
sys/netpfil/pf/pf.c
7185
&pd->hdr.udp.uh_dport,
sys/netpfil/pf/pf.c
7186
&nk->addr[pd->didx],
sys/netpfil/pf/pf.c
7187
nk->port[pd->didx]);
sys/netpfil/pf/pf.c
7188
pd->dport = &pd->hdr.udp.uh_dport;
sys/netpfil/pf/pf.c
7189
pd->ndport = pd->hdr.udp.uh_dport;
sys/netpfil/pf/pf.c
7190
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7195
if (PF_ANEQ(&pd->nsaddr, &nk->addr[pd->sidx], pd->af) ||
sys/netpfil/pf/pf.c
7196
nk->port[pd->sidx] != pd->nsport) {
sys/netpfil/pf/pf.c
7197
pf_change_ap(pd, pd->src,
sys/netpfil/pf/pf.c
7198
&pd->hdr.sctp.src_port,
sys/netpfil/pf/pf.c
7199
&nk->addr[pd->sidx],
sys/netpfil/pf/pf.c
7200
nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
7201
pd->sport = &pd->hdr.sctp.src_port;
sys/netpfil/pf/pf.c
7202
pd->nsport = pd->hdr.sctp.src_port;
sys/netpfil/pf/pf.c
7203
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
7205
if (PF_ANEQ(&pd->ndaddr, &nk->addr[pd->didx], pd->af) ||
sys/netpfil/pf/pf.c
7206
nk->port[pd->didx] != pd->ndport) {
sys/netpfil/pf/pf.c
7207
pf_change_ap(pd, pd->dst,
sys/netpfil/pf/pf.c
7208
&pd->hdr.sctp.dest_port,
sys/netpfil/pf/pf.c
7209
&nk->addr[pd->didx],
sys/netpfil/pf/pf.c
7210
nk->port[pd->didx]);
sys/netpfil/pf/pf.c
7211
pd->dport = &pd->hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
7212
pd->ndport = pd->hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
7213
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7219
if (PF_ANEQ(&pd->nsaddr, &nk->addr[pd->sidx], AF_INET)) {
sys/netpfil/pf/pf.c
7220
pf_change_a(&pd->src->v4.s_addr, pd->ip_sum,
sys/netpfil/pf/pf.c
7221
nk->addr[pd->sidx].v4.s_addr, 0);
sys/netpfil/pf/pf.c
7222
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
7225
if (PF_ANEQ(&pd->ndaddr, &nk->addr[pd->didx], AF_INET)) {
sys/netpfil/pf/pf.c
7226
pf_change_a(&pd->dst->v4.s_addr, pd->ip_sum,
sys/netpfil/pf/pf.c
7227
nk->addr[pd->didx].v4.s_addr, 0);
sys/netpfil/pf/pf.c
7228
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7232
nk->port[pd->sidx] != pd->hdr.icmp.icmp_id) {
sys/netpfil/pf/pf.c
7233
pd->hdr.icmp.icmp_cksum = pf_cksum_fixup(
sys/netpfil/pf/pf.c
7234
pd->hdr.icmp.icmp_cksum, pd->nsport,
sys/netpfil/pf/pf.c
7235
nk->port[pd->sidx], 0);
sys/netpfil/pf/pf.c
7236
pd->hdr.icmp.icmp_id = nk->port[pd->sidx];
sys/netpfil/pf/pf.c
7237
pd->sport = &pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
7239
m_copyback(pd->m, pd->off, ICMP_MINLEN, (caddr_t)&pd->hdr.icmp);
sys/netpfil/pf/pf.c
7244
if (PF_ANEQ(&pd->nsaddr, &nk->addr[pd->sidx], AF_INET6)) {
sys/netpfil/pf/pf.c
7245
pf_change_a6(pd->src, &pd->hdr.icmp6.icmp6_cksum,
sys/netpfil/pf/pf.c
7246
&nk->addr[pd->sidx], 0);
sys/netpfil/pf/pf.c
7247
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
7250
if (PF_ANEQ(&pd->ndaddr, &nk->addr[pd->didx], AF_INET6)) {
sys/netpfil/pf/pf.c
7251
pf_change_a6(pd->dst, &pd->hdr.icmp6.icmp6_cksum,
sys/netpfil/pf/pf.c
7252
&nk->addr[pd->didx], 0);
sys/netpfil/pf/pf.c
7253
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7259
switch (pd->af) {
sys/netpfil/pf/pf.c
7262
if (PF_ANEQ(&pd->nsaddr,
sys/netpfil/pf/pf.c
7263
&nk->addr[pd->sidx], AF_INET)) {
sys/netpfil/pf/pf.c
7264
pf_change_a(&pd->src->v4.s_addr,
sys/netpfil/pf/pf.c
7265
pd->ip_sum,
sys/netpfil/pf/pf.c
7266
nk->addr[pd->sidx].v4.s_addr, 0);
sys/netpfil/pf/pf.c
7267
pf_addrcpy(&pd->nsaddr, pd->src, pd->af);
sys/netpfil/pf/pf.c
727
if (st->rule->rt == PF_REPLYTO || (pd->af != pd->naf && st->direction == PF_IN))
sys/netpfil/pf/pf.c
7270
if (PF_ANEQ(&pd->ndaddr,
sys/netpfil/pf/pf.c
7271
&nk->addr[pd->didx], AF_INET)) {
sys/netpfil/pf/pf.c
7272
pf_change_a(&pd->dst->v4.s_addr,
sys/netpfil/pf/pf.c
7273
pd->ip_sum,
sys/netpfil/pf/pf.c
7274
nk->addr[pd->didx].v4.s_addr, 0);
sys/netpfil/pf/pf.c
7275
pf_addrcpy(&pd->ndaddr, pd->dst, pd->af);
sys/netpfil/pf/pf.c
7281
if (PF_ANEQ(&pd->nsaddr,
sys/netpfil/pf/pf.c
7282
&nk->addr[pd->sidx], AF_INET6)) {
sys/netpfil/pf/pf.c
7283
pf_addrcpy(&pd->nsaddr, &nk->addr[pd->sidx],
sys/netpfil/pf/pf.c
7284
pd->af);
sys/netpfil/pf/pf.c
7285
pf_addrcpy(pd->src, &nk->addr[pd->sidx], pd->af);
sys/netpfil/pf/pf.c
7288
if (PF_ANEQ(&pd->ndaddr,
sys/netpfil/pf/pf.c
7289
&nk->addr[pd->didx], AF_INET6)) {
sys/netpfil/pf/pf.c
7290
pf_addrcpy(&pd->ndaddr, &nk->addr[pd->didx],
sys/netpfil/pf/pf.c
7291
pd->af);
sys/netpfil/pf/pf.c
7292
pf_addrcpy(pd->dst, &nk->addr[pd->didx],
sys/netpfil/pf/pf.c
7293
pd->af);
sys/netpfil/pf/pf.c
7304
pf_tcp_track_full(struct pf_kstate *state, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
7308
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7332
if (pf_normalize_tcp_init(pd, th, src)) {
sys/netpfil/pf/pf.c
7344
pf_change_proto_a(pd->m, &th->th_seq, &th->th_sum, htonl(seq +
sys/netpfil/pf/pf.c
7346
pf_change_proto_a(pd->m, &th->th_ack, &th->th_sum, htonl(ack), 0);
sys/netpfil/pf/pf.c
735
if (pd->related_rule)
sys/netpfil/pf/pf.c
7352
end = seq + pd->p_len;
sys/netpfil/pf/pf.c
7356
src->wscale = pf_get_wscale(pd);
sys/netpfil/pf/pf.c
7397
pf_change_proto_a(pd->m, &th->th_seq, &th->th_sum, htonl(seq +
sys/netpfil/pf/pf.c
7399
pf_change_proto_a(pd->m, &th->th_ack, &th->th_sum, htonl(ack), 0);
sys/netpfil/pf/pf.c
7402
end = seq + pd->p_len;
sys/netpfil/pf/pf.c
7445
if (pf_modulate_sack(pd, th, dst))
sys/netpfil/pf/pf.c
7467
if (pf_normalize_tcp_stateful(pd, reason, th,
sys/netpfil/pf/pf.c
7560
pd->p_len, ackskew, (unsigned long long)state->packets[0],
sys/netpfil/pf/pf.c
7562
pd->dir == PF_IN ? "in" : "out",
sys/netpfil/pf/pf.c
7563
pd->dir == state->direction ? "fwd" : "rev");
sys/netpfil/pf/pf.c
7567
if (pf_normalize_tcp_stateful(pd, reason, th,
sys/netpfil/pf/pf.c
7600
pf_send_tcp(state->rule, pd->af,
sys/netpfil/pf/pf.c
7601
pd->dst, pd->src, th->th_dport,
sys/netpfil/pf/pf.c
7615
seq, orig_seq, ack, pd->p_len, ackskew,
sys/netpfil/pf/pf.c
7618
pd->dir == PF_IN ? "in" : "out",
sys/netpfil/pf/pf.c
7619
pd->dir == state->direction ? "fwd" : "rev");
sys/netpfil/pf/pf.c
7637
pf_tcp_track_sloppy(struct pf_kstate *state, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
7641
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7711
pf_synproxy(struct pf_pdesc *pd, struct pf_kstate *state, u_short *reason)
sys/netpfil/pf/pf.c
7713
struct pf_state_key *sk = state->key[pd->didx];
sys/netpfil/pf/pf.c
7714
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7717
if (pd->dir != state->direction) {
sys/netpfil/pf/pf.c
7726
pf_send_tcp(state->rule, pd->af, pd->dst,
sys/netpfil/pf/pf.c
7727
pd->src, th->th_dport, th->th_sport,
sys/netpfil/pf/pf.c
7748
if (pd->dir == state->direction) {
sys/netpfil/pf/pf.c
7758
pf_send_tcp(state->rule, pd->af,
sys/netpfil/pf/pf.c
7759
&sk->addr[pd->sidx], &sk->addr[pd->didx],
sys/netpfil/pf/pf.c
7760
sk->port[pd->sidx], sk->port[pd->didx],
sys/netpfil/pf/pf.c
7776
pf_send_tcp(state->rule, pd->af, pd->dst,
sys/netpfil/pf/pf.c
7777
pd->src, th->th_dport, th->th_sport,
sys/netpfil/pf/pf.c
7782
pf_send_tcp(state->rule, pd->af,
sys/netpfil/pf/pf.c
7783
&sk->addr[pd->sidx], &sk->addr[pd->didx],
sys/netpfil/pf/pf.c
7784
sk->port[pd->sidx], sk->port[pd->didx],
sys/netpfil/pf/pf.c
7809
pf_synproxy_ack(struct pf_krule *r, struct pf_pdesc *pd, struct pf_kstate **sm,
sys/netpfil/pf/pf.c
7812
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7827
mss = pf_get_mss(pd);
sys/netpfil/pf/pf.c
7828
mss = pf_calc_mss(pd->src, pd->af, rtid, mss);
sys/netpfil/pf/pf.c
7829
mss = pf_calc_mss(pd->dst, pd->af, rtid, mss);
sys/netpfil/pf/pf.c
7832
pf_send_tcp(r, pd->af, pd->dst, pd->src, th->th_dport,
sys/netpfil/pf/pf.c
7841
pf_test_state(struct pf_kstate **state, struct pf_pdesc *pd, u_short *reason)
sys/netpfil/pf/pf.c
7850
key.af = pd->af;
sys/netpfil/pf/pf.c
7851
key.proto = pd->virtual_proto;
sys/netpfil/pf/pf.c
7852
pf_addrcpy(&key.addr[pd->sidx], pd->src, key.af);
sys/netpfil/pf/pf.c
7853
pf_addrcpy(&key.addr[pd->didx], pd->dst, key.af);
sys/netpfil/pf/pf.c
7854
key.port[pd->sidx] = pd->osport;
sys/netpfil/pf/pf.c
7855
key.port[pd->didx] = pd->odport;
sys/netpfil/pf/pf.c
7857
action = pf_find_state(pd, &key, state);
sys/netpfil/pf/pf.c
7862
if (pd->dir == (*state)->direction) {
sys/netpfil/pf/pf.c
7863
if (PF_REVERSED_KEY(*state, pd->af)) {
sys/netpfil/pf/pf.c
7875
if (PF_REVERSED_KEY(*state, pd->af)) {
sys/netpfil/pf/pf.c
7888
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
7890
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
7892
if ((action = pf_synproxy(pd, *state, reason)) != PF_PASS)
sys/netpfil/pf/pf.c
7896
pf_syncookie_check(pd) && pd->dir == PF_IN)) {
sys/netpfil/pf/pf.c
7920
pf_send_challenge_ack(pd, *state, src, dst, reason);
sys/netpfil/pf/pf.c
7925
if (pf_tcp_track_sloppy(*state, pd, reason, src, dst,
sys/netpfil/pf/pf.c
7931
ret = pf_tcp_track_full(*state, pd, reason,
sys/netpfil/pf/pf.c
7955
pd->sctp_flags & PFDESC_SCTP_INIT) {
sys/netpfil/pf/pf.c
7962
if (pf_sctp_track(*state, pd, reason) != PF_PASS)
sys/netpfil/pf/pf.c
7966
if (pd->sctp_flags & PFDESC_SCTP_INIT) {
sys/netpfil/pf/pf.c
7972
if (pd->sctp_flags & PFDESC_SCTP_INIT_ACK) {
sys/netpfil/pf/pf.c
7975
dst->scrub->pfss_v_tag = pd->sctp_initiate_tag;
sys/netpfil/pf/pf.c
7985
(*state)->kif = pd->kif;
sys/netpfil/pf/pf.c
7987
if (pd->sctp_flags & (PFDESC_SCTP_COOKIE | PFDESC_SCTP_HEARTBEAT_ACK)) {
sys/netpfil/pf/pf.c
7993
if (pd->sctp_flags & (PFDESC_SCTP_SHUTDOWN |
sys/netpfil/pf/pf.c
8000
if (pd->sctp_flags & (PFDESC_SCTP_SHUTDOWN_COMPLETE | PFDESC_SCTP_ABORT)) {
sys/netpfil/pf/pf.c
8028
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
8029
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
8031
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
8033
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
8036
sidx = pd->didx;
sys/netpfil/pf/pf.c
8037
didx = pd->sidx;
sys/netpfil/pf/pf.c
8039
sidx = pd->sidx;
sys/netpfil/pf/pf.c
8040
didx = pd->didx;
sys/netpfil/pf/pf.c
8044
pf_addrcpy(&pd->nsaddr, &nk->addr[sidx], nk->af);
sys/netpfil/pf/pf.c
8045
pf_addrcpy(&pd->ndaddr, &nk->addr[didx], nk->af);
sys/netpfil/pf/pf.c
8046
pd->naf = nk->af;
sys/netpfil/pf/pf.c
8050
if (afto || PF_ANEQ(pd->src, &nk->addr[sidx], pd->af) ||
sys/netpfil/pf/pf.c
8051
nk->port[sidx] != pd->osport)
sys/netpfil/pf/pf.c
8052
pf_change_ap(pd, pd->src, pd->sport,
sys/netpfil/pf/pf.c
8055
if (afto || PF_ANEQ(pd->dst, &nk->addr[didx], pd->af) ||
sys/netpfil/pf/pf.c
8056
nk->port[didx] != pd->odport)
sys/netpfil/pf/pf.c
8057
pf_change_ap(pd, pd->dst, pd->dport,
sys/netpfil/pf/pf.c
8063
if (copyback && pd->hdrlen > 0)
sys/netpfil/pf/pf.c
8064
m_copyback(pd->m, pd->off, pd->hdrlen, pd->hdr.any);
sys/netpfil/pf/pf.c
8070
pf_sctp_track(struct pf_kstate *state, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
8074
if (pd->dir == state->direction) {
sys/netpfil/pf/pf.c
8075
if (PF_REVERSED_KEY(state, pd->af))
sys/netpfil/pf/pf.c
8080
if (PF_REVERSED_KEY(state, pd->af))
sys/netpfil/pf/pf.c
8092
src->scrub->pfss_v_tag = pd->hdr.sctp.v_tag;
sys/netpfil/pf/pf.c
8093
else if (src->scrub->pfss_v_tag != pd->hdr.sctp.v_tag)
sys/netpfil/pf/pf.c
8160
pf_sctp_multihome_add_addr(struct pf_pdesc *pd, struct pf_addr *a, uint32_t v_tag)
sys/netpfil/pf/pf.c
8189
if (pf_addr_cmp(&i->addr, a, pd->af) == 0) {
sys/netpfil/pf/pf.c
8207
i->af = pd->af;
sys/netpfil/pf/pf.c
8216
pf_sctp_multihome_delayed(struct pf_pdesc *pd, struct pfi_kkif *kif,
sys/netpfil/pf/pf.c
8233
TAILQ_FOREACH_SAFE(j, &pd->sctp_multihome_jobs, next, tmp) {
sys/netpfil/pf/pf.c
8238
MPASS(! (pd->sctp_flags & PFDESC_SCTP_ADD_IP));
sys/netpfil/pf/pf.c
8242
uint32_t v_tag = pd->sctp_initiate_tag;
sys/netpfil/pf/pf.c
8245
if (s->direction == pd->dir)
sys/netpfil/pf/pf.c
8257
if (pf_addr_cmp(&j->src, pd->src, pd->af) == 0) {
sys/netpfil/pf/pf.c
8261
j->pd.sctp_flags |= PFDESC_SCTP_ADD_IP;
sys/netpfil/pf/pf.c
8265
j->pd.related_rule = s->rule;
sys/netpfil/pf/pf.c
8269
&j->pd, &ra, &rs, &reason, NULL, &match_rules);
sys/netpfil/pf/pf.c
8276
SDT_PROBE4(pf, sctp, multihome, test, kif, r, j->pd.m, ret);
sys/netpfil/pf/pf.c
8293
pf_sctp_multihome_add_addr(pd, &j->src, v_tag);
sys/netpfil/pf/pf.c
8303
.v_tag = pd->hdr.sctp.v_tag,
sys/netpfil/pf/pf.c
8319
if (i->af != pd->af)
sys/netpfil/pf/pf.c
8326
memcpy(&nj->pd, &j->pd, sizeof(j->pd));
sys/netpfil/pf/pf.c
8328
nj->pd.src = &nj->src;
sys/netpfil/pf/pf.c
8331
nj->pd.dst = &nj->dst;
sys/netpfil/pf/pf.c
8332
nj->pd.m = j->pd.m;
sys/netpfil/pf/pf.c
8335
MPASS(nj->pd.pcksum);
sys/netpfil/pf/pf.c
8336
TAILQ_INSERT_TAIL(&pd->sctp_multihome_jobs, nj, next);
sys/netpfil/pf/pf.c
8348
key.af = j->pd.af;
sys/netpfil/pf/pf.c
8350
if (j->pd.dir == PF_IN) { /* wire side, straight */
sys/netpfil/pf/pf.c
8351
pf_addrcpy(&key.addr[0], j->pd.src, key.af);
sys/netpfil/pf/pf.c
8352
pf_addrcpy(&key.addr[1], j->pd.dst, key.af);
sys/netpfil/pf/pf.c
8353
key.port[0] = j->pd.hdr.sctp.src_port;
sys/netpfil/pf/pf.c
8354
key.port[1] = j->pd.hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
8356
pf_addrcpy(&key.addr[1], j->pd.src, key.af);
sys/netpfil/pf/pf.c
8357
pf_addrcpy(&key.addr[0], j->pd.dst, key.af);
sys/netpfil/pf/pf.c
8358
key.port[1] = j->pd.hdr.sctp.src_port;
sys/netpfil/pf/pf.c
8359
key.port[0] = j->pd.hdr.sctp.dest_port;
sys/netpfil/pf/pf.c
8362
action = pf_find_state(&j->pd, &key, &sm);
sys/netpfil/pf/pf.c
8365
if (j->pd.dir == sm->direction) {
sys/netpfil/pf/pf.c
8381
TAILQ_REMOVE(&pd->sctp_multihome_jobs, j, next);
sys/netpfil/pf/pf.c
8386
if (! TAILQ_EMPTY(&pd->sctp_multihome_jobs)) {
sys/netpfil/pf/pf.c
8393
pf_multihome_scan(int start, int len, struct pf_pdesc *pd, int op)
sys/netpfil/pf/pf.c
8398
SDT_PROBE4(pf, sctp, multihome_scan, entry, start, len, pd, op);
sys/netpfil/pf/pf.c
8403
if (!pf_pull_hdr(pd->m, start + off, &h, sizeof(h), NULL,
sys/netpfil/pf/pf.c
8404
pd->af))
sys/netpfil/pf/pf.c
8422
if (!pf_pull_hdr(pd->m, start + off + sizeof(h), &t, sizeof(t),
sys/netpfil/pf/pf.c
8423
NULL, pd->af))
sys/netpfil/pf/pf.c
8427
t.s_addr = pd->src->v4.s_addr;
sys/netpfil/pf/pf.c
8445
memcpy(&job->pd, pd, sizeof(*pd));
sys/netpfil/pf/pf.c
8449
job->pd.src = &job->src;
sys/netpfil/pf/pf.c
8450
memcpy(&job->dst, pd->dst, sizeof(job->dst));
sys/netpfil/pf/pf.c
8451
job->pd.dst = &job->dst;
sys/netpfil/pf/pf.c
8452
job->pd.m = pd->m;
sys/netpfil/pf/pf.c
8455
MPASS(job->pd.pcksum);
sys/netpfil/pf/pf.c
8456
TAILQ_INSERT_TAIL(&pd->sctp_multihome_jobs, job, next);
sys/netpfil/pf/pf.c
8467
if (!pf_pull_hdr(pd->m, start + off + sizeof(h), &t, sizeof(t),
sys/netpfil/pf/pf.c
8468
NULL, pd->af))
sys/netpfil/pf/pf.c
8470
if (memcmp(&t, &pd->src->v6, sizeof(t)) == 0)
sys/netpfil/pf/pf.c
8473
memcpy(&t, &pd->src->v6, sizeof(t));
sys/netpfil/pf/pf.c
8481
memcpy(&job->pd, pd, sizeof(*pd));
sys/netpfil/pf/pf.c
8483
job->pd.src = &job->src;
sys/netpfil/pf/pf.c
8484
memcpy(&job->dst, pd->dst, sizeof(job->dst));
sys/netpfil/pf/pf.c
8485
job->pd.dst = &job->dst;
sys/netpfil/pf/pf.c
8486
job->pd.m = pd->m;
sys/netpfil/pf/pf.c
8489
MPASS(job->pd.pcksum);
sys/netpfil/pf/pf.c
8490
TAILQ_INSERT_TAIL(&pd->sctp_multihome_jobs, job, next);
sys/netpfil/pf/pf.c
8498
if (!pf_pull_hdr(pd->m, start + off, &ah, sizeof(ah),
sys/netpfil/pf/pf.c
8499
NULL, pd->af))
sys/netpfil/pf/pf.c
8503
ntohs(ah.ph.param_length) - sizeof(ah), pd,
sys/netpfil/pf/pf.c
8513
if (!pf_pull_hdr(pd->m, start + off, &ah, sizeof(ah),
sys/netpfil/pf/pf.c
8514
NULL, pd->af))
sys/netpfil/pf/pf.c
8517
ntohs(ah.ph.param_length) - sizeof(ah), pd,
sys/netpfil/pf/pf.c
8534
pf_multihome_scan_init(int start, int len, struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
8539
return (pf_multihome_scan(start, len, pd, SCTP_ADD_IP_ADDRESS));
sys/netpfil/pf/pf.c
8543
pf_multihome_scan_asconf(int start, int len, struct pf_pdesc *pd)
sys/netpfil/pf/pf.c
8548
return (pf_multihome_scan(start, len, pd, SCTP_ADD_IP_ADDRESS));
sys/netpfil/pf/pf.c
8552
pf_icmp_state_lookup(struct pf_state_key_cmp *key, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
8556
int action, direction = pd->dir;
sys/netpfil/pf/pf.c
8558
key->af = pd->af;
sys/netpfil/pf/pf.c
8559
key->proto = pd->proto;
sys/netpfil/pf/pf.c
8561
*iidx = pd->sidx;
sys/netpfil/pf/pf.c
8562
key->port[pd->sidx] = icmpid;
sys/netpfil/pf/pf.c
8563
key->port[pd->didx] = type;
sys/netpfil/pf/pf.c
8565
*iidx = pd->didx;
sys/netpfil/pf/pf.c
8566
key->port[pd->sidx] = type;
sys/netpfil/pf/pf.c
8567
key->port[pd->didx] = icmpid;
sys/netpfil/pf/pf.c
8569
if (pf_state_key_addr_setup(pd, key, multi))
sys/netpfil/pf/pf.c
8572
action = pf_find_state(pd, key, state);
sys/netpfil/pf/pf.c
8581
direction = (pd->af == (*state)->key[PF_SK_WIRE]->af) ?
sys/netpfil/pf/pf.c
8586
(((!inner && direction == pd->dir) ||
sys/netpfil/pf/pf.c
8587
(inner && direction != pd->dir)) ?
sys/netpfil/pf/pf.c
8603
pf_test_state_icmp(struct pf_kstate **state, struct pf_pdesc *pd,
sys/netpfil/pf/pf.c
8606
struct pf_addr *saddr = pd->src, *daddr = pd->dst;
sys/netpfil/pf/pf.c
8618
switch (pd->proto) {
sys/netpfil/pf/pf.c
8621
icmptype = pd->hdr.icmp.icmp_type;
sys/netpfil/pf/pf.c
8622
icmpcode = pd->hdr.icmp.icmp_code;
sys/netpfil/pf/pf.c
8623
icmpid = pd->hdr.icmp.icmp_id;
sys/netpfil/pf/pf.c
8624
icmpsum = &pd->hdr.icmp.icmp_cksum;
sys/netpfil/pf/pf.c
8629
icmptype = pd->hdr.icmp6.icmp6_type;
sys/netpfil/pf/pf.c
8630
icmpcode = pd->hdr.icmp6.icmp6_code;
sys/netpfil/pf/pf.c
8632
icmpid = pd->hdr.icmp6.icmp6_id;
sys/netpfil/pf/pf.c
8634
icmpsum = &pd->hdr.icmp6.icmp6_cksum;
sys/netpfil/pf/pf.c
8638
panic("unhandled proto %d", pd->proto);
sys/netpfil/pf/pf.c
8641
if (pf_icmp_mapping(pd, icmptype, &icmp_dir, &virtual_id,
sys/netpfil/pf/pf.c
8647
ret = pf_icmp_state_lookup(&key, pd, state, virtual_id,
sys/netpfil/pf/pf.c
8650
if (ret == PF_DROP && pd->af == AF_INET6 && icmp_dir == PF_OUT) {
sys/netpfil/pf/pf.c
8652
ret = pf_icmp_state_lookup(&key, pd, state,
sys/netpfil/pf/pf.c
8669
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
8670
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
8672
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
8674
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
8677
sidx = pd->didx;
sys/netpfil/pf/pf.c
8678
didx = pd->sidx;
sys/netpfil/pf/pf.c
8681
sidx = pd->sidx;
sys/netpfil/pf/pf.c
8682
didx = pd->didx;
sys/netpfil/pf/pf.c
8685
switch (pd->af) {
sys/netpfil/pf/pf.c
8691
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
8693
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
8697
PF_ANEQ(pd->src, &nk->addr[sidx], AF_INET))
sys/netpfil/pf/pf.c
8699
pd->ip_sum,
sys/netpfil/pf/pf.c
8703
if (!afto && PF_ANEQ(pd->dst,
sys/netpfil/pf/pf.c
8706
pd->ip_sum,
sys/netpfil/pf/pf.c
8710
pd->hdr.icmp.icmp_id) {
sys/netpfil/pf/pf.c
8711
pd->hdr.icmp.icmp_cksum =
sys/netpfil/pf/pf.c
8713
pd->hdr.icmp.icmp_cksum, icmpid,
sys/netpfil/pf/pf.c
8715
pd->hdr.icmp.icmp_id =
sys/netpfil/pf/pf.c
8719
m_copyback(pd->m, pd->off, ICMP_MINLEN,
sys/netpfil/pf/pf.c
8720
(caddr_t )&pd->hdr.icmp);
sys/netpfil/pf/pf.c
8728
&pd->hdr.icmp6))
sys/netpfil/pf/pf.c
8730
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
8734
PF_ANEQ(pd->src, &nk->addr[sidx], AF_INET6))
sys/netpfil/pf/pf.c
8736
&pd->hdr.icmp6.icmp6_cksum,
sys/netpfil/pf/pf.c
8739
if (!afto && PF_ANEQ(pd->dst,
sys/netpfil/pf/pf.c
8742
&pd->hdr.icmp6.icmp6_cksum,
sys/netpfil/pf/pf.c
8745
if (nk->port[iidx] != pd->hdr.icmp6.icmp6_id)
sys/netpfil/pf/pf.c
8746
pd->hdr.icmp6.icmp6_id =
sys/netpfil/pf/pf.c
8749
m_copyback(pd->m, pd->off, sizeof(struct icmp6_hdr),
sys/netpfil/pf/pf.c
8750
(caddr_t )&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
8755
pf_addrcpy(&pd->nsaddr, &nk->addr[sidx],
sys/netpfil/pf/pf.c
8757
pf_addrcpy(&pd->ndaddr, &nk->addr[didx],
sys/netpfil/pf/pf.c
8759
pd->naf = nk->af;
sys/netpfil/pf/pf.c
8781
pd2.af = pd->af;
sys/netpfil/pf/pf.c
8782
pd2.dir = pd->dir;
sys/netpfil/pf/pf.c
8784
pd2.sidx = (pd->dir == PF_IN) ? 1 : 0;
sys/netpfil/pf/pf.c
8785
pd2.didx = (pd->dir == PF_IN) ? 0 : 1;
sys/netpfil/pf/pf.c
8786
pd2.m = pd->m;
sys/netpfil/pf/pf.c
8787
pd2.pf_mtag = pd->pf_mtag;
sys/netpfil/pf/pf.c
8788
pd2.kif = pd->kif;
sys/netpfil/pf/pf.c
8789
switch (pd->af) {
sys/netpfil/pf/pf.c
8793
ipoff2 = pd->off + ICMP_MINLEN;
sys/netpfil/pf/pf.c
8795
if (!pf_pull_hdr(pd->m, ipoff2, &h2, sizeof(h2),
sys/netpfil/pf/pf.c
8825
ipoff2 = pd->off + sizeof(struct icmp6_hdr);
sys/netpfil/pf/pf.c
8827
if (!pf_pull_hdr(pd->m, ipoff2, &h2_6, sizeof(h2_6),
sys/netpfil/pf/pf.c
8847
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
8850
if (PF_ANEQ(pd->dst, pd2.src, pd->af)) {
sys/netpfil/pf/pf.c
8854
pf_print_host(pd->src, 0, pd->af);
sys/netpfil/pf/pf.c
8856
pf_print_host(pd->dst, 0, pd->af);
sys/netpfil/pf/pf.c
8881
if (!pf_pull_hdr(pd->m, pd2.off, th, 8, reason,
sys/netpfil/pf/pf.c
8901
if (pd->dir == (*state)->direction) {
sys/netpfil/pf/pf.c
8902
if (PF_REVERSED_KEY(*state, pd->af)) {
sys/netpfil/pf/pf.c
8910
if (PF_REVERSED_KEY(*state, pd->af)) {
sys/netpfil/pf/pf.c
8938
pf_print_host(pd->src, 0, pd->af);
sys/netpfil/pf/pf.c
894
pf_packet_rework_nat(struct pf_pdesc *pd, int off, struct pf_state_key *nk)
sys/netpfil/pf/pf.c
8940
pf_print_host(pd->dst, 0, pd->af);
sys/netpfil/pf/pf.c
8951
pf_print_host(pd->src, 0, pd->af);
sys/netpfil/pf/pf.c
8953
pf_print_host(pd->dst, 0, pd->af);
sys/netpfil/pf/pf.c
8966
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
8967
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
8969
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
897
switch (pd->virtual_proto) {
sys/netpfil/pf/pf.c
8974
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
8986
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
8988
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
899
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf.c
8990
(c_caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
8991
if (pf_change_icmp_af(pd->m, ipoff2, pd,
sys/netpfil/pf/pf.c
8993
&nk->addr[didx], pd->af,
sys/netpfil/pf/pf.c
8996
pf_addrcpy(&pd->nsaddr,
sys/netpfil/pf/pf.c
8998
pf_addrcpy(&pd->ndaddr,
sys/netpfil/pf/pf.c
9001
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
9003
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
901
if (PF_ANEQ(pd->src, &nk->addr[pd->sidx], pd->af))
sys/netpfil/pf/pf.c
9011
pd->nsaddr.addr32[3] =
sys/netpfil/pf/pf.c
9012
pd->src->addr32[0];
sys/netpfil/pf/pf.c
9014
pd->naf = pd2.naf = nk->af;
sys/netpfil/pf/pf.c
902
pf_change_ap(pd, pd->src, &th->th_sport,
sys/netpfil/pf/pf.c
903
&nk->addr[pd->sidx], nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
9031
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
904
if (PF_ANEQ(pd->dst, &nk->addr[pd->didx], pd->af))
sys/netpfil/pf/pf.c
9040
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
9048
m_copyback(pd->m, pd->off, ICMP_MINLEN,
sys/netpfil/pf/pf.c
9049
(caddr_t )&pd->hdr.icmp);
sys/netpfil/pf/pf.c
905
pf_change_ap(pd, pd->dst, &th->th_dport,
sys/netpfil/pf/pf.c
9050
m_copyback(pd->m, ipoff2, sizeof(h2),
sys/netpfil/pf/pf.c
9056
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9058
(caddr_t )&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9059
m_copyback(pd->m, ipoff2, sizeof(h2_6),
sys/netpfil/pf/pf.c
906
&nk->addr[pd->didx], nk->port[pd->didx]);
sys/netpfil/pf/pf.c
9064
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
9066
m_copyback(pd->m, pd2.off, 8, (caddr_t)th);
sys/netpfil/pf/pf.c
907
m_copyback(pd->m, off, sizeof(*th), (caddr_t)th);
sys/netpfil/pf/pf.c
9076
if (!pf_pull_hdr(pd->m, pd2.off, uh, sizeof(*uh),
sys/netpfil/pf/pf.c
9101
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
9102
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
9104
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
9109
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
911
struct udphdr *uh = &pd->hdr.udp;
sys/netpfil/pf/pf.c
9121
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
9123
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9125
(c_caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9126
if (pf_change_icmp_af(pd->m, ipoff2, pd,
sys/netpfil/pf/pf.c
9128
&nk->addr[didx], pd->af,
sys/netpfil/pf/pf.c
913
if (PF_ANEQ(pd->src, &nk->addr[pd->sidx], pd->af))
sys/netpfil/pf/pf.c
9131
pf_addrcpy(&pd->nsaddr,
sys/netpfil/pf/pf.c
9133
pf_addrcpy(&pd->ndaddr,
sys/netpfil/pf/pf.c
9136
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
9138
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
914
pf_change_ap(pd, pd->src, &uh->uh_sport,
sys/netpfil/pf/pf.c
9146
pd->nsaddr.addr32[3] =
sys/netpfil/pf/pf.c
9147
pd->src->addr32[0];
sys/netpfil/pf/pf.c
9149
pd->naf = pd2.naf = nk->af;
sys/netpfil/pf/pf.c
915
&nk->addr[pd->sidx], nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
916
if (PF_ANEQ(pd->dst, &nk->addr[pd->didx], pd->af))
sys/netpfil/pf/pf.c
9167
pd->ip_sum, 1, pd2.af);
sys/netpfil/pf/pf.c
917
pf_change_ap(pd, pd->dst, &uh->uh_dport,
sys/netpfil/pf/pf.c
9176
pd->ip_sum, 1, pd2.af);
sys/netpfil/pf/pf.c
918
&nk->addr[pd->didx], nk->port[pd->didx]);
sys/netpfil/pf/pf.c
9181
m_copyback(pd->m, pd->off, ICMP_MINLEN,
sys/netpfil/pf/pf.c
9182
(caddr_t )&pd->hdr.icmp);
sys/netpfil/pf/pf.c
9183
m_copyback(pd->m, ipoff2, sizeof(h2), (caddr_t)&h2);
sys/netpfil/pf/pf.c
9188
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
919
m_copyback(pd->m, off, sizeof(*uh), (caddr_t)uh);
sys/netpfil/pf/pf.c
9190
(caddr_t )&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9191
m_copyback(pd->m, ipoff2, sizeof(h2_6),
sys/netpfil/pf/pf.c
9196
m_copyback(pd->m, pd2.off, sizeof(*uh), (caddr_t)uh);
sys/netpfil/pf/pf.c
9208
if (! pf_pull_hdr(pd->m, pd2.off, sh, sizeof(*sh), reason,
sys/netpfil/pf/pf.c
9228
if (pd->dir == (*state)->direction) {
sys/netpfil/pf/pf.c
9229
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
923
struct sctphdr *sh = &pd->hdr.sctp;
sys/netpfil/pf/pf.c
9234
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
925
if (PF_ANEQ(pd->src, &nk->addr[pd->sidx], pd->af)) {
sys/netpfil/pf/pf.c
9253
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
9254
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
9256
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
926
pf_change_ap(pd, pd->src, &sh->src_port,
sys/netpfil/pf/pf.c
9261
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
927
&nk->addr[pd->sidx], nk->port[pd->sidx]);
sys/netpfil/pf/pf.c
9273
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
9275
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9277
(c_caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9278
if (pf_change_icmp_af(pd->m, ipoff2, pd,
sys/netpfil/pf/pf.c
9280
&nk->addr[didx], pd->af,
sys/netpfil/pf/pf.c
9286
pf_addrcpy(&pd->nsaddr,
sys/netpfil/pf/pf.c
9288
pf_addrcpy(&pd->ndaddr,
sys/netpfil/pf/pf.c
929
if (PF_ANEQ(pd->dst, &nk->addr[pd->didx], pd->af)) {
sys/netpfil/pf/pf.c
9291
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
9293
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
930
pf_change_ap(pd, pd->dst, &sh->dest_port,
sys/netpfil/pf/pf.c
9301
pd->nsaddr.addr32[3] =
sys/netpfil/pf/pf.c
9302
pd->src->addr32[0];
sys/netpfil/pf/pf.c
9304
pd->naf = nk->af;
sys/netpfil/pf/pf.c
931
&nk->addr[pd->didx], nk->port[pd->didx]);
sys/netpfil/pf/pf.c
9316
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
9325
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
9333
m_copyback(pd->m, pd->off, ICMP_MINLEN,
sys/netpfil/pf/pf.c
9334
(caddr_t )&pd->hdr.icmp);
sys/netpfil/pf/pf.c
9335
m_copyback(pd->m, ipoff2, sizeof(h2),
sys/netpfil/pf/pf.c
9341
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9343
(caddr_t )&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9344
m_copyback(pd->m, ipoff2, sizeof(h2_6),
sys/netpfil/pf/pf.c
9349
m_copyback(pd->m, pd2.off, sizeof(*sh), (caddr_t)sh);
sys/netpfil/pf/pf.c
9363
if (!pf_pull_hdr(pd->m, pd2.off, iih, ICMP_MINLEN,
sys/netpfil/pf/pf.c
937
struct icmp *ih = &pd->hdr.icmp;
sys/netpfil/pf/pf.c
9388
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
9389
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
939
if (nk->port[pd->sidx] != ih->icmp_id) {
sys/netpfil/pf/pf.c
9391
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
9396
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
940
pd->hdr.icmp.icmp_cksum = pf_cksum_fixup(
sys/netpfil/pf/pf.c
9411
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
9413
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9415
(c_caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9416
if (pf_change_icmp_af(pd->m, ipoff2, pd,
sys/netpfil/pf/pf.c
9418
&nk->addr[didx], pd->af,
sys/netpfil/pf/pf.c
942
nk->port[pd->sidx], 0);
sys/netpfil/pf/pf.c
9421
pd->proto = IPPROTO_ICMPV6;
sys/netpfil/pf/pf.c
9429
pf_addrcpy(&pd->nsaddr,
sys/netpfil/pf/pf.c
943
ih->icmp_id = nk->port[pd->sidx];
sys/netpfil/pf/pf.c
9431
pf_addrcpy(&pd->ndaddr,
sys/netpfil/pf/pf.c
9439
pd->nsaddr.addr32[3] =
sys/netpfil/pf/pf.c
944
pd->sport = &ih->icmp_id;
sys/netpfil/pf/pf.c
9440
pd->src->addr32[0];
sys/netpfil/pf/pf.c
9441
pd->naf = nk->af;
sys/netpfil/pf/pf.c
9457
pd->ip_sum, 0, AF_INET);
sys/netpfil/pf/pf.c
946
m_copyback(pd->m, off, ICMP_MINLEN, (caddr_t)ih);
sys/netpfil/pf/pf.c
9463
pd2.ip_sum, icmpsum, pd->ip_sum, 0,
sys/netpfil/pf/pf.c
9466
m_copyback(pd->m, pd->off, ICMP_MINLEN, (caddr_t)&pd->hdr.icmp);
sys/netpfil/pf/pf.c
9467
m_copyback(pd->m, ipoff2, sizeof(h2), (caddr_t)&h2);
sys/netpfil/pf/pf.c
9468
m_copyback(pd->m, pd2.off, ICMP_MINLEN, (caddr_t)iih);
sys/netpfil/pf/pf.c
9483
if (!pf_pull_hdr(pd->m, pd2.off, iih,
sys/netpfil/pf/pf.c
951
if (PF_ANEQ(pd->src, &nk->addr[pd->sidx], pd->af)) {
sys/netpfil/pf/pf.c
9515
if (PF_REVERSED_KEY(*state, pd->af))
sys/netpfil/pf/pf.c
9516
nk = (*state)->key[pd->sidx];
sys/netpfil/pf/pf.c
9518
nk = (*state)->key[pd->didx];
sys/netpfil/pf/pf.c
952
switch (pd->af) {
sys/netpfil/pf/pf.c
9523
afto = pd->af != nk->af;
sys/netpfil/pf/pf.c
9538
&pd->hdr.icmp))
sys/netpfil/pf/pf.c
954
pf_change_a(&pd->src->v4.s_addr,
sys/netpfil/pf/pf.c
9540
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9542
(c_caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9543
if (pf_change_icmp_af(pd->m, ipoff2, pd,
sys/netpfil/pf/pf.c
9545
&nk->addr[didx], pd->af,
sys/netpfil/pf/pf.c
9548
pd->proto = IPPROTO_ICMP;
sys/netpfil/pf/pf.c
955
pd->ip_sum, nk->addr[pd->sidx].v4.s_addr,
sys/netpfil/pf/pf.c
9557
pf_addrcpy(&pd->nsaddr,
sys/netpfil/pf/pf.c
9559
pf_addrcpy(&pd->ndaddr,
sys/netpfil/pf/pf.c
9561
pd->naf = nk->af;
sys/netpfil/pf/pf.c
9577
pd->ip_sum, 0, AF_INET6);
sys/netpfil/pf/pf.c
9584
pd->ip_sum, 0, AF_INET6);
sys/netpfil/pf/pf.c
9586
m_copyback(pd->m, pd->off, sizeof(struct icmp6_hdr),
sys/netpfil/pf/pf.c
9587
(caddr_t)&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9588
m_copyback(pd->m, ipoff2, sizeof(h2_6), (caddr_t)&h2_6);
sys/netpfil/pf/pf.c
9589
m_copyback(pd->m, pd2.off, sizeof(struct icmp6_hdr),
sys/netpfil/pf/pf.c
959
pf_addrcpy(pd->src, &nk->addr[pd->sidx],
sys/netpfil/pf/pf.c
960
pd->af);
sys/netpfil/pf/pf.c
9603
pd->pcksum = &pd->sctp_dummy_sum;
sys/netpfil/pf/pf.c
9618
(*state)->key[pd->didx];
sys/netpfil/pf/pf.c
9625
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
963
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
9632
pd->ip_sum, 0, pd2.af);
sys/netpfil/pf/pf.c
9637
m_copyback(pd->m, pd->off, ICMP_MINLEN,
sys/netpfil/pf/pf.c
9638
(caddr_t)&pd->hdr.icmp);
sys/netpfil/pf/pf.c
9639
m_copyback(pd->m, ipoff2, sizeof(h2), (caddr_t)&h2);
sys/netpfil/pf/pf.c
9644
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf.c
9646
(caddr_t )&pd->hdr.icmp6);
sys/netpfil/pf/pf.c
9647
m_copyback(pd->m, ipoff2, sizeof(h2_6),
sys/netpfil/pf/pf.c
966
if (PF_ANEQ(pd->dst, &nk->addr[pd->didx], pd->af)) {
sys/netpfil/pf/pf.c
967
switch (pd->af) {
sys/netpfil/pf/pf.c
969
pf_change_a(&pd->dst->v4.s_addr,
sys/netpfil/pf/pf.c
970
pd->ip_sum, nk->addr[pd->didx].v4.s_addr,
sys/netpfil/pf/pf.c
974
pf_addrcpy(pd->dst, &nk->addr[pd->didx],
sys/netpfil/pf/pf.c
9744
struct pf_kstate *s, struct pf_pdesc *pd, struct inpcb *inp)
sys/netpfil/pf/pf.c
975
pd->af);
sys/netpfil/pf/pf.c
9760
KASSERT(pd->m && r && oifp, ("%s: invalid parameters", __func__));
sys/netpfil/pf/pf.c
9762
SDT_PROBE4(pf, ip, route_to, entry, pd->m, pd, s, oifp);
sys/netpfil/pf/pf.c
9770
KASSERT(pd->dir == PF_IN || pd->dir == PF_OUT ||
sys/netpfil/pf/pf.c
9774
if ((pd->pf_mtag == NULL &&
sys/netpfil/pf/pf.c
9775
((pd->pf_mtag = pf_get_mtag(pd->m)) == NULL)) ||
sys/netpfil/pf/pf.c
9776
pd->pf_mtag->routed++ > 3) {
sys/netpfil/pf/pf.c
9777
m0 = pd->m;
sys/netpfil/pf/pf.c
9778
pd->m = NULL;
sys/netpfil/pf/pf.c
978
unhandled_af(pd->af);
sys/netpfil/pf/pf.c
9784
if (pd->act.rt_kif != NULL)
sys/netpfil/pf/pf.c
9785
ifp = pd->act.rt_kif->pfik_ifp;
sys/netpfil/pf/pf.c
9787
if (pd->act.rt == PF_DUPTO) {
sys/netpfil/pf/pf.c
9788
if ((pd->pf_mtag->flags & PF_MTAG_FLAG_DUPLICATED)) {
sys/netpfil/pf/pf.c
9796
m0 = pd->m;
sys/netpfil/pf/pf.c
9797
pd->m = NULL;
sys/netpfil/pf/pf.c
9803
pd->pf_mtag->flags |= PF_MTAG_FLAG_DUPLICATED;
sys/netpfil/pf/pf.c
9804
if (((m0 = m_dup(pd->m, M_NOWAIT)) == NULL)) {
sys/netpfil/pf/pf.c
9811
if ((pd->act.rt == PF_REPLYTO) == (r_dir == pd->dir)) {
sys/netpfil/pf/pf.c
9812
if (pd->af == pd->naf) {
sys/netpfil/pf/pf.c
9813
pf_dummynet(pd, s, r, &pd->m);
sys/netpfil/pf/pf.c
9828
if (pd->act.rt_kif && pd->act.rt_kif->pfik_ifp &&
sys/netpfil/pf/pf.c
9829
pd->af != pd->naf) {
sys/netpfil/pf/pf.c
9830
if (pd->act.rt == PF_ROUTETO && r->naf != AF_INET) {
sys/netpfil/pf/pf.c
9834
if (pd->act.rt == PF_REPLYTO && r->naf != AF_INET6) {
sys/netpfil/pf/pf.c
9839
m0 = pd->m;
sys/netpfil/pf/pf.c
9853
switch (pd->act.rt_af) {
sys/netpfil/pf/pf.c
9858
rt_gw.sin.sin_addr.s_addr = pd->act.rt_addr.v4.s_addr;
sys/netpfil/pf/pf.c
9866
&pd->act.rt_addr, AF_INET6);
sys/netpfil/pf/pf.c
9875
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
9877
if (r->rt != PF_DUPTO && pd->naf == pd->af)
sys/netpfil/pf/pf.c
9879
ICMP_TIMXCEED_INTRANS, 0, pd->af, r,
sys/netpfil/pf/pf.c
9880
pd->act.rtableid);
sys/netpfil/pf/pf.c
9888
if (ifp == NULL && (pd->af != pd->naf)) {
sys/netpfil/pf/pf.c
9911
m0 = pd->m;
sys/netpfil/pf/pf.c
9912
pd->m = NULL;
sys/netpfil/pf/pf.c
9925
MPASS(r->rt == PF_REPLYTO || (pd->af != pd->naf && s->direction == PF_IN));
sys/netpfil/pf/pf.c
9927
if (pd->act.rt == PF_REPLYTO) {
sys/netpfil/pf/pf.c
9932
if (r->rt == PF_DUPTO || (pd->af != pd->naf && s->direction == PF_IN))
sys/netpfil/pf/pf.c
9935
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
9941
&out_kif->pfik_bytes[pd->naf == AF_INET6][1]
sys/netpfil/pf/pf.c
9942
[action != PF_PASS && action != PF_AFRT], pd->tot_len);
sys/netpfil/pf/pf.c
9944
&out_kif->pfik_packets[pd->naf == AF_INET6][1]
sys/netpfil/pf/pf.c
9949
&pd->act) != PF_PASS) {
sys/netpfil/pf/pf.c
9986
if (pd->dir == PF_IN) {
sys/netpfil/pf/pf.c
9991
pd->dir = PF_OUT;
sys/netpfil/pf/pf_lb.c
1064
if (ctx->pd->dir == PF_OUT) {
sys/netpfil/pf/pf_lb.c
1095
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf_lb.c
1107
if (pf_state_key_setup(pd, pd->nsport, pd->ndport, &ctx->sk,
sys/netpfil/pf/pf_lb.c
1114
idx = pd->sidx;
sys/netpfil/pf/pf_lb.c
1120
idx = pd->didx;
sys/netpfil/pf/pf_lb.c
1128
if (pd->proto == IPPROTO_ICMP) {
sys/netpfil/pf/pf_lb.c
1136
if (pf_get_mape_sport(pd, r, naddr, nportp,
sys/netpfil/pf/pf_lb.c
1147
} else if (pf_get_sport(pd, r, naddr, nportp, low, high,
sys/netpfil/pf/pf_lb.c
1157
switch (pd->dir) {
sys/netpfil/pf/pf_lb.c
1160
switch (pd->af) {
sys/netpfil/pf/pf_lb.c
1171
&pd->nsaddr, AF_INET);
sys/netpfil/pf/pf_lb.c
1184
&pd->nsaddr, AF_INET6);
sys/netpfil/pf/pf_lb.c
1191
&rpool->cur->addr.v.a.mask, &pd->nsaddr,
sys/netpfil/pf/pf_lb.c
1192
pd->af);
sys/netpfil/pf/pf_lb.c
1196
switch (pd->af) {
sys/netpfil/pf/pf_lb.c
1206
&pd->ndaddr, AF_INET);
sys/netpfil/pf/pf_lb.c
1218
&pd->ndaddr, AF_INET6);
sys/netpfil/pf/pf_lb.c
1224
&r->src.addr.v.a.mask, &pd->ndaddr, pd->af);
sys/netpfil/pf/pf_lb.c
1233
reason = pf_map_addr_sn(pd->af, r, &pd->nsaddr, naddr,
sys/netpfil/pf/pf_lb.c
1234
&(pd->naf), NULL, NULL, rpool, PF_SN_NAT);
sys/netpfil/pf/pf_lb.c
1240
&pd->ndaddr, pd->af);
sys/netpfil/pf/pf_lb.c
1243
if (pd->proto == IPPROTO_SCTP)
sys/netpfil/pf/pf_lb.c
1253
tmp_nport = ((ntohs(pd->ndport) - ntohs(r->dst.port[0])) % div) +
sys/netpfil/pf/pf_lb.c
1263
nport = pd->ndport;
sys/netpfil/pf/pf_lb.c
1277
key.af = pd->af;
sys/netpfil/pf/pf_lb.c
1278
key.proto = pd->proto;
sys/netpfil/pf/pf_lb.c
1279
key.port[0] = pd->nsport;
sys/netpfil/pf/pf_lb.c
1280
pf_addrcpy(&key.addr[0], &pd->nsaddr, key.af);
sys/netpfil/pf/pf_lb.c
1327
ntohs(pd->nsport), ntohs(ctx->nk->port[0]));
sys/netpfil/pf/pf_lb.c
1349
pf_get_transaddr_af(struct pf_krule *r, struct pf_pdesc *pd)
sys/netpfil/pf/pf_lb.c
1361
pd->naf == AF_INET ? "inet" : "inet6",
sys/netpfil/pf/pf_lb.c
1363
pf_print_host(&pd->nsaddr, pd->nsport, pd->af);
sys/netpfil/pf/pf_lb.c
1365
pf_print_host(&pd->ndaddr, pd->ndport, pd->af);
sys/netpfil/pf/pf_lb.c
1373
if (pf_get_sport(pd, r, &nsaddr, &nport, r->nat.proxy_port[0],
sys/netpfil/pf/pf_lb.c
1381
if (pd->proto == IPPROTO_ICMPV6 && pd->naf == AF_INET) {
sys/netpfil/pf/pf_lb.c
1382
pd->ndport = ntohs(pd->ndport);
sys/netpfil/pf/pf_lb.c
1383
if (pd->ndport == ICMP6_ECHO_REQUEST)
sys/netpfil/pf/pf_lb.c
1384
pd->ndport = ICMP_ECHO;
sys/netpfil/pf/pf_lb.c
1385
else if (pd->ndport == ICMP6_ECHO_REPLY)
sys/netpfil/pf/pf_lb.c
1386
pd->ndport = ICMP_ECHOREPLY;
sys/netpfil/pf/pf_lb.c
1387
pd->ndport = htons(pd->ndport);
sys/netpfil/pf/pf_lb.c
1388
} else if (pd->proto == IPPROTO_ICMP && pd->naf == AF_INET6) {
sys/netpfil/pf/pf_lb.c
1389
pd->nsport = ntohs(pd->nsport);
sys/netpfil/pf/pf_lb.c
1390
if (pd->ndport == ICMP_ECHO)
sys/netpfil/pf/pf_lb.c
1391
pd->ndport = ICMP6_ECHO_REQUEST;
sys/netpfil/pf/pf_lb.c
1392
else if (pd->ndport == ICMP_ECHOREPLY)
sys/netpfil/pf/pf_lb.c
1393
pd->ndport = ICMP6_ECHO_REPLY;
sys/netpfil/pf/pf_lb.c
1394
pd->nsport = htons(pd->nsport);
sys/netpfil/pf/pf_lb.c
1399
if (pf_map_addr_sn(pd->naf, r, &nsaddr, &naddr, &(pd->naf),
sys/netpfil/pf/pf_lb.c
140
struct pf_pdesc *pd = ctx->pd;
sys/netpfil/pf/pf_lb.c
1403
pd->ndport = htons(r->rdr.proxy_port[0]);
sys/netpfil/pf/pf_lb.c
1405
if (pd->naf == AF_INET) {
sys/netpfil/pf/pf_lb.c
1409
inet_nat46(pd->naf, &pd->ndaddr, &ndaddr, &naddr,
sys/netpfil/pf/pf_lb.c
1415
inet_nat64(pd->naf, &pd->ndaddr, &ndaddr, &naddr,
sys/netpfil/pf/pf_lb.c
1419
if (pd->naf == AF_INET) {
sys/netpfil/pf/pf_lb.c
1425
inet_nat64(pd->naf, &pd->ndaddr, &ndaddr, &pd->ndaddr,
sys/netpfil/pf/pf_lb.c
1436
inet_nat64(pd->naf, &pd->ndaddr, &ndaddr, &nsaddr,
sys/netpfil/pf/pf_lb.c
1441
pf_addrcpy(&pd->nsaddr, &nsaddr, pd->naf);
sys/netpfil/pf/pf_lb.c
1442
pf_addrcpy(&pd->ndaddr, &ndaddr, pd->naf);
sys/netpfil/pf/pf_lb.c
1446
pd->naf == AF_INET ? "inet" : "inet6",
sys/netpfil/pf/pf_lb.c
1448
pf_print_host(&pd->nsaddr, pd->nsport, pd->naf);
sys/netpfil/pf/pf_lb.c
1450
pf_print_host(&pd->ndaddr, pd->ndport, pd->naf);
sys/netpfil/pf/pf_lb.c
148
if (r->action == PF_BINAT && pd->dir == PF_IN) {
sys/netpfil/pf/pf_lb.c
158
PF_TEST_ATTRIB(pfi_kkif_match(r->kif, pd->kif) == r->ifnot,
sys/netpfil/pf/pf_lb.c
160
PF_TEST_ATTRIB(r->direction && r->direction != pd->dir,
sys/netpfil/pf/pf_lb.c
162
PF_TEST_ATTRIB(r->af && r->af != pd->af,
sys/netpfil/pf/pf_lb.c
164
PF_TEST_ATTRIB(r->proto && r->proto != pd->proto,
sys/netpfil/pf/pf_lb.c
166
PF_TEST_ATTRIB(PF_MISMATCHAW(&src->addr, &pd->nsaddr, pd->af,
sys/netpfil/pf/pf_lb.c
167
src->neg, pd->kif, M_GETFIB(pd->m)),
sys/netpfil/pf/pf_lb.c
171
src->port[0], src->port[1], pd->nsport),
sys/netpfil/pf/pf_lb.c
175
PF_MISMATCHAW(&dst->addr, &pd->ndaddr, pd->af, dst->neg, NULL,
sys/netpfil/pf/pf_lb.c
176
M_GETFIB(pd->m)),
sys/netpfil/pf/pf_lb.c
178
PF_TEST_ATTRIB(xdst != NULL && PF_MISMATCHAW(xdst, &pd->ndaddr, pd->af,
sys/netpfil/pf/pf_lb.c
179
0, NULL, M_GETFIB(pd->m)),
sys/netpfil/pf/pf_lb.c
183
dst->port[1], pd->ndport),
sys/netpfil/pf/pf_lb.c
185
PF_TEST_ATTRIB(r->match_tag && !pf_match_tag(pd->m, r, &ctx->tag,
sys/netpfil/pf/pf_lb.c
186
pd->pf_mtag ? pd->pf_mtag->tag : 0),
sys/netpfil/pf/pf_lb.c
188
PF_TEST_ATTRIB(r->os_fingerprint != PF_OSFP_ANY && (pd->proto !=
sys/netpfil/pf/pf_lb.c
189
IPPROTO_TCP || !pf_osfp_match(pf_osfp_fingerprint(pd,
sys/netpfil/pf/pf_lb.c
190
&pd->hdr.tcp), r->os_fingerprint)),
sys/netpfil/pf/pf_lb.c
231
if (ctx->tag > 0 && pf_tag_packet(pd, ctx->tag))
sys/netpfil/pf/pf_lb.c
234
M_SETFIB(pd->m, rtableid);
sys/netpfil/pf/pf_lb.c
292
pf_get_sport(struct pf_pdesc *pd, struct pf_krule *r, struct pf_addr *naddr,
sys/netpfil/pf/pf_lb.c
298
int dir = (pd->dir == PF_IN) ? PF_OUT : PF_IN;
sys/netpfil/pf/pf_lb.c
299
int sidx = pd->sidx;
sys/netpfil/pf/pf_lb.c
300
int didx = pd->didx;
sys/netpfil/pf/pf_lb.c
312
if (pd->proto == IPPROTO_UDP && (rpool->opts & PF_POOL_ENDPI)) {
sys/netpfil/pf/pf_lb.c
316
udp_source.af = pd->af;
sys/netpfil/pf/pf_lb.c
317
pf_addrcpy(&udp_source.addr, &pd->nsaddr, pd->af);
sys/netpfil/pf/pf_lb.c
318
udp_source.port = pd->nsport;
sys/netpfil/pf/pf_lb.c
326
pd->af);
sys/netpfil/pf/pf_lb.c
334
sn = pf_find_src_node(&pd->nsaddr, r,
sys/netpfil/pf/pf_lb.c
335
pd->af, &sh, sn_type, false);
sys/netpfil/pf/pf_lb.c
340
*udp_mapping = pf_udp_mapping_create(pd->af, &pd->nsaddr,
sys/netpfil/pf/pf_lb.c
341
pd->nsport, &init_addr, 0);
sys/netpfil/pf/pf_lb.c
348
if (pf_map_addr_sn(pd->naf, r, &pd->nsaddr, naddr, &(pd->naf), NULL,
sys/netpfil/pf/pf_lb.c
352
if (pd->proto == IPPROTO_ICMP) {
sys/netpfil/pf/pf_lb.c
353
if (pd->ndport == htons(ICMP_ECHO)) {
sys/netpfil/pf/pf_lb.c
363
if (pd->proto == IPPROTO_ICMPV6) {
sys/netpfil/pf/pf_lb.c
364
if (pd->ndport == htons(ICMP6_ECHO_REQUEST)) {
sys/netpfil/pf/pf_lb.c
376
key.af = pd->naf;
sys/netpfil/pf/pf_lb.c
377
key.proto = pd->proto;
sys/netpfil/pf/pf_lb.c
380
pf_addrcpy(&key.addr[didx], &pd->ndaddr, key.af);
sys/netpfil/pf/pf_lb.c
382
key.port[didx] = pd->ndport;
sys/netpfil/pf/pf_lb.c
386
pd->af);
sys/netpfil/pf/pf_lb.c
392
if (pd->proto == IPPROTO_SCTP) {
sys/netpfil/pf/pf_lb.c
393
key.port[sidx] = pd->nsport;
sys/netpfil/pf/pf_lb.c
397
*nport = pd->nsport;
sys/netpfil/pf/pf_lb.c
402
} else if (!(pd->proto == IPPROTO_TCP || pd->proto == IPPROTO_UDP ||
sys/netpfil/pf/pf_lb.c
403
pd->proto == IPPROTO_ICMP) || (low == 0 && high == 0)) {
sys/netpfil/pf/pf_lb.c
408
key.port[sidx] = pd->nsport;
sys/netpfil/pf/pf_lb.c
411
(*udp_mapping)->endpoints[1].port = pd->nsport;
sys/netpfil/pf/pf_lb.c
413
*nport = pd->nsport;
sys/netpfil/pf/pf_lb.c
419
*nport = pd->nsport;
sys/netpfil/pf/pf_lb.c
470
if (pd->proto == IPPROTO_UDP &&
sys/netpfil/pf/pf_lb.c
497
if (pf_map_addr_sn(pd->naf, r, &pd->nsaddr, naddr,
sys/netpfil/pf/pf_lb.c
498
&(pd->naf), NULL, &init_addr, rpool, sn_type))
sys/netpfil/pf/pf_lb.c
507
} while (! PF_AEQ(&init_addr, naddr, pd->naf) );
sys/netpfil/pf/pf_lb.c
527
pf_get_mape_sport(struct pf_pdesc *pd, struct pf_krule *r,
sys/netpfil/pf/pf_lb.c
548
if (!pf_get_sport(pd, r, naddr, nport, low, low | highmask,
sys/netpfil/pf/pf_lb.c
554
if (!pf_get_sport(pd, r, naddr, nport, low, low | highmask,
sys/netpfil/pf/pf_norm.c
1052
struct pf_pdesc pd;
sys/netpfil/pf/pf_norm.c
1116
memset(&pd, 0, sizeof(pd));
sys/netpfil/pf/pf_norm.c
1117
pd.pf_mtag = pf_find_mtag(m);
sys/netpfil/pf/pf_norm.c
1153
pf_normalize_ip(u_short *reason, struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
1156
struct ip *h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_norm.c
1182
if (pfi_kkif_match(r->kif, pd->kif) == r->ifnot)
sys/netpfil/pf/pf_norm.c
1184
else if (r->direction && r->direction != pd->dir)
sys/netpfil/pf/pf_norm.c
1192
r->src.neg, pd->kif, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1196
r->dst.neg, NULL, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1198
else if (r->match_tag && !pf_match_tag(pd->m, r, &tag,
sys/netpfil/pf/pf_norm.c
1199
pd->pf_mtag ? pd->pf_mtag->tag : 0))
sys/netpfil/pf/pf_norm.c
1212
pf_counter_u64_add_protected(&r->packets[pd->dir == PF_OUT], 1);
sys/netpfil/pf/pf_norm.c
1213
pf_counter_u64_add_protected(&r->bytes[pd->dir == PF_OUT], pd->tot_len);
sys/netpfil/pf/pf_norm.c
1215
pf_rule_to_actions(r, &pd->act);
sys/netpfil/pf/pf_norm.c
1277
verdict = pf_reassemble(&pd->m, reason);
sys/netpfil/pf/pf_norm.c
1283
if (pd->m == NULL)
sys/netpfil/pf/pf_norm.c
1286
h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_norm.c
1287
pd->tot_len = htons(h->ip_len);
sys/netpfil/pf/pf_norm.c
1306
PFLOG_PACKET(PF_DROP, *reason, r, NULL, NULL, pd, 1, NULL);
sys/netpfil/pf/pf_norm.c
1315
struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
1336
if (pfi_kkif_match(r->kif, pd->kif) == r->ifnot)
sys/netpfil/pf/pf_norm.c
1338
else if (r->direction && r->direction != pd->dir)
sys/netpfil/pf/pf_norm.c
1342
else if (r->proto && r->proto != pd->proto)
sys/netpfil/pf/pf_norm.c
1345
(struct pf_addr *)&pd->src, AF_INET6,
sys/netpfil/pf/pf_norm.c
1346
r->src.neg, pd->kif, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1349
(struct pf_addr *)&pd->dst, AF_INET6,
sys/netpfil/pf/pf_norm.c
1350
r->dst.neg, NULL, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1363
pf_counter_u64_add_protected(&r->packets[pd->dir == PF_OUT], 1);
sys/netpfil/pf/pf_norm.c
1364
pf_counter_u64_add_protected(&r->bytes[pd->dir == PF_OUT], pd->tot_len);
sys/netpfil/pf/pf_norm.c
1366
pf_rule_to_actions(r, &pd->act);
sys/netpfil/pf/pf_norm.c
1369
if (!pf_pull_hdr(pd->m, off, &frag, sizeof(frag), reason, AF_INET6))
sys/netpfil/pf/pf_norm.c
1375
if (pd->virtual_proto == PF_VPROTO_FRAGMENT) {
sys/netpfil/pf/pf_norm.c
1378
if (pf_reassemble6(&pd->m, &frag, off, pd->extoff, reason) != PF_PASS)
sys/netpfil/pf/pf_norm.c
1380
if (pd->m == NULL)
sys/netpfil/pf/pf_norm.c
1382
h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf_norm.c
1383
pd->tot_len = ntohs(h->ip6_plen) + sizeof(struct ip6_hdr);
sys/netpfil/pf/pf_norm.c
1391
pf_normalize_tcp(struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
1394
struct tcphdr *th = &pd->hdr.tcp;
sys/netpfil/pf/pf_norm.c
1398
sa_family_t af = pd->af;
sys/netpfil/pf/pf_norm.c
1409
if (pfi_kkif_match(r->kif, pd->kif) == r->ifnot)
sys/netpfil/pf/pf_norm.c
1411
else if (r->direction && r->direction != pd->dir)
sys/netpfil/pf/pf_norm.c
1415
else if (r->proto && r->proto != pd->proto)
sys/netpfil/pf/pf_norm.c
1417
else if (PF_MISMATCHAW(&r->src.addr, pd->src, af,
sys/netpfil/pf/pf_norm.c
1418
r->src.neg, pd->kif, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1423
else if (PF_MISMATCHAW(&r->dst.addr, pd->dst, af,
sys/netpfil/pf/pf_norm.c
1424
r->dst.neg, NULL, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
1430
pf_osfp_fingerprint(pd, th),
sys/netpfil/pf/pf_norm.c
1446
pf_counter_u64_add_protected(&r->packets[pd->dir == PF_OUT], 1);
sys/netpfil/pf/pf_norm.c
1447
pf_counter_u64_add_protected(&r->bytes[pd->dir == PF_OUT], pd->tot_len);
sys/netpfil/pf/pf_norm.c
1449
pf_rule_to_actions(rm, &pd->act);
sys/netpfil/pf/pf_norm.c
1453
pd->flags |= PFDESC_TCP_NORM;
sys/netpfil/pf/pf_norm.c
1489
th->th_sum = pf_proto_cksum_fixup(pd->m, th->th_sum, ov, nv, 0);
sys/netpfil/pf/pf_norm.c
1495
th->th_sum = pf_proto_cksum_fixup(pd->m, th->th_sum, th->th_urp,
sys/netpfil/pf/pf_norm.c
1503
m_copyback(pd->m, pd->off, sizeof(*th), (caddr_t)th);
sys/netpfil/pf/pf_norm.c
1510
PFLOG_PACKET(PF_DROP, reason, r, NULL, NULL, pd, 1, NULL);
sys/netpfil/pf/pf_norm.c
1515
pf_normalize_tcp_init(struct pf_pdesc *pd, struct tcphdr *th,
sys/netpfil/pf/pf_norm.c
1529
switch (pd->af) {
sys/netpfil/pf/pf_norm.c
1532
struct ip *h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_norm.c
1539
struct ip6_hdr *h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf_norm.c
1545
unhandled_af(pd->af);
sys/netpfil/pf/pf_norm.c
1556
if (olen < TCPOLEN_TIMESTAMP || !pf_pull_hdr(pd->m,
sys/netpfil/pf/pf_norm.c
1557
pd->off + sizeof(*th), opts, olen, NULL, pd->af))
sys/netpfil/pf/pf_norm.c
1589
pf_normalize_sctp_init(struct pf_pdesc *pd, struct pf_state_peer *src,
sys/netpfil/pf/pf_norm.c
1602
dst->scrub->pfss_v_tag = pd->sctp_initiate_tag;
sys/netpfil/pf/pf_norm.c
1608
pf_normalize_tcp_stateful(struct pf_pdesc *pd,
sys/netpfil/pf/pf_norm.c
1628
switch (pd->af) {
sys/netpfil/pf/pf_norm.c
1632
struct ip *h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_norm.c
1643
struct ip6_hdr *h = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf_norm.c
1652
unhandled_af(pd->af);
sys/netpfil/pf/pf_norm.c
1660
pf_pull_hdr(pd->m, pd->off + sizeof(*th), opts, olen, NULL, pd->af)) {
sys/netpfil/pf/pf_norm.c
1689
pf_patch_32(pd,
sys/netpfil/pf/pf_norm.c
1700
pf_patch_32(pd, tsr, htonl(tsecr),
sys/netpfil/pf/pf_norm.c
1712
m_copyback(pd->m, pd->off + sizeof(*th), olen, opts);
sys/netpfil/pf/pf_norm.c
1875
|| pd->p_len > 0 || (tcp_get_flags(th) & TH_SYN)) &&
sys/netpfil/pf/pf_norm.c
1905
if (pd->p_len > 0 && (src->scrub->pfss_flags & PFSS_DATA_TS)) {
sys/netpfil/pf/pf_norm.c
1930
if (pd->p_len > 0 && src->scrub && (src->scrub->pfss_flags &
sys/netpfil/pf/pf_norm.c
1982
pf_normalize_mss(struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
1987
olen = (pd->hdr.tcp.th_off << 2) - sizeof(struct tcphdr);
sys/netpfil/pf/pf_norm.c
1988
optsoff = pd->off + sizeof(struct tcphdr);
sys/netpfil/pf/pf_norm.c
1990
!pf_pull_hdr(pd->m, optsoff, opts, olen, NULL, pd->af))
sys/netpfil/pf/pf_norm.c
1999
if (ntohs(mss) > pd->act.max_mss) {
sys/netpfil/pf/pf_norm.c
2001
pf_patch_16(pd, &mss,
sys/netpfil/pf/pf_norm.c
2002
htons(pd->act.max_mss), PF_ALGNMNT(mssoffopts));
sys/netpfil/pf/pf_norm.c
2003
m_copyback(pd->m, optsoff + mssoffopts,
sys/netpfil/pf/pf_norm.c
2005
m_copyback(pd->m, pd->off,
sys/netpfil/pf/pf_norm.c
2006
sizeof(struct tcphdr), (caddr_t)&pd->hdr.tcp);
sys/netpfil/pf/pf_norm.c
2016
pf_scan_sctp(struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
2023
while (pd->off + chunk_off < pd->tot_len) {
sys/netpfil/pf/pf_norm.c
2024
if (!pf_pull_hdr(pd->m, pd->off + chunk_off, &ch, sizeof(ch),
sys/netpfil/pf/pf_norm.c
2025
NULL, pd->af))
sys/netpfil/pf/pf_norm.c
2040
if (!pf_pull_hdr(pd->m, pd->off + chunk_start, &init,
sys/netpfil/pf/pf_norm.c
2041
sizeof(init), NULL, pd->af))
sys/netpfil/pf/pf_norm.c
2062
pd->hdr.sctp.v_tag != 0)
sys/netpfil/pf/pf_norm.c
2065
pd->sctp_initiate_tag = init.init.initiate_tag;
sys/netpfil/pf/pf_norm.c
2068
pd->sctp_flags |= PFDESC_SCTP_INIT;
sys/netpfil/pf/pf_norm.c
2070
pd->sctp_flags |= PFDESC_SCTP_INIT_ACK;
sys/netpfil/pf/pf_norm.c
2072
ret = pf_multihome_scan_init(pd->off + chunk_start,
sys/netpfil/pf/pf_norm.c
2073
ntohs(init.ch.chunk_length), pd);
sys/netpfil/pf/pf_norm.c
2080
pd->sctp_flags |= PFDESC_SCTP_ABORT;
sys/netpfil/pf/pf_norm.c
2084
pd->sctp_flags |= PFDESC_SCTP_SHUTDOWN;
sys/netpfil/pf/pf_norm.c
2087
pd->sctp_flags |= PFDESC_SCTP_SHUTDOWN_COMPLETE;
sys/netpfil/pf/pf_norm.c
2090
pd->sctp_flags |= PFDESC_SCTP_COOKIE;
sys/netpfil/pf/pf_norm.c
2093
pd->sctp_flags |= PFDESC_SCTP_COOKIE_ACK;
sys/netpfil/pf/pf_norm.c
2096
pd->sctp_flags |= PFDESC_SCTP_DATA;
sys/netpfil/pf/pf_norm.c
2099
pd->sctp_flags |= PFDESC_SCTP_HEARTBEAT;
sys/netpfil/pf/pf_norm.c
2102
pd->sctp_flags |= PFDESC_SCTP_HEARTBEAT_ACK;
sys/netpfil/pf/pf_norm.c
2105
pd->sctp_flags |= PFDESC_SCTP_ASCONF;
sys/netpfil/pf/pf_norm.c
2107
ret = pf_multihome_scan_asconf(pd->off + chunk_start,
sys/netpfil/pf/pf_norm.c
2108
ntohs(ch.chunk_length), pd);
sys/netpfil/pf/pf_norm.c
2113
pd->sctp_flags |= PFDESC_SCTP_OTHER;
sys/netpfil/pf/pf_norm.c
2119
if (pd->off + chunk_off != pd->tot_len)
sys/netpfil/pf/pf_norm.c
2126
if ((pd->sctp_flags & PFDESC_SCTP_INIT) &&
sys/netpfil/pf/pf_norm.c
2127
(pd->sctp_flags & ~PFDESC_SCTP_INIT))
sys/netpfil/pf/pf_norm.c
2129
if ((pd->sctp_flags & PFDESC_SCTP_INIT_ACK) &&
sys/netpfil/pf/pf_norm.c
2130
(pd->sctp_flags & ~PFDESC_SCTP_INIT_ACK))
sys/netpfil/pf/pf_norm.c
2132
if ((pd->sctp_flags & PFDESC_SCTP_SHUTDOWN_COMPLETE) &&
sys/netpfil/pf/pf_norm.c
2133
(pd->sctp_flags & ~PFDESC_SCTP_SHUTDOWN_COMPLETE))
sys/netpfil/pf/pf_norm.c
2135
if ((pd->sctp_flags & PFDESC_SCTP_ABORT) &&
sys/netpfil/pf/pf_norm.c
2136
(pd->sctp_flags & PFDESC_SCTP_DATA)) {
sys/netpfil/pf/pf_norm.c
2148
pf_normalize_sctp(struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
2151
struct sctphdr *sh = &pd->hdr.sctp;
sys/netpfil/pf/pf_norm.c
2153
sa_family_t af = pd->af;
sys/netpfil/pf/pf_norm.c
2164
if (pfi_kkif_match(r->kif, pd->kif) == r->ifnot)
sys/netpfil/pf/pf_norm.c
2166
else if (r->direction && r->direction != pd->dir)
sys/netpfil/pf/pf_norm.c
2170
else if (r->proto && r->proto != pd->proto)
sys/netpfil/pf/pf_norm.c
2172
else if (PF_MISMATCHAW(&r->src.addr, pd->src, af,
sys/netpfil/pf/pf_norm.c
2173
r->src.neg, pd->kif, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
2178
else if (PF_MISMATCHAW(&r->dst.addr, pd->dst, af,
sys/netpfil/pf/pf_norm.c
2179
r->dst.neg, NULL, M_GETFIB(pd->m)))
sys/netpfil/pf/pf_norm.c
2197
pf_counter_u64_add_protected(&r->packets[pd->dir == PF_OUT], 1);
sys/netpfil/pf/pf_norm.c
2198
pf_counter_u64_add_protected(&r->bytes[pd->dir == PF_OUT], pd->tot_len);
sys/netpfil/pf/pf_norm.c
2203
if ((pd->tot_len - pd->off - sizeof(struct sctphdr)) % 4)
sys/netpfil/pf/pf_norm.c
2207
if (pd->sctp_flags & PFDESC_SCTP_INIT)
sys/netpfil/pf/pf_norm.c
2208
if (pd->sctp_flags & ~PFDESC_SCTP_INIT)
sys/netpfil/pf/pf_norm.c
2216
PFLOG_PACKET(PF_DROP, reason, r, NULL, NULL, pd,
sys/netpfil/pf/pf_norm.c
2224
pf_scrub(struct pf_pdesc *pd)
sys/netpfil/pf/pf_norm.c
2227
struct ip *h = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_norm.c
2229
struct ip6_hdr *h6 = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf_norm.c
2233
if (pd->af == AF_INET && pd->act.flags & PFSTATE_NODF &&
sys/netpfil/pf/pf_norm.c
2243
if (pd->af == AF_INET && pd->act.min_ttl &&
sys/netpfil/pf/pf_norm.c
2244
h->ip_ttl < pd->act.min_ttl) {
sys/netpfil/pf/pf_norm.c
2247
pd->ttl = h->ip_ttl = pd->act.min_ttl;
sys/netpfil/pf/pf_norm.c
2252
if (pd->af == AF_INET6 && pd->act.min_ttl &&
sys/netpfil/pf/pf_norm.c
2253
h6->ip6_hlim < pd->act.min_ttl)
sys/netpfil/pf/pf_norm.c
2254
pd->ttl = h6->ip6_hlim = pd->act.min_ttl;
sys/netpfil/pf/pf_norm.c
2257
if (pd->act.flags & PFSTATE_SETTOS) {
sys/netpfil/pf/pf_norm.c
2258
switch (pd->af) {
sys/netpfil/pf/pf_norm.c
2263
h->ip_tos = pd->act.set_tos | (h->ip_tos & IPTOS_ECN_MASK);
sys/netpfil/pf/pf_norm.c
2264
pd->tos = h->ip_tos & ~IPTOS_ECN_MASK;
sys/netpfil/pf/pf_norm.c
2273
h6->ip6_flow |= htonl((pd->act.set_tos | IPV6_ECN(h6)) << 20);
sys/netpfil/pf/pf_norm.c
2274
pd->tos = IPV6_DSCP(h6);
sys/netpfil/pf/pf_norm.c
2282
if (pd->af == AF_INET &&
sys/netpfil/pf/pf_norm.c
2283
pd->act.flags & PFSTATE_RANDOMID && !(h->ip_off & ~htons(IP_DF))) {
sys/netpfil/pf/pf_osfp.c
67
pf_osfp_fingerprint(struct pf_pdesc *pd, const struct tcphdr *tcp)
sys/netpfil/pf/pf_osfp.c
73
if (pd->proto != IPPROTO_TCP || (tcp->th_off << 2) < sizeof(*tcp))
sys/netpfil/pf/pf_osfp.c
76
switch (pd->af) {
sys/netpfil/pf/pf_osfp.c
78
ip = mtod(pd->m, struct ip *);
sys/netpfil/pf/pf_osfp.c
82
ip6 = mtod(pd->m, struct ip6_hdr *);
sys/netpfil/pf/pf_osfp.c
85
if (!pf_pull_hdr(pd->m, pd->off, hdr, tcp->th_off << 2, NULL,
sys/netpfil/pf/pf_osfp.c
86
pd->af)) return (NULL);
sys/netpfil/pf/pf_syncookies.c
261
pf_synflood_check(struct pf_pdesc *pd)
sys/netpfil/pf/pf_syncookies.c
263
MPASS(pd->proto == IPPROTO_TCP);
sys/netpfil/pf/pf_syncookies.c
266
if (pd->pf_mtag && (pd->pf_mtag->flags & PF_MTAG_FLAG_SYNCOOKIE_RECREATED))
sys/netpfil/pf/pf_syncookies.c
290
pf_syncookie_send(struct pf_pdesc *pd, u_short *reason)
sys/netpfil/pf/pf_syncookies.c
295
mss = max(V_tcp_mssdflt, pf_get_mss(pd));
sys/netpfil/pf/pf_syncookies.c
296
iss = pf_syncookie_generate(pd, mss);
sys/netpfil/pf/pf_syncookies.c
297
pf_send_tcp(NULL, pd->af, pd->dst, pd->src, *pd->dport, *pd->sport,
sys/netpfil/pf/pf_syncookies.c
298
iss, ntohl(pd->hdr.tcp.th_seq) + 1, TH_SYN|TH_ACK, 0, mss,
sys/netpfil/pf/pf_syncookies.c
299
0, M_SKIP_FIREWALL | (pd->m->m_flags & M_LOOP), 0, 0,
sys/netpfil/pf/pf_syncookies.c
300
pd->act.rtableid, reason);
sys/netpfil/pf/pf_syncookies.c
308
pf_syncookie_check(struct pf_pdesc *pd)
sys/netpfil/pf/pf_syncookies.c
313
MPASS(pd->proto == IPPROTO_TCP);
sys/netpfil/pf/pf_syncookies.c
316
seq = ntohl(pd->hdr.tcp.th_seq) - 1;
sys/netpfil/pf/pf_syncookies.c
317
ack = ntohl(pd->hdr.tcp.th_ack) - 1;
sys/netpfil/pf/pf_syncookies.c
325
hash = pf_syncookie_mac(pd, cookie, seq);
sys/netpfil/pf/pf_syncookies.c
333
pf_syncookie_validate(struct pf_pdesc *pd)
sys/netpfil/pf/pf_syncookies.c
338
if (! pf_syncookie_check(pd))
sys/netpfil/pf/pf_syncookies.c
341
ack = ntohl(pd->hdr.tcp.th_ack) - 1;
sys/netpfil/pf/pf_syncookies.c
424
pf_syncookie_mac(struct pf_pdesc *pd, union pf_syncookie cookie, uint32_t seq)
sys/netpfil/pf/pf_syncookies.c
430
MPASS(pd->proto == IPPROTO_TCP);
sys/netpfil/pf/pf_syncookies.c
435
switch (pd->af) {
sys/netpfil/pf/pf_syncookies.c
437
SipHash_Update(&ctx, pd->src, sizeof(pd->src->v4));
sys/netpfil/pf/pf_syncookies.c
438
SipHash_Update(&ctx, pd->dst, sizeof(pd->dst->v4));
sys/netpfil/pf/pf_syncookies.c
441
SipHash_Update(&ctx, pd->src, sizeof(pd->src->v6));
sys/netpfil/pf/pf_syncookies.c
442
SipHash_Update(&ctx, pd->dst, sizeof(pd->dst->v6));
sys/netpfil/pf/pf_syncookies.c
448
SipHash_Update(&ctx, pd->sport, sizeof(*pd->sport));
sys/netpfil/pf/pf_syncookies.c
449
SipHash_Update(&ctx, pd->dport, sizeof(*pd->dport));
sys/netpfil/pf/pf_syncookies.c
458
pf_syncookie_generate(struct pf_pdesc *pd, uint16_t mss)
sys/netpfil/pf/pf_syncookies.c
475
wscale = pf_get_wscale(pd);
sys/netpfil/pf/pf_syncookies.c
483
hash = pf_syncookie_mac(pd, cookie, ntohl(pd->hdr.tcp.th_seq));
sys/netpfil/pf/pf_syncookies.c
498
pf_syncookie_recreate_syn(struct pf_pdesc *pd, u_short *reason)
sys/netpfil/pf/pf_syncookies.c
505
seq = ntohl(pd->hdr.tcp.th_seq) - 1;
sys/netpfil/pf/pf_syncookies.c
506
ack = ntohl(pd->hdr.tcp.th_ack) - 1;
sys/netpfil/pf/pf_syncookies.c
516
return (pf_build_tcp(NULL, pd->af, pd->src, pd->dst, *pd->sport,
sys/netpfil/pf/pf_syncookies.c
517
*pd->dport, seq, 0, TH_SYN, wscale, mss, pd->ttl,
sys/netpfil/pf/pf_syncookies.c
518
(pd->m->m_flags & M_LOOP), 0, PF_MTAG_FLAG_SYNCOOKIE_RECREATED,
sys/netpfil/pf/pf_syncookies.c
519
cookie.flags.sack_ok, pd->act.rtableid, reason));
sys/ofed/drivers/infiniband/core/core_priv.h
149
struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/core_priv.h
159
qp = dev->create_qp(pd, attr, udata);
sys/ofed/drivers/infiniband/core/core_priv.h
164
qp->pd = pd;
sys/ofed/drivers/infiniband/core/ib_agent.c
104
ah = ib_create_ah_from_wc(agent->qp->pd, wc, grh, port_num);
sys/ofed/drivers/infiniband/core/ib_cm.c
352
ah = ib_create_ah(mad_agent->qp->pd, &av->ah_attr, 0);
sys/ofed/drivers/infiniband/core/ib_cm.c
397
ah = ib_create_ah_from_wc(port->mad_agent->qp->pd, mad_recv_wc->wc,
sys/ofed/drivers/infiniband/core/ib_cma.c
836
int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_cma.c
844
if (id->device != pd->device)
sys/ofed/drivers/infiniband/core/ib_cma.c
848
qp = ib_create_qp(pd, qp_init_attr);
sys/ofed/drivers/infiniband/core/ib_fmr_pool.c
213
struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_fmr_pool.c
225
device = pd->device;
sys/ofed/drivers/infiniband/core/ib_fmr_pool.c
304
fmr->fmr = ib_alloc_fmr(pd, params->access, &fmr_attr);
sys/ofed/drivers/infiniband/core/ib_mad.c
1033
mad_send_wr->sg_list[0].lkey = mad_agent->qp->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/core/ib_mad.c
1042
mad_send_wr->sg_list[1].lkey = mad_agent->qp->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/core/ib_mad.c
2848
sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/core/ib_mad.c
3078
qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr);
sys/ofed/drivers/infiniband/core/ib_mad.c
3141
port_priv->pd = ib_alloc_pd(device, 0);
sys/ofed/drivers/infiniband/core/ib_mad.c
3142
if (IS_ERR(port_priv->pd)) {
sys/ofed/drivers/infiniband/core/ib_mad.c
3144
ret = PTR_ERR(port_priv->pd);
sys/ofed/drivers/infiniband/core/ib_mad.c
3199
ib_dealloc_pd(port_priv->pd);
sys/ofed/drivers/infiniband/core/ib_mad.c
3230
ib_dealloc_pd(port_priv->pd);
sys/ofed/drivers/infiniband/core/ib_mad_rmpp.c
166
ah = ib_create_ah_from_wc(agent->qp->pd, recv_wc->wc,
sys/ofed/drivers/infiniband/core/ib_mad_rmpp.c
297
rmpp_recv->ah = ib_create_ah_from_wc(agent->agent.qp->pd,
sys/ofed/drivers/infiniband/core/ib_sa_query.c
536
new_ah->ah = ib_create_ah(port->agent->qp->pd, &ah_attr, RDMA_CREATE_AH_SLEEPABLE);
sys/ofed/drivers/infiniband/core/ib_user_mad.c
510
ah = ib_create_user_ah(agent->qp->pd, &ah_attr, NULL);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1275
struct ib_pd *pd = NULL;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1366
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle,
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1368
if (!pd || (!scq && has_sq)) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1373
device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1418
qp = ib_create_qp(pd, &attr);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1420
qp = _ib_create_qp(device, pd, &attr, &attrs->driver_udata,
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1429
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1466
if (pd)
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1467
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1488
if (pd)
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
1489
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2400
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2419
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2420
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2444
ah = ib_create_user_ah(pd, &attr, &attrs->driver_udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2460
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2469
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2903
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2920
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2921
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2941
wq = pd->device->create_wq(pd, &wq_init_attr, &attrs->driver_udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2951
wq->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2952
wq->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2955
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2970
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
2982
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3373
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3415
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3416
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3436
srq->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3437
srq->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3443
ret = pd->device->create_srq(srq, &attr, udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3457
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3481
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
3492
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
422
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
434
pd = rdma_zalloc_drv_obj(ib_dev, ib_pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
435
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
440
pd->device = ib_dev;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
441
pd->uobject = uobj;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
442
pd->__internal_mr = NULL;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
443
atomic_set(&pd->usecnt, 0);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
445
ret = ib_dev->alloc_pd(pd, &attrs->driver_udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
449
uobj->object = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
461
ib_dealloc_pd_user(pd, uverbs_get_cleared_udata(attrs));
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
462
pd = NULL;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
464
kfree(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
717
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
737
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
738
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
744
if (!(pd->device->attrs.device_cap_flags &
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
752
mr = pd->device->reg_user_mr(pd, cmd.start, cmd.length, cmd.hca_va,
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
760
mr->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
761
mr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
766
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
779
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
788
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
799
struct ib_pd *pd = NULL;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
835
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle,
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
837
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
843
old_pd = mr->pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
846
cmd.access_flags, pd,
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
852
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
853
mr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
865
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
890
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
903
pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
904
if (!pd) {
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
914
mw = pd->device->alloc_mw(pd, cmd.mw_type, &attrs->driver_udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
920
mw->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
921
mw->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
923
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
935
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_cmd.c
942
uobj_put_obj_read(pd);
sys/ofed/drivers/infiniband/core/ib_uverbs_main.c
108
struct ib_pd *pd = mw->pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_main.c
113
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types.c
192
struct ib_pd *pd = uobject->object;
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types.c
195
ret = ib_destroy_usecnt(&pd->usecnt, why, uobject);
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types.c
199
ib_dealloc_pd_user(pd, &attrs->driver_udata);
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
125
mr = pd->device->reg_dm_mr(pd, dm, &attr, attrs);
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
129
mr->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
130
mr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
134
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
48
struct ib_pd *pd =
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
51
struct ib_device *ib_dev = pd->device;
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
77
return ib_dev->advise_mr(pd, advice, flags, sg_list, num_sge,
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
89
struct ib_pd *pd =
sys/ofed/drivers/infiniband/core/ib_uverbs_std_types_mr.c
91
struct ib_device *ib_dev = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1488
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1500
pd = qp->pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1508
if (pd)
sys/ofed/drivers/infiniband/core/ib_verbs.c
1509
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1586
struct ib_pd *pd = mr->pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1593
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1616
struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/ofed/drivers/infiniband/core/ib_verbs.c
1621
if (!pd->device->alloc_mr) {
sys/ofed/drivers/infiniband/core/ib_verbs.c
1632
mr = pd->device->alloc_mr(pd, mr_type, max_num_sg, udata);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1634
mr->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1635
mr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1638
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1651
struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
1657
if (!pd->device->alloc_fmr)
sys/ofed/drivers/infiniband/core/ib_verbs.c
1660
fmr = pd->device->alloc_fmr(pd, mr_access_flags, fmr_attr);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1662
fmr->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1663
fmr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1664
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1685
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1688
pd = fmr->pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1691
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1827
struct ib_wq *ib_create_wq(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
1832
if (!pd->device->create_wq)
sys/ofed/drivers/infiniband/core/ib_verbs.c
1835
wq = pd->device->create_wq(pd, wq_attr, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1841
wq->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1842
wq->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1844
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
1860
struct ib_pd *pd = wq->pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
1866
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
267
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
271
pd = rdma_zalloc_drv_obj(device, ib_pd);
sys/ofed/drivers/infiniband/core/ib_verbs.c
272
if (!pd)
sys/ofed/drivers/infiniband/core/ib_verbs.c
275
pd->device = device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
276
pd->uobject = NULL;
sys/ofed/drivers/infiniband/core/ib_verbs.c
277
pd->__internal_mr = NULL;
sys/ofed/drivers/infiniband/core/ib_verbs.c
278
atomic_set(&pd->usecnt, 0);
sys/ofed/drivers/infiniband/core/ib_verbs.c
279
pd->flags = flags;
sys/ofed/drivers/infiniband/core/ib_verbs.c
281
ret = device->alloc_pd(pd, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
283
kfree(pd);
sys/ofed/drivers/infiniband/core/ib_verbs.c
288
pd->local_dma_lkey = device->local_dma_lkey;
sys/ofed/drivers/infiniband/core/ib_verbs.c
300
mr = pd->device->get_dma_mr(pd, mr_access_flags);
sys/ofed/drivers/infiniband/core/ib_verbs.c
302
ib_dealloc_pd(pd);
sys/ofed/drivers/infiniband/core/ib_verbs.c
306
mr->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
307
mr->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
312
pd->__internal_mr = mr;
sys/ofed/drivers/infiniband/core/ib_verbs.c
315
pd->local_dma_lkey = pd->__internal_mr->lkey;
sys/ofed/drivers/infiniband/core/ib_verbs.c
318
pd->unsafe_global_rkey = pd->__internal_mr->rkey;
sys/ofed/drivers/infiniband/core/ib_verbs.c
321
return pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
334
void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata)
sys/ofed/drivers/infiniband/core/ib_verbs.c
338
if (pd->__internal_mr) {
sys/ofed/drivers/infiniband/core/ib_verbs.c
339
ret = pd->device->dereg_mr(pd->__internal_mr, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
341
pd->__internal_mr = NULL;
sys/ofed/drivers/infiniband/core/ib_verbs.c
346
WARN_ON(atomic_read(&pd->usecnt));
sys/ofed/drivers/infiniband/core/ib_verbs.c
348
pd->device->dealloc_pd(pd, udata);
sys/ofed/drivers/infiniband/core/ib_verbs.c
349
kfree(pd);
sys/ofed/drivers/infiniband/core/ib_verbs.c
355
static struct ib_ah *_ib_create_ah(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
360
struct ib_device *device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
376
ah->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
384
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
399
struct ib_ah *ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr,
sys/ofed/drivers/infiniband/core/ib_verbs.c
404
ah = _ib_create_ah(pd, ah_attr, flags, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
424
struct ib_ah *ib_create_user_ah(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
430
if (rdma_protocol_roce(pd->device, ah_attr->port_num)) {
sys/ofed/drivers/infiniband/core/ib_verbs.c
431
err = ib_resolve_eth_dmac(pd->device, ah_attr);
sys/ofed/drivers/infiniband/core/ib_verbs.c
436
return _ib_create_ah(pd, ah_attr, RDMA_CREATE_AH_SLEEPABLE, udata);
sys/ofed/drivers/infiniband/core/ib_verbs.c
656
struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc,
sys/ofed/drivers/infiniband/core/ib_verbs.c
662
ret = ib_init_ah_from_wc(pd->device, port_num, wc, grh, &ah_attr);
sys/ofed/drivers/infiniband/core/ib_verbs.c
666
return ib_create_ah(pd, &ah_attr, RDMA_CREATE_AH_SLEEPABLE);
sys/ofed/drivers/infiniband/core/ib_verbs.c
688
struct ib_pd *pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
692
pd = ah->pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
694
atomic_dec(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
703
struct ib_srq *ib_create_srq(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
709
if (!pd->device->create_srq)
sys/ofed/drivers/infiniband/core/ib_verbs.c
712
srq = rdma_zalloc_drv_obj(pd->device, ib_srq);
sys/ofed/drivers/infiniband/core/ib_verbs.c
716
srq->device = pd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
717
srq->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
730
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
732
ret = pd->device->create_srq(srq, srq_init_attr, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
734
atomic_dec(&srq->pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
772
atomic_dec(&srq->pd->usecnt);
sys/ofed/drivers/infiniband/core/ib_verbs.c
859
qp->pd = NULL;
sys/ofed/drivers/infiniband/core/ib_verbs.c
875
struct ib_qp *ib_create_qp(struct ib_pd *pd,
sys/ofed/drivers/infiniband/core/ib_verbs.c
878
struct ib_device *device = pd ? pd->device : qp_init_attr->xrcd->device;
sys/ofed/drivers/infiniband/core/ib_verbs.c
887
qp = _ib_create_qp(device, pd, qp_init_attr, NULL, NULL);
sys/ofed/drivers/infiniband/core/ib_verbs.c
917
qp->pd = pd;
sys/ofed/drivers/infiniband/core/ib_verbs.c
921
atomic_inc(&pd->usecnt);
sys/ofed/drivers/infiniband/core/mad_priv.h
204
struct ib_pd *pd;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
353
struct ib_pd *pd;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
443
struct ib_pd *pd, struct ib_ah_attr *attr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1361
priv->cm.srq = ib_create_srq(priv->pd, &srq_init_attr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
229
return ib_create_qp(priv->pd, &attr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
293
sge[i].lkey = priv->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
939
return ib_create_qp(priv->pd, &attr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
59
struct ib_pd *pd, struct ib_ah_attr *attr)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
71
ah->ah = ib_create_ah(pd, attr, RDMA_CREATE_AH_SLEEPABLE);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
533
ah = ipoib_create_ah(priv, priv->pd, &av);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
213
ah = ipoib_create_ah(priv, priv->pd, &av);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
150
priv->pd = ib_alloc_pd(priv->ca, 0);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
151
if (IS_ERR(priv->pd)) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
208
priv->qp = ib_create_qp(priv->pd, &init_attr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
220
priv->tx_sge[i].lkey = priv->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
227
priv->rx_sge[i].lkey = priv->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
242
ib_dealloc_pd(priv->pd);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_verbs.c
263
ib_dealloc_pd(priv->pd);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
297
struct ib_pd *pd;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_cma.c
120
rc = rdma_create_qp(id, ssk->sdp_dev->pd, &qp_init_attr);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1695
sdp_dev->pd = ib_alloc_pd(device, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1696
if (IS_ERR(sdp_dev->pd))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1705
sdp_dev->fmr_pool = ib_create_fmr_pool(sdp_dev->pd, &param);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1712
ib_dealloc_pd(sdp_dev->pd);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1742
ib_dealloc_pd(sdp_dev->pd);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
137
sge->lkey = ssk->sdp_dev->pd->local_dma_lkey;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
136
sge->lkey = ssk->sdp_dev->pd->local_dma_lkey;
sys/ofed/include/rdma/ib_fmr_pool.h
81
struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
1454
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1480
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1525
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1584
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1617
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1636
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
1644
struct ib_pd *pd;
sys/ofed/include/rdma/ib_verbs.h
2228
int (*alloc_pd)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2230
void (*dealloc_pd)(struct ib_pd *pd, struct ib_udata *udata);
sys/ofed/include/rdma/ib_verbs.h
2251
struct ib_qp * (*create_qp)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2284
struct ib_mr * (*get_dma_mr)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2286
struct ib_mr * (*reg_user_mr)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2296
struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2299
struct ib_mr * (*alloc_mr)(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/ofed/include/rdma/ib_verbs.h
2301
int (*advise_mr)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2309
struct ib_mw * (*alloc_mw)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2313
struct ib_fmr * (*alloc_fmr)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2374
struct ib_wq * (*create_wq)(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
2391
struct ib_mr *(*reg_dm_mr)(struct ib_pd *pd, struct ib_dm *dm,
sys/ofed/include/rdma/ib_verbs.h
2976
void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata);
sys/ofed/include/rdma/ib_verbs.h
2984
static inline void ib_dealloc_pd(struct ib_pd *pd)
sys/ofed/include/rdma/ib_verbs.h
2986
ib_dealloc_pd_user(pd, NULL);
sys/ofed/include/rdma/ib_verbs.h
3003
struct ib_ah *ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr,
sys/ofed/include/rdma/ib_verbs.h
3018
struct ib_ah *ib_create_user_ah(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3049
struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc,
sys/ofed/include/rdma/ib_verbs.h
3108
struct ib_srq *ib_create_srq(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3176
struct ib_qp *ib_create_qp(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3776
struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type,
sys/ofed/include/rdma/ib_verbs.h
3779
static inline struct ib_mr *ib_alloc_mr(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3782
return ib_alloc_mr_user(pd, mr_type, max_num_sg, NULL);
sys/ofed/include/rdma/ib_verbs.h
3785
struct ib_mr *ib_alloc_mr_integrity(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3821
struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd,
sys/ofed/include/rdma/ib_verbs.h
3937
struct ib_wq *ib_create_wq(struct ib_pd *pd,
sys/ofed/include/rdma/rdma_cm.h
232
int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd,
sys/ofed/include/rdma/rdma_vt.h
497
int rvt_lkey_ok(struct rvt_lkey_table *rkt, struct rvt_pd *pd,
sys/ofed/include/rdma/rdmavt_mr.h
74
struct ib_pd *pd; /* shares refcnt of ibmr.pd */
sys/powerpc/aim/mmu_radix.c
4436
pml3_entry_t *pd;
sys/powerpc/aim/mmu_radix.c
4447
pd = pmap_pml3e(pmap, va);
sys/powerpc/aim/mmu_radix.c
4453
if (pd != NULL && (be64toh(*pd) & (RPTE_LEAF | PG_V)) == (RPTE_LEAF | PG_V)) {
sys/powerpc/aim/mmu_radix.c
4454
if (!pmap_demote_l3e_locked(pmap, pd, va, lockp)) {
sys/powerpc/aim/mmu_radix.c
4459
pd = NULL;
sys/powerpc/aim/mmu_radix.c
4467
if (pd != NULL && (be64toh(*pd) & PG_V) != 0) {
sys/powerpc/aim/mmu_radix.c
4468
m = PHYS_TO_VM_PAGE(be64toh(*pd) & PG_FRAME);
sys/sys/procdesc.h
79
#define PROCDESC_LOCK_DESTROY(pd) mtx_destroy(&(pd)->pd_lock)
sys/sys/procdesc.h
80
#define PROCDESC_LOCK_INIT(pd) mtx_init(&(pd)->pd_lock, "procdesc", NULL, \
sys/sys/procdesc.h
82
#define PROCDESC_LOCK(pd) mtx_lock(&(pd)->pd_lock)
sys/sys/procdesc.h
83
#define PROCDESC_UNLOCK(pd) mtx_unlock(&(pd)->pd_lock)
sys/teken/teken.h
104
#define TP_SETBELLPD_PITCH(pd) ((pd) >> 16)
sys/teken/teken.h
105
#define TP_SETBELLPD_DURATION(pd) ((pd) & 0xffff)
tests/sys/capsicum/linux.cc
719
int pd;
tests/sys/capsicum/linux.cc
720
pid_t child = pdfork(&pd, 0);
tests/sys/capsicum/linux.cc
736
pd, child, ProcessState(child));
tests/sys/capsicum/linux.cc
739
EXPECT_OK(pdgetpid(pd, &pid));
tests/sys/capsicum/linux.cc
745
SendFD(pd, shared_sock_fds[1]);
tests/sys/capsicum/linux.cc
841
int pd;
tests/sys/capsicum/linux.cc
842
pid_t child = pdfork(&pd, 0);
tests/sys/capsicum/linux.cc
856
pd, child, ProcessState(child));
tests/sys/capsicum/linux.cc
860
SendFD(pd, shared_sock_fds[1]);
tests/sys/capsicum/linux.cc
861
close(pd);
tests/sys/capsicum/procdesc.cc
1007
int pd = *(int*)CMSG_DATA(cmptr);
tests/sys/capsicum/procdesc.cc
1011
if (verbose) fprintf(stderr, " [%d] got process descriptor %d on socket\n", getpid_(), pd);
tests/sys/capsicum/procdesc.cc
1016
EXPECT_OK(pdgetpid(pd, &other));
tests/sys/capsicum/procdesc.cc
1017
if (verbose) fprintf(stderr, " [%d] process descriptor %d is pid %d\n", getpid_(), pd, other);
tests/sys/capsicum/procdesc.cc
102
int pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
1022
if (verbose) fprintf(stderr, " [%d] close process descriptor %d\n", getpid_(), pd);
tests/sys/capsicum/procdesc.cc
1023
close(pd);
tests/sys/capsicum/procdesc.cc
106
EXPECT_EQ(-1, pd);
tests/sys/capsicum/procdesc.cc
120
EXPECT_NE(-1, pd);
tests/sys/capsicum/procdesc.cc
123
EXPECT_OK(pdgetpid(pd, &pid_got));
tests/sys/capsicum/procdesc.cc
137
int waitrc = pdwait4_(pd, &status, 0, &ru);
tests/sys/capsicum/procdesc.cc
140
fprintf(stderr, "For pd %d pid %d:\n", pd, pid);
tests/sys/capsicum/procdesc.cc
147
EXPECT_EQ(-1, pdwait4_(pd, &status, 0, &ru));
tests/sys/capsicum/procdesc.cc
150
EXPECT_OK(close(pd));
tests/sys/capsicum/procdesc.cc
154
int pd = -1;
tests/sys/capsicum/procdesc.cc
155
int pid = pdfork(&pd, PD_DAEMON<<5);
tests/sys/capsicum/procdesc.cc
169
int pd = -1;
tests/sys/capsicum/procdesc.cc
170
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
174
EXPECT_EQ(-1, pdgetpid(pd, &pid));
tests/sys/capsicum/procdesc.cc
182
EXPECT_OK(fstat(pd, &stat));
tests/sys/capsicum/procdesc.cc
193
EXPECT_OK(pdgetpid(pd, &pd_pid));
tests/sys/capsicum/procdesc.cc
199
int pd = -1;
tests/sys/capsicum/procdesc.cc
200
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
233
int pd;
tests/sys/capsicum/procdesc.cc
236
pid_t child = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
26
static pid_t pdwait4_(int pd, int *status, int options, struct rusage *ru) {
tests/sys/capsicum/procdesc.cc
262
if (verbose) fprintf(stderr, " thread generated pd %d\n", pd);
tests/sys/capsicum/procdesc.cc
268
return reinterpret_cast<void *>(pd);
tests/sys/capsicum/procdesc.cc
277
int pd = reinterpret_cast<intptr_t>(data);
tests/sys/capsicum/procdesc.cc
278
if (verbose) fprintf(stderr, "retrieved pd %d from terminated thread\n", pd);
tests/sys/capsicum/procdesc.cc
28
return pdwait4(pd, status, options, ru);
tests/sys/capsicum/procdesc.cc
282
EXPECT_OK(pdgetpid(pd, &pid));
tests/sys/capsicum/procdesc.cc
283
EXPECT_OK(pdkill(pd, SIGKILL));
tests/sys/capsicum/procdesc.cc
285
EXPECT_EQ(pid, pdwait4_(pd, &status, 0, NULL));
tests/sys/capsicum/procdesc.cc
32
int rc = pdgetpid(pd, &pid);
tests/sys/capsicum/procdesc.cc
539
int pd;
tests/sys/capsicum/procdesc.cc
541
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
570
EXPECT_OK(close(pd));
tests/sys/capsicum/procdesc.cc
573
EXPECT_SYSCALL_FAIL(EBADF, pdwait4_(pd, &status, WNOHANG, NULL));
tests/sys/capsicum/procdesc.cc
576
EXPECT_OK(pdkill(pd, SIGKILL));
tests/sys/capsicum/procdesc.cc
579
int rc = pdwait4_(pd, &status, WNOHANG, NULL);
tests/sys/capsicum/procdesc.cc
626
static void TestPdkill(pid_t pid, int pd) {
tests/sys/capsicum/procdesc.cc
629
EXPECT_OK(pdkill(pd, SIGCONT));
tests/sys/capsicum/procdesc.cc
633
EXPECT_OK(pdkill(pd, SIGINT));
tests/sys/capsicum/procdesc.cc
638
EXPECT_EQ(0, pdkill(pd, SIGINT));
tests/sys/capsicum/procdesc.cc
643
EXPECT_EQ(-1, pdkill(pd, SIGINT));
tests/sys/capsicum/procdesc.cc
656
int pd = -1;
tests/sys/capsicum/procdesc.cc
659
int pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
680
EXPECT_EQ(-1, pdkill(pd, 0xFFFF));
tests/sys/capsicum/procdesc.cc
685
pdkill(pd, SIGUSR1);
tests/sys/capsicum/procdesc.cc
709
int pd;
tests/sys/capsicum/procdesc.cc
713
pid_t grandchild = pdfork(&pd, pdfork_flags);
tests/sys/capsicum/procdesc.cc
840
int pd;
tests/sys/capsicum/procdesc.cc
843
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
867
EXPECT_OK(pdgetpid(pd, &pd_pid));
tests/sys/capsicum/procdesc.cc
874
EXPECT_OK(pdkill(pd, SIGINT));
tests/sys/capsicum/procdesc.cc
882
int pd;
tests/sys/capsicum/procdesc.cc
885
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
905
EXPECT_OK(pdkill(pd, SIGINT));
tests/sys/capsicum/procdesc.cc
939
int pd = -1;
tests/sys/capsicum/procdesc.cc
940
pid_t pid = pdfork(&pd, 0);
tests/sys/capsicum/procdesc.cc
950
int cap_incapable = dup(pd);
tests/sys/capsicum/procdesc.cc
955
int cap_capable = dup(pd);
tests/sys/capsicum/procdesc.cc
968
int rc = pdwait4_(pd, &status, 0, NULL);
tests/sys/capsicum/procdesc.cc
98
int pd = -1;
tests/sys/capsicum/smoketest.c
53
int pd = -1;
tests/sys/capsicum/smoketest.c
54
rc = pdfork(&pd, 0);
tests/sys/capsicum/smoketest.c
66
fprintf(stderr, "[%d] pdfork() rc=%d pd=%d\n", getpid(), rc, pd);
tests/sys/capsicum/smoketest.c
71
rc = pdgetpid(pd, &got_pid);
tests/sys/capsicum/smoketest.c
72
if (rc < 0) fprintf(stderr, "*** pdgetpid(pd=%d) failed: errno=%d %s\n", pd, errno, strerror(errno));
tests/sys/capsicum/smoketest.c
73
fprintf(stderr, "[%d] pdgetpid(pd=%d)=%d, pdfork returned %d\n", getpid(), pd, got_pid, actual_pid);
tests/sys/capsicum/smoketest.c
78
rc = pdkill(pd, SIGKILL);
tests/sys/capsicum/smoketest.c
79
fprintf(stderr, "[%d] pdkill(pd=%d, SIGKILL) -> rc=%d\n", getpid(), pd, rc);
tests/sys/kern/pdrfork.c
101
close(pd);
tests/sys/kern/pdrfork.c
130
int pd = -1;
tests/sys/kern/pdrfork.c
132
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, -1, RFSPAWN) < 0);
tests/sys/kern/pdrfork.c
133
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, -1) < 0);
tests/sys/kern/pdrfork.c
134
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, RFSPAWN | RFNOWAIT) < 0);
tests/sys/kern/pdrfork.c
135
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, RFPROC | RFFDG| RFCFDG) < 0);
tests/sys/kern/pdrfork.c
136
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, RFPROCDESC) < 0);
tests/sys/kern/pdrfork.c
137
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, RFPROC) < 0);
tests/sys/kern/pdrfork.c
138
ATF_CHECK_ERRNO(EINVAL, pdrfork(&pd, 0, 0) < 0);
tests/sys/kern/pdrfork.c
160
int pd = -1;
tests/sys/kern/pdrfork.c
170
pid = pdrfork_thread(&pd, 0, RFSPAWN, stack, rfspawn_fn, NULL);
tests/sys/kern/pdrfork.c
171
basic_usage_tail(pd, pid);
tests/sys/kern/pdrfork.c
40
basic_usage_tail(int pd, pid_t pid)
tests/sys/kern/pdrfork.c
45
ATF_REQUIRE_MSG(pd >= 0, "rfork did not return a process descriptor");
tests/sys/kern/pdrfork.c
46
r = pdgetpid(pd, &pd_pid);
tests/sys/kern/pdrfork.c
55
close(pd);
tests/sys/kern/pdrfork.c
64
int pd = -1;
tests/sys/kern/pdrfork.c
67
pid = pdrfork(&pd, 0, rfflags);
tests/sys/kern/pdrfork.c
73
basic_usage_tail(pd, pid);
tests/sys/kern/pdrfork.c
80
int pd = -1;
tests/sys/kern/pdrfork.c
84
pid = pdrfork(&pd, 0, RFPROC | RFPROCDESC);
tests/sys/kern/pdrfork.c
91
_exit(pd == -1);
tests/sys/kern/pdrfork.c
93
ATF_REQUIRE_MSG(pd >= 0, "rfork did not return a process descriptor");
tests/sys/kern/pdrfork.c
94
r = pdgetpid(pd, &pd_pid);
tests/sys/kern/prace.c
102
(void)read(pd[0], &pd[0], sizeof(pd[0]));
tests/sys/kern/prace.c
103
close(pd[0]);
tests/sys/kern/prace.c
86
int pd[2], status;
tests/sys/kern/prace.c
90
if (pipe(pd) != 0)
tests/sys/kern/prace.c
95
close(pd[0]);
tests/sys/kern/prace.c
96
child(pd[1], poll);
tests/sys/kern/prace.c
99
close(pd[1]);
tests/sys/kern/procdesc.c
53
int i, pd, pid_max;
tests/sys/kern/procdesc.c
62
dangle_pid = pdfork(&pd, PD_CLOEXEC | PD_DAEMON);
tests/sys/kern/procdesc.c
90
close(pd);
tests/sys/kern/ptrace_test.c
4223
int pd, status;
tests/sys/kern/ptrace_test.c
4225
traced = pdfork(&pd, 0);
tests/sys/kern/ptrace_test.c
4231
ATF_REQUIRE(pd >= 0);
tests/sys/kern/ptrace_test.c
4242
REQUIRE_EQ(close(pd), 0);
tests/sys/kern/ptrace_test.c
4253
REQUIRE_EQ(close(pd), 0);
tests/sys/kern/ptrace_test.c
4272
int pd, status;
tests/sys/kern/ptrace_test.c
4274
child = pdfork(&pd, 0);
tests/sys/kern/ptrace_test.c
4305
ATF_REQUIRE(close(pd) != -1);
tests/sys/kern/ptrace_test.c
4316
int pd, status;
tests/sys/kern/ptrace_test.c
4318
traced = pdfork(&pd, 0);
tests/sys/kern/ptrace_test.c
4324
ATF_REQUIRE(pd >= 0);
tests/sys/kern/ptrace_test.c
4346
REQUIRE_EQ(close(pd), 0);
tests/sys/kern/ptrace_test.c
4362
REQUIRE_EQ(close(pd), 0);
tests/sys/kern/reaper.c
747
int pd, r, status;
tests/sys/kern/reaper.c
753
child = pdfork(&pd, 0);
tests/sys/kern/reaper.c
756
grandchild = pdfork(&pd, 0);
tools/regression/security/cap_test/cap_test_pdfork.c
58
int pd, error;
tools/regression/security/cap_test/cap_test_pdfork.c
64
pid = pdfork(&pd, 0);
tools/regression/security/cap_test/cap_test_pdfork.c
74
error = pdgetpid(pd, &pid);
tools/regression/security/cap_test/cap_test_pdfork.c
84
REQUIRE(fstat(pd, &stat));
tools/regression/security/cap_test/cap_test_pdfork.c
96
error = pdgetpid(pd, &pid);
tools/regression/security/cap_test/cap_test_pdkill.c
61
int pd, error;
tools/regression/security/cap_test/cap_test_pdkill.c
66
error = pdfork(&pd, 0);
tools/regression/security/cap_test/cap_test_pdkill.c
77
error = pdgetpid(pd, &pid);
tools/regression/security/cap_test/cap_test_pdkill.c
83
error = pdkill(pd, SIGINT);
usr.bin/calendar/dates.c
149
pd = pm->days;
usr.bin/calendar/dates.c
150
while (pd != NULL) {
usr.bin/calendar/dates.c
151
pdp = pd;
usr.bin/calendar/dates.c
152
pd = pd->nextday;
usr.bin/calendar/dates.c
155
if (pd == NULL) { /* Always true */
usr.bin/calendar/dates.c
156
pd = (struct cal_day *)calloc(1, sizeof(struct cal_day));
usr.bin/calendar/dates.c
157
pd->month = pm;
usr.bin/calendar/dates.c
158
pd->year = py;
usr.bin/calendar/dates.c
159
pd->dayofmonth = d;
usr.bin/calendar/dates.c
160
pd->julianday = pm->firstdayjulian + d - 1;
usr.bin/calendar/dates.c
161
pd->dayofweek = (pm->firstdayofweek + d - 1) % 7;
usr.bin/calendar/dates.c
163
pdp->nextday = pd;
usr.bin/calendar/dates.c
166
pm->days = pd;
usr.bin/calendar/dates.c
90
struct cal_day *pd, *pdp;
usr.bin/diff3/diff3.c
312
int pd, error;
usr.bin/diff3/diff3.c
319
posix_spawnattr_setprocdescp_np(&sa, &pd, 0);
usr.bin/diff3/diff3.c
329
return (pd);
usr.bin/diff3/diff3.c
939
wait_and_check(int pd)
usr.bin/diff3/diff3.c
943
while (pdwait(pd, &status, WEXITED, NULL, NULL) == -1) {
usr.bin/diff3/diff3.c
947
close(pd);
usr.bin/rpcgen/rpc_main.c
385
int pd[2];
usr.bin/rpcgen/rpc_main.c
388
(void) pipe(pd);
usr.bin/rpcgen/rpc_main.c
397
(void) dup2(pd[1], 1);
usr.bin/rpcgen/rpc_main.c
398
(void) close(pd[0]);
usr.bin/rpcgen/rpc_main.c
404
(void) close(pd[1]);
usr.bin/rpcgen/rpc_main.c
405
fin = fdopen(pd[0], "r");
usr.bin/sed/process.c
105
pd = 0;
usr.bin/sed/process.c
133
pd = 1;
usr.bin/sed/process.c
139
pd = 1;
usr.bin/sed/process.c
142
if (pd)
usr.bin/sed/process.c
146
pd = 1;
usr.bin/sed/process.c
174
if (!nflag && !pd)
usr.bin/sed/process.c
179
pd = 0;
usr.bin/sed/process.c
188
if (pd)
usr.bin/sed/process.c
193
if (pd)
usr.bin/sed/process.c
209
if (!nflag && !pd)
usr.bin/sed/process.c
238
if (pd)
usr.bin/sed/process.c
263
if (pd || psl == 0)
usr.bin/sed/process.c
276
new: if (!nflag && !pd)
usr.bin/sed/process.c
475
if (cp->u.s->wfile && !pd) {
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2875
uint8_t data[32], *pd = NULL;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2889
pd = data;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2894
pd = data;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2903
if (wlan_ioctl(wif->wname, op, &val, pd, &argsize, 0) < 0)
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2945
uint8_t data[32], *pd = NULL;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2968
pd = data;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2977
pd = data;
usr.sbin/bsnmpd/modules/snmp_wlan/wlan_sys.c
2984
if (wlan_ioctl(wif->wname, op, &val, pd, &argsize, 1) < 0)
usr.sbin/cxgbetool/cxgbetool.c
3461
parse_offload_policy_line(size_t lno, char *line, size_t llen, pcap_t *pd,
usr.sbin/cxgbetool/cxgbetool.c
3524
if (pcap_compile(pd, &r->bpf_prog, expr, 1,
usr.sbin/cxgbetool/cxgbetool.c
3527
lno, pcap_geterr(pd));
usr.sbin/cxgbetool/cxgbetool.c
3577
pcap_t *pd;
usr.sbin/cxgbetool/cxgbetool.c
3584
pd = pcap_open_dead(DLT_EN10MB, 128);
usr.sbin/cxgbetool/cxgbetool.c
3585
if (pd == NULL) {
usr.sbin/cxgbetool/cxgbetool.c
3620
rc = parse_offload_policy_line(lno, line, llen, pd, r);
usr.sbin/cxgbetool/cxgbetool.c
3642
pcap_close(pd);
usr.sbin/mfiutil/mfi_config.c
1136
device_id = ar->pd[j].ref.v.device_id;
usr.sbin/mfiutil/mfi_config.c
1141
mfi_pdstate(ar->pd[j].fw_state));
usr.sbin/mfiutil/mfi_config.c
442
ar->pd[i].ref = array_info->drives[i].ref;
usr.sbin/mfiutil/mfi_config.c
443
ar->pd[i].fw_state = MFI_PD_STATE_ONLINE;
usr.sbin/mfiutil/mfi_evt.c
342
pdrive_location(struct mfi_evt_pd *pd)
usr.sbin/mfiutil/mfi_evt.c
346
if (pd->enclosure_index == 0)
usr.sbin/mfiutil/mfi_evt.c
347
snprintf(buffer, sizeof(buffer), "%02d(s%d)", pd->device_id,
usr.sbin/mfiutil/mfi_evt.c
348
pd->slot_number);
usr.sbin/mfiutil/mfi_evt.c
350
snprintf(buffer, sizeof(buffer), "%02d(e%d/s%d)", pd->device_id,
usr.sbin/mfiutil/mfi_evt.c
351
pd->enclosure_index, pd->slot_number);
usr.sbin/mfiutil/mfi_evt.c
376
pdrive_location(&detail->args.cdb_sense.pd)
usr.sbin/mfiutil/mfi_evt.c
419
pdrive_location(&detail->args.ld_lba_pd_lba.pd),
usr.sbin/mfiutil/mfi_evt.c
453
pdrive_location(&detail->args.pd));
usr.sbin/mfiutil/mfi_evt.c
459
pdrive_location(&detail->args.pd_err.pd),
usr.sbin/mfiutil/mfi_evt.c
466
pdrive_location(&detail->args.pd_lba.pd),
usr.sbin/mfiutil/mfi_evt.c
473
pdrive_location(&detail->args.pd_lba_ld.pd),
usr.sbin/mfiutil/mfi_evt.c
481
pdrive_location(&detail->args.pd_prog.pd),
usr.sbin/mfiutil/mfi_evt.c
489
pdrive_location(&detail->args.pd_state.pd),
usr.sbin/mfiutil/mfi_foreign.c
192
device_id = ar->pd[k].ref.v.device_id;
usr.sbin/mfiutil/mfi_foreign.c
197
mfi_pdstate(ar->pd[k].fw_state));
usr.sbin/mfiutil/mfi_show.c
386
device_id = ar->pd[j].ref.v.device_id;
usr.sbin/mfiutil/mfi_show.c
394
mfi_pdstate(ar->pd[j].fw_state));
usr.sbin/mlxcontrol/command.c
292
struct mlx_phys_drv pd;
usr.sbin/mlxcontrol/command.c
369
if ((mlx_get_device_state(unit, channel, target, &pd) == 0) &&
usr.sbin/mlxcontrol/command.c
370
(pd.pd_flags1 & MLX_PHYS_DRV_PRESENT)) {
usr.sbin/mlxcontrol/command.c
371
mlx_print_phys_drv(&pd, channel, target, " ", verbosity - 1);
usr.sbin/mptutil/mpt_config.c
1085
MPT_TABLE(top, pd);
usr.sbin/mptutil/mpt_config.c
1134
MPT_COMMAND(pd, create, pd_create);
usr.sbin/mptutil/mpt_config.c
1188
MPT_COMMAND(pd, delete, pd_delete);
usr.sbin/mptutil/mpt_show.c
583
MPT_COMMAND(show, pd, show_physdisks);
usr.sbin/pmccontrol/pmccontrol.c
217
struct pmc_info *pd;
usr.sbin/pmccontrol/pmccontrol.c
253
pd = &pi->pm_pmcs[n];
usr.sbin/pmccontrol/pmccontrol.c
257
pd->pm_name,
usr.sbin/pmccontrol/pmccontrol.c
258
pmc_name_of_class(pd->pm_class),
usr.sbin/pmccontrol/pmccontrol.c
259
pd->pm_enabled ? "ENABLED" : "DISABLED",
usr.sbin/pmccontrol/pmccontrol.c
260
pmc_name_of_disposition(pd->pm_rowdisp));
usr.sbin/pmccontrol/pmccontrol.c
262
if (pd->pm_ownerpid != -1) {
usr.sbin/pmccontrol/pmccontrol.c
263
printf(" (pid %d)", pd->pm_ownerpid);
usr.sbin/pmccontrol/pmccontrol.c
265
pmc_name_of_event(pd->pm_event));
usr.sbin/pmccontrol/pmccontrol.c
266
if (PMC_IS_SAMPLING_MODE(pd->pm_mode))
usr.sbin/pmccontrol/pmccontrol.c
268
pd->pm_reloadcount);
usr.sbin/syslogd/syslogd.c
3715
int pfd[2], pd;
usr.sbin/syslogd/syslogd.c
3722
switch ((pid = pdfork(&pd, PD_CLOEXEC))) {
usr.sbin/syslogd/syslogd.c
3765
*rpd = pd;
usr.sbin/syslogd/syslogd.c
3770
deadq_enter(int pd)
usr.sbin/syslogd/syslogd.c
3774
if (pd == -1)
usr.sbin/syslogd/syslogd.c
3783
dq->dq_procdesc = pd;
usr.sbin/syslogd/syslogd.c
3797
log_deadchild(int pd, int status, const struct filed *f)
usr.sbin/syslogd/syslogd.c
3814
if (pdgetpid(pd, &pid) == -1)