Symbol: pf
arch/mips/include/asm/octeon/cvmx-pci-defs.h
279
uint32_t pf:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
285
uint32_t pf:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
309
uint32_t pf:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
315
uint32_t pf:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
338
uint32_t pf:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
344
uint32_t pf:1;
arch/s390/include/asm/timex.h
30
__uint128_t pf : 16; /* programmable field */
arch/s390/kernel/vdso/getcpu.c
15
*cpu = clk.pf;
arch/x86/include/asm/microcode.h
41
unsigned int pf;
arch/x86/include/asm/microcode.h
9
unsigned int pf;
arch/x86/include/asm/xor.h
52
#define BLK64(pf, op, i) \
arch/x86/include/asm/xor.h
53
pf(i) \
arch/x86/kernel/cpu/microcode/core.c
797
return sprintf(buf, "0x%x\n", uci->cpu_sig.pf);
arch/x86/kernel/cpu/microcode/intel.c
126
sig->pf = 0;
arch/x86/kernel/cpu/microcode/intel.c
134
sig->pf = 1 << ((val[1] >> 18) & 7);
arch/x86/kernel/cpu/microcode/intel.c
146
return ((!s1->pf && !pf2) || (s1->pf & pf2));
arch/x86/kernel/cpu/microcode/intel.c
156
if (cpu_signatures_match(sig, mc_hdr->sig, mc_hdr->pf))
arch/x86/kernel/cpu/microcode/intel.c
167
if (cpu_signatures_match(sig, ext_sig->sig, ext_sig->pf))
arch/x86/kernel/cpu/microcode/intel.c
277
sum = (mc_header->sig + mc_header->pf + mc_header->cksum) -
arch/x86/kernel/cpu/microcode/intel.c
278
(ext_sig->sig + ext_sig->pf + ext_sig->cksum);
arch/x86/kernel/cpu/microcode/intel.c
81
unsigned int pf;
arch/x86/platform/efi/efi_64.c
181
unsigned long pfn, text, pf, rodata, tramp;
arch/x86/platform/efi/efi_64.c
193
pf = _PAGE_NX | _PAGE_RW | _PAGE_ENC;
arch/x86/platform/efi/efi_64.c
194
if (kernel_map_pages_in_pgd(pgd, pfn, pa_memmap, num_pages, pf)) {
arch/x86/platform/efi/efi_64.c
210
if (kernel_map_pages_in_pgd(pgd, 0x0, 0x0, 1, pf)) {
arch/x86/platform/efi/efi_64.c
253
pf = _PAGE_NX | _PAGE_ENC;
arch/x86/platform/efi/efi_64.c
254
if (kernel_map_pages_in_pgd(pgd, pfn, rodata, npages, pf)) {
arch/x86/platform/efi/efi_64.c
262
pf = _PAGE_ENC;
arch/x86/platform/efi/efi_64.c
263
if (kernel_map_pages_in_pgd(pgd, pfn, tramp, 1, pf)) {
arch/x86/platform/efi/efi_64.c
370
static int __init efi_update_mappings(efi_memory_desc_t *md, unsigned long pf)
arch/x86/platform/efi/efi_64.c
378
err1 = kernel_map_pages_in_pgd(pgd, pfn, md->phys_addr, md->num_pages, pf);
arch/x86/platform/efi/efi_64.c
384
err2 = kernel_map_pages_in_pgd(pgd, pfn, md->virt_addr, md->num_pages, pf);
arch/x86/platform/efi/efi_64.c
398
unsigned long pf = 0;
arch/x86/platform/efi/efi_64.c
403
pf |= _PAGE_NX;
arch/x86/platform/efi/efi_64.c
406
pf |= _PAGE_RW;
arch/x86/platform/efi/efi_64.c
409
pf |= _PAGE_ENC;
arch/x86/platform/efi/efi_64.c
411
return efi_update_mappings(md, pf);
drivers/android/binder.c
2473
struct binder_ptr_fixup *pf =
drivers/android/binder.c
2488
copy_size = pf ? min(bytes_left, (size_t)pf->offset - offset)
drivers/android/binder.c
2498
BUG_ON(!pf);
drivers/android/binder.c
2500
if (pf->skip_size) {
drivers/android/binder.c
2507
bytes_copied += pf->skip_size;
drivers/android/binder.c
2513
pf->offset,
drivers/android/binder.c
2514
&pf->fixup_data,
drivers/android/binder.c
2515
sizeof(pf->fixup_data));
drivers/android/binder.c
2516
bytes_copied += sizeof(pf->fixup_data);
drivers/android/binder.c
2518
list_del(&pf->node);
drivers/android/binder.c
2519
kfree(pf);
drivers/android/binder.c
2520
pf = list_first_entry_or_null(pf_head,
drivers/android/binder.c
2527
list_for_each_entry_safe(pf, tmppf, pf_head, node) {
drivers/android/binder.c
2528
BUG_ON(pf->skip_size == 0);
drivers/android/binder.c
2529
list_del(&pf->node);
drivers/android/binder.c
2530
kfree(pf);
drivers/android/binder.c
2549
struct binder_ptr_fixup *pf, *tmppf;
drivers/android/binder.c
2555
list_for_each_entry_safe(pf, tmppf, pf_head, node) {
drivers/android/binder.c
2556
list_del(&pf->node);
drivers/android/binder.c
2557
kfree(pf);
drivers/android/binder.c
2625
struct binder_ptr_fixup *pf = kzalloc_obj(*pf);
drivers/android/binder.c
2628
if (!pf)
drivers/android/binder.c
2631
pf->offset = offset;
drivers/android/binder.c
2632
pf->fixup_data = fixup;
drivers/android/binder.c
2633
pf->skip_size = skip_size;
drivers/android/binder.c
2634
INIT_LIST_HEAD(&pf->node);
drivers/android/binder.c
2640
if (tmppf->offset < pf->offset) {
drivers/android/binder.c
2641
list_add(&pf->node, &tmppf->node);
drivers/android/binder.c
2649
list_add(&pf->node, pf_head);
drivers/block/drbd/drbd_state.c
2113
static void conn_old_common_state(struct drbd_connection *connection, union drbd_state *pcs, enum chg_state_flags *pf)
drivers/block/drbd/drbd_state.c
2154
*pf |= CS_DC_MASK;
drivers/block/drbd/drbd_state.c
2155
*pf &= flags;
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
103
if (adf_devmgr_add_dev(accel_dev, pf)) {
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
47
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
70
pf = adf_devmgr_pci_to_accel_dev(accel_pci_dev->pci_dev->physfn);
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
71
adf_devmgr_rm_dev(accel_dev, pf);
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
77
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_c3xxxvf/adf_drv.c
98
pf = adf_devmgr_pci_to_accel_dev(pdev->physfn);
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
103
if (adf_devmgr_add_dev(accel_dev, pf)) {
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
47
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
70
pf = adf_devmgr_pci_to_accel_dev(accel_pci_dev->pci_dev->physfn);
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
71
adf_devmgr_rm_dev(accel_dev, pf);
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
77
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_c62xvf/adf_drv.c
98
pf = adf_devmgr_pci_to_accel_dev(pdev->physfn);
drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
477
} pf;
drivers/crypto/intel/qat/qat_common/adf_aer.c
249
if (accel_dev->pf.vf_info)
drivers/crypto/intel/qat/qat_common/adf_common_drv.h
67
struct adf_accel_dev *pf);
drivers/crypto/intel/qat/qat_common/adf_common_drv.h
69
struct adf_accel_dev *pf);
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
140
struct adf_accel_dev *pf)
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
155
if (!accel_dev->is_vf || !pf) {
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
185
} else if (accel_dev->is_vf && pf) {
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
249
struct adf_accel_dev *pf)
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
253
if (!accel_dev->is_vf || !pf) {
drivers/crypto/intel/qat/qat_common/adf_dev_mgr.c
256
} else if (accel_dev->is_vf && pf) {
drivers/crypto/intel/qat/qat_common/adf_gen2_hw_data.c
108
val = accel_dev->pf.vf_info ? 0 : BIT_ULL(GET_MAX_BANKS(accel_dev)) - 1;
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
102
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
135
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
407
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
436
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
485
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
535
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
60
vf_info = &accel_dev->pf.vf_info[mdev->vf_id];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
639
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
668
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
732
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
781
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
830
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
84
vf_info = &accel_dev->pf.vf_info[mdev->vf_id];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
879
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
931
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_gen4_vf_mig.c
960
vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_isr.c
107
vf_info = accel_dev->pf.vf_info + i;
drivers/crypto/intel/qat/qat_common/adf_isr.c
161
if (accel_dev->pf.vf_info && adf_handle_vf2pf_int(accel_dev))
drivers/crypto/intel/qat/qat_common/adf_isr.c
213
if (!accel_dev->pf.vf_info) {
drivers/crypto/intel/qat/qat_common/adf_isr.c
277
if (!accel_dev->pf.vf_info)
drivers/crypto/intel/qat/qat_common/adf_isr.c
64
spin_lock_irqsave(&accel_dev->pf.vf2pf_ints_lock, flags);
drivers/crypto/intel/qat/qat_common/adf_isr.c
66
spin_unlock_irqrestore(&accel_dev->pf.vf2pf_ints_lock, flags);
drivers/crypto/intel/qat/qat_common/adf_isr.c
74
spin_lock_irqsave(&accel_dev->pf.vf2pf_ints_lock, flags);
drivers/crypto/intel/qat/qat_common/adf_isr.c
76
spin_unlock_irqrestore(&accel_dev->pf.vf2pf_ints_lock, flags);
drivers/crypto/intel/qat/qat_common/adf_isr.c
84
spin_lock(&accel_dev->pf.vf2pf_ints_lock);
drivers/crypto/intel/qat/qat_common/adf_isr.c
86
spin_unlock(&accel_dev->pf.vf2pf_ints_lock);
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_msg.c
20
for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) {
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_msg.c
45
for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++)
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_msg.c
64
for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) {
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_msg.c
79
for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) {
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_proto.c
243
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_proto.c
356
spin_lock_init(&accel_dev->pf.vf2pf_ints_lock);
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_proto.c
38
&accel_dev->pf.vf_info[vf_nr].pf2vf_lock);
drivers/crypto/intel/qat/qat_common/adf_pfvf_pf_proto.c
52
struct adf_accel_vf_info *vf_info = &accel_dev->pf.vf_info[vf_nr];
drivers/crypto/intel/qat/qat_common/adf_sriov.c
176
accel_dev->pf.vf_info = kzalloc_objs(struct adf_accel_vf_info, totalvfs);
drivers/crypto/intel/qat/qat_common/adf_sriov.c
178
if (!accel_dev->pf.vf_info)
drivers/crypto/intel/qat/qat_common/adf_sriov.c
202
kfree(accel_dev->pf.vf_info);
drivers/crypto/intel/qat/qat_common/adf_sriov.c
203
accel_dev->pf.vf_info = NULL;
drivers/crypto/intel/qat/qat_common/adf_sriov.c
219
if (!accel_dev->pf.vf_info)
drivers/crypto/intel/qat/qat_common/adf_sriov.c
244
if (!accel_dev->pf.vf_info)
drivers/crypto/intel/qat/qat_common/adf_sriov.c
258
for (i = 0, vf = accel_dev->pf.vf_info; i < totalvfs; i++, vf++) {
drivers/crypto/intel/qat/qat_common/adf_sriov.c
264
kfree(accel_dev->pf.vf_info);
drivers/crypto/intel/qat/qat_common/adf_sriov.c
265
accel_dev->pf.vf_info = NULL;
drivers/crypto/intel/qat/qat_common/adf_sriov.c
60
for (i = 0, vf_info = accel_dev->pf.vf_info; i < totalvfs;
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
103
if (adf_devmgr_add_dev(accel_dev, pf)) {
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
47
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
70
pf = adf_devmgr_pci_to_accel_dev(accel_pci_dev->pci_dev->physfn);
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
71
adf_devmgr_rm_dev(accel_dev, pf);
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
77
struct adf_accel_dev *pf;
drivers/crypto/intel/qat/qat_dh895xccvf/adf_drv.c
98
pf = adf_devmgr_pci_to_accel_dev(pdev->physfn);
drivers/crypto/marvell/octeontx2/otx2_cpt_common.h
22
#define OTX2_CPT_RVU_PFFUNC(pdev, pf, func) rvu_make_pcifunc(pdev, pf, func)
drivers/crypto/marvell/octeontx2/otx2_cptpf.h
23
struct otx2_cptpf_dev *pf;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
137
struct otx2_cptpf_dev *pf;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
143
pf = flr_work->pf;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
144
mbox = &pf->afpf_mbox;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
146
vf = flr_work - pf->flr_work;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
148
mutex_lock(&pf->lock);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
152
mutex_unlock(&pf->lock);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
161
otx2_cpt_send_mbox_msg(mbox, pf->pdev);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
162
if (!otx2_cpt_sync_mbox_msg(&pf->afpf_mbox)) {
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
169
otx2_cpt_write64(pf->reg_base, BLKADDR_RVUM, 0,
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
171
otx2_cpt_write64(pf->reg_base, BLKADDR_RVUM, 0,
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
174
mutex_unlock(&pf->lock);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
326
static void cptpf_flr_wq_destroy(struct otx2_cptpf_dev *pf)
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
328
if (!pf->flr_wq)
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
330
destroy_workqueue(pf->flr_wq);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
331
pf->flr_wq = NULL;
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
332
kfree(pf->flr_work);
drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c
348
cptpf->flr_work[vf].pf = cptpf;
drivers/firmware/efi/libstub/gop.c
142
int pf;
drivers/firmware/efi/libstub/gop.c
162
pf = info->pixel_format;
drivers/firmware/efi/libstub/gop.c
163
if (pf == PIXEL_BLT_ONLY || pf >= PIXEL_FORMAT_MAX) {
drivers/firmware/efi/libstub/gop.c
208
int pf = info->pixel_format;
drivers/firmware/efi/libstub/gop.c
210
if (pf == PIXEL_BLT_ONLY || pf >= PIXEL_FORMAT_MAX)
drivers/firmware/efi/libstub/gop.c
215
(cmdline.res.format < 0 || cmdline.res.format == pf) &&
drivers/firmware/efi/libstub/gop.c
216
(!cmdline.res.depth || cmdline.res.depth == pixel_bpp(pf, pi));
drivers/firmware/efi/libstub/gop.c
240
int pf = info->pixel_format;
drivers/firmware/efi/libstub/gop.c
241
u8 depth = pixel_bpp(pf, pi);
drivers/firmware/efi/libstub/gop.c
244
if (pf == PIXEL_BLT_ONLY || pf >= PIXEL_FORMAT_MAX)
drivers/firmware/efi/libstub/gop.c
266
int pf = info->pixel_format;
drivers/firmware/efi/libstub/gop.c
271
valid = !(pf == PIXEL_BLT_ONLY || pf >= PIXEL_FORMAT_MAX);
drivers/firmware/efi/libstub/gop.c
273
switch (pf) {
drivers/firmware/efi/libstub/gop.c
282
depth = pixel_bpp(pf, pi);
drivers/firmware/efi/libstub/gop.c
59
int pf = -1;
drivers/firmware/efi/libstub/gop.c
71
pf = PIXEL_RGB_RESERVED_8BIT_PER_COLOR;
drivers/firmware/efi/libstub/gop.c
74
pf = PIXEL_BGR_RESERVED_8BIT_PER_COLOR;
drivers/firmware/efi/libstub/gop.c
85
cmdline.res.format = pf;
drivers/gpu/drm/stm/ltdc.c
1280
enum ltdc_pix_fmt pf;
drivers/gpu/drm/stm/ltdc.c
1315
pf = to_ltdc_pixelformat(fb->format->format);
drivers/gpu/drm/stm/ltdc.c
1317
if (ldev->caps.pix_fmt_hw[val] == pf)
drivers/gpu/drm/stm/ltdc.c
1322
val = ltdc_set_flexible_pixel_format(plane, pf);
drivers/gpu/drm/stm/ltdc.c
502
enum ltdc_pix_fmt pf;
drivers/gpu/drm/stm/ltdc.c
507
pf = PF_ARGB8888;
drivers/gpu/drm/stm/ltdc.c
511
pf = PF_ABGR8888;
drivers/gpu/drm/stm/ltdc.c
515
pf = PF_RGBA8888;
drivers/gpu/drm/stm/ltdc.c
519
pf = PF_BGRA8888;
drivers/gpu/drm/stm/ltdc.c
522
pf = PF_RGB888;
drivers/gpu/drm/stm/ltdc.c
525
pf = PF_BGR888;
drivers/gpu/drm/stm/ltdc.c
528
pf = PF_RGB565;
drivers/gpu/drm/stm/ltdc.c
531
pf = PF_BGR565;
drivers/gpu/drm/stm/ltdc.c
535
pf = PF_ARGB1555;
drivers/gpu/drm/stm/ltdc.c
539
pf = PF_ARGB4444;
drivers/gpu/drm/stm/ltdc.c
542
pf = PF_L8;
drivers/gpu/drm/stm/ltdc.c
545
pf = PF_NONE;
drivers/gpu/drm/stm/ltdc.c
550
return pf;
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
54
xe->sriov.pf.device_total_vfs = TEST_MAX_VFS;
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
55
xe->sriov.pf.driver_max_vfs = TEST_MAX_VFS;
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
100
KUNIT_ASSERT_LE(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
110
if (!xe->sriov.pf.service.version.base.minor)
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
115
xe->sriov.pf.service.version.base.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
116
xe->sriov.pf.service.version.base.minor - 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
127
xe->sriov.pf.service.version.base.major - 1, 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
138
xe->sriov.pf.service.version.latest.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
139
xe->sriov.pf.service.version.latest.minor,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
141
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
142
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
152
xe->sriov.pf.service.version.latest.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
153
xe->sriov.pf.service.version.latest.minor + 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
155
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
156
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
166
xe->sriov.pf.service.version.latest.major + 1, 0,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
168
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
169
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
177
if (!xe->sriov.pf.service.version.latest.minor)
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
182
xe->sriov.pf.service.version.latest.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
183
xe->sriov.pf.service.version.latest.minor - 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
185
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
186
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.latest.minor - 1);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
194
if (xe->sriov.pf.service.version.base.major == xe->sriov.pf.service.version.latest.major)
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
199
xe->sriov.pf.service.version.latest.major - 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
200
xe->sriov.pf.service.version.base.minor + 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
202
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major - 1);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
203
KUNIT_ASSERT_GE(test, major, xe->sriov.pf.service.version.base.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
33
KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.base.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
34
KUNIT_ASSERT_NE(test, 0, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
35
KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
36
xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
37
if (xe->sriov.pf.service.version.base.major == xe->sriov.pf.service.version.latest.major)
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
38
KUNIT_ASSERT_LE(test, xe->sriov.pf.service.version.base.minor,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
39
xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
52
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
53
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
63
xe->sriov.pf.service.version.base.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
64
xe->sriov.pf.service.version.base.minor,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
66
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.base.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
67
KUNIT_ASSERT_EQ(test, minor, xe->sriov.pf.service.version.base.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
77
xe->sriov.pf.service.version.base.major,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
78
xe->sriov.pf.service.version.base.minor + 1,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
80
KUNIT_ASSERT_EQ(test, major, xe->sriov.pf.service.version.base.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
81
KUNIT_ASSERT_GE(test, minor, xe->sriov.pf.service.version.base.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
82
if (xe->sriov.pf.service.version.base.major == xe->sriov.pf.service.version.latest.major)
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
83
KUNIT_ASSERT_LE(test, minor, xe->sriov.pf.service.version.latest.minor);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
95
xe->sriov.pf.service.version.base.major + 1, 0,
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
97
KUNIT_ASSERT_GE(test, major, xe->sriov.pf.service.version.base.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
98
KUNIT_ASSERT_LE(test, major, xe->sriov.pf.service.version.latest.major);
drivers/gpu/drm/xe/tests/xe_sriov_pf_service_kunit.c
99
if (major == xe->sriov.pf.service.version.latest.major)
drivers/gpu/drm/xe/xe_device_types.h
222
} pf;
drivers/gpu/drm/xe/xe_device_types.h
452
struct xe_device_pf pf;
drivers/gpu/drm/xe/xe_gt.c
546
xe_lmtt_init(&gt_to_tile(gt)->sriov.pf.lmtt);
drivers/gpu/drm/xe/xe_gt.c
642
xe_lmtt_init_hw(&gt_to_tile(gt)->sriov.pf.lmtt);
drivers/gpu/drm/xe/xe_gt.c
840
xe_lmtt_init_hw(&gt_to_tile(gt)->sriov.pf.lmtt);
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
198
if (cancel_work_sync(&gt->sriov.pf.workers.restart)) {
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
233
struct xe_gt *gt = container_of(w, typeof(*gt), sriov.pf.workers.restart);
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
247
if (!queue_work(xe->sriov.wq, &gt->sriov.pf.workers.restart)) {
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
267
flush_work(&gt->sriov.pf.workers.restart);
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
42
gt->sriov.pf.vfs = drmm_kcalloc(&gt_to_xe(gt)->drm, 1 + num_vfs,
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
43
sizeof(*gt->sriov.pf.vfs), GFP_KERNEL);
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
44
if (!gt->sriov.pf.vfs)
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
53
INIT_WORK(&gt->sriov.pf.workers.restart, pf_worker_restart_func);
drivers/gpu/drm/xe/xe_gt_sriov_pf.c
60
if (disable_work_sync(&gt->sriov.pf.workers.restart)) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1218
spare = gt->sriov.pf.spare.num_dbs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1235
gt->sriov.pf.spare.num_dbs = spare;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1489
spare = gt->sriov.pf.spare.lmem_size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1503
gt->sriov.pf.spare.lmem_size = size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1549
lmtt = &tile->sriov.pf.lmtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1564
lmtt = &tile->sriov.pf.lmtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1589
lmtt = &tile->sriov.pf.lmtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1621
lmtt = &tile->sriov.pf.lmtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
256
return &gt->sriov.pf.vfs[vfid].config;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
3037
config = &gt->sriov.pf.vfs[n].config;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
3073
config = &gt->sriov.pf.vfs[n].config;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
3108
config = &gt->sriov.pf.vfs[n].config;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
3144
config = &gt->sriov.pf.vfs[n].config;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
452
spare = gt->sriov.pf.spare.ggtt_size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
468
gt->sriov.pf.spare.ggtt_size = size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
870
spare = gt->sriov.pf.spare.num_ctxs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
887
gt->sriov.pf.spare.num_ctxs = spare;
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2053
return container_of(cs, struct xe_gt_sriov_metadata, control) - gt->sriov.pf.vfs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2058
struct xe_gt_sriov_pf_control *pfc = &gt->sriov.pf.control;
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2087
struct xe_gt *gt = container_of(w, struct xe_gt, sriov.pf.control.worker);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2096
cancel_work_sync(&gt->sriov.pf.control.worker);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2129
spin_lock_init(&gt->sriov.pf.control.lock);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2130
INIT_LIST_HEAD(&gt->sriov.pf.control.list);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
2131
INIT_WORK(&gt->sriov.pf.control.worker, control_worker_func);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
237
return &gt->sriov.pf.vfs[vfid].control;
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
362
queue_work(xe->sriov.wq, &gt->sriov.pf.control.worker);
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
367
struct xe_gt_sriov_pf_control *pfc = &gt->sriov.pf.control;
drivers/gpu/drm/xe/xe_gt_sriov_pf_control.c
372
list_move_tail(&gt->sriov.pf.vfs[vfid].control.link, &pfc->list);
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
199
gt->sriov.pf.policy.guc.sched_groups.current_mode;
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
372
struct xe_gt_sriov_scheduler_groups *info = &gt->sriov.pf.policy.guc.sched_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
38
return &gt->sriov.pf.vfs[vfid].migration;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
127
data = &gt->sriov.pf.vfs[n].monitor;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
32
gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
41
gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
165
gt->sriov.pf.vfs[n].config.sched_priority = priority;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
176
&gt->sriov.pf.policy.guc.sched_if_idle,
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
190
return pf_provision_sched_if_idle(gt, gt->sriov.pf.policy.guc.sched_if_idle);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
198
gt->sriov.pf.policy.guc.sched_if_idle = false;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
236
enable = gt->sriov.pf.policy.guc.sched_if_idle;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
248
&gt->sriov.pf.policy.guc.reset_engine, enable);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
256
return pf_provision_reset_engine(gt, gt->sriov.pf.policy.guc.reset_engine);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
264
gt->sriov.pf.policy.guc.reset_engine = false;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
302
enable = gt->sriov.pf.policy.guc.reset_engine;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
314
&gt->sriov.pf.policy.guc.sample_period, value);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
322
return pf_provision_sample_period(gt, gt->sriov.pf.policy.guc.sample_period);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
330
gt->sriov.pf.policy.guc.sample_period = 0;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
368
value = gt->sriov.pf.policy.guc.sample_period;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
422
if (group > gt->sriov.pf.policy.guc.sched_groups.max_groups) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
424
group, gt->sriov.pf.policy.guc.sched_groups.max_groups);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
497
gt->sriov.pf.policy.guc.sched_groups.max_groups = 2;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
500
u32 *num_groups = &gt->sriov.pf.policy.guc.sched_groups.modes[m].num_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
502
&gt->sriov.pf.policy.guc.sched_groups.modes[m].groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
526
gt->sriov.pf.policy.guc.sched_groups.supported_modes |= BIT(m);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
541
return gt->sriov.pf.policy.guc.sched_groups.supported_modes;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
560
return gt->sriov.pf.policy.guc.sched_groups.supported_modes & BIT(mode);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
565
struct guc_sched_group *groups = gt->sriov.pf.policy.guc.sched_groups.modes[mode].groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
566
u32 num_groups = gt->sriov.pf.policy.guc.sched_groups.modes[mode].num_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
583
if (gt->sriov.pf.policy.guc.sched_groups.current_mode == mode)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
610
gt->sriov.pf.policy.guc.sched_groups.current_mode = mode;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
624
return __pf_provision_sched_groups(gt, gt->sriov.pf.policy.guc.sched_groups.current_mode);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
632
gt->sriov.pf.policy.guc.sched_groups.current_mode = XE_SRIOV_SCHED_GROUPS_DISABLED;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
665
return gt->sriov.pf.policy.guc.sched_groups.current_mode != XE_SRIOV_SCHED_GROUPS_DISABLED;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
759
print_guc_policies(p, &gt->sriov.pf.policy.guc);
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
158
xe_gt_assert(gt, !gt->sriov.pf.service.runtime.size);
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
159
xe_gt_assert(gt, !gt->sriov.pf.service.runtime.regs);
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
160
xe_gt_assert(gt, !gt->sriov.pf.service.runtime.values);
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
173
gt->sriov.pf.service.runtime.size = size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
174
gt->sriov.pf.service.runtime.regs = regs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
175
gt->sriov.pf.service.runtime.values = values;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
193
if (!gt->sriov.pf.service.runtime.size)
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
196
size = gt->sriov.pf.service.runtime.size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
197
regs = gt->sriov.pf.service.runtime.regs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
198
values = gt->sriov.pf.service.runtime.values;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
298
runtime = &gt->sriov.pf.service.runtime;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
414
size = gt->sriov.pf.service.runtime.size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
415
regs = gt->sriov.pf.service.runtime.regs;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
416
values = gt->sriov.pf.service.runtime.values;
drivers/gpu/drm/xe/xe_gt_types.h
183
struct xe_gt_sriov_pf pf;
drivers/gpu/drm/xe/xe_guc_pagefault.c
12
static void guc_ack_fault(struct xe_pagefault *pf, int err)
drivers/gpu/drm/xe/xe_guc_pagefault.c
14
u32 vfid = FIELD_GET(PFD_VFID, pf->producer.msg[2]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
15
u32 engine_instance = FIELD_GET(PFD_ENG_INSTANCE, pf->producer.msg[0]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
16
u32 engine_class = FIELD_GET(PFD_ENG_CLASS, pf->producer.msg[0]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
17
u32 pdata = FIELD_GET(PFD_PDATA_LO, pf->producer.msg[0]) |
drivers/gpu/drm/xe/xe_guc_pagefault.c
18
(FIELD_GET(PFD_PDATA_HI, pf->producer.msg[1]) <<
drivers/gpu/drm/xe/xe_guc_pagefault.c
27
FIELD_PREP(PFR_ASID, pf->consumer.asid),
drivers/gpu/drm/xe/xe_guc_pagefault.c
34
struct xe_guc *guc = pf->producer.private;
drivers/gpu/drm/xe/xe_guc_pagefault.c
56
struct xe_pagefault pf;
drivers/gpu/drm/xe/xe_guc_pagefault.c
67
pf.gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_pagefault.c
73
pf.consumer.page_addr = ((u64)FIELD_GET(PFD_VIRTUAL_ADDR_HI, msg[3])
drivers/gpu/drm/xe/xe_guc_pagefault.c
77
pf.consumer.asid = FIELD_GET(PFD_ASID, msg[1]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
78
pf.consumer.access_type = FIELD_GET(PFD_ACCESS_TYPE, msg[2]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
79
pf.consumer.fault_type = FIELD_GET(PFD_FAULT_TYPE, msg[2]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
81
pf.consumer.fault_level = XE_PAGEFAULT_LEVEL_NACK;
drivers/gpu/drm/xe/xe_guc_pagefault.c
83
pf.consumer.fault_level = FIELD_GET(PFD_FAULT_LEVEL, msg[0]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
84
pf.consumer.engine_class = FIELD_GET(PFD_ENG_CLASS, msg[0]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
85
pf.consumer.engine_instance = FIELD_GET(PFD_ENG_INSTANCE, msg[0]);
drivers/gpu/drm/xe/xe_guc_pagefault.c
87
pf.producer.private = guc;
drivers/gpu/drm/xe/xe_guc_pagefault.c
88
pf.producer.ops = &guc_pagefault_ops;
drivers/gpu/drm/xe/xe_guc_pagefault.c
90
pf.producer.msg[i] = msg[i];
drivers/gpu/drm/xe/xe_guc_pagefault.c
94
return xe_pagefault_handler(guc_to_xe(guc), &pf);
drivers/gpu/drm/xe/xe_lmtt.c
101
lmtt_debug(&pt->bo->tile->sriov.pf.lmtt, "level=%u addr=%llx\n",
drivers/gpu/drm/xe/xe_lmtt.c
47
return container_of(lmtt, struct xe_tile, sriov.pf.lmtt);
drivers/gpu/drm/xe/xe_pagefault.c
157
static int xe_pagefault_service(struct xe_pagefault *pf)
drivers/gpu/drm/xe/xe_pagefault.c
159
struct xe_gt *gt = pf->gt;
drivers/gpu/drm/xe/xe_pagefault.c
167
if (pf->consumer.fault_level == XE_PAGEFAULT_LEVEL_NACK)
drivers/gpu/drm/xe/xe_pagefault.c
170
vm = xe_pagefault_asid_to_vm(xe, pf->consumer.asid);
drivers/gpu/drm/xe/xe_pagefault.c
184
vma = xe_vm_find_vma_by_addr(vm, pf->consumer.page_addr);
drivers/gpu/drm/xe/xe_pagefault.c
191
pf->consumer.access_type != XE_PAGEFAULT_ACCESS_TYPE_READ) {
drivers/gpu/drm/xe/xe_pagefault.c
196
atomic = xe_pagefault_access_is_atomic(pf->consumer.access_type);
drivers/gpu/drm/xe/xe_pagefault.c
200
pf->consumer.page_addr, atomic);
drivers/gpu/drm/xe/xe_pagefault.c
214
struct xe_pagefault *pf)
drivers/gpu/drm/xe/xe_pagefault.c
220
memcpy(pf, pf_queue->data + pf_queue->tail, sizeof(*pf));
drivers/gpu/drm/xe/xe_pagefault.c
230
static void xe_pagefault_print(struct xe_pagefault *pf)
drivers/gpu/drm/xe/xe_pagefault.c
232
xe_gt_info(pf->gt, "\n\tASID: %d\n"
drivers/gpu/drm/xe/xe_pagefault.c
239
pf->consumer.asid,
drivers/gpu/drm/xe/xe_pagefault.c
240
upper_32_bits(pf->consumer.page_addr),
drivers/gpu/drm/xe/xe_pagefault.c
241
lower_32_bits(pf->consumer.page_addr),
drivers/gpu/drm/xe/xe_pagefault.c
242
pf->consumer.fault_type,
drivers/gpu/drm/xe/xe_pagefault.c
243
pf->consumer.access_type,
drivers/gpu/drm/xe/xe_pagefault.c
244
pf->consumer.fault_level,
drivers/gpu/drm/xe/xe_pagefault.c
245
pf->consumer.engine_class,
drivers/gpu/drm/xe/xe_pagefault.c
246
xe_hw_engine_class_to_str(pf->consumer.engine_class),
drivers/gpu/drm/xe/xe_pagefault.c
247
pf->consumer.engine_instance);
drivers/gpu/drm/xe/xe_pagefault.c
254
struct xe_pagefault pf;
drivers/gpu/drm/xe/xe_pagefault.c
260
while (xe_pagefault_queue_pop(pf_queue, &pf)) {
drivers/gpu/drm/xe/xe_pagefault.c
263
if (!pf.gt) /* Fault squashed during reset */
drivers/gpu/drm/xe/xe_pagefault.c
266
err = xe_pagefault_service(&pf);
drivers/gpu/drm/xe/xe_pagefault.c
268
xe_pagefault_print(&pf);
drivers/gpu/drm/xe/xe_pagefault.c
269
xe_gt_info(pf.gt, "Fault response: Unsuccessful %pe\n",
drivers/gpu/drm/xe/xe_pagefault.c
273
pf.producer.ops->ack_fault(&pf, err);
drivers/gpu/drm/xe/xe_pagefault.c
276
queue_work(gt_to_xe(pf.gt)->usm.pf_wq, w);
drivers/gpu/drm/xe/xe_pagefault.c
386
struct xe_pagefault *pf = pf_queue->data + i;
drivers/gpu/drm/xe/xe_pagefault.c
388
if (pf->gt == gt)
drivers/gpu/drm/xe/xe_pagefault.c
389
pf->gt = NULL;
drivers/gpu/drm/xe/xe_pagefault.c
428
int xe_pagefault_handler(struct xe_device *xe, struct xe_pagefault *pf)
drivers/gpu/drm/xe/xe_pagefault.c
431
(pf->consumer.asid % XE_PAGEFAULT_QUEUE_COUNT);
drivers/gpu/drm/xe/xe_pagefault.c
438
memcpy(pf_queue->data + pf_queue->head, pf, sizeof(*pf));
drivers/gpu/drm/xe/xe_pagefault.c
445
pf->consumer.asid % XE_PAGEFAULT_QUEUE_COUNT);
drivers/gpu/drm/xe/xe_pagefault.h
17
int xe_pagefault_handler(struct xe_device *xe, struct xe_pagefault *pf);
drivers/gpu/drm/xe/xe_pagefault_types.h
44
void (*ack_fault)(struct xe_pagefault *pf, int err);
drivers/gpu/drm/xe/xe_pci_sriov.c
101
return xe_sriov_pf_arm_guard(xe, &xe->sriov.pf.guard_vfs_enabling, false, NULL);
drivers/gpu/drm/xe/xe_pci_sriov.c
108
xe_sriov_pf_disarm_guard(xe, &xe->sriov.pf.guard_vfs_enabling, false, NULL);
drivers/gpu/drm/xe/xe_sriov_packet.c
20
return &xe->sriov.pf.vfs[vfid].migration.lock;
drivers/gpu/drm/xe/xe_sriov_packet.c
29
return &xe->sriov.pf.vfs[vfid].migration.pending;
drivers/gpu/drm/xe/xe_sriov_packet.c
39
return &xe->sriov.pf.vfs[vfid].migration.descriptor;
drivers/gpu/drm/xe/xe_sriov_packet.c
48
return &xe->sriov.pf.vfs[vfid].migration.trailer;
drivers/gpu/drm/xe/xe_sriov_pf.c
102
err = drmm_mutex_init(&xe->drm, &xe->sriov.pf.master_lock);
drivers/gpu/drm/xe/xe_sriov_pf.c
110
xe_guard_init(&xe->sriov.pf.guard_vfs_enabling, "vfs_enabling");
drivers/gpu/drm/xe/xe_sriov_pf.c
250
return xe_sriov_pf_arm_guard(xe, &xe->sriov.pf.guard_vfs_enabling, true,
drivers/gpu/drm/xe/xe_sriov_pf.c
265
xe_sriov_pf_disarm_guard(xe, &xe->sriov.pf.guard_vfs_enabling, true,
drivers/gpu/drm/xe/xe_sriov_pf.c
282
drm_printf(p, "total: %u\n", xe->sriov.pf.device_total_vfs);
drivers/gpu/drm/xe/xe_sriov_pf.c
283
drm_printf(p, "supported: %u\n", xe->sriov.pf.driver_max_vfs);
drivers/gpu/drm/xe/xe_sriov_pf.c
79
xe->sriov.pf.device_total_vfs = totalvfs;
drivers/gpu/drm/xe/xe_sriov_pf.c
80
xe->sriov.pf.driver_max_vfs = newlimit;
drivers/gpu/drm/xe/xe_sriov_pf.c
97
xe->sriov.pf.vfs = drmm_kcalloc(&xe->drm, 1 + xe_sriov_pf_get_totalvfs(xe),
drivers/gpu/drm/xe/xe_sriov_pf.c
98
sizeof(*xe->sriov.pf.vfs), GFP_KERNEL);
drivers/gpu/drm/xe/xe_sriov_pf.c
99
if (!xe->sriov.pf.vfs)
drivers/gpu/drm/xe/xe_sriov_pf_helpers.h
37
return xe->sriov.pf.driver_max_vfs;
drivers/gpu/drm/xe/xe_sriov_pf_helpers.h
65
return &xe->sriov.pf.master_lock;
drivers/gpu/drm/xe/xe_sriov_pf_migration.c
24
return &xe->sriov.pf.vfs[vfid].migration;
drivers/gpu/drm/xe/xe_sriov_pf_migration.c
49
return IS_ENABLED(CONFIG_DRM_XE_DEBUG) || !xe->sriov.pf.migration.disabled;
drivers/gpu/drm/xe/xe_sriov_pf_migration.c
73
xe->sriov.pf.migration.disabled = true;
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
138
if (mode == xe->sriov.pf.provision.mode)
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
151
mode_to_string(xe->sriov.pf.provision.mode),
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
153
xe->sriov.pf.provision.mode = mode;
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
32
return xe->sriov.pf.provision.mode == XE_SRIOV_PROVISIONING_MODE_AUTO;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
100
xe->sriov.pf.vfs[vfid].version.major = 0;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
101
xe->sriov.pf.vfs[vfid].version.minor = 0;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
121
return major == xe->sriov.pf.vfs[vfid].version.major &&
drivers/gpu/drm/xe/xe_sriov_pf_service.c
122
minor <= xe->sriov.pf.vfs[vfid].version.minor;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
202
print_pf_version(p, "base", &xe->sriov.pf.service.version.base);
drivers/gpu/drm/xe/xe_sriov_pf_service.c
203
print_pf_version(p, "latest", &xe->sriov.pf.service.version.latest);
drivers/gpu/drm/xe/xe_sriov_pf_service.c
206
version = &xe->sriov.pf.vfs[n].version;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
32
xe->sriov.pf.service.version.base.major = GUC_RELAY_VERSION_BASE_MAJOR;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
33
xe->sriov.pf.service.version.base.minor = GUC_RELAY_VERSION_BASE_MINOR;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
36
xe->sriov.pf.service.version.latest.major = GUC_RELAY_VERSION_LATEST_MAJOR;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
37
xe->sriov.pf.service.version.latest.minor = GUC_RELAY_VERSION_LATEST_MINOR;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
45
struct xe_sriov_pf_service_version base = xe->sriov.pf.service.version.base;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
46
struct xe_sriov_pf_service_version latest = xe->sriov.pf.service.version.latest;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
92
xe->sriov.pf.vfs[vfid].version.major = major;
drivers/gpu/drm/xe/xe_sriov_pf_service.c
93
xe->sriov.pf.vfs[vfid].version.minor = minor;
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
496
xe_assert(xe, !xe->sriov.pf.sysfs.root);
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
497
xe->sriov.pf.sysfs.root = root;
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
509
root = xe->sriov.pf.sysfs.root;
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
523
xe_assert(xe, !xe->sriov.pf.vfs[n].kobj);
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
524
xe->sriov.pf.vfs[n].kobj = kobj;
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
539
struct kobject *kobj = xe->sriov.pf.vfs[PFID].kobj;
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
609
err = sysfs_create_link(xe->sriov.pf.vfs[VFID(n)].kobj,
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
636
sysfs_remove_link(xe->sriov.pf.vfs[VFID(n)].kobj, "device");
drivers/iio/accel/adxl355_core.c
663
struct iio_poll_func *pf = p;
drivers/iio/accel/adxl355_core.c
664
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/adxl355_core.c
704
sizeof(data->buffer), pf->timestamp);
drivers/iio/accel/adxl372.c
634
struct iio_poll_func *pf = p;
drivers/iio/accel/adxl372.c
635
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/bma180.c
863
struct iio_poll_func *pf = p;
drivers/iio/accel/bma180.c
864
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/bma220_core.c
231
struct iio_poll_func *pf = p;
drivers/iio/accel/bma220_core.c
232
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/bma400_core.c
1609
struct iio_poll_func *pf = p;
drivers/iio/accel/bma400_core.c
1610
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/bmc150-accel-core.c
1185
struct iio_poll_func *pf = p;
drivers/iio/accel/bmc150-accel-core.c
1186
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/bmc150-accel-core.c
1198
pf->timestamp);
drivers/iio/accel/kionix-kx022a.c
1027
struct iio_poll_func *pf = p;
drivers/iio/accel/kionix-kx022a.c
1028
struct iio_dev *idev = pf->indio_dev;
drivers/iio/accel/kxcjk-1013.c
1187
struct iio_poll_func *pf = p;
drivers/iio/accel/kxcjk-1013.c
1188
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/kxsd9.c
208
const struct iio_poll_func *pf = p;
drivers/iio/accel/kxsd9.c
209
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/mma7455_core.c
91
struct iio_poll_func *pf = p;
drivers/iio/accel/mma7455_core.c
92
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/mma8452.c
1094
struct iio_poll_func *pf = p;
drivers/iio/accel/mma8452.c
1095
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/msa311.c
885
struct iio_poll_func *pf = p;
drivers/iio/accel/msa311.c
886
struct msa311_priv *msa311 = iio_priv(pf->indio_dev);
drivers/iio/accel/msa311.c
887
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/mxc4005.c
327
struct iio_poll_func *pf = private;
drivers/iio/accel/mxc4005.c
328
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/mxc4005.c
337
pf->timestamp);
drivers/iio/accel/sca3300.c
476
struct iio_poll_func *pf = p;
drivers/iio/accel/sca3300.c
477
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/stk8312.c
430
struct iio_poll_func *pf = p;
drivers/iio/accel/stk8312.c
431
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/stk8312.c
463
pf->timestamp);
drivers/iio/accel/stk8ba50.c
313
struct iio_poll_func *pf = p;
drivers/iio/accel/stk8ba50.c
314
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/accel/stk8ba50.c
343
pf->timestamp);
drivers/iio/adc/ad4000.c
787
struct iio_poll_func *pf = p;
drivers/iio/adc/ad4000.c
788
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad4000.c
797
pf->timestamp);
drivers/iio/adc/ad4030.c
697
struct iio_poll_func *pf = p;
drivers/iio/adc/ad4030.c
698
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad4030.c
707
pf->timestamp);
drivers/iio/adc/ad4062.c
635
struct iio_poll_func *pf = p;
drivers/iio/adc/ad4062.c
636
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad4170-4.c
2791
struct iio_poll_func *pf = p;
drivers/iio/adc/ad4170-4.c
2792
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad4695.c
791
struct iio_poll_func *pf = p;
drivers/iio/adc/ad4695.c
792
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad4695.c
801
pf->timestamp);
drivers/iio/adc/ad7266.c
83
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7266.c
84
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7266.c
91
pf->timestamp);
drivers/iio/adc/ad7298.c
149
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7298.c
150
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7380.c
1356
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7380.c
1357
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7380.c
1367
pf->timestamp);
drivers/iio/adc/ad7476.c
82
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7476.c
83
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7606.c
633
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7606.c
634
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7766.c
68
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7766.c
69
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7766.c
78
pf->timestamp);
drivers/iio/adc/ad7768-1.c
1292
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7768-1.c
1293
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7779.c
583
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7779.c
584
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7779.c
601
pf->timestamp);
drivers/iio/adc/ad7887.c
113
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7887.c
114
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7923.c
201
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7923.c
202
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7944.c
525
struct iio_poll_func *pf = p;
drivers/iio/adc/ad7944.c
526
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad7944.c
536
pf->timestamp);
drivers/iio/adc/ad7944.c
539
pf->timestamp);
drivers/iio/adc/ad799x.c
192
struct iio_poll_func *pf = p;
drivers/iio/adc/ad799x.c
193
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad_sigma_delta.c
615
struct iio_poll_func *pf = p;
drivers/iio/adc/ad_sigma_delta.c
616
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ad_sigma_delta.c
639
iio_push_to_buffers_with_timestamp(indio_dev, data, pf->timestamp);
drivers/iio/adc/ad_sigma_delta.c
666
pf->timestamp);
drivers/iio/adc/at91-sama5d2_adc.c
1424
struct iio_poll_func *pf)
drivers/iio/adc/at91-sama5d2_adc.c
1477
pf->timestamp);
drivers/iio/adc/at91-sama5d2_adc.c
1531
struct iio_poll_func *pf = p;
drivers/iio/adc/at91-sama5d2_adc.c
1532
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/at91-sama5d2_adc.c
1545
at91_adc_trigger_handler_nodma(indio_dev, pf);
drivers/iio/adc/at91_adc.c
266
struct iio_poll_func *pf = p;
drivers/iio/adc/at91_adc.c
267
struct iio_dev *idev = pf->indio_dev;
drivers/iio/adc/at91_adc.c
278
iio_push_to_buffers_with_timestamp(idev, st->buffer, pf->timestamp);
drivers/iio/adc/cc10001_adc.c
137
struct iio_poll_func *pf = p;
drivers/iio/adc/cc10001_adc.c
146
indio_dev = pf->indio_dev;
drivers/iio/adc/dln2-adc.c
465
struct iio_poll_func *pf = p;
drivers/iio/adc/dln2-adc.c
466
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/hi8435.c
437
struct iio_poll_func *pf = private;
drivers/iio/adc/hi8435.c
438
struct iio_dev *idev = pf->indio_dev;
drivers/iio/adc/hx711.c
360
struct iio_poll_func *pf = p;
drivers/iio/adc/hx711.c
361
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/hx711.c
376
pf->timestamp);
drivers/iio/adc/max1027.c
492
struct iio_poll_func *pf = private;
drivers/iio/adc/max1027.c
493
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/max1118.c
169
struct iio_poll_func *pf = p;
drivers/iio/adc/max1118.c
170
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/max11410.c
624
struct iio_poll_func *pf = p;
drivers/iio/adc/max11410.c
625
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/max1363.c
1479
struct iio_poll_func *pf = p;
drivers/iio/adc/max1363.c
1480
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/mcp3911.c
512
struct iio_poll_func *pf = p;
drivers/iio/adc/mcp3911.c
513
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/mt6360-adc.c
261
struct iio_poll_func *pf = p;
drivers/iio/adc/mt6360-adc.c
262
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/mxs-lradc-adc.c
413
struct iio_poll_func *pf = p;
drivers/iio/adc/mxs-lradc-adc.c
414
struct iio_dev *iio = pf->indio_dev;
drivers/iio/adc/mxs-lradc-adc.c
429
pf->timestamp);
drivers/iio/adc/nxp-sar-adc.c
833
struct iio_poll_func *pf = p;
drivers/iio/adc/nxp-sar-adc.c
834
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/pac1921.c
1021
struct iio_poll_func *pf = p;
drivers/iio/adc/pac1921.c
1022
struct iio_dev *idev = pf->indio_dev;
drivers/iio/adc/pac1921.c
1033
ret = pac1921_check_push_overflow(idev, pf->timestamp);
drivers/iio/adc/pac1921.c
1048
pf->timestamp);
drivers/iio/adc/rockchip_saradc.c
397
struct iio_poll_func *pf = p;
drivers/iio/adc/rockchip_saradc.c
398
struct iio_dev *i_dev = pf->indio_dev;
drivers/iio/adc/rtq6056.c
640
struct iio_poll_func *pf = p;
drivers/iio/adc/rtq6056.c
641
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/stm32-adc.c
1987
struct iio_poll_func *pf = p;
drivers/iio/adc/stm32-adc.c
1988
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/stm32-adc.c
1996
pf->timestamp);
drivers/iio/adc/ti-adc081c.c
130
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-adc081c.c
131
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-adc0832.c
206
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-adc0832.c
207
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-adc084s021.c
145
struct iio_poll_func *pf = pollfunc;
drivers/iio/adc/ti-adc084s021.c
146
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-adc108s102.c
141
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-adc108s102.c
142
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-adc12138.c
335
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-adc12138.c
336
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-adc161s626.c
108
struct iio_poll_func *pf = private;
drivers/iio/adc/ti-adc161s626.c
109
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads1015.c
444
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-ads1015.c
445
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads1018.c
551
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-ads1018.c
552
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads1018.c
571
pf->timestamp);
drivers/iio/adc/ti-ads1119.c
505
struct iio_poll_func *pf = private;
drivers/iio/adc/ti-ads1119.c
506
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads124s08.c
276
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-ads124s08.c
277
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads124s08.c
301
pf->timestamp);
drivers/iio/adc/ti-ads131e08.c
613
struct iio_poll_func *pf = private;
drivers/iio/adc/ti-ads131e08.c
614
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads7950.c
306
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-ads7950.c
307
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-ads8688.c
381
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-ads8688.c
382
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-lmp92064.c
197
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-lmp92064.c
198
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-tlc4541.c
93
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-tlc4541.c
94
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/ti-tsc2046.c
434
struct iio_poll_func *pf = p;
drivers/iio/adc/ti-tsc2046.c
435
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/adc/xilinx-xadc-core.c
674
struct iio_poll_func *pf = p;
drivers/iio/adc/xilinx-xadc-core.c
675
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/addac/ad74115.c
611
struct iio_poll_func *pf = p;
drivers/iio/addac/ad74115.c
612
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/addac/ad74413r.c
788
struct iio_poll_func *pf = p;
drivers/iio/addac/ad74413r.c
789
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/atlas-sensor.c
449
struct iio_poll_func *pf = private;
drivers/iio/chemical/atlas-sensor.c
450
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/bme680_core.c
1055
struct iio_poll_func *pf = p;
drivers/iio/chemical/bme680_core.c
1056
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/ccs811.c
332
struct iio_poll_func *pf = p;
drivers/iio/chemical/ccs811.c
333
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/ens160_core.c
258
struct iio_poll_func *pf = p;
drivers/iio/chemical/ens160_core.c
259
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/ens160_core.c
271
pf->timestamp);
drivers/iio/chemical/pms7003.c
109
struct iio_poll_func *pf = p;
drivers/iio/chemical/pms7003.c
110
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/scd30_core.c
584
struct iio_poll_func *pf = p;
drivers/iio/chemical/scd30_core.c
585
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/scd4x.c
665
struct iio_poll_func *pf = p;
drivers/iio/chemical/scd4x.c
666
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/chemical/sps30.c
105
struct iio_poll_func *pf = p;
drivers/iio/chemical/sps30.c
106
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/common/cros_ec_sensors/cros_ec_activity.c
188
struct iio_poll_func *pf = p;
drivers/iio/common/cros_ec_sensors/cros_ec_activity.c
189
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/common/cros_ec_sensors/cros_ec_sensors_core.c
713
struct iio_poll_func *pf = p;
drivers/iio/common/cros_ec_sensors/cros_ec_sensors_core.c
714
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/common/st_sensors/st_sensors_buffer.c
51
struct iio_poll_func *pf = p;
drivers/iio/common/st_sensors/st_sensors_buffer.c
52
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/dac/ad3552r.c
291
struct iio_poll_func *pf = p;
drivers/iio/dac/ad3552r.c
292
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/dac/ad5766.c
574
struct iio_poll_func *pf = p;
drivers/iio/dac/ad5766.c
575
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/dummy/iio_simple_dummy_buffer.c
51
struct iio_poll_func *pf = p;
drivers/iio/dummy/iio_simple_dummy_buffer.c
52
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/gyro/adxrs290.c
503
struct iio_poll_func *pf = p;
drivers/iio/gyro/adxrs290.c
504
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/gyro/adxrs290.c
518
pf->timestamp);
drivers/iio/gyro/bmg160_core.c
876
struct iio_poll_func *pf = p;
drivers/iio/gyro/bmg160_core.c
877
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/gyro/bmg160_core.c
889
pf->timestamp);
drivers/iio/gyro/fxas21002c_core.c
725
struct iio_poll_func *pf = p;
drivers/iio/gyro/fxas21002c_core.c
726
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/gyro/itg3200_buffer.c
46
struct iio_poll_func *pf = p;
drivers/iio/gyro/itg3200_buffer.c
47
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/gyro/itg3200_buffer.c
62
iio_push_to_buffers_with_timestamp(indio_dev, &scan, pf->timestamp);
drivers/iio/gyro/mpu3050-core.c
472
const struct iio_poll_func *pf = p;
drivers/iio/gyro/mpu3050-core.c
473
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/health/afe4403.c
310
struct iio_poll_func *pf = private;
drivers/iio/health/afe4403.c
311
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/health/afe4403.c
339
pf->timestamp);
drivers/iio/health/afe4404.c
329
struct iio_poll_func *pf = private;
drivers/iio/health/afe4404.c
330
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/health/afe4404.c
342
pf->timestamp);
drivers/iio/humidity/am2315.c
162
struct iio_poll_func *pf = p;
drivers/iio/humidity/am2315.c
163
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/humidity/am2315.c
186
pf->timestamp);
drivers/iio/humidity/hdc100x.c
301
struct iio_poll_func *pf = p;
drivers/iio/humidity/hdc100x.c
302
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/humidity/hts221_buffer.c
163
struct iio_poll_func *pf = p;
drivers/iio/humidity/hts221_buffer.c
164
struct iio_dev *iio_dev = pf->indio_dev;
drivers/iio/iio_core_trigger.h
25
struct iio_poll_func *pf);
drivers/iio/iio_core_trigger.h
27
struct iio_poll_func *pf);
drivers/iio/iio_core_trigger.h
49
struct iio_poll_func *pf)
drivers/iio/iio_core_trigger.h
54
struct iio_poll_func *pf)
drivers/iio/imu/adis16400.c
608
struct iio_poll_func *pf = p;
drivers/iio/imu/adis16400.c
609
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16400.c
628
indio_dev->scan_bytes - sizeof(pf->timestamp),
drivers/iio/imu/adis16400.c
629
pf->timestamp);
drivers/iio/imu/adis16400.c
633
pf->timestamp);
drivers/iio/imu/adis16475.c
1559
static int adis16475_push_single_sample(struct iio_poll_func *pf)
drivers/iio/imu/adis16475.c
1561
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16475.c
1650
iio_push_to_buffers_with_timestamp(indio_dev, st->data, pf->timestamp);
drivers/iio/imu/adis16475.c
1657
struct iio_poll_func *pf = p;
drivers/iio/imu/adis16475.c
1658
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16475.c
1661
adis16475_push_single_sample(pf);
drivers/iio/imu/adis16475.c
1692
static int adis16575_custom_burst_read(struct iio_poll_func *pf, u8 burst_req)
drivers/iio/imu/adis16475.c
1694
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16475.c
1703
return adis16475_push_single_sample(pf);
drivers/iio/imu/adis16475.c
1722
struct iio_poll_func *pf = p;
drivers/iio/imu/adis16475.c
1723
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16475.c
1746
ret = adis16575_custom_burst_read(pf, adis->data->burst_reg_cmd);
drivers/iio/imu/adis16475.c
1751
ret = adis16475_push_single_sample(pf);
drivers/iio/imu/adis16475.c
1757
ret = adis16575_custom_burst_read(pf, 0);
drivers/iio/imu/adis16480.c
1395
struct iio_poll_func *pf = p;
drivers/iio/imu/adis16480.c
1396
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16480.c
1488
iio_push_to_buffers_with_timestamp(indio_dev, st->data, pf->timestamp);
drivers/iio/imu/adis16550.c
838
struct iio_poll_func *pf = p;
drivers/iio/imu/adis16550.c
840
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis16550.c
867
iio_push_to_buffers_with_timestamp(indio_dev, data, pf->timestamp);
drivers/iio/imu/adis_buffer.c
148
struct iio_poll_func *pf = p;
drivers/iio/imu/adis_buffer.c
149
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/adis_buffer.c
163
pf->timestamp);
drivers/iio/imu/bmi160/bmi160_core.c
432
struct iio_poll_func *pf = p;
drivers/iio/imu/bmi160/bmi160_core.c
433
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/bmi160/bmi160_core.c
446
iio_push_to_buffers_with_timestamp(indio_dev, data->buf, pf->timestamp);
drivers/iio/imu/bmi270/bmi270_core.c
840
struct iio_poll_func *pf = p;
drivers/iio/imu/bmi270/bmi270_core.c
841
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/bmi270/bmi270_core.c
855
pf->timestamp);
drivers/iio/imu/bmi323/bmi323_core.c
1414
struct iio_poll_func *pf = p;
drivers/iio/imu/bmi323/bmi323_core.c
1415
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/bno055/bno055.c
1448
struct iio_poll_func *pf = p;
drivers/iio/imu/bno055/bno055.c
1449
struct iio_dev *iio_dev = pf->indio_dev;
drivers/iio/imu/bno055/bno055.c
1526
&priv->buf, pf->timestamp);
drivers/iio/imu/inv_mpu6050/inv_mpu_ring.c
104
inv_sensors_timestamp_interrupt(&st->timestamp, 1, pf->timestamp);
drivers/iio/imu/inv_mpu6050/inv_mpu_ring.c
45
struct iio_poll_func *pf = p;
drivers/iio/imu/inv_mpu6050/inv_mpu_ring.c
46
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/kmx61.c
1186
struct iio_poll_func *pf = p;
drivers/iio/imu/kmx61.c
1187
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/smi240.c
370
struct iio_poll_func *pf = p;
drivers/iio/imu/smi240.c
371
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/smi240.c
386
iio_push_to_buffers_with_timestamp(indio_dev, data->buf, pf->timestamp);
drivers/iio/imu/smi330/smi330_core.c
622
struct iio_poll_func *pf = p;
drivers/iio/imu/smi330/smi330_core.c
623
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/imu/smi330/smi330_core.c
631
iio_push_to_buffers_with_timestamp(indio_dev, data->buf, pf->timestamp);
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_core.c
2714
struct iio_poll_func *pf = private;
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_core.c
2715
struct iio_dev *iio_dev = pf->indio_dev;
drivers/iio/industrialio-trigger.c
281
struct iio_poll_func *pf)
drivers/iio/industrialio-trigger.c
283
struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(pf->indio_dev);
drivers/iio/industrialio-trigger.c
292
pf->irq = iio_trigger_get_irq(trig);
drivers/iio/industrialio-trigger.c
293
if (pf->irq < 0) {
drivers/iio/industrialio-trigger.c
300
ret = request_threaded_irq(pf->irq, pf->h, pf->thread,
drivers/iio/industrialio-trigger.c
301
pf->type, pf->name,
drivers/iio/industrialio-trigger.c
302
pf);
drivers/iio/industrialio-trigger.c
318
if (!iio_validate_own_trigger(pf->indio_dev, trig))
drivers/iio/industrialio-trigger.c
324
free_irq(pf->irq, pf);
drivers/iio/industrialio-trigger.c
326
iio_trigger_put_irq(trig, pf->irq);
drivers/iio/industrialio-trigger.c
333
struct iio_poll_func *pf)
drivers/iio/industrialio-trigger.c
335
struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(pf->indio_dev);
drivers/iio/industrialio-trigger.c
345
if (pf->indio_dev->dev.parent == trig->dev.parent)
drivers/iio/industrialio-trigger.c
347
iio_trigger_put_irq(trig, pf->irq);
drivers/iio/industrialio-trigger.c
348
free_irq(pf->irq, pf);
drivers/iio/industrialio-trigger.c
350
pf->irq = 0;
drivers/iio/industrialio-trigger.c
357
struct iio_poll_func *pf = p;
drivers/iio/industrialio-trigger.c
359
pf->timestamp = iio_get_time_ns(pf->indio_dev);
drivers/iio/industrialio-trigger.c
373
struct iio_poll_func *pf;
drivers/iio/industrialio-trigger.c
375
pf = kmalloc_obj(*pf);
drivers/iio/industrialio-trigger.c
376
if (!pf)
drivers/iio/industrialio-trigger.c
379
pf->name = kvasprintf(GFP_KERNEL, fmt, vargs);
drivers/iio/industrialio-trigger.c
381
if (pf->name == NULL) {
drivers/iio/industrialio-trigger.c
382
kfree(pf);
drivers/iio/industrialio-trigger.c
385
pf->h = h;
drivers/iio/industrialio-trigger.c
386
pf->thread = thread;
drivers/iio/industrialio-trigger.c
387
pf->type = type;
drivers/iio/industrialio-trigger.c
388
pf->indio_dev = indio_dev;
drivers/iio/industrialio-trigger.c
390
return pf;
drivers/iio/industrialio-trigger.c
394
void iio_dealloc_pollfunc(struct iio_poll_func *pf)
drivers/iio/industrialio-trigger.c
396
kfree(pf->name);
drivers/iio/industrialio-trigger.c
397
kfree(pf);
drivers/iio/light/acpi-als.c
142
struct iio_poll_func *pf = p;
drivers/iio/light/acpi-als.c
143
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/acpi-als.c
167
if (!pf->timestamp)
drivers/iio/light/acpi-als.c
168
pf->timestamp = iio_get_time_ns(indio_dev);
drivers/iio/light/acpi-als.c
170
iio_push_to_buffers_with_ts(indio_dev, &scan, sizeof(scan), pf->timestamp);
drivers/iio/light/adjd_s311.c
114
struct iio_poll_func *pf = p;
drivers/iio/light/adjd_s311.c
115
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/as73211.c
636
struct iio_poll_func *pf = p;
drivers/iio/light/as73211.c
637
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/bh1745.c
737
struct iio_poll_func *pf = p;
drivers/iio/light/bh1745.c
738
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/gp2ap020a00f.c
962
struct iio_poll_func *pf = data;
drivers/iio/light/gp2ap020a00f.c
963
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/gp2ap020a00f.c
988
pf->timestamp);
drivers/iio/light/isl29125.c
173
struct iio_poll_func *pf = p;
drivers/iio/light/isl29125.c
174
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/ltr501.c
1276
struct iio_poll_func *pf = p;
drivers/iio/light/ltr501.c
1277
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/max44000.c
488
struct iio_poll_func *pf = p;
drivers/iio/light/max44000.c
489
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/opt4060.c
1062
struct iio_poll_func *pf = p;
drivers/iio/light/opt4060.c
1063
struct iio_dev *idev = pf->indio_dev;
drivers/iio/light/opt4060.c
1089
iio_push_to_buffers_with_ts(idev, &raw, sizeof(raw), pf->timestamp);
drivers/iio/light/rpr0521.c
440
struct iio_poll_func *pf = p;
drivers/iio/light/rpr0521.c
441
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/rpr0521.c
447
pf->timestamp = data->irq_timestamp;
drivers/iio/light/rpr0521.c
451
if (!pf->timestamp)
drivers/iio/light/rpr0521.c
452
pf->timestamp = iio_get_time_ns(indio_dev);
drivers/iio/light/rpr0521.c
459
sizeof(data->scan), pf->timestamp);
drivers/iio/light/rpr0521.c
463
pf->timestamp = 0;
drivers/iio/light/si1145.c
448
struct iio_poll_func *pf = private;
drivers/iio/light/si1145.c
449
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/st_uvis25_core.c
232
struct iio_poll_func *pf = p;
drivers/iio/light/st_uvis25_core.c
233
struct iio_dev *iio_dev = pf->indio_dev;
drivers/iio/light/tcs3414.c
198
struct iio_poll_func *pf = p;
drivers/iio/light/tcs3414.c
199
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/tcs3472.c
371
struct iio_poll_func *pf = p;
drivers/iio/light/tcs3472.c
372
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/vcnl4000.c
1643
struct iio_poll_func *pf = p;
drivers/iio/light/vcnl4000.c
1644
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/vcnl4035.c
102
struct iio_poll_func *pf = p;
drivers/iio/light/vcnl4035.c
103
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/light/veml6030.c
887
struct iio_poll_func *pf = p;
drivers/iio/light/veml6030.c
888
struct iio_dev *iio = pf->indio_dev;
drivers/iio/light/veml6030.c
906
iio_push_to_buffers_with_ts(iio, &scan, sizeof(scan), pf->timestamp);
drivers/iio/light/veml6046x00.c
835
struct iio_poll_func *pf = p;
drivers/iio/light/veml6046x00.c
836
struct iio_dev *iio = pf->indio_dev;
drivers/iio/light/vl6180.c
538
struct iio_poll_func *pf = priv;
drivers/iio/light/vl6180.c
539
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/af8133j.c
357
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/af8133j.c
358
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/ak8974.c
685
const struct iio_poll_func *pf = p;
drivers/iio/magnetometer/ak8974.c
686
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/ak8975.c
896
const struct iio_poll_func *pf = p;
drivers/iio/magnetometer/ak8975.c
897
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/als31300.c
226
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/als31300.c
227
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/als31300.c
246
iio_push_to_buffers_with_ts(indio_dev, &scan, sizeof(scan), pf->timestamp);
drivers/iio/magnetometer/bmc150_magn.c
663
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/bmc150_magn.c
664
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/bmc150_magn.c
674
pf->timestamp);
drivers/iio/magnetometer/hmc5843_core.c
436
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/hmc5843_core.c
437
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/mag3110.c
390
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/mag3110.c
391
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/rm3100-core.c
463
struct iio_poll_func *pf = p;
drivers/iio/magnetometer/rm3100-core.c
464
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/rm3100-core.c
519
pf->timestamp);
drivers/iio/magnetometer/tlv493d.c
344
struct iio_poll_func *pf = ptr;
drivers/iio/magnetometer/tlv493d.c
345
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/magnetometer/tlv493d.c
364
iio_push_to_buffers_with_ts(indio_dev, &scan, sizeof(scan), pf->timestamp);
drivers/iio/magnetometer/yamaha-yas530.c
681
const struct iio_poll_func *pf = p;
drivers/iio/magnetometer/yamaha-yas530.c
682
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/potentiostat/lmp91000.c
136
struct iio_poll_func *pf = private;
drivers/iio/potentiostat/lmp91000.c
137
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/abp2030pa.c
321
struct iio_poll_func *pf = private;
drivers/iio/pressure/abp2030pa.c
322
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/bmp280-core.c
1107
struct iio_poll_func *pf = p;
drivers/iio/pressure/bmp280-core.c
1108
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/bmp280-core.c
1228
struct iio_poll_func *pf = p;
drivers/iio/pressure/bmp280-core.c
1229
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/bmp280-core.c
1904
struct iio_poll_func *pf = p;
drivers/iio/pressure/bmp280-core.c
1905
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/bmp280-core.c
2612
struct iio_poll_func *pf = p;
drivers/iio/pressure/bmp280-core.c
2613
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/bmp280-core.c
2957
struct iio_poll_func *pf = p;
drivers/iio/pressure/bmp280-core.c
2958
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/dlhl60d.c
243
struct iio_poll_func *pf = private;
drivers/iio/pressure/dlhl60d.c
244
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/hsc030pa.c
305
struct iio_poll_func *pf = private;
drivers/iio/pressure/hsc030pa.c
306
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/mpl3115.c
298
struct iio_poll_func *pf = p;
drivers/iio/pressure/mpl3115.c
299
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/mprls0025pa.c
272
struct iio_poll_func *pf = p;
drivers/iio/pressure/mprls0025pa.c
273
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/ms5611_core.c
210
struct iio_poll_func *pf = p;
drivers/iio/pressure/ms5611_core.c
211
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/pressure/rohm-bm1390.c
626
struct iio_poll_func *pf = p;
drivers/iio/pressure/rohm-bm1390.c
627
struct iio_dev *idev = pf->indio_dev;
drivers/iio/proximity/as3935.c
224
struct iio_poll_func *pf = private;
drivers/iio/proximity/as3935.c
225
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/hx9023s.c
934
struct iio_poll_func *pf = private;
drivers/iio/proximity/hx9023s.c
935
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/hx9023s.c
960
sizeof(data->buffer), pf->timestamp);
drivers/iio/proximity/isl29501.c
937
struct iio_poll_func *pf = p;
drivers/iio/proximity/isl29501.c
938
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/isl29501.c
952
iio_push_to_buffers_with_timestamp(indio_dev, &scan, pf->timestamp);
drivers/iio/proximity/mb1232.c
115
struct iio_poll_func *pf = p;
drivers/iio/proximity/mb1232.c
116
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/mb1232.c
128
pf->timestamp);
drivers/iio/proximity/pulsedlight-lidar-lite-v2.c
227
struct iio_poll_func *pf = private;
drivers/iio/proximity/pulsedlight-lidar-lite-v2.c
228
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/srf08.c
176
struct iio_poll_func *pf = p;
drivers/iio/proximity/srf08.c
177
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/srf08.c
191
pf->timestamp);
drivers/iio/proximity/sx9500.c
630
struct iio_poll_func *pf = private;
drivers/iio/proximity/sx9500.c
631
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/sx_common.c
364
struct iio_poll_func *pf = private;
drivers/iio/proximity/sx_common.c
365
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/proximity/sx_common.c
383
sizeof(data->buffer), pf->timestamp);
drivers/iio/proximity/vl53l0x-i2c.c
77
struct iio_poll_func *pf = priv;
drivers/iio/proximity/vl53l0x-i2c.c
78
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/resolver/ad2s1210.c
1281
struct iio_poll_func *pf = p;
drivers/iio/resolver/ad2s1210.c
1282
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/resolver/ad2s1210.c
1342
ad2s1210_push_events(indio_dev, st->sample.fault, pf->timestamp);
drivers/iio/resolver/ad2s1210.c
1344
pf->timestamp);
drivers/iio/temperature/maxim_thermocouple.c
170
struct iio_poll_func *pf = private;
drivers/iio/temperature/maxim_thermocouple.c
171
struct iio_dev *indio_dev = pf->indio_dev;
drivers/iio/temperature/tmp006.c
251
struct iio_poll_func *pf = p;
drivers/iio/temperature/tmp006.c
252
struct iio_dev *indio_dev = pf->indio_dev;
drivers/infiniband/hw/cxgb4/cm.c
618
(ep->com.dev->rdev.lldi.pf));
drivers/infiniband/hw/usnic/usnic_ib.h
92
struct usnic_ib_dev *pf;
drivers/infiniband/hw/usnic/usnic_ib_main.c
550
struct usnic_ib_dev *pf;
drivers/infiniband/hw/usnic/usnic_ib_main.c
588
pf = usnic_ib_discover_pf(vf->vnic);
drivers/infiniband/hw/usnic/usnic_ib_main.c
589
if (IS_ERR(pf)) {
drivers/infiniband/hw/usnic/usnic_ib_main.c
590
err = PTR_ERR(pf);
drivers/infiniband/hw/usnic/usnic_ib_main.c
596
vf->pf = pf;
drivers/infiniband/hw/usnic/usnic_ib_main.c
598
mutex_lock(&pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_main.c
599
list_add_tail(&vf->link, &pf->vf_dev_list);
drivers/infiniband/hw/usnic/usnic_ib_main.c
607
pf->vf_res_cnt[res_type] = usnic_vnic_res_cnt(vf->vnic,
drivers/infiniband/hw/usnic/usnic_ib_main.c
611
mutex_unlock(&pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_main.c
614
dev_name(&pf->ib_dev.dev));
drivers/infiniband/hw/usnic/usnic_ib_main.c
633
struct usnic_ib_dev *pf = vf->pf;
drivers/infiniband/hw/usnic/usnic_ib_main.c
635
mutex_lock(&pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_main.c
637
mutex_unlock(&pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_main.c
639
kref_put(&pf->vf_cnt, usnic_ib_undiscover_pf);
drivers/infiniband/hw/usnic/usnic_ib_main.c
79
return scnprintf(buf, buf_sz, "PF: %s ", dev_name(&vf->pf->ib_dev.dev));
drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c
485
ib_event.device = &qp_grp->vf->pf->ib_dev;
drivers/infiniband/hw/usnic/usnic_ib_sysfs.c
276
us_ibdev = qp_grp->vf->pf;
drivers/infiniband/hw/usnic/usnic_ib_sysfs.c
291
us_ibdev = qp_grp->vf->pf;
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
400
mutex_lock(&vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
415
mutex_unlock(&vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
419
mutex_unlock(&vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
536
mutex_lock(&vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
544
mutex_unlock(&vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
561
mutex_lock(&qp_grp->vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
575
mutex_unlock(&qp_grp->vf->pf->usdev_lock);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
95
us_ibdev = qp_grp->vf->pf;
drivers/isdn/mISDN/layer2.c
1100
enquiry_cr(struct layer2 *l2, u_char typ, u_char cr, u_char pf)
drivers/isdn/mISDN/layer2.c
1109
tmp[i++] = (l2->vr << 1) | (pf ? 1 : 0);
drivers/isdn/mISDN/layer2.c
1111
tmp[i++] = (l2->vr << 5) | typ | (pf ? 0x10 : 0);
drivers/md/dm-thin.c
1357
if (len < pool->sectors_per_block && pool->pf.zero_new_blocks) {
drivers/md/dm-thin.c
1396
if (pool->pf.zero_new_blocks) {
drivers/md/dm-thin.c
1601
return pool->pf.error_if_no_space ? BLK_STS_NOSPC : 0;
drivers/md/dm-thin.c
245
struct pool_features pf;
drivers/md/dm-thin.c
2453
if (get_pool_mode(pool) == PM_OUT_OF_DATA_SPACE && !pool->pf.error_if_no_space) {
drivers/md/dm-thin.c
2454
pool->pf.error_if_no_space = true;
drivers/md/dm-thin.c
2609
if (!pool->pf.error_if_no_space && no_space_timeout)
drivers/md/dm-thin.c
2617
pool->pf.error_if_no_space = pt->requested_pf.error_if_no_space;
drivers/md/dm-thin.c
2627
pool->pf.mode = new_mode;
drivers/md/dm-thin.c
2879
pool->pf = pt->adjusted_pf;
drivers/md/dm-thin.c
2899
static void pool_features_init(struct pool_features *pf)
drivers/md/dm-thin.c
2901
pf->mode = PM_WRITE;
drivers/md/dm-thin.c
2902
pf->zero_new_blocks = true;
drivers/md/dm-thin.c
2903
pf->discard_enabled = true;
drivers/md/dm-thin.c
2904
pf->discard_passdown = true;
drivers/md/dm-thin.c
2905
pf->error_if_no_space = false;
drivers/md/dm-thin.c
294
return pool->pf.mode;
drivers/md/dm-thin.c
2966
pool_features_init(&pool->pf);
drivers/md/dm-thin.c
310
if (!pool->pf.error_if_no_space)
drivers/md/dm-thin.c
3141
static int parse_pool_features(struct dm_arg_set *as, struct pool_features *pf,
drivers/md/dm-thin.c
3167
pf->zero_new_blocks = false;
drivers/md/dm-thin.c
3170
pf->discard_enabled = false;
drivers/md/dm-thin.c
3173
pf->discard_passdown = false;
drivers/md/dm-thin.c
3176
pf->mode = PM_READ_ONLY;
drivers/md/dm-thin.c
3179
pf->error_if_no_space = true;
drivers/md/dm-thin.c
3288
struct pool_features pf;
drivers/md/dm-thin.c
3320
pool_features_init(&pf);
drivers/md/dm-thin.c
3323
r = parse_pool_features(&as, &pf, ti);
drivers/md/dm-thin.c
3328
((pf.mode == PM_READ_ONLY) ? 0 : BLK_OPEN_WRITE);
drivers/md/dm-thin.c
3364
block_size, pf.mode == PM_READ_ONLY, &ti->error, &pool_created);
drivers/md/dm-thin.c
3376
if (!pool_created && pf.discard_enabled != pool->pf.discard_enabled) {
drivers/md/dm-thin.c
3387
pt->adjusted_pf = pt->requested_pf = pf;
drivers/md/dm-thin.c
3396
if (pf.discard_enabled && pf.discard_passdown) {
drivers/md/dm-thin.c
3882
static void emit_flags(struct pool_features *pf, char *result,
drivers/md/dm-thin.c
3885
unsigned int count = !pf->zero_new_blocks + !pf->discard_enabled +
drivers/md/dm-thin.c
3886
!pf->discard_passdown + (pf->mode == PM_READ_ONLY) +
drivers/md/dm-thin.c
3887
pf->error_if_no_space;
drivers/md/dm-thin.c
3890
if (!pf->zero_new_blocks)
drivers/md/dm-thin.c
3893
if (!pf->discard_enabled)
drivers/md/dm-thin.c
3896
if (!pf->discard_passdown)
drivers/md/dm-thin.c
3899
if (pf->mode == PM_READ_ONLY)
drivers/md/dm-thin.c
3902
if (pf->error_if_no_space)
drivers/md/dm-thin.c
4002
if (!pool->pf.discard_enabled)
drivers/md/dm-thin.c
4004
else if (pool->pf.discard_passdown)
drivers/md/dm-thin.c
4009
if (pool->pf.error_if_no_space)
drivers/md/dm-thin.c
4275
if (tc->pool->pf.discard_enabled) {
drivers/md/dm-thin.c
4487
if (pool->pf.discard_enabled) {
drivers/media/pci/mgb4/mgb4_trigger.c
100
iio_push_to_buffers_with_ts(indio_dev, &scan, sizeof(scan), pf->timestamp);
drivers/media/pci/mgb4/mgb4_trigger.c
89
struct iio_poll_func *pf = p;
drivers/media/pci/mgb4/mgb4_trigger.c
90
struct iio_dev *indio_dev = pf->indio_dev;
drivers/media/platform/qcom/venus/venc_ctrls.c
25
static int venc_calc_bpframes(u32 gop_size, u32 conseq_b, u32 *bf, u32 *pf)
drivers/media/platform/qcom/venus/venc_ctrls.c
34
*bf = *pf = 0;
drivers/media/platform/qcom/venus/venc_ctrls.c
37
*pf = gop_size - 1;
drivers/media/platform/qcom/venus/venc_ctrls.c
65
*pf = p;
drivers/misc/mrvl_cn10k_dpi.c
169
struct dpipf *pf;
drivers/misc/mrvl_cn10k_dpi.c
323
dpi = mbox->pf;
drivers/misc/mrvl_cn10k_dpi.c
368
dpi->mbox[vf]->pf = dpi;
drivers/net/ethernet/amd/pds_core/auxbus.c
100
struct pdsc *pf;
drivers/net/ethernet/amd/pds_core/auxbus.c
105
pf = pci_get_drvdata(pf_pdev);
drivers/net/ethernet/amd/pds_core/auxbus.c
107
dev_dbg(pf->dev, "%s: %s opcode %d\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
116
err = pdsc_adminq_post(pf, &cmd, resp,
drivers/net/ethernet/amd/pds_core/auxbus.c
119
dev_info(pf->dev, "client admin cmd failed: %pe\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
135
struct pdsc *pf,
drivers/net/ethernet/amd/pds_core/auxbus.c
17
int pds_client_register(struct pdsc *pf, char *devname)
drivers/net/ethernet/amd/pds_core/auxbus.c
175
void pdsc_auxbus_dev_del(struct pdsc *cf, struct pdsc *pf,
drivers/net/ethernet/amd/pds_core/auxbus.c
183
mutex_lock(&pf->config_lock);
drivers/net/ethernet/amd/pds_core/auxbus.c
186
pds_client_unregister(pf, padev->client_id);
drivers/net/ethernet/amd/pds_core/auxbus.c
191
mutex_unlock(&pf->config_lock);
drivers/net/ethernet/amd/pds_core/auxbus.c
194
int pdsc_auxbus_dev_add(struct pdsc *cf, struct pdsc *pf,
drivers/net/ethernet/amd/pds_core/auxbus.c
211
mutex_lock(&pf->config_lock);
drivers/net/ethernet/amd/pds_core/auxbus.c
216
dev_err(pf->dev, "%s: can't add dev, VF client in bad state %#lx\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
226
vt_support = !!le16_to_cpu(pf->dev_ident.vif_types[vt]);
drivers/net/ethernet/amd/pds_core/auxbus.c
228
pf->viftype_status[vt].supported &&
drivers/net/ethernet/amd/pds_core/auxbus.c
229
pf->viftype_status[vt].enabled))
drivers/net/ethernet/amd/pds_core/auxbus.c
237
PDS_CORE_DRV_NAME, pf->viftype_status[vt].name, cf->uid);
drivers/net/ethernet/amd/pds_core/auxbus.c
238
client_id = pds_client_register(pf, devname);
drivers/net/ethernet/amd/pds_core/auxbus.c
244
padev = pdsc_auxbus_dev_register(cf, pf, client_id,
drivers/net/ethernet/amd/pds_core/auxbus.c
245
pf->viftype_status[vt].name);
drivers/net/ethernet/amd/pds_core/auxbus.c
247
pds_client_unregister(pf, client_id);
drivers/net/ethernet/amd/pds_core/auxbus.c
254
mutex_unlock(&pf->config_lock);
drivers/net/ethernet/amd/pds_core/auxbus.c
28
err = pdsc_adminq_post(pf, &cmd, &comp, false);
drivers/net/ethernet/amd/pds_core/auxbus.c
30
dev_info(pf->dev, "register dev_name %s with DSC failed, status %d: %pe\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
37
dev_err(pf->dev, "%s: device returned null client_id\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
42
dev_dbg(pf->dev, "%s: device returned client_id %d for %s\n",
drivers/net/ethernet/amd/pds_core/auxbus.c
57
int pds_client_unregister(struct pdsc *pf, u16 client_id)
drivers/net/ethernet/amd/pds_core/auxbus.c
66
err = pdsc_adminq_post(pf, &cmd, &comp, false);
drivers/net/ethernet/amd/pds_core/auxbus.c
68
dev_info(pf->dev, "unregister client_id %d failed, status %d: %pe\n",
drivers/net/ethernet/amd/pds_core/core.h
308
int pdsc_auxbus_dev_add(struct pdsc *cf, struct pdsc *pf,
drivers/net/ethernet/amd/pds_core/core.h
311
void pdsc_auxbus_dev_del(struct pdsc *cf, struct pdsc *pf,
drivers/net/ethernet/amd/pds_core/main.c
177
struct pdsc *pf;
drivers/net/ethernet/amd/pds_core/main.c
180
pf = pdsc_get_pf_struct(vf->pdev);
drivers/net/ethernet/amd/pds_core/main.c
181
if (IS_ERR_OR_NULL(pf))
drivers/net/ethernet/amd/pds_core/main.c
182
return PTR_ERR(pf) ?: -1;
drivers/net/ethernet/amd/pds_core/main.c
191
pf->vfs[vf->vf_id].vf = vf;
drivers/net/ethernet/amd/pds_core/main.c
192
err = pdsc_auxbus_dev_add(vf, pf, PDS_DEV_TYPE_VDPA,
drivers/net/ethernet/amd/pds_core/main.c
193
&pf->vfs[vf->vf_id].padev);
drivers/net/ethernet/amd/pds_core/main.c
422
struct pdsc *pf;
drivers/net/ethernet/amd/pds_core/main.c
424
pf = pdsc_get_pf_struct(pdsc->pdev);
drivers/net/ethernet/amd/pds_core/main.c
425
if (!IS_ERR(pf)) {
drivers/net/ethernet/amd/pds_core/main.c
426
pdsc_auxbus_dev_del(pdsc, pf, &pf->vfs[pdsc->vf_id].padev);
drivers/net/ethernet/amd/pds_core/main.c
427
pf->vfs[pdsc->vf_id].vf = NULL;
drivers/net/ethernet/amd/pds_core/main.c
488
struct pdsc *pf;
drivers/net/ethernet/amd/pds_core/main.c
490
pf = pdsc_get_pf_struct(pdsc->pdev);
drivers/net/ethernet/amd/pds_core/main.c
491
if (!IS_ERR(pf))
drivers/net/ethernet/amd/pds_core/main.c
492
pdsc_auxbus_dev_del(pdsc, pf,
drivers/net/ethernet/amd/pds_core/main.c
493
&pf->vfs[pdsc->vf_id].padev);
drivers/net/ethernet/amd/pds_core/main.c
536
struct pdsc *pf;
drivers/net/ethernet/amd/pds_core/main.c
538
pf = pdsc_get_pf_struct(pdsc->pdev);
drivers/net/ethernet/amd/pds_core/main.c
539
if (!IS_ERR(pf))
drivers/net/ethernet/amd/pds_core/main.c
540
pdsc_auxbus_dev_add(pdsc, pf, PDS_DEV_TYPE_VDPA,
drivers/net/ethernet/amd/pds_core/main.c
541
&pf->vfs[pdsc->vf_id].padev);
drivers/net/ethernet/broadcom/bcmsysport.c
219
STAT_MIB_RX("rx_pause", mib.rx.pf),
drivers/net/ethernet/broadcom/bcmsysport.c
246
STAT_MIB_TX("tx_pause", mib.tx.pf),
drivers/net/ethernet/broadcom/bcmsysport.h
519
u32 pf; /* RO # of Received pause frame pkt */
drivers/net/ethernet/broadcom/bcmsysport.h
540
u32 pf; /* RO # of xmited pause frame count */
drivers/net/ethernet/broadcom/bnge/bnge.h
139
struct bnge_pf_info pf;
drivers/net/ethernet/broadcom/bnge/bnge_auxr.c
178
auxr_dev->pf_port_id = bd->pf.port_id;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
579
struct bnge_pf_info *pf = &bd->pf;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
599
pf->fw_fid = le16_to_cpu(resp->fid);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
600
pf->port_id = le16_to_cpu(resp->port_id);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
601
memcpy(pf->mac_addr, resp->mac_address, ETH_ALEN);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
2572
eth_hw_addr_set(bd->netdev, bd->pf.mac_addr);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
101
if (!bd->pf.port_id && bd->port_count > 1)
drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
1272
struct per_pf_stats pf;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c
1483
offsetof(struct bnx2x_fw_stats_data, pf);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.c
922
&bp->fw_stats_data->pf.tstorm_pf_statistics;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
10483
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
10497
req->port_id = cpu_to_le16(pf->port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
10510
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
10525
req_qs->port_id = cpu_to_le16(pf->port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12320
bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12679
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12690
req->port_id = cpu_to_le16(pf->port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12728
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12750
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12768
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13178
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13179
int n = pf->active_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13323
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13354
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14466
if (bp->pf.registered_vfs)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14467
return bp->pf.registered_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14493
if (bp->pf.active_vfs &&
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15514
return bnxt_tc_setup_flower(bp, bp->pf.fw_fid, type_data);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16680
eth_hw_addr_set(bp->dev, bp->pf.mac_addr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2767
if (bp->pf.port_id != port_id)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2962
if ((vf_id < bp->pf.first_vf_id) ||
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2963
(vf_id >= bp->pf.first_vf_id + bp->pf.active_vfs)) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2969
set_bit(vf_id - bp->pf.first_vf_id, bp->pf.vf_event_bmap);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6140
static u16 bnxt_vf_target_id(struct bnxt_pf_info *pf, u16 vf_idx)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6143
struct bnxt_vf_info *vf = &pf->vf[vf_idx];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6158
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6160
if (fltr->base.vf_idx >= pf->active_vfs)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6163
target_id = bnxt_vf_target_id(pf, fltr->base.vf_idx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6185
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6187
if (fltr->base.vf_idx >= pf->active_vfs)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
6190
target_id = bnxt_vf_target_id(pf, fltr->base.vf_idx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8567
bp->pf.registered_vfs = le16_to_cpu(resp->registered_vfs);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9635
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9637
pf->vf_resv_strategy =
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9639
if (pf->vf_resv_strategy > BNXT_VF_RESV_STRATEGY_MINIMAL_STATIC)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9640
pf->vf_resv_strategy = BNXT_VF_RESV_STRATEGY_MAXIMAL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9664
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9821
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9823
pf->fw_fid = le16_to_cpu(resp->fid);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9824
pf->port_id = le16_to_cpu(resp->port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9825
memcpy(pf->mac_addr, resp->mac_address, ETH_ALEN);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9826
pf->first_vf_id = le16_to_cpu(resp->first_vf_id);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9827
pf->max_vfs = le16_to_cpu(resp->max_vfs);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2656
struct bnxt_pf_info pf;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2897
return BNXT_PF(bp) && (bp->pf.active_vfs || bp->sriov_cfg);
drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c
1033
idx = bp->pf.port_id;
drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c
1035
idx = bp->pf.fw_fid - BNXT_FIRST_PF_FID;
drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c
1296
attrs.phys.port_number = bp->pf.port_id;
drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c
1300
rc = devlink_port_register(dl, &bp->dl_port, bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
1531
if (BNXT_PF(bp) && vf > bp->pf.active_vfs)
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3252
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4619
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4827
req->port_id = cpu_to_le16(bp->pf.port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4892
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4914
req->port_id = cpu_to_le16(pf->port_id);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5228
if (bp->pf.active_vfs || !BNXT_SINGLE_PF(bp)) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
1045
if (num_vfs && num_vfs == bp->pf.active_vfs)
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
1288
u32 i = 0, active_vfs = bp->pf.active_vfs, vf_id;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
1292
vf_id = find_next_bit(bp->pf.vf_event_bmap, active_vfs, i);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
1296
clear_bit(vf_id, bp->pf.vf_event_bmap);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
1297
bnxt_vf_req_validate_snd(bp, &bp->pf.vf[vf_id]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
170
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
192
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
231
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
270
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
304
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
344
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
377
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
408
vf = &bp->pf.vf[i];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
417
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
425
for (i = pf->first_vf_id; i < pf->first_vf_id + num_vfs; i++) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
440
kfree(bp->pf.vf_event_bmap);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
441
bp->pf.vf_event_bmap = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
444
if (bp->pf.hwrm_cmd_req_addr[i]) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
446
bp->pf.hwrm_cmd_req_addr[i],
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
447
bp->pf.hwrm_cmd_req_dma_addr[i]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
448
bp->pf.hwrm_cmd_req_addr[i] = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
452
bp->pf.active_vfs = 0;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
453
kfree(bp->pf.vf);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
454
bp->pf.vf = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
462
bp->pf.vf = kzalloc_objs(struct bnxt_vf_info, num_vfs);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
463
if (!bp->pf.vf)
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
474
bp->pf.hwrm_cmd_req_addr[i] =
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
476
&bp->pf.hwrm_cmd_req_dma_addr[i],
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
479
if (!bp->pf.hwrm_cmd_req_addr[i])
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
483
struct bnxt_vf_info *vf = &bp->pf.vf[k];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
485
vf->hwrm_cmd_req_addr = bp->pf.hwrm_cmd_req_addr[i] +
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
488
bp->pf.hwrm_cmd_req_dma_addr[i] + j *
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
495
bp->pf.vf_event_bmap = kzalloc(16, GFP_KERNEL);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
496
if (!bp->pf.vf_event_bmap)
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
499
bp->pf.hwrm_cmd_req_pages = nr_pages;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
512
req->req_buf_num_pages = cpu_to_le16(bp->pf.hwrm_cmd_req_pages);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
515
req->req_buf_page_addr0 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[0]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
516
req->req_buf_page_addr1 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[1]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
517
req->req_buf_page_addr2 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[2]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
518
req->req_buf_page_addr3 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[3]);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
533
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
59
if (!bp->pf.active_vfs) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
622
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
63
if (vf_id >= bp->pf.active_vfs) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
648
if (pf->vf_resv_strategy == BNXT_VF_RESV_STRATEGY_MINIMAL_STATIC) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
652
if (pf->vf_resv_strategy == BNXT_VF_RESV_STRATEGY_MINIMAL ||
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
653
pf->vf_resv_strategy == BNXT_VF_RESV_STRATEGY_MINIMAL_STATIC) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
667
vf_vnics >= pf->max_vfs) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
671
vf_vnics = (vf_vnics - pf->max_vfs + num_vfs) / num_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
702
struct bnxt_vf_info *vf = &pf->vf[i];
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
704
vf->fw_fid = pf->first_vf_id + i;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
716
pf->active_vfs = i + 1;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
719
if (pf->active_vfs) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
720
u16 n = pf->active_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
734
rc = pf->active_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
747
struct bnxt_pf_info *pf = &bp->pf;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
808
req->fid = cpu_to_le16(pf->first_vf_id + i);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
812
pf->active_vfs = i + 1;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
813
pf->vf[i].fw_fid = le16_to_cpu(req->fid);
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
814
rc = __bnxt_hwrm_get_tx_rings(bp, pf->vf[i].fw_fid,
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
821
if (pf->active_vfs) {
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
829
rc = pf->active_vfs;
drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c
86
vf = &bp->pf.vf[vf_id];
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1502
flow->l2_key.dir = (bp->pf.fw_fid == src_fid) ? BNXT_DIR_RX : BNXT_DIR_TX;
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1509
flow->src_fid = bp->pf.fw_fid;
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1880
return bnxt_tc_setup_flower(bp, bp->pf.fw_fid, flower);
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
355
actions->dst_fid = bp->pf.fw_fid;
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
64
return bp->pf.fw_fid;
drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c
100
if (BNXT_PF(bp) && !bp->pf.port_id &&
drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c
448
edev->pf_port_id = bp->pf.port_id;
drivers/net/ethernet/broadcom/bnxt/bnxt_vfr.c
154
int vf_fid = bp->pf.vf[vf_rep->vf_idx].fw_fid;
drivers/net/ethernet/broadcom/bnxt/bnxt_vfr.c
480
bnxt_vf_rep_eth_addr_gen(bp->pf.mac_addr, vf_rep->vf_idx,
drivers/net/ethernet/broadcom/bnxt/bnxt_vfr.c
82
req->fid = cpu_to_le16(bp->pf.vf[vf_rep->vf_idx].fw_fid);
drivers/net/ethernet/broadcom/bnxt/bnxt_vfr.h
31
return bp->pf.vf[vf_rep->vf_idx].fw_fid;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1101
STAT_GENET_MIB_RX("rx_pause", mib.rx.pf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1128
STAT_GENET_MIB_TX("tx_pause", mib.tx.pf),
drivers/net/ethernet/broadcom/genet/bcmgenet.h
102
u32 pf; /* RO # of Received pause frame pkt */
drivers/net/ethernet/broadcom/genet/bcmgenet.h
123
u32 pf; /* RO # of xmited pause frame count */
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
131
#define CN23XX_SLI_PKT_MAC_RINFO64(mac, pf) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
134
((pf) * CN23XX_PF_RINFO_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
343
#define CN23XX_SLI_MAC_PF_MBOX_INT(mac, pf) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
346
(pf) * CN23XX_PF_INT_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
407
#define CN23XX_SLI_MAC_PF_INT_SUM64(mac, pf) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
410
((pf) * CN23XX_PF_INT_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
412
#define CN23XX_SLI_MAC_PF_INT_ENB64(mac, pf) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
415
((pf) * CN23XX_PF_INT_OFFSET))
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1311
win_pf = is_t4(adap->params.chip) ? 0 : PFNUM_V(adap->pf);
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2158
rc = t4_query_params(padap, padap->mbox, padap->pf, 0, 2, para, val);
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2176
rc = t4_query_params(padap, padap->mbox, padap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
520
vaddr = t4_eeprom_ptov(addr, padap->pf, EEPROMPFSIZE);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1112
unsigned int pf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1351
uint32_t pf:PF_BITWIDTH; /* PCI-E PF ID */
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1889
int port, int pf, int vf, u8 mac[]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1890
int t4_port_init(struct adapter *adap, int mbox, int pf, int vf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1891
int t4_init_port_mirror(struct port_info *pi, u8 mbox, u8 port, u8 pf, u8 vf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1969
int t4_query_params(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1972
int t4_query_params_ns(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1975
int t4_query_params_rw(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1979
unsigned int pf, unsigned int vf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1982
int t4_set_params(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1985
int t4_cfg_pfvf(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1991
unsigned int pf, unsigned int vf, unsigned int nmac, u8 *mac,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1994
unsigned int pf, unsigned int vf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2034
int t4_iq_stop(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2037
int t4_iq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2040
int t4_eth_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2042
int t4_ctrl_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2044
int t4_ofld_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1955
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2292
int pf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2301
for (pf = 0; pf < 8; pf++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2302
pfconf[pf].rss_pf_map = rss_pf_map;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2303
pfconf[pf].rss_pf_mask = rss_pf_mask;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2304
t4_read_rss_pf_config(adapter, pf, &pfconf[pf].rss_pf_config,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1042
ret = t4_set_params(adap, adap->mbox, adap->pf, 0, 1, &param, &val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1100
ret = t4_set_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1193
int vaddr = t4_eeprom_ptov(phys_addr, adap->pf, EEPROMPFSIZE);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1202
int vaddr = t4_eeprom_ptov(phys_addr, adap->pf, EEPROMPFSIZE);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1245
if (adapter->pf > 0) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1246
u32 start = 1024 + adapter->pf * EEPROMPFSIZE;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1513
if (master_vld && (master != adap->pf)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
399
t4_restart_aneg(p->adapter, p->adapter->pf, p->tx_chan);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
416
return t4_identify_port(adap, adap->pf, netdev2pinfo(dev)->viid, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1084
if (fs->val.pf && !fs->mask.pf)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1085
fs->mask.pf |= ~0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1271
(fs->val.pf << 13) |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1438
f->fs.val.ovlan = (fs->val.pf << 13) | fs->val.vf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1439
f->fs.mask.ovlan = (fs->mask.pf << 13) | fs->mask.vf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1685
f->fs.val.ovlan = (fs->val.pf << 13) | fs->val.vf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1686
f->fs.mask.ovlan = (fs->mask.pf << 13) | fs->mask.vf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
296
if (fs->val.pf > 0x7 || fs->val.vf > 0x7f)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
298
fs->mask.pf &= 0x7;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1259
err = t4_set_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1528
ret = t4_init_port_mirror(pi, adap->mbox, pi->port_id, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1553
t4_free_vi(adap, adap->mbox, adap->pf, 0, pi->viid_mirror);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1579
t4_free_vi(adap, adap->mbox, adap->pf, 0, pi->viid_mirror);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
258
err = t4_set_params_timeout(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2886
ret = t4_enable_pi_params(adapter, adapter->pf, pi,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3146
mbox = pi->adapter->pf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3219
macaddr[5] = adap->pf * nvfs + vf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3305
ret = t4_set_params(adap, adap->mbox, adap->pf, vf + 1, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3310
ret, adap->pf, vf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3315
adap->pf, vf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3365
ret = t4_set_params(adap, adap->mbox, adap->pf, vf + 1, 1, &fw_pfvf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3370
ret, adap->pf, vf, class_id);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3374
adap->pf, vf, class_id);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3399
ret, (vlan ? "setting" : "clearing"), adap->pf, vf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3433
ret = t4_set_params(adap, adap->mbox, adap->pf, vf + 1, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3438
ret, adap->pf, vf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4042
ret = t4_query_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4200
ret = t4_config_glbl_rss(adap, adap->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4207
ret = t4_cfg_pfvf(adap, adap->mbox, adap->pf, 0, adap->sge.egr_sz, 64,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4251
return t4_early_init(adap, adap->pf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4516
adapter->pf, 0, 1, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4569
ret = t4_set_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4865
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4928
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1, &v, &port_vec);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4948
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4973
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 6, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4986
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4999
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5018
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5062
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5070
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5084
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5101
(void)t4_set_params(adap, adap->mbox, adap->pf, 0, 1, params, val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5113
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5120
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5129
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5139
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5173
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 6,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5211
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5228
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 6,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5241
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5259
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 6, params,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5272
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2, params,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5289
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1, params,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5295
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1, params,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5303
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5312
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5332
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5346
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5476
if (t4_fw_hello(adap, adap->mbox, adap->pf, MASTER_MUST, NULL) < 0)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5486
ret = t4_alloc_vi(adap, adap->mbox, pi->tx_chan, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5540
if (adapter->pf != 4)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5565
if (adapter->pf != 4)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5587
err = t4_port_init(adapter, adapter->pf, adapter->pf, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6187
t4_free_vi(adapter, adapter->mbox, adapter->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6193
t4_fw_bye(adapter, adapter->pf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6317
FW_PFVF_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6328
adap->pf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6633
adapter->pf = func;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6719
adapter->pf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
7089
if (adapter->pf == 4) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
7174
if (adapter->pf == 4) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
785
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1, &param, &val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
796
ret = t4_set_params(adap, adap->mbox, adap->pf, 0, 1, &param, &val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
814
ret = t4_set_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
334
fs->val.pf = adap->pf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_thermal.c
23
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_thermal.c
53
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
195
ret = t4_set_params(adap, adap->mbox, adap->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
226
t4_set_params(adap, adap->mbox, adap->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
410
t4_ofld_eq_free(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
596
lld->pf = adap->pf;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
694
ret = t4_set_params(adap, adap->mbox, adap->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
712
ret = t4_set_params(adap, adap->mbox, adap->pf,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
456
unsigned int pf; /* Physical Function we're using */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
497
FT_VNID_ID_PF_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/sched.c
76
unsigned int pf = adap->pf;
drivers/net/ethernet/chelsio/cxgb4/sched.c
95
pf = adap->pf;
drivers/net/ethernet/chelsio/cxgb4/sched.c
98
err = t4_set_params(adap, adap->mbox, pf, vf, 1,
drivers/net/ethernet/chelsio/cxgb4/sge.c
1694
TXPKT_PF_V(adap->pf);
drivers/net/ethernet/chelsio/cxgb4/sge.c
2323
TXPKT_PF_V(adap->pf));
drivers/net/ethernet/chelsio/cxgb4/sge.c
2581
flowc->mnemval[0].val = cpu_to_be32(FW_PFVF_CMD_PFN_V(adap->pf));
drivers/net/ethernet/chelsio/cxgb4/sge.c
2695
ctrl0 = TXPKT_OPCODE_V(CPL_TX_PKT_XT) | TXPKT_PF_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/sge.c
4383
FW_IQ_CMD_PFN_V(adap->pf) | FW_IQ_CMD_VFN_V(0));
drivers/net/ethernet/chelsio/cxgb4/sge.c
4525
ret = t4_set_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4600
FW_EQ_ETH_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/sge.c
4691
FW_EQ_CTRL_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/sge.c
4737
return t4_set_params(adap, adap->mbox, adap->pf, 0, 1, &param, &val);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4766
FW_EQ_OFLD_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/sge.c
4858
t4_iq_free(adap, adap->mbox, adap->pf, 0, FW_IQ_TYPE_FL_INT_CAP,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4881
t4_ofld_eq_free(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4905
t4_iq_stop(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4925
t4_eth_eq_free(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4941
t4_ctrl_eq_free(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4964
t4_eth_eq_free(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
10231
FW_ACL_MAC_CMD_PFN_V(adapter->pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
10482
FW_ACL_VLAN_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
154
u32 req = FUNCTION_V(adap->pf) | REGISTER_V(reg);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
2840
ret = t4_query_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
2871
FW_PFVF_CMD_PFN_V(adapter->pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3208
ret = t4_query_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3245
ret = t4_query_params(adapter, adapter->mbox, adapter->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3722
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3790
ret = t4_query_params_rw(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3817
ret = t4_set_params_timeout(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3855
FW_PARAMS_CMD_PFN_V(adap->pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5080
u32 pf = CHELSIO_CHIP_VERSION(adapter->params.chip) <= CHELSIO_T5 ?
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5093
t4_set_reg_field(adapter, PL_INT_MAP0_A, 0, 1 << pf);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5106
u32 whoami, pf;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5112
pf = CHELSIO_CHIP_VERSION(adapter->params.chip) <= CHELSIO_T5 ?
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5116
t4_set_reg_field(adapter, PL_INT_MAP0_A, 1 << pf, 0);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
6176
ret = t4_query_params_ns(adapter, adapter->mbox, adapter->pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
6222
ret = t4_query_params_ns(adapter, adapter->mbox, adapter->pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
644
win_pf = is_t4(adap->params.chip) ? 0 : PFNUM_V(adap->pf);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
6817
ret = t4_query_params(adap, adap->mbox, adap->pf, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7444
int t4_query_params_rw(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7458
FW_PARAMS_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7476
int t4_query_params(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7480
return t4_query_params_rw(adap, mbox, pf, vf, nparams, params, val, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7484
int t4_query_params_ns(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7488
return t4_query_params_rw(adap, mbox, pf, vf, nparams, params, val, 0,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
750
(FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(adap->pf));
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7507
unsigned int pf, unsigned int vf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7520
FW_PARAMS_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7545
int t4_set_params(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7549
return t4_set_params_timeout(adap, mbox, pf, vf, nparams, params, val,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7574
int t4_cfg_pfvf(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7584
FW_CMD_WRITE_F | FW_PFVF_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7621
unsigned int pf, unsigned int vf, unsigned int nmac, u8 *mac,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7630
FW_VI_CMD_PFN_V(pf) | FW_VI_CMD_VFN_V(vf));
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7677
int t4_free_vi(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
7686
FW_VI_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8323
int t4_iq_stop(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8331
FW_CMD_EXEC_F | FW_IQ_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8354
int t4_iq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8362
FW_CMD_EXEC_F | FW_IQ_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8382
int t4_eth_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8390
FW_EQ_ETH_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8407
int t4_ctrl_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8415
FW_EQ_CTRL_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8432
int t4_ofld_eq_free(struct adapter *adap, unsigned int mbox, unsigned int pf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
8440
FW_EQ_OFLD_CMD_PFN_V(pf) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9387
(HOSTPAGESIZEPF1_S - HOSTPAGESIZEPF0_S) * adapter->pf);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9393
(QUEUESPERPAGEPF1_S - QUEUESPERPAGEPF0_S) * adapter->pf);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9431
ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9601
int port, int pf, int vf, u8 mac[])
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9625
ret = t4_set_params(adapter, mbox, pf, vf, 1, &param, &val);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9665
ret = t4_alloc_vi(pi->adapter, mbox, port, pf, vf, 1, mac, &rss_size,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9696
int t4_port_init(struct adapter *adap, int mbox, int pf, int vf)
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9707
ret = t4_init_portinfo(pi, mbox, j, pf, vf, addr);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9717
int t4_init_port_mirror(struct port_info *pi, u8 mbox, u8 port, u8 pf, u8 vf,
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9722
ret = t4_alloc_vi(pi->adapter, mbox, port, pf, vf, 1, NULL, NULL,
drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h
2526
} pf;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
916
unsigned int pf, s_hps, s_qpp;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
939
pf = t4vf_get_pf_from_vf(adapter);
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
941
(HOSTPAGESIZEPF1_S - HOSTPAGESIZEPF0_S) * pf);
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
947
(QUEUESPERPAGEPF1_S - QUEUESPERPAGEPF0_S) * pf);
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
484
TXPKT_PF_V(adap->pf);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
1043
TXPKT_PF_V(tx_info->adap->pf));
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
1628
TXPKT_PF_V(tx_info->adap->pf));
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
179
FLOWC_PARAM(PFNVFN, FW_PFVF_CMD_PFN_V(csk->cdev->lldi->pf));
drivers/net/ethernet/freescale/enetc/enetc4_debugfs.c
34
struct enetc_pf *pf;
drivers/net/ethernet/freescale/enetc/enetc4_debugfs.c
38
pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc4_debugfs.c
39
num_si = pf->caps.num_vsi + 1;
drivers/net/ethernet/freescale/enetc/enetc4_debugfs.c
53
if (!pf->num_mfe)
drivers/net/ethernet/freescale/enetc/enetc4_debugfs.c
58
for (i = 0; i < pf->num_mfe; i++) {
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1000
enetc_mdiobus_destroy(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1007
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1010
enetc_mdiobus_destroy(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1117
struct enetc_pf *pf;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1120
err = enetc_pci_probe(pdev, KBUILD_MODNAME, sizeof(*pf));
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1145
pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1146
err = enetc4_pf_init(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1161
enetc4_pf_free(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1169
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
1173
enetc4_pf_free(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
129
static void enetc4_pf_clear_maft_entries(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
133
for (i = 0; i < pf->num_mfe; i++)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
134
ntmp_maft_delete_entry(&pf->si->ntmp_user, i);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
136
pf->num_mfe = 0;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
139
static int enetc4_pf_add_maft_entries(struct enetc_pf *pf,
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
150
err = ntmp_maft_add_entry(&pf->si->ntmp_user, i, &maft);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
152
pf->num_mfe = i;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
157
pf->num_mfe = mac_cnt;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
162
enetc4_pf_clear_maft_entries(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
167
static int enetc4_pf_set_uc_exact_filter(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
169
int max_num_mfe = pf->caps.mac_filter_num;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
171
struct net_device *ndev = pf->si->ndev;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
172
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
185
enetc4_pf_clear_maft_entries(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
212
enetc4_pf_clear_maft_entries(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
214
if (!enetc4_pf_add_maft_entries(pf, mac_tbl, i))
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
227
static void enetc4_pf_set_mac_hash_filter(struct enetc_pf *pf, int type)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
229
struct net_device *ndev = pf->si->ndev;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
231
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
236
mac_filter = &pf->mac_filter[UC];
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
24
static void enetc4_get_port_caps(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
246
mac_filter = &pf->mac_filter[MC];
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
257
static void enetc4_pf_set_mac_filter(struct enetc_pf *pf, int type)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
26
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
265
if (type & ENETC_MAC_FILTER_TYPE_UC && enetc4_pf_set_uc_exact_filter(pf)) {
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
267
enetc4_pf_set_mac_hash_filter(pf, ENETC_MAC_FILTER_TYPE_UC);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
269
enetc4_pf_clear_maft_entries(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
273
enetc4_pf_set_mac_hash_filter(pf, ENETC_MAC_FILTER_TYPE_MC);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
283
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
285
pf->si = si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
286
pf->total_vfs = pci_sriov_get_totalvfs(si->pdev);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
287
pf->ops = &enetc4_pf_ops;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
289
enetc4_get_port_caps(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
30
pf->caps.num_vsi = (val & ECAPR1_NUM_VSI) >> 24;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
309
static void enetc4_default_rings_allocation(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
31
pf->caps.num_msix = ((val & ECAPR1_NUM_MSIX) >> 12) + 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
311
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
316
if (pf->caps.num_rx_bdr < ENETC_SI_MAX_RING_NUM + pf->caps.num_vsi)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
317
num_rx_bdr = pf->caps.num_rx_bdr - pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
321
if (pf->caps.num_tx_bdr < ENETC_SI_MAX_RING_NUM + pf->caps.num_vsi)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
322
num_tx_bdr = pf->caps.num_tx_bdr - pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
329
num_rx_bdr = pf->caps.num_rx_bdr - num_rx_bdr;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
330
rx_rem = num_rx_bdr % pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
331
num_rx_bdr = num_rx_bdr / pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
333
num_tx_bdr = pf->caps.num_tx_bdr - num_tx_bdr;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
334
tx_rem = num_tx_bdr % pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
335
num_tx_bdr = num_tx_bdr / pf->caps.num_vsi;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
337
for (i = 0; i < pf->caps.num_vsi; i++) {
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
34
pf->caps.num_rx_bdr = (val & ECAPR2_NUM_RX_BDR) >> 16;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
345
static void enetc4_allocate_si_rings(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
347
enetc4_default_rings_allocation(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
35
pf->caps.num_tx_bdr = val & ECAPR2_NUM_TX_BDR;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
362
static void enetc4_set_default_si_vlan_promisc(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
364
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
365
int num_si = pf->caps.num_vsi + 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
374
static void enetc4_set_si_msix_num(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
376
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
38
pf->caps.half_duplex = (val & PMCAPR_HD) ? 1 : 0;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
380
total_si = pf->caps.num_vsi + 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
382
num_msix = pf->caps.num_msix / total_si +
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
383
pf->caps.num_msix % total_si - 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
387
num_msix = pf->caps.num_msix / total_si - 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
389
for (i = 0; i < pf->caps.num_vsi; i++)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
393
static void enetc4_enable_all_si(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
395
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
396
int num_si = pf->caps.num_vsi + 1;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
407
static void enetc4_configure_port_si(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
409
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
41
pf->caps.mac_filter_num = val & PSIMAFCAPR_NUM_MAC_AFTE;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
411
enetc4_allocate_si_rings(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
416
enetc4_set_default_si_vlan_promisc(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
421
enetc4_set_si_msix_num(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
423
enetc4_enable_all_si(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
437
static void enetc4_set_trx_frame_size(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
439
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
447
static void enetc4_configure_port(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
449
enetc4_configure_port_si(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
450
enetc4_set_trx_frame_size(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
451
enetc_set_default_rss_key(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
469
static int enetc4_pf_init(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
471
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
475
err = enetc_setup_mac_addresses(dev->of_node, pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
481
err = enetc4_init_ntmp_user(pf->si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
487
enetc4_configure_port(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
492
static void enetc4_pf_free(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
494
enetc4_free_ntmp_user(pf->si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
500
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
523
enetc4_pf_clear_maft_entries(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
530
enetc4_pf_set_mac_filter(pf, type);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
582
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
584
return pf->pcs;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
587
static void enetc4_mac_config(struct enetc_pf *pf, unsigned int mode,
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
590
struct enetc_ndev_priv *priv = netdev_priv(pf->si->ndev);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
591
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
636
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
638
enetc4_mac_config(pf, mode, state->interface);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
668
static void enetc4_set_rgmii_mac(struct enetc_pf *pf, int speed, int duplex)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
670
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
696
static void enetc4_set_rmii_mac(struct enetc_pf *pf, int speed, int duplex)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
698
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
721
static void enetc4_set_hd_flow_control(struct enetc_pf *pf, bool enable)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
723
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
726
if (!pf->caps.half_duplex)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
737
static void enetc4_set_rx_pause(struct enetc_pf *pf, bool rx_pause)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
739
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
750
static void enetc4_set_tx_pause(struct enetc_pf *pf, int num_rxbdr, bool tx_pause)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
754
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
789
enetc_port_mac_wr(pf->si, ENETC4_PM_PAUSE_QUANTA(0), init_quanta);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
790
enetc_port_mac_wr(pf->si, ENETC4_PM_PAUSE_THRESH(0), refresh_quanta);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
807
static void enetc4_mac_tx_graceful_stop(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
809
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
810
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
829
static void enetc4_mac_tx_enable(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
831
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
832
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
856
static void enetc4_mac_rx_graceful_stop(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
858
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
859
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
889
static void enetc4_mac_rx_enable(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
891
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
892
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
909
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
910
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
919
enetc4_set_rgmii_mac(pf, speed, duplex);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
922
enetc4_set_rmii_mac(pf, speed, duplex);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
941
enetc4_set_hd_flow_control(pf, hd_fc);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
942
enetc4_set_tx_pause(pf, priv->num_rx_rings, tx_pause);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
943
enetc4_set_rx_pause(pf, rx_pause);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
944
enetc4_mac_tx_enable(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
945
enetc4_mac_rx_enable(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
952
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
954
enetc4_mac_rx_graceful_stop(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
955
enetc4_mac_tx_graceful_stop(pf);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
975
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
979
err = of_get_phy_mode(node, &pf->if_mode);
drivers/net/ethernet/freescale/enetc/enetc4_pf.c
985
err = enetc_mdiobus_create(pf, node);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
107
int enetc_msg_psi_init(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_msg.c
109
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc_msg.c
113
snprintf(pf->msg_int_name, sizeof(pf->msg_int_name), "%s-vfmsg",
drivers/net/ethernet/freescale/enetc/enetc_msg.c
116
err = request_irq(vector, enetc_msg_psi_msix, 0, pf->msg_int_name, si);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
127
INIT_WORK(&pf->msg_task, enetc_msg_task);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
129
for (i = 0; i < pf->num_vfs; i++) {
drivers/net/ethernet/freescale/enetc/enetc_msg.c
149
void enetc_msg_psi_free(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_msg.c
151
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc_msg.c
154
cancel_work_sync(&pf->msg_task);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
159
for (i = 0; i < pf->num_vfs; i++)
drivers/net/ethernet/freescale/enetc/enetc_msg.c
23
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
26
schedule_work(&pf->msg_task);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
33
struct enetc_pf *pf = container_of(work, struct enetc_pf, msg_task);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
34
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_msg.c
47
for (i = 0; i < pf->num_vfs; i++) {
drivers/net/ethernet/freescale/enetc/enetc_msg.c
54
enetc_msg_handle_rxmsg(pf, i, &msg_code);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
66
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
72
msg = &pf->rxmsg[idx];
drivers/net/ethernet/freescale/enetc/enetc_msg.c
95
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_msg.c
99
msg = &pf->rxmsg[idx];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1008
err = of_get_phy_mode(node, &pf->if_mode);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1014
err = enetc_mdiobus_create(pf, node);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1031
enetc_mdiobus_destroy(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1043
kfree(pf->vf_state);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1053
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1058
if (pf->num_vfs)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1064
enetc_mdiobus_destroy(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
1071
kfree(pf->vf_state);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
114
static void enetc_sync_mac_filters(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
116
struct enetc_mac_filter *f = pf->mac_filter;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
117
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
161
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
183
filter = &pf->mac_filter[UC];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
199
filter = &pf->mac_filter[MC];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
212
enetc_sync_mac_filters(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
245
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
248
if (vf >= pf->total_vfs)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
254
vf_state = &pf->vf_state[vf];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
264
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
269
if (vf >= pf->total_vfs)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
283
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
286
if (vf >= pf->total_vfs)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
298
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
306
vf_entries = num_entries / (pf->total_vfs + 1);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
308
for (i = 0; i < pf->total_vfs; i++)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
311
num_entries - vf_entries * pf->total_vfs);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
336
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
34
static struct phylink_pcs *enetc_pf_create_pcs(struct enetc_pf *pf,
drivers/net/ethernet/freescale/enetc/enetc_pf.c
372
num_rings /= pf->total_vfs;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
377
for (i = 0; i < pf->total_vfs; i++)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
455
static void enetc_configure_port(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
457
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
459
enetc_configure_port_mac(pf->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
461
enetc_port_si_configure(pf->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
464
enetc_set_default_rss_key(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
467
enetc_port_assign_rfs_entries(pf->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
470
pf->vlan_promisc_simap = ENETC_VLAN_PROMISC_MAP_ALL;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
471
enetc_set_vlan_promisc(hw, pf->vlan_promisc_simap);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
480
static u16 enetc_msg_pf_set_vf_primary_mac_addr(struct enetc_pf *pf,
drivers/net/ethernet/freescale/enetc/enetc_pf.c
483
struct enetc_vf_state *vf_state = &pf->vf_state[vf_id];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
484
struct enetc_msg_swbd *msg = &pf->rxmsg[vf_id];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
486
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
500
enetc_pf_set_primary_mac_addr(&pf->si->hw, vf_id + 1, addr);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
505
void enetc_msg_handle_rxmsg(struct enetc_pf *pf, int vf_id, u16 *status)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
507
struct enetc_msg_swbd *msg = &pf->rxmsg[vf_id];
drivers/net/ethernet/freescale/enetc/enetc_pf.c
508
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
518
*status = enetc_msg_pf_set_vf_primary_mac_addr(pf, vf_id);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
53
static void enetc_enable_si_vlan_promisc(struct enetc_pf *pf, int si_idx)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
530
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
534
enetc_msg_psi_free(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
535
pf->num_vfs = 0;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
538
pf->num_vfs = num_vfs;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
540
err = enetc_msg_psi_init(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
55
pf->vlan_promisc_simap |= BIT(si_idx);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
556
enetc_msg_psi_free(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
558
pf->num_vfs = 0;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
56
enetc_set_vlan_promisc(&pf->si->hw, pf->vlan_promisc_simap);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
580
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
583
enetc_disable_si_vlan_promisc(pf, 0);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
585
enetc_enable_si_vlan_promisc(pf, 0);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
59
static void enetc_disable_si_vlan_promisc(struct enetc_pf *pf, int si_idx)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
61
pf->vlan_promisc_simap &= ~BIT(si_idx);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
62
enetc_set_vlan_promisc(&pf->si->hw, pf->vlan_promisc_simap);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
641
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
643
return pf->pcs;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
650
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
652
enetc_mac_config(pf->si, state->interface);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
688
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
691
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
692
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
697
priv = netdev_priv(pf->si->ndev);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
699
if (pf->si->hw_features & ENETC_SI_F_QBV)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
764
struct enetc_pf *pf = phylink_to_enetc_pf(config);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
765
struct enetc_si *si = pf->si;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
938
struct enetc_pf *pf;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
955
pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
956
pf->si = si;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
957
pf->ops = &enetc_pf_ops;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
959
pf->total_vfs = pci_sriov_get_totalvfs(pdev);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
960
if (pf->total_vfs) {
drivers/net/ethernet/freescale/enetc/enetc_pf.c
961
pf->vf_state = kzalloc_objs(struct enetc_vf_state,
drivers/net/ethernet/freescale/enetc/enetc_pf.c
962
pf->total_vfs);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
963
if (!pf->vf_state)
drivers/net/ethernet/freescale/enetc/enetc_pf.c
967
err = enetc_setup_mac_addresses(node, pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
971
enetc_configure_port(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.h
34
struct phylink_pcs *(*create_pcs)(struct enetc_pf *pf, struct mii_bus *bus);
drivers/net/ethernet/freescale/enetc/enetc_pf.h
71
int enetc_msg_psi_init(struct enetc_pf *pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.h
72
void enetc_msg_psi_free(struct enetc_pf *pf);
drivers/net/ethernet/freescale/enetc/enetc_pf.h
73
void enetc_msg_handle_rxmsg(struct enetc_pf *pf, int mbox_id, u16 *status);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
10
static void enetc_set_si_hw_addr(struct enetc_pf *pf, int si,
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
13
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
147
if (si->hw_features & ENETC_SI_F_PSFP && pf->ops->enable_psfp &&
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
148
!pf->ops->enable_psfp(priv)) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
15
pf->ops->set_si_primary_mac(hw, si, mac_addr);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
160
static int enetc_mdio_probe(struct enetc_pf *pf, struct device_node *np)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
162
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
178
mdio_priv->hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
18
static void enetc_get_si_hw_addr(struct enetc_pf *pf, int si, u8 *mac_addr)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
180
if (is_enetc_rev1(pf->si))
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
191
pf->mdio = bus;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
196
static void enetc_mdio_remove(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
198
if (pf->mdio)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
199
mdiobus_unregister(pf->mdio);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
20
struct enetc_hw *hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
202
static int enetc_imdio_create(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
204
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
210
if (!pf->ops->create_pcs) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
22
pf->ops->get_si_primary_mac(hw, si, mac_addr);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
228
mdio_priv->hw = &pf->si->hw;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
230
if (is_enetc_rev1(pf->si))
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
243
phylink_pcs = pf->ops->create_pcs(pf, bus);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
250
pf->imdio = bus;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
251
pf->pcs = phylink_pcs;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
262
static void enetc_imdio_remove(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
264
if (pf->pcs && pf->ops->destroy_pcs)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
265
pf->ops->destroy_pcs(pf->pcs);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
267
if (pf->imdio) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
268
mdiobus_unregister(pf->imdio);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
269
mdiobus_free(pf->imdio);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
273
static bool enetc_port_has_pcs(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
275
return (pf->if_mode == PHY_INTERFACE_MODE_SGMII ||
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
276
pf->if_mode == PHY_INTERFACE_MODE_1000BASEX ||
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
277
pf->if_mode == PHY_INTERFACE_MODE_2500BASEX ||
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
278
pf->if_mode == PHY_INTERFACE_MODE_USXGMII);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
28
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
281
int enetc_mdiobus_create(struct enetc_pf *pf, struct device_node *node)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
288
err = enetc_mdio_probe(pf, mdio_np);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
295
if (enetc_port_has_pcs(pf)) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
296
err = enetc_imdio_create(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
298
enetc_mdio_remove(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
307
void enetc_mdiobus_destroy(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
309
enetc_mdio_remove(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
310
enetc_imdio_remove(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
317
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
321
pf->phylink_config.dev = &priv->ndev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
322
pf->phylink_config.type = PHYLINK_NETDEV;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
323
pf->phylink_config.mac_capabilities = MAC_ASYM_PAUSE | MAC_SYM_PAUSE |
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
327
pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
329
pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
331
pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
333
pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
335
pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
336
phy_interface_set_rgmii(pf->phylink_config.supported_interfaces);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
338
phylink = phylink_create(&pf->phylink_config, of_fwnode_handle(node),
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
339
pf->if_mode, ops);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
35
enetc_set_si_hw_addr(pf, 0, saddr->sa_data);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
357
void enetc_set_default_rss_key(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
363
enetc_set_rss_key(pf->si, hash_key);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
378
static void enetc_refresh_vlan_ht_filter(struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
382
bitmap_zero(pf->vlan_ht_filter, ENETC_VLAN_HT_SIZE);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
383
for_each_set_bit(i, pf->active_vlans, VLAN_N_VID) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
386
__set_bit(hidx, pf->vlan_ht_filter);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
41
static int enetc_setup_mac_address(struct device_node *np, struct enetc_pf *pf,
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
411
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
414
__set_bit(vid, pf->active_vlans);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
417
if (!__test_and_set_bit(idx, pf->vlan_ht_filter))
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
418
enetc_set_si_vlan_ht_filter(pf->si, 0, *pf->vlan_ht_filter);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
427
struct enetc_pf *pf = enetc_si_priv(priv->si);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
429
if (__test_and_clear_bit(vid, pf->active_vlans)) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
430
enetc_refresh_vlan_ht_filter(pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
431
enetc_set_si_vlan_ht_filter(pf->si, 0, *pf->vlan_ht_filter);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
44
struct device *dev = &pf->si->pdev->dev;
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
57
enetc_get_si_hw_addr(pf, si, mac_addr);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
66
enetc_set_si_hw_addr(pf, si, mac_addr);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
71
int enetc_setup_mac_addresses(struct device_node *np, struct enetc_pf *pf)
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
76
err = enetc_setup_mac_address(np, pf, 0);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
80
for (i = 0; i < pf->total_vfs; i++) {
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
81
err = enetc_setup_mac_address(NULL, pf, i + 1);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.c
94
struct enetc_pf *pf = enetc_si_priv(si);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.h
10
int enetc_mdiobus_create(struct enetc_pf *pf, struct device_node *node);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.h
11
void enetc_mdiobus_destroy(struct enetc_pf *pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.h
15
void enetc_set_default_rss_key(struct enetc_pf *pf);
drivers/net/ethernet/freescale/enetc/enetc_pf_common.h
7
int enetc_setup_mac_addresses(struct device_node *np, struct enetc_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1103
static inline int i40e_get_fd_cnt_all(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1105
return pf->hw.fdir_shared_filter_count + pf->fdir_pf_filter_count;
drivers/net/ethernet/intel/i40e/i40e.h
1116
static inline u64 i40e_read_fd_input_set(struct i40e_pf *pf, u16 addr)
drivers/net/ethernet/intel/i40e/i40e.h
1120
val = i40e_read_rx_ctl(&pf->hw, I40E_PRTQF_FD_INSET(addr, 1));
drivers/net/ethernet/intel/i40e/i40e.h
1122
val += i40e_read_rx_ctl(&pf->hw, I40E_PRTQF_FD_INSET(addr, 0));
drivers/net/ethernet/intel/i40e/i40e.h
1136
static inline void i40e_write_fd_input_set(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e.h
1139
i40e_write_rx_ctl(&pf->hw, I40E_PRTQF_FD_INSET(addr, 1),
drivers/net/ethernet/intel/i40e/i40e.h
1141
i40e_write_rx_ctl(&pf->hw, I40E_PRTQF_FD_INSET(addr, 0),
drivers/net/ethernet/intel/i40e/i40e.h
1164
void i40e_do_reset_safe(struct i40e_pf *pf, u32 reset_flags);
drivers/net/ethernet/intel/i40e/i40e.h
1165
void i40e_do_reset(struct i40e_pf *pf, u32 reset_flags, bool lock_acquired);
drivers/net/ethernet/intel/i40e/i40e.h
1168
void i40e_fill_rss_lut(struct i40e_pf *pf, u8 *lut,
drivers/net/ethernet/intel/i40e/i40e.h
1170
struct i40e_vsi *i40e_find_vsi_from_id(struct i40e_pf *pf, u16 id);
drivers/net/ethernet/intel/i40e/i40e.h
1177
i40e_find_vsi_by_type(struct i40e_pf *pf, u16 type)
drivers/net/ethernet/intel/i40e/i40e.h
1182
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e.h
1192
int i40e_fetch_switch_configuration(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e.h
1197
void i40e_fdir_check_and_reenable(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1198
u32 i40e_get_current_fd_count(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1199
u32 i40e_get_current_atr_cnt(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1200
u32 i40e_get_global_fd_count(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1201
bool i40e_set_ntuple(struct i40e_pf *pf, netdev_features_t features);
drivers/net/ethernet/intel/i40e/i40e.h
1207
struct i40e_vsi *i40e_vsi_setup(struct i40e_pf *pf, u8 type,
drivers/net/ethernet/intel/i40e/i40e.h
1210
void i40e_service_event_schedule(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1214
int i40e_control_wait_tx_q(int seid, struct i40e_pf *pf, int pf_q, bool is_xdp,
drivers/net/ethernet/intel/i40e/i40e.h
1216
int i40e_control_wait_rx_q(struct i40e_pf *pf, int pf_q, bool enable);
drivers/net/ethernet/intel/i40e/i40e.h
1221
int i40e_reconfig_rss_queues(struct i40e_pf *pf, int queue_count);
drivers/net/ethernet/intel/i40e/i40e.h
1222
struct i40e_veb *i40e_veb_setup(struct i40e_pf *pf, u16 uplink_seid,
drivers/net/ethernet/intel/i40e/i40e.h
1230
void i40e_pf_reset_stats(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1232
void i40e_dbg_pf_init(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1233
void i40e_dbg_pf_exit(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1237
static inline void i40e_dbg_pf_init(struct i40e_pf *pf) {}
drivers/net/ethernet/intel/i40e/i40e.h
1238
static inline void i40e_dbg_pf_exit(struct i40e_pf *pf) {}
drivers/net/ethernet/intel/i40e/i40e.h
1243
int i40e_lan_add_device(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1244
int i40e_lan_del_device(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1245
void i40e_client_subtask(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1246
void i40e_notify_client_of_l2_param_changes(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1247
void i40e_notify_client_of_netdev_close(struct i40e_pf *pf, bool reset);
drivers/net/ethernet/intel/i40e/i40e.h
1248
void i40e_notify_client_of_vf_enable(struct i40e_pf *pf, u32 num_vfs);
drivers/net/ethernet/intel/i40e/i40e.h
1249
void i40e_notify_client_of_vf_reset(struct i40e_pf *pf, u32 vf_id);
drivers/net/ethernet/intel/i40e/i40e.h
1250
void i40e_client_update_msix_info(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1251
int i40e_vf_client_capable(struct i40e_pf *pf, u32 vf_id);
drivers/net/ethernet/intel/i40e/i40e.h
1259
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e.h
1260
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e.h
1270
void i40e_irq_dynamic_disable_icr0(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1271
void i40e_irq_dynamic_enable_icr0(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1289
static inline bool i40e_is_sw_dcb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1291
return test_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags);
drivers/net/ethernet/intel/i40e/i40e.h
1295
void i40e_dcbnl_flush_apps(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e.h
1300
bool i40e_dcb_need_reconfig(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e.h
1303
int i40e_hw_dcb_config(struct i40e_pf *pf, struct i40e_dcbx_config *new_cfg);
drivers/net/ethernet/intel/i40e/i40e.h
1304
int i40e_dcb_sw_default_config(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1306
void i40e_ptp_rx_hang(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1307
void i40e_ptp_tx_hang(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1308
void i40e_ptp_tx_hwtstamp(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1309
void i40e_ptp_rx_hwtstamp(struct i40e_pf *pf, struct sk_buff *skb, u8 index);
drivers/net/ethernet/intel/i40e/i40e.h
1310
void i40e_ptp_set_increment(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1316
void i40e_ptp_save_hw_time(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1317
void i40e_ptp_restore_hw_time(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1318
void i40e_ptp_init(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1319
void i40e_ptp_stop(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1320
int i40e_ptp_alloc_pins(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1323
int i40e_get_partition_bw_setting(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1324
int i40e_set_partition_bw_setting(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e.h
1351
static inline bool i40e_is_tc_mqprio_enabled(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1353
return test_bit(I40E_FLAG_TC_MQPRIO_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e.h
1373
i40e_pf_get_vsi_by_seid(struct i40e_pf *pf, u16 seid)
drivers/net/ethernet/intel/i40e/i40e.h
1378
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e.h
1391
static inline struct i40e_vsi *i40e_pf_get_main_vsi(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1393
return (pf->lan_vsi != I40E_NO_VSI) ? pf->vsi[pf->lan_vsi] : NULL;
drivers/net/ethernet/intel/i40e/i40e.h
1402
i40e_pf_get_veb_by_seid(struct i40e_pf *pf, u16 seid)
drivers/net/ethernet/intel/i40e/i40e.h
1407
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e.h
1420
static inline struct i40e_veb *i40e_pf_get_main_veb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1422
return (pf->lan_veb != I40E_NO_VEB) ? pf->veb[pf->lan_veb] : NULL;
drivers/net/ethernet/intel/i40e/i40e.h
1425
static inline u32 i40e_get_max_num_descriptors(const struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e.h
1427
const struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e.h
37
#define i40e_default_queues_per_vmdq(pf) \
drivers/net/ethernet/intel/i40e/i40e.h
38
(test_bit(I40E_HW_CAP_RSS_AQ, (pf)->hw.caps) ? 4 : 1)
drivers/net/ethernet/intel/i40e/i40e.h
41
#define i40e_pf_get_max_q_per_tc(pf) \
drivers/net/ethernet/intel/i40e/i40e.h
42
(test_bit(I40E_HW_CAP_128_QP_RSS, (pf)->hw.caps) ? 128 : 64)
drivers/net/ethernet/intel/i40e/i40e.h
706
__i40e_pf_next_vsi(struct i40e_pf *pf, int *idx)
drivers/net/ethernet/intel/i40e/i40e.h
708
while (*idx < pf->num_alloc_vsi) {
drivers/net/ethernet/intel/i40e/i40e.h
709
if (pf->vsi[*idx])
drivers/net/ethernet/intel/i40e/i40e.h
710
return pf->vsi[*idx];
drivers/net/ethernet/intel/i40e/i40e.h
730
__i40e_pf_next_veb(struct i40e_pf *pf, int *idx)
drivers/net/ethernet/intel/i40e/i40e.h
733
if (pf->veb[*idx])
drivers/net/ethernet/intel/i40e/i40e.h
734
return pf->veb[*idx];
drivers/net/ethernet/intel/i40e/i40e.h
793
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e.h
973
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
108
void i40e_notify_client_of_l2_param_changes(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
110
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
111
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
117
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
122
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
147
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
153
wr32(&pf->hw, reg_idx, I40E_PFINT_LNKLSTN_FIRSTQ_INDX_MASK);
drivers/net/ethernet/intel/i40e/i40e_client.c
166
void i40e_notify_client_of_netdev_close(struct i40e_pf *pf, bool reset)
drivers/net/ethernet/intel/i40e/i40e_client.c
168
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
173
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
178
dev_dbg(&pf->pdev->dev, "Client is not open, abort close\n");
drivers/net/ethernet/intel/i40e/i40e_client.c
193
void i40e_notify_client_of_vf_reset(struct i40e_pf *pf, u32 vf_id)
drivers/net/ethernet/intel/i40e/i40e_client.c
195
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
200
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
205
dev_dbg(&pf->pdev->dev, "Client is not open, abort vf-reset\n");
drivers/net/ethernet/intel/i40e/i40e_client.c
218
void i40e_notify_client_of_vf_enable(struct i40e_pf *pf, u32 num_vfs)
drivers/net/ethernet/intel/i40e/i40e_client.c
220
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
225
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
231
dev_dbg(&pf->pdev->dev, "Client is not open, abort vf-enable\n");
drivers/net/ethernet/intel/i40e/i40e_client.c
245
int i40e_vf_client_capable(struct i40e_pf *pf, u32 vf_id)
drivers/net/ethernet/intel/i40e/i40e_client.c
247
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
253
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
267
void i40e_client_update_msix_info(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
269
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
274
cdev->lan_info.msix_count = pf->num_iwarp_msix;
drivers/net/ethernet/intel/i40e/i40e_client.c
275
cdev->lan_info.msix_entries = &pf->msix_entries[pf->iwarp_base_vector];
drivers/net/ethernet/intel/i40e/i40e_client.c
334
static void i40e_client_add_instance(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
336
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
344
cdev->lan_info.pf = (void *)pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
346
cdev->lan_info.pcidev = pf->pdev;
drivers/net/ethernet/intel/i40e/i40e_client.c
347
cdev->lan_info.fid = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_client.c
349
cdev->lan_info.hw_addr = pf->hw.hw_addr;
drivers/net/ethernet/intel/i40e/i40e_client.c
354
cdev->lan_info.fw_maj_ver = pf->hw.aq.fw_maj_ver;
drivers/net/ethernet/intel/i40e/i40e_client.c
355
cdev->lan_info.fw_min_ver = pf->hw.aq.fw_min_ver;
drivers/net/ethernet/intel/i40e/i40e_client.c
356
cdev->lan_info.fw_build = pf->hw.aq.fw_build;
drivers/net/ethernet/intel/i40e/i40e_client.c
367
dev_err(&pf->pdev->dev, "MAC address list is empty!\n");
drivers/net/ethernet/intel/i40e/i40e_client.c
369
pf->cinst = cdev;
drivers/net/ethernet/intel/i40e/i40e_client.c
371
cdev->lan_info.msix_count = pf->num_iwarp_msix;
drivers/net/ethernet/intel/i40e/i40e_client.c
372
cdev->lan_info.msix_entries = &pf->msix_entries[pf->iwarp_base_vector];
drivers/net/ethernet/intel/i40e/i40e_client.c
381
pf->cinst = NULL;
drivers/net/ethernet/intel/i40e/i40e_client.c
390
void i40e_client_del_instance(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
392
kfree(pf->cinst);
drivers/net/ethernet/intel/i40e/i40e_client.c
393
pf->cinst = NULL;
drivers/net/ethernet/intel/i40e/i40e_client.c
400
void i40e_client_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
402
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
407
if (!test_and_clear_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_client.c
409
cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
412
if (test_bit(__I40E_DOWN, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_client.c
413
test_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/i40e/i40e_client.c
457
int i40e_lan_add_device(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
464
if (ldev->pf == pf) {
drivers/net/ethernet/intel/i40e/i40e_client.c
474
ldev->pf = pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
477
dev_info(&pf->pdev->dev, "Added LAN device PF%d bus=0x%02x dev=0x%02x func=0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_client.c
478
pf->hw.pf_id, pf->hw.bus.bus_id,
drivers/net/ethernet/intel/i40e/i40e_client.c
479
pf->hw.bus.device, pf->hw.bus.func);
drivers/net/ethernet/intel/i40e/i40e_client.c
481
i40e_client_add_instance(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
483
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_client.c
484
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
497
int i40e_lan_del_device(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_client.c
499
struct auxiliary_device *aux_dev = pf->cinst->lan_info.aux_dev;
drivers/net/ethernet/intel/i40e/i40e_client.c
507
i40e_client_del_instance(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
511
if (ldev->pf == pf) {
drivers/net/ethernet/intel/i40e/i40e_client.c
512
dev_info(&pf->pdev->dev, "Deleted LAN device PF%d bus=0x%02x dev=0x%02x func=0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_client.c
513
pf->hw.pf_id, pf->hw.bus.bus_id,
drivers/net/ethernet/intel/i40e/i40e_client.c
514
pf->hw.bus.device, pf->hw.bus.func);
drivers/net/ethernet/intel/i40e/i40e_client.c
539
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
540
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_client.c
546
dev_err(&pf->pdev->dev, "Unable to send iWarp message to VF, error %d, aq status %d\n",
drivers/net/ethernet/intel/i40e/i40e_client.c
564
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
565
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_client.c
580
if ((v_idx >= (pf->iwarp_base_vector + pf->num_iwarp_msix)) ||
drivers/net/ethernet/intel/i40e/i40e_client.c
581
(v_idx < pf->iwarp_base_vector))
drivers/net/ethernet/intel/i40e/i40e_client.c
633
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
637
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_client.c
640
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_client.c
643
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
645
pf->hw.pf_id, reset_level);
drivers/net/ethernet/intel/i40e/i40e_client.c
649
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
668
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(ldev->pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
669
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
678
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_client.c
679
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_client.c
680
err = i40e_aq_get_vsi_params(&pf->hw, &ctxt, NULL);
drivers/net/ethernet/intel/i40e/i40e_client.c
683
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
685
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_client.c
701
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
703
pf->hw.pf_id, flag);
drivers/net/ethernet/intel/i40e/i40e_client.c
709
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
712
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_client.c
720
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
722
pf->cinst->client = client;
drivers/net/ethernet/intel/i40e/i40e_client.c
723
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_client.c
724
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_client.c
730
struct i40e_pf *pf = ldev->pf;
drivers/net/ethernet/intel/i40e/i40e_client.c
731
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
736
while (test_and_set_bit(__I40E_SERVICE_SCHED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_client.c
745
pf->cinst->client = NULL;
drivers/net/ethernet/intel/i40e/i40e_client.c
746
clear_bit(__I40E_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_client.c
84
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_client.c
85
struct i40e_client_instance *cdev = pf->cinst;
drivers/net/ethernet/intel/i40e/i40e_client.c
90
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_client.c
95
dev_dbg(&pf->pdev->dev, "Client is not open, abort virtchnl_receive\n");
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
1001
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags) && !(pf->hw.func_caps.iscsi))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
1008
i40e_dcbnl_del_app(pf, &app);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
1021
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
1024
if (!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
108
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
112
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
113
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
116
old_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
118
pf->tmp_cfg = *old_cfg;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
121
pf->tmp_cfg.etscfg.willing = ets->willing;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
122
pf->tmp_cfg.etscfg.maxtcs = I40E_MAX_TRAFFIC_CLASS;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
123
pf->tmp_cfg.etscfg.cbs = ets->cbs;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
125
pf->tmp_cfg.etscfg.tcbwtable[i] = ets->tc_tx_bw[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
126
pf->tmp_cfg.etscfg.tsatable[i] = ets->tc_tsa[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
127
pf->tmp_cfg.etscfg.prioritytable[i] = ets->prio_tc[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
128
pf->tmp_cfg.etsrec.tcbwtable[i] = ets->tc_reco_bw[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
129
pf->tmp_cfg.etsrec.tsatable[i] = ets->tc_reco_tsa[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
130
pf->tmp_cfg.etsrec.prioritytable[i] = ets->reco_prio_tc[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
134
ret = i40e_hw_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
136
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
139
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
156
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
160
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
161
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
164
old_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
166
pf->tmp_cfg = *old_cfg;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
168
pf->tmp_cfg.pfc.pfccap = pfc->pfc_cap;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
170
pf->tmp_cfg.pfc.pfccap = I40E_MAX_TRAFFIC_CLASS;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
171
pf->tmp_cfg.pfc.pfcenable = pfc->pfc_en;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
173
ret = i40e_hw_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
175
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
178
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
195
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
200
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
201
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
204
old_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
220
pf->tmp_cfg = *old_cfg;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
222
pf->tmp_cfg.app[pf->tmp_cfg.numapps++] = new_app;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
224
ret = i40e_hw_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
226
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
229
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
247
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
251
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
252
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
259
old_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
265
pf->tmp_cfg = *old_cfg;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
268
for (i = 1; i < pf->tmp_cfg.numapps; i++) {
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
269
if (app->selector == pf->tmp_cfg.app[i].selector &&
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
270
app->protocol == pf->tmp_cfg.app[i].protocolid &&
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
271
app->priority == pf->tmp_cfg.app[i].priority) {
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
273
pf->tmp_cfg.app[i].selector = 0;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
274
pf->tmp_cfg.app[i].protocolid = 0;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
275
pf->tmp_cfg.app[i].priority = 0;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
281
if (i == pf->tmp_cfg.numapps)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
284
pf->tmp_cfg.numapps--;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
286
for (j = i; j < pf->tmp_cfg.numapps; j++)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
287
pf->tmp_cfg.app[j] = old_cfg->app[j + 1];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
289
ret = i40e_hw_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
291
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
294
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
309
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
311
dev_dbg(&pf->pdev->dev, "DCB state=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
312
test_bit(I40E_FLAG_DCB_ENA, pf->flags) ? 1 : 0);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
313
return test_bit(I40E_FLAG_DCB_ENA, pf->flags) ? 1 : 0;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
325
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
328
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
329
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
332
dev_dbg(&pf->pdev->dev, "new state=%d current state=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
333
state, test_bit(I40E_FLAG_DCB_ENA, pf->flags) ? 1 : 0);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
335
if (!state == !test_bit(I40E_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
338
if (i40e_is_sw_dcb(pf)) {
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
340
set_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
341
memcpy(&pf->hw.desired_dcbx_config,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
342
&pf->hw.local_dcbx_config,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
345
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
369
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
37
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
372
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
373
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
385
pf->tmp_cfg.etscfg.prioritytable[i] = tc;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
387
pf->tmp_cfg.etscfg.tsatable[tc] = I40E_IEEE_TSA_ETS;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
388
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
40
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
404
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
406
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
407
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
414
pf->tmp_cfg.etscfg.tcbwtable[pgid] = bw_pct;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
415
dev_dbg(&pf->pdev->dev, "Set PG BW config tc=%d bw_pct=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
43
dcbxcfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
438
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
440
dev_dbg(&pf->pdev->dev, "Rx TC PG Config Not Supported.\n");
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
455
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
457
dev_dbg(&pf->pdev->dev, "Rx BWG PG Config Not Supported.\n");
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
477
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
479
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
480
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
486
*pgid = pf->hw.local_dcbx_config.etscfg.prioritytable[prio];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
487
dev_dbg(&pf->pdev->dev, "Get PG config prio=%d tc=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
502
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
504
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
505
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
511
*bw_pct = pf->hw.local_dcbx_config.etscfg.tcbwtable[pgid];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
512
dev_dbg(&pf->pdev->dev, "Get PG BW config tc=%d bw_pct=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
533
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
535
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
536
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
542
*pgid = pf->hw.local_dcbx_config.etscfg.prioritytable[prio];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
558
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
560
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
561
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
577
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
579
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
580
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
586
pf->tmp_cfg.pfc.pfccap = I40E_MAX_TRAFFIC_CLASS;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
588
pf->tmp_cfg.pfc.pfcenable |= BIT(prio);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
590
pf->tmp_cfg.pfc.pfcenable &= ~BIT(prio);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
591
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
593
prio, setting, pf->tmp_cfg.pfc.pfcenable);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
607
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
609
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
610
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
616
*setting = (pf->hw.local_dcbx_config.pfc.pfcenable >> prio) & 0x1;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
617
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
619
prio, *setting, pf->hw.local_dcbx_config.pfc.pfcenable);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
630
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
633
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
634
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
637
dev_dbg(&pf->pdev->dev, "Commit DCB Configuration to the hardware\n");
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
638
err = i40e_hw_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
653
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
655
if (!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
668
*cap = pf->dcbx_cap;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
678
dev_dbg(&pf->pdev->dev, "Get Capability cap=%d capval=0x%x\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
693
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
695
if (!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
723
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
726
if (pf->hw.local_dcbx_config.pfc.pfcenable)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
742
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
744
dev_dbg(&pf->pdev->dev, "PFC State is modified via PFC config.\n");
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
75
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
757
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
763
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE) ||
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
764
(pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
77
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
779
struct i40e_pf *pf = i40e_netdev_to_pf(netdev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
782
if (pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
792
if (mode == pf->dcbx_cap)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
795
pf->dcbx_cap = mode;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
797
pf->hw.local_dcbx_config.dcbx_mode = I40E_DCBX_MODE_CEE;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
799
pf->hw.local_dcbx_config.dcbx_mode = I40E_DCBX_MODE_IEEE;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
80
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
801
dev_dbg(&pf->pdev->dev, "mode=%d\n", mode);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
813
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
815
return pf->dcbx_cap;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
828
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
834
perm_addr[i] = pf->hw.mac.perm_addr[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
878
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
880
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
886
if (pf->dcbx_cap & DCB_CAP_DCBX_HOST)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
890
if (!test_bit(I40E_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
894
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags) && !(hw->func_caps.iscsi))
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
91
pfc->requests[i] = pf->stats.priority_xoff_tx[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
92
pfc->indications[i] = pf->stats.priority_xoff_rx[i];
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
947
static void i40e_dcbnl_del_app(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
953
i40e_pf_for_each_vsi(pf, v, vsi)
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
956
dev_dbg(&pf->pdev->dev, "Deleting app for VSI seid=%d err=%d sel=%d proto=0x%x prio=%d\n",
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
993
void i40e_dcbnl_flush_apps(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ddp.c
293
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ddp.c
334
istatus = i40e_ddp_does_profile_exist(&pf->hw, &pinfo);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
344
istatus = i40e_ddp_does_profile_overlap(&pf->hw, &pinfo);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
363
status = i40e_write_profile(&pf->hw, profile_hdr, track_id);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
374
status = i40e_rollback_profile(&pf->hw, profile_hdr, track_id);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
383
status = i40e_add_pinfo(&pf->hw, profile_hdr, profile_info_sec,
drivers/net/ethernet/intel/i40e/i40e_ddp.c
390
status = i40e_del_pinfo(&pf->hw, profile_hdr, profile_info_sec,
drivers/net/ethernet/intel/i40e/i40e_ddp.c
408
static int i40e_ddp_restore(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ddp.c
410
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
415
if (!list_empty(&pf->ddp_old_prof)) {
drivers/net/ethernet/intel/i40e/i40e_ddp.c
416
entry = list_first_entry(&pf->ddp_old_prof,
drivers/net/ethernet/intel/i40e/i40e_ddp.c
439
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ddp.c
447
if (pf->hw.bus.func != 0) {
drivers/net/ethernet/intel/i40e/i40e_ddp.c
485
list_add(&list_entry->list, &pf->ddp_old_prof);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
491
if (!list_empty(&pf->ddp_old_prof)) {
drivers/net/ethernet/intel/i40e/i40e_ddp.c
492
status = i40e_ddp_restore(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1001
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1003
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1008
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
101
dev_info(&pf->pdev->dev, " active_filters %u, promisc_threshold %u, overflow promisc %s\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1015
dev_info(&pf->pdev->dev, "port bw: tc_bw_share=%d tc_bw_limit=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1023
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1025
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1030
dev_info(&pf->pdev->dev, "port ets_cfg: %d prio_tc=%d tcbw=%d tctsa=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1036
dev_info(&pf->pdev->dev, "port ets_rec: %d prio_tc=%d tcbw=%d tctsa=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1041
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1045
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1048
dev_info(&pf->pdev->dev, "port app_table: %d prio=%d selector=%d protocol=0x%x\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1054
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1059
dev_info(&pf->pdev->dev, "remote port ets_cfg: %d prio_tc=%d tcbw=%d tctsa=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
106
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1065
dev_info(&pf->pdev->dev, "remote port ets_rec: %d prio_tc=%d tcbw=%d tctsa=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1070
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1076
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1080
dev_info(&pf->pdev->dev, "remote port app_table: %d prio=%d selector=%d protocol=0x%x\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1097
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1102
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1109
ret = i40e_aq_debug_dump(&pf->hw, cluster_id, table_id,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1114
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1116
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
112
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1121
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1130
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1132
dev_info(&pf->pdev->dev, "dump switch\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1133
dev_info(&pf->pdev->dev, "dump vsi [seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1134
dev_info(&pf->pdev->dev, "dump reset stats\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1135
dev_info(&pf->pdev->dev, "dump port\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1136
dev_info(&pf->pdev->dev, "dump vf [vf_id]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1137
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1141
dev_info(&pf->pdev->dev, "debugfs: forcing PFR\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1142
i40e_do_reset_safe(pf, BIT(__I40E_PF_RESET_REQUESTED));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1145
dev_info(&pf->pdev->dev, "debugfs: forcing CoreR\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1146
i40e_do_reset_safe(pf, BIT(__I40E_CORE_RESET_REQUESTED));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1149
dev_info(&pf->pdev->dev, "debugfs: forcing GlobR\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1150
i40e_do_reset_safe(pf, BIT(__I40E_GLOBAL_RESET_REQUESTED));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1158
dev_info(&pf->pdev->dev, "read <reg>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1163
if (address > (pf->ioremap_len - sizeof(u32))) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1164
dev_info(&pf->pdev->dev, "read reg address 0x%08x too large, max=0x%08lx\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1165
address, (unsigned long int)(pf->ioremap_len - sizeof(u32)));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1169
value = rd32(&pf->hw, address);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1170
dev_info(&pf->pdev->dev, "read: 0x%08x = 0x%08x\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1178
dev_info(&pf->pdev->dev, "write <reg> <value>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
118
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1183
if (address > (pf->ioremap_len - sizeof(u32))) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1184
dev_info(&pf->pdev->dev, "write reg address 0x%08x too large, max=0x%08lx\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1185
address, (unsigned long int)(pf->ioremap_len - sizeof(u32)));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1188
wr32(&pf->hw, address, value);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1189
value = rd32(&pf->hw, address);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1190
dev_info(&pf->pdev->dev, "write: 0x%08x = 0x%08x\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1198
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1200
dev_info(&pf->pdev->dev, "vsi clear stats called for all vsi's\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1202
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1204
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1210
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1214
dev_info(&pf->pdev->dev, "clear_stats vsi [seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1217
if (pf->hw.partition_id == 1) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1218
i40e_pf_reset_stats(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1219
dev_info(&pf->pdev->dev, "port stats cleared\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
122
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1221
dev_info(&pf->pdev->dev, "clear port stats not allowed on this port partition\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1224
dev_info(&pf->pdev->dev, "clear_stats vsi [seid] or clear_stats port\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1243
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1250
ret = i40e_asq_send_command(&pf->hw, desc, NULL, 0, NULL);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1252
dev_info(&pf->pdev->dev, "AQ command sent Status : Success\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1254
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1256
desc->opcode, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1258
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1262
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
127
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1292
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1310
ret = i40e_asq_send_command(&pf->hw, desc, buff,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1313
dev_info(&pf->pdev->dev, "AQ command sent Status : Success\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1315
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1317
desc->opcode, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1319
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
132
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1323
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1339
dev_info(&pf->pdev->dev, "FD current total filter count for this interface: %d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1340
i40e_get_current_fd_count(pf));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1343
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1348
ret = i40e_aq_stop_lldp(&pf->hw, false, false, NULL);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1350
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1352
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1355
ret = i40e_aq_add_rem_control_packet_filter(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1356
pf->hw.mac.addr, ETH_P_LLDP, 0,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1360
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1362
__func__, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1366
pf->dcbx_cap = DCB_CAP_DCBX_HOST |
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
137
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1372
ret = i40e_aq_add_rem_control_packet_filter(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1373
pf->hw.mac.addr, ETH_P_LLDP, 0,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1377
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1379
__func__, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1383
ret = i40e_aq_start_lldp(&pf->hw, false, NULL);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1385
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1387
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1391
pf->dcbx_cap = DCB_CAP_DCBX_LLD_MANAGED |
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1404
ret = i40e_aq_get_lldp_mib(&pf->hw, 0,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1409
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
141
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1411
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1416
dev_info(&pf->pdev->dev, "LLDP MIB (local)\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1431
ret = i40e_aq_get_lldp_mib(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1437
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1439
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1444
dev_info(&pf->pdev->dev, "LLDP MIB (remote)\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
145
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1453
ret = i40e_aq_cfg_lldp_mib_change_event(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1456
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1458
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1464
ret = i40e_aq_cfg_lldp_mib_change_event(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1467
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1469
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1492
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1508
ret = i40e_acquire_nvm(&pf->hw, I40E_RESOURCE_READ);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
151
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1510
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1512
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1517
ret = i40e_aq_read_nvm(&pf->hw, module, (2 * offset),
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1519
i40e_release_nvm(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1521
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1523
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1525
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1536
dev_info(&pf->pdev->dev, "unknown command '%s'\n", cmd_buf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1537
dev_info(&pf->pdev->dev, "available commands\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1538
dev_info(&pf->pdev->dev, " add vsi [relay_seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1539
dev_info(&pf->pdev->dev, " del vsi [vsi_seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1540
dev_info(&pf->pdev->dev, " add relay <uplink_seid> <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1541
dev_info(&pf->pdev->dev, " del relay <relay_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1542
dev_info(&pf->pdev->dev, " add pvid <vsi_seid> <vid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1543
dev_info(&pf->pdev->dev, " del pvid <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1544
dev_info(&pf->pdev->dev, " dump switch\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1545
dev_info(&pf->pdev->dev, " dump vsi [seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1546
dev_info(&pf->pdev->dev, " dump desc tx <vsi_seid> <ring_id> [<desc_n>]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1547
dev_info(&pf->pdev->dev, " dump desc rx <vsi_seid> <ring_id> [<desc_n>]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1548
dev_info(&pf->pdev->dev, " dump desc xdp <vsi_seid> <ring_id> [<desc_n>]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1549
dev_info(&pf->pdev->dev, " dump desc aq\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1550
dev_info(&pf->pdev->dev, " dump reset stats\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1551
dev_info(&pf->pdev->dev, " dump debug fwdata <cluster_id> <table_id> <index>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1552
dev_info(&pf->pdev->dev, " read <reg>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1553
dev_info(&pf->pdev->dev, " write <reg> <value>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1554
dev_info(&pf->pdev->dev, " clear_stats vsi [seid]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1555
dev_info(&pf->pdev->dev, " clear_stats port\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1556
dev_info(&pf->pdev->dev, " pfr\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1557
dev_info(&pf->pdev->dev, " corer\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1558
dev_info(&pf->pdev->dev, " globr\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1559
dev_info(&pf->pdev->dev, " send aq_cmd <flags> <opcode> <datalen> <retval> <cookie_h> <cookie_l> <param0> <param1> <param2> <param3>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1560
dev_info(&pf->pdev->dev, " send indirect aq_cmd <flags> <opcode> <datalen> <retval> <cookie_h> <cookie_l> <param0> <param1> <param2> <param3> <buffer_len>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1561
dev_info(&pf->pdev->dev, " fd current cnt");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1562
dev_info(&pf->pdev->dev, " lldp start\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1563
dev_info(&pf->pdev->dev, " lldp stop\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1564
dev_info(&pf->pdev->dev, " lldp get local\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1565
dev_info(&pf->pdev->dev, " lldp get remote\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1566
dev_info(&pf->pdev->dev, " lldp event on\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1567
dev_info(&pf->pdev->dev, " lldp event off\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1568
dev_info(&pf->pdev->dev, " nvm read [module] [word_offset] [word_count]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
157
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1600
struct i40e_pf *pf = filp->private_data;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
161
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1633
dev_info(&pf->pdev->dev, "change_mtu <vsi_seid> <mtu>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1636
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1638
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1641
dev_info(&pf->pdev->dev, "change_mtu: no netdev for VSI %d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1647
dev_info(&pf->pdev->dev, "change_mtu called\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1649
dev_info(&pf->pdev->dev, "Could not acquire RTNL - please try again\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1655
dev_info(&pf->pdev->dev, "set_rx_mode <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1658
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
166
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1660
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1663
dev_info(&pf->pdev->dev, "set_rx_mode: no netdev for VSI %d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1668
dev_info(&pf->pdev->dev, "set_rx_mode called\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1670
dev_info(&pf->pdev->dev, "Could not acquire RTNL - please try again\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1676
dev_info(&pf->pdev->dev, "napi <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1679
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1681
dev_info(&pf->pdev->dev, "napi: VSI %d not found\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1684
dev_info(&pf->pdev->dev, "napi: no netdev for VSI %d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1689
dev_info(&pf->pdev->dev, "napi called\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1692
dev_info(&pf->pdev->dev, "unknown command '%s'\n", cmd_buf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1693
dev_info(&pf->pdev->dev, "available commands\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1694
dev_info(&pf->pdev->dev, " change_mtu <vsi_seid> <mtu>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1695
dev_info(&pf->pdev->dev, " set_rx_mode <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1696
dev_info(&pf->pdev->dev, " napi <vsi_seid>\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
171
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1713
void i40e_dbg_pf_init(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1715
const char *name = pci_name(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1717
pf->i40e_dbg_pf = debugfs_create_dir(name, i40e_dbg_root);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1719
debugfs_create_file("command", 0600, pf->i40e_dbg_pf, pf,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1722
debugfs_create_file("netdev_ops", 0600, pf->i40e_dbg_pf, pf,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1730
void i40e_dbg_pf_exit(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1732
debugfs_remove_recursive(pf->i40e_dbg_pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1733
pf->i40e_dbg_pf = NULL;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
176
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
180
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
184
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
195
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
200
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
203
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
209
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
214
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
219
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
223
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
226
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
237
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
242
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
248
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
25
static struct i40e_vsi *i40e_dbg_find_vsi(struct i40e_pf *pf, int seid)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
253
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
259
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
262
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
265
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
277
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
28
dev_info(&pf->pdev->dev, "%d: bad seid\n", seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
282
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
288
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
293
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
298
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
301
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
304
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
312
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
315
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
318
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
321
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
324
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
328
dev_info(&pf->pdev->dev, " type = %i\n", vsi->type);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
33
return i40e_pf_get_vsi_by_seid(pf, seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
330
dev_info(&pf->pdev->dev, " VF ID = %i\n", vsi->vf_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
331
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
334
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
337
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
340
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
344
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
348
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
351
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
355
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
361
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
367
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
373
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
379
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
382
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
385
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
391
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
397
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
400
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
408
dev_info(&pf->pdev->dev, " idx = %d\n", vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
409
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
413
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
419
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
423
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
435
static void i40e_dbg_dump_aq_desc(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
438
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
443
dev_driver_string(&pf->pdev->dev),
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
444
dev_name(&pf->pdev->dev));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
447
dev_info(&pf->pdev->dev, "AdminQ Tx Ring\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
452
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
460
dev_info(&pf->pdev->dev, "AdminQ Rx Ring\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
465
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
484
struct i40e_pf *pf, enum ring_type type)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
493
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
495
dev_info(&pf->pdev->dev, "vsi %d not found\n", vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
501
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
507
dev_info(&pf->pdev->dev, "XDP not enabled on VSI %d\n", vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
511
dev_info(&pf->pdev->dev, "ring %d not found\n", ring_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
515
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
541
dev_info(&pf->pdev->dev, "VSI = %02i Rx ring = %02i\n", vsi_seid, ring_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
544
dev_info(&pf->pdev->dev, "VSI = %02i Tx ring = %02i\n", vsi_seid, ring_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
547
dev_info(&pf->pdev->dev, "VSI = %02i XDP ring = %02i\n", vsi_seid, ring_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
553
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
559
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
567
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
573
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
579
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
58
static void i40e_dbg_dump_vsi_seid(struct i40e_pf *pf, int seid)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
585
dev_info(&pf->pdev->dev, "dump desc rx/tx/xdp <vsi_seid> <ring_id> [<desc_n>]\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
596
static void i40e_dbg_dump_vsi_no_seid(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
601
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
602
dev_info(&pf->pdev->dev, "dump vsi[%d]: %d\n", i, vsi->seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
610
static void i40e_dbg_dump_eth_stats(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
613
dev_info(&pf->pdev->dev, " ethstats:\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
614
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
617
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
620
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
623
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
626
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
636
static void i40e_dbg_dump_veb_seid(struct i40e_pf *pf, int seid)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
640
veb = i40e_pf_get_veb_by_seid(pf, seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
642
dev_info(&pf->pdev->dev, "can't find veb %d\n", seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
645
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
649
i40e_dbg_dump_eth_stats(pf, &veb->stats);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
65
vsi = i40e_dbg_find_vsi(pf, seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
656
static void i40e_dbg_dump_veb_all(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
661
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
662
i40e_dbg_dump_veb_seid(pf, veb->seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
67
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
670
static void i40e_dbg_dump_vf(struct i40e_pf *pf, int vf_id)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
675
if (!pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
676
dev_info(&pf->pdev->dev, "no VFs allocated\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
677
} else if ((vf_id >= 0) && (vf_id < pf->num_alloc_vfs)) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
678
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
679
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
680
dev_info(&pf->pdev->dev, "vf %2d: VSI id=%d, seid=%d, qps=%d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
682
dev_info(&pf->pdev->dev, " num MDD=%lld\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
685
dev_info(&pf->pdev->dev, "invalid VF id %d\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
693
static void i40e_dbg_dump_vf_all(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
697
if (!pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
698
dev_info(&pf->pdev->dev, "no VFs enabled!\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
700
for (i = 0; i < pf->num_alloc_vfs; i++)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
701
i40e_dbg_dump_vf(pf, i);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
71
dev_info(&pf->pdev->dev, "vsi seid %d\n", seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
715
struct i40e_pf *pf = filp->private_data;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
749
vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
75
dev_info(&pf->pdev->dev, " netdev: name = %s, state = %lu, flags = 0x%08x\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
752
dev_info(&pf->pdev->dev, "add VSI %d: bad vsi seid\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
760
if (!test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
761
set_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
762
i40e_do_reset_safe(pf, I40E_PF_RESET_FLAG);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
765
vsi = i40e_vsi_setup(pf, I40E_VSI_VMDQ2, vsi_seid, 0);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
767
dev_info(&pf->pdev->dev, "added VSI %d to relay %d\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
77
dev_info(&pf->pdev->dev, " features = 0x%08lx\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
770
dev_info(&pf->pdev->dev, "'%s' failed\n", cmd_buf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
775
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
780
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
782
dev_info(&pf->pdev->dev, "del VSI %d: seid not found\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
787
dev_info(&pf->pdev->dev, "deleting VSI %d\n", vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
79
dev_info(&pf->pdev->dev, " hw_features = 0x%08lx\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
800
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
805
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
81
dev_info(&pf->pdev->dev, " vlan_features = 0x%08lx\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
811
if (uplink_seid != 0 && uplink_seid != pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
812
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
817
vsi = i40e_pf_get_vsi_by_seid(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
819
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
826
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
831
veb = i40e_veb_setup(pf, uplink_seid, vsi_seid, enabled_tc);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
833
dev_info(&pf->pdev->dev, "added relay %d\n", veb->seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
835
dev_info(&pf->pdev->dev, "add relay failed\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
84
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
843
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
848
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
854
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
859
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
864
dev_info(&pf->pdev->dev, "deleting relay %d\n", veb_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
873
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
878
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
88
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
880
dev_info(&pf->pdev->dev, "add pvid: VSI %d not found\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
888
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
892
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
900
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
906
vsi = i40e_dbg_find_vsi(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
908
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
914
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
919
i40e_fetch_switch_configuration(pf, true);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
92
dev_info(&pf->pdev->dev, " MAC address: %pM Port MAC: %pM\n",
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
923
i40e_dbg_dump_vsi_seid(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
925
i40e_dbg_dump_vsi_no_seid(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
929
i40e_dbg_dump_veb_seid(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
93
pf->hw.mac.addr,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
931
i40e_dbg_dump_veb_all(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
935
i40e_dbg_dump_vf(pf, vf_id);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
937
i40e_dbg_dump_vf_all(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
94
pf->hw.mac.port_addr);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
944
desc_n, pf, RING_TYPE_RX);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
950
desc_n, pf, RING_TYPE_TX);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
956
desc_n, pf, RING_TYPE_XDP);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
958
i40e_dbg_dump_aq_desc(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
96
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
960
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
962
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
964
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
966
dev_info(&pf->pdev->dev, "dump desc aq\n");
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
969
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
970
"core reset count: %d\n", pf->corer_count);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
971
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
972
"global reset count: %d\n", pf->globr_count);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
973
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
974
"emp reset count: %d\n", pf->empr_count);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
975
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
976
"pf reset count: %d\n", pf->pfr_count);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
980
&pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
982
&pf->hw.remote_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
992
vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
997
ret = i40e_aq_query_port_ets_config(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_devlink.c
106
struct i40e_pf *pf = devlink_priv(dl);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
107
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
111
i40e_info_get_dsn(pf, buf, sizeof(buf));
drivers/net/ethernet/intel/i40e/i40e_devlink.c
13
struct i40e_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
15
if (pf->num_alloc_vfs > 0) {
drivers/net/ethernet/intel/i40e/i40e_devlink.c
189
void i40e_free_pf(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
191
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
202
void i40e_devlink_register(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
204
struct devlink *dl = priv_to_devlink(pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
205
struct device *dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
21
pf->max_mac_per_vf = ctx->val.vu32;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
224
void i40e_devlink_unregister(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
226
struct devlink *dl = priv_to_devlink(pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
238
static void i40e_devlink_set_switch_id(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_devlink.c
241
u64 id = pci_get_dsn(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
257
int i40e_devlink_create_port(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
259
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
261
struct device *dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
265
attrs.phys.port_number = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
266
i40e_devlink_set_switch_id(pf, &attrs.switch_id);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
267
devlink_port_attrs_set(&pf->devlink_port, &attrs);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
268
err = devlink_port_register(devlink, &pf->devlink_port, pf->hw.pf_id);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
283
void i40e_devlink_destroy_port(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
285
devlink_port_unregister(&pf->devlink_port);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
30
struct i40e_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
32
ctx->val.vu32 = pf->max_mac_per_vf;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
44
static void i40e_info_get_dsn(struct i40e_pf *pf, char *buf, size_t len)
drivers/net/ethernet/intel/i40e/i40e_devlink.c
48
put_unaligned_be64(pci_get_dsn(pf->pdev), dsn);
drivers/net/ethernet/intel/i40e/i40e_devlink.h
12
void i40e_free_pf(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.h
13
void i40e_devlink_register(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.h
14
void i40e_devlink_unregister(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.h
15
int i40e_devlink_create_port(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_devlink.h
16
void i40e_devlink_destroy_port(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1003
i40e_phy_type_to_ethtool(pf, &cap_ksettings);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1049
struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1054
i40e_phy_type_to_ethtool(pf, ks);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1072
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1073
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1081
i40e_get_settings_link_up(hw, ks, netdev, pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1083
i40e_get_settings_link_down(hw, ks, pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1229
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1232
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1244
i40e_partition_setting_complaint(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1300
while (test_and_set_bit(__I40E_CONFIG_BUSY, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1482
clear_bit(__I40E_CONFIG_BUSY, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1491
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1492
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1527
i40e_set_fec_in_flags(fec_cfg, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1549
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1550
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1591
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1592
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1623
dev_warn(&pf->pdev->dev, "Unsupported FEC mode: %d",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1635
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1636
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1662
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1663
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1697
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1699
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1712
i40e_partition_setting_complaint(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1726
if (!test_bit(__I40E_DOWN, pf->state) && !is_an) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1777
if (!test_bit(__I40E_DOWN, pf->state) && is_an) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1780
if (!test_bit(__I40E_DOWN, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1790
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1791
u32 debug_mask = pf->hw.debug_mask;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1796
return pf->msg_enable;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1802
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1805
pf->hw.debug_mask = data;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1807
pf->msg_enable = data;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1825
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1826
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1857
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1877
else if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1878
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1884
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1902
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1922
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1928
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1933
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1970
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1983
else if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1984
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1990
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2004
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2007
i40e_nvm_version_str(&pf->hw, drvinfo->fw_version,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2009
strscpy(drvinfo->bus_info, pci_name(pf->pdev),
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2012
if (pf->hw.pf_id == 0)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2022
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2023
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2025
ring->rx_max_pending = i40e_get_max_num_descriptors(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2026
ring->tx_max_pending = i40e_get_max_num_descriptors(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2056
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2064
max_num_descriptors = i40e_get_max_num_descriptors(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2091
while (test_and_set_bit(__I40E_CONFIG_BUSY, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2259
clear_bit(__I40E_CONFIG_BUSY, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2282
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2285
if (vsi->type == I40E_VSI_MAIN && pf->hw.partition_id == 1)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2313
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2322
(pf->hw.pf_id == 0 ? I40E_GL_PRIV_FLAGS_STR_LEN : 0);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2361
i40e_get_pfc_stats(struct i40e_pf *pf, unsigned int i)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2364
.stat = pf->stats.stat[priority]
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2395
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2415
if (vsi->type != I40E_VSI_MAIN || pf->hw.partition_id != 1)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2418
veb = i40e_pf_get_main_veb(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2419
veb_stats = veb && test_bit(I40E_FLAG_VEB_STATS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2443
i40e_add_ethtool_stats(&data, pf, i40e_gstrings_stats);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2446
struct i40e_pfc_stats pfc = i40e_get_pfc_stats(pf, i);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2470
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2485
if (vsi->type != I40E_VSI_MAIN || pf->hw.partition_id != 1)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2507
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2513
if (pf->hw.pf_id != 0)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2541
struct i40e_pf *pf = i40e_netdev_to_pf(dev);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2544
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2552
if (pf->ptp_clock)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2553
info->phc_index = ptp_clock_index(pf->ptp_clock);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2562
if (test_bit(I40E_HW_CAP_PTP_L4, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2578
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2582
netif_info(pf, hw, netdev, "link test\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2583
status = i40e_get_link_status(&pf->hw, &link_up);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2585
netif_err(pf, drv, netdev, "link query timed out, please retry test\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2601
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2603
netif_info(pf, hw, netdev, "register test\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2604
*data = i40e_diag_reg_test(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2612
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2614
netif_info(pf, hw, netdev, "eeprom test\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2615
*data = i40e_diag_eeprom_test(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2618
pf->hw.nvmupd_state = I40E_NVMUPD_STATE_INIT;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2626
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2627
u16 swc_old = pf->sw_int_count;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2629
netif_info(pf, hw, netdev, "interrupt test\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2630
wr32(&pf->hw, I40E_PFINT_DYN_CTL0,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2637
*data = (swc_old == pf->sw_int_count);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2642
static inline bool i40e_active_vfs(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2644
struct i40e_vf *vfs = pf->vf;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2647
for (i = 0; i < pf->num_alloc_vfs; i++)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2653
static inline bool i40e_active_vmdqs(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2655
return !!i40e_find_vsi_by_type(pf, I40E_VSI_VMDQ2);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2663
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2667
netif_info(pf, drv, netdev, "offline testing starting\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2669
set_bit(__I40E_TESTING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2671
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2672
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2673
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2678
if (i40e_active_vfs(pf) || i40e_active_vmdqs(pf)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2679
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2694
i40e_do_reset(pf, BIT(__I40E_PF_RESET_REQUESTED), true);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2709
clear_bit(__I40E_TESTING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2710
i40e_do_reset(pf, BIT(__I40E_PF_RESET_REQUESTED), true);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2716
netif_info(pf, drv, netdev, "online testing starting\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2727
netif_info(pf, drv, netdev, "testing finished\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2736
clear_bit(__I40E_TESTING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2737
netif_info(pf, drv, netdev, "testing failed\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2744
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2746
stats->link_down_events = pf->link_down_events;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2753
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2754
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2764
wol->wolopts = (pf->wol_en ? WAKE_MAGIC : 0);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2776
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2778
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2783
i40e_partition_setting_complaint(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2800
if (pf->wol_en != !!wol->wolopts) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2801
pf->wol_en = !!wol->wolopts;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2802
device_set_wakeup_enable(&pf->pdev->dev, pf->wol_en);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2812
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2813
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2820
if (!test_bit(I40E_HW_CAP_PHY_CONTROLS_LEDS, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2821
pf->led_status = i40e_led_get(hw);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2827
&pf->phy_led_val);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2828
pf->led_status = temp_status;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2832
if (!test_bit(I40E_HW_CAP_PHY_CONTROLS_LEDS, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2835
ret = i40e_led_set_phy(hw, true, pf->led_status, 0);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2838
if (!test_bit(I40E_HW_CAP_PHY_CONTROLS_LEDS, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2841
ret = i40e_led_set_phy(hw, false, pf->led_status, 0);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2844
if (!test_bit(I40E_HW_CAP_PHY_CONTROLS_LEDS, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2845
i40e_led_set(hw, pf->led_status, false);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2847
ret = i40e_led_set_phy(hw, false, pf->led_status,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2848
(pf->phy_led_val |
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2967
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2968
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3017
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3030
netif_info(pf, drv, netdev, "Invalid queue value, queue range is 0 - %d\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3040
netif_info(pf, drv, netdev, "tx-usecs-high is not used, please program rx-usecs-high\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3045
netif_info(pf, drv, netdev, "Invalid value, rx-usecs-high range is 0-%lu\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3052
netif_info(pf, drv, netdev, "RX interrupt moderation cannot be changed if adaptive-rx is enabled.\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3057
netif_info(pf, drv, netdev, "Invalid value, rx-usecs range is 0-8160\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3063
netif_info(pf, drv, netdev, "TX interrupt moderation cannot be changed if adaptive-tx is enabled.\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3068
netif_info(pf, drv, netdev, "Invalid value, tx-usecs range is 0-8160\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3081
netif_info(pf, drv, netdev, "Interrupt rate limit rounded down to %d\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3134
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3135
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3319
static int i40e_get_ethtool_fdir_all(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3328
cmd->data = i40e_get_fd_cnt_all(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3331
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3354
static int i40e_get_ethtool_fdir_entry(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3367
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3444
input_set = i40e_read_fd_input_set(pf, index);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3486
vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3488
vsi = i40e_find_vsi_from_id(pf, rule->dest_vsi);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3538
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3543
cmd->rule_cnt = pf->fdir_pf_active_filters;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3545
cmd->data = i40e_get_fd_cnt_all(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3549
ret = i40e_get_ethtool_fdir_entry(pf, cmd);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3552
ret = i40e_get_ethtool_fdir_all(pf, cmd, rule_locs);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3621
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3622
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3630
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3631
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3647
pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3654
pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3661
pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3672
pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3716
i_set = i40e_get_rss_hash_bits(&pf->hw, nfc, i_setc);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3751
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3759
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3774
pf->fdir_pf_active_filters--;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3791
&pf->fdir_filter_list);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3794
pf->fdir_pf_active_filters++;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3807
static void i40e_prune_flex_pit_list(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3813
list_for_each_entry_safe(entry, tmp, &pf->l3_flex_pit_list, list) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3816
hlist_for_each_entry(rule, &pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3836
list_for_each_entry_safe(entry, tmp, &pf->l4_flex_pit_list, list) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3839
hlist_for_each_entry(rule, &pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3877
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3880
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3881
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3884
if (test_bit(__I40E_FD_FLUSH_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3889
i40e_prune_flex_pit_list(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3891
i40e_fdir_check_and_reenable(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3904
static u8 i40e_unused_pit_index(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3914
list_for_each_entry(entry, &pf->l4_flex_pit_list, list)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3917
list_for_each_entry(entry, &pf->l3_flex_pit_list, list)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4040
static void __i40e_reprogram_flex_pit(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4074
i40e_write_rx_ctl(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4082
i40e_write_rx_ctl(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4100
i40e_write_rx_ctl(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4115
static void i40e_reprogram_flex_pit(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4117
__i40e_reprogram_flex_pit(pf, &pf->l3_flex_pit_list,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4120
__i40e_reprogram_flex_pit(pf, &pf->l4_flex_pit_list,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4124
i40e_write_rx_ctl(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4129
i40e_write_rx_ctl(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4209
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4216
netif_info(pf, drv, vsi->netdev, "L3 source address: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4223
netif_info(pf, drv, vsi->netdev, "L3 destination address: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4230
netif_info(pf, drv, vsi->netdev, "L4 source port: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4237
netif_info(pf, drv, vsi->netdev, "L4 destination port: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4244
netif_info(pf, drv, vsi->netdev, "SCTP verification tag: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4255
netif_info(pf, drv, vsi->netdev, "FLEX index %d: %s -> %s\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4261
netif_info(pf, drv, vsi->netdev, " Current input set: %0llx\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4263
netif_info(pf, drv, vsi->netdev, "Requested input set: %0llx\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4303
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4315
fdir_filter_count = &pf->fd_sctp4_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4319
fdir_filter_count = &pf->fd_tcp4_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4323
fdir_filter_count = &pf->fd_udp4_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4327
fdir_filter_count = &pf->fd_sctp6_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4331
fdir_filter_count = &pf->fd_tcp6_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4335
fdir_filter_count = &pf->fd_udp6_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4339
fdir_filter_count = &pf->fd_ip4_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4344
fdir_filter_count = &pf->fd_ip6_filter_cnt;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4352
current_mask = i40e_read_fd_input_set(pf, index);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4559
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4568
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4578
flex_pit = i40e_find_flex_offset(&pf->l4_flex_pit_list,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4592
i40e_find_flex_offset(&pf->l3_flex_pit_list,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4622
pit_index = i40e_unused_pit_index(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4638
netif_info(pf, drv, vsi->netdev, "Input set change requested for %s flows:\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4642
netif_info(pf, drv, vsi->netdev, "FLEX index %d: Offset -> %d",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4650
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4651
netif_err(pf, drv, vsi->netdev, "Cannot change Flow Director input sets while MFP is enabled\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4665
netif_err(pf, drv, vsi->netdev, "Cannot change input set for %s flows until %d preexisting filters are removed\n",
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4671
i40e_write_fd_input_set(pf, index, new_mask);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4680
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_FRAG_IPV4,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4685
err = i40e_add_flex_offset(&pf->l4_flex_pit_list, src_offset,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4691
err = i40e_add_flex_offset(&pf->l3_flex_pit_list,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4698
i40e_reprogram_flex_pit(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4759
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4765
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4777
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4802
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4808
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
481
static void i40e_partition_setting_complaint(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4810
if (!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4813
if (test_bit(__I40E_FD_SB_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4816
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4817
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4820
if (test_bit(__I40E_FD_FLUSH_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
483
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4837
if (fsp->location >= (pf->hw.func_caps.fd_filters_best_effort +
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4838
pf->hw.func_caps.fd_filters_guaranteed)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4859
if (vf >= pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4861
if (ring >= pf->vf[vf].num_queue_pairs)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4863
dest_vsi = pf->vf[vf].lan_vsi_id;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4879
input->cnt_index = I40E_FD_SB_STAT_IDX(pf->hw.pf_id);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
493
static void i40e_phy_type_to_ethtool(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4939
pf->fdir_pf_active_filters--;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
496
struct i40e_link_status *hw_link_info = &pf->hw.phy.link_info;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
497
u64 phy_types = pf->hw.phy.phy_types;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
4997
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5003
ch->other_count = test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) ? 1 : 0;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5025
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5038
if (i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5046
if (ch->other_count != (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) ? 1 : 0))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5057
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5059
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5067
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5079
new_count = i40e_reconfig_rss_queues(pf, count);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
508
if (test_bit(I40E_HW_CAP_100M_SGMII, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5161
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5190
i40e_fill_rss_lut(pf, vsi->rss_lut_user, I40E_HLUT_ARRAY_SIZE,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5211
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5219
if (test_bit(priv_flag->bitno, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5223
if (pf->hw.pf_id != 0)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5231
if (test_bit(priv_flag->bitno, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5250
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5256
bitmap_copy(orig_flags, pf->flags, I40E_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5257
bitmap_copy(new_flags, pf->flags, I40E_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5277
if (pf->hw.pf_id != 0)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5315
!test_bit(I40E_HW_CAP_ATR_EVICT, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5327
!test_bit(I40E_HW_CAP_FW_LLDP_STOPPABLE, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5328
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5334
pf->hw.device_id != I40E_DEV_ID_25G_SFP28 &&
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5335
pf->hw.device_id != I40E_DEV_ID_25G_B) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5336
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5342
pf->hw.device_id != I40E_DEV_ID_25G_SFP28 &&
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5343
pf->hw.device_id != I40E_DEV_ID_25G_B &&
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5344
pf->hw.device_id != I40E_DEV_ID_KX_X722) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5345
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5358
set_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5359
set_bit(__I40E_FD_FLUSH_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5369
ret = i40e_aq_set_switch_config(&pf->hw, sw_flags, valid_flags,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5371
if (ret && pf->hw.aq.asq_last_status != LIBIE_AQ_RC_ESRCH) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5372
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5375
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5395
dev_warn(&pf->pdev->dev, "Cannot change FEC config\n");
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5400
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5406
pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5407
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5414
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5422
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5428
i40e_dcb_sw_default_config(pf);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5430
i40e_aq_cfg_lldp_mib_change_event(&pf->hw, false, NULL);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5431
i40e_aq_stop_lldp(&pf->hw, true, false, NULL);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5433
status = i40e_aq_start_lldp(&pf->hw, false, NULL);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5435
adq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5438
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5443
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5447
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5451
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5466
bitmap_copy(pf->flags, new_flags, I40E_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5472
i40e_do_reset(pf, reset_needed, true);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5487
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5488
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5594
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5595
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5660
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5661
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5688
edata->tx_lpi_enabled = pf->stats.tx_lpi_status;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5690
edata->eee_active = pf->stats.tx_lpi_status && pf->stats.rx_lpi_status;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5700
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5707
{edata->tx_lpi_enabled != pf->stats.tx_lpi_status, "tx-lpi"}
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5729
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5730
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5854
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5856
if (!test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
607
!test_bit(I40E_HW_CAP_CRT_RETIMER, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
615
!test_bit(I40E_HW_CAP_CRT_RETIMER, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
775
struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
923
if (test_bit(I40E_HW_CAP_100M_SGMII, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10003
if (test_bit(__I40E_DOWN, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
10004
test_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10008
if (time_before(jiffies, (pf->service_timer_previous +
drivers/net/ethernet/intel/i40e/i40e_main.c
10009
pf->service_timer_period)))
drivers/net/ethernet/intel/i40e/i40e_main.c
1001
ns->rx_crc_errors = pf->stats.crc_errors;
drivers/net/ethernet/intel/i40e/i40e_main.c
10011
pf->service_timer_previous = jiffies;
drivers/net/ethernet/intel/i40e/i40e_main.c
10013
if (test_bit(I40E_FLAG_LINK_POLLING_ENA, pf->flags) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
10014
test_bit(__I40E_TEMP_LINK_POLLING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10015
i40e_link_event(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1002
ns->rx_errors = pf->stats.crc_errors + pf->stats.illegal_bytes;
drivers/net/ethernet/intel/i40e/i40e_main.c
10020
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
10024
if (test_bit(I40E_FLAG_VEB_STATS_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10026
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e_main.c
1003
ns->rx_length_errors = pf->stats.rx_length_errors;
drivers/net/ethernet/intel/i40e/i40e_main.c
10030
i40e_ptp_rx_hang(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10031
i40e_ptp_tx_hang(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10038
static void i40e_reset_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10042
if (test_bit(__I40E_REINIT_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10044
clear_bit(__I40E_REINIT_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10046
if (test_bit(__I40E_PF_RESET_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10048
clear_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10050
if (test_bit(__I40E_CORE_RESET_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10052
clear_bit(__I40E_CORE_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10054
if (test_bit(__I40E_GLOBAL_RESET_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10056
clear_bit(__I40E_GLOBAL_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10058
if (test_bit(__I40E_DOWN_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10060
clear_bit(__I40E_DOWN_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10066
if (test_bit(__I40E_RESET_INTR_RECEIVED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10067
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10068
i40e_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10069
i40e_rebuild(pf, false, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
10074
!test_bit(__I40E_DOWN, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
10075
!test_bit(__I40E_CONFIG_BUSY, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10076
i40e_do_reset(pf, reset_flags, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
10085
static void i40e_handle_link_event(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
10096
i40e_link_event(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10100
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10102
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1011
static void i40e_update_pf_stats(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10111
(!test_bit(I40E_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags))) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10112
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10114
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10124
static void i40e_clean_adminq_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10127
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
1013
struct i40e_hw_port_stats *osd = &pf->stats_offsets;
drivers/net/ethernet/intel/i40e/i40e_main.c
10135
if (test_bit(__I40E_RESET_FAILED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10139
val = rd32(&pf->hw, I40E_PF_ARQLEN);
drivers/net/ethernet/intel/i40e/i40e_main.c
1014
struct i40e_hw_port_stats *nsd = &pf->stats;
drivers/net/ethernet/intel/i40e/i40e_main.c
10143
dev_info(&pf->pdev->dev, "ARQ VF Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10148
dev_info(&pf->pdev->dev, "ARQ Overflow Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
1015
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
10150
pf->arq_overflows++;
drivers/net/ethernet/intel/i40e/i40e_main.c
10154
dev_info(&pf->pdev->dev, "ARQ Critical Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10158
wr32(&pf->hw, I40E_PF_ARQLEN, val);
drivers/net/ethernet/intel/i40e/i40e_main.c
10160
val = rd32(&pf->hw, I40E_PF_ATQLEN);
drivers/net/ethernet/intel/i40e/i40e_main.c
10163
if (pf->hw.debug_mask & I40E_DEBUG_AQ)
drivers/net/ethernet/intel/i40e/i40e_main.c
10164
dev_info(&pf->pdev->dev, "ASQ VF Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10168
if (pf->hw.debug_mask & I40E_DEBUG_AQ)
drivers/net/ethernet/intel/i40e/i40e_main.c
10169
dev_info(&pf->pdev->dev, "ASQ Overflow Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10173
if (pf->hw.debug_mask & I40E_DEBUG_AQ)
drivers/net/ethernet/intel/i40e/i40e_main.c
10174
dev_info(&pf->pdev->dev, "ASQ Critical Error detected\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10178
wr32(&pf->hw, I40E_PF_ATQLEN, val);
drivers/net/ethernet/intel/i40e/i40e_main.c
10190
dev_info(&pf->pdev->dev, "ARQ event error %d\n", ret);
drivers/net/ethernet/intel/i40e/i40e_main.c
10199
i40e_handle_link_event(pf, &event);
drivers/net/ethernet/intel/i40e/i40e_main.c
10203
ret = i40e_vc_process_vf_msg(pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
1021
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10211
dev_dbg(&pf->pdev->dev, "ARQ: Update LLDP MIB event received\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10214
i40e_handle_lldp_event(pf, &event);
drivers/net/ethernet/intel/i40e/i40e_main.c
10219
dev_dbg(&pf->pdev->dev, "ARQ LAN queue overflow event received\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10220
i40e_handle_lan_overflow_event(pf, &event);
drivers/net/ethernet/intel/i40e/i40e_main.c
10223
dev_info(&pf->pdev->dev, "ARQ: Msg from other pf\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10228
i40e_debug(&pf->hw, I40E_DEBUG_NVM,
drivers/net/ethernet/intel/i40e/i40e_main.c
10233
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10241
clear_bit(__I40E_ADMINQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
1025
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10256
static void i40e_verify_eeprom(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10260
err = i40e_diag_eeprom_test(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
10263
err = i40e_diag_eeprom_test(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
10265
dev_info(&pf->pdev->dev, "eeprom check failed (%d), Tx/Rx traffic disabled\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
10267
set_bit(__I40E_BAD_EEPROM, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10271
if (!err && test_bit(__I40E_BAD_EEPROM, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10272
dev_info(&pf->pdev->dev, "eeprom check passed, Tx/Rx traffic enabled\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10273
clear_bit(__I40E_BAD_EEPROM, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
1028
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10283
static void i40e_enable_pf_switch_lb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10285
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10289
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
10290
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_main.c
10292
ret = i40e_aq_get_vsi_params(&pf->hw, &ctxt, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
10294
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10296
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10305
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10307
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10317
static void i40e_disable_pf_switch_lb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10319
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10323
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
10324
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_main.c
10326
ret = i40e_aq_get_vsi_params(&pf->hw, &ctxt, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
10328
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1033
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10330
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10339
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10341
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10355
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
10357
if (pf->hw.debug_mask & I40E_DEBUG_LAN)
drivers/net/ethernet/intel/i40e/i40e_main.c
10358
dev_info(&pf->pdev->dev, "enabling bridge mode: %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
10361
i40e_disable_pf_switch_lb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10363
i40e_enable_pf_switch_lb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10377
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
1038
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10385
if (WARN_ON(veb->uplink_seid && veb->uplink_seid != pf->mac_seid)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10386
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10391
if (veb->uplink_seid == pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10393
ctl_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10397
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10405
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10420
if (test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
10428
i40e_pf_for_each_vsi(pf, v, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
1043
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10436
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10453
static int i40e_get_capabilities(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
10468
err = i40e_aq_discover_capabilities(&pf->hw, cap_buf, buf_len,
drivers/net/ethernet/intel/i40e/i40e_main.c
10474
if (pf->hw.aq.asq_last_status == LIBIE_AQ_RC_ENOMEM) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10477
} else if (pf->hw.aq.asq_last_status != LIBIE_AQ_RC_OK || err) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10478
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1048
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10481
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10486
if (pf->hw.debug_mask & I40E_DEBUG_USER) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10488
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10490
pf->hw.pf_id, pf->hw.func_caps.num_vfs,
drivers/net/ethernet/intel/i40e/i40e_main.c
10491
pf->hw.func_caps.num_msix_vectors,
drivers/net/ethernet/intel/i40e/i40e_main.c
10492
pf->hw.func_caps.num_msix_vectors_vf,
drivers/net/ethernet/intel/i40e/i40e_main.c
10493
pf->hw.func_caps.fd_filters_guaranteed,
drivers/net/ethernet/intel/i40e/i40e_main.c
10494
pf->hw.func_caps.fd_filters_best_effort,
drivers/net/ethernet/intel/i40e/i40e_main.c
10495
pf->hw.func_caps.num_tx_qp,
drivers/net/ethernet/intel/i40e/i40e_main.c
10496
pf->hw.func_caps.num_vsis);
drivers/net/ethernet/intel/i40e/i40e_main.c
10498
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10500
pf->hw.dev_caps.switch_mode,
drivers/net/ethernet/intel/i40e/i40e_main.c
10501
pf->hw.dev_caps.valid_functions);
drivers/net/ethernet/intel/i40e/i40e_main.c
10502
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10504
pf->hw.dev_caps.sr_iov_1_1,
drivers/net/ethernet/intel/i40e/i40e_main.c
10505
pf->hw.dev_caps.num_vfs);
drivers/net/ethernet/intel/i40e/i40e_main.c
10506
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10508
pf->hw.dev_caps.num_vsis,
drivers/net/ethernet/intel/i40e/i40e_main.c
10509
pf->hw.dev_caps.num_rx_qp,
drivers/net/ethernet/intel/i40e/i40e_main.c
10510
pf->hw.dev_caps.num_tx_qp);
drivers/net/ethernet/intel/i40e/i40e_main.c
10514
#define DEF_NUM_VSI (1 + (pf->hw.func_caps.fcoe ? 1 : 0) \
drivers/net/ethernet/intel/i40e/i40e_main.c
10515
+ pf->hw.func_caps.num_vfs)
drivers/net/ethernet/intel/i40e/i40e_main.c
10516
if (pf->hw.revision_id == 0 &&
drivers/net/ethernet/intel/i40e/i40e_main.c
10517
pf->hw.func_caps.num_vsis < DEF_NUM_VSI) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10518
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10520
pf->hw.func_caps.num_vsis, DEF_NUM_VSI);
drivers/net/ethernet/intel/i40e/i40e_main.c
10521
pf->hw.func_caps.num_vsis = DEF_NUM_VSI;
drivers/net/ethernet/intel/i40e/i40e_main.c
1053
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10533
static void i40e_fdir_sb_setup(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10540
if (!rd32(&pf->hw, I40E_GLQF_HKEY(0))) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10549
wr32(&pf->hw, I40E_GLQF_HKEY(i), hkey[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
10552
if (!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
10556
vsi = i40e_find_vsi_by_type(pf, I40E_VSI_FDIR);
drivers/net/ethernet/intel/i40e/i40e_main.c
10560
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10561
vsi = i40e_vsi_setup(pf, I40E_VSI_FDIR, main_vsi->seid, 0);
drivers/net/ethernet/intel/i40e/i40e_main.c
10563
dev_info(&pf->pdev->dev, "Couldn't create FDir VSI\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10564
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
10565
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
10577
static void i40e_fdir_teardown(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
1058
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10581
i40e_fdir_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10582
vsi = i40e_find_vsi_by_type(pf, I40E_VSI_FDIR);
drivers/net/ethernet/intel/i40e/i40e_main.c
10598
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
10603
hlist_for_each_entry_safe(cfilter, node, &pf->cloud_filter_list,
drivers/net/ethernet/intel/i40e/i40e_main.c
10615
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10618
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
1063
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
1068
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10707
static void i40e_prep_for_reset(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10709
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
10714
clear_bit(__I40E_RESET_INTR_RECEIVED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10715
if (test_and_set_bit(__I40E_RESET_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10717
if (i40e_check_asq_alive(&pf->hw))
drivers/net/ethernet/intel/i40e/i40e_main.c
10718
i40e_vc_notify_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1072
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10720
dev_dbg(&pf->pdev->dev, "Tearing down internal switch for reset\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10723
i40e_pf_quiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10725
i40e_pf_for_each_vsi(pf, v, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10730
i40e_shutdown_adminq(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
10736
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10743
i40e_ptp_save_hw_time(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10750
static void i40e_send_version(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10759
i40e_aq_send_driver_version(&pf->hw, &dv, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
1076
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
1080
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10813
static int i40e_reset(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
10815
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
10820
dev_info(&pf->pdev->dev, "PF reset failed, %d\n", ret);
drivers/net/ethernet/intel/i40e/i40e_main.c
10821
set_bit(__I40E_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10822
clear_bit(__I40E_RESET_RECOVERY_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
10824
pf->pfr_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
10836
static void i40e_rebuild(struct i40e_pf *pf, bool reinit, bool lock_acquired)
drivers/net/ethernet/intel/i40e/i40e_main.c
10838
const bool is_recovery_mode_reported = i40e_check_recovery_mode(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10839
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10840
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
10846
if (test_bit(__I40E_EMP_RESET_INTR_RECEIVED, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
1085
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10850
if (test_bit(__I40E_DOWN, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
10851
!test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10853
dev_dbg(&pf->pdev->dev, "Rebuilding internal switch\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10856
ret = i40e_init_adminq(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
10858
dev_info(&pf->pdev->dev, "Rebuild AdminQ failed, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
10859
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10862
i40e_get_oem_version(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
10864
if (test_and_clear_bit(__I40E_EMP_RESET_INTR_RECEIVED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10870
if (test_and_clear_bit(__I40E_EMP_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
10871
i40e_verify_eeprom(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10877
if (test_bit(__I40E_RECOVERY_MODE, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10878
if (i40e_get_capabilities(pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
10886
if (i40e_setup_misc_vector_for_recovery_mode(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
10895
free_irq(pf->pdev->irq, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10896
i40e_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10897
if (i40e_restore_interrupt_scheme(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
1090
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10902
i40e_send_version(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10911
ret = i40e_get_capabilities(pf, i40e_aqc_opc_list_func_capabilities);
drivers/net/ethernet/intel/i40e/i40e_main.c
10918
dev_info(&pf->pdev->dev, "init_lan_hmc failed: %d\n", ret);
drivers/net/ethernet/intel/i40e/i40e_main.c
10923
dev_info(&pf->pdev->dev, "configure_lan_hmc failed: %d\n", ret);
drivers/net/ethernet/intel/i40e/i40e_main.c
1093
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10932
if (i40e_is_tc_mqprio_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10939
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10941
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
10944
ret = i40e_init_pf_dcb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
10946
dev_info(&pf->pdev->dev, "DCB init failed %d, disabled\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
10948
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
10957
ret = i40e_setup_pf_switch(pf, reinit, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
1096
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10964
ret = i40e_aq_set_phy_int_mask(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
10969
dev_info(&pf->pdev->dev, "set phy mask fail, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
10970
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10979
if (vsi->uplink_seid != pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10980
dev_dbg(&pf->pdev->dev, "attempting to rebuild switch\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
10983
i40e_pf_for_each_veb(pf, v, veb) {
drivers/net/ethernet/intel/i40e/i40e_main.c
1099
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
10994
if (veb->uplink_seid == pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10995
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
10998
vsi->uplink_seid = pf->mac_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
11001
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11008
if (vsi->uplink_seid == pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11009
dev_dbg(&pf->pdev->dev, "attempting to rebuild PF VSI\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
11013
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1104
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11062
if (test_bit(I40E_HW_CAP_RESTART_AUTONEG, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11064
ret = i40e_aq_set_link_restart_an(&pf->hw, true, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
11066
dev_info(&pf->pdev->dev, "link restart failed, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
11068
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
11071
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11072
ret = i40e_setup_misc_vector(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1108
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11083
i40e_add_filter_to_drop_tx_flow_control_frames(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
11084
pf->main_vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
11087
i40e_pf_unquiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11094
ret = i40e_set_promiscuous(pf, pf->cur_promisc);
drivers/net/ethernet/intel/i40e/i40e_main.c
11096
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11098
pf->cur_promisc ? "on" : "off",
drivers/net/ethernet/intel/i40e/i40e_main.c
11099
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
11101
i40e_reset_all_vfs(pf, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
11104
i40e_send_version(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11113
clear_bit(__I40E_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
11115
clear_bit(__I40E_RESET_RECOVERY_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
11116
clear_bit(__I40E_TIMEOUT_RECOVERY_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
1112
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11126
static void i40e_reset_and_rebuild(struct i40e_pf *pf, bool reinit,
drivers/net/ethernet/intel/i40e/i40e_main.c
11131
if (test_bit(__I40E_IN_REMOVE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11137
ret = i40e_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11139
i40e_rebuild(pf, reinit, lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
11141
dev_err(&pf->pdev->dev, "%s: i40e_reset() FAILED", __func__);
drivers/net/ethernet/intel/i40e/i40e_main.c
11153
static void i40e_handle_reset_warning(struct i40e_pf *pf, bool lock_acquired)
drivers/net/ethernet/intel/i40e/i40e_main.c
11155
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11156
i40e_reset_and_rebuild(pf, false, lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
1116
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11165
static void i40e_print_vf_mdd_event(struct i40e_pf *pf, struct i40e_vf *vf,
drivers/net/ethernet/intel/i40e/i40e_main.c
11168
dev_err(&pf->pdev->dev, is_tx ?
drivers/net/ethernet/intel/i40e/i40e_main.c
11172
pf->hw.pf_id,
drivers/net/ethernet/intel/i40e/i40e_main.c
11175
str_on_off(test_bit(I40E_FLAG_MDD_AUTO_RESET_VF, pf->flags)));
drivers/net/ethernet/intel/i40e/i40e_main.c
11184
static void i40e_print_vfs_mdd_events(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
11189
if (!test_and_clear_bit(__I40E_MDD_VF_PRINT_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11192
if (!__ratelimit(&pf->mdd_message_rate_limit))
drivers/net/ethernet/intel/i40e/i40e_main.c
11195
for (i = 0; i < pf->num_alloc_vfs; i++) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11196
struct i40e_vf *vf = &pf->vf[i];
drivers/net/ethernet/intel/i40e/i40e_main.c
11202
i40e_print_vf_mdd_event(pf, vf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
11209
i40e_print_vf_mdd_event(pf, vf, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
1121
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11213
if (is_printed && !test_bit(I40E_FLAG_MDD_AUTO_RESET_VF, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
11214
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11226
static void i40e_handle_mdd_event(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
11228
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
11234
if (!test_and_clear_bit(__I40E_MDD_EVENT_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11238
i40e_print_vfs_mdd_events(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11249
pf->hw.func_caps.base_queue;
drivers/net/ethernet/intel/i40e/i40e_main.c
11250
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
11251
dev_info(&pf->pdev->dev, "Malicious Driver Detection event 0x%02x on TX queue %d PF number 0x%02x VF number 0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
11261
pf->hw.func_caps.base_queue;
drivers/net/ethernet/intel/i40e/i40e_main.c
11262
if (netif_msg_rx_err(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
11263
dev_info(&pf->pdev->dev, "Malicious Driver Detection event 0x%02x on RX queue %d of function 0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
11273
dev_dbg(&pf->pdev->dev, "TX driver issue detected on PF\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
11278
dev_dbg(&pf->pdev->dev, "RX driver issue detected on PF\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
1128
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11283
for (i = 0; i < pf->num_alloc_vfs && mdd_detected; i++) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11287
vf = &(pf->vf[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
11290
set_bit(__I40E_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
11299
set_bit(__I40E_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
11307
test_bit(I40E_FLAG_MDD_AUTO_RESET_VF, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11312
i40e_print_vf_mdd_event(pf, vf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
11314
i40e_print_vf_mdd_event(pf, vf, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
1132
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11325
i40e_print_vfs_mdd_events(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11334
struct i40e_pf *pf = container_of(work,
drivers/net/ethernet/intel/i40e/i40e_main.c
11340
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
11341
test_bit(__I40E_SUSPENDED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11344
if (test_and_set_bit(__I40E_SERVICE_SCHED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11347
if (!test_bit(__I40E_RECOVERY_MODE, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11348
i40e_detect_recover_hung(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11349
i40e_sync_filters_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11350
i40e_reset_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11351
i40e_handle_mdd_event(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11352
i40e_vc_process_vflr_event(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11353
i40e_watchdog_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11354
i40e_fdir_reinit_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11355
if (test_and_clear_bit(__I40E_CLIENT_RESET, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11357
i40e_notify_client_of_netdev_close(pf, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
11359
i40e_client_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1136
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11361
pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11362
i40e_notify_client_of_l2_param_changes(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11364
i40e_sync_filters_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11366
i40e_reset_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11369
i40e_clean_adminq_subtask(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11373
clear_bit(__I40E_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
11379
if (time_after(jiffies, (start_time + pf->service_timer_period)) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
11380
test_bit(__I40E_ADMINQ_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
11381
test_bit(__I40E_MDD_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
11382
test_bit(__I40E_VFLR_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
11383
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
11392
struct i40e_pf *pf = timer_container_of(pf, t, service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
11394
mod_timer(&pf->service_timer,
drivers/net/ethernet/intel/i40e/i40e_main.c
11395
round_jiffies(jiffies + pf->service_timer_period));
drivers/net/ethernet/intel/i40e/i40e_main.c
11396
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1140
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11405
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
11409
vsi->alloc_queue_pairs = pf->num_lan_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
11416
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
11417
vsi->num_q_vectors = pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11429
vsi->num_q_vectors = pf->num_fdsb_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11433
vsi->alloc_queue_pairs = pf->num_vmdq_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
1144
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11440
vsi->num_q_vectors = pf->num_vmdq_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11444
vsi->alloc_queue_pairs = pf->num_vf_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
1148
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11517
static int i40e_vsi_mem_alloc(struct i40e_pf *pf, enum i40e_vsi_type type)
drivers/net/ethernet/intel/i40e/i40e_main.c
1152
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11525
mutex_lock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
11533
i = pf->next_vsi;
drivers/net/ethernet/intel/i40e/i40e_main.c
11534
while (i < pf->num_alloc_vsi && pf->vsi[i])
drivers/net/ethernet/intel/i40e/i40e_main.c
11536
if (i >= pf->num_alloc_vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11538
while (i < pf->next_vsi && pf->vsi[i])
drivers/net/ethernet/intel/i40e/i40e_main.c
11542
if (i < pf->num_alloc_vsi && !pf->vsi[i]) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11548
pf->next_vsi = ++i;
drivers/net/ethernet/intel/i40e/i40e_main.c
11556
vsi->back = pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
11562
pf->rss_table_size : 64;
drivers/net/ethernet/intel/i40e/i40e_main.c
11569
vsi->af_xdp_zc_qps = bitmap_zalloc(pf->num_lan_qps, GFP_KERNEL);
drivers/net/ethernet/intel/i40e/i40e_main.c
1157
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11587
pf->vsi[vsi_idx] = vsi;
drivers/net/ethernet/intel/i40e/i40e_main.c
11593
pf->next_vsi = i - 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11596
mutex_unlock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
1161
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11644
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
1165
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11651
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
11653
mutex_lock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
11654
if (!pf->vsi[vsi->idx]) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11655
dev_err(&pf->pdev->dev, "pf->vsi[%d] is NULL, just free vsi[%d](type %d)\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
11660
if (pf->vsi[vsi->idx] != vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11661
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11663
pf->vsi[vsi->idx]->idx,
drivers/net/ethernet/intel/i40e/i40e_main.c
11664
pf->vsi[vsi->idx]->type,
drivers/net/ethernet/intel/i40e/i40e_main.c
11670
i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_main.c
11671
i40e_put_lump(pf->irq_pile, vsi->base_vector, vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_main.c
11677
pf->vsi[vsi->idx] = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
11678
if (vsi->idx < pf->next_vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
11679
pf->next_vsi = vsi->idx;
drivers/net/ethernet/intel/i40e/i40e_main.c
11682
mutex_unlock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
1169
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11715
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
1173
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11730
ring->dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_main.c
11736
ring->itr_setting = pf->tx_itr_default;
drivers/net/ethernet/intel/i40e/i40e_main.c
11747
ring->dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_main.c
11754
ring->itr_setting = pf->tx_itr_default;
drivers/net/ethernet/intel/i40e/i40e_main.c
11763
ring->dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_main.c
11767
ring->itr_setting = pf->rx_itr_default;
drivers/net/ethernet/intel/i40e/i40e_main.c
1177
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11785
static int i40e_reserve_msix_vectors(struct i40e_pf *pf, int vectors)
drivers/net/ethernet/intel/i40e/i40e_main.c
11787
vectors = pci_enable_msix_range(pf->pdev, pf->msix_entries,
drivers/net/ethernet/intel/i40e/i40e_main.c
11790
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11806
static int i40e_init_msix(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
11808
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
1181
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11815
if (!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
1185
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11850
pf->num_lan_msix = min_t(int, cpus, vectors_left / 2);
drivers/net/ethernet/intel/i40e/i40e_main.c
11851
vectors_left -= pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11854
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11856
pf->num_fdsb_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11860
pf->num_fdsb_msix = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
11865
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11866
iwarp_requested = pf->num_iwarp_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11869
pf->num_iwarp_msix = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
11870
else if (vectors_left < pf->num_iwarp_msix)
drivers/net/ethernet/intel/i40e/i40e_main.c
11871
pf->num_iwarp_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11872
v_budget += pf->num_iwarp_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11873
vectors_left -= pf->num_iwarp_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11877
if (test_bit(I40E_FLAG_VMDQ_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11879
pf->num_vmdq_msix = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
1188
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11880
pf->num_vmdq_qps = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
11883
pf->num_vmdq_vsis * pf->num_vmdq_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
11894
pf->num_vmdq_qps = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11895
vmdq_vecs_wanted = pf->num_vmdq_vsis;
drivers/net/ethernet/intel/i40e/i40e_main.c
11900
pf->num_vmdq_msix = pf->num_vmdq_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
1191
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11916
extra_vectors = min_t(int, cpus - pf->num_lan_msix, vectors_left);
drivers/net/ethernet/intel/i40e/i40e_main.c
11917
pf->num_lan_msix += extra_vectors;
drivers/net/ethernet/intel/i40e/i40e_main.c
11923
v_budget += pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
11924
pf->msix_entries = kzalloc_objs(struct msix_entry, v_budget);
drivers/net/ethernet/intel/i40e/i40e_main.c
11925
if (!pf->msix_entries)
drivers/net/ethernet/intel/i40e/i40e_main.c
11929
pf->msix_entries[i].entry = i;
drivers/net/ethernet/intel/i40e/i40e_main.c
11930
v_actual = i40e_reserve_msix_vectors(pf, v_budget);
drivers/net/ethernet/intel/i40e/i40e_main.c
11933
clear_bit(I40E_FLAG_MSIX_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
11934
kfree(pf->msix_entries);
drivers/net/ethernet/intel/i40e/i40e_main.c
11935
pf->msix_entries = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
11936
pci_disable_msix(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
1194
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
11941
pf->num_vmdq_vsis = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
11942
pf->num_vmdq_qps = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
11943
pf->num_lan_qps = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11944
pf->num_lan_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11954
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
11961
pf->num_vmdq_msix = 1; /* force VMDqs to only one vector */
drivers/net/ethernet/intel/i40e/i40e_main.c
11962
pf->num_vmdq_vsis = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11963
pf->num_vmdq_qps = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11968
pf->num_lan_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11971
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11972
pf->num_lan_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11973
pf->num_iwarp_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11975
pf->num_lan_msix = 2;
drivers/net/ethernet/intel/i40e/i40e_main.c
11979
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11980
pf->num_iwarp_msix = min_t(int, (vec / 3),
drivers/net/ethernet/intel/i40e/i40e_main.c
11982
pf->num_vmdq_vsis = min_t(int, (vec / 3),
drivers/net/ethernet/intel/i40e/i40e_main.c
11985
pf->num_vmdq_vsis = min_t(int, (vec / 2),
drivers/net/ethernet/intel/i40e/i40e_main.c
11988
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
11989
pf->num_fdsb_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
11992
pf->num_lan_msix = min_t(int,
drivers/net/ethernet/intel/i40e/i40e_main.c
11993
(vec - (pf->num_iwarp_msix + pf->num_vmdq_vsis)),
drivers/net/ethernet/intel/i40e/i40e_main.c
11994
pf->num_lan_msix);
drivers/net/ethernet/intel/i40e/i40e_main.c
11995
pf->num_lan_qps = pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
12000
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) && pf->num_fdsb_msix == 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12001
dev_info(&pf->pdev->dev, "Sideband Flowdir disabled, not enough MSI-X vectors\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12002
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12003
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12005
if (test_bit(I40E_FLAG_VMDQ_ENA, pf->flags) && pf->num_vmdq_msix == 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12006
dev_info(&pf->pdev->dev, "VMDq disabled, not enough MSI-X vectors\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12007
clear_bit(I40E_FLAG_VMDQ_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12010
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12011
pf->num_iwarp_msix == 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12012
dev_info(&pf->pdev->dev, "IWARP disabled, not enough MSI-X vectors\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12013
clear_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12015
i40e_debug(&pf->hw, I40E_DEBUG_INIT,
drivers/net/ethernet/intel/i40e/i40e_main.c
12017
pf->num_lan_msix,
drivers/net/ethernet/intel/i40e/i40e_main.c
12018
pf->num_vmdq_msix * pf->num_vmdq_vsis,
drivers/net/ethernet/intel/i40e/i40e_main.c
12019
pf->num_fdsb_msix,
drivers/net/ethernet/intel/i40e/i40e_main.c
12020
pf->num_iwarp_msix);
drivers/net/ethernet/intel/i40e/i40e_main.c
12063
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12067
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
12093
static int i40e_init_interrupt_scheme(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12098
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12099
vectors = i40e_init_msix(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12101
clear_bit(I40E_FLAG_MSIX_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12102
clear_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12103
clear_bit(I40E_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12104
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12105
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12106
clear_bit(I40E_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12107
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12108
clear_bit(I40E_FLAG_FD_ATR_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12109
clear_bit(I40E_FLAG_VMDQ_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12110
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12113
i40e_determine_queue_usage(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12117
if (!test_bit(I40E_FLAG_MSIX_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12118
test_bit(I40E_FLAG_MSI_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12119
dev_info(&pf->pdev->dev, "MSI-X not available, trying MSI\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12120
vectors = pci_enable_msi(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
12122
dev_info(&pf->pdev->dev, "MSI init failed - %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
12124
clear_bit(I40E_FLAG_MSI_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12129
if (!test_bit(I40E_FLAG_MSI_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12130
!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
12131
dev_info(&pf->pdev->dev, "MSI-X and MSI not available, falling back to Legacy IRQ\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12135
pf->irq_pile = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/intel/i40e/i40e_main.c
12136
if (!pf->irq_pile)
drivers/net/ethernet/intel/i40e/i40e_main.c
12139
pf->irq_pile->num_entries = vectors;
drivers/net/ethernet/intel/i40e/i40e_main.c
1214
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
12142
(void)i40e_get_lump(pf, pf->irq_pile, 1, I40E_PILE_VALID_BIT - 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
12155
static int i40e_restore_interrupt_scheme(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12164
set_bit(I40E_FLAG_MSI_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12165
set_bit(I40E_FLAG_MSIX_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12167
err = i40e_init_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1217
pf->stat_offsets_loaded,
drivers/net/ethernet/intel/i40e/i40e_main.c
12174
i40e_pf_for_each_vsi(pf, i, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12182
err = i40e_setup_misc_vector(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12186
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
12187
i40e_client_update_msix_info(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12193
if (pf->vsi[i])
drivers/net/ethernet/intel/i40e/i40e_main.c
12194
i40e_vsi_free_q_vectors(pf->vsi[i]);
drivers/net/ethernet/intel/i40e/i40e_main.c
1220
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
1221
!test_bit(__I40E_FD_SB_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
12210
static int i40e_setup_misc_vector_for_recovery_mode(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12214
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12215
err = i40e_setup_misc_vector(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12218
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12224
u32 flags = test_bit(I40E_FLAG_MSI_ENA, pf->flags) ? 0 : IRQF_SHARED;
drivers/net/ethernet/intel/i40e/i40e_main.c
12226
err = request_irq(pf->pdev->irq, i40e_intr, flags,
drivers/net/ethernet/intel/i40e/i40e_main.c
12227
pf->int_name, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12230
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12235
i40e_enable_misc_int_causes(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12236
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12250
static int i40e_setup_misc_vector(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12252
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12256
if (!test_and_set_bit(__I40E_MISC_IRQ_REQUESTED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12257
err = request_irq(pf->msix_entries[0].vector,
drivers/net/ethernet/intel/i40e/i40e_main.c
12258
i40e_intr, 0, pf->int_name, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1226
if (test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12260
clear_bit(__I40E_MISC_IRQ_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
12261
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12263
pf->int_name, err);
drivers/net/ethernet/intel/i40e/i40e_main.c
12268
i40e_enable_misc_int_causes(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1227
!test_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
12276
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12293
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12294
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12301
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12304
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
12314
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12317
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
1232
pf->stat_offsets_loaded = true;
drivers/net/ethernet/intel/i40e/i40e_main.c
12337
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12338
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12353
dev_err(&pf->pdev->dev, "Cannot set RSS seed - invalid VSI type\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12371
dev_err(&pf->pdev->dev, "Cannot set RSS LUT - invalid VSI type\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12391
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12392
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12424
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12426
if (test_bit(I40E_HW_CAP_RSS_AQ, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
1243
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12443
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12445
if (test_bit(I40E_HW_CAP_RSS_AQ, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
12458
void i40e_fill_rss_lut(struct i40e_pf *pf, u8 *lut,
drivers/net/ethernet/intel/i40e/i40e_main.c
1246
i40e_update_pf_stats(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12471
static int i40e_pf_config_rss(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12473
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12476
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12484
hena |= i40e_pf_get_default_rss_hashcfg(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12491
reg_val = (pf->rss_table_size == 512) ?
drivers/net/ethernet/intel/i40e/i40e_main.c
12506
vsi->rss_size = min_t(int, pf->alloc_rss_size, qcount);
drivers/net/ethernet/intel/i40e/i40e_main.c
12519
i40e_fill_rss_lut(pf, lut, vsi->rss_table_size, vsi->rss_size);
drivers/net/ethernet/intel/i40e/i40e_main.c
12543
int i40e_reconfig_rss_queues(struct i40e_pf *pf, int queue_count)
drivers/net/ethernet/intel/i40e/i40e_main.c
12545
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12548
if (!test_bit(I40E_FLAG_RSS_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
12552
new_rss_size = min_t(int, queue_count, pf->rss_size_max);
drivers/net/ethernet/intel/i40e/i40e_main.c
12558
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12559
if (test_bit(__I40E_IN_REMOVE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
12560
return pf->alloc_rss_size;
drivers/net/ethernet/intel/i40e/i40e_main.c
12562
pf->alloc_rss_size = new_rss_size;
drivers/net/ethernet/intel/i40e/i40e_main.c
12564
i40e_reset_and_rebuild(pf, true, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
12571
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12577
vsi->rss_size = min_t(int, pf->alloc_rss_size, qcount);
drivers/net/ethernet/intel/i40e/i40e_main.c
12579
i40e_pf_config_rss(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12581
dev_info(&pf->pdev->dev, "User requested queue count/HW max RSS count: %d/%d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
12582
vsi->req_queue_pairs, pf->rss_size_max);
drivers/net/ethernet/intel/i40e/i40e_main.c
12583
return pf->alloc_rss_size;
drivers/net/ethernet/intel/i40e/i40e_main.c
12590
int i40e_get_partition_bw_setting(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12596
status = i40e_read_bw_from_alt_ram(&pf->hw, &max_bw, &min_bw,
drivers/net/ethernet/intel/i40e/i40e_main.c
12601
pf->min_bw = min_bw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12603
pf->max_bw = max_bw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12613
int i40e_set_partition_bw_setting(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12621
bw_data.pf_valid_bits = cpu_to_le16(BIT(pf->hw.pf_id));
drivers/net/ethernet/intel/i40e/i40e_main.c
12622
bw_data.max_bw[pf->hw.pf_id] = pf->max_bw & I40E_ALT_BW_VALUE_MASK;
drivers/net/ethernet/intel/i40e/i40e_main.c
12623
bw_data.min_bw[pf->hw.pf_id] = pf->min_bw & I40E_ALT_BW_VALUE_MASK;
drivers/net/ethernet/intel/i40e/i40e_main.c
12626
status = i40e_aq_configure_partition_bw(&pf->hw, &bw_data, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
12636
static bool i40e_is_total_port_shutdown_enabled(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12651
read_status = i40e_read_nvm_word(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
12656
read_status = i40e_read_nvm_word(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
12663
read_status = i40e_read_nvm_module_data(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
12671
link_behavior >>= (pf->hw.port * I40E_LINK_BEHAVIOR_PORT_BIT_LENGTH);
drivers/net/ethernet/intel/i40e/i40e_main.c
12677
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12691
static int i40e_sw_init(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
12698
bitmap_zero(pf->flags, I40E_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/i40e/i40e_main.c
12699
set_bit(I40E_FLAG_MSI_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12700
set_bit(I40E_FLAG_MSIX_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12703
pf->rx_itr_default = I40E_ITR_RX_DEF;
drivers/net/ethernet/intel/i40e/i40e_main.c
12704
pf->tx_itr_default = I40E_ITR_TX_DEF;
drivers/net/ethernet/intel/i40e/i40e_main.c
12709
pf->rss_size_max = BIT(pf->hw.func_caps.rss_table_entry_width);
drivers/net/ethernet/intel/i40e/i40e_main.c
12710
pf->alloc_rss_size = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
12711
pf->rss_table_size = pf->hw.func_caps.rss_table_size;
drivers/net/ethernet/intel/i40e/i40e_main.c
12712
pf->rss_size_max = min_t(int, pf->rss_size_max,
drivers/net/ethernet/intel/i40e/i40e_main.c
12713
pf->hw.func_caps.num_tx_qp);
drivers/net/ethernet/intel/i40e/i40e_main.c
12717
pf->rss_size_max = min_t(int, pf->rss_size_max, pow);
drivers/net/ethernet/intel/i40e/i40e_main.c
12719
if (pf->hw.func_caps.rss) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12720
set_bit(I40E_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12721
pf->alloc_rss_size = min_t(int, pf->rss_size_max,
drivers/net/ethernet/intel/i40e/i40e_main.c
12726
if (pf->hw.func_caps.npar_enable || pf->hw.func_caps.flex10_enable) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12727
set_bit(I40E_FLAG_MFP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12728
dev_info(&pf->pdev->dev, "MFP mode Enabled\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12729
if (i40e_get_partition_bw_setting(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12730
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12733
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12735
pf->min_bw, pf->max_bw);
drivers/net/ethernet/intel/i40e/i40e_main.c
12738
i40e_set_partition_bw_setting(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12742
if ((pf->hw.func_caps.fd_filters_guaranteed > 0) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
12743
(pf->hw.func_caps.fd_filters_best_effort > 0)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12744
set_bit(I40E_FLAG_FD_ATR_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12745
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12746
pf->hw.num_partitions > 1)
drivers/net/ethernet/intel/i40e/i40e_main.c
12747
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12750
set_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12751
pf->fdir_pf_filter_count =
drivers/net/ethernet/intel/i40e/i40e_main.c
12752
pf->hw.func_caps.fd_filters_guaranteed;
drivers/net/ethernet/intel/i40e/i40e_main.c
12753
pf->hw.fdir_shared_filter_count =
drivers/net/ethernet/intel/i40e/i40e_main.c
12754
pf->hw.func_caps.fd_filters_best_effort;
drivers/net/ethernet/intel/i40e/i40e_main.c
12758
if (test_bit(I40E_HW_CAP_ATR_EVICT, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
12759
set_bit(I40E_FLAG_HW_ATR_EVICT_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12761
if (pf->hw.func_caps.vmdq && num_online_cpus() != 1) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12762
pf->num_vmdq_vsis = I40E_DEFAULT_NUM_VMDQ_VSI;
drivers/net/ethernet/intel/i40e/i40e_main.c
12763
set_bit(I40E_FLAG_VMDQ_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12764
pf->num_vmdq_qps = i40e_default_queues_per_vmdq(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12767
if (pf->hw.func_caps.iwarp && num_online_cpus() != 1) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12768
set_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12770
pf->num_iwarp_msix = (int)num_online_cpus() + 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
12777
if (pf->hw.mac.type == I40E_MAC_XL710 &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12778
pf->hw.func_caps.npar_enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
12779
clear_bit(I40E_HW_CAP_FW_LLDP_STOPPABLE, pf->hw.caps);
drivers/net/ethernet/intel/i40e/i40e_main.c
12782
if (pf->hw.func_caps.num_vfs && pf->hw.partition_id == 1) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12783
pf->num_vf_qps = I40E_DEFAULT_QUEUES_PER_VF;
drivers/net/ethernet/intel/i40e/i40e_main.c
12784
set_bit(I40E_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12785
pf->num_req_vfs = min_t(int,
drivers/net/ethernet/intel/i40e/i40e_main.c
12786
pf->hw.func_caps.num_vfs,
drivers/net/ethernet/intel/i40e/i40e_main.c
12790
pf->lan_veb = I40E_NO_VEB;
drivers/net/ethernet/intel/i40e/i40e_main.c
12791
pf->lan_vsi = I40E_NO_VSI;
drivers/net/ethernet/intel/i40e/i40e_main.c
12794
clear_bit(I40E_FLAG_VEB_STATS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12798
+ (sizeof(u16) * pf->hw.func_caps.num_tx_qp);
drivers/net/ethernet/intel/i40e/i40e_main.c
12799
pf->qp_pile = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/intel/i40e/i40e_main.c
12800
if (!pf->qp_pile) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12804
pf->qp_pile->num_entries = pf->hw.func_caps.num_tx_qp;
drivers/net/ethernet/intel/i40e/i40e_main.c
12806
pf->tx_timeout_recovery_level = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
12808
if (pf->hw.mac.type != I40E_MAC_X722 &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12809
i40e_is_total_port_shutdown_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12813
set_bit(I40E_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12814
set_bit(I40E_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12815
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12818
mutex_init(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
12831
bool i40e_set_ntuple(struct i40e_pf *pf, netdev_features_t features)
drivers/net/ethernet/intel/i40e/i40e_main.c
12840
if (!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
12845
if (pf->num_fdsb_msix > 0 && !pf->num_cloud_filters) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12846
set_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12847
clear_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12851
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12853
i40e_fdir_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12855
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12856
clear_bit(__I40E_FD_SB_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
12857
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
12860
pf->fd_add_err = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
12861
pf->fd_atr_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
12863
if (test_and_clear_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
12864
if (test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
12865
(I40E_DEBUG_FD & pf->hw.debug_mask))
drivers/net/ethernet/intel/i40e/i40e_main.c
12866
dev_info(&pf->pdev->dev, "ATR re-enabled.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12877
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12878
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
12889
dev_err(&pf->pdev->dev, "Cannot set RSS LUT - invalid VSI type\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
12927
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
12931
i40e_pf_config_rss(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
12942
(netdev->features & NETIF_F_HW_TC) && pf->num_cloud_filters) {
drivers/net/ethernet/intel/i40e/i40e_main.c
12943
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
12951
need_reset = i40e_set_ntuple(pf, features);
drivers/net/ethernet/intel/i40e/i40e_main.c
12954
i40e_do_reset(pf, I40E_PF_RESET_FLAG, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
13008
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13009
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
13011
if (!test_bit(I40E_HW_CAP_PORT_ID_VALID, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
13038
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13041
if (!test_bit(I40E_FLAG_SRIOV_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
13094
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13104
veb = i40e_pf_get_veb_by_seid(pf, vsi->uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
13119
veb = i40e_veb_setup(pf, vsi->uplink_seid, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
13134
set_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
13136
clear_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
13137
i40e_do_reset(pf, I40E_PF_RESET_FLAG, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
13164
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13172
veb = i40e_pf_get_veb_by_seid(pf, vsi->uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
13248
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13254
if (prog && test_bit(__I40E_IN_REMOVE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
13266
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
13276
i40e_reset_and_rebuild(pf, true, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
13315
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13318
while (test_and_set_bit(__I40E_CONFIG_BUSY, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
13334
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13336
clear_bit(__I40E_CONFIG_BUSY, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
13409
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13413
ret = i40e_control_wait_tx_q(vsi->seid, pf, pf_q,
drivers/net/ethernet/intel/i40e/i40e_main.c
13416
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13422
i40e_control_rx_q(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
13423
ret = i40e_pf_rxq_wait(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
13425
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13440
ret = i40e_control_wait_tx_q(vsi->seid, pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
13444
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13460
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13461
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
13464
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
13467
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
13480
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13481
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
13489
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
13494
synchronize_irq(pf->msix_entries[intpf].vector);
drivers/net/ethernet/intel/i40e/i40e_main.c
13500
synchronize_irq(pf->pdev->irq);
drivers/net/ethernet/intel/i40e/i40e_main.c
13636
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13637
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
13675
if (!test_bit(I40E_HW_CAP_OUTER_UDP_CSUM, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
13678
netdev->udp_tunnel_nic_info = &pf->udp_tunnel_nic;
drivers/net/ethernet/intel/i40e/i40e_main.c
13711
if (!test_bit(I40E_FLAG_MFP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
13722
SET_NETDEV_DEV(netdev, &pf->pdev->dev);
drivers/net/ethernet/intel/i40e/i40e_main.c
13750
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
13825
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13831
veb = pf->veb[vsi->veb_idx];
drivers/net/ethernet/intel/i40e/i40e_main.c
13833
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13860
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
13861
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
13878
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
13879
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_main.c
13881
ret = i40e_aq_get_vsi_params(&pf->hw, &ctxt, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
13884
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13887
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13896
enabled_tc = i40e_pf_get_tc_map(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
13902
if (test_bit(I40E_FLAG_SOURCE_PRUNING_DIS, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
13904
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
13905
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_main.c
13913
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13916
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13923
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
13924
!(pf->hw.func_caps.iscsi)) { /* NIC type PF */
drivers/net/ethernet/intel/i40e/i40e_main.c
13926
ctxt.seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
13927
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_main.c
13932
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13935
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13954
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
13958
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13969
if (test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
14027
if (pf->vf[vsi->vf_id].spoofchk) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14052
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14074
set_bit(__I40E_MACVLAN_SYNC_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
14080
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14082
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14102
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14106
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
14110
dev_info(&pf->pdev->dev, "VSI %d has existing VEB %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14114
if (vsi->type == I40E_VSI_MAIN && !test_bit(__I40E_DOWN, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14115
dev_info(&pf->pdev->dev, "Can't remove PF VSI\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
14135
i40e_devlink_destroy_port(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14170
veb = i40e_pf_get_veb_by_seid(pf, uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14175
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
14203
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
14206
dev_info(&pf->pdev->dev, "VSI %d has existing q_vectors\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14212
dev_info(&pf->pdev->dev, "VSI %d has non-zero base vector %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14219
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14229
if (!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
14232
vsi->base_vector = i40e_get_lump(pf, pf->irq_pile,
drivers/net/ethernet/intel/i40e/i40e_main.c
14235
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14260
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14266
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
14268
i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_main.c
14280
ret = i40e_get_lump(pf, pf->qp_pile, alloc_queue_pairs, vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_main.c
14282
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14292
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14293
main_vsi->seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
14297
i40e_rm_default_mac_filter(vsi, pf->hw.mac.perm_addr);
drivers/net/ethernet/intel/i40e/i40e_main.c
14317
i40e_devlink_destroy_port(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14318
i40e_aq_delete_element(&pf->hw, vsi->seid, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
14337
struct i40e_vsi *i40e_vsi_setup(struct i40e_pf *pf, u8 type,
drivers/net/ethernet/intel/i40e/i40e_main.c
14359
veb = i40e_pf_get_veb_by_seid(pf, uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14360
if (!veb && uplink_seid != pf->mac_seid) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14361
vsi = i40e_pf_get_vsi_by_seid(pf, uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14363
dev_info(&pf->pdev->dev, "no such uplink_seid %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14368
if (vsi->uplink_seid == pf->mac_seid)
drivers/net/ethernet/intel/i40e/i40e_main.c
14369
veb = i40e_veb_setup(pf, pf->mac_seid, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
14372
veb = i40e_veb_setup(pf, vsi->uplink_seid, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
14384
if (!test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14386
clear_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
14390
veb = i40e_pf_get_veb_by_seid(pf, vsi->uplink_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14392
dev_info(&pf->pdev->dev, "couldn't add VEB\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
14401
v_idx = i40e_vsi_mem_alloc(pf, type);
drivers/net/ethernet/intel/i40e/i40e_main.c
14404
vsi = pf->vsi[v_idx];
drivers/net/ethernet/intel/i40e/i40e_main.c
14411
pf->lan_vsi = v_idx;
drivers/net/ethernet/intel/i40e/i40e_main.c
14418
ret = i40e_get_lump(pf, pf->qp_pile, alloc_queue_pairs, vsi->idx);
drivers/net/ethernet/intel/i40e/i40e_main.c
14420
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14444
ret = i40e_devlink_create_port(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14447
SET_NETDEV_DEVLINK_PORT(vsi->netdev, &pf->devlink_port);
drivers/net/ethernet/intel/i40e/i40e_main.c
14479
if (test_bit(I40E_HW_CAP_RSS_AQ, pf->hw.caps) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
145
struct i40e_pf *pf = i40e_hw_to_pf(hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
14500
i40e_devlink_destroy_port(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14502
i40e_aq_delete_element(&pf->hw, vsi->seid, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
14519
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14520
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
14528
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14537
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14567
static int i40e_veb_mem_alloc(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
14574
mutex_lock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
14583
while ((i < I40E_MAX_VEB) && (pf->veb[i] != NULL))
drivers/net/ethernet/intel/i40e/i40e_main.c
14595
veb->pf = pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14599
pf->veb[i] = veb;
drivers/net/ethernet/intel/i40e/i40e_main.c
14602
mutex_unlock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
14615
struct i40e_pf *pf = branch->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14623
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e_main.c
14632
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
14642
if (pf->veb[veb_idx])
drivers/net/ethernet/intel/i40e/i40e_main.c
14643
i40e_veb_release(pf->veb[veb_idx]);
drivers/net/ethernet/intel/i40e/i40e_main.c
14655
if (veb->pf) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14656
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14658
mutex_lock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
14659
if (pf->veb[veb->idx] == veb)
drivers/net/ethernet/intel/i40e/i40e_main.c
14660
pf->veb[veb->idx] = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
14661
mutex_unlock(&pf->switch_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
14674
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14677
pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14680
i40e_pf_for_each_vsi(pf, i, vsi_it)
drivers/net/ethernet/intel/i40e/i40e_main.c
14691
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
147
return &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_main.c
14704
i40e_aq_delete_element(&pf->hw, veb->seid, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
14715
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14716
bool enable_stats = !!test_bit(I40E_FLAG_VEB_STATS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
14719
ret = i40e_aq_add_veb(&pf->hw, veb->uplink_seid, vsi ? vsi->seid : 0,
drivers/net/ethernet/intel/i40e/i40e_main.c
14725
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14727
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14732
ret = i40e_aq_get_veb_parameters(&pf->hw, veb->seid, NULL, NULL,
drivers/net/ethernet/intel/i40e/i40e_main.c
14735
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14737
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14742
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14744
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14745
i40e_aq_delete_element(&pf->hw, veb->seid, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
14773
struct i40e_veb *i40e_veb_setup(struct i40e_pf *pf, u16 uplink_seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
14784
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14792
vsi = i40e_pf_get_vsi_by_seid(pf, vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14794
dev_err(&pf->pdev->dev, "vsi seid %d not found\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14801
veb_idx = i40e_veb_mem_alloc(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14804
veb = pf->veb[veb_idx];
drivers/net/ethernet/intel/i40e/i40e_main.c
14813
if (vsi && vsi->idx == pf->lan_vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
14814
pf->lan_veb = veb->idx;
drivers/net/ethernet/intel/i40e/i40e_main.c
14833
static void i40e_setup_pf_switch_element(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
14844
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14850
pf->mac_seid = seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
14854
if (uplink_seid != pf->mac_seid)
drivers/net/ethernet/intel/i40e/i40e_main.c
14856
veb = i40e_pf_get_main_veb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14861
veb = i40e_pf_get_veb_by_seid(pf, seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14863
pf->lan_veb = veb->idx;
drivers/net/ethernet/intel/i40e/i40e_main.c
14865
v = i40e_veb_mem_alloc(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14868
pf->lan_veb = v;
drivers/net/ethernet/intel/i40e/i40e_main.c
14873
veb = i40e_pf_get_main_veb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14878
veb->uplink_seid = pf->mac_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
14879
veb->pf = pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
14887
pf->mac_seid = uplink_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
14888
pf->main_vsi_seid = seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
14890
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14892
downlink_seid, pf->main_vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
14903
dev_info(&pf->pdev->dev, "unknown element type=%d seid=%d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
14917
int i40e_fetch_switch_configuration(struct i40e_pf *pf, bool printconfig)
drivers/net/ethernet/intel/i40e/i40e_main.c
14933
ret = i40e_aq_get_switch_config(&pf->hw, sw_config,
drivers/net/ethernet/intel/i40e/i40e_main.c
14937
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14939
ret, libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14948
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14956
i40e_setup_pf_switch_element(pf, ele, num_reported,
drivers/net/ethernet/intel/i40e/i40e_main.c
14973
static int i40e_setup_pf_switch(struct i40e_pf *pf, bool reinit, bool lock_acquired)
drivers/net/ethernet/intel/i40e/i40e_main.c
14980
ret = i40e_fetch_switch_configuration(pf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
14982
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
14984
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14987
i40e_pf_reset_stats(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
14995
if ((pf->hw.pf_id == 0) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
14996
!test_bit(I40E_FLAG_TRUE_PROMISC_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
14998
pf->last_sw_conf_flags = flags;
drivers/net/ethernet/intel/i40e/i40e_main.c
15001
if (pf->hw.pf_id == 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15005
ret = i40e_aq_set_switch_config(&pf->hw, flags, valid_flags, 0,
drivers/net/ethernet/intel/i40e/i40e_main.c
15007
if (ret && pf->hw.aq.asq_last_status != LIBIE_AQ_RC_ESRCH) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15008
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
15011
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15014
pf->last_sw_conf_valid_flags = valid_flags;
drivers/net/ethernet/intel/i40e/i40e_main.c
15018
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15026
veb = i40e_pf_get_main_veb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15030
uplink_seid = pf->mac_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
15032
main_vsi = i40e_vsi_setup(pf, I40E_VSI_MAIN,
drivers/net/ethernet/intel/i40e/i40e_main.c
15037
dev_info(&pf->pdev->dev, "setup of MAIN VSI failed\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15038
i40e_cloud_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15039
i40e_fdir_teardown(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15044
main_vsi->seid = pf->main_vsi_seid;
drivers/net/ethernet/intel/i40e/i40e_main.c
15049
i40e_fdir_sb_setup(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15052
ret = i40e_setup_pf_filter_control(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15054
dev_info(&pf->pdev->dev, "setup_pf_filter_control failed: %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
15062
if (test_bit(I40E_FLAG_RSS_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15063
i40e_pf_config_rss(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15066
i40e_link_event(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15068
i40e_ptp_init(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15086
static void i40e_determine_queue_usage(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15091
pf->num_lan_qps = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
15097
queues_left = pf->hw.func_caps.num_tx_qp;
drivers/net/ethernet/intel/i40e/i40e_main.c
1510
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
15100
!test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15103
pf->alloc_rss_size = pf->num_lan_qps = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
15106
clear_bit(I40E_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15107
clear_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15108
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15109
clear_bit(I40E_FLAG_FD_ATR_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15110
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15111
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15112
clear_bit(I40E_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15113
clear_bit(I40E_FLAG_VMDQ_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15114
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15115
} else if (!test_bit(I40E_FLAG_RSS_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15116
!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15117
!test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15118
!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15120
pf->alloc_rss_size = pf->num_lan_qps = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
15121
queues_left -= pf->num_lan_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
15123
clear_bit(I40E_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15124
clear_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15125
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15126
clear_bit(I40E_FLAG_FD_ATR_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15127
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15128
clear_bit(I40E_FLAG_VMDQ_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15129
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15132
if (test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15134
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15135
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15136
dev_info(&pf->pdev->dev, "not enough queues for DCB. DCB is disabled.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15140
q_max = max_t(int, pf->rss_size_max, num_online_cpus());
drivers/net/ethernet/intel/i40e/i40e_main.c
15141
q_max = min_t(int, q_max, pf->hw.func_caps.num_tx_qp);
drivers/net/ethernet/intel/i40e/i40e_main.c
15142
q_max = min_t(int, q_max, pf->hw.func_caps.num_msix_vectors);
drivers/net/ethernet/intel/i40e/i40e_main.c
15143
pf->num_lan_qps = q_max;
drivers/net/ethernet/intel/i40e/i40e_main.c
15145
queues_left -= pf->num_lan_qps;
drivers/net/ethernet/intel/i40e/i40e_main.c
15148
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15152
clear_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15153
set_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15154
dev_info(&pf->pdev->dev, "not enough queues for Flow Director. Flow Director feature is disabled\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15158
if (test_bit(I40E_FLAG_SRIOV_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15159
pf->num_vf_qps && pf->num_req_vfs && queues_left) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15160
pf->num_req_vfs = min_t(int, pf->num_req_vfs,
drivers/net/ethernet/intel/i40e/i40e_main.c
15161
(queues_left / pf->num_vf_qps));
drivers/net/ethernet/intel/i40e/i40e_main.c
15162
queues_left -= (pf->num_req_vfs * pf->num_vf_qps);
drivers/net/ethernet/intel/i40e/i40e_main.c
15165
if (test_bit(I40E_FLAG_VMDQ_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15166
pf->num_vmdq_vsis && pf->num_vmdq_qps && queues_left) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15167
pf->num_vmdq_vsis = min_t(int, pf->num_vmdq_vsis,
drivers/net/ethernet/intel/i40e/i40e_main.c
15168
(queues_left / pf->num_vmdq_qps));
drivers/net/ethernet/intel/i40e/i40e_main.c
15169
queues_left -= (pf->num_vmdq_vsis * pf->num_vmdq_qps);
drivers/net/ethernet/intel/i40e/i40e_main.c
15172
pf->queues_left = queues_left;
drivers/net/ethernet/intel/i40e/i40e_main.c
15173
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
15175
pf->hw.func_caps.num_tx_qp,
drivers/net/ethernet/intel/i40e/i40e_main.c
15176
!!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags),
drivers/net/ethernet/intel/i40e/i40e_main.c
15177
pf->num_lan_qps, pf->alloc_rss_size, pf->num_req_vfs,
drivers/net/ethernet/intel/i40e/i40e_main.c
15178
pf->num_vf_qps, pf->num_vmdq_vsis, pf->num_vmdq_qps,
drivers/net/ethernet/intel/i40e/i40e_main.c
15193
static int i40e_setup_pf_filter_control(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15195
struct i40e_filter_control_settings *settings = &pf->filter_settings;
drivers/net/ethernet/intel/i40e/i40e_main.c
1520
!test_bit(I40E_FLAG_VF_VLAN_PRUNING_ENA, pf->flags));
drivers/net/ethernet/intel/i40e/i40e_main.c
15200
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
15201
test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15208
if (i40e_set_filter_control(&pf->hw, settings))
drivers/net/ethernet/intel/i40e/i40e_main.c
15216
static void i40e_print_features(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15218
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15219
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
15229
i += scnprintf(&buf[i], REMAIN(i), " VFs: %d", pf->num_req_vfs);
drivers/net/ethernet/intel/i40e/i40e_main.c
15232
pf->hw.func_caps.num_vsis, main_vsi->num_queue_pairs);
drivers/net/ethernet/intel/i40e/i40e_main.c
15233
if (test_bit(I40E_FLAG_RSS_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15235
if (test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15237
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15241
if (test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15245
if (test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15247
if (test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
15252
dev_info(&pf->pdev->dev, "%s\n", buf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15267
static void i40e_get_platform_mac_addr(struct pci_dev *pdev, struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15269
if (eth_platform_get_mac_address(&pdev->dev, pf->hw.mac.addr))
drivers/net/ethernet/intel/i40e/i40e_main.c
15270
i40e_get_mac_addr(&pf->hw, pf->hw.mac.addr);
drivers/net/ethernet/intel/i40e/i40e_main.c
15309
static bool i40e_check_recovery_mode(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15311
u32 val = rd32(&pf->hw, I40E_GL_FWSTS);
drivers/net/ethernet/intel/i40e/i40e_main.c
15314
dev_crit(&pf->pdev->dev, "Firmware recovery mode detected. Limiting functionality.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15315
dev_crit(&pf->pdev->dev, "Refer to the Intel(R) Ethernet Adapters and Devices User Guide for details on firmware recovery mode.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15316
set_bit(__I40E_RECOVERY_MODE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
15320
if (test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
15321
dev_info(&pf->pdev->dev, "Please do Power-On Reset to initialize adapter in normal mode with full functionality.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15347
static int i40e_pf_loop_reset(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15351
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
15361
pf->pfr_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
15363
dev_info(&pf->pdev->dev, "PF reset failed: %d\n", ret);
drivers/net/ethernet/intel/i40e/i40e_main.c
15379
static bool i40e_check_fw_empr(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15381
const u32 fw_sts = rd32(&pf->hw, I40E_GL_FWSTS) &
drivers/net/ethernet/intel/i40e/i40e_main.c
15398
static int i40e_handle_resets(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
15400
const int pfr = i40e_pf_loop_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15401
const bool is_empr = i40e_check_fw_empr(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15404
dev_crit(&pf->pdev->dev, "Entering recovery mode due to repeated FW resets. This may take several minutes. Refer to the Intel(R) Ethernet Adapters and Devices User Guide.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
15419
static int i40e_init_recovery_mode(struct i40e_pf *pf, struct i40e_hw *hw)
drivers/net/ethernet/intel/i40e/i40e_main.c
15425
pci_set_drvdata(pf->pdev, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15426
pci_save_state(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
15429
timer_setup(&pf->service_timer, i40e_service_timer, 0);
drivers/net/ethernet/intel/i40e/i40e_main.c
15430
pf->service_timer_period = HZ;
drivers/net/ethernet/intel/i40e/i40e_main.c
15432
INIT_WORK(&pf->service_task, i40e_service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
15433
clear_bit(__I40E_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
15435
err = i40e_init_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15444
if (pf->hw.func_caps.num_vsis < I40E_MIN_VSI_ALLOC)
drivers/net/ethernet/intel/i40e/i40e_main.c
15445
pf->num_alloc_vsi = I40E_MIN_VSI_ALLOC;
drivers/net/ethernet/intel/i40e/i40e_main.c
15447
pf->num_alloc_vsi = pf->hw.func_caps.num_vsis;
drivers/net/ethernet/intel/i40e/i40e_main.c
15450
pf->vsi = kzalloc_objs(struct i40e_vsi *, pf->num_alloc_vsi);
drivers/net/ethernet/intel/i40e/i40e_main.c
15451
if (!pf->vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15459
v_idx = i40e_vsi_mem_alloc(pf, I40E_VSI_MAIN);
drivers/net/ethernet/intel/i40e/i40e_main.c
15464
pf->lan_vsi = v_idx;
drivers/net/ethernet/intel/i40e/i40e_main.c
15465
vsi = pf->vsi[v_idx];
drivers/net/ethernet/intel/i40e/i40e_main.c
15478
i40e_dbg_pf_init(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15480
err = i40e_setup_misc_vector_for_recovery_mode(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15485
i40e_send_version(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15488
mod_timer(&pf->service_timer,
drivers/net/ethernet/intel/i40e/i40e_main.c
15489
round_jiffies(jiffies + pf->service_timer_period));
drivers/net/ethernet/intel/i40e/i40e_main.c
15494
i40e_reset_interrupt_capability(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15495
timer_shutdown_sync(&pf->service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
15498
pci_release_mem_regions(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
15499
pci_disable_device(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
15500
i40e_free_pf(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15514
struct i40e_pf *pf = i40e_hw_to_pf(hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
15516
hw->subsystem_device_id = pf->pdev->subsystem_device ?
drivers/net/ethernet/intel/i40e/i40e_main.c
15517
pf->pdev->subsystem_device :
drivers/net/ethernet/intel/i40e/i40e_main.c
15539
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
15577
pf = i40e_alloc_pf(&pdev->dev);
drivers/net/ethernet/intel/i40e/i40e_main.c
15578
if (!pf) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15582
pf->next_vsi = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
15583
pf->pdev = pdev;
drivers/net/ethernet/intel/i40e/i40e_main.c
15584
set_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
15586
hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
15588
pf->ioremap_len = min_t(int, pci_resource_len(pdev, 0),
drivers/net/ethernet/intel/i40e/i40e_main.c
15595
if (pf->ioremap_len < I40E_GLGEN_STAT_CLEAR) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15597
pf->ioremap_len);
drivers/net/ethernet/intel/i40e/i40e_main.c
15601
hw->hw_addr = ioremap(pci_resource_start(pdev, 0), pf->ioremap_len);
drivers/net/ethernet/intel/i40e/i40e_main.c
15606
pf->ioremap_len, err);
drivers/net/ethernet/intel/i40e/i40e_main.c
15625
INIT_LIST_HEAD(&pf->l3_flex_pit_list);
drivers/net/ethernet/intel/i40e/i40e_main.c
15626
INIT_LIST_HEAD(&pf->l4_flex_pit_list);
drivers/net/ethernet/intel/i40e/i40e_main.c
15627
INIT_LIST_HEAD(&pf->ddp_old_prof);
drivers/net/ethernet/intel/i40e/i40e_main.c
15635
pf->msg_enable = netif_msg_init(debug,
drivers/net/ethernet/intel/i40e/i40e_main.c
15640
pf->hw.debug_mask = debug;
drivers/net/ethernet/intel/i40e/i40e_main.c
15648
pf->corer_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
15663
err = i40e_handle_resets(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15667
i40e_check_recovery_mode(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15679
snprintf(pf->int_name, sizeof(pf->int_name) - 1,
drivers/net/ethernet/intel/i40e/i40e_main.c
15681
dev_driver_string(&pf->pdev->dev), dev_name(&pdev->dev));
drivers/net/ethernet/intel/i40e/i40e_main.c
15691
pf->hw.fc.requested_mode = I40E_FC_NONE;
drivers/net/ethernet/intel/i40e/i40e_main.c
15735
i40e_verify_eeprom(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15743
err = i40e_get_capabilities(pf, i40e_aqc_opc_list_func_capabilities);
drivers/net/ethernet/intel/i40e/i40e_main.c
15747
err = i40e_sw_init(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15753
if (test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
15754
return i40e_init_recovery_mode(pf, hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
15774
if (test_bit(I40E_HW_CAP_STOP_FW_LLDP, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15780
i40e_get_platform_mac_addr(pdev, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15791
set_bit(I40E_HW_CAP_PORT_ID_VALID, pf->hw.caps);
drivers/net/ethernet/intel/i40e/i40e_main.c
15793
i40e_ptp_alloc_pins(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15794
pci_set_drvdata(pdev, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15798
status = i40e_get_fw_lldp_status(&pf->hw, &lldp_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
15801
(clear_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags)) :
drivers/net/ethernet/intel/i40e/i40e_main.c
15802
(set_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags));
drivers/net/ethernet/intel/i40e/i40e_main.c
15804
test_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags) ?
drivers/net/ethernet/intel/i40e/i40e_main.c
15811
err = i40e_init_pf_dcb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15814
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15815
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15821
timer_setup(&pf->service_timer, i40e_service_timer, 0);
drivers/net/ethernet/intel/i40e/i40e_main.c
15822
pf->service_timer_period = HZ;
drivers/net/ethernet/intel/i40e/i40e_main.c
15824
INIT_WORK(&pf->service_task, i40e_service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
15825
clear_bit(__I40E_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
15830
pf->wol_en = false;
drivers/net/ethernet/intel/i40e/i40e_main.c
15832
pf->wol_en = true;
drivers/net/ethernet/intel/i40e/i40e_main.c
15833
device_set_wakeup_enable(&pf->pdev->dev, pf->wol_en);
drivers/net/ethernet/intel/i40e/i40e_main.c
15836
i40e_determine_queue_usage(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15837
err = i40e_init_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15847
pf->num_lan_msix = 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
15849
pf->udp_tunnel_nic.set_port = i40e_udp_tunnel_set_port;
drivers/net/ethernet/intel/i40e/i40e_main.c
15850
pf->udp_tunnel_nic.unset_port = i40e_udp_tunnel_unset_port;
drivers/net/ethernet/intel/i40e/i40e_main.c
15851
pf->udp_tunnel_nic.shared = &pf->udp_tunnel_shared;
drivers/net/ethernet/intel/i40e/i40e_main.c
15852
pf->udp_tunnel_nic.tables[0].n_entries = I40E_MAX_PF_UDP_OFFLOAD_PORTS;
drivers/net/ethernet/intel/i40e/i40e_main.c
15853
pf->udp_tunnel_nic.tables[0].tunnel_types = UDP_TUNNEL_TYPE_VXLAN |
drivers/net/ethernet/intel/i40e/i40e_main.c
15861
if (pf->hw.func_caps.num_vsis < I40E_MIN_VSI_ALLOC)
drivers/net/ethernet/intel/i40e/i40e_main.c
15862
pf->num_alloc_vsi = I40E_MIN_VSI_ALLOC;
drivers/net/ethernet/intel/i40e/i40e_main.c
15864
pf->num_alloc_vsi = pf->hw.func_caps.num_vsis;
drivers/net/ethernet/intel/i40e/i40e_main.c
15865
if (pf->num_alloc_vsi > UDP_TUNNEL_NIC_MAX_SHARING_DEVICES) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15866
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
15868
pf->num_alloc_vsi, UDP_TUNNEL_NIC_MAX_SHARING_DEVICES);
drivers/net/ethernet/intel/i40e/i40e_main.c
15869
pf->num_alloc_vsi = UDP_TUNNEL_NIC_MAX_SHARING_DEVICES;
drivers/net/ethernet/intel/i40e/i40e_main.c
15873
pf->vsi = kzalloc_objs(struct i40e_vsi *, pf->num_alloc_vsi);
drivers/net/ethernet/intel/i40e/i40e_main.c
15874
if (!pf->vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15881
if (test_bit(I40E_FLAG_SRIOV_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15882
test_bit(I40E_FLAG_MSIX_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15883
!test_bit(__I40E_BAD_EEPROM, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15885
set_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15888
err = i40e_setup_pf_switch(pf, false, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
15894
vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15898
vsi = i40e_find_vsi_by_type(pf, I40E_VSI_FDIR);
drivers/net/ethernet/intel/i40e/i40e_main.c
15905
err = i40e_aq_set_phy_int_mask(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
15910
dev_info(&pf->pdev->dev, "set phy mask fail, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
15911
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15914
ratelimit_state_init(&pf->mdd_message_rate_limit, 1 * HZ, 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
15927
if (test_bit(I40E_HW_CAP_RESTART_AUTONEG, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15929
err = i40e_aq_set_link_restart_an(&pf->hw, true, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
15931
dev_info(&pf->pdev->dev, "link restart failed, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
15933
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15939
clear_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
15946
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15947
err = i40e_setup_misc_vector(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15951
i40e_cloud_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15952
i40e_fdir_teardown(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15959
if (test_bit(I40E_FLAG_SRIOV_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15960
test_bit(I40E_FLAG_MSIX_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
15961
!test_bit(__I40E_BAD_EEPROM, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15971
err = i40e_alloc_vfs(pf, pci_num_vf(pdev));
drivers/net/ethernet/intel/i40e/i40e_main.c
15980
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15981
pf->iwarp_base_vector = i40e_get_lump(pf, pf->irq_pile,
drivers/net/ethernet/intel/i40e/i40e_main.c
15982
pf->num_iwarp_msix,
drivers/net/ethernet/intel/i40e/i40e_main.c
15984
if (pf->iwarp_base_vector < 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15987
pf->num_iwarp_msix, pf->iwarp_base_vector);
drivers/net/ethernet/intel/i40e/i40e_main.c
15988
clear_bit(I40E_FLAG_IWARP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
15992
i40e_dbg_pf_init(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15995
i40e_send_version(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
15998
mod_timer(&pf->service_timer,
drivers/net/ethernet/intel/i40e/i40e_main.c
15999
round_jiffies(jiffies + pf->service_timer_period));
drivers/net/ethernet/intel/i40e/i40e_main.c
160
struct i40e_pf *pf = i40e_hw_to_pf(hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
16002
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16003
err = i40e_lan_add_device(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16015
if (!test_bit(I40E_HW_CAP_NO_PCI_LINK_CHECK, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16022
pcie_capability_read_word(pf->pdev, PCI_EXP_LNKSTA,
drivers/net/ethernet/intel/i40e/i40e_main.c
16063
dev_dbg(&pf->pdev->dev, "get requested speeds ret = %pe last_status = %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
16064
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16065
pf->hw.phy.link_info.requested_speeds = abilities.link_speed;
drivers/net/ethernet/intel/i40e/i40e_main.c
16068
i40e_set_fec_in_flags(abilities.fec_cfg_curr_mod_ext_info, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
16073
dev_dbg(&pf->pdev->dev, "get supported phy types ret = %pe last_status = %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
16074
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16081
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16094
i40e_add_filter_to_drop_tx_flow_control_frames(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
16095
pf->main_vsi_seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
16097
if ((pf->hw.device_id == I40E_DEV_ID_10G_BASE_T) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
16098
(pf->hw.device_id == I40E_DEV_ID_10G_BASE_T4))
drivers/net/ethernet/intel/i40e/i40e_main.c
16099
set_bit(I40E_HW_CAP_PHY_CONTROLS_LEDS, pf->hw.caps);
drivers/net/ethernet/intel/i40e/i40e_main.c
1610
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
16100
if (pf->hw.device_id == I40E_DEV_ID_SFP_I_X722)
drivers/net/ethernet/intel/i40e/i40e_main.c
16101
set_bit(I40E_HW_CAP_CRT_RETIMER, pf->hw.caps);
drivers/net/ethernet/intel/i40e/i40e_main.c
16103
i40e_print_features(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16105
i40e_devlink_register(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16111
set_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16112
i40e_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16113
kfree(pf->vsi);
drivers/net/ethernet/intel/i40e/i40e_main.c
16115
i40e_reset_interrupt_capability(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16116
timer_shutdown_sync(&pf->service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
16121
kfree(pf->qp_pile);
drivers/net/ethernet/intel/i40e/i40e_main.c
16127
i40e_free_pf(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16147
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16148
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
16154
i40e_devlink_unregister(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16156
i40e_dbg_pf_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16158
i40e_ptp_stop(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16168
while (test_and_set_bit(__I40E_RESET_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16170
set_bit(__I40E_IN_REMOVE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16172
if (test_bit(I40E_FLAG_SRIOV_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16173
set_bit(__I40E_VF_RESETS_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16174
i40e_free_vfs(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16175
clear_bit(I40E_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
16178
set_bit(__I40E_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16179
set_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16180
if (pf->service_timer.function)
drivers/net/ethernet/intel/i40e/i40e_main.c
16181
timer_shutdown_sync(&pf->service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
16182
if (pf->service_task.func)
drivers/net/ethernet/intel/i40e/i40e_main.c
16183
cancel_work_sync(&pf->service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
16185
if (test_bit(__I40E_RECOVERY_MODE, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16186
struct i40e_vsi *vsi = pf->vsi[0];
drivers/net/ethernet/intel/i40e/i40e_main.c
16201
i40e_notify_client_of_netdev_close(pf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
16203
i40e_fdir_teardown(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16208
i40e_pf_for_each_veb(pf, i, veb)
drivers/net/ethernet/intel/i40e/i40e_main.c
16209
if (veb->uplink_seid == pf->mac_seid ||
drivers/net/ethernet/intel/i40e/i40e_main.c
1621
i40e_aq_remove_macvlan(&pf->hw, vsi->seid, &element, 1, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
16216
i40e_pf_for_each_vsi(pf, i, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16219
pf->vsi[i] = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
16222
i40e_cloud_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16225
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16226
ret_code = i40e_lan_del_device(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16243
if (test_bit(__I40E_RECOVERY_MODE, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
16244
!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
16245
free_irq(pf->pdev->irq, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16256
i40e_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16257
i40e_pf_for_each_vsi(pf, i, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16258
if (!test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16262
pf->vsi[i] = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
16266
i40e_pf_for_each_veb(pf, i, veb) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16268
pf->veb[i] = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
16271
kfree(pf->qp_pile);
drivers/net/ethernet/intel/i40e/i40e_main.c
16272
kfree(pf->vsi);
drivers/net/ethernet/intel/i40e/i40e_main.c
16275
i40e_free_pf(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16286
static void i40e_enable_mc_magic_wake(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
16288
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16289
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
1629
i40e_aq_remove_macvlan(&pf->hw, vsi->seid, &element, 1, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
16298
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
163
mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa,
drivers/net/ethernet/intel/i40e/i40e_main.c
16314
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
16324
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
16333
static int i40e_io_suspend(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
16335
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
16337
set_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16340
timer_delete_sync(&pf->service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
16341
cancel_work_sync(&pf->service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
16346
i40e_notify_client_of_netdev_close(pf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
16348
if (test_bit(I40E_HW_CAP_WOL_MC_MAGIC_PKT_WAKE, pf->hw.caps) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
16349
pf->wol_en)
drivers/net/ethernet/intel/i40e/i40e_main.c
16350
i40e_enable_mc_magic_wake(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16358
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16360
wr32(hw, I40E_PFPM_APM, (pf->wol_en ? I40E_PFPM_APM_APME_MASK : 0));
drivers/net/ethernet/intel/i40e/i40e_main.c
16361
wr32(hw, I40E_PFPM_WUFC, (pf->wol_en ? I40E_PFPM_WUFC_MAG_MASK : 0));
drivers/net/ethernet/intel/i40e/i40e_main.c
16368
i40e_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16380
static int i40e_io_resume(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
16382
struct device *dev = &pf->pdev->dev;
drivers/net/ethernet/intel/i40e/i40e_main.c
16393
err = i40e_restore_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16399
clear_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16400
i40e_reset_and_rebuild(pf, false, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
16405
clear_bit(__I40E_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16408
mod_timer(&pf->service_timer,
drivers/net/ethernet/intel/i40e/i40e_main.c
16409
round_jiffies(jiffies + pf->service_timer_period));
drivers/net/ethernet/intel/i40e/i40e_main.c
16426
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16430
if (!pf) {
drivers/net/ethernet/intel/i40e/i40e_main.c
16437
if (!test_bit(__I40E_SUSPENDED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16438
i40e_io_suspend(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16455
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16470
reg = rd32(&pf->hw, I40E_GLGEN_RTRIG);
drivers/net/ethernet/intel/i40e/i40e_main.c
16486
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16488
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16497
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16499
if (test_bit(__I40E_IN_REMOVE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16502
i40e_reset_and_rebuild(pf, false, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
16517
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16520
if (test_bit(__I40E_SUSPENDED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16523
i40e_io_resume(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16532
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16533
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
16535
set_bit(__I40E_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16536
set_bit(__I40E_DOWN, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
16538
timer_delete_sync(&pf->service_timer);
drivers/net/ethernet/intel/i40e/i40e_main.c
16539
cancel_work_sync(&pf->service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
16540
i40e_cloud_filter_exit(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16541
i40e_fdir_teardown(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16546
i40e_notify_client_of_netdev_close(pf, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
16548
if (test_bit(I40E_HW_CAP_WOL_MC_MAGIC_PKT_WAKE, pf->hw.caps) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
16549
pf->wol_en)
drivers/net/ethernet/intel/i40e/i40e_main.c
16550
i40e_enable_mc_magic_wake(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16552
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16555
(pf->wol_en ? I40E_PFPM_APM_APME_MASK : 0));
drivers/net/ethernet/intel/i40e/i40e_main.c
16557
(pf->wol_en ? I40E_PFPM_WUFC_MAG_MASK : 0));
drivers/net/ethernet/intel/i40e/i40e_main.c
16560
if (test_bit(__I40E_RECOVERY_MODE, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
16561
!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
16562
free_irq(pf->pdev->irq, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16569
i40e_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16573
pci_wake_from_d3(pdev, pf->wol_en);
drivers/net/ethernet/intel/i40e/i40e_main.c
16584
struct i40e_pf *pf = dev_get_drvdata(dev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16587
if (test_and_set_bit(__I40E_SUSPENDED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16589
return i40e_io_suspend(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
16598
struct i40e_pf *pf = dev_get_drvdata(dev);
drivers/net/ethernet/intel/i40e/i40e_main.c
16601
if (!test_bit(__I40E_SUSPENDED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
16603
return i40e_io_resume(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
178
struct i40e_pf *pf = i40e_hw_to_pf(hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
180
dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa);
drivers/net/ethernet/intel/i40e/i40e_main.c
1805
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
1806
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
1812
if (test_bit(__I40E_DOWN, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
1813
test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
1848
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
1862
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
1863
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
1871
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1883
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
1899
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
1904
if (!test_bit(I40E_HW_CAP_RSS_AQ, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
1907
vsi->rss_size = min_t(int, pf->alloc_rss_size,
drivers/net/ethernet/intel/i40e/i40e_main.c
1921
i40e_fill_rss_lut(pf, lut, vsi->rss_table_size, vsi->rss_size);
drivers/net/ethernet/intel/i40e/i40e_main.c
2032
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
2059
else if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
2060
vsi->num_queue_pairs = pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
2079
dev_warn(&pf->pdev->dev, "DCB is enabled but no TC enabled, forcing TC0\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
2084
i40e_pf_get_max_q_per_tc(pf));
drivers/net/ethernet/intel/i40e/i40e_main.c
2091
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
2092
num_tc_qps = min_t(int, num_tc_qps, pf->num_lan_msix);
drivers/net/ethernet/intel/i40e/i40e_main.c
2104
pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
2106
pf->flags)) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
2108
qcount = min_t(int, pf->alloc_rss_size,
drivers/net/ethernet/intel/i40e/i40e_main.c
230
static int i40e_get_lump(struct i40e_pf *pf, struct i40e_lump_tracking *pile,
drivers/net/ethernet/intel/i40e/i40e_main.c
237
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
246
if (pile == pf->qp_pile && pf->vsi[id]->type == I40E_VSI_FDIR) {
drivers/net/ethernet/intel/i40e/i40e_main.c
248
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2482
static int i40e_set_promiscuous(struct i40e_pf *pf, bool promisc)
drivers/net/ethernet/intel/i40e/i40e_main.c
2484
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
2485
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
2489
i40e_pf_get_main_veb(pf) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
2490
!test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
2505
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2517
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2527
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2535
pf->cur_promisc = promisc;
drivers/net/ethernet/intel/i40e/i40e_main.c
2561
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
2576
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
2634
else if (pf->vf)
drivers/net/ethernet/intel/i40e/i40e_main.c
2637
vlan_filters, pf->vf[vsi->vf_id].trusted);
drivers/net/ethernet/intel/i40e/i40e_main.c
2800
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2808
if (vsi->type == I40E_VSI_SRIOV && pf->vf &&
drivers/net/ethernet/intel/i40e/i40e_main.c
2809
!pf->vf[vsi->vf_id].trusted) {
drivers/net/ethernet/intel/i40e/i40e_main.c
2834
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2840
dev_info(&pf->pdev->dev, "%s allmulti mode.\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
2850
aq_ret = i40e_set_promiscuous(pf, cur_promisc);
drivers/net/ethernet/intel/i40e/i40e_main.c
2854
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
2887
static void i40e_sync_filters_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
2892
if (!pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
2894
if (!test_and_clear_bit(__I40E_MACVLAN_SYNC_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
2896
if (test_bit(__I40E_VF_DISABLE, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
2897
set_bit(__I40E_MACVLAN_SYNC_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
2901
i40e_pf_for_each_vsi(pf, v, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
2909
pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
2959
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
2974
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
2975
set_bit(__I40E_CLIENT_L2_CHANGE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
319
struct i40e_vsi *i40e_find_vsi_from_id(struct i40e_pf *pf, u16 id)
drivers/net/ethernet/intel/i40e/i40e_main.c
324
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
337
void i40e_service_event_schedule(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
339
if ((!test_bit(__I40E_DOWN, pf->state) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
340
!test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state)) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
341
test_bit(__I40E_RECOVERY_MODE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
342
queue_work(i40e_wq, &pf->service_task);
drivers/net/ethernet/intel/i40e/i40e_main.c
358
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
36
static void i40e_handle_reset_warning(struct i40e_pf *pf, bool lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
363
pf->tx_timeout_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
376
if (time_after(jiffies, (pf->tx_timeout_last_recovery + HZ*20)))
drivers/net/ethernet/intel/i40e/i40e_main.c
377
pf->tx_timeout_recovery_level = 1; /* reset after some time */
drivers/net/ethernet/intel/i40e/i40e_main.c
379
(pf->tx_timeout_last_recovery + netdev->watchdog_timeo)))
drivers/net/ethernet/intel/i40e/i40e_main.c
3817
static void i40e_reset_fdir_filter_cnt(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
3819
pf->fd_tcp4_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3820
pf->fd_udp4_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3821
pf->fd_sctp4_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3822
pf->fd_ip4_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3823
pf->fd_tcp6_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3824
pf->fd_udp6_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3825
pf->fd_sctp6_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
3826
pf->fd_ip6_filter_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
383
if (test_and_set_bit(__I40E_TIMEOUT_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
3839
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
3842
if (!test_bit(I40E_FLAG_FD_SB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
3846
i40e_reset_fdir_filter_cnt(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
3849
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_main.c
3879
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
3880
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
389
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
39
static int i40e_setup_pf_switch(struct i40e_pf *pf, bool reinit, bool lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
390
val = rd32(&pf->hw,
drivers/net/ethernet/intel/i40e/i40e_main.c
394
val = rd32(&pf->hw, I40E_PFINT_DYN_CTL0);
drivers/net/ethernet/intel/i40e/i40e_main.c
3968
static void i40e_enable_misc_int_causes(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
3970
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
3986
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
3989
if (test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
40
static int i40e_setup_misc_vector(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4010
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4011
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
402
pf->tx_timeout_last_recovery = jiffies;
drivers/net/ethernet/intel/i40e/i40e_main.c
4023
i40e_enable_misc_int_causes(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
404
pf->tx_timeout_recovery_level, txqueue);
drivers/net/ethernet/intel/i40e/i40e_main.c
4049
void i40e_irq_dynamic_disable_icr0(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
4051
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
406
switch (pf->tx_timeout_recovery_level) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4062
void i40e_irq_dynamic_enable_icr0(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
4064
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
408
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
41
static void i40e_determine_queue_usage(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
411
set_bit(__I40E_CORE_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4129
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
414
set_bit(__I40E_GLOBAL_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4140
irq_num = pf->msix_entries[base + vector].vector;
drivers/net/ethernet/intel/i40e/i40e_main.c
4162
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
418
set_bit(__I40E_DOWN_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4188
irq_num = pf->msix_entries[base + vector].vector;
drivers/net/ethernet/intel/i40e/i40e_main.c
42
static int i40e_setup_pf_filter_control(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4202
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4203
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
4225
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
423
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4232
synchronize_irq(pf->msix_entries[i + base].vector);
drivers/net/ethernet/intel/i40e/i40e_main.c
4238
synchronize_irq(pf->pdev->irq);
drivers/net/ethernet/intel/i40e/i40e_main.c
424
pf->tx_timeout_recovery_level++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4248
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4251
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4255
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4258
i40e_flush(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
4266
static void i40e_free_misc_vector(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
4269
wr32(&pf->hw, I40E_PFINT_ICR0_ENA, 0);
drivers/net/ethernet/intel/i40e/i40e_main.c
4270
i40e_flush(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
4272
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags) && pf->msix_entries) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4273
free_irq(pf->msix_entries[0].vector, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4274
clear_bit(__I40E_MISC_IRQ_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4289
struct i40e_pf *pf = (struct i40e_pf *)data;
drivers/net/ethernet/intel/i40e/i40e_main.c
4290
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
43
static void i40e_prep_for_reset(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4305
pf->sw_int_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4307
if (test_bit(I40E_FLAG_IWARP_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
4310
dev_dbg(&pf->pdev->dev, "cleared PE_CRITERR\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
4311
set_bit(__I40E_CORE_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4316
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4325
if (!test_bit(__I40E_DOWN, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
4331
set_bit(__I40E_ADMINQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4332
i40e_debug(&pf->hw, I40E_DEBUG_NVM, "AdminQ event\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
4337
set_bit(__I40E_MDD_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4342
if (test_bit(__I40E_VF_RESETS_DISABLED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4349
set_bit(__I40E_VFLR_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4354
if (!test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
4355
set_bit(__I40E_RESET_INTR_RECEIVED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4360
pf->corer_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4362
pf->globr_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4364
pf->empr_count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4365
set_bit(__I40E_EMP_RESET_INTR_RECEIVED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4371
dev_info(&pf->pdev->dev, "HMC error interrupt\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
4372
dev_info(&pf->pdev->dev, "HMC error info 0x%x, HMC error data 0x%x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
4381
schedule_work(&pf->ptp_extts0_work);
drivers/net/ethernet/intel/i40e/i40e_main.c
4384
i40e_ptp_tx_hwtstamp(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4395
dev_info(&pf->pdev->dev, "unhandled interrupt icr0=0x%08x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
44
static void i40e_reset_and_rebuild(struct i40e_pf *pf, bool reinit,
drivers/net/ethernet/intel/i40e/i40e_main.c
4400
dev_info(&pf->pdev->dev, "device will be reset\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
4401
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
4402
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4411
if (!test_bit(__I40E_DOWN, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
4412
test_bit(__I40E_RECOVERY_MODE, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4413
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4414
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
46
static int i40e_reset(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4609
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4612
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
4614
else if (test_bit(I40E_FLAG_MSI_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
4615
err = request_irq(pf->pdev->irq, i40e_intr, 0,
drivers/net/ethernet/intel/i40e/i40e_main.c
4616
pf->int_name, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4618
err = request_irq(pf->pdev->irq, i40e_intr, IRQF_SHARED,
drivers/net/ethernet/intel/i40e/i40e_main.c
4619
pf->int_name, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4622
dev_info(&pf->pdev->dev, "request_irq failed, Error %d\n", err);
drivers/net/ethernet/intel/i40e/i40e_main.c
4639
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4646
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4650
i40e_intr(pf->pdev->irq, netdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
4668
static int i40e_pf_txq_wait(struct i40e_pf *pf, int pf_q, bool enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
4674
tx_reg = rd32(&pf->hw, I40E_QTX_ENA(pf_q));
drivers/net/ethernet/intel/i40e/i40e_main.c
4696
static void i40e_control_tx_q(struct i40e_pf *pf, int pf_q, bool enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
4698
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
47
static void i40e_rebuild(struct i40e_pf *pf, bool reinit, bool lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
4703
i40e_pre_tx_queue_cfg(&pf->hw, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
4738
int i40e_control_wait_tx_q(int seid, struct i40e_pf *pf, int pf_q,
drivers/net/ethernet/intel/i40e/i40e_main.c
4743
i40e_control_tx_q(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
4746
ret = i40e_pf_txq_wait(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
4748
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
4763
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4768
ret = i40e_control_wait_tx_q(vsi->seid, pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
4777
ret = i40e_control_wait_tx_q(vsi->seid, pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
4797
static int i40e_pf_rxq_wait(struct i40e_pf *pf, int pf_q, bool enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
48
static int i40e_setup_misc_vector_for_recovery_mode(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4803
rx_reg = rd32(&pf->hw, I40E_QRX_ENA(pf_q));
drivers/net/ethernet/intel/i40e/i40e_main.c
4825
static void i40e_control_rx_q(struct i40e_pf *pf, int pf_q, bool enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
4827
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
4862
int i40e_control_wait_rx_q(struct i40e_pf *pf, int pf_q, bool enable)
drivers/net/ethernet/intel/i40e/i40e_main.c
4866
i40e_control_rx_q(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
4869
ret = i40e_pf_rxq_wait(pf, pf_q, enable);
drivers/net/ethernet/intel/i40e/i40e_main.c
4882
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4887
ret = i40e_control_wait_rx_q(pf, pf_q, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
4889
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
49
static int i40e_restore_interrupt_scheme(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
4924
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4934
i40e_pre_tx_queue_cfg(&pf->hw, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
4938
i40e_control_rx_q(pf, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
4942
wr32(&pf->hw, I40E_QTX_ENA(pf_q), 0);
drivers/net/ethernet/intel/i40e/i40e_main.c
4960
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4965
i40e_control_tx_q(pf, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
4966
i40e_control_rx_q(pf, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
4976
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
4977
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
4982
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4995
irq_num = pf->msix_entries[vector].vector;
drivers/net/ethernet/intel/i40e/i40e_main.c
50
static bool i40e_check_recovery_mode(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5055
free_irq(pf->pdev->irq, pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
51
static int i40e_init_recovery_mode(struct i40e_pf *pf, struct i40e_hw *hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
5140
static void i40e_reset_interrupt_capability(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5143
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5144
pci_disable_msix(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
5145
kfree(pf->msix_entries);
drivers/net/ethernet/intel/i40e/i40e_main.c
5146
pf->msix_entries = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
5147
kfree(pf->irq_pile);
drivers/net/ethernet/intel/i40e/i40e_main.c
5148
pf->irq_pile = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
5149
} else if (test_bit(I40E_FLAG_MSI_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5150
pci_disable_msi(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_main.c
5152
clear_bit(I40E_FLAG_MSI_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
5153
clear_bit(I40E_FLAG_MSIX_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
5163
static void i40e_clear_interrupt_scheme(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5168
if (test_bit(__I40E_MISC_IRQ_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
5169
i40e_free_misc_vector(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5171
i40e_put_lump(pf->irq_pile, pf->iwarp_base_vector,
drivers/net/ethernet/intel/i40e/i40e_main.c
5174
i40e_put_lump(pf->irq_pile, 0, I40E_PILE_VALID_BIT-1);
drivers/net/ethernet/intel/i40e/i40e_main.c
5176
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
5179
i40e_reset_interrupt_capability(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
52
static void i40e_fdir_sb_setup(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5226
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5233
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
5234
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
5235
set_bit(__I40E_CLIENT_RESET, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
5273
static void i40e_pf_quiesce_all_vsi(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5278
i40e_pf_for_each_vsi(pf, v, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
5286
static void i40e_pf_unquiesce_all_vsi(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5291
i40e_pf_for_each_vsi(pf, v, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
5303
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5309
ret = i40e_pf_txq_wait(pf, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
5311
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5321
ret = i40e_pf_txq_wait(pf, pf_q + vsi->alloc_queue_pairs,
drivers/net/ethernet/intel/i40e/i40e_main.c
5324
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5331
ret = i40e_pf_rxq_wait(pf, pf_q, false);
drivers/net/ethernet/intel/i40e/i40e_main.c
5333
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5351
static int i40e_pf_wait_queues_disabled(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5356
i40e_pf_for_each_vsi(pf, v, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5374
static u8 i40e_get_iscsi_tc_map(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5377
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
54
static int i40e_get_capabilities(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
5464
static u8 i40e_mqprio_get_enabled_tc(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5466
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5481
static u8 i40e_pf_get_num_tc(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5486
if (i40e_is_tc_mqprio_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5487
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5493
if (!test_bit(I40E_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
5497
if (!test_bit(I40E_FLAG_MFP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
5498
return i40e_dcb_get_num_tc(&pf->hw.local_dcbx_config);
drivers/net/ethernet/intel/i40e/i40e_main.c
5501
if (pf->hw.func_caps.iscsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
5502
enabled_tc = i40e_get_iscsi_tc_map(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5519
static u8 i40e_pf_get_tc_map(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5521
if (i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
5522
return i40e_mqprio_get_enabled_tc(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5527
if (!test_bit(I40E_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
5531
if (!test_bit(I40E_FLAG_MFP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
5532
return i40e_dcb_get_enabled_tc(&pf->hw.local_dcbx_config);
drivers/net/ethernet/intel/i40e/i40e_main.c
5535
if (pf->hw.func_caps.iscsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
5536
return i40e_get_iscsi_tc_map(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5551
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5552
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
5560
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5563
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5571
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5574
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5579
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
56
static bool i40e_is_total_port_shutdown_enabled(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
5613
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5618
if (i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
562
void i40e_pf_reset_stats(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
5620
if (!vsi->mqprio_qopt.qopt.hw && !test_bit(I40E_FLAG_DCB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5623
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5633
ret = i40e_aq_config_vsi_tc_bw(&pf->hw, vsi->seid, &bw_data, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
5635
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5637
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
5656
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5657
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
567
memset(&pf->stats, 0, sizeof(pf->stats));
drivers/net/ethernet/intel/i40e/i40e_main.c
568
memset(&pf->stats_offsets, 0, sizeof(pf->stats_offsets));
drivers/net/ethernet/intel/i40e/i40e_main.c
569
pf->stat_offsets_loaded = false;
drivers/net/ethernet/intel/i40e/i40e_main.c
5690
if (i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
571
i40e_pf_for_each_veb(pf, i, veb) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5730
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
5736
pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5737
hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
5750
vsi->rss_size = min_t(int, pf->alloc_rss_size,
drivers/net/ethernet/intel/i40e/i40e_main.c
5754
dev_info(&pf->pdev->dev, "Failed to reconfig rss for num_queues\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
5762
dev_info(&pf->pdev->dev, "Update vsi config failed, err %pe aq_err %s\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
578
pf->hw_csum_rx_error = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
5790
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5791
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
5811
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5817
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5830
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5838
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5851
if (i40e_is_tc_mqprio_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5884
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5897
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
5939
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5941
switch (pf->hw.phy.link_info.link_speed) {
drivers/net/ethernet/intel/i40e/i40e_main.c
5987
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
5994
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6000
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6008
ret = i40e_aq_config_vsi_bw_limit(&pf->hw, seid, credits,
drivers/net/ethernet/intel/i40e/i40e_main.c
6011
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6014
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6029
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
6073
&pf->cloud_filter_list, cloud_node) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6085
last_aq_status = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_main.c
6087
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6139
static int i40e_validate_num_queues(struct i40e_pf *pf, int num_queues,
drivers/net/ethernet/intel/i40e/i40e_main.c
6150
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6156
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6171
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6191
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
6193
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6210
i40e_fill_rss_lut(pf, lut, vsi->rss_table_size, local_rss_size);
drivers/net/ethernet/intel/i40e/i40e_main.c
6222
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6247
static void i40e_channel_setup_queue_map(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
6258
qcount = min_t(int, ch->num_queue_pairs, pf->num_lan_msix);
drivers/net/ethernet/intel/i40e/i40e_main.c
6286
static int i40e_add_channel(struct i40e_pf *pf, u16 uplink_seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
6289
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6295
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6308
if (test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6316
i40e_channel_setup_queue_map(pf, &ctxt, ch);
drivers/net/ethernet/intel/i40e/i40e_main.c
6321
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6324
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6385
static int i40e_channel_config_tx_ring(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
6439
static inline int i40e_setup_hw_channel(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
6451
ret = i40e_add_channel(pf, uplink_seid, ch);
drivers/net/ethernet/intel/i40e/i40e_main.c
6453
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6463
ret = i40e_channel_config_tx_ring(pf, vsi, ch);
drivers/net/ethernet/intel/i40e/i40e_main.c
6465
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6473
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6490
static bool i40e_setup_channel(struct i40e_pf *pf, struct i40e_vsi *vsi,
drivers/net/ethernet/intel/i40e/i40e_main.c
6501
dev_err(&pf->pdev->dev, "unsupported parent vsi type(%d)\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
6507
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6511
ret = i40e_setup_hw_channel(pf, vsi, ch, seid, vsi_type);
drivers/net/ethernet/intel/i40e/i40e_main.c
6513
dev_err(&pf->pdev->dev, "failed to setup hw_channel\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
6530
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
6531
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6534
ret = i40e_get_capabilities(pf, i40e_aqc_opc_list_dev_capabilities);
drivers/net/ethernet/intel/i40e/i40e_main.c
6547
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6564
ret = i40e_aq_set_switch_config(hw, pf->last_sw_conf_flags,
drivers/net/ethernet/intel/i40e/i40e_main.c
6565
pf->last_sw_conf_valid_flags,
drivers/net/ethernet/intel/i40e/i40e_main.c
6568
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6586
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
6594
dev_err(&pf->pdev->dev, "Invalid num_queues requested: %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
6600
err = i40e_validate_num_queues(pf, ch->num_queue_pairs, vsi,
drivers/net/ethernet/intel/i40e/i40e_main.c
6603
dev_info(&pf->pdev->dev, "Failed to validate num_queues (%d)\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
6612
if (!test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6613
set_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
6616
if (i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
6617
i40e_do_reset(pf, I40E_PF_RESET_FLAG, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
6619
i40e_do_reset_safe(pf, I40E_PF_RESET_FLAG);
drivers/net/ethernet/intel/i40e/i40e_main.c
6630
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6640
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6647
if (!i40e_setup_channel(pf, vsi, ch)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6648
dev_info(&pf->pdev->dev, "Failed to setup channel\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
6652
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6664
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6747
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
6764
ret = i40e_aq_config_switch_comp_bw_config(&pf->hw, veb->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
6767
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6769
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6776
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6778
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6794
static void i40e_dcb_reconfigure(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
6803
tc_map = i40e_pf_get_tc_map(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6807
i40e_pf_for_each_veb(pf, v, veb) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6810
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6818
i40e_pf_for_each_vsi(pf, v, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6823
tc_map = i40e_pf_get_tc_map(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6829
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6849
static int i40e_resume_port_tx(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
6851
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6856
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6859
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6861
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
6862
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6874
static int i40e_suspend_port_tx(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
6876
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6879
ret = i40e_aq_suspend_port_tx(hw, pf->mac_seid, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
6881
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6883
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6885
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
6886
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6900
static int i40e_hw_set_dcb_config(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
6903
struct i40e_dcbx_config *old_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_main.c
6908
dev_dbg(&pf->pdev->dev, "No Change in DCB Config required.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
6913
i40e_pf_quiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6918
ret = i40e_set_dcb_config(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
6920
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
6922
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6927
i40e_dcb_reconfigure(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6930
if (!test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
6932
ret = i40e_resume_port_tx(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6936
i40e_pf_unquiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
6950
int i40e_hw_dcb_config(struct i40e_pf *pf, struct i40e_dcbx_config *new_cfg)
drivers/net/ethernet/intel/i40e/i40e_main.c
6958
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
6967
dev_dbg(&pf->pdev->dev, "Configuring DCB registers directly\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7009
need_reconfig = i40e_dcb_need_reconfig(pf, old_cfg, new_cfg);
drivers/net/ethernet/intel/i40e/i40e_main.c
7017
set_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7019
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7021
set_bit(__I40E_PORT_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7023
i40e_pf_quiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7024
ret = i40e_suspend_port_tx(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7033
(hw, pf->mac_seid, &ets_data,
drivers/net/ethernet/intel/i40e/i40e_main.c
7036
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7038
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7058
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7067
i40e_dcb_hw_rx_pb_config(hw, &pf->pb_cfg, &pb_cfg);
drivers/net/ethernet/intel/i40e/i40e_main.c
7070
pf->pb_cfg = pb_cfg;
drivers/net/ethernet/intel/i40e/i40e_main.c
7073
ret = i40e_aq_dcb_updated(&pf->hw, NULL);
drivers/net/ethernet/intel/i40e/i40e_main.c
7075
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7077
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7085
i40e_dcb_reconfigure(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7089
ret = i40e_resume_port_tx(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7091
clear_bit(__I40E_PORT_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7097
ret = i40e_pf_wait_queues_disabled(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7100
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7101
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7104
i40e_pf_unquiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7105
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7106
set_bit(__I40E_CLIENT_L2_CHANGE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7109
if (test_bit(I40E_HW_CAP_USE_SET_LLDP_MIB, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_main.c
7110
ret = i40e_hw_set_dcb_config(pf, new_cfg);
drivers/net/ethernet/intel/i40e/i40e_main.c
7123
int i40e_dcb_sw_default_config(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
7125
struct i40e_dcbx_config *dcb_cfg = &pf->hw.local_dcbx_config;
drivers/net/ethernet/intel/i40e/i40e_main.c
7127
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7130
if (test_bit(I40E_HW_CAP_USE_SET_LLDP_MIB, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7132
memset(&pf->tmp_cfg, 0, sizeof(struct i40e_dcbx_config));
drivers/net/ethernet/intel/i40e/i40e_main.c
7133
pf->tmp_cfg.etscfg.willing = I40E_IEEE_DEFAULT_ETS_WILLING;
drivers/net/ethernet/intel/i40e/i40e_main.c
7134
pf->tmp_cfg.etscfg.maxtcs = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
7135
pf->tmp_cfg.etscfg.tcbwtable[0] = I40E_IEEE_DEFAULT_ETS_TCBW;
drivers/net/ethernet/intel/i40e/i40e_main.c
7136
pf->tmp_cfg.etscfg.tsatable[0] = I40E_IEEE_TSA_ETS;
drivers/net/ethernet/intel/i40e/i40e_main.c
7137
pf->tmp_cfg.pfc.willing = I40E_IEEE_DEFAULT_PFC_WILLING;
drivers/net/ethernet/intel/i40e/i40e_main.c
7138
pf->tmp_cfg.pfc.pfccap = I40E_MAX_TRAFFIC_CLASS;
drivers/net/ethernet/intel/i40e/i40e_main.c
7140
pf->tmp_cfg.numapps = I40E_IEEE_DEFAULT_NUM_APPS;
drivers/net/ethernet/intel/i40e/i40e_main.c
7141
pf->tmp_cfg.app[0].selector = I40E_APP_SEL_ETHTYPE;
drivers/net/ethernet/intel/i40e/i40e_main.c
7142
pf->tmp_cfg.app[0].priority = I40E_IEEE_DEFAULT_APP_PRIO;
drivers/net/ethernet/intel/i40e/i40e_main.c
7143
pf->tmp_cfg.app[0].protocolid = I40E_APP_PROTOID_FCOE;
drivers/net/ethernet/intel/i40e/i40e_main.c
7145
return i40e_hw_set_dcb_config(pf, &pf->tmp_cfg);
drivers/net/ethernet/intel/i40e/i40e_main.c
7155
(hw, pf->mac_seid, &ets_data,
drivers/net/ethernet/intel/i40e/i40e_main.c
7158
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7160
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7182
static int i40e_init_pf_dcb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
7184
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7190
if (test_bit(I40E_HW_CAP_NO_DCB_SUPPORT, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7191
dev_info(&pf->pdev->dev, "DCB is not supported.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7195
if (test_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7196
dev_info(&pf->pdev->dev, "FW LLDP is disabled, attempting SW DCB\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7197
err = i40e_dcb_sw_default_config(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7199
dev_info(&pf->pdev->dev, "Could not initialize SW DCB\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7202
dev_info(&pf->pdev->dev, "SW DCB initialization succeeded.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7203
pf->dcbx_cap = DCB_CAP_DCBX_HOST |
drivers/net/ethernet/intel/i40e/i40e_main.c
7206
set_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7207
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7215
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7219
pf->dcbx_cap = DCB_CAP_DCBX_LLD_MANAGED |
drivers/net/ethernet/intel/i40e/i40e_main.c
7222
set_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7227
set_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7229
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7230
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7233
} else if (pf->hw.aq.asq_last_status == LIBIE_AQ_RC_EPERM) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7234
dev_info(&pf->pdev->dev, "FW LLDP disabled for this PF.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
7235
set_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
7237
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7239
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
724
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
725
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7275
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
7283
new_speed = pf->hw.phy.link_info.link_speed;
drivers/net/ethernet/intel/i40e/i40e_main.c
7299
if (pf->hw.func_caps.npar_enable &&
drivers/net/ethernet/intel/i40e/i40e_main.c
7300
(pf->hw.phy.link_info.link_speed == I40E_LINK_SPEED_1GB ||
drivers/net/ethernet/intel/i40e/i40e_main.c
7301
pf->hw.phy.link_info.link_speed == I40E_LINK_SPEED_100MB))
drivers/net/ethernet/intel/i40e/i40e_main.c
7305
switch (pf->hw.phy.link_info.link_speed) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7334
switch (pf->hw.fc.current_mode) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7349
if (pf->hw.phy.link_info.link_speed == I40E_LINK_SPEED_25GB) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7354
if (pf->hw.phy.link_info.an_info & I40E_AQ_AN_COMPLETED)
drivers/net/ethernet/intel/i40e/i40e_main.c
7357
if (pf->hw.phy.link_info.fec_info &
drivers/net/ethernet/intel/i40e/i40e_main.c
7360
else if (pf->hw.phy.link_info.fec_info &
drivers/net/ethernet/intel/i40e/i40e_main.c
7378
} else if (pf->hw.device_id == I40E_DEV_ID_KX_X722) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7383
if (pf->hw.phy.link_info.an_info & I40E_AQ_AN_COMPLETED)
drivers/net/ethernet/intel/i40e/i40e_main.c
7386
if (pf->hw.phy.link_info.fec_info &
drivers/net/ethernet/intel/i40e/i40e_main.c
7390
if (pf->hw.phy.link_info.req_fec_info &
drivers/net/ethernet/intel/i40e/i40e_main.c
7409
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
7412
if (test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
7426
if ((pf->hw.phy.link_info.link_info & I40E_AQ_LINK_UP) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
7436
pf->fd_add_err = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
7437
pf->fd_atr_cnt = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
7444
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7445
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
7459
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
7461
while (test_and_set_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
7466
clear_bit(__I40E_CONFIG_BUSY, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
7474
static int i40e_force_link_state(struct i40e_pf *pf, bool is_up)
drivers/net/ethernet/intel/i40e/i40e_main.c
7479
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7493
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7504
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7514
if (test_bit(I40E_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
7530
if (test_bit(I40E_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7548
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7550
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7835
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
7836
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
786
struct i40e_pf *pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
787
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7888
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7907
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
7908
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
7944
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7960
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
7979
if (!i40e_setup_channel(pf, vsi, ch)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7993
dev_info(&pf->pdev->dev, "Failed to setup macvlans\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
8009
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8013
if (test_bit(I40E_FLAG_DCB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8017
if (i40e_is_tc_mqprio_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8021
if (pf->num_lan_msix < I40E_MIN_MACVLAN_VECTORS) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8040
vectors = pf->num_lan_msix;
drivers/net/ethernet/intel/i40e/i40e_main.c
8117
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8118
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
8154
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8155
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
8175
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8194
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8207
clear_bit(I40E_FLAG_TC_MQPRIO_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8213
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8220
clear_bit(I40E_FLAG_TC_MQPRIO_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8223
if (!test_bit(I40E_FLAG_DCB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8230
if (num_tc > i40e_pf_get_num_tc(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8237
if (test_bit(I40E_FLAG_DCB_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8242
if (!test_bit(I40E_FLAG_MSIX_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
8249
set_bit(I40E_FLAG_TC_MQPRIO_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8250
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8269
if (!hw && !i40e_is_tc_mqprio_enabled(pf))
drivers/net/ethernet/intel/i40e/i40e_main.c
8293
if (i40e_is_tc_mqprio_enabled(pf)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8391
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8430
ret = i40e_aq_add_cloud_filters(&pf->hw, filter->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
8433
ret = i40e_aq_rem_cloud_filters(&pf->hw, filter->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
8436
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8439
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
8441
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8461
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8519
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8532
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8538
ret = i40e_aq_add_cloud_filters_bb(&pf->hw, filter->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
8541
ret = i40e_aq_rem_cloud_filters_bb(&pf->hw, filter->seid,
drivers/net/ethernet/intel/i40e/i40e_main.c
8546
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8548
add ? "add" : "delete", ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
8550
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8571
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8583
dev_err(&pf->pdev->dev, "Unsupported key used: 0x%llx\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8623
dev_err(&pf->pdev->dev, "Bad ether dest mask %pM\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8633
dev_err(&pf->pdev->dev, "Bad ether src mask %pM\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8651
dev_err(&pf->pdev->dev, "Bad vlan mask 0x%04x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8679
dev_err(&pf->pdev->dev, "Bad ip dst mask %pI4b\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8689
dev_err(&pf->pdev->dev, "Bad ip src mask %pI4b\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8696
dev_err(&pf->pdev->dev, "Tenant id not allowed for ip filter\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
8713
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
872
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8735
dev_err(&pf->pdev->dev, "Bad src port mask 0x%04x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8745
dev_err(&pf->pdev->dev, "Bad dst port mask 0x%04x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8759
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8814
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8823
dev_err(&pf->pdev->dev, "Unable to add filter because of invalid destination");
drivers/net/ethernet/intel/i40e/i40e_main.c
8827
if (test_bit(__I40E_RESET_RECOVERY_PENDING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
8828
test_bit(__I40E_RESET_INTR_RECEIVED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
8831
if (pf->fdir_pf_active_filters ||
drivers/net/ethernet/intel/i40e/i40e_main.c
8832
(!hlist_empty(&pf->fdir_filter_list))) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8866
dev_err(&pf->pdev->dev, "Failed to add cloud filter, err %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
8874
hlist_add_head(&filter->cloud_node, &pf->cloud_filter_list);
drivers/net/ethernet/intel/i40e/i40e_main.c
8876
pf->num_cloud_filters++;
drivers/net/ethernet/intel/i40e/i40e_main.c
889
test_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
8913
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
8930
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
8933
return i40e_aq_rc_to_posix(err, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
8936
pf->num_cloud_filters--;
drivers/net/ethernet/intel/i40e/i40e_main.c
8937
if (!pf->num_cloud_filters)
drivers/net/ethernet/intel/i40e/i40e_main.c
8938
if (test_bit(I40E_FLAG_FD_SB_TO_CLOUD_FILTER, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
8939
!test_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
8940
set_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8941
clear_bit(I40E_FLAG_FD_SB_TO_CLOUD_FILTER, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
8942
clear_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9022
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
9026
if (test_bit(__I40E_TESTING, pf->state) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
9027
test_bit(__I40E_BAD_EEPROM, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
9032
if (i40e_force_link_state(pf, true))
drivers/net/ethernet/intel/i40e/i40e_main.c
9040
wr32(&pf->hw, I40E_GLLAN_TSOMSK_F, be32_to_cpu(TCP_FLAG_PSH |
drivers/net/ethernet/intel/i40e/i40e_main.c
9042
wr32(&pf->hw, I40E_GLLAN_TSOMSK_M, be32_to_cpu(TCP_FLAG_PSH |
drivers/net/ethernet/intel/i40e/i40e_main.c
9045
wr32(&pf->hw, I40E_GLLAN_TSOMSK_L, be32_to_cpu(TCP_FLAG_CWR) >> 16);
drivers/net/ethernet/intel/i40e/i40e_main.c
9083
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_main.c
9101
dev_driver_string(&pf->pdev->dev), vsi->netdev->name);
drivers/net/ethernet/intel/i40e/i40e_main.c
9113
dev_driver_string(&pf->pdev->dev),
drivers/net/ethernet/intel/i40e/i40e_main.c
9114
dev_name(&pf->pdev->dev));
drivers/net/ethernet/intel/i40e/i40e_main.c
9139
i40e_do_reset(pf, I40E_PF_RESET_FLAG, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
9151
static void i40e_fdir_filter_exit(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9158
&pf->fdir_filter_list, fdir_node) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9163
list_for_each_entry_safe(pit_entry, tmp, &pf->l3_flex_pit_list, list) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9167
INIT_LIST_HEAD(&pf->l3_flex_pit_list);
drivers/net/ethernet/intel/i40e/i40e_main.c
9169
list_for_each_entry_safe(pit_entry, tmp, &pf->l4_flex_pit_list, list) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9173
INIT_LIST_HEAD(&pf->l4_flex_pit_list);
drivers/net/ethernet/intel/i40e/i40e_main.c
9175
pf->fdir_pf_active_filters = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
9176
i40e_reset_fdir_filter_cnt(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9179
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV4_TCP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9184
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV6_TCP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9189
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV4_UDP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9194
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV6_UDP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9199
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV4_SCTP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9204
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV6_SCTP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9209
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV4_OTHER,
drivers/net/ethernet/intel/i40e/i40e_main.c
9212
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_FRAG_IPV4,
drivers/net/ethernet/intel/i40e/i40e_main.c
9216
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV6_OTHER,
drivers/net/ethernet/intel/i40e/i40e_main.c
9219
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_FRAG_IPV6,
drivers/net/ethernet/intel/i40e/i40e_main.c
9230
static void i40e_cloud_filter_exit(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9236
&pf->cloud_filter_list, cloud_node) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9240
pf->num_cloud_filters = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
9242
if (test_bit(I40E_FLAG_FD_SB_TO_CLOUD_FILTER, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
9243
!test_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9244
set_bit(I40E_FLAG_FD_SB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9245
clear_bit(I40E_FLAG_FD_SB_TO_CLOUD_FILTER, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9246
clear_bit(I40E_FLAG_FD_SB_INACTIVE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9281
void i40e_do_reset(struct i40e_pf *pf, u32 reset_flags, bool lock_acquired)
drivers/net/ethernet/intel/i40e/i40e_main.c
9298
dev_dbg(&pf->pdev->dev, "GlobalR requested\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9299
val = rd32(&pf->hw, I40E_GLGEN_RTRIG);
drivers/net/ethernet/intel/i40e/i40e_main.c
9301
wr32(&pf->hw, I40E_GLGEN_RTRIG, val);
drivers/net/ethernet/intel/i40e/i40e_main.c
9309
dev_dbg(&pf->pdev->dev, "CoreR requested\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9310
val = rd32(&pf->hw, I40E_GLGEN_RTRIG);
drivers/net/ethernet/intel/i40e/i40e_main.c
9312
wr32(&pf->hw, I40E_GLGEN_RTRIG, val);
drivers/net/ethernet/intel/i40e/i40e_main.c
9313
i40e_flush(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
9325
dev_dbg(&pf->pdev->dev, "PFR requested\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9326
i40e_handle_reset_warning(pf, lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
9333
i40e_prep_for_reset(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9334
i40e_reset_and_rebuild(pf, true, lock_acquired);
drivers/net/ethernet/intel/i40e/i40e_main.c
9335
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
9336
test_bit(I40E_FLAG_FW_LLDP_DIS, pf->flags) ?
drivers/net/ethernet/intel/i40e/i40e_main.c
9342
dev_info(&pf->pdev->dev, "VSI reinit requested\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9344
i40e_pf_for_each_vsi(pf, i, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9351
dev_info(&pf->pdev->dev, "VSI down requested\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9353
i40e_pf_for_each_vsi(pf, i, vsi) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9361
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
9373
bool i40e_dcb_need_reconfig(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
9388
dev_dbg(&pf->pdev->dev, "ETS UP2TC changed.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9394
dev_dbg(&pf->pdev->dev, "ETS TC BW Table changed.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9399
dev_dbg(&pf->pdev->dev, "ETS TSA Table changed.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9407
dev_dbg(&pf->pdev->dev, "PFC config change detected.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9415
dev_dbg(&pf->pdev->dev, "APP Table change detected.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9418
dev_dbg(&pf->pdev->dev, "dcb need_reconfig=%d\n", need_reconfig);
drivers/net/ethernet/intel/i40e/i40e_main.c
9427
static int i40e_handle_lldp_event(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
9431
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
9441
!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
9443
set_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9446
if (!test_bit(I40E_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
9452
dev_dbg(&pf->pdev->dev, "LLDP event mib bridge type 0x%x\n", type);
drivers/net/ethernet/intel/i40e/i40e_main.c
9458
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
9474
ret = i40e_get_dcb_config(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
9480
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
9482
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9484
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_main.c
9487
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
9495
dev_dbg(&pf->pdev->dev, "No change detected in DCBX configuration.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9499
need_reconfig = i40e_dcb_need_reconfig(pf, &tmp_dcbx_cfg,
drivers/net/ethernet/intel/i40e/i40e_main.c
9502
i40e_dcbnl_flush_apps(pf, &tmp_dcbx_cfg, &hw->local_dcbx_config);
drivers/net/ethernet/intel/i40e/i40e_main.c
9509
set_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9511
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9513
set_bit(__I40E_PORT_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9515
i40e_pf_quiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9518
i40e_dcb_reconfigure(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9520
ret = i40e_resume_port_tx(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9522
clear_bit(__I40E_PORT_SUSPENDED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9528
ret = i40e_pf_wait_queues_disabled(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9531
set_bit(__I40E_PF_RESET_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9532
i40e_service_event_schedule(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9534
i40e_pf_unquiesce_all_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9535
set_bit(__I40E_CLIENT_SERVICE_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9536
set_bit(__I40E_CLIENT_L2_CHANGE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9550
void i40e_do_reset_safe(struct i40e_pf *pf, u32 reset_flags)
drivers/net/ethernet/intel/i40e/i40e_main.c
9553
i40e_do_reset(pf, reset_flags, true);
drivers/net/ethernet/intel/i40e/i40e_main.c
9565
static void i40e_handle_lan_overflow_event(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
9571
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_main.c
9575
dev_dbg(&pf->pdev->dev, "overflow Rx Queue Number = %d QTX_CTL=0x%08x\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
9585
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_main.c
9596
u32 i40e_get_current_fd_count(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9600
val = rd32(&pf->hw, I40E_PFQF_FDSTAT);
drivers/net/ethernet/intel/i40e/i40e_main.c
9610
u32 i40e_get_global_fd_count(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9614
val = rd32(&pf->hw, I40E_GLQF_FDCNT_0);
drivers/net/ethernet/intel/i40e/i40e_main.c
9624
static void i40e_reenable_fdir_sb(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9626
if (test_and_clear_bit(__I40E_FD_SB_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
9627
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
9628
(I40E_DEBUG_FD & pf->hw.debug_mask))
drivers/net/ethernet/intel/i40e/i40e_main.c
9629
dev_info(&pf->pdev->dev, "FD Sideband/ntuple is being enabled since we have space in the table now\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9636
static void i40e_reenable_fdir_atr(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9638
if (test_and_clear_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9644
i40e_write_fd_input_set(pf, LIBIE_FILTER_PCTYPE_NONF_IPV4_TCP,
drivers/net/ethernet/intel/i40e/i40e_main.c
9648
if (test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_main.c
9649
(I40E_DEBUG_FD & pf->hw.debug_mask))
drivers/net/ethernet/intel/i40e/i40e_main.c
9650
dev_info(&pf->pdev->dev, "ATR is being enabled since we have space in the table and there are no conflicting ntuple rules\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9659
static void i40e_delete_invalid_filter(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_main.c
9663
pf->fdir_pf_active_filters--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9664
pf->fd_inv = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
9668
pf->fd_tcp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9671
pf->fd_udp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9674
pf->fd_sctp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9677
pf->fd_tcp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9680
pf->fd_udp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9683
pf->fd_udp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9688
pf->fd_tcp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9691
pf->fd_udp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9694
pf->fd_sctp4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9697
pf->fd_ip4_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9704
pf->fd_tcp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9707
pf->fd_udp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9710
pf->fd_sctp6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9713
pf->fd_ip6_filter_cnt--;
drivers/net/ethernet/intel/i40e/i40e_main.c
9728
void i40e_fdir_check_and_reenable(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9734
if (test_bit(__I40E_FD_FLUSH_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
9738
fcnt_prog = i40e_get_global_fd_count(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9739
fcnt_avail = pf->fdir_pf_filter_count;
drivers/net/ethernet/intel/i40e/i40e_main.c
9741
(pf->fd_add_err == 0) ||
drivers/net/ethernet/intel/i40e/i40e_main.c
9742
(i40e_get_current_atr_cnt(pf) < pf->fd_atr_cnt))
drivers/net/ethernet/intel/i40e/i40e_main.c
9743
i40e_reenable_fdir_sb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9750
pf->fd_tcp4_filter_cnt == 0 && pf->fd_tcp6_filter_cnt == 0)
drivers/net/ethernet/intel/i40e/i40e_main.c
9751
i40e_reenable_fdir_atr(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9754
if (pf->fd_inv > 0) {
drivers/net/ethernet/intel/i40e/i40e_main.c
9756
&pf->fdir_filter_list, fdir_node)
drivers/net/ethernet/intel/i40e/i40e_main.c
9757
if (filter->fd_id == pf->fd_inv)
drivers/net/ethernet/intel/i40e/i40e_main.c
9758
i40e_delete_invalid_filter(pf, filter);
drivers/net/ethernet/intel/i40e/i40e_main.c
9768
static void i40e_fdir_flush_and_replay(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9776
if (!time_after(jiffies, pf->fd_flush_timestamp +
drivers/net/ethernet/intel/i40e/i40e_main.c
9783
min_flush_time = pf->fd_flush_timestamp +
drivers/net/ethernet/intel/i40e/i40e_main.c
9785
fd_room = pf->fdir_pf_filter_count - pf->fdir_pf_active_filters;
drivers/net/ethernet/intel/i40e/i40e_main.c
9789
if (I40E_DEBUG_FD & pf->hw.debug_mask)
drivers/net/ethernet/intel/i40e/i40e_main.c
9790
dev_info(&pf->pdev->dev, "ATR disabled, not enough FD filter space.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9794
pf->fd_flush_timestamp = jiffies;
drivers/net/ethernet/intel/i40e/i40e_main.c
9795
set_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9797
wr32(&pf->hw, I40E_PFQF_CTL_1,
drivers/net/ethernet/intel/i40e/i40e_main.c
9799
i40e_flush(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_main.c
9800
pf->fd_flush_cnt++;
drivers/net/ethernet/intel/i40e/i40e_main.c
9801
pf->fd_add_err = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
9805
reg = rd32(&pf->hw, I40E_PFQF_CTL_1);
drivers/net/ethernet/intel/i40e/i40e_main.c
9810
dev_warn(&pf->pdev->dev, "FD table did not flush, needs more time\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9813
i40e_fdir_filter_restore(i40e_pf_get_main_vsi(pf));
drivers/net/ethernet/intel/i40e/i40e_main.c
9814
if (!disable_atr && !pf->fd_tcp4_filter_cnt)
drivers/net/ethernet/intel/i40e/i40e_main.c
9815
clear_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9816
clear_bit(__I40E_FD_FLUSH_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9817
if (I40E_DEBUG_FD & pf->hw.debug_mask)
drivers/net/ethernet/intel/i40e/i40e_main.c
9818
dev_info(&pf->pdev->dev, "FD Filter table flushed and FD-SB replayed.\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9826
u32 i40e_get_current_atr_cnt(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9828
return i40e_get_current_fd_count(pf) - pf->fdir_pf_active_filters;
drivers/net/ethernet/intel/i40e/i40e_main.c
9835
static void i40e_fdir_reinit_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9839
if (test_bit(__I40E_DOWN, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
9842
if (test_bit(__I40E_FD_FLUSH_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_main.c
9843
i40e_fdir_flush_and_replay(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9845
i40e_fdir_check_and_reenable(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9892
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
9895
if (!veb || !veb->pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9897
pf = veb->pf;
drivers/net/ethernet/intel/i40e/i40e_main.c
9900
i40e_pf_for_each_vsi(pf, i, vsi)
drivers/net/ethernet/intel/i40e/i40e_main.c
9909
static void i40e_link_event(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9911
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9912
struct i40e_veb *veb = i40e_pf_get_main_veb(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9921
pf->hw.phy.get_link_info = true;
drivers/net/ethernet/intel/i40e/i40e_main.c
9922
old_link = (pf->hw.phy.link_info_old.link_info & I40E_AQ_LINK_UP);
drivers/net/ethernet/intel/i40e/i40e_main.c
9923
status = i40e_get_link_status(&pf->hw, &new_link);
drivers/net/ethernet/intel/i40e/i40e_main.c
9927
clear_bit(__I40E_TEMP_LINK_POLLING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9932
set_bit(__I40E_TEMP_LINK_POLLING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_main.c
9933
dev_dbg(&pf->pdev->dev, "couldn't get link state, status: %d\n",
drivers/net/ethernet/intel/i40e/i40e_main.c
9938
old_link_speed = pf->hw.phy.link_info_old.link_speed;
drivers/net/ethernet/intel/i40e/i40e_main.c
9939
new_link_speed = pf->hw.phy.link_info.link_speed;
drivers/net/ethernet/intel/i40e/i40e_main.c
9948
pf->link_down_events++;
drivers/net/ethernet/intel/i40e/i40e_main.c
9960
if (pf->vf)
drivers/net/ethernet/intel/i40e/i40e_main.c
9961
i40e_vc_notify_link_state(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9963
if (test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_main.c
9964
i40e_ptp_set_increment(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9969
if (pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED)
drivers/net/ethernet/intel/i40e/i40e_main.c
9976
dev_dbg(&pf->pdev->dev, "Reconfig DCB to single TC as result of Link Down\n");
drivers/net/ethernet/intel/i40e/i40e_main.c
9977
memset(&pf->tmp_cfg, 0, sizeof(pf->tmp_cfg));
drivers/net/ethernet/intel/i40e/i40e_main.c
9978
err = i40e_dcb_sw_default_config(pf);
drivers/net/ethernet/intel/i40e/i40e_main.c
9980
clear_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9981
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9983
pf->dcbx_cap = DCB_CAP_DCBX_HOST |
drivers/net/ethernet/intel/i40e/i40e_main.c
9985
set_bit(I40E_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9986
clear_bit(I40E_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_main.c
9996
static void i40e_watchdog_subtask(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1040
static void i40e_ptp_set_pins_hw(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1042
const struct i40e_ptp_pins_settings *pins = pf->ptp_pins;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1043
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1059
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1074
static int i40e_ptp_set_pins(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1077
enum i40e_can_set_pins pin_caps = i40e_can_set_pins(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1086
pins->sdp3_2 = pf->ptp_pins->sdp3_2;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1088
pins->sdp3_3 = pf->ptp_pins->sdp3_3;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1090
pins->gpio_4 = pf->ptp_pins->gpio_4;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1108
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1116
memcpy(pf->ptp_pins, pins, sizeof(*pins));
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1117
i40e_ptp_set_pins_hw(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1118
i40_ptp_reset_timing_events(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1129
int i40e_ptp_alloc_pins(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1131
if (!i40e_is_ptp_pin_dev(&pf->hw))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1134
pf->ptp_pins =
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1137
if (!pf->ptp_pins) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1138
dev_warn(&pf->pdev->dev, "Cannot allocate memory for PTP pins structure.\n");
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1142
pf->ptp_pins->sdp3_2 = off;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1143
pf->ptp_pins->sdp3_3 = off;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1144
pf->ptp_pins->gpio_4 = off;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1145
pf->ptp_pins->led2_0 = high;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1146
pf->ptp_pins->led2_1 = high;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1147
pf->ptp_pins->led3_0 = high;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1148
pf->ptp_pins->led3_1 = high;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1151
if (pf->hw.pf_id)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1154
i40e_ptp_init_leds_hw(&pf->hw);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1155
i40e_ptp_set_pins_hw(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1172
static int i40e_ptp_set_timestamp_mode(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1175
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1191
INIT_WORK(&pf->ptp_extts0_work, i40e_ptp_extts0_work);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1195
pf->ptp_tx = false;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1198
pf->ptp_tx = true;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1206
pf->ptp_rx = false;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1217
if (!test_bit(I40E_HW_CAP_PTP_L4, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1219
pf->ptp_rx = true;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1231
if (!test_bit(I40E_HW_CAP_PTP_L4, pf->hw.caps))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1237
pf->ptp_rx = true;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1240
if (test_bit(I40E_HW_CAP_PTP_L4, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1254
spin_lock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1261
pf->latch_event_flags = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1262
spin_unlock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1266
if (pf->ptp_tx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1273
if (pf->ptp_tx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1315
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1318
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1321
err = i40e_ptp_set_timestamp_mode(pf, config);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1326
pf->tstamp_config = *config;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1338
static int i40e_init_pin_config(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1342
pf->ptp_caps.n_pins = 3;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1343
pf->ptp_caps.n_ext_ts = 2;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1344
pf->ptp_caps.pps = 1;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1345
pf->ptp_caps.n_per_out = 2;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1347
pf->ptp_caps.pin_config = kzalloc_objs(*pf->ptp_caps.pin_config,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1348
pf->ptp_caps.n_pins);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1349
if (!pf->ptp_caps.pin_config)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1352
for (i = 0; i < pf->ptp_caps.n_pins; i++) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1353
snprintf(pf->ptp_caps.pin_config[i].name,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1354
sizeof(pf->ptp_caps.pin_config[i].name),
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1356
pf->ptp_caps.pin_config[i].index = sdp_desc[i].index;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1357
pf->ptp_caps.pin_config[i].func = PTP_PF_NONE;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1358
pf->ptp_caps.pin_config[i].chan = sdp_desc[i].chan;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1361
pf->ptp_caps.verify = i40e_ptp_verify;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1362
pf->ptp_caps.enable = i40e_ptp_feature_enable;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1364
pf->ptp_caps.pps = 1;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1379
static long i40e_ptp_create_clock(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1382
if (!IS_ERR_OR_NULL(pf->ptp_clock))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1385
strscpy(pf->ptp_caps.name, i40e_driver_name,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1386
sizeof(pf->ptp_caps.name) - 1);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1387
pf->ptp_caps.owner = THIS_MODULE;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1388
pf->ptp_caps.max_adj = 999999999;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1389
pf->ptp_caps.adjfine = i40e_ptp_adjfine;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1390
pf->ptp_caps.adjtime = i40e_ptp_adjtime;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1391
pf->ptp_caps.gettimex64 = i40e_ptp_gettimex;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1392
pf->ptp_caps.settime64 = i40e_ptp_settime;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1393
if (i40e_is_ptp_pin_dev(&pf->hw)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1394
int err = i40e_init_pin_config(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1401
pf->ptp_clock = ptp_clock_register(&pf->ptp_caps, &pf->pdev->dev);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1402
if (IS_ERR(pf->ptp_clock))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1403
return PTR_ERR(pf->ptp_clock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1409
pf->tstamp_config.rx_filter = HWTSTAMP_FILTER_NONE;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
141
static int i40e_ptp_set_pins(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1410
pf->tstamp_config.tx_type = HWTSTAMP_TX_OFF;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1413
ktime_get_real_ts64(&pf->ptp_prev_hw_time);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1414
pf->ptp_reset_start = ktime_get();
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1428
void i40e_ptp_save_hw_time(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1431
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1434
i40e_ptp_gettimex(&pf->ptp_caps, &pf->ptp_prev_hw_time, NULL);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1436
pf->ptp_reset_start = ktime_get();
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1451
void i40e_ptp_restore_hw_time(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1453
ktime_t delta = ktime_sub(ktime_get(), pf->ptp_reset_start);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1456
timespec64_add_ns(&pf->ptp_prev_hw_time, ktime_to_ns(delta));
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1459
i40e_ptp_settime(&pf->ptp_caps, &pf->ptp_prev_hw_time);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1475
void i40e_ptp_init(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1477
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1479
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1489
clear_bit(I40E_FLAG_PTP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1490
dev_info(&pf->pdev->dev, "%s: PTP not supported on %s\n",
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1496
mutex_init(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1497
spin_lock_init(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1500
err = i40e_ptp_create_clock(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1502
pf->ptp_clock = NULL;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1503
dev_err(&pf->pdev->dev, "%s: ptp_clock_register failed\n",
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1505
} else if (pf->ptp_clock) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1508
if (pf->hw.debug_mask & I40E_DEBUG_LAN)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1509
dev_info(&pf->pdev->dev, "PHC enabled\n");
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1510
set_bit(I40E_FLAG_PTP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
152
struct i40e_pf *pf = container_of(work, struct i40e_pf,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1521
i40e_ptp_set_increment(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1524
i40e_ptp_set_timestamp_mode(pf, &pf->tstamp_config);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1527
i40e_ptp_restore_hw_time(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1530
i40e_ptp_set_1pps_signal_hw(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
154
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1540
void i40e_ptp_stop(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1542
struct i40e_vsi *main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1543
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1546
clear_bit(I40E_FLAG_PTP_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1547
pf->ptp_tx = false;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1548
pf->ptp_rx = false;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1550
if (pf->ptp_tx_skb) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1551
struct sk_buff *skb = pf->ptp_tx_skb;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1553
pf->ptp_tx_skb = NULL;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1554
clear_bit_unlock(__I40E_PTP_TX_IN_PROGRESS, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1558
if (pf->ptp_clock) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1559
ptp_clock_unregister(pf->ptp_clock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1560
pf->ptp_clock = NULL;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1561
dev_info(&pf->pdev->dev, "%s: removed PHC on %s\n", __func__,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1565
if (i40e_is_ptp_pin_dev(&pf->hw)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
1580
i40e_ptp_free_pins(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
172
ptp_clock_event(pf->ptp_clock, &event);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
196
static enum i40e_can_set_pins i40e_can_set_pins(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
198
if (!i40e_is_ptp_pin_dev(&pf->hw)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
199
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
204
if (!pf->ptp_pins) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
205
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
210
if (pf->hw.pf_id) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
211
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
225
static void i40_ptp_reset_timing_events(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
229
spin_lock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
232
rd32(&pf->hw, I40E_PRTTSYN_RXTIME_L(i));
drivers/net/ethernet/intel/i40e/i40e_ptp.c
233
rd32(&pf->hw, I40E_PRTTSYN_RXTIME_H(i));
drivers/net/ethernet/intel/i40e/i40e_ptp.c
234
pf->latch_events[i] = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
237
rd32(&pf->hw, I40E_PRTTSYN_TXTIME_L);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
238
rd32(&pf->hw, I40E_PRTTSYN_TXTIME_H);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
240
pf->tx_hwtstamp_timeouts = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
241
pf->tx_hwtstamp_skipped = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
242
pf->rx_hwtstamp_cleared = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
243
pf->latch_event_flags = 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
244
spin_unlock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
281
static void i40e_ptp_read(struct i40e_pf *pf, struct timespec64 *ts,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
284
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
308
static void i40e_ptp_write(struct i40e_pf *pf, const struct timespec64 *ts)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
310
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
349
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
350
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
354
base_adj = I40E_PTP_40GB_INCVAL * READ_ONCE(pf->ptp_adj_mult);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
370
static void i40e_ptp_set_1pps_signal_hw(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
372
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
380
i40e_ptp_read(pf, &now, NULL);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
404
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
405
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
407
mutex_lock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
429
i40e_ptp_read(pf, &now, NULL);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
431
i40e_ptp_write(pf, (const struct timespec64 *)&now);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
432
i40e_ptp_set_1pps_signal_hw(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
435
mutex_unlock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
452
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
454
mutex_lock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
455
i40e_ptp_read(pf, ts, sts);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
456
mutex_unlock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
472
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
474
mutex_lock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
475
i40e_ptp_write(pf, ts);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
476
mutex_unlock(&pf->tmreg_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
494
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
497
i40e_ptp_set_1pps_signal_hw(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
536
static int i40e_ptp_enable_pin(struct i40e_pf *pf, unsigned int chan,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
544
if (pf->hw.pf_id)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
548
pins.sdp3_2 = pf->ptp_pins->sdp3_2;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
549
pins.sdp3_3 = pf->ptp_pins->sdp3_3;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
550
pins.gpio_4 = pf->ptp_pins->gpio_4;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
559
pin_index = ptp_find_pin(pf->ptp_clock, func, chan);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
584
return i40e_ptp_set_pins(pf, &pins) ? -EINVAL : 0;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
599
struct i40e_pf *pf = container_of(ptp, struct i40e_pf, ptp_caps);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
620
return i40e_ptp_enable_pin(pf, chan, func, on);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
635
static u32 i40e_ptp_get_rx_events(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
637
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
642
new_latch_events = prttsyn_stat & ~pf->latch_event_flags;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
655
pf->latch_events[i] = jiffies;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
659
pf->latch_event_flags = prttsyn_stat;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
673
void i40e_ptp_rx_hang(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
675
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
683
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags) || !pf->ptp_rx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
686
spin_lock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
689
i40e_ptp_get_rx_events(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
698
if ((pf->latch_event_flags & BIT(i)) &&
drivers/net/ethernet/intel/i40e/i40e_ptp.c
699
time_is_before_jiffies(pf->latch_events[i] + HZ)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
701
pf->latch_event_flags &= ~BIT(i);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
706
spin_unlock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
715
dev_dbg(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
720
pf->rx_hwtstamp_cleared += cleared;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
732
void i40e_ptp_tx_hang(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
736
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags) || !pf->ptp_tx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
740
if (!test_bit(__I40E_PTP_TX_IN_PROGRESS, pf->state))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
747
if (time_is_before_jiffies(pf->ptp_tx_start + HZ)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
748
skb = pf->ptp_tx_skb;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
749
pf->ptp_tx_skb = NULL;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
750
clear_bit_unlock(__I40E_PTP_TX_IN_PROGRESS, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
754
pf->tx_hwtstamp_timeouts++;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
766
void i40e_ptp_tx_hwtstamp(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
769
struct sk_buff *skb = pf->ptp_tx_skb;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
770
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
774
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags) || !pf->ptp_tx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
778
if (!pf->ptp_tx_skb)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
792
pf->ptp_tx_skb = NULL;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
793
clear_bit_unlock(__I40E_PTP_TX_IN_PROGRESS, pf->state);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
812
void i40e_ptp_rx_hwtstamp(struct i40e_pf *pf, struct sk_buff *skb, u8 index)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
821
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags) || !pf->ptp_rx)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
824
hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
826
spin_lock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
829
prttsyn_stat = i40e_ptp_get_rx_events(pf);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
833
spin_unlock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
838
pf->latch_event_flags &= ~BIT(index);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
843
spin_unlock_bh(&pf->ptp_rx_lock);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
858
void i40e_ptp_set_increment(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
861
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
867
i40e_aq_get_link_info(&pf->hw, true, NULL, NULL);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
884
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_ptp.c
910
WRITE_ONCE(pf->ptp_adj_mult, mult);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
927
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
929
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_ptp.c
932
*config = pf->tstamp_config;
drivers/net/ethernet/intel/i40e/i40e_ptp.c
943
static void i40e_ptp_free_pins(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_ptp.c
945
if (i40e_is_ptp_pin_dev(&pf->hw)) {
drivers/net/ethernet/intel/i40e/i40e_ptp.c
946
kfree(pf->ptp_pins);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
947
kfree(pf->ptp_caps.pin_config);
drivers/net/ethernet/intel/i40e/i40e_ptp.c
948
pf->ptp_pins = NULL;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
27
struct i40e_pf *pf = tx_ring->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2862
struct i40e_pf *pf = tx_ring->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2875
if (!test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2878
if (test_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2918
if (th->syn && test_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2920
if (test_bit(I40E_FLAG_HW_ATR_EVICT_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2974
I40E_FD_ATR_STAT_IDX(pf->hw.pf_id));
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2978
I40E_FD_ATR_TUNNEL_STAT_IDX(pf->hw.pf_id));
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2980
if (test_bit(I40E_FLAG_HW_ATR_EVICT_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
303
static int i40e_prepare_fdir_filter(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3201
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3213
pf = i40e_netdev_to_pf(tx_ring->netdev);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3214
if (!test_bit(I40E_FLAG_PTP_ENA, pf->flags))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3217
if (pf->ptp_tx &&
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3218
!test_and_set_bit_lock(__I40E_PTP_TX_IN_PROGRESS, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3220
pf->ptp_tx_start = jiffies;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3221
pf->ptp_tx_skb = skb_get(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3223
pf->tx_hwtstamp_skipped++;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
325
ret = i40e_program_fdir_filter(fd_data, packet_addr, pf, add);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
327
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
332
} else if (I40E_DEBUG_FD & pf->hw.debug_mask) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
334
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
338
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
387
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3959
struct i40e_pf *pf = i40e_netdev_to_pf(tx_ring->netdev);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3961
dev_kfree_skb_any(pf->ptp_tx_skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3962
pf->ptp_tx_skb = NULL;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3963
clear_bit_unlock(__I40E_PTP_TX_IN_PROGRESS, pf->state);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
399
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
4010
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
4019
test_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
404
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
413
i40e_change_filter_num(ipv4, add, &pf->fd_udp4_filter_cnt,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
414
&pf->fd_udp6_filter_cnt);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
435
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
446
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
451
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
46
vsi_id = fdata->dest_vsi ? : i40e_pf_get_main_vsi(pf)->id;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
460
i40e_change_filter_num(ipv4, add, &pf->fd_tcp4_filter_cnt,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
461
&pf->fd_tcp6_filter_cnt);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
464
if (test_bit(I40E_FLAG_FD_ATR_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_txrx.c
465
I40E_DEBUG_FD & pf->hw.debug_mask)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
466
dev_info(&pf->pdev->dev, "Forcing ATR off, sideband rules for TCP/IPv4 flow being applied\n");
drivers/net/ethernet/intel/i40e/i40e_txrx.c
467
set_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
489
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
501
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
506
(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
515
i40e_change_filter_num(ipv4, add, &pf->fd_sctp4_filter_cnt,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
516
&pf->fd_sctp6_filter_cnt);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
538
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
566
ret = i40e_prepare_fdir_filter(pf, fd_data, add, raw_packet,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
572
i40e_change_filter_num(ipv4, add, &pf->fd_ip4_filter_cnt,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
573
&pf->fd_ip6_filter_cnt);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
592
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
630
dev_info(&pf->pdev->dev, "Unsupported IPv4 protocol 0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_txrx.c
651
dev_info(&pf->pdev->dev, "Unsupported IPv6 protocol 0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_txrx.c
657
dev_info(&pf->pdev->dev, "Unsupported flow type 0x%02x\n",
drivers/net/ethernet/intel/i40e/i40e_txrx.c
684
struct i40e_pf *pf = rx_ring->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
685
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
694
pf->fd_inv = le32_to_cpu(qw0->hi_dword.fd_id);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
696
(I40E_DEBUG_FD & pf->hw.debug_mask))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
698
pf->fd_inv);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
706
if (test_bit(__I40E_FD_FLUSH_REQUESTED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
709
pf->fd_add_err++;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
711
pf->fd_atr_cnt = i40e_get_current_atr_cnt(pf);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
714
test_bit(__I40E_FD_SB_AUTO_DISABLED, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
721
set_bit(__I40E_FD_ATR_AUTO_DISABLED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
722
set_bit(__I40E_FD_FLUSH_REQUESTED, pf->state);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
726
fcnt_prog = i40e_get_global_fd_count(pf);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
727
fcnt_avail = pf->fdir_pf_filter_count;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
733
if (test_bit(I40E_FLAG_FD_SB_ENA, pf->flags) &&
drivers/net/ethernet/intel/i40e/i40e_txrx.c
735
pf->state))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
736
if (I40E_DEBUG_FD & pf->hw.debug_mask)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
740
if (I40E_DEBUG_FD & pf->hw.debug_mask)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
83
u8 *raw_packet, struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
871
void i40e_detect_recover_hung(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
873
struct i40e_vsi *vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
96
vsi = i40e_find_vsi_by_type(pf, I40E_VSI_FDIR);
drivers/net/ethernet/intel/i40e/i40e_txrx.h
474
void i40e_detect_recover_hung(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_txrx.h
96
#define i40e_pf_get_default_rss_hashcfg(pf) \
drivers/net/ethernet/intel/i40e/i40e_txrx.h
97
(test_bit(I40E_HW_CAP_MULTI_TCP_UDP_RSS_PCTYPE, (pf)->hw.caps) ? \
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1016
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1017
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1036
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1037
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1050
pf->queues_left += vf->num_queue_pairs -
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1056
i40e_vsi_release(pf->vsi[vf->lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1069
i40e_vsi_release(pf->vsi[vf->ch[j].vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1074
msix_vf = pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1117
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1122
vf->num_req_queues <= pf->queues_left + I40E_DEFAULT_QUEUES_PER_VF)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1123
pf->num_vf_qps = vf->num_req_queues;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1125
pf->num_vf_qps = I40E_DEFAULT_QUEUES_PER_VF;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
113
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1131
total_queue_pairs += pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1135
if (pf->queues_left >=
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
114
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1146
dev_info(&pf->pdev->dev, "VF %d: Not enough queues to allocate, disabling ADq\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
115
struct i40e_link_status *ls = &pf->hw.phy.link_info;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1158
pf->queues_left -=
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1192
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1193
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1292
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1293
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1303
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1305
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1318
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1320
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
133
void i40e_vc_notify_link_state(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1334
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1336
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1349
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1351
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
137
for (i = 0; i < pf->num_alloc_vfs; i++)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
138
i40e_vc_notify_vf_link_state(&pf->vf[i]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1382
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1388
vsi = i40e_find_vsi_from_id(pf, vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1451
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1452
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
147
void i40e_vc_notify_reset(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1479
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1498
dev_err(&pf->pdev->dev, "VF %d PCI transactions stuck\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1512
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1513
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
153
i40e_vc_vf_broadcast(pf, VIRTCHNL_OP_EVENT, 0,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1545
i40e_notify_client_of_vf_reset(pf, abs_vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1566
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1567
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1572
if (test_bit(__I40E_VF_RESETS_DISABLED, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1576
if (test_bit(__I40E_VF_DISABLE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1606
dev_err(&pf->pdev->dev, "VF reset check timeout on VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1612
i40e_vsi_stop_rings(pf->vsi[vf->lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1635
bool i40e_reset_all_vfs(struct i40e_pf *pf, bool flr)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1637
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1643
if (!pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1647
if (test_and_set_bit(__I40E_VF_DISABLE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1651
for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1663
for (i = 0, vf = &pf->vf[0]; i < 10 && vf < &pf->vf[pf->num_alloc_vfs]; ++i) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1669
while (vf < &pf->vf[pf->num_alloc_vfs]) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1689
if (vf < &pf->vf[pf->num_alloc_vfs])
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1690
dev_err(&pf->pdev->dev, "VF reset check timeout on VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1697
for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1706
i40e_vsi_stop_rings_no_wait(pf->vsi[vf->lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1712
for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1721
i40e_vsi_wait_queues_disabled(pf->vsi[vf->lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1730
for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1740
clear_bit(__I40E_VF_DISABLE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1751
void i40e_free_vfs(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1753
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1757
if (!pf->vf)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1760
set_bit(__I40E_VFS_RELEASING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1761
while (test_and_set_bit(__I40E_VF_DISABLE, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1764
i40e_notify_client_of_vf_enable(pf, 0);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1770
if (!pci_vfs_assigned(pf->pdev))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1771
pci_disable_sriov(pf->pdev);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1773
dev_warn(&pf->pdev->dev, "VFs are assigned - not disabling SR-IOV\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1776
for (i = 0; i < pf->num_alloc_vfs; i++) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1777
if (test_bit(I40E_VF_STATE_INIT, &pf->vf[i].vf_states))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1780
i40e_vsi_stop_rings_no_wait(pf->vsi[pf->vf[i].lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1783
for (i = 0; i < pf->num_alloc_vfs; i++) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1784
if (test_bit(I40E_VF_STATE_INIT, &pf->vf[i].vf_states))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1787
i40e_vsi_wait_queues_disabled(pf->vsi[pf->vf[i].lan_vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1791
tmp = pf->num_alloc_vfs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1792
pf->num_alloc_vfs = 0;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1794
if (test_bit(I40E_VF_STATE_INIT, &pf->vf[i].vf_states))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1795
i40e_free_vf_res(&pf->vf[i]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1797
i40e_disable_vf_mappings(&pf->vf[i]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1800
kfree(pf->vf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1801
pf->vf = NULL;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1807
if (!pci_vfs_assigned(pf->pdev)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1817
clear_bit(__I40E_VF_DISABLE, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1818
clear_bit(__I40E_VFS_RELEASING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1829
int i40e_alloc_vfs(struct i40e_pf *pf, u16 num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1835
i40e_irq_dynamic_disable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1838
if (pci_num_vf(pf->pdev) != num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1839
ret = pci_enable_sriov(pf->pdev, num_alloc_vfs);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1841
clear_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1842
pf->num_alloc_vfs = 0;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1852
pf->vf = vfs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1856
vfs[i].pf = pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1867
pf->num_alloc_vfs = num_alloc_vfs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1870
i40e_reset_all_vfs(pf, false);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1872
i40e_notify_client_of_vf_enable(pf, num_alloc_vfs);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1876
i40e_free_vfs(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1879
i40e_irq_dynamic_enable_icr0(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1894
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1898
if (test_bit(__I40E_TESTING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1906
i40e_free_vfs(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1910
if (num_vfs > pf->num_req_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1912
num_vfs, pf->num_req_vfs);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1918
err = i40e_alloc_vfs(pf, num_vfs);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1943
struct i40e_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1946
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
195
if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1952
if (!test_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1953
set_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1954
i40e_do_reset_safe(pf, I40E_PF_RESET_AND_REBUILD_FLAG);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1960
if (!pci_vfs_assigned(pf->pdev)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1961
i40e_free_vfs(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1962
clear_bit(I40E_FLAG_VEB_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1963
i40e_do_reset_safe(pf, I40E_PF_RESET_AND_REBUILD_FLAG);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1970
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1989
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1995
if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1998
pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1999
hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
20
static void i40e_vc_vf_broadcast(struct i40e_pf *pf,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2005
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2007
vf->vf_id, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
203
abs_vf_id = vf->vf_id + (int)vf->pf->hw.func_caps.vf_base_id;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
207
i40e_aq_send_msg_to_vf(&vf->pf->hw, abs_vf_id, VIRTCHNL_OP_EVENT,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2083
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2091
i40e_vsi_release(pf->vsi[vf->ch[i].vsi_idx]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2109
u16 max_frame_size = vf->pf->hw.phy.link_info.max_frame_size;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2127
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2158
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2162
if (i40e_vf_client_capable(pf, vf->vf_id) &&
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2173
if (test_bit(I40E_HW_CAP_RSS_AQ, pf->hw.caps) &&
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2180
if (test_bit(I40E_HW_CAP_MULTI_TCP_UDP_RSS_PCTYPE, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2189
if (test_bit(I40E_HW_CAP_OUTER_UDP_CSUM, pf->hw.caps) &&
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2194
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2195
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2204
if (test_bit(I40E_HW_CAP_WB_ON_ITR, pf->hw.caps)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
221
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2218
vfres->max_vectors = pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2263
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2273
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2308
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2313
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2320
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2325
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2350
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
236
if (test_bit(__I40E_VFS_RELEASING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
244
dev_warn(&vf->pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2440
pf->vsi[vf->lan_vsi_idx]->num_queue_pairs =
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2444
vsi = pf->vsi[vf->ch[i].vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
248
dev_dbg(&vf->pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
25
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2513
vf->pf->hw.func_caps.num_msix_vectors_vf) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2555
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2560
ret = i40e_control_wait_tx_q(vsi->seid, pf,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2578
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2583
ret = i40e_control_wait_rx_q(pf, vsi->base_queue + q_id,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
26
struct i40e_vf *vf = pf->vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2618
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
262
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2624
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
263
struct i40e_vsi *vsi = i40e_find_vsi_from_id(pf, vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2646
if (i40e_ctrl_vf_rx_rings(pf->vsi[vf->lan_vsi_idx], vqs->rx_queues,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2651
if (i40e_ctrl_vf_tx_rings(pf->vsi[vf->lan_vsi_idx], vqs->tx_queues,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2661
if (i40e_vsi_start_rings(pf->vsi[vf->ch[i].vsi_idx]))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2684
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2703
if (i40e_ctrl_vf_tx_rings(pf->vsi[vf->lan_vsi_idx], vqs->tx_queues,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2708
if (i40e_ctrl_vf_rx_rings(pf->vsi[vf->lan_vsi_idx], vqs->rx_queues,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2730
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2733
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2740
pile = pf->qp_pile;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2788
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
279
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2794
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2799
} else if (req_pairs - cur_pairs > pf->queues_left) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
280
struct i40e_vsi *vsi = i40e_find_vsi_from_id(pf, vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2800
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2804
pf->queues_left);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2805
vfres->num_queue_pairs = pf->queues_left + cur_pairs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2807
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2834
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2851
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
29
for (i = 0; i < pf->num_alloc_vfs; i++, vf++) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2900
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2901
struct i40e_vsi *vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2902
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2914
dev_err(&pf->pdev->dev, "invalid VF MAC addr %pM\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2928
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
294
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2955
mac_add_max = I40E_VC_MAX_MACVLAN_PER_TRUSTED_VF(pf->num_alloc_vfs, hw->num_ports);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2957
if (pf->max_mac_per_vf > 0)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2958
mac_add_max = pf->max_mac_per_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
296
return vector_id < pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2966
if (pf->max_mac_per_vf == mac_add_max && mac_add_max > 0) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2967
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2973
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2977
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3063
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3074
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3096
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3111
dev_err(&pf->pdev->dev, "Unable to program VF %d MAC filters, error %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
312
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
313
struct i40e_vsi *vsi = i40e_find_vsi_from_id(pf, vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3132
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3146
dev_err(&pf->pdev->dev, "Invalid MAC addr %pM for VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3152
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3184
dev_err(&pf->pdev->dev, "Unable to program VF %d MAC filters, error %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3219
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3226
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3239
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3244
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3258
i40e_aq_set_vsi_uc_promisc_on_vlan(&pf->hw, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3263
i40e_aq_set_vsi_mc_promisc_on_vlan(&pf->hw, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3269
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3290
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3308
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3320
i40e_aq_set_vsi_uc_promisc_on_vlan(&pf->hw, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3325
i40e_aq_set_vsi_mc_promisc_on_vlan(&pf->hw, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3346
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3357
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3358
abs_vf_id = vf->vf_id + pf->hw.func_caps.vf_base_id;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3413
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3424
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3443
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3461
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3479
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3495
vrh->hashcfg = i40e_pf_get_default_rss_hashcfg(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3515
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3516
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3550
vsi = vf->pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3576
vsi = vf->pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3597
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3605
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3614
dev_info(&pf->pdev->dev, "VF %d: Invalid TC number %u\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3629
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3633
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3648
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3657
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3667
dev_info(&pf->pdev->dev, "VF %d: Invalid Dest MAC addr %pM\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3676
dev_info(&pf->pdev->dev, "VF %d: Invalid Source MAC addr %pM\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3684
dev_info(&pf->pdev->dev, "VF %d: Invalid Dest port\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3692
dev_info(&pf->pdev->dev, "VF %d: Invalid Source port\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3700
dev_info(&pf->pdev->dev, "VF %d: Invalid Flow type\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3707
dev_info(&pf->pdev->dev, "VF %d: invalid VLAN ID\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3725
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3730
vsi = i40e_find_vsi_from_id(pf, vf->ch[i].vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
374
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3746
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
375
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3756
dev_err(&pf->pdev->dev, "VF %d: no VSI found for matching %u seid, can't delete cloud filter\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3767
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3770
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3791
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3803
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3811
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3854
dev_info(&pf->pdev->dev, "VF %d: Flow type not configured\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3859
vsi = pf->vsi[vf->ch[vcf->action_meta].vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3869
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3872
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
388
((pf->hw.func_caps.num_msix_vectors_vf - 1) * vf->vf_id) +
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3925
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3936
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3944
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3952
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4000
dev_info(&pf->pdev->dev, "VF %d: Flow type not configured\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4005
vsi = pf->vsi[vf->ch[vcf->action_meta].vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4015
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4018
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4043
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4044
struct i40e_link_status *ls = &pf->hw.phy.link_info;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4056
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4063
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4072
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4083
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4094
if (pf->queues_left < adq_request_qps) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4095
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4111
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4122
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4157
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4170
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4174
dev_info(&pf->pdev->dev, "VF %d trying to delete queue channels but ADq isn't enabled\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4201
int i40e_vc_process_vf_msg(struct i40e_pf *pf, s16 vf_id, u32 v_opcode,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4204
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4209
pf->vf_aq_requests++;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4210
if (local_vf_id < 0 || local_vf_id >= pf->num_alloc_vfs)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4212
vf = &(pf->vf[local_vf_id]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4223
dev_err(&pf->pdev->dev, "Invalid message from VF %d, opcode %d, len %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4315
dev_err(&pf->pdev->dev, "Unsupported opcode %d from VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4332
int i40e_vc_process_vflr_event(struct i40e_pf *pf)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4334
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4339
if (!test_bit(__I40E_VFLR_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4352
clear_bit(__I40E_VFLR_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4353
for (vf_id = 0; vf_id < pf->num_alloc_vfs; vf_id++) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4357
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4363
set_bit(__I40E_VFLR_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4379
static int i40e_validate_vf(struct i40e_pf *pf, int vf_id)
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4385
if (vf_id >= pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4386
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4391
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4392
vsi = i40e_find_vsi_from_id(pf, vf->lan_vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4423
dev_err(&vf->pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4443
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4450
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4451
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4456
ret = i40e_validate_vf(pf, vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4460
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4465
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4468
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4493
dev_err(&pf->pdev->dev, "Unable to program ucast filters\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4501
dev_info(&pf->pdev->dev, "Removing MAC on VF %d\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4504
dev_info(&pf->pdev->dev, "Setting MAC %pM on VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4512
dev_info(&pf->pdev->dev, "Bring down and up the VF interface to make this change effective.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4515
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4535
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4540
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4541
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4546
ret = i40e_validate_vf(pf, vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4551
dev_err(&pf->pdev->dev, "Invalid VF Parameters\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4557
dev_err(&pf->pdev->dev, "VF VLAN protocol is not supported\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4562
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4567
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4611
dev_err(&pf->pdev->dev, "Unable to config VF promiscuous mode\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4622
dev_info(&pf->pdev->dev, "Setting VLAN %d, QOS 0x%x on VF %d\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4651
dev_err(&pf->pdev->dev, "Unable to update VF vsi context\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4662
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4666
dev_err(&pf->pdev->dev, "Unable to config vf promiscuous mode\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4673
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4690
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4695
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4696
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4701
ret = i40e_validate_vf(pf, vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4706
dev_err(&pf->pdev->dev, "Invalid min tx rate (%d) (greater than 0) specified for VF %d.\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4712
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4717
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4725
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4742
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4746
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4747
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4752
ret = i40e_validate_vf(pf, vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4756
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4758
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
478
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4783
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4798
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4799
struct i40e_link_status *ls = &pf->hw.phy.link_info;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4801
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4810
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4811
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4816
if (vf_id >= pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4817
dev_err(&pf->pdev->dev, "Invalid VF Identifier %d\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4822
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
486
msix_vf = pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4860
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4884
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
490
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4900
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4902
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4906
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4907
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4912
if (vf_id >= pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4913
dev_err(&pf->pdev->dev, "Invalid VF Identifier %d\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4918
vf = &(pf->vf[vf_id]);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4929
ctxt.seid = pf->vsi[vf->lan_vsi_idx]->seid;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4930
ctxt.pf_num = pf->hw.pf_id;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4937
dev_err(&pf->pdev->dev, "Error %d updating VSI parameters\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4942
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4957
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4961
if (test_and_set_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4962
dev_warn(&pf->pdev->dev, "Unable to configure VFs, other operation is pending.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4967
if (vf_id >= pf->num_alloc_vfs) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4968
dev_err(&pf->pdev->dev, "Invalid VF Identifier %d\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4973
if (test_bit(I40E_FLAG_MFP_ENA, pf->flags)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4974
dev_err(&pf->pdev->dev, "Trusted VF not supported in MFP mode.\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4979
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4987
set_bit(__I40E_MACVLAN_SYNC_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4988
pf->vsi[vf->lan_vsi_idx]->flags |= I40E_VSI_FLAG_FILTER_CHANGED;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4991
dev_info(&pf->pdev->dev, "VF %u is now %strusted\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4996
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5004
clear_bit(__I40E_VIRTCHNL_OP_PENDING, pf->state);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5018
struct i40e_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5024
if (i40e_validate_vf(pf, vf_id))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5027
vf = &pf->vf[vf_id];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5029
dev_err(&pf->pdev->dev, "VF %d in reset. Try again.\n", vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
5033
vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
530
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
531
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
539
msix_vf = pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
542
dev_warn(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
560
msix_vf = pf->hw.func_caps.num_msix_vectors_vf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
632
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
633
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
645
vsi = i40e_find_vsi_from_id(pf, vsi_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
659
info->ring_len > i40e_get_max_num_descriptors(pf)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
672
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
682
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
715
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
716
struct i40e_vsi *vsi = pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
717
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
729
info->ring_len > i40e_get_max_num_descriptors(pf)) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
781
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
791
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
813
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
817
main_vsi = i40e_pf_get_main_vsi(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
818
vsi = i40e_vsi_setup(pf, I40E_VSI_SRIOV, main_vsi->seid, vf->vf_id);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
821
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
823
vf->vf_id, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
829
u64 hashcfg = i40e_pf_get_default_rss_hashcfg(pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
848
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
855
dev_info(&pf->pdev->dev,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
858
wr32(&pf->hw, I40E_VFQF_HENA1(0, vf->vf_id), (u32)hashcfg);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
859
wr32(&pf->hw, I40E_VFQF_HENA1(1, vf->vf_id),
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
864
dev_err(&pf->pdev->dev, "Unable to program ucast filters\n");
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
882
ret = i40e_aq_config_vsi_bw_limit(&pf->hw, vsi->seid,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
885
dev_err(&pf->pdev->dev, "Unable to set tx rate, VF %d, error code %d.\n",
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
902
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
903
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
916
qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
949
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
950
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
964
qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
987
struct i40e_pf *pf = vf->pf;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
988
struct i40e_hw *hw = &pf->hw;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
126
void i40e_free_vfs(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
128
int i40e_alloc_vfs(struct i40e_pf *pf, u16 num_alloc_vfs);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
129
int i40e_vc_process_vf_msg(struct i40e_pf *pf, s16 vf_id, u32 v_opcode,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
131
int i40e_vc_process_vflr_event(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
134
bool i40e_reset_all_vfs(struct i40e_pf *pf, bool flr);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
149
void i40e_vc_notify_link_state(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
150
void i40e_vc_notify_reset(struct i40e_pf *pf);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.h
76
struct i40e_pf *pf;
drivers/net/ethernet/intel/i40e/i40e_xsk.c
695
struct i40e_pf *pf = vsi->back;
drivers/net/ethernet/intel/i40e/i40e_xsk.c
698
if (test_bit(__I40E_CONFIG_BUSY, pf->state))
drivers/net/ethernet/intel/ice/devlink/devlink.c
100
struct ice_nvm_info *nvm = &pf->hw.flash.nvm;
drivers/net/ethernet/intel/ice/devlink/devlink.c
106
ice_info_pending_nvm_ver(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
116
static void ice_info_eetrack(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
118
struct ice_nvm_info *nvm = &pf->hw.flash.nvm;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1212
static void ice_set_min_max_msix(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1214
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1222
pf->msix.min = val.vu32;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1228
pf->msix.max = val.vu32;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1235
static int ice_devlink_reinit_up(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1237
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1238
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
124
ice_info_pending_eetrack(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1242
err = ice_init_hw(&pf->hw);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1249
ice_set_min_max_msix(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1251
err = ice_init_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1255
err = ice_init_pf(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1270
err = ice_load(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1281
ice_deinit_pf(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1285
ice_deinit_hw(&pf->hw);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1287
ice_deinit_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1306
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1311
return ice_devlink_reinit_up(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1314
return ice_devlink_reload_empr_finish(pf, extack);
drivers/net/ethernet/intel/ice/devlink/devlink.c
132
static void ice_info_ddp_pkg_name(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
134
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1358
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1361
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1374
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1379
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1384
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1390
ret = ice_plug_aux_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
140
ice_info_ddp_pkg_version(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1402
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1405
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1409
if (!test_bit(ICE_FLAG_RDMA_ENA, pf->flags))
drivers/net/ethernet/intel/ice/devlink/devlink.c
142
struct ice_pkg_ver *pkg = &pf->hw.active_pkg_ver;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1425
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1428
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1441
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1446
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1451
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1457
ret = ice_plug_aux_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1469
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1471
if (!test_bit(ICE_FLAG_RDMA_ENA, pf->flags))
drivers/net/ethernet/intel/ice/devlink/devlink.c
1474
if (pf->cdev_info->rdma_protocol & IIDC_RDMA_PROTOCOL_ROCEV2) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
149
ice_info_ddp_pkg_bundle_id(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
151
snprintf(ctx->buf, sizeof(ctx->buf), "0x%08x", pf->hw.active_track_id);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1538
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
154
static void ice_info_netlist_ver(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1542
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
156
struct ice_netlist_info *netlist = &pf->hw.flash.netlist;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1563
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1564
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1567
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1571
ice_schedule_reset(pf, ICE_RESET_CORER);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1608
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1610
if (val.vu32 > pf->hw.func_caps.common_cap.num_msix_vectors)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1631
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1634
if (new_state && !test_bit(ICE_FLAG_RDMA_ENA, pf->flags))
drivers/net/ethernet/intel/ice/devlink/devlink.c
165
static void ice_info_netlist_build(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
167
struct ice_netlist_info *netlist = &pf->hw.flash.netlist;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1721
struct ice_sf_priv *ice_allocate_sf(struct device *dev, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
173
ice_info_pending_netlist_ver(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
1731
err = devl_nested_devlink_set(priv_to_devlink(pf), devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1748
void ice_devlink_register(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1750
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1761
void ice_devlink_unregister(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1763
devl_unregister(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/devlink/devlink.c
1766
int ice_devlink_register_params(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1768
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1770
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1789
value.vu32 = pf->msix.max;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1793
value.vu32 = pf->msix.min;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1798
value.vbool = test_bit(ICE_FLAG_RDMA_ENA, pf->flags);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1814
void ice_devlink_unregister_params(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
1816
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1817
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
1855
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1856
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1857
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
187
ice_info_pending_netlist_build(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
1944
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1945
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
1946
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
196
static void ice_info_cgu_fw_build(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
200
if (!ice_is_feature_supported(pf, ICE_F_CGU))
drivers/net/ethernet/intel/ice/devlink/devlink.c
2008
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2009
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2010
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
202
if (ice_aq_get_cgu_info(&pf->hw, &id, &cfg_ver, &fw_ver))
drivers/net/ethernet/intel/ice/devlink/devlink.c
2060
void ice_devlink_init_regions(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2062
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2063
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2066
nvm_size = pf->hw.flash.flash_size;
drivers/net/ethernet/intel/ice/devlink/devlink.c
2067
pf->nvm_region = devl_region_create(devlink, &ice_nvm_region_ops, 1,
drivers/net/ethernet/intel/ice/devlink/devlink.c
2069
if (IS_ERR(pf->nvm_region)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
207
static void ice_info_cgu_id(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2071
PTR_ERR(pf->nvm_region));
drivers/net/ethernet/intel/ice/devlink/devlink.c
2072
pf->nvm_region = NULL;
drivers/net/ethernet/intel/ice/devlink/devlink.c
2075
sram_size = pf->hw.flash.sr_words * 2u;
drivers/net/ethernet/intel/ice/devlink/devlink.c
2076
pf->sram_region = devl_region_create(devlink, &ice_sram_region_ops,
drivers/net/ethernet/intel/ice/devlink/devlink.c
2078
if (IS_ERR(pf->sram_region)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
2080
PTR_ERR(pf->sram_region));
drivers/net/ethernet/intel/ice/devlink/devlink.c
2081
pf->sram_region = NULL;
drivers/net/ethernet/intel/ice/devlink/devlink.c
2084
pf->devcaps_region = devl_region_create(devlink,
drivers/net/ethernet/intel/ice/devlink/devlink.c
2087
if (IS_ERR(pf->devcaps_region)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
2089
PTR_ERR(pf->devcaps_region));
drivers/net/ethernet/intel/ice/devlink/devlink.c
209
if (!ice_is_feature_supported(pf, ICE_F_CGU))
drivers/net/ethernet/intel/ice/devlink/devlink.c
2090
pf->devcaps_region = NULL;
drivers/net/ethernet/intel/ice/devlink/devlink.c
2100
void ice_devlink_destroy_regions(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2102
if (pf->nvm_region)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2103
devl_region_destroy(pf->nvm_region);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2105
if (pf->sram_region)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2106
devl_region_destroy(pf->sram_region);
drivers/net/ethernet/intel/ice/devlink/devlink.c
2108
if (pf->devcaps_region)
drivers/net/ethernet/intel/ice/devlink/devlink.c
2109
devl_region_destroy(pf->devcaps_region);
drivers/net/ethernet/intel/ice/devlink/devlink.c
211
snprintf(ctx->buf, sizeof(ctx->buf), "%u", pf->hw.cgu_part_number);
drivers/net/ethernet/intel/ice/devlink/devlink.c
241
void (*getter)(struct ice_pf *pf, struct ice_info_ctx *ctx);
drivers/net/ethernet/intel/ice/devlink/devlink.c
242
void (*fallback)(struct ice_pf *pf, struct ice_info_ctx *ctx);
drivers/net/ethernet/intel/ice/devlink/devlink.c
275
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
276
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
277
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
282
err = ice_wait_for_reset(pf, 10 * HZ);
drivers/net/ethernet/intel/ice/devlink/devlink.c
334
ice_info_get_dsn(pf, ctx);
drivers/net/ethernet/intel/ice/devlink/devlink.c
34
static void ice_info_get_dsn(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
348
ice_devlink_versions[i].getter(pf, ctx);
drivers/net/ethernet/intel/ice/devlink/devlink.c
356
ice_devlink_versions[i].fallback(pf, ctx);
drivers/net/ethernet/intel/ice/devlink/devlink.c
39
put_unaligned_be64(pci_get_dsn(pf->pdev), dsn);
drivers/net/ethernet/intel/ice/devlink/devlink.c
411
ice_devlink_reload_empr_start(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
414
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
415
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
419
err = ice_get_pending_updates(pf, &pending, extack);
drivers/net/ethernet/intel/ice/devlink/devlink.c
433
if (pf->fw_emp_reset_disabled) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
44
static void ice_info_pba(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
455
static void ice_devlink_reinit_down(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
458
ice_unload(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
46
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
460
ice_vsi_decfg(ice_get_main_vsi(pf));
drivers/net/ethernet/intel/ice/devlink/devlink.c
462
ice_deinit_pf(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
463
ice_deinit_hw(&pf->hw);
drivers/net/ethernet/intel/ice/devlink/devlink.c
464
ice_deinit_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
481
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
485
if (ice_is_eswitch_mode_switchdev(pf)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
490
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
495
if (ice_has_vfs(pf)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
500
ice_devlink_reinit_down(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
503
return ice_devlink_reload_empr_start(pf, extack);
drivers/net/ethernet/intel/ice/devlink/devlink.c
52
dev_dbg(ice_pf_to_dev(pf), "Failed to read Product Board Assembly string, status %d\n",
drivers/net/ethernet/intel/ice/devlink/devlink.c
520
ice_devlink_reload_empr_finish(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
525
err = ice_wait_for_reset(pf, 60 * HZ);
drivers/net/ethernet/intel/ice/devlink/devlink.c
543
static int ice_get_tx_topo_user_sel(struct ice_pf *pf, uint8_t *layers)
drivers/net/ethernet/intel/ice/devlink/devlink.c
546
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
56
static void ice_info_fw_mgmt(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
58
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
580
static int ice_update_tx_topo_user_sel(struct ice_pf *pf, int layers)
drivers/net/ethernet/intel/ice/devlink/devlink.c
583
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
600
err = ice_write_one_nvm_block(pf, ICE_AQC_NVM_TX_TOPO_MOD_ID, 2,
drivers/net/ethernet/intel/ice/devlink/devlink.c
622
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
625
err = ice_get_tx_topo_user_sel(pf, &ctx->val.vu8);
drivers/net/ethernet/intel/ice/devlink/devlink.c
64
static void ice_info_fw_api(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
645
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
648
err = ice_update_tx_topo_user_sel(pf, ctx->val.vu8);
drivers/net/ethernet/intel/ice/devlink/devlink.c
66
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
692
void ice_tear_down_devlink_rate_tree(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
698
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
701
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/devlink/devlink.c
702
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
706
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/devlink/devlink.c
719
static bool ice_enable_custom_tx(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
72
static void ice_info_fw_build(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
721
struct ice_port_info *pi = ice_get_main_vsi(pf)->port_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
722
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
728
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
733
if (ice_is_dcb_active(pf)) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
74
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/devlink/devlink.c
754
struct ice_sched_node *tc_node, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/devlink.c
769
pf->vsi[node->vsi_handle]->type == ICE_VSI_VF &&
drivers/net/ethernet/intel/ice/devlink/devlink.c
770
pf->vsi[node->vsi_handle]->vf) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
771
vf = pf->vsi[node->vsi_handle]->vf;
drivers/net/ethernet/intel/ice/devlink/devlink.c
779
pf->vsi[node->vsi_handle]->type == ICE_VSI_SF &&
drivers/net/ethernet/intel/ice/devlink/devlink.c
780
pf->vsi[node->vsi_handle]->sf) {
drivers/net/ethernet/intel/ice/devlink/devlink.c
781
sf = pf->vsi[node->vsi_handle]->sf;
drivers/net/ethernet/intel/ice/devlink/devlink.c
79
static void ice_info_orom_ver(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.c
799
ice_traverse_tx_tree(devlink, node->children[i], tc_node, pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
81
struct ice_orom_info *orom = &pf->hw.flash.orom;
drivers/net/ethernet/intel/ice/devlink/devlink.c
814
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/devlink/devlink.c
820
ice_traverse_tx_tree(devlink, tc_node->children[i], tc_node, pf);
drivers/net/ethernet/intel/ice/devlink/devlink.c
88
ice_info_pending_orom_ver(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/devlink/devlink.c
970
struct ice_pf *pf = devlink_priv(rate_node->devlink);
drivers/net/ethernet/intel/ice/devlink/devlink.c
972
return ice_get_main_vsi(pf)->port_info;
drivers/net/ethernet/intel/ice/devlink/devlink.c
98
static void ice_info_nvm_ver(struct ice_pf *pf, struct ice_info_ctx *ctx)
drivers/net/ethernet/intel/ice/devlink/devlink.h
10
void ice_devlink_register(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
11
void ice_devlink_unregister(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
12
int ice_devlink_register_params(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
13
void ice_devlink_unregister_params(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
14
int ice_devlink_create_pf_port(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
15
void ice_devlink_destroy_pf_port(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
19
void ice_devlink_init_regions(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
20
void ice_devlink_destroy_regions(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
23
void ice_tear_down_devlink_rate_tree(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/devlink.h
8
struct ice_sf_priv *ice_allocate_sf(struct device *dev, struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/health.c
158
struct ice_pf *pf = devlink_health_reporter_priv(reporter);
drivers/net/ethernet/intel/ice/devlink/health.c
160
ice_describe_status_code(fmsg, &pf->health_reporters.port_status);
drivers/net/ethernet/intel/ice/devlink/health.c
168
struct ice_pf *pf = devlink_health_reporter_priv(reporter);
drivers/net/ethernet/intel/ice/devlink/health.c
170
ice_describe_status_code(fmsg, &pf->health_reporters.port_status);
drivers/net/ethernet/intel/ice/devlink/health.c
178
struct ice_pf *pf = devlink_health_reporter_priv(reporter);
drivers/net/ethernet/intel/ice/devlink/health.c
180
ice_describe_status_code(fmsg, &pf->health_reporters.fw_status);
drivers/net/ethernet/intel/ice/devlink/health.c
188
struct ice_pf *pf = devlink_health_reporter_priv(reporter);
drivers/net/ethernet/intel/ice/devlink/health.c
190
ice_describe_status_code(fmsg, &pf->health_reporters.fw_status);
drivers/net/ethernet/intel/ice/devlink/health.c
194
static void ice_config_health_events(struct ice_pf *pf, bool enable)
drivers/net/ethernet/intel/ice/devlink/health.c
203
ret = ice_aq_set_health_status_cfg(&pf->hw, enable_bits);
drivers/net/ethernet/intel/ice/devlink/health.c
205
dev_err(ice_pf_to_dev(pf), "Failed to %s firmware health events, err %d aq_err %s\n",
drivers/net/ethernet/intel/ice/devlink/health.c
207
libie_aq_str(pf->hw.adminq.sq_last_status));
drivers/net/ethernet/intel/ice/devlink/health.c
217
void ice_process_health_status_event(struct ice_pf *pf, struct ice_rq_event_info *event)
drivers/net/ethernet/intel/ice/devlink/health.c
228
dev_err(ice_pf_to_dev(pf), "Received a health status event with invalid element count\n");
drivers/net/ethernet/intel/ice/devlink/health.c
242
pf->health_reporters.fw_status = *health_info;
drivers/net/ethernet/intel/ice/devlink/health.c
243
devlink_health_report(pf->health_reporters.fw,
drivers/net/ethernet/intel/ice/devlink/health.c
248
pf->health_reporters.port_status = *health_info;
drivers/net/ethernet/intel/ice/devlink/health.c
249
devlink_health_report(pf->health_reporters.port,
drivers/net/ethernet/intel/ice/devlink/health.c
253
dev_err(ice_pf_to_dev(pf), "Health code with unknown source\n");
drivers/net/ethernet/intel/ice/devlink/health.c
262
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/devlink/health.c
348
void ice_report_mdd_event(struct ice_pf *pf, enum ice_mdd_src src, u8 pf_num,
drivers/net/ethernet/intel/ice/devlink/health.c
359
ice_devlink_health_report(pf->health_reporters.mdd, "MDD event", &ev);
drivers/net/ethernet/intel/ice/devlink/health.c
418
void ice_prep_tx_hang_report(struct ice_pf *pf, struct ice_tx_ring *tx_ring,
drivers/net/ethernet/intel/ice/devlink/health.c
421
struct ice_health_tx_hang_buf *buf = &pf->health_reporters.tx_hang_buf;
drivers/net/ethernet/intel/ice/devlink/health.c
429
void ice_report_tx_hang(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/health.c
431
struct ice_health_tx_hang_buf *buf = &pf->health_reporters.tx_hang_buf;
drivers/net/ethernet/intel/ice/devlink/health.c
444
ice_devlink_health_report(pf->health_reporters.tx_hang, "Tx hang", &ev);
drivers/net/ethernet/intel/ice/devlink/health.c
448
ice_init_devlink_rep(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/health.c
451
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/health.c
454
rep = devl_health_reporter_create(devlink, ops, pf);
drivers/net/ethernet/intel/ice/devlink/health.c
456
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/health.c
492
void ice_health_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/health.c
494
struct ice_health *reps = &pf->health_reporters;
drivers/net/ethernet/intel/ice/devlink/health.c
496
reps->mdd = ice_init_devlink_rep(pf, &ice_mdd_reporter_ops);
drivers/net/ethernet/intel/ice/devlink/health.c
497
reps->tx_hang = ice_init_devlink_rep(pf, &ice_tx_hang_reporter_ops);
drivers/net/ethernet/intel/ice/devlink/health.c
499
if (ice_is_fw_health_report_supported(&pf->hw)) {
drivers/net/ethernet/intel/ice/devlink/health.c
500
reps->fw = ice_init_devlink_rep(pf, &ice_fw_reporter_ops);
drivers/net/ethernet/intel/ice/devlink/health.c
501
reps->port = ice_init_devlink_rep(pf, &ice_port_reporter_ops);
drivers/net/ethernet/intel/ice/devlink/health.c
502
ice_config_health_events(pf, true);
drivers/net/ethernet/intel/ice/devlink/health.c
522
void ice_health_deinit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/health.c
524
ice_deinit_devl_reporter(pf->health_reporters.mdd);
drivers/net/ethernet/intel/ice/devlink/health.c
525
ice_deinit_devl_reporter(pf->health_reporters.tx_hang);
drivers/net/ethernet/intel/ice/devlink/health.c
526
if (ice_is_fw_health_report_supported(&pf->hw)) {
drivers/net/ethernet/intel/ice/devlink/health.c
527
ice_deinit_devl_reporter(pf->health_reporters.fw);
drivers/net/ethernet/intel/ice/devlink/health.c
528
ice_deinit_devl_reporter(pf->health_reporters.port);
drivers/net/ethernet/intel/ice/devlink/health.c
529
ice_config_health_events(pf, false);
drivers/net/ethernet/intel/ice/devlink/health.c
547
void ice_health_clear(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/health.c
549
ice_health_assign_healthy_state(pf->health_reporters.mdd);
drivers/net/ethernet/intel/ice/devlink/health.c
550
ice_health_assign_healthy_state(pf->health_reporters.tx_hang);
drivers/net/ethernet/intel/ice/devlink/health.h
58
void ice_process_health_status_event(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/health.h
61
void ice_health_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/health.h
62
void ice_health_deinit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/health.h
63
void ice_health_clear(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/health.h
65
void ice_prep_tx_hang_report(struct ice_pf *pf, struct ice_tx_ring *tx_ring,
drivers/net/ethernet/intel/ice/devlink/health.h
67
void ice_report_mdd_event(struct ice_pf *pf, enum ice_mdd_src src, u8 pf_num,
drivers/net/ethernet/intel/ice/devlink/health.h
69
void ice_report_tx_hang(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/port.c
1000
return ice_alloc_dynamic_port(pf, new_attr, extack, devlink_port);
drivers/net/ethernet/intel/ice/devlink/port.c
126
ice_devlink_aq_set_port_option(struct ice_pf *pf, u8 option_idx,
drivers/net/ethernet/intel/ice/devlink/port.c
129
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
132
status = ice_aq_set_port_option(&pf->hw, 0, true, option_idx);
drivers/net/ethernet/intel/ice/devlink/port.c
135
status, pf->hw.adminq.sq_last_status);
drivers/net/ethernet/intel/ice/devlink/port.c
140
status = ice_acquire_nvm(&pf->hw, ICE_RES_WRITE);
drivers/net/ethernet/intel/ice/devlink/port.c
143
status, pf->hw.adminq.sq_last_status);
drivers/net/ethernet/intel/ice/devlink/port.c
148
status = ice_nvm_write_activate(&pf->hw, ICE_AQC_NVM_ACTIV_REQ_EMPR, NULL);
drivers/net/ethernet/intel/ice/devlink/port.c
151
status, pf->hw.adminq.sq_last_status);
drivers/net/ethernet/intel/ice/devlink/port.c
153
ice_release_nvm(&pf->hw);
drivers/net/ethernet/intel/ice/devlink/port.c
157
ice_release_nvm(&pf->hw);
drivers/net/ethernet/intel/ice/devlink/port.c
187
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/port.c
189
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
193
status = ice_aq_get_port_options(&pf->hw, options, &option_count,
drivers/net/ethernet/intel/ice/devlink/port.c
222
ice_devlink_port_options_print(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
229
ice_devlink_port_options_print(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
233
status = ice_devlink_aq_set_port_option(pf, new_option, extack);
drivers/net/ethernet/intel/ice/devlink/port.c
237
ice_devlink_port_options_print(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
269
ice_devlink_set_port_split_options(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/port.c
277
status = ice_aq_get_port_options(&pf->hw, options, &option_count,
drivers/net/ethernet/intel/ice/devlink/port.c
281
dev_dbg(ice_pf_to_dev(pf), "Couldn't read port split options, err = %d\n",
drivers/net/ethernet/intel/ice/devlink/port.c
305
ice_devlink_set_switch_id(struct ice_pf *pf, struct netdev_phys_item_id *ppid)
drivers/net/ethernet/intel/ice/devlink/port.c
307
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/intel/ice/devlink/port.c
325
int ice_devlink_create_pf_port(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/port.c
334
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
336
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
338
devlink_port = &pf->devlink_port;
drivers/net/ethernet/intel/ice/devlink/port.c
340
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
345
attrs.phys.port_number = pf->hw.pf_id;
drivers/net/ethernet/intel/ice/devlink/port.c
350
if (pf->hw.pf_id == 0)
drivers/net/ethernet/intel/ice/devlink/port.c
351
ice_devlink_set_port_split_options(pf, &attrs);
drivers/net/ethernet/intel/ice/devlink/port.c
353
ice_devlink_set_switch_id(pf, &attrs.switch_id);
drivers/net/ethernet/intel/ice/devlink/port.c
361
pf->hw.pf_id, err);
drivers/net/ethernet/intel/ice/devlink/port.c
375
void ice_devlink_destroy_pf_port(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/port.c
377
devl_port_unregister(&pf->devlink_port);
drivers/net/ethernet/intel/ice/devlink/port.c
421
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/devlink/port.c
424
pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/port.c
428
return __ice_set_vf_mac(pf, vf_id, hw_addr);
drivers/net/ethernet/intel/ice/devlink/port.c
451
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/devlink/port.c
454
pf = vf->pf;
drivers/net/ethernet/intel/ice/devlink/port.c
455
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
463
attrs.pci_vf.pf = pf->hw.pf_id;
drivers/net/ethernet/intel/ice/devlink/port.c
466
ice_devlink_set_switch_id(pf, &attrs.switch_id);
drivers/net/ethernet/intel/ice/devlink/port.c
469
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
51
static void ice_devlink_port_options_print(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/port.c
55
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
587
struct ice_pf *pf = dyn_port->pf;
drivers/net/ethernet/intel/ice/devlink/port.c
591
xa_erase(&pf->sf_nums, devlink_port->attrs.pci_sf.sf);
drivers/net/ethernet/intel/ice/devlink/port.c
592
ice_eswitch_detach_sf(pf, dyn_port);
drivers/net/ethernet/intel/ice/devlink/port.c
594
xa_erase(&pf->dyn_ports, dyn_port->vsi->idx);
drivers/net/ethernet/intel/ice/devlink/port.c
602
void ice_dealloc_all_dynamic_ports(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/devlink/port.c
607
xa_for_each(&pf->dyn_ports, index, dyn_port)
drivers/net/ethernet/intel/ice/devlink/port.c
623
ice_devlink_port_new_check_attr(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/port.c
642
if (new_attr->pfnum != pf->hw.pf_id) {
drivers/net/ethernet/intel/ice/devlink/port.c
647
if (!pci_msix_can_alloc_dyn(pf->pdev)) {
drivers/net/ethernet/intel/ice/devlink/port.c
71
status = ice_aq_get_port_options(&pf->hw, opt, &options_count,
drivers/net/ethernet/intel/ice/devlink/port.c
827
ice_reserve_sf_num(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/port.c
835
return xa_alloc(&pf->sf_nums, sfnum, NULL, xa_limit_32b,
drivers/net/ethernet/intel/ice/devlink/port.c
839
err = xa_insert(&pf->sf_nums, new_attr->sfnum, NULL, GFP_KERNEL);
drivers/net/ethernet/intel/ice/devlink/port.c
866
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/devlink/port.c
869
pf = dyn_port->pf;
drivers/net/ethernet/intel/ice/devlink/port.c
874
attrs.pci_sf.pf = pf->hw.pf_id;
drivers/net/ethernet/intel/ice/devlink/port.c
878
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
909
ice_alloc_dynamic_port(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/devlink/port.c
919
err = ice_reserve_sf_num(pf, new_attr, extack, &sfnum);
drivers/net/ethernet/intel/ice/devlink/port.c
929
vsi = ice_vsi_alloc(pf);
drivers/net/ethernet/intel/ice/devlink/port.c
937
dyn_port->pf = pf;
drivers/net/ethernet/intel/ice/devlink/port.c
941
err = xa_insert(&pf->dyn_ports, vsi->idx, dyn_port, GFP_KERNEL);
drivers/net/ethernet/intel/ice/devlink/port.c
947
err = ice_eswitch_attach_sf(pf, dyn_port);
drivers/net/ethernet/intel/ice/devlink/port.c
958
xa_erase(&pf->dyn_ports, vsi->idx);
drivers/net/ethernet/intel/ice/devlink/port.c
964
xa_erase(&pf->sf_nums, sfnum);
drivers/net/ethernet/intel/ice/devlink/port.c
987
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/devlink/port.c
990
err = ice_devlink_port_new_check_attr(pf, new_attr, extack);
drivers/net/ethernet/intel/ice/devlink/port.c
994
if (!ice_is_eswitch_mode_switchdev(pf)) {
drivers/net/ethernet/intel/ice/devlink/port.h
29
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/devlink/port.h
39
void ice_dealloc_all_dynamic_ports(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/port.h
41
int ice_devlink_create_pf_port(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/devlink/port.h
42
void ice_devlink_destroy_pf_port(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1008
int ice_schedule_reset(struct ice_pf *pf, enum ice_reset_req reset);
drivers/net/ethernet/intel/ice/ice.h
1010
int ice_plug_aux_dev(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1011
void ice_unplug_aux_dev(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1012
void ice_rdma_finalize_setup(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1013
int ice_init_rdma(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1014
void ice_deinit_rdma(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1018
ice_fdir_write_fltr(struct ice_pf *pf, struct ice_fdir_fltr *input, bool add,
drivers/net/ethernet/intel/ice/ice.h
1030
void ice_fdir_replay_fltrs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1031
int ice_fdir_create_dflt_rules(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1047
void ice_aq_prep_for_event(struct ice_pf *pf, struct ice_aq_task *task,
drivers/net/ethernet/intel/ice/ice.h
1049
int ice_aq_wait_for_event(struct ice_pf *pf, struct ice_aq_task *task,
drivers/net/ethernet/intel/ice/ice.h
1054
void ice_service_task_schedule(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1055
void ice_start_service_task(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1056
int ice_load(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1057
void ice_unload(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1059
void ice_init_dev_hw(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1060
int ice_init_dev(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1061
void ice_deinit_dev(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1062
int ice_init_pf(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1063
void ice_deinit_pf(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
1077
static inline void ice_set_rdma_cap(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
1079
if (pf->hw.func_caps.common_cap.rdma && pf->num_rdma_msix) {
drivers/net/ethernet/intel/ice/ice.h
1080
set_bit(ICE_FLAG_RDMA_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice.h
1081
set_bit(ICE_FLAG_PLUG_AUX_DEV, pf->flags);
drivers/net/ethernet/intel/ice/ice.h
1089
static inline void ice_clear_rdma_cap(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
1094
clear_bit(ICE_FLAG_PLUG_AUX_DEV, pf->flags);
drivers/net/ethernet/intel/ice/ice.h
1095
set_bit(ICE_FLAG_UNPLUG_AUX_DEV, pf->flags);
drivers/net/ethernet/intel/ice/ice.h
1096
clear_bit(ICE_FLAG_RDMA_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice.h
1138
static inline bool ice_pf_src_tmr_owned(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
1140
return pf->hw.func_caps.ts_func_info.src_tmr_owned &&
drivers/net/ethernet/intel/ice/ice.h
1141
ice_is_primary(&pf->hw);
drivers/net/ethernet/intel/ice/ice.h
1151
static inline struct ice_hw *ice_get_primary_hw(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
1153
if (!pf->adapter->ctrl_pf)
drivers/net/ethernet/intel/ice/ice.h
1154
return &pf->hw;
drivers/net/ethernet/intel/ice/ice.h
1156
return &pf->adapter->ctrl_pf->hw;
drivers/net/ethernet/intel/ice/ice.h
161
#define ice_for_each_vsi(pf, i) \
drivers/net/ethernet/intel/ice/ice.h
162
for ((i) = 0; (i) < (pf)->num_alloc_vsi; (i)++)
drivers/net/ethernet/intel/ice/ice.h
199
#define ice_pf_to_dev(pf) (&((pf)->pdev->dev))
drivers/net/ethernet/intel/ice/ice.h
267
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice.h
707
static inline bool ice_ptp_pf_handles_tx_interrupt(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
709
return pf->ptp.tx_interrupt_mode != ICE_PTP_TX_INTERRUPT_NONE;
drivers/net/ethernet/intel/ice/ice.h
768
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice.h
770
return test_bit(ring->q_index, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice.h
848
static inline int ice_get_max_txq(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
850
return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_txq);
drivers/net/ethernet/intel/ice/ice.h
859
static inline int ice_get_max_rxq(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
861
return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_rxq);
drivers/net/ethernet/intel/ice/ice.h
870
static inline struct ice_vsi *ice_get_main_vsi(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
872
if (pf->vsi)
drivers/net/ethernet/intel/ice/ice.h
873
return pf->vsi[0];
drivers/net/ethernet/intel/ice/ice.h
895
static inline struct ice_vsi *ice_get_ctrl_vsi(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
898
if (!pf->vsi || pf->ctrl_vsi_idx == ICE_NO_VSI)
drivers/net/ethernet/intel/ice/ice.h
901
return pf->vsi[pf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/ice.h
909
static inline struct ice_vsi *ice_find_vsi(struct ice_pf *pf, u16 vsi_num)
drivers/net/ethernet/intel/ice/ice.h
913
ice_for_each_vsi(pf, i)
drivers/net/ethernet/intel/ice/ice.h
914
if (pf->vsi[i] && pf->vsi[i]->vsi_num == vsi_num)
drivers/net/ethernet/intel/ice/ice.h
915
return pf->vsi[i];
drivers/net/ethernet/intel/ice/ice.h
926
static inline bool ice_is_switchdev_running(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
928
return pf->eswitch.is_running;
drivers/net/ethernet/intel/ice/ice.h
947
static inline bool ice_is_adq_active(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice.h
951
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice.h
957
test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice.h
963
int ice_debugfs_pf_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
964
void ice_debugfs_pf_deinit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
977
u16 ice_get_avail_txq_count(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
978
u16 ice_get_avail_rxq_count(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
981
void ice_update_pf_stats(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice.h
986
struct ice_vsi *ice_lb_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi);
drivers/net/ethernet/intel/ice/ice_arfs.c
263
void ice_sync_arfs_fltrs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_arfs.c
270
pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_arfs.c
28
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_arfs.c
31
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_arfs.c
448
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_arfs.c
458
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_arfs.c
479
if (!ice_arfs_is_perfect_flow_set(&pf->hw, n_proto, ip_proto))
drivers/net/ethernet/intel/ice/ice_arfs.c
522
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_arfs.c
625
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_arfs.c
630
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_arfs.c
632
if (!pf || !netdev || !vsi->num_q_vectors)
drivers/net/ethernet/intel/ice/ice_arfs.c
645
void ice_remove_arfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_arfs.c
649
pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_arfs.c
660
void ice_rebuild_arfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_arfs.c
664
pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_arfs.c
668
ice_remove_arfs(pf);
drivers/net/ethernet/intel/ice/ice_arfs.h
49
void ice_sync_arfs_fltrs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_arfs.h
51
void ice_remove_arfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_arfs.h
52
void ice_rebuild_arfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_arfs.h
59
static inline void ice_sync_arfs_fltrs(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_arfs.h
60
static inline void ice_remove_arfs(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_arfs.h
61
static inline void ice_rebuild_arfs(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_base.c
1011
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1012
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
1018
dev_err(ice_pf_to_dev(pf), "Failed to setup Tx time queue context for queue %d, error: %d\n",
drivers/net/ethernet/intel/ice/ice_base.c
1046
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1047
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
105
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1058
dev_err(ice_pf_to_dev(pf), "Failed to setup Tx context for queue %d, error: %d\n",
drivers/net/ethernet/intel/ice/ice_base.c
1091
dev_err(ice_pf_to_dev(pf), "Failed to set LAN Tx queue context, error: %d\n",
drivers/net/ethernet/intel/ice/ice_base.c
1107
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_base.c
1115
dev_err(ice_pf_to_dev(pf), "Failed to set Tx Time queue context, error: %d\n",
drivers/net/ethernet/intel/ice/ice_base.c
1240
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1241
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
1272
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1273
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
130
struct ice_vsi *ctrl_vsi = ice_get_vf_ctrl_vsi(pf, vsi);
drivers/net/ethernet/intel/ice/ice_base.c
1312
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1314
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
1397
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_base.c
1399
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
1400
if (!pf->vsi_stats)
drivers/net/ethernet/intel/ice/ice_base.c
1403
vsi_stat = pf->vsi_stats[vsi->idx];
drivers/net/ethernet/intel/ice/ice_base.c
143
q_vector->irq = ice_alloc_irq(pf, vsi->irq_dyn_alloc);
drivers/net/ethernet/intel/ice/ice_base.c
181
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
186
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_base.c
209
ice_get_vf_ctrl_vsi(pf, vsi))
drivers/net/ethernet/intel/ice/ice_base.c
212
ice_free_irq(pf, q_vector->irq);
drivers/net/ethernet/intel/ice/ice_base.c
80
static int ice_pf_rxq_wait(struct ice_pf *pf, int pf_q, bool ena)
drivers/net/ethernet/intel/ice/ice_base.c
843
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
844
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_base.c
85
if (ena == !!(rd32(&pf->hw, QRX_CTRL(pf_q)) &
drivers/net/ethernet/intel/ice/ice_base.c
864
return ice_pf_rxq_wait(pf, pf_q, ena);
drivers/net/ethernet/intel/ice/ice_base.c
881
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_base.c
883
return ice_pf_rxq_wait(pf, pf_q, ena);
drivers/net/ethernet/intel/ice/ice_common.c
1576
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_common.c
1593
spin_lock(&pf->adapter->txq_ctx_lock);
drivers/net/ethernet/intel/ice/ice_common.c
1605
spin_unlock(&pf->adapter->txq_ctx_lock);
drivers/net/ethernet/intel/ice/ice_common.c
1618
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_common.c
1634
spin_lock(&pf->adapter->txq_ctx_lock);
drivers/net/ethernet/intel/ice/ice_common.c
1648
spin_unlock(&pf->adapter->txq_ctx_lock);
drivers/net/ethernet/intel/ice/ice_common.c
962
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_common.c
964
.pdev = pf->pdev,
drivers/net/ethernet/intel/ice/ice_common.c
974
err = ice_debugfs_pf_init(pf);
drivers/net/ethernet/intel/ice/ice_common.c
978
api.debugfs_root = pf->ice_debugfs_pf;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1000
ice_dcb_process_lldp_set_mib_change(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1004
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1014
if (!(test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags)))
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1017
if (pf->dcbx_cap & DCB_CAP_DCBX_HOST) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1022
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1058
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1085
pf->dcbx_cap = ice_dcb_get_mode(pi, false);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1087
need_reconfig = ice_dcb_need_recfg(pf, &tmp_dcbx_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1089
ice_dcbnl_flush_apps(pf, &tmp_dcbx_cfg, &pi->qos_cfg.local_dcbx_cfg);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1096
set_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1099
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1104
ice_lldp_execute_pending_mib(&pf->hw);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1110
ice_dcb_ena_dis_vsi(pf, false, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1119
ice_pf_dcb_recfg(pf, false);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1122
ice_dcb_ena_dis_vsi(pf, true, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1126
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
1130
ice_lldp_execute_pending_mib(&pf->hw);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
283
static void ice_dcb_ena_dis_vsi(struct ice_pf *pf, bool ena, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
287
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
288
struct ice_vsi *vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
31
bool ice_is_pfc_causing_hung_q(struct ice_pf *pf, unsigned int txqueue)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
312
int ice_dcb_bwchk(struct ice_pf *pf, struct ice_dcbx_cfg *dcbcfg)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
337
dev_err(ice_pf_to_dev(pf), "Invalid config, total bandwidth must equal 100\n");
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
350
int ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
354
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
359
curr_cfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
362
if (!pf->hw.port_info->qos_cfg.is_sw_lldp)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
368
if (pf->hw.port_info->is_custom_tx_enabled) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
372
ice_tear_down_devlink_rate_tree(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
374
set_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
377
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
38
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
385
if (ice_dcb_bwchk(pf, new_cfg))
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
394
pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
409
ice_send_event_to_aux(pf, event);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
419
ice_dcb_ena_dis_vsi(pf, false, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
428
if (pf->hw.port_info->qos_cfg.is_sw_lldp) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
429
ret = ice_set_dcb_cfg(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
438
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
444
ice_pf_dcb_recfg(pf, false);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
448
ice_dcb_ena_dis_vsi(pf, true, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
485
ice_dcb_need_recfg(struct ice_pf *pf, struct ice_dcbx_cfg *old_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
488
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
533
void ice_dcb_rebuild(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
536
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
540
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
546
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
548
if (!pf->hw.port_info->qos_cfg.is_sw_lldp)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
549
ice_cfg_etsrec_defaults(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
551
ret = ice_set_dcb_cfg(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
557
if (!pf->hw.port_info->qos_cfg.is_sw_lldp) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
558
ret = ice_cfg_lldp_mib_change(&pf->hw, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
559
if (ret && !pf->hw.port_info->qos_cfg.is_sw_lldp) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
56
up2tc = rd32(&pf->hw, PRTDCB_TUP2TC);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
566
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
572
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
580
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
594
ice_pf_dcb_cfg(pf, err_cfg, false);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
597
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
605
static int ice_dcb_init_cfg(struct ice_pf *pf, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
611
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
619
dev_info(ice_pf_to_dev(pf), "Configuring initial DCB values\n");
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
620
if (ice_pf_dcb_cfg(pf, newcfg, locked))
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
634
int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
642
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
667
ret = ice_pf_dcb_cfg(pf, dcbcfg, locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
70
ref_prio_xoff[i] = pf->stats.priority_xoff_rx[i];
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
709
static int ice_dcb_noncontig_cfg(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
711
struct ice_dcbx_cfg *dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
712
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
716
ret = ice_dcb_sw_dflt_cfg(pf, false, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
72
ice_update_dcb_stats(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
724
ret = ice_set_dcb_cfg(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
740
void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
742
struct ice_dcbx_cfg *dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
750
ice_for_each_vsi(pf, v) {
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
751
struct ice_vsi *vsi = pf->vsi[v];
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
76
if (pf->stats.priority_xoff_rx[i] > ref_prio_xoff[i])
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
764
ice_dcb_noncontig_cfg(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
774
dev_err(ice_pf_to_dev(pf), "Failed to config TC for VSI index: %d\n",
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
789
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
792
ice_setup_dcb_qos_info(pf, &privd->qos_info);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
799
ice_send_event_to_aux(pf, event);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
809
int ice_init_pf_dcb(struct ice_pf *pf, bool locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
811
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
813
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
825
pf->hw.func_caps.common_cap.maxtc);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
831
clear_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
832
err = ice_aq_set_pfc_mode(&pf->hw, ICE_AQC_PFC_VLAN_BASED_PFC,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
837
err = ice_dcb_sw_dflt_cfg(pf, true, locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
848
pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
855
ice_cfg_sw_rx_lldp(pf, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
857
pf->dcbx_cap = ice_dcb_get_mode(port_info, true);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
861
set_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
864
pf->dcbx_cap = ice_dcb_get_mode(port_info, false);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
866
err = ice_dcb_init_cfg(pf, locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
881
void ice_update_dcb_stats(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
884
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
889
prev_ps = &pf->stats_prev;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
890
cur_ps = &pf->stats;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
892
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
893
pf->stat_prev_loaded = false;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
897
pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
901
pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
905
pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
909
pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
913
pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
958
void ice_setup_dcb_qos_info(struct ice_pf *pf, struct iidc_rdma_qos_params *qos_info)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
964
if (!pf || !qos_info)
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
967
dcbx_cfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_lib.c
968
up2tc = rd32(&pf->hw, PRTDCB_TUP2TC);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
103
ice_pf_dcb_cfg(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
117
static inline bool ice_is_dcb_active(struct ice_pf __always_unused *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
123
ice_is_pfc_causing_hung_q(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
129
static inline u8 ice_get_pfc_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
134
static inline void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked) { }
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
136
static inline void ice_update_dcb_stats(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
138
ice_dcb_process_lldp_set_mib_change(struct ice_pf *pf, struct ice_rq_event_info *event) { }
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
141
ice_setup_dcb_qos_info(struct ice_pf *pf, struct iidc_rdma_qos_params *qos_info)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
17
void ice_dcb_rebuild(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
18
int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
21
bool ice_is_pfc_causing_hung_q(struct ice_pf *pf, unsigned int txqueue);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
24
ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
25
int ice_dcb_bwchk(struct ice_pf *pf, struct ice_dcbx_cfg *dcbcfg);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
26
void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
28
int ice_init_pf_dcb(struct ice_pf *pf, bool locked);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
29
void ice_update_dcb_stats(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
34
ice_setup_dcb_qos_info(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
37
ice_dcb_process_lldp_set_mib_change(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
58
static inline bool ice_is_dcb_active(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
60
return (test_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags) ||
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
61
test_bit(ICE_FLAG_DCB_ENA, pf->flags));
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
64
static inline u8 ice_get_pfc_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
66
return pf->hw.port_info->qos_cfg.local_dcbx_cfg.pfc_mode;
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
70
static inline void ice_dcb_rebuild(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
96
ice_init_pf_dcb(struct ice_pf *pf, bool __always_unused locked)
drivers/net/ethernet/intel/ice/ice_dcb_lib.h
98
dev_dbg(ice_pf_to_dev(pf), "DCB not supported\n");
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
103
new_cfg->etscfg.maxtcs = pf->hw.func_caps.common_cap.maxtc;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1044
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1050
pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1051
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1054
if (pf->dcbx_cap & DCB_CAP_DCBX_HOST)
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1058
if (!test_bit(ICE_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
111
err = ice_pf_dcb_cfg(pf, new_cfg, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1113
ice_dcbnl_flush_apps(struct ice_pf *pf, struct ice_dcbx_cfg *old_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1116
struct ice_vsi *main_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1138
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1140
pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
1141
if (!test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
119
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
134
struct ice_pf *pf = ice_netdev_to_pf(dev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
136
if (!test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
139
*num = pf->hw.func_caps.common_cap.maxtc;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
149
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
151
return pf->dcbx_cap;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
16
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
161
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
165
if (test_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
175
if (mode == pf->dcbx_cap)
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
178
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
18
while (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
183
qos_cfg = &pf->hw.port_info->qos_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
189
pf->dcbx_cap = mode;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
196
dev_info(ice_pf_to_dev(pf), "DCBx mode = 0x%x\n", mode);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
207
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
208
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
240
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
241
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
249
ice_get_pfc_delay(&pf->hw, &pfc->delay);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
252
pfc->requests[i] = pf->stats.priority_xoff_tx[i];
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
253
pfc->indications[i] = pf->stats.priority_xoff_rx[i];
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
266
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
270
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
271
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
274
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
279
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
281
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
286
new_cfg->pfc.pfccap = pf->hw.func_caps.common_cap.maxtc;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
290
err = ice_pf_dcb_cfg(pf, new_cfg, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
295
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
308
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
309
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
311
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
312
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
319
dev_dbg(ice_pf_to_dev(pf), "Get PFC Config up=%d, setting=%d, pfcenable=0x%x\n",
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
331
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
334
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
335
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
341
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
346
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
348
new_cfg->pfc.pfccap = pf->hw.func_caps.common_cap.maxtc;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
35
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
354
dev_dbg(ice_pf_to_dev(pf), "Set PFC config UP:%d set:%d pfcena:0x%x\n",
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
364
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
365
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
37
pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
38
dcbxcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
380
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
383
state = test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
385
dev_dbg(ice_pf_to_dev(pf), "DCB enabled state = %d\n", state);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
396
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
398
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
399
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
402
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
408
if (!!state == test_bit(ICE_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
412
set_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
413
memcpy(&pf->hw.port_info->qos_cfg.desired_dcbx_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
414
&pf->hw.port_info->qos_cfg.local_dcbx_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
417
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
438
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
439
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
441
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
442
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
449
dev_dbg(ice_pf_to_dev(pf), "Get PG config prio=%d tc=%d\n", prio,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
468
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
472
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
473
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
479
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
484
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
504
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
505
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
507
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
508
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
515
dev_dbg(ice_pf_to_dev(pf), "Get PG BW config tc=%d bw_pct=%d\n",
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
528
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
531
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
532
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
538
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
543
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
563
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
564
struct ice_port_info *pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
566
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
567
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
595
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
597
dev_dbg(ice_pf_to_dev(pf), "Rx TC PG Config Not Supported.\n");
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
610
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
612
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
613
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
631
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
633
dev_dbg(ice_pf_to_dev(pf), "Rx BWG PG Config Not Supported.\n");
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
64
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
644
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
646
if (!(test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags)))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
672
*cap = pf->dcbx_cap;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
679
dev_dbg(ice_pf_to_dev(pf), "DCBX Get Capability cap=%d capval=0x%x\n",
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
69
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
692
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
698
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
699
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
70
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
73
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
735
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
746
if (pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
751
if (!(pf->dcbx_cap & DCB_CAP_DCBX_VER_IEEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
754
if (!ice_is_feature_supported(pf, ICE_F_DSCP))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
763
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
768
max_tc = pf->hw.func_caps.common_cap.maxtc;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
776
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
778
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
779
old_cfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
78
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
80
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
804
ret = ice_aq_set_pfc_mode(&pf->hw, ICE_AQC_PFC_DSCP_BASED_PFC,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
848
ret = ice_pf_dcb_cfg(pf, new_cfg, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
856
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
869
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
874
if (pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
879
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
884
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
885
old_cfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
891
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
920
!ice_is_feature_supported(pf, ICE_F_DSCP)) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
936
ret = ice_aq_set_pfc_mode(&pf->hw,
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
948
ret = ice_dcb_sw_dflt_cfg(pf, true, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
950
ret = ice_pf_dcb_cfg(pf, new_cfg, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
967
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
977
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
98
if (ice_dcb_bwchk(pf, new_cfg)) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
981
if ((pf->dcbx_cap & DCB_CAP_DCBX_LLD_MANAGED) ||
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
982
!(pf->dcbx_cap & DCB_CAP_DCBX_VER_CEE))
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
985
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
990
new_cfg = &pf->hw.port_info->qos_cfg.desired_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
992
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
994
err = ice_pf_dcb_cfg(pf, new_cfg, true);
drivers/net/ethernet/intel/ice/ice_dcb_nl.c
996
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_dcb_nl.h
11
ice_dcbnl_flush_apps(struct ice_pf *pf, struct ice_dcbx_cfg *old_cfg,
drivers/net/ethernet/intel/ice/ice_dcb_nl.h
17
ice_dcbnl_flush_apps(struct ice_pf *pf, struct ice_dcbx_cfg *old_cfg,
drivers/net/ethernet/intel/ice/ice_debugfs.c
11
const char *name = pci_name(pf->pdev);
drivers/net/ethernet/intel/ice/ice_debugfs.c
13
pf->ice_debugfs_pf = debugfs_create_dir(name, ice_debugfs_root);
drivers/net/ethernet/intel/ice/ice_debugfs.c
14
if (IS_ERR(pf->ice_debugfs_pf))
drivers/net/ethernet/intel/ice/ice_debugfs.c
15
return PTR_ERR(pf->ice_debugfs_pf);
drivers/net/ethernet/intel/ice/ice_debugfs.c
24
void ice_debugfs_pf_deinit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_debugfs.c
26
debugfs_remove_recursive(pf->ice_debugfs_pf);
drivers/net/ethernet/intel/ice/ice_debugfs.c
27
pf->ice_debugfs_pf = NULL;
drivers/net/ethernet/intel/ice/ice_debugfs.c
9
int ice_debugfs_pf_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
101
if (input && pf->hw.device_id == ICE_DEV_ID_E810C_QSFP)
drivers/net/ethernet/intel/ice/ice_dpll.c
1085
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1088
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1091
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1092
ret = ice_dpll_pin_state_update(pf, p, pin_type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1100
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1179
ret = ice_read_sma_ctrl(&p->pf->hw, &data);
drivers/net/ethernet/intel/ice/ice_dpll.c
1200
ret = ice_write_sma_ctrl(&p->pf->hw, data);
drivers/net/ethernet/intel/ice/ice_dpll.c
1202
ret = ice_dpll_pin_state_update(p->pf, p,
drivers/net/ethernet/intel/ice/ice_dpll.c
1234
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
124
static bool ice_dpll_is_reset(struct ice_pf *pf, struct netlink_ext_ack *extack)
drivers/net/ethernet/intel/ice/ice_dpll.c
1240
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1243
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1244
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dpll.c
126
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_dpll.c
1285
ret = ice_dpll_pin_state_update(pf, p, ICE_DPLL_PIN_TYPE_SOFTWARE,
drivers/net/ethernet/intel/ice/ice_dpll.c
1295
ret = ice_dpll_pin_state_update(pf, target, type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1298
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1327
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1330
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1332
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1339
ret = ice_dpll_pin_state_update(pf, p->input,
drivers/net/ethernet/intel/ice/ice_dpll.c
1346
ret = ice_dpll_pin_state_update(pf, p->output,
drivers/net/ethernet/intel/ice/ice_dpll.c
1354
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1383
struct ice_pf *pf = sma->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1388
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1391
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1408
ret = ice_dpll_pin_enable(&pf->hw, target, d->dpll_idx, type,
drivers/net/ethernet/intel/ice/ice_dpll.c
1411
ret = ice_dpll_pin_disable(&pf->hw, target, type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1413
ret = ice_dpll_pin_state_update(pf, target, type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1416
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1444
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1446
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1448
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1476
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1479
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1482
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1483
ret = ice_dpll_hw_input_prio_set(pf, d, p, prio, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1484
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
149
ice_dpll_pin_freq_set(struct ice_pf *pf, struct ice_dpll_pin *pin,
drivers/net/ethernet/intel/ice/ice_dpll.c
1496
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1498
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1503
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1515
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1520
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1523
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1524
ret = ice_dpll_hw_input_prio_set(pf, d, p->input, prio, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
1525
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
159
ret = ice_aq_set_input_pin_cfg(&pf->hw, pin->idx, flags,
drivers/net/ethernet/intel/ice/ice_dpll.c
1603
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1606
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1609
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1611
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1639
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
164
ret = ice_aq_set_output_pin_cfg(&pf->hw, pin->idx, flags,
drivers/net/ethernet/intel/ice/ice_dpll.c
1641
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1643
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1645
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1673
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1675
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1677
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1709
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1713
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1716
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1724
ret = ice_aq_set_input_pin_cfg(&pf->hw, p->idx, flag, flags_en,
drivers/net/ethernet/intel/ice/ice_dpll.c
1733
ret = ice_aq_set_output_pin_cfg(&pf->hw, p->idx, flag, 0, 0,
drivers/net/ethernet/intel/ice/ice_dpll.c
174
libie_aq_str(pf->hw.adminq.sq_last_status),
drivers/net/ethernet/intel/ice/ice_dpll.c
1741
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1746
libie_aq_str(pf->hw.adminq.sq_last_status),
drivers/net/ethernet/intel/ice/ice_dpll.c
1911
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1913
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1921
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1984
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
1988
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
1990
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
1998
ret = ice_aq_set_output_pin_cfg(&pf->hw, p->idx, flags,
drivers/net/ethernet/intel/ice/ice_dpll.c
2006
ret = ice_aq_set_output_pin_cfg(&pf->hw, p->idx, flags,
drivers/net/ethernet/intel/ice/ice_dpll.c
2010
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2040
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2042
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2044
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2047
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2059
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2088
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
209
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2092
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2094
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2102
ret = ice_aq_set_input_pin_cfg(&pf->hw, p->idx, 0,
drivers/net/ethernet/intel/ice/ice_dpll.c
2110
ret = ice_aq_set_input_pin_cfg(&pf->hw, p->idx, 0,
drivers/net/ethernet/intel/ice/ice_dpll.c
2114
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
212
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2144
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2146
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2148
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
215
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2151
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
216
ret = ice_dpll_pin_freq_set(pf, p, pin_type, frequency, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
2163
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
217
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2262
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2266
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2268
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2274
ret = ice_aq_set_input_pin_cfg(&pf->hw, p->idx, 0, flags_en, 0, 0);
drivers/net/ethernet/intel/ice/ice_dpll.c
2276
ret = ice_dpll_pin_state_update(pf, p, ICE_DPLL_PIN_TYPE_INPUT,
drivers/net/ethernet/intel/ice/ice_dpll.c
2278
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2307
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2309
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2311
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2316
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2388
if (pin->pf->dplls.inputs[pin->parent_idx[i]].pin == parent)
drivers/net/ethernet/intel/ice/ice_dpll.c
2428
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2433
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dpll.c
2435
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2438
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2442
hw_idx -= pf->dplls.base_rclk_idx;
drivers/net/ethernet/intel/ice/ice_dpll.c
2455
pf->ptp.port.port_num,
drivers/net/ethernet/intel/ice/ice_dpll.c
2466
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2495
struct ice_pf *pf = p->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2499
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
2502
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2507
ret = ice_dpll_pin_state_update(pf, p, ICE_DPLL_PIN_TYPE_RCLK_INPUT,
drivers/net/ethernet/intel/ice/ice_dpll.c
2515
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2607
static u64 ice_generate_clock_id(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
2609
return pci_get_dsn(pf->pdev);
drivers/net/ethernet/intel/ice/ice_dpll.c
2653
static bool ice_dpll_is_pps_phase_monitor(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
2656
int ret = ice_aq_get_cgu_input_pin_measure(&pf->hw, DPLL_TYPE_PPS, meas,
drivers/net/ethernet/intel/ice/ice_dpll.c
2659
if (ret && pf->hw.adminq.sq_last_status == LIBIE_AQ_RC_ESRCH)
drivers/net/ethernet/intel/ice/ice_dpll.c
2701
static int ice_dpll_pps_update_phase_offsets(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dpll.c
2710
ret = ice_aq_get_cgu_input_pin_measure(&pf->hw, DPLL_TYPE_PPS, meas,
drivers/net/ethernet/intel/ice/ice_dpll.c
2712
if (ret && pf->hw.adminq.sq_last_status == LIBIE_AQ_RC_EAGAIN) {
drivers/net/ethernet/intel/ice/ice_dpll.c
2715
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
2718
libie_aq_str(pf->hw.adminq.sq_last_status));
drivers/net/ethernet/intel/ice/ice_dpll.c
2721
for (i = 0; i < pf->dplls.num_inputs; i++) {
drivers/net/ethernet/intel/ice/ice_dpll.c
2722
p = &pf->dplls.inputs[i];
drivers/net/ethernet/intel/ice/ice_dpll.c
2735
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
2760
ice_dpll_update_state(struct ice_pf *pf, struct ice_dpll *d, bool init)
drivers/net/ethernet/intel/ice/ice_dpll.c
2765
ret = ice_get_cgu_state(&pf->hw, d->dpll_idx, d->prev_dpll_state,
drivers/net/ethernet/intel/ice/ice_dpll.c
2769
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
2774
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
2777
libie_aq_str(pf->hw.adminq.sq_last_status));
drivers/net/ethernet/intel/ice/ice_dpll.c
2783
d->active_input = pf->dplls.inputs[d->input_idx].pin;
drivers/net/ethernet/intel/ice/ice_dpll.c
2784
p = &pf->dplls.inputs[d->input_idx];
drivers/net/ethernet/intel/ice/ice_dpll.c
2785
return ice_dpll_pin_state_update(pf, p,
drivers/net/ethernet/intel/ice/ice_dpll.c
2792
p = &pf->dplls.inputs[d->input_idx];
drivers/net/ethernet/intel/ice/ice_dpll.c
2797
ret = ice_dpll_pin_state_update(pf, p,
drivers/net/ethernet/intel/ice/ice_dpll.c
2801
p = &pf->dplls.inputs[d->prev_input_idx];
drivers/net/ethernet/intel/ice/ice_dpll.c
2802
ice_dpll_pin_state_update(pf, p,
drivers/net/ethernet/intel/ice/ice_dpll.c
2807
p = &pf->dplls.inputs[d->input_idx];
drivers/net/ethernet/intel/ice/ice_dpll.c
2809
ice_dpll_pin_state_update(pf, p,
drivers/net/ethernet/intel/ice/ice_dpll.c
2829
struct ice_pf *pf = container_of(d, struct ice_pf, dplls);
drivers/net/ethernet/intel/ice/ice_dpll.c
2830
struct ice_dpll *de = &pf->dplls.eec;
drivers/net/ethernet/intel/ice/ice_dpll.c
2831
struct ice_dpll *dp = &pf->dplls.pps;
drivers/net/ethernet/intel/ice/ice_dpll.c
2835
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_dpll.c
2837
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2839
ret = ice_dpll_update_state(pf, de, false);
drivers/net/ethernet/intel/ice/ice_dpll.c
2841
ret = ice_dpll_update_state(pf, dp, false);
drivers/net/ethernet/intel/ice/ice_dpll.c
2844
ret = ice_dpll_pps_update_phase_offsets(pf, &phase_offset_ntf);
drivers/net/ethernet/intel/ice/ice_dpll.c
2850
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
2852
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2856
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
2881
static int ice_dpll_init_ref_sync_inputs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
2883
struct ice_dpll_pin *inputs = pf->dplls.inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
2884
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dpll.c
2890
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
297
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
2984
ice_dpll_get_pins(struct ice_pf *pf, struct ice_dpll_pin *pins,
drivers/net/ethernet/intel/ice/ice_dpll.c
299
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
301
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
3111
ice_dpll_deinit_direct_pins(struct ice_pf *pf, bool cgu,
drivers/net/ethernet/intel/ice/ice_dpll.c
3143
ice_dpll_init_direct_pins(struct ice_pf *pf, bool cgu,
drivers/net/ethernet/intel/ice/ice_dpll.c
3150
ret = ice_dpll_get_pins(pf, pins, start_idx, count, pf->dplls.clock_id);
drivers/net/ethernet/intel/ice/ice_dpll.c
3177
static void ice_dpll_deinit_rclk_pin(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3179
struct ice_dpll_pin *rclk = &pf->dplls.rclk;
drivers/net/ethernet/intel/ice/ice_dpll.c
3180
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
3185
parent = &pf->dplls.inputs[rclk->parent_idx[i]];
drivers/net/ethernet/intel/ice/ice_dpll.c
3208
struct ice_pf *pf = parent->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3211
wait_for_completion(&pf->dplls.dpll_init);
drivers/net/ethernet/intel/ice/ice_dpll.c
3212
if (!test_bit(ICE_FLAG_DPLL, pf->flags))
drivers/net/ethernet/intel/ice/ice_dpll.c
3226
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3232
pin = &pf->dplls.rclk;
drivers/net/ethernet/intel/ice/ice_dpll.c
3236
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3250
pin = &pf->dplls.rclk;
drivers/net/ethernet/intel/ice/ice_dpll.c
3287
queue_work(pin->pf->dplls.wq, &work->work);
drivers/net/ethernet/intel/ice/ice_dpll.c
3307
ice_dpll_init_pin_common(struct ice_pf *pf, struct ice_dpll_pin *pin,
drivers/net/ethernet/intel/ice/ice_dpll.c
3313
ret = ice_dpll_get_pins(pf, pin, start_idx, 1, pf->dplls.clock_id);
drivers/net/ethernet/intel/ice/ice_dpll.c
3318
parent = &pf->dplls.inputs[pin->parent_idx[i]];
drivers/net/ethernet/intel/ice/ice_dpll.c
3327
dev_info(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3341
parent = &pf->dplls.inputs[pin->parent_idx[--i]];
drivers/net/ethernet/intel/ice/ice_dpll.c
3365
ice_dpll_init_rclk_pin(struct ice_pf *pf, int start_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
3368
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
3371
ret = ice_dpll_init_pin_common(pf, &pf->dplls.rclk, start_idx, ops);
drivers/net/ethernet/intel/ice/ice_dpll.c
3375
dpll_netdev_pin_set(vsi->netdev, pf->dplls.rclk.pin);
drivers/net/ethernet/intel/ice/ice_dpll.c
3384
flush_workqueue(pin->pf->dplls.wq);
drivers/net/ethernet/intel/ice/ice_dpll.c
3394
ice_dpll_deinit_fwnode_pins(struct ice_pf *pf, struct ice_dpll_pin *pins,
drivers/net/ethernet/intel/ice/ice_dpll.c
3399
for (i = 0; i < pf->dplls.rclk.num_parents; i++)
drivers/net/ethernet/intel/ice/ice_dpll.c
3401
destroy_workqueue(pf->dplls.wq);
drivers/net/ethernet/intel/ice/ice_dpll.c
3412
static void ice_dpll_deinit_pins(struct ice_pf *pf, bool cgu)
drivers/net/ethernet/intel/ice/ice_dpll.c
3414
struct ice_dpll_pin *outputs = pf->dplls.outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3415
struct ice_dpll_pin *inputs = pf->dplls.inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3416
int num_outputs = pf->dplls.num_outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3417
int num_inputs = pf->dplls.num_inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3418
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
3422
ice_dpll_deinit_rclk_pin(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
3423
if (pf->hw.mac_type == ICE_MAC_GENERIC_3K_E825)
drivers/net/ethernet/intel/ice/ice_dpll.c
3424
ice_dpll_deinit_fwnode_pins(pf, pf->dplls.inputs, 0);
drivers/net/ethernet/intel/ice/ice_dpll.c
3438
if (!pf->dplls.generic) {
drivers/net/ethernet/intel/ice/ice_dpll.c
3439
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.ufl,
drivers/net/ethernet/intel/ice/ice_dpll.c
3442
pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3443
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3444
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.sma,
drivers/net/ethernet/intel/ice/ice_dpll.c
3447
pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3448
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3454
ice_dpll_pin_node_get(struct ice_pf *pf, const char *name)
drivers/net/ethernet/intel/ice/ice_dpll.c
3456
struct fwnode_handle *fwnode = dev_fwnode(ice_pf_to_dev(pf));
drivers/net/ethernet/intel/ice/ice_dpll.c
3469
struct ice_pf *pf = pin->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3472
pin->fwnode = ice_dpll_pin_node_get(pf, name);
drivers/net/ethernet/intel/ice/ice_dpll.c
3474
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3481
dev_dbg(ice_pf_to_dev(pf), "Found fwnode node for %s\n", name);
drivers/net/ethernet/intel/ice/ice_dpll.c
3485
dev_info(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3494
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3527
ice_dpll_init_fwnode_pins(struct ice_pf *pf, struct ice_dpll_pin *pins,
drivers/net/ethernet/intel/ice/ice_dpll.c
3533
pf->dplls.wq = create_singlethread_workqueue("ice_dpll_wq");
drivers/net/ethernet/intel/ice/ice_dpll.c
3534
if (!pf->dplls.wq)
drivers/net/ethernet/intel/ice/ice_dpll.c
3537
for (i = 0; i < pf->dplls.rclk.num_parents; i++) {
drivers/net/ethernet/intel/ice/ice_dpll.c
3538
pins[start_idx + i].pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3550
destroy_workqueue(pf->dplls.wq);
drivers/net/ethernet/intel/ice/ice_dpll.c
3567
static int ice_dpll_init_pins_e825(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3571
ret = ice_dpll_init_fwnode_pins(pf, pf->dplls.inputs, 0);
drivers/net/ethernet/intel/ice/ice_dpll.c
3575
ret = ice_dpll_init_rclk_pin(pf, DPLL_PIN_IDX_UNSPEC,
drivers/net/ethernet/intel/ice/ice_dpll.c
3581
complete_all(&pf->dplls.dpll_init);
drivers/net/ethernet/intel/ice/ice_dpll.c
3582
ice_dpll_deinit_fwnode_pins(pf, pf->dplls.inputs, 0);
drivers/net/ethernet/intel/ice/ice_dpll.c
3600
static int ice_dpll_init_pins(struct ice_pf *pf, bool cgu)
drivers/net/ethernet/intel/ice/ice_dpll.c
3609
ret = ice_dpll_init_direct_pins(pf, cgu, pf->dplls.inputs, 0,
drivers/net/ethernet/intel/ice/ice_dpll.c
3610
pf->dplls.num_inputs, input_ops,
drivers/net/ethernet/intel/ice/ice_dpll.c
3611
pf->dplls.eec.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3612
pf->dplls.pps.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3615
count = pf->dplls.num_inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3617
ret = ice_dpll_init_direct_pins(pf, cgu, pf->dplls.outputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3618
count, pf->dplls.num_outputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3619
output_ops, pf->dplls.eec.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3620
pf->dplls.pps.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3623
count += pf->dplls.num_outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3624
if (!pf->dplls.generic) {
drivers/net/ethernet/intel/ice/ice_dpll.c
3625
ret = ice_dpll_init_direct_pins(pf, cgu, pf->dplls.sma,
drivers/net/ethernet/intel/ice/ice_dpll.c
3629
pf->dplls.eec.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3630
pf->dplls.pps.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3634
ret = ice_dpll_init_direct_pins(pf, cgu, pf->dplls.ufl,
drivers/net/ethernet/intel/ice/ice_dpll.c
3638
pf->dplls.eec.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3639
pf->dplls.pps.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3644
ret = ice_dpll_pin_ref_sync_register(pf->dplls.inputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3645
pf->dplls.num_inputs);
drivers/net/ethernet/intel/ice/ice_dpll.c
3648
ret = ice_dpll_pin_ref_sync_register(pf->dplls.sma,
drivers/net/ethernet/intel/ice/ice_dpll.c
3653
count += pf->dplls.num_outputs + 2 * ICE_DPLL_PIN_SW_NUM;
drivers/net/ethernet/intel/ice/ice_dpll.c
3656
ret = ice_dpll_init_rclk_pin(pf, count + pf->ptp.port.port_num,
drivers/net/ethernet/intel/ice/ice_dpll.c
3663
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.ufl, ICE_DPLL_PIN_SW_NUM,
drivers/net/ethernet/intel/ice/ice_dpll.c
3664
&ice_dpll_pin_ufl_ops, pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3665
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3667
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.sma, ICE_DPLL_PIN_SW_NUM,
drivers/net/ethernet/intel/ice/ice_dpll.c
3668
&ice_dpll_pin_sma_ops, pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3669
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3671
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.outputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3672
pf->dplls.num_outputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3673
output_ops, pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3674
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3676
ice_dpll_deinit_direct_pins(pf, cgu, pf->dplls.inputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3677
pf->dplls.num_inputs,
drivers/net/ethernet/intel/ice/ice_dpll.c
3678
input_ops, pf->dplls.pps.dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
3679
pf->dplls.eec.dpll);
drivers/net/ethernet/intel/ice/ice_dpll.c
3693
ice_dpll_deinit_dpll(struct ice_pf *pf, struct ice_dpll *d, bool cgu)
drivers/net/ethernet/intel/ice/ice_dpll.c
3715
ice_dpll_init_dpll(struct ice_pf *pf, struct ice_dpll *d, bool cgu,
drivers/net/ethernet/intel/ice/ice_dpll.c
3718
u64 clock_id = pf->dplls.clock_id;
drivers/net/ethernet/intel/ice/ice_dpll.c
3725
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
3729
d->pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3733
if (type == DPLL_TYPE_PPS && ice_dpll_is_pps_phase_monitor(pf))
drivers/net/ethernet/intel/ice/ice_dpll.c
3735
ice_dpll_update_state(pf, d, true);
drivers/net/ethernet/intel/ice/ice_dpll.c
3754
static void ice_dpll_deinit_worker(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3756
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
3773
static int ice_dpll_init_worker(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3775
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
3780
dev_name(ice_pf_to_dev(pf)));
drivers/net/ethernet/intel/ice/ice_dpll.c
3813
static int ice_dpll_init_info_pins_generic(struct ice_pf *pf, bool input)
drivers/net/ethernet/intel/ice/ice_dpll.c
3815
struct ice_dpll *de = &pf->dplls.eec, *dp = &pf->dplls.pps;
drivers/net/ethernet/intel/ice/ice_dpll.c
3826
pin_num = pf->dplls.num_inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3827
pins = pf->dplls.inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3828
phase_adj_max = pf->dplls.input_phase_adj_max;
drivers/net/ethernet/intel/ice/ice_dpll.c
3832
pin_num = pf->dplls.num_outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3833
pins = pf->dplls.outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3834
phase_adj_max = pf->dplls.output_phase_adj_max;
drivers/net/ethernet/intel/ice/ice_dpll.c
3846
pins[i].pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3847
ret = ice_dpll_pin_state_update(pf, &pins[i], pin_type, NULL);
drivers/net/ethernet/intel/ice/ice_dpll.c
3856
ret = ice_aq_get_cgu_ref_prio(&pf->hw, de->dpll_idx, i,
drivers/net/ethernet/intel/ice/ice_dpll.c
3860
ret = ice_aq_get_cgu_ref_prio(&pf->hw, dp->dpll_idx, i,
drivers/net/ethernet/intel/ice/ice_dpll.c
3882
ice_dpll_init_info_direct_pins(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dpll.c
3885
struct ice_dpll *de = &pf->dplls.eec, *dp = &pf->dplls.pps;
drivers/net/ethernet/intel/ice/ice_dpll.c
3887
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dpll.c
3896
pins = pf->dplls.inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3897
num_pins = pf->dplls.num_inputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3898
phase_adj_max = pf->dplls.input_phase_adj_max;
drivers/net/ethernet/intel/ice/ice_dpll.c
3902
pins = pf->dplls.outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3903
num_pins = pf->dplls.num_outputs;
drivers/net/ethernet/intel/ice/ice_dpll.c
3904
phase_adj_max = pf->dplls.output_phase_adj_max;
drivers/net/ethernet/intel/ice/ice_dpll.c
3911
pf->dplls.generic = true;
drivers/net/ethernet/intel/ice/ice_dpll.c
3912
return ice_dpll_init_info_pins_generic(pf, input);
drivers/net/ethernet/intel/ice/ice_dpll.c
3931
if (ice_dpll_is_sw_pin(pf, i, true))
drivers/net/ethernet/intel/ice/ice_dpll.c
3937
if (ice_dpll_is_sw_pin(pf, i, false))
drivers/net/ethernet/intel/ice/ice_dpll.c
3943
ret = ice_dpll_pin_state_update(pf, &pins[i], pin_type, NULL);
drivers/net/ethernet/intel/ice/ice_dpll.c
3949
pins[i].pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3952
ret = ice_dpll_init_ref_sync_inputs(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
3966
static int ice_dpll_init_info_pin_on_pin_e825c(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3968
struct ice_dpll_pin *rclk_pin = &pf->dplls.rclk;
drivers/net/ethernet/intel/ice/ice_dpll.c
3972
rclk_pin->pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3987
static int ice_dpll_init_info_rclk_pin(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
3989
struct ice_dpll_pin *pin = &pf->dplls.rclk;
drivers/net/ethernet/intel/ice/ice_dpll.c
3993
pin->pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
3995
return ice_dpll_pin_state_update(pf, pin,
drivers/net/ethernet/intel/ice/ice_dpll.c
4010
static int ice_dpll_init_info_sw_pins(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4013
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
4018
if (pf->hw.device_id == ICE_DEV_ID_E810C_QSFP)
drivers/net/ethernet/intel/ice/ice_dpll.c
4028
ice_cgu_get_pin_freq_supp(&pf->hw, pin_abs_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
4035
pin->pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
4048
pin->pf = pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
4054
ice_cgu_get_pin_freq_supp(&pf->hw, pin_abs_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
4067
ice_cgu_get_pin_freq_supp(&pf->hw, pin_abs_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
4076
ret = ice_dpll_pin_state_update(pf, pin, ICE_DPLL_PIN_TYPE_SOFTWARE,
drivers/net/ethernet/intel/ice/ice_dpll.c
4096
ice_dpll_init_pins_info(struct ice_pf *pf, enum ice_dpll_pin_type pin_type)
drivers/net/ethernet/intel/ice/ice_dpll.c
4101
return ice_dpll_init_info_direct_pins(pf, pin_type);
drivers/net/ethernet/intel/ice/ice_dpll.c
4103
if (pf->hw.mac_type == ICE_MAC_GENERIC_3K_E825)
drivers/net/ethernet/intel/ice/ice_dpll.c
4104
return ice_dpll_init_info_pin_on_pin_e825c(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4106
return ice_dpll_init_info_rclk_pin(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4108
return ice_dpll_init_info_sw_pins(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4120
static void ice_dpll_deinit_info(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4122
kfree(pf->dplls.inputs);
drivers/net/ethernet/intel/ice/ice_dpll.c
4123
kfree(pf->dplls.outputs);
drivers/net/ethernet/intel/ice/ice_dpll.c
4124
kfree(pf->dplls.eec.input_prio);
drivers/net/ethernet/intel/ice/ice_dpll.c
4125
kfree(pf->dplls.pps.input_prio);
drivers/net/ethernet/intel/ice/ice_dpll.c
4139
static int ice_dpll_init_info_e825c(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4141
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
4145
d->clock_id = ice_generate_clock_id(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4152
ret = ice_get_cgu_rclk_pin_info(&pf->hw, &d->base_rclk_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
4153
&pf->dplls.rclk.num_parents);
drivers/net/ethernet/intel/ice/ice_dpll.c
4157
for (i = 0; i < pf->dplls.rclk.num_parents; i++)
drivers/net/ethernet/intel/ice/ice_dpll.c
4158
pf->dplls.rclk.parent_idx[i] = d->base_rclk_idx + i;
drivers/net/ethernet/intel/ice/ice_dpll.c
4160
ret = ice_dpll_init_pins_info(pf, ICE_DPLL_PIN_TYPE_RCLK_INPUT);
drivers/net/ethernet/intel/ice/ice_dpll.c
4163
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
4168
ice_dpll_deinit_info(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4183
static int ice_dpll_init_info(struct ice_pf *pf, bool cgu)
drivers/net/ethernet/intel/ice/ice_dpll.c
4186
struct ice_dpll *de = &pf->dplls.eec;
drivers/net/ethernet/intel/ice/ice_dpll.c
4187
struct ice_dpll *dp = &pf->dplls.pps;
drivers/net/ethernet/intel/ice/ice_dpll.c
4188
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
4189
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_dpll.c
4192
d->clock_id = ice_generate_clock_id(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4195
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
4224
ret = ice_dpll_init_pins_info(pf, ICE_DPLL_PIN_TYPE_INPUT);
drivers/net/ethernet/intel/ice/ice_dpll.c
4236
ret = ice_dpll_init_pins_info(pf, ICE_DPLL_PIN_TYPE_OUTPUT);
drivers/net/ethernet/intel/ice/ice_dpll.c
4239
ret = ice_dpll_init_pins_info(pf, ICE_DPLL_PIN_TYPE_SOFTWARE);
drivers/net/ethernet/intel/ice/ice_dpll.c
4244
ret = ice_get_cgu_rclk_pin_info(&pf->hw, &d->base_rclk_idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
4245
&pf->dplls.rclk.num_parents);
drivers/net/ethernet/intel/ice/ice_dpll.c
4248
for (i = 0; i < pf->dplls.rclk.num_parents; i++)
drivers/net/ethernet/intel/ice/ice_dpll.c
4249
pf->dplls.rclk.parent_idx[i] = d->base_rclk_idx + i;
drivers/net/ethernet/intel/ice/ice_dpll.c
4250
ret = ice_dpll_init_pins_info(pf, ICE_DPLL_PIN_TYPE_RCLK_INPUT);
drivers/net/ethernet/intel/ice/ice_dpll.c
4256
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
4263
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
4267
ice_dpll_deinit_info(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4282
void ice_dpll_deinit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4284
bool cgu = ice_is_feature_supported(pf, ICE_F_CGU);
drivers/net/ethernet/intel/ice/ice_dpll.c
4286
clear_bit(ICE_FLAG_DPLL, pf->flags);
drivers/net/ethernet/intel/ice/ice_dpll.c
4288
ice_dpll_deinit_worker(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4290
ice_dpll_deinit_pins(pf, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4291
if (!IS_ERR_OR_NULL(pf->dplls.pps.dpll))
drivers/net/ethernet/intel/ice/ice_dpll.c
4292
ice_dpll_deinit_dpll(pf, &pf->dplls.pps, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4293
if (!IS_ERR_OR_NULL(pf->dplls.eec.dpll))
drivers/net/ethernet/intel/ice/ice_dpll.c
4294
ice_dpll_deinit_dpll(pf, &pf->dplls.eec, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4295
ice_dpll_deinit_info(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4296
mutex_destroy(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
4309
static void ice_dpll_init_e825(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4311
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
4317
err = ice_dpll_init_info_e825c(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4320
err = ice_dpll_init_pins_e825(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4323
set_bit(ICE_FLAG_DPLL, pf->flags);
drivers/net/ethernet/intel/ice/ice_dpll.c
4329
ice_dpll_deinit_info(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4332
dev_warn(ice_pf_to_dev(pf), "DPLLs init failure err:%d\n", err);
drivers/net/ethernet/intel/ice/ice_dpll.c
4345
static void ice_dpll_init_e810(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4347
bool cgu = ice_is_feature_supported(pf, ICE_F_CGU);
drivers/net/ethernet/intel/ice/ice_dpll.c
4348
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
4352
err = ice_dpll_init_info(pf, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4355
err = ice_dpll_init_dpll(pf, &pf->dplls.eec, cgu, DPLL_TYPE_EEC);
drivers/net/ethernet/intel/ice/ice_dpll.c
4358
err = ice_dpll_init_dpll(pf, &pf->dplls.pps, cgu, DPLL_TYPE_PPS);
drivers/net/ethernet/intel/ice/ice_dpll.c
4361
err = ice_dpll_init_pins(pf, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4365
err = ice_dpll_init_worker(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4369
set_bit(ICE_FLAG_DPLL, pf->flags);
drivers/net/ethernet/intel/ice/ice_dpll.c
4374
ice_dpll_deinit_pins(pf, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4376
ice_dpll_deinit_dpll(pf, &pf->dplls.pps, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4378
ice_dpll_deinit_dpll(pf, &pf->dplls.eec, cgu);
drivers/net/ethernet/intel/ice/ice_dpll.c
4380
ice_dpll_deinit_info(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4383
dev_warn(ice_pf_to_dev(pf), "DPLLs init failure err:%d\n", err);
drivers/net/ethernet/intel/ice/ice_dpll.c
4386
void ice_dpll_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
4388
switch (pf->hw.mac_type) {
drivers/net/ethernet/intel/ice/ice_dpll.c
4390
ice_dpll_init_e825(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
4393
ice_dpll_init_e810(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
557
static int ice_dpll_rclk_update_e825c(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_dpll.c
564
if (pf->dplls.rclk.num_parents > ICE_SYNCE_CLK_NUM)
drivers/net/ethernet/intel/ice/ice_dpll.c
567
err = ice_read_cgu_reg(&pf->hw, ICE_CGU_R10, &reg);
drivers/net/ethernet/intel/ice/ice_dpll.c
573
(pf->ptp.port.port_num + ICE_CGU_BYPASS_MUX_OFFSET_E825C));
drivers/net/ethernet/intel/ice/ice_dpll.c
575
err = ice_read_cgu_reg(&pf->hw, ICE_CGU_R11, &reg);
drivers/net/ethernet/intel/ice/ice_dpll.c
581
(pf->ptp.port.port_num + ICE_CGU_BYPASS_MUX_OFFSET_E825C));
drivers/net/ethernet/intel/ice/ice_dpll.c
599
static int ice_dpll_rclk_update(struct ice_pf *pf, struct ice_dpll_pin *pin,
drivers/net/ethernet/intel/ice/ice_dpll.c
604
for (u8 parent = 0; parent < pf->dplls.rclk.num_parents; parent++) {
drivers/net/ethernet/intel/ice/ice_dpll.c
607
ret = ice_aq_get_phy_rec_clk_out(&pf->hw, &p, &port_num,
drivers/net/ethernet/intel/ice/ice_dpll.c
633
ice_dpll_sw_pins_update(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_dpll.c
635
struct ice_dplls *d = &pf->dplls;
drivers/net/ethernet/intel/ice/ice_dpll.c
640
ret = ice_read_sma_ctrl(&pf->hw, &data);
drivers/net/ethernet/intel/ice/ice_dpll.c
698
ice_dpll_pin_state_update(struct ice_pf *pf, struct ice_dpll_pin *pin,
drivers/net/ethernet/intel/ice/ice_dpll.c
707
ret = ice_aq_get_input_pin_cfg(&pf->hw, pin->idx, &pin->status,
drivers/net/ethernet/intel/ice/ice_dpll.c
714
pin->state[pf->dplls.eec.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
715
pin->pin == pf->dplls.eec.active_input ?
drivers/net/ethernet/intel/ice/ice_dpll.c
718
pin->state[pf->dplls.pps.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
719
pin->pin == pf->dplls.pps.active_input ?
drivers/net/ethernet/intel/ice/ice_dpll.c
723
pin->state[pf->dplls.eec.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
725
pin->state[pf->dplls.pps.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
729
pin->state[pf->dplls.eec.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
731
pin->state[pf->dplls.pps.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
736
ret = ice_aq_get_output_pin_cfg(&pf->hw, pin->idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
744
pin->state[pf->dplls.eec.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
745
parent == pf->dplls.eec.dpll_idx ?
drivers/net/ethernet/intel/ice/ice_dpll.c
748
pin->state[pf->dplls.pps.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
749
parent == pf->dplls.pps.dpll_idx ?
drivers/net/ethernet/intel/ice/ice_dpll.c
753
pin->state[pf->dplls.eec.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
755
pin->state[pf->dplls.pps.dpll_idx] =
drivers/net/ethernet/intel/ice/ice_dpll.c
760
if (pf->hw.mac_type == ICE_MAC_GENERIC_3K_E825) {
drivers/net/ethernet/intel/ice/ice_dpll.c
761
ret = ice_dpll_rclk_update_e825c(pf, pin);
drivers/net/ethernet/intel/ice/ice_dpll.c
765
ret = ice_dpll_rclk_update(pf, pin, port_num);
drivers/net/ethernet/intel/ice/ice_dpll.c
771
ret = ice_dpll_sw_pins_update(pf);
drivers/net/ethernet/intel/ice/ice_dpll.c
785
libie_aq_str(pf->hw.adminq.sq_last_status),
drivers/net/ethernet/intel/ice/ice_dpll.c
788
dev_err_ratelimited(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_dpll.c
791
libie_aq_str(pf->hw.adminq.sq_last_status),
drivers/net/ethernet/intel/ice/ice_dpll.c
812
ice_dpll_hw_input_prio_set(struct ice_pf *pf, struct ice_dpll *dpll,
drivers/net/ethernet/intel/ice/ice_dpll.c
818
ret = ice_aq_set_cgu_ref_prio(&pf->hw, dpll->dpll_idx, pin->idx,
drivers/net/ethernet/intel/ice/ice_dpll.c
824
libie_aq_str(pf->hw.adminq.sq_last_status),
drivers/net/ethernet/intel/ice/ice_dpll.c
854
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
856
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
858
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
882
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
884
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
886
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
909
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
911
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
916
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
940
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
942
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
947
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
977
struct ice_pf *pf = d->pf;
drivers/net/ethernet/intel/ice/ice_dpll.c
980
if (ice_dpll_is_reset(pf, extack))
drivers/net/ethernet/intel/ice/ice_dpll.c
983
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.c
985
ret = ice_dpll_pin_enable(&pf->hw, p, d->dpll_idx, pin_type,
drivers/net/ethernet/intel/ice/ice_dpll.c
988
ret = ice_dpll_pin_disable(&pf->hw, p, pin_type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
99
static bool ice_dpll_is_sw_pin(struct ice_pf *pf, u8 index, bool input)
drivers/net/ethernet/intel/ice/ice_dpll.c
990
ret = ice_dpll_pin_state_update(pf, p, pin_type, extack);
drivers/net/ethernet/intel/ice/ice_dpll.c
991
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_dpll.h
152
void ice_dpll_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_dpll.h
153
void ice_dpll_deinit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_dpll.h
155
static inline void ice_dpll_init(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_dpll.h
156
static inline void ice_dpll_deinit(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_dpll.h
46
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_dpll.h
89
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_eswitch.c
109
static int ice_eswitch_setup_repr(struct ice_pf *pf, struct ice_repr *repr)
drivers/net/ethernet/intel/ice/ice_eswitch.c
111
struct ice_vsi *uplink_vsi = pf->eswitch.uplink_vsi;
drivers/net/ethernet/intel/ice/ice_eswitch.c
178
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_eswitch.c
182
if (!ice_is_switchdev_running(pf))
drivers/net/ethernet/intel/ice/ice_eswitch.c
185
repr = xa_load(&pf->eswitch.reprs, *repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
197
dev_err(ice_pf_to_dev(pf), "Failed to update VSI of port representor %d",
drivers/net/ethernet/intel/ice/ice_eswitch.c
20
static int ice_eswitch_setup_env(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
202
xa_erase(&pf->eswitch.reprs, repr->id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
204
if (xa_insert(&pf->eswitch.reprs, repr->id, repr, GFP_KERNEL))
drivers/net/ethernet/intel/ice/ice_eswitch.c
205
dev_err(ice_pf_to_dev(pf), "Failed to reload port representor %d",
drivers/net/ethernet/intel/ice/ice_eswitch.c
22
struct ice_vsi *uplink_vsi = pf->eswitch.uplink_vsi;
drivers/net/ethernet/intel/ice/ice_eswitch.c
270
static void ice_eswitch_release_env(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
272
struct ice_vsi *uplink_vsi = pf->eswitch.uplink_vsi;
drivers/net/ethernet/intel/ice/ice_eswitch.c
293
static int ice_eswitch_enable_switchdev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
297
uplink_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
302
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_eswitch.c
307
pf->eswitch.uplink_vsi = uplink_vsi;
drivers/net/ethernet/intel/ice/ice_eswitch.c
309
if (ice_eswitch_setup_env(pf))
drivers/net/ethernet/intel/ice/ice_eswitch.c
31
ice_remove_vsi_fltr(&pf->hw, uplink_vsi->idx);
drivers/net/ethernet/intel/ice/ice_eswitch.c
312
if (ice_eswitch_br_offloads_init(pf))
drivers/net/ethernet/intel/ice/ice_eswitch.c
315
pf->eswitch.is_running = true;
drivers/net/ethernet/intel/ice/ice_eswitch.c
320
ice_eswitch_release_env(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
328
static void ice_eswitch_disable_switchdev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
330
ice_eswitch_br_offloads_deinit(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
331
ice_eswitch_release_env(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
333
pf->eswitch.is_running = false;
drivers/net/ethernet/intel/ice/ice_eswitch.c
346
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/ice_eswitch.c
348
if (pf->eswitch_mode == mode)
drivers/net/ethernet/intel/ice/ice_eswitch.c
351
if (ice_has_vfs(pf)) {
drivers/net/ethernet/intel/ice/ice_eswitch.c
352
dev_info(ice_pf_to_dev(pf), "Changing eswitch mode is allowed only if there is no VFs created");
drivers/net/ethernet/intel/ice/ice_eswitch.c
359
dev_info(ice_pf_to_dev(pf), "PF %d changed eswitch mode to legacy",
drivers/net/ethernet/intel/ice/ice_eswitch.c
360
pf->hw.pf_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
361
xa_destroy(&pf->eswitch.reprs);
drivers/net/ethernet/intel/ice/ice_eswitch.c
366
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/ice_eswitch.c
367
dev_err(ice_pf_to_dev(pf), "Couldn't change eswitch mode to switchdev - ADQ is active. Delete ADQ configs and try again, e.g. tc qdisc del dev $PF root");
drivers/net/ethernet/intel/ice/ice_eswitch.c
372
dev_info(ice_pf_to_dev(pf), "PF %d changed eswitch mode to switchdev",
drivers/net/ethernet/intel/ice/ice_eswitch.c
373
pf->hw.pf_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
374
xa_init(&pf->eswitch.reprs);
drivers/net/ethernet/intel/ice/ice_eswitch.c
383
pf->eswitch_mode = mode;
drivers/net/ethernet/intel/ice/ice_eswitch.c
394
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/ice_eswitch.c
396
*mode = pf->eswitch_mode;
drivers/net/ethernet/intel/ice/ice_eswitch.c
407
bool ice_is_eswitch_mode_switchdev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
409
return pf->eswitch_mode == DEVLINK_ESWITCH_MODE_SWITCHDEV;
drivers/net/ethernet/intel/ice/ice_eswitch.c
416
static void ice_eswitch_start_all_tx_queues(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
421
if (test_bit(ICE_DOWN, pf->state))
drivers/net/ethernet/intel/ice/ice_eswitch.c
424
xa_for_each(&pf->eswitch.reprs, id, repr)
drivers/net/ethernet/intel/ice/ice_eswitch.c
432
void ice_eswitch_stop_all_tx_queues(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
437
if (test_bit(ICE_DOWN, pf->state))
drivers/net/ethernet/intel/ice/ice_eswitch.c
440
xa_for_each(&pf->eswitch.reprs, id, repr)
drivers/net/ethernet/intel/ice/ice_eswitch.c
444
static void ice_eswitch_stop_reprs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
446
ice_eswitch_stop_all_tx_queues(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
449
static void ice_eswitch_start_reprs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
451
ice_eswitch_start_all_tx_queues(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
455
ice_eswitch_attach(struct ice_pf *pf, struct ice_repr *repr, unsigned long *id)
drivers/net/ethernet/intel/ice/ice_eswitch.c
459
if (pf->eswitch_mode == DEVLINK_ESWITCH_MODE_LEGACY)
drivers/net/ethernet/intel/ice/ice_eswitch.c
462
if (xa_empty(&pf->eswitch.reprs)) {
drivers/net/ethernet/intel/ice/ice_eswitch.c
463
err = ice_eswitch_enable_switchdev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
468
ice_eswitch_stop_reprs(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
474
err = ice_eswitch_setup_repr(pf, repr);
drivers/net/ethernet/intel/ice/ice_eswitch.c
478
err = xa_insert(&pf->eswitch.reprs, repr->id, repr, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_eswitch.c
484
ice_eswitch_start_reprs(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
489
ice_eswitch_release_repr(pf, repr);
drivers/net/ethernet/intel/ice/ice_eswitch.c
493
if (xa_empty(&pf->eswitch.reprs))
drivers/net/ethernet/intel/ice/ice_eswitch.c
494
ice_eswitch_disable_switchdev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
495
ice_eswitch_start_reprs(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
509
int ice_eswitch_attach_vf(struct ice_pf *pf, struct ice_vf *vf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
511
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
515
if (!ice_is_eswitch_mode_switchdev(pf))
drivers/net/ethernet/intel/ice/ice_eswitch.c
523
err = ice_eswitch_attach(pf, repr, &vf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
540
int ice_eswitch_attach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
548
err = ice_eswitch_attach(pf, repr, &sf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
555
static void ice_eswitch_detach(struct ice_pf *pf, struct ice_repr *repr)
drivers/net/ethernet/intel/ice/ice_eswitch.c
557
ice_eswitch_stop_reprs(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
560
xa_erase(&pf->eswitch.reprs, repr->id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
562
if (xa_empty(&pf->eswitch.reprs))
drivers/net/ethernet/intel/ice/ice_eswitch.c
563
ice_eswitch_disable_switchdev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
565
ice_eswitch_release_repr(pf, repr);
drivers/net/ethernet/intel/ice/ice_eswitch.c
568
if (xa_empty(&pf->eswitch.reprs)) {
drivers/net/ethernet/intel/ice/ice_eswitch.c
569
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
574
ice_devlink_rate_clear_tx_topology(ice_get_main_vsi(pf));
drivers/net/ethernet/intel/ice/ice_eswitch.c
577
ice_eswitch_start_reprs(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
586
void ice_eswitch_detach_vf(struct ice_pf *pf, struct ice_vf *vf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
588
struct ice_repr *repr = xa_load(&pf->eswitch.reprs, vf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
589
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_eswitch.c
595
ice_eswitch_detach(pf, repr);
drivers/net/ethernet/intel/ice/ice_eswitch.c
604
void ice_eswitch_detach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf)
drivers/net/ethernet/intel/ice/ice_eswitch.c
606
struct ice_repr *repr = xa_load(&pf->eswitch.reprs, sf->repr_id);
drivers/net/ethernet/intel/ice/ice_eswitch.c
611
ice_eswitch_detach(pf, repr);
drivers/net/ethernet/intel/ice/ice_eswitch.c
89
ice_eswitch_release_repr(struct ice_pf *pf, struct ice_repr *repr)
drivers/net/ethernet/intel/ice/ice_eswitch.h
11
void ice_eswitch_detach_vf(struct ice_pf *pf, struct ice_vf *vf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
12
void ice_eswitch_detach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
13
int ice_eswitch_attach_vf(struct ice_pf *pf, struct ice_vf *vf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
14
int ice_eswitch_attach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
20
bool ice_is_eswitch_mode_switchdev(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
24
void ice_eswitch_stop_all_tx_queues(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_eswitch.h
37
ice_eswitch_detach_vf(struct ice_pf *pf, struct ice_vf *vf) { }
drivers/net/ethernet/intel/ice/ice_eswitch.h
40
ice_eswitch_detach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf) { }
drivers/net/ethernet/intel/ice/ice_eswitch.h
43
ice_eswitch_attach_vf(struct ice_pf *pf, struct ice_vf *vf)
drivers/net/ethernet/intel/ice/ice_eswitch.h
49
ice_eswitch_attach_sf(struct ice_pf *pf, struct ice_dynamic_port *sf)
drivers/net/ethernet/intel/ice/ice_eswitch.h
54
static inline void ice_eswitch_stop_all_tx_queues(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_eswitch.h
75
static inline bool ice_is_eswitch_mode_switchdev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1124
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1134
pf = ice_netdev_to_pf(ice_dev);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1136
err = ice_eswitch_br_uplink_port_init(bridge, pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1137
trace_ice_eswitch_br_port_link(pf->br_port);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1195
ice_eswitch_br_offloads_dealloc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1197
struct ice_esw_br_offloads *br_offloads = pf->eswitch.br_offloads;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1206
pf->eswitch.br_offloads = NULL;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1211
ice_eswitch_br_offloads_alloc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1217
if (pf->eswitch.br_offloads)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1224
pf->eswitch.br_offloads = br_offloads;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1225
br_offloads->pf = pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1231
ice_eswitch_br_offloads_deinit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1235
br_offloads = pf->eswitch.br_offloads;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1250
ice_eswitch_br_offloads_dealloc(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1289
ice_eswitch_br_offloads_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1292
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1296
br_offloads = ice_eswitch_br_offloads_alloc(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
1351
ice_eswitch_br_offloads_dealloc(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
286
ice_eswitch_br_flow_delete(struct ice_pf *pf, struct ice_esw_br_flow *flow)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
288
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
291
err = ice_eswitch_br_rule_delete(&pf->hw, flow->fwd_rule);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
296
err = ice_eswitch_br_rule_delete(&pf->hw, flow->guard_rule);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
307
struct ice_pf *pf = bridge->br_offloads->pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
308
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
332
struct ice_pf *pf = bridge->br_offloads->pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
338
ice_eswitch_br_flow_delete(pf, fdb_entry->flow);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
372
struct ice_pf *pf = bridge->br_offloads->pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
374
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
394
struct ice_pf *pf = bridge->br_offloads->pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
395
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
399
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
460
ice_eswitch_br_flow_delete(pf, flow);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
51
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
61
pf = ice_netdev_to_pf(ice_dev);
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
63
return pf->br_port;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
957
ice_eswitch_br_uplink_port_init(struct ice_esw_br *bridge, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
959
struct ice_vsi *vsi = pf->eswitch.uplink_vsi;
drivers/net/ethernet/intel/ice/ice_eswitch_br.c
970
pf->br_port = br_port;
drivers/net/ethernet/intel/ice/ice_eswitch_br.h
117
ice_eswitch_br_offloads_deinit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.h
119
ice_eswitch_br_offloads_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_eswitch_br.h
70
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1126
static int ice_lbtest_create_frame(struct ice_pf *pf, u8 **ret_data, u16 size)
drivers/net/ethernet/intel/ice/ice_ethtool.c
1130
if (!pf)
drivers/net/ethernet/intel/ice/ice_ethtool.c
1272
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1283
test_vsi = ice_lb_vsi_setup(pf, pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1308
if (ice_aq_set_mac_loopback(&pf->hw, true, NULL)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1320
if (ice_lbtest_create_frame(pf, &tx_frame, ICE_LB_FRAME_SIZE)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1344
if (ice_aq_set_mac_loopback(&pf->hw, false, NULL))
drivers/net/ethernet/intel/ice/ice_ethtool.c
1366
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1367
u16 swic_old = pf->sw_int_count;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1371
wr32(&pf->hw, GLINT_DYN_CTL(pf->oicr_irq.index),
drivers/net/ethernet/intel/ice/ice_ethtool.c
1377
return (swic_old == pf->sw_int_count);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1394
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1398
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1403
set_bit(ICE_TESTING, pf->state);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1405
if (ice_active_vfs(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1413
clear_bit(ICE_TESTING, pf->state);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1434
clear_bit(ICE_TESTING, pf->state);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1441
pf->int_name, status);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1719
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1727
if (test_bit(priv_flag->bitno, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool.c
1745
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1753
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1754
set_bit(ICE_FLAG_ETHTOOL_CTXT, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1756
bitmap_copy(orig_flags, pf->flags, ICE_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1763
set_bit(priv_flag->bitno, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1765
clear_bit(priv_flag->bitno, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1768
bitmap_xor(change_flags, pf->flags, orig_flags, ICE_PF_FLAGS_NBITS);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1774
test_bit(ICE_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1776
set_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1782
if (!test_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1786
status = ice_cfg_lldp_mib_change(&pf->hw, false);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1798
status = ice_aq_stop_lldp(&pf->hw, true, true, NULL);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1805
status = ice_init_pf_dcb(pf, true);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1809
pf->dcbx_cap &= ~DCB_CAP_DCBX_LLD_MANAGED;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1810
pf->dcbx_cap |= DCB_CAP_DCBX_HOST;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1815
if (ice_get_pfc_mode(pf) == ICE_QOS_MODE_DSCP) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1816
clear_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1830
status = ice_aq_start_lldp(&pf->hw, true, NULL);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1837
status = ice_aq_start_stop_dcbx(&pf->hw, true,
drivers/net/ethernet/intel/ice/ice_ethtool.c
1851
status = ice_init_pf_dcb(pf, true);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1856
status = ice_cfg_lldp_mib_change(&pf->hw, true);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1860
pf->dcbx_cap &= ~DCB_CAP_DCBX_HOST;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1861
pf->dcbx_cap |= DCB_CAP_DCBX_LLD_MANAGED;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1870
ice_is_any_vf_in_unicast_promisc(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1873
change_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1878
ice_has_vfs(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
1881
change_bit(ICE_FLAG_VF_VLAN_PRUNING, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1885
clear_bit(ICE_FLAG_ETHTOOL_CTXT, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1926
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
1944
ice_update_pf_stats(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1994
p = (char *)pf + ice_gstrings_pf_stats[j].stat_offset;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2000
data[i++] = pf->stats.priority_xon_tx[j];
drivers/net/ethernet/intel/ice/ice_ethtool.c
2001
data[i++] = pf->stats.priority_xoff_tx[j];
drivers/net/ethernet/intel/ice/ice_ethtool.c
2005
data[i++] = pf->stats.priority_xon_rx[j];
drivers/net/ethernet/intel/ice/ice_ethtool.c
2006
data[i++] = pf->stats.priority_xoff_rx[j];
drivers/net/ethernet/intel/ice/ice_ethtool.c
2124
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2144
if (test_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
2145
phy_types_low = le64_to_cpu(pf->nvm_phy_type_lo);
drivers/net/ethernet/intel/ice/ice_ethtool.c
2146
phy_types_high = le64_to_cpu(pf->nvm_phy_type_hi);
drivers/net/ethernet/intel/ice/ice_ethtool.c
2148
ice_mask_min_supported_speeds(&pf->hw, phy_types_high,
drivers/net/ethernet/intel/ice/ice_ethtool.c
2154
if (ice_fw_supports_link_override(&pf->hw) &&
drivers/net/ethernet/intel/ice/ice_ethtool.c
2155
!ice_fw_supports_report_dflt_cfg(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
2158
ldo = &pf->link_dflt_override;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2166
le64_to_cpu(pf->nvm_phy_type_lo) &
drivers/net/ethernet/intel/ice/ice_ethtool.c
2169
le64_to_cpu(pf->nvm_phy_type_hi) &
drivers/net/ethernet/intel/ice/ice_ethtool.c
2624
struct ice_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2670
if (!test_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool.c
2696
while (test_and_set_bit(ICE_CFG_BUSY, pf->state)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
2771
if (test_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
2773
pf->nvm_phy_type_hi;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2775
pf->nvm_phy_type_lo;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2794
err = ice_aq_set_phy_cfg(&pf->hw, pi, &config, NULL);
drivers/net/ethernet/intel/ice/ice_ethtool.c
2804
clear_bit(ICE_CFG_BUSY, pf->state);
drivers/net/ethernet/intel/ice/ice_ethtool.c
2978
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
2986
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
2987
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3013
status = ice_add_rss_cfg(&pf->hw, vsi, &cfg);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3028
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3034
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3037
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3050
hash_flds = ice_get_rss_cfg(&pf->hw, vsi->idx, hdrs, &symm);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3202
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3204
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3249
while (test_and_set_bit(ICE_CFG_BUSY, pf->state)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3293
tx_rings[i].tx_tstamps = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3350
rx_rings[i].cached_phctime = pf->ptp.cached_phc_time;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3439
clear_bit(ICE_CFG_BUSY, pf->state);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3505
struct ice_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3508
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3557
if (!test_bit(ICE_DOWN, pf->state) &&
drivers/net/ethernet/intel/ice/ice_ethtool.c
3634
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3639
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3662
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3692
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3696
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3701
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3707
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3764
struct ice_pf *pf = ice_netdev_to_pf(dev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3767
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ethtool.c
3775
info->phc_index = ice_ptp_clock_index(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3815
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3818
ch->max_rx = ice_get_max_rxq(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3819
ch->max_tx = ice_get_max_txq(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3828
ch->other_count = test_bit(ICE_FLAG_FD_ENA, pf->flags) ? 1 : 0;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3853
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3859
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3860
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3870
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool.c
3895
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3901
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3906
if (ch->other_count != (test_bit(ICE_FLAG_FD_ENA, pf->flags) ? 1U : 0U))
drivers/net/ethernet/intel/ice/ice_ethtool.c
3909
if (ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3914
if (test_bit(ICE_FLAG_FD_ENA, pf->flags) && pf->hw.fdir_active_fltr) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3937
if (new_rx > ice_get_max_rxq(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3939
ice_get_max_rxq(pf));
drivers/net/ethernet/intel/ice/ice_ethtool.c
3942
if (new_tx > ice_get_max_txq(pf)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3944
ice_get_max_txq(pf));
drivers/net/ethernet/intel/ice/ice_ethtool.c
3948
if (pf->cdev_info && pf->cdev_info->adev) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3949
mutex_lock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3950
device_lock(&pf->cdev_info->adev->dev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3952
if (pf->cdev_info->adev->dev.driver) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3967
vsi->rss_size = ice_get_valid_rss_size(&pf->hw, new_rx);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3971
device_unlock(&pf->cdev_info->adev->dev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3972
mutex_unlock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_ethtool.c
3985
struct ice_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3991
if (ice_is_wol_supported(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
3993
wol->wolopts = pf->wol_ena ? WAKE_MAGIC : 0;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4009
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4011
if (vsi->type != ICE_VSI_PF || !ice_is_wol_supported(&pf->hw))
drivers/net/ethernet/intel/ice/ice_ethtool.c
4019
if (pf->wol_ena != !!wol->wolopts) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
4020
pf->wol_ena = !!wol->wolopts;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4021
device_set_wakeup_enable(ice_pf_to_dev(pf), pf->wol_ena);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4023
pf->wol_ena ? "en" : "dis");
drivers/net/ethernet/intel/ice/ice_ethtool.c
4157
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4170
ec->rx_coalesce_usecs_high < pf->hw.intrl_gran)) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
4172
c_type_str, pf->hw.intrl_gran,
drivers/net/ethernet/intel/ice/ice_ethtool.c
4196
dev_dbg(ice_pf_to_dev(pf), "Invalid container type %d\n",
drivers/net/ethernet/intel/ice/ice_ethtool.c
4226
ice_flush(&pf->hw);
drivers/net/ethernet/intel/ice/ice_ethtool.c
434
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
435
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4432
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4433
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4504
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4508
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
451
strscpy(drvinfo->bus_info, pci_name(pf->pdev),
drivers/net/ethernet/intel/ice/ice_ethtool.c
4638
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4642
pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4643
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4668
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4669
struct ice_hw_port_stats *ps = &pf->stats;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4691
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4692
struct ice_hw_port_stats *ps = &pf->stats;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4713
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4714
struct ice_hw_port_stats *ps = &pf->stats;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4747
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4748
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4778
struct ice_pf *pf = ice_netdev_to_pf(dev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
4796
ice_schedule_reset(pf, reset);
drivers/net/ethernet/intel/ice/ice_ethtool.c
759
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ethtool.c
764
pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
765
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
797
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
798
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
812
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
815
if (pf->hw.debug_mask)
drivers/net/ethernet/intel/ice/ice_ethtool.c
817
pf->hw.debug_mask);
drivers/net/ethernet/intel/ice/ice_ethtool.c
820
return pf->msg_enable;
drivers/net/ethernet/intel/ice/ice_ethtool.c
825
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
829
pf->hw.debug_mask = data;
drivers/net/ethernet/intel/ice/ice_ethtool.c
831
pf->msg_enable = data;
drivers/net/ethernet/intel/ice/ice_ethtool.c
833
pf->msg_enable = data;
drivers/net/ethernet/intel/ice/ice_ethtool.c
840
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
842
stats->link_down_events = pf->link_down_events;
drivers/net/ethernet/intel/ice/ice_ethtool.c
847
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
849
return (int)pf->hw.flash.flash_size;
drivers/net/ethernet/intel/ice/ice_ethtool.c
856
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
857
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool.c
862
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool.c
901
static bool ice_active_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ethtool.c
908
ice_for_each_vf_rcu(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_ethtool.c
955
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ethtool.c
958
return !!(ice_nvm_validate_checksum(&pf->hw));
drivers/net/ethernet/intel/ice/ice_ethtool.c
969
struct ice_pf *pf = (struct ice_pf *)hw->back;
drivers/net/ethernet/intel/ice/ice_ethtool.c
970
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1329
ice_cfg_fdir_xtrct_seq(struct ice_pf *pf, struct ethtool_rx_flow_spec *fsp,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1333
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1335
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1425
ret = ice_fdir_set_hw_fltr_rule(pf, seg, fltr_idx,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1438
ret = ice_fdir_set_hw_fltr_rule(pf, tun_seg, fltr_idx,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1493
ice_fdir_write_fltr(struct ice_pf *pf, struct ice_fdir_fltr *input, bool add,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1496
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1497
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1504
ctrl_vsi = ice_get_ctrl_vsi(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1563
ice_fdir_write_all_fltr(struct ice_pf *pf, struct ice_fdir_fltr *input,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1573
if (is_tun && !ice_get_open_tunnel_port(&pf->hw, &port_num, TNL_ALL))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1575
err = ice_fdir_write_fltr(pf, input, add, is_tun);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1586
void ice_fdir_replay_fltrs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1589
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1592
int err = ice_fdir_write_all_fltr(pf, f_rule, true);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1595
dev_dbg(ice_pf_to_dev(pf), "Flow Director error %d, could not reprogram filter %d\n",
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1606
int ice_fdir_create_dflt_rules(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1616
err = ice_create_init_fdir_rule(pf, dflt_rules[i]);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1634
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1635
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1638
ice_fdir_write_all_fltr(pf, f_rule, false);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1641
devm_kfree(ice_pf_to_dev(pf), f_rule);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1652
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1653
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1657
set_bit(ICE_FLAG_FD_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1658
ice_fdir_create_dflt_rules(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1663
if (!test_and_clear_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1684
ice_fdir_do_rem_flow(struct ice_pf *pf, enum ice_fltr_ptype flow_type)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1686
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1700
ice_create_init_fdir_rule(pf, flow_type);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1712
ice_fdir_update_list_entry(struct ice_pf *pf, struct ice_fdir_fltr *input,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1716
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1721
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1724
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1730
err = ice_fdir_write_all_fltr(pf, old_fltr, false);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1740
ice_fdir_do_rem_flow(pf, old_fltr->flow_type);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1764
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1765
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1768
if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1772
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1773
dev_err(ice_pf_to_dev(pf), "Device is resetting - deleting Flow Director filters not supported during reset\n");
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1777
if (test_bit(ICE_FD_FLUSH_REQ, pf->state))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1781
val = ice_fdir_update_list_entry(pf, NULL, fsp->location);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1832
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1841
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1842
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1852
dev_err(ice_pf_to_dev(pf), "Failed to add filter. Flow director filters are not supported on VF queues.\n");
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1983
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1993
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1994
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1995
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1997
if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
2001
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
2014
ret = ice_cfg_fdir_xtrct_seq(pf, fsp, &userdata);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
2028
ice_fdir_num_avail_fltr(hw, pf->vsi[vsi->idx]) < fltrs_needed) {
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
2058
ret = ice_fdir_update_list_entry(pf, input, fsp->location);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
2062
ret = ice_fdir_write_all_fltr(pf, input, true);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
632
ice_fdir_set_hw_fltr_rule(struct ice_pf *pf, struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
635
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
640
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
647
main_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
651
ctrl_vsi = ice_get_ctrl_vsi(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
719
if (!ice_is_adq_active(pf) || !main_vsi->tc_map_vsi[idx])
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
849
ice_create_init_fdir_rule(struct ice_pf *pf, enum ice_fltr_ptype flow)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
852
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
853
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
890
ret = ice_fdir_set_hw_fltr_rule(pf, seg, flow, ICE_FD_HW_SEG_NON_TUN);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
897
ret = ice_fdir_set_hw_fltr_rule(pf, tun_seg, flow, ICE_FD_HW_SEG_TUN);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
577
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
583
index = ice_tunnel_idx_to_entry(&pf->hw, tnl_type, idx);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
585
status = ice_create_tunnel(&pf->hw, index, tnl_type, ntohs(ti->port));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
599
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
605
status = ice_destroy_tunnel(&pf->hw, ti->hw_priv, tnl_type,
drivers/net/ethernet/intel/ice/ice_fltr.c
108
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_fltr.c
113
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_fltr.c
131
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_fltr.c
136
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_fltr.c
61
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_fltr.c
66
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_fltr.c
85
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_fltr.c
90
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_fw_update.c
1037
priv.pf = pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
1042
err = ice_cancel_pending_update(pf, NULL, extack);
drivers/net/ethernet/intel/ice/ice_fw_update.c
130
if (ice_is_recovery_mode(&pf->hw))
drivers/net/ethernet/intel/ice/ice_fw_update.c
15
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
220
struct ice_pf *pf = priv->pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
221
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
265
return ice_check_component_response(pf, component->identifier, comp_response,
drivers/net/ethernet/intel/ice/ice_fw_update.c
294
int ice_write_one_nvm_block(struct ice_pf *pf, u16 module, u32 offset,
drivers/net/ethernet/intel/ice/ice_fw_update.c
299
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
301
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
310
ice_aq_prep_for_event(pf, &task, ice_aqc_opc_nvm_write);
drivers/net/ethernet/intel/ice/ice_fw_update.c
328
err = ice_aq_wait_for_event(pf, &task, 15 * HZ);
drivers/net/ethernet/intel/ice/ice_fw_update.c
406
ice_write_nvm_module(struct ice_pf *pf, u16 module, const char *component,
drivers/net/ethernet/intel/ice/ice_fw_update.c
410
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
419
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
440
err = ice_write_one_nvm_block(pf, module, offset, block_size,
drivers/net/ethernet/intel/ice/ice_fw_update.c
485
ice_erase_nvm_module(struct ice_pf *pf, u16 module, const char *component,
drivers/net/ethernet/intel/ice/ice_fw_update.c
489
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
491
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
499
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
503
ice_aq_prep_for_event(pf, &task, ice_aqc_opc_nvm_erase);
drivers/net/ethernet/intel/ice/ice_fw_update.c
515
err = ice_aq_wait_for_event(pf, &task, ICE_FW_ERASE_TIMEOUT * HZ);
drivers/net/ethernet/intel/ice/ice_fw_update.c
54
struct ice_pf *pf = priv->pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
55
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
571
ice_switch_flash_banks(struct ice_pf *pf, u8 activate_flags,
drivers/net/ethernet/intel/ice/ice_fw_update.c
574
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
576
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
581
ice_aq_prep_for_event(pf, &task, ice_aqc_opc_nvm_write_activate);
drivers/net/ethernet/intel/ice/ice_fw_update.c
607
err = ice_aq_wait_for_event(pf, &task, 30 * HZ);
drivers/net/ethernet/intel/ice/ice_fw_update.c
644
struct ice_pf *pf = priv->pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
682
err = ice_erase_nvm_module(pf, module, name, extack);
drivers/net/ethernet/intel/ice/ice_fw_update.c
686
return ice_write_nvm_module(pf, module, name, component->component_data,
drivers/net/ethernet/intel/ice/ice_fw_update.c
705
struct ice_pf *pf = priv->pf;
drivers/net/ethernet/intel/ice/ice_fw_update.c
710
err = ice_switch_flash_banks(pf, priv->activate_flags,
drivers/net/ethernet/intel/ice/ice_fw_update.c
715
devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
722
dev_dbg(ice_pf_to_dev(pf), "Firmware indicated EMP reset as sufficient, but EMP reset is disabled\n");
drivers/net/ethernet/intel/ice/ice_fw_update.c
745
pf->fw_emp_reset_disabled = !priv->emp_reset_available;
drivers/net/ethernet/intel/ice/ice_fw_update.c
857
int ice_get_pending_updates(struct ice_pf *pf, u8 *pending,
drivers/net/ethernet/intel/ice/ice_fw_update.c
860
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
862
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
915
ice_cancel_pending_update(struct ice_pf *pf, const char *component,
drivers/net/ethernet/intel/ice/ice_fw_update.c
918
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
919
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
920
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.c
924
err = ice_get_pending_updates(pf, &pending, extack);
drivers/net/ethernet/intel/ice/ice_fw_update.c
95
ice_check_component_response(struct ice_pf *pf, u16 id, u8 response, u8 code,
drivers/net/ethernet/intel/ice/ice_fw_update.c
962
err = ice_switch_flash_banks(pf, pending, NULL, extack);
drivers/net/ethernet/intel/ice/ice_fw_update.c
969
pf->fw_emp_reset_disabled = false;
drivers/net/ethernet/intel/ice/ice_fw_update.c
98
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
994
struct ice_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/intel/ice/ice_fw_update.c
995
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_fw_update.c
996
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_fw_update.h
10
int ice_get_pending_updates(struct ice_pf *pf, u8 *pending,
drivers/net/ethernet/intel/ice/ice_fw_update.h
12
int ice_write_one_nvm_block(struct ice_pf *pf, u16 module, u32 offset,
drivers/net/ethernet/intel/ice/ice_gnss.c
102
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_gnss.c
147
count = gnss_insert_raw(pf->gnss_dev, buf, i);
drivers/net/ethernet/intel/ice/ice_gnss.c
149
dev_warn(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_gnss.c
158
dev_dbg(ice_pf_to_dev(pf), "GNSS failed to read err=%d\n", err);
drivers/net/ethernet/intel/ice/ice_gnss.c
171
static struct gnss_serial *ice_gnss_struct_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
173
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
181
gnss->back = pf;
drivers/net/ethernet/intel/ice/ice_gnss.c
182
pf->gnss_serial = gnss;
drivers/net/ethernet/intel/ice/ice_gnss.c
20
ice_gnss_do_write(struct ice_pf *pf, const unsigned char *buf, unsigned int size)
drivers/net/ethernet/intel/ice/ice_gnss.c
208
struct ice_pf *pf = gnss_get_drvdata(gdev);
drivers/net/ethernet/intel/ice/ice_gnss.c
211
if (!pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
214
if (!test_bit(ICE_FLAG_GNSS, pf->flags))
drivers/net/ethernet/intel/ice/ice_gnss.c
217
gnss = pf->gnss_serial;
drivers/net/ethernet/intel/ice/ice_gnss.c
23
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_gnss.c
234
struct ice_pf *pf = gnss_get_drvdata(gdev);
drivers/net/ethernet/intel/ice/ice_gnss.c
237
if (!pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
240
gnss = pf->gnss_serial;
drivers/net/ethernet/intel/ice/ice_gnss.c
261
struct ice_pf *pf = gnss_get_drvdata(gdev);
drivers/net/ethernet/intel/ice/ice_gnss.c
268
if (!pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
271
if (!test_bit(ICE_FLAG_GNSS, pf->flags))
drivers/net/ethernet/intel/ice/ice_gnss.c
274
gnss = pf->gnss_serial;
drivers/net/ethernet/intel/ice/ice_gnss.c
278
return ice_gnss_do_write(pf, buf, count);
drivers/net/ethernet/intel/ice/ice_gnss.c
297
static int ice_gnss_register(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
302
gdev = gnss_allocate_device(ice_pf_to_dev(pf));
drivers/net/ethernet/intel/ice/ice_gnss.c
304
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_gnss.c
311
gnss_set_drvdata(gdev, pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
314
dev_err(ice_pf_to_dev(pf), "gnss_register_device err=%d\n",
drivers/net/ethernet/intel/ice/ice_gnss.c
318
pf->gnss_dev = gdev;
drivers/net/ethernet/intel/ice/ice_gnss.c
331
static void ice_gnss_deregister(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
333
if (pf->gnss_dev) {
drivers/net/ethernet/intel/ice/ice_gnss.c
334
gnss_deregister_device(pf->gnss_dev);
drivers/net/ethernet/intel/ice/ice_gnss.c
335
gnss_put_device(pf->gnss_dev);
drivers/net/ethernet/intel/ice/ice_gnss.c
336
pf->gnss_dev = NULL;
drivers/net/ethernet/intel/ice/ice_gnss.c
344
void ice_gnss_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
348
pf->gnss_serial = ice_gnss_struct_init(pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
349
if (!pf->gnss_serial)
drivers/net/ethernet/intel/ice/ice_gnss.c
352
ret = ice_gnss_register(pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
354
set_bit(ICE_FLAG_GNSS, pf->flags);
drivers/net/ethernet/intel/ice/ice_gnss.c
355
dev_info(ice_pf_to_dev(pf), "GNSS init successful\n");
drivers/net/ethernet/intel/ice/ice_gnss.c
357
ice_gnss_exit(pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
358
dev_err(ice_pf_to_dev(pf), "GNSS init failure\n");
drivers/net/ethernet/intel/ice/ice_gnss.c
366
void ice_gnss_exit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_gnss.c
368
ice_gnss_deregister(pf);
drivers/net/ethernet/intel/ice/ice_gnss.c
369
clear_bit(ICE_FLAG_GNSS, pf->flags);
drivers/net/ethernet/intel/ice/ice_gnss.c
371
if (pf->gnss_serial) {
drivers/net/ethernet/intel/ice/ice_gnss.c
372
struct gnss_serial *gnss = pf->gnss_serial;
drivers/net/ethernet/intel/ice/ice_gnss.c
379
pf->gnss_serial = NULL;
drivers/net/ethernet/intel/ice/ice_gnss.c
72
dev_err(ice_pf_to_dev(pf), "GNSS failed to write, offset=%u, size=%u, err=%d\n",
drivers/net/ethernet/intel/ice/ice_gnss.c
91
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_gnss.c
98
pf = gnss->back;
drivers/net/ethernet/intel/ice/ice_gnss.c
99
if (!pf || !test_bit(ICE_FLAG_GNSS, pf->flags))
drivers/net/ethernet/intel/ice/ice_gnss.h
38
void ice_gnss_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_gnss.h
39
void ice_gnss_exit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_gnss.h
42
static inline void ice_gnss_init(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_gnss.h
43
static inline void ice_gnss_exit(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_hwmon.c
102
void ice_hwmon_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_hwmon.c
104
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_hwmon.c
107
if (!ice_is_internal_reading_supported(pf))
drivers/net/ethernet/intel/ice/ice_hwmon.c
110
hdev = hwmon_device_register_with_info(dev, "ice", pf, &ice_chip_info,
drivers/net/ethernet/intel/ice/ice_hwmon.c
118
pf->hwmon_dev = hdev;
drivers/net/ethernet/intel/ice/ice_hwmon.c
121
void ice_hwmon_exit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_hwmon.c
123
if (!pf->hwmon_dev)
drivers/net/ethernet/intel/ice/ice_hwmon.c
125
hwmon_device_unregister(pf->hwmon_dev);
drivers/net/ethernet/intel/ice/ice_hwmon.c
23
struct ice_pf *pf = dev_get_drvdata(dev);
drivers/net/ethernet/intel/ice/ice_hwmon.c
29
ret = ice_aq_get_sensor_reading(&pf->hw, &resp);
drivers/net/ethernet/intel/ice/ice_hwmon.c
88
static bool ice_is_internal_reading_supported(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_hwmon.c
94
if (pf->hw.pf_id)
drivers/net/ethernet/intel/ice/ice_hwmon.c
97
unsigned long sensors = pf->hw.dev_caps.supported_sensors;
drivers/net/ethernet/intel/ice/ice_hwmon.h
11
static inline void ice_hwmon_init(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_hwmon.h
12
static inline void ice_hwmon_exit(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_hwmon.h
8
void ice_hwmon_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_hwmon.h
9
void ice_hwmon_exit(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_idc.c
133
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
140
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
141
vsi = ice_find_vsi(pf, qset->vport_id);
drivers/net/ethernet/intel/ice/ice_idc.c
143
dev_err(ice_pf_to_dev(pf), "RDMA Invalid VSI\n");
drivers/net/ethernet/intel/ice/ice_idc.c
165
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
170
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
183
return ice_schedule_reset(pf, reset);
drivers/net/ethernet/intel/ice/ice_idc.c
199
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
205
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
206
vsi = ice_find_vsi(pf, vsi_id);
drivers/net/ethernet/intel/ice/ice_idc.c
210
status = ice_cfg_rdma_fltr(&pf->hw, vsi->idx, enable);
drivers/net/ethernet/intel/ice/ice_idc.c
212
dev_err(ice_pf_to_dev(pf), "Failed to %sable RDMA filtering\n",
drivers/net/ethernet/intel/ice/ice_idc.c
236
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
241
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
242
map = ice_alloc_irq(pf, true);
drivers/net/ethernet/intel/ice/ice_idc.c
262
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
267
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
271
ice_free_irq(pf, map);
drivers/net/ethernet/intel/ice/ice_idc.c
294
int ice_plug_aux_dev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_idc.c
304
if (!ice_is_rdma_ena(pf))
drivers/net/ethernet/intel/ice/ice_idc.c
307
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/ice_idc.c
318
adev->id = pf->aux_idx;
drivers/net/ethernet/intel/ice/ice_idc.c
320
adev->dev.parent = &pf->pdev->dev;
drivers/net/ethernet/intel/ice/ice_idc.c
336
mutex_lock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
338
mutex_unlock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
339
set_bit(ICE_FLAG_AUX_DEV_CREATED, pf->flags);
drivers/net/ethernet/intel/ice/ice_idc.c
347
void ice_unplug_aux_dev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_idc.c
351
if (!test_and_clear_bit(ICE_FLAG_AUX_DEV_CREATED, pf->flags))
drivers/net/ethernet/intel/ice/ice_idc.c
354
mutex_lock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
355
adev = pf->cdev_info->adev;
drivers/net/ethernet/intel/ice/ice_idc.c
356
pf->cdev_info->adev = NULL;
drivers/net/ethernet/intel/ice/ice_idc.c
357
mutex_unlock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
371
void ice_rdma_finalize_setup(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_idc.c
373
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_idc.c
377
if (!ice_is_rdma_ena(pf) || !pf->cdev_info)
drivers/net/ethernet/intel/ice/ice_idc.c
38
void ice_send_event_to_aux(struct ice_pf *pf, struct iidc_rdma_event *event)
drivers/net/ethernet/intel/ice/ice_idc.c
380
privd = pf->cdev_info->iidc_priv;
drivers/net/ethernet/intel/ice/ice_idc.c
381
if (!privd || !pf->vsi || !pf->vsi[0] || !pf->vsi[0]->netdev)
drivers/net/ethernet/intel/ice/ice_idc.c
385
privd->netdev = pf->vsi[0]->netdev;
drivers/net/ethernet/intel/ice/ice_idc.c
386
privd->vport_id = pf->vsi[0]->vsi_num;
drivers/net/ethernet/intel/ice/ice_idc.c
389
ice_setup_dcb_qos_info(pf, &privd->qos_info);
drivers/net/ethernet/intel/ice/ice_idc.c
391
ret = ice_plug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_idc.c
400
int ice_init_rdma(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_idc.c
403
struct device *dev = &pf->pdev->dev;
drivers/net/ethernet/intel/ice/ice_idc.c
407
if (!ice_is_rdma_ena(pf)) {
drivers/net/ethernet/intel/ice/ice_idc.c
416
pf->cdev_info = cdev;
drivers/net/ethernet/intel/ice/ice_idc.c
424
privd->pf_id = pf->hw.pf_id;
drivers/net/ethernet/intel/ice/ice_idc.c
425
ret = xa_alloc(&ice_aux_id, &pf->aux_idx, NULL, XA_LIMIT(1, INT_MAX),
drivers/net/ethernet/intel/ice/ice_idc.c
435
privd->hw_addr = (u8 __iomem *)pf->hw.hw_addr;
drivers/net/ethernet/intel/ice/ice_idc.c
436
cdev->pdev = pf->pdev;
drivers/net/ethernet/intel/ice/ice_idc.c
438
pf->cdev_info->rdma_protocol |= IIDC_RDMA_PROTOCOL_ROCEV2;
drivers/net/ethernet/intel/ice/ice_idc.c
446
pf->cdev_info = NULL;
drivers/net/ethernet/intel/ice/ice_idc.c
455
void ice_deinit_rdma(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_idc.c
457
if (!ice_is_rdma_ena(pf))
drivers/net/ethernet/intel/ice/ice_idc.c
46
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/ice_idc.c
460
xa_erase(&ice_aux_id, pf->aux_idx);
drivers/net/ethernet/intel/ice/ice_idc.c
461
kfree(pf->cdev_info->iidc_priv);
drivers/net/ethernet/intel/ice/ice_idc.c
462
kfree(pf->cdev_info);
drivers/net/ethernet/intel/ice/ice_idc.c
463
pf->cdev_info = NULL;
drivers/net/ethernet/intel/ice/ice_idc.c
50
mutex_lock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
60
mutex_unlock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_idc.c
76
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_idc.c
85
pf = pci_get_drvdata(cdev->pdev);
drivers/net/ethernet/intel/ice/ice_idc.c
86
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_idc.c
88
if (!ice_is_rdma_ena(pf))
drivers/net/ethernet/intel/ice/ice_idc.c
91
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_idc_int.h
12
void ice_send_event_to_aux(struct ice_pf *pf, struct iidc_rdma_event *event);
drivers/net/ethernet/intel/ice/ice_irq.c
107
static int ice_get_default_msix_amount(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_irq.c
110
(test_bit(ICE_FLAG_FD_ENA, pf->flags) ? ICE_FDIR_MSIX : 0) +
drivers/net/ethernet/intel/ice/ice_irq.c
111
(ice_is_rdma_ena(pf) ? netif_get_num_default_rss_queues() +
drivers/net/ethernet/intel/ice/ice_irq.c
119
void ice_clear_interrupt_scheme(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_irq.c
121
pci_free_irq_vectors(pf->pdev);
drivers/net/ethernet/intel/ice/ice_irq.c
122
ice_deinit_irq_tracker(pf);
drivers/net/ethernet/intel/ice/ice_irq.c
123
ice_deinit_virt_irq_tracker(pf);
drivers/net/ethernet/intel/ice/ice_irq.c
130
int ice_init_interrupt_scheme(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_irq.c
132
int total_vectors = pf->hw.func_caps.common_cap.num_msix_vectors;
drivers/net/ethernet/intel/ice/ice_irq.c
136
if (!pf->msix.min)
drivers/net/ethernet/intel/ice/ice_irq.c
137
pf->msix.min = ICE_MIN_MSIX;
drivers/net/ethernet/intel/ice/ice_irq.c
139
if (!pf->msix.max)
drivers/net/ethernet/intel/ice/ice_irq.c
140
pf->msix.max = min(total_vectors,
drivers/net/ethernet/intel/ice/ice_irq.c
141
ice_get_default_msix_amount(pf));
drivers/net/ethernet/intel/ice/ice_irq.c
143
pf->msix.total = total_vectors;
drivers/net/ethernet/intel/ice/ice_irq.c
144
pf->msix.rest = total_vectors - pf->msix.max;
drivers/net/ethernet/intel/ice/ice_irq.c
146
if (pci_msix_can_alloc_dyn(pf->pdev))
drivers/net/ethernet/intel/ice/ice_irq.c
147
vectors = pf->msix.min;
drivers/net/ethernet/intel/ice/ice_irq.c
149
vectors = pf->msix.max;
drivers/net/ethernet/intel/ice/ice_irq.c
15
ice_init_irq_tracker(struct ice_pf *pf, unsigned int max_vectors,
drivers/net/ethernet/intel/ice/ice_irq.c
151
vectors = pci_alloc_irq_vectors(pf->pdev, pf->msix.min, vectors,
drivers/net/ethernet/intel/ice/ice_irq.c
156
ice_init_irq_tracker(pf, pf->msix.max, vectors);
drivers/net/ethernet/intel/ice/ice_irq.c
158
return ice_init_virt_irq_tracker(pf, pf->msix.max, pf->msix.rest);
drivers/net/ethernet/intel/ice/ice_irq.c
18
pf->irq_tracker.num_entries = max_vectors;
drivers/net/ethernet/intel/ice/ice_irq.c
183
struct msi_map ice_alloc_irq(struct ice_pf *pf, bool dyn_allowed)
drivers/net/ethernet/intel/ice/ice_irq.c
186
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_irq.c
189
entry = ice_get_irq_res(pf, dyn_allowed);
drivers/net/ethernet/intel/ice/ice_irq.c
19
pf->irq_tracker.num_static = num_static;
drivers/net/ethernet/intel/ice/ice_irq.c
193
if (pci_msix_can_alloc_dyn(pf->pdev) && entry->dynamic) {
drivers/net/ethernet/intel/ice/ice_irq.c
194
map = pci_msix_alloc_irq_at(pf->pdev, entry->index, NULL);
drivers/net/ethernet/intel/ice/ice_irq.c
20
xa_init_flags(&pf->irq_tracker.entries, XA_FLAGS_ALLOC);
drivers/net/ethernet/intel/ice/ice_irq.c
200
map.virq = pci_irq_vector(pf->pdev, map.index);
drivers/net/ethernet/intel/ice/ice_irq.c
207
ice_free_irq_res(pf, entry->index);
drivers/net/ethernet/intel/ice/ice_irq.c
219
void ice_free_irq(struct ice_pf *pf, struct msi_map map)
drivers/net/ethernet/intel/ice/ice_irq.c
223
entry = xa_load(&pf->irq_tracker.entries, map.index);
drivers/net/ethernet/intel/ice/ice_irq.c
226
dev_err(ice_pf_to_dev(pf), "Failed to get MSIX interrupt entry at index %d",
drivers/net/ethernet/intel/ice/ice_irq.c
231
dev_dbg(ice_pf_to_dev(pf), "Free irq at index %d\n", map.index);
drivers/net/ethernet/intel/ice/ice_irq.c
234
pci_msix_free_irq(pf->pdev, map);
drivers/net/ethernet/intel/ice/ice_irq.c
236
ice_free_irq_res(pf, map.index);
drivers/net/ethernet/intel/ice/ice_irq.c
24
ice_init_virt_irq_tracker(struct ice_pf *pf, u32 base, u32 num_entries)
drivers/net/ethernet/intel/ice/ice_irq.c
251
int ice_virt_get_irqs(struct ice_pf *pf, u32 needed)
drivers/net/ethernet/intel/ice/ice_irq.c
253
int res = bitmap_find_next_zero_area(pf->virt_irq_tracker.bm,
drivers/net/ethernet/intel/ice/ice_irq.c
254
pf->virt_irq_tracker.num_entries,
drivers/net/ethernet/intel/ice/ice_irq.c
257
if (res >= pf->virt_irq_tracker.num_entries)
drivers/net/ethernet/intel/ice/ice_irq.c
26
pf->virt_irq_tracker.bm = bitmap_zalloc(num_entries, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_irq.c
260
bitmap_set(pf->virt_irq_tracker.bm, res, needed);
drivers/net/ethernet/intel/ice/ice_irq.c
263
return res + pf->virt_irq_tracker.base;
drivers/net/ethernet/intel/ice/ice_irq.c
27
if (!pf->virt_irq_tracker.bm)
drivers/net/ethernet/intel/ice/ice_irq.c
272
void ice_virt_free_irqs(struct ice_pf *pf, u32 index, u32 irqs)
drivers/net/ethernet/intel/ice/ice_irq.c
274
bitmap_clear(pf->virt_irq_tracker.bm, index - pf->virt_irq_tracker.base,
drivers/net/ethernet/intel/ice/ice_irq.c
30
pf->virt_irq_tracker.num_entries = num_entries;
drivers/net/ethernet/intel/ice/ice_irq.c
31
pf->virt_irq_tracker.base = base;
drivers/net/ethernet/intel/ice/ice_irq.c
40
static void ice_deinit_irq_tracker(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_irq.c
42
xa_destroy(&pf->irq_tracker.entries);
drivers/net/ethernet/intel/ice/ice_irq.c
45
static void ice_deinit_virt_irq_tracker(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_irq.c
47
bitmap_free(pf->virt_irq_tracker.bm);
drivers/net/ethernet/intel/ice/ice_irq.c
55
static void ice_free_irq_res(struct ice_pf *pf, u16 index)
drivers/net/ethernet/intel/ice/ice_irq.c
59
entry = xa_erase(&pf->irq_tracker.entries, index);
drivers/net/ethernet/intel/ice/ice_irq.c
74
static struct ice_irq_entry *ice_get_irq_res(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_irq.c
77
struct xa_limit limit = { .max = pf->irq_tracker.num_entries - 1,
drivers/net/ethernet/intel/ice/ice_irq.c
79
unsigned int num_static = pf->irq_tracker.num_static - 1;
drivers/net/ethernet/intel/ice/ice_irq.c
92
ret = xa_alloc(&pf->irq_tracker.entries, &index, entry, limit,
drivers/net/ethernet/intel/ice/ice_irq.h
28
int ice_init_interrupt_scheme(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_irq.h
29
void ice_clear_interrupt_scheme(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_irq.h
31
struct msi_map ice_alloc_irq(struct ice_pf *pf, bool dyn_only);
drivers/net/ethernet/intel/ice/ice_irq.h
32
void ice_free_irq(struct ice_pf *pf, struct msi_map map);
drivers/net/ethernet/intel/ice/ice_irq.h
34
int ice_virt_get_irqs(struct ice_pf *pf, u32 needed);
drivers/net/ethernet/intel/ice/ice_irq.h
35
void ice_virt_free_irqs(struct ice_pf *pf, u32 index, u32 irqs);
drivers/net/ethernet/intel/ice/ice_lag.c
100
pf->pdev->slot != test_pf->pdev->slot)
drivers/net/ethernet/intel/ice/ice_lag.c
1014
pri_hw = &lag->pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
1015
if (e_pf && lag->pf != e_pf)
drivers/net/ethernet/intel/ice/ice_lag.c
1026
vsi = ice_get_main_vsi(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1215
ice_for_each_vsi(lag->pf, i)
drivers/net/ethernet/intel/ice/ice_lag.c
1216
if (lag->pf->vsi[i] && lag->pf->vsi[i]->type == ICE_VSI_VF)
drivers/net/ethernet/intel/ice/ice_lag.c
122
xa_for_each(&lag->pf->eswitch.reprs, id, repr)
drivers/net/ethernet/intel/ice/ice_lag.c
1232
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1236
struct ice_hw *hw = &lag->pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
1302
if (ice_aq_move_sched_elems(&lag->pf->hw, buf, buf_size, &num_moved))
drivers/net/ethernet/intel/ice/ice_lag.c
1326
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1332
pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1333
ice_for_each_vsi(pf, i)
drivers/net/ethernet/intel/ice/ice_lag.c
1334
if (pf->vsi[i] && pf->vsi[i]->type == ICE_VSI_VF)
drivers/net/ethernet/intel/ice/ice_lag.c
1345
struct ice_pf *pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1348
dev_warn(ice_pf_to_dev(pf), "%s Already part of a bond\n",
drivers/net/ethernet/intel/ice/ice_lag.c
1364
struct ice_pf *pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1373
pri_port = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1387
pri_port = primary_lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1388
loc_port = pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1392
&lag->pf->hw);
drivers/net/ethernet/intel/ice/ice_lag.c
1417
ice_lag_aa_failover(pri_lag, ICE_LAGP_IDX, lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1476
dev_err(ice_pf_to_dev(local_lag->pf), "-ENOMEM error setting SWID\n");
drivers/net/ethernet/intel/ice/ice_lag.c
1485
status = ice_aq_alloc_free_res(&local_lag->pf->hw, buf,
drivers/net/ethernet/intel/ice/ice_lag.c
1488
dev_err(ice_pf_to_dev(local_lag->pf), "Error freeing SWID during LAG unlink\n");
drivers/net/ethernet/intel/ice/ice_lag.c
1500
cpu_to_le16(local_lag->pf->hw.port_info->sw_id);
drivers/net/ethernet/intel/ice/ice_lag.c
1503
status = ice_aq_alloc_free_res(&local_lag->pf->hw, buf, buf_len,
drivers/net/ethernet/intel/ice/ice_lag.c
1506
dev_err(ice_pf_to_dev(local_lag->pf), "Error subscribing to SWID 0x%04X\n",
drivers/net/ethernet/intel/ice/ice_lag.c
1515
swid = local_lag->pf->hw.port_info->sw_id;
drivers/net/ethernet/intel/ice/ice_lag.c
1527
status = ice_aq_send_cmd(&local_lag->pf->hw, &desc, NULL, 0,
drivers/net/ethernet/intel/ice/ice_lag.c
1536
dev_err(ice_pf_to_dev(local_lag->pf), "Error setting SWID in port params %d\n",
drivers/net/ethernet/intel/ice/ice_lag.c
1549
struct ice_hw *hw = &lag->pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
1553
dev_warn(ice_pf_to_dev(lag->pf), "Failure to set primary interface shared status\n");
drivers/net/ethernet/intel/ice/ice_lag.c
1567
dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1569
prim_vsi_idx = lag->pf->vsi[0]->idx;
drivers/net/ethernet/intel/ice/ice_lag.c
1571
if (!ice_find_vsi_list_entry(&lag->pf->hw, ICE_SW_LKUP_VLAN,
drivers/net/ethernet/intel/ice/ice_lag.c
1606
dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1608
vsi_idx = lag->pf->vsi[0]->idx;
drivers/net/ethernet/intel/ice/ice_lag.c
1610
if (!ice_find_vsi_list_entry(&lag->pf->hw, ICE_SW_LKUP_VLAN,
drivers/net/ethernet/intel/ice/ice_lag.c
1639
static void ice_lag_init_feature_support_flag(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lag.c
1643
caps = &pf->hw.dev_caps.common_cap;
drivers/net/ethernet/intel/ice/ice_lag.c
1645
ice_set_feature_support(pf, ICE_F_ROCE_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1647
ice_clear_feature_support(pf, ICE_F_ROCE_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1649
if (caps->sriov_lag && ice_pkg_has_lport_extract(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lag.c
1650
ice_set_feature_support(pf, ICE_F_SRIOV_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1652
ice_clear_feature_support(pf, ICE_F_SRIOV_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1654
if (caps->sriov_aa_lag && ice_pkg_has_lport_extract(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lag.c
1655
ice_set_feature_support(pf, ICE_F_SRIOV_AA_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1657
ice_clear_feature_support(pf, ICE_F_SRIOV_AA_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
1685
if (!ice_is_switchdev_running(lag->pf))
drivers/net/ethernet/intel/ice/ice_lag.c
1691
lag->bond_lport_pri = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1697
if (!ice_is_switchdev_running(primary_lag->pf))
drivers/net/ethernet/intel/ice/ice_lag.c
1700
swid = primary_lag->pf->hw.port_info->sw_id;
drivers/net/ethernet/intel/ice/ice_lag.c
1702
ice_lag_add_prune_list(primary_lag, lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1704
lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1729
ice_lag_del_prune_list(primary_lag, lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
1749
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1756
if (!netif_is_same_ice(lag->pf, event_netdev))
drivers/net/ethernet/intel/ice/ice_lag.c
1765
pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1766
prim_hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
1811
struct ice_pf *pf, *event_pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1814
pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
1815
if (!pf)
drivers/net/ethernet/intel/ice/ice_lag.c
1821
prim_port = pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1885
prim_port = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
1922
ice_for_each_vf(lag->pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_lag.c
2014
dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2019
if (!ice_is_switchdev_running(lag->pf)) {
drivers/net/ethernet/intel/ice/ice_lag.c
2024
if (lag->bond_aa && !ice_is_feature_supported(lag->pf,
drivers/net/ethernet/intel/ice/ice_lag.c
2051
vsi = ice_get_main_vsi(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2053
if (lag->pf->pdev->bus != peer_vsi->back->pdev->bus ||
drivers/net/ethernet/intel/ice/ice_lag.c
2054
lag->pf->pdev->slot != peer_vsi->back->pdev->slot) {
drivers/net/ethernet/intel/ice/ice_lag.c
2097
if (p_lag->active_port != p_lag->pf->hw.port_info->lport &&
drivers/net/ethernet/intel/ice/ice_lag.c
2139
ice_clear_rdma_cap(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2141
ice_set_rdma_cap(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2165
!ice_is_feature_supported(prim_lag->pf, ICE_F_SRIOV_LAG)) {
drivers/net/ethernet/intel/ice/ice_lag.c
2166
ice_clear_feature_support(lag->pf, ICE_F_SRIOV_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
2170
ice_lag_init_feature_support_flag(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2180
struct ice_pf *pf = ice_netdev_to_pf(lag->netdev);
drivers/net/ethernet/intel/ice/ice_lag.c
2182
ice_clear_feature_support(pf, ICE_F_SRIOV_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
2183
ice_clear_feature_support(pf, ICE_F_SRIOV_AA_LAG);
drivers/net/ethernet/intel/ice/ice_lag.c
2215
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lag.c
2218
pf = lag_work->lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
2220
mutex_lock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_lag.c
2227
if (ice_is_feature_supported(pf, ICE_F_SRIOV_LAG)) {
drivers/net/ethernet/intel/ice/ice_lag.c
2235
if (ice_is_feature_supported(pf, ICE_F_SRIOV_LAG)) {
drivers/net/ethernet/intel/ice/ice_lag.c
2253
if (ice_is_feature_supported(pf, ICE_F_SRIOV_LAG)) {
drivers/net/ethernet/intel/ice/ice_lag.c
2275
mutex_unlock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_lag.c
2373
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2394
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
240
struct ice_hw *hw = &lag->pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
2453
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2457
struct ice_hw *hw = &lag->pf->hw;
drivers/net/ethernet/intel/ice/ice_lag.c
251
dev_err(ice_pf_to_dev(lag->pf), "error allocating rule for LAG\n");
drivers/net/ethernet/intel/ice/ice_lag.c
2520
if (ice_aq_move_sched_elems(&lag->pf->hw, buf, buf_size, &num_moved))
drivers/net/ethernet/intel/ice/ice_lag.c
2547
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lag.c
2553
pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
2554
ice_for_each_vsi(pf, i)
drivers/net/ethernet/intel/ice/ice_lag.c
2555
if (pf->vsi[i] && pf->vsi[i]->type == ICE_VSI_VF)
drivers/net/ethernet/intel/ice/ice_lag.c
2568
int ice_init_lag(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lag.c
2570
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2576
ice_lag_init_feature_support_flag(pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2577
if (!ice_is_feature_supported(pf, ICE_F_SRIOV_LAG))
drivers/net/ethernet/intel/ice/ice_lag.c
2580
pf->lag = kzalloc_obj(*lag);
drivers/net/ethernet/intel/ice/ice_lag.c
2581
if (!pf->lag)
drivers/net/ethernet/intel/ice/ice_lag.c
2583
lag = pf->lag;
drivers/net/ethernet/intel/ice/ice_lag.c
2585
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2592
lag->pf = pf;
drivers/net/ethernet/intel/ice/ice_lag.c
2610
err = ice_create_lag_recipe(&pf->hw, &lag->pf_recipe,
drivers/net/ethernet/intel/ice/ice_lag.c
2615
err = ice_create_lag_recipe(&pf->hw, &lag->lport_recipe,
drivers/net/ethernet/intel/ice/ice_lag.c
2620
err = ice_create_lag_recipe(&pf->hw, &lag->act_act_recipe,
drivers/net/ethernet/intel/ice/ice_lag.c
2627
err = ice_aq_get_recipe_to_profile(&pf->hw, n,
drivers/net/ethernet/intel/ice/ice_lag.c
2636
ice_aq_map_recipe_to_profile(&pf->hw, n,
drivers/net/ethernet/intel/ice/ice_lag.c
2647
ice_free_hw_res(&pf->hw, ICE_AQC_RES_TYPE_RECIPE, 1,
drivers/net/ethernet/intel/ice/ice_lag.c
2651
ice_free_hw_res(&pf->hw, ICE_AQC_RES_TYPE_RECIPE, 1,
drivers/net/ethernet/intel/ice/ice_lag.c
2655
pf->lag = NULL;
drivers/net/ethernet/intel/ice/ice_lag.c
2666
void ice_deinit_lag(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lag.c
2668
struct ice_lag *lag = pf->lag;
drivers/net/ethernet/intel/ice/ice_lag.c
2673
if (lag->pf)
drivers/net/ethernet/intel/ice/ice_lag.c
2678
ice_free_hw_res(&pf->hw, ICE_AQC_RES_TYPE_RECIPE, 1,
drivers/net/ethernet/intel/ice/ice_lag.c
2679
&pf->lag->pf_recipe);
drivers/net/ethernet/intel/ice/ice_lag.c
2680
ice_free_hw_res(&pf->hw, ICE_AQC_RES_TYPE_RECIPE, 1,
drivers/net/ethernet/intel/ice/ice_lag.c
2681
&pf->lag->lport_recipe);
drivers/net/ethernet/intel/ice/ice_lag.c
2685
pf->lag = NULL;
drivers/net/ethernet/intel/ice/ice_lag.c
2700
void ice_lag_rebuild(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lag.c
2706
if (!pf->lag || !pf->lag->bonded)
drivers/net/ethernet/intel/ice/ice_lag.c
2709
mutex_lock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_lag.c
2711
lag = pf->lag;
drivers/net/ethernet/intel/ice/ice_lag.c
2720
dev_dbg(ice_pf_to_dev(pf), "No primary interface in aggregate, can't rebuild\n");
drivers/net/ethernet/intel/ice/ice_lag.c
2725
loc_port = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
2731
ice_lag_set_swid(prim_lag->pf->hw.port_info->sw_id, lag, true);
drivers/net/ethernet/intel/ice/ice_lag.c
2732
ice_lag_add_prune_list(prim_lag, pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2734
ice_lag_move_vf_nodes_sync(prim_lag, &pf->hw);
drivers/net/ethernet/intel/ice/ice_lag.c
2741
dev_err(ice_pf_to_dev(pf), "Error adding default VSI rule in rebuild\n");
drivers/net/ethernet/intel/ice/ice_lag.c
2747
ice_clear_rdma_cap(pf);
drivers/net/ethernet/intel/ice/ice_lag.c
2750
mutex_unlock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_lag.c
2759
bool ice_lag_is_switchdev_running(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lag.c
2761
struct ice_lag *lag = pf->lag;
drivers/net/ethernet/intel/ice/ice_lag.c
2764
if (!ice_is_feature_supported(pf, ICE_F_SRIOV_LAG) ||
drivers/net/ethernet/intel/ice/ice_lag.c
279
err = ice_aq_sw_rules(&lag->pf->hw, s_rule, s_rule_sz, 1, opc, NULL);
drivers/net/ethernet/intel/ice/ice_lag.c
351
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
381
struct ice_vsi *vsi = lag->pf->vsi[0];
drivers/net/ethernet/intel/ice/ice_lag.c
39
struct ice_pf *pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
41
if (!pf)
drivers/net/ethernet/intel/ice/ice_lag.c
419
if (ice_aq_sw_rules(&lag->pf->hw, s_rule, buf_len, 1, opc, NULL)) {
drivers/net/ethernet/intel/ice/ice_lag.c
45
dev_warn(ice_pf_to_dev(pf), "%s: Attempt to be Primary, but incompatible state.\n",
drivers/net/ethernet/intel/ice/ice_lag.c
477
struct device *dev = &lag->pf->pdev->dev;
drivers/net/ethernet/intel/ice/ice_lag.c
519
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_lag.c
542
qid = pf->vsi[vsi_num]->txq_map[q_ctx->q_handle];
drivers/net/ethernet/intel/ice/ice_lag.c
59
struct ice_pf *pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
61
if (!pf)
drivers/net/ethernet/intel/ice/ice_lag.c
634
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
644
ctx = ice_get_vsi_ctx(&lag->pf->hw, vsi_num);
drivers/net/ethernet/intel/ice/ice_lag.c
65
dev_dbg(ice_pf_to_dev(pf), "%s: Attempt to be Backup, but incompatible state\n",
drivers/net/ethernet/intel/ice/ice_lag.c
672
if (ice_sched_suspend_resume_elems(&lag->pf->hw, 1, &tmp_teid, true))
drivers/net/ethernet/intel/ice/ice_lag.c
686
valq = ice_lag_qbuf_recfg(&lag->pf->hw, qbuf, vsi_num, numq, tc);
drivers/net/ethernet/intel/ice/ice_lag.c
692
if (ice_aq_cfg_lan_txq(&lag->pf->hw, qbuf, qbuf_size, valq, oldport,
drivers/net/ethernet/intel/ice/ice_lag.c
715
if (ice_aq_move_sched_elems(&lag->pf->hw, buf, buf_size, &num_moved))
drivers/net/ethernet/intel/ice/ice_lag.c
727
if (ice_sched_suspend_resume_elems(&lag->pf->hw, 1, &tmp_teid, false))
drivers/net/ethernet/intel/ice/ice_lag.c
78
static bool netif_is_same_ice(struct ice_pf *pf, struct net_device *netdev)
drivers/net/ethernet/intel/ice/ice_lag.c
800
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lag.c
806
pf = lag->pf;
drivers/net/ethernet/intel/ice/ice_lag.c
807
ice_for_each_vsi(pf, i)
drivers/net/ethernet/intel/ice/ice_lag.c
808
if (pf->vsi[i] && pf->vsi[i]->type == ICE_VSI_VF)
drivers/net/ethernet/intel/ice/ice_lag.c
849
pri_prt = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
881
pri_prt = lag->pf->hw.port_info->lport;
drivers/net/ethernet/intel/ice/ice_lag.c
944
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_lag.c
964
q_id = pf->vsi[vsi_num]->txq_map[q_ctx->q_handle];
drivers/net/ethernet/intel/ice/ice_lag.c
989
struct device *dev = ice_pf_to_dev(lag->pf);
drivers/net/ethernet/intel/ice/ice_lag.c
99
if (pf->pdev->bus != test_pf->pdev->bus ||
drivers/net/ethernet/intel/ice/ice_lag.c
997
vf_id = lag->pf->vsi[vsi_num]->vf->vf_id;
drivers/net/ethernet/intel/ice/ice_lag.h
38
struct ice_pf *pf; /* backlink to PF struct */
drivers/net/ethernet/intel/ice/ice_lag.h
86
int ice_init_lag(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lag.h
87
void ice_deinit_lag(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lag.h
88
void ice_lag_rebuild(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lag.h
89
bool ice_lag_is_switchdev_running(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1158
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lib.c
1160
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1161
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1231
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1232
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lib.c
1237
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1268
main_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1278
if (test_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
1285
if (test_bit(ICE_FLAG_RSS_ENA, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_lib.c
1392
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1396
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1410
ring->tx_tstamps = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_lib.c
1435
ring->cached_phctime = pf->ptp.cached_phc_time;
drivers/net/ethernet/intel/ice/ice_lib.c
1437
if (ice_is_feature_supported(pf, ICE_F_GCS))
drivers/net/ethernet/intel/ice/ice_lib.c
1501
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1506
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1508
(test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))) {
drivers/net/ethernet/intel/ice/ice_lib.c
156
static u16 ice_get_rxq_count(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
1573
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1577
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1578
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_lib.c
158
return min(ice_get_avail_rxq_count(pf),
drivers/net/ethernet/intel/ice/ice_lib.c
1584
status = ice_add_avf_rss_cfg(&pf->hw, vsi, ICE_DEFAULT_RSS_HASHCFG);
drivers/net/ethernet/intel/ice/ice_lib.c
162
static u16 ice_get_txq_count(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
164
return min(ice_get_avail_txq_count(pf),
drivers/net/ethernet/intel/ice/ice_lib.c
1673
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1674
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lib.c
1679
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1680
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_lib.c
1706
bool ice_pf_state_is_nominal(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
1710
if (!pf)
drivers/net/ethernet/intel/ice/ice_lib.c
1714
if (bitmap_intersects(pf->state, check_bits, ICE_STATE_NBITS))
drivers/net/ethernet/intel/ice/ice_lib.c
1734
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1740
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_lib.c
177
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
189
vsi->alloc_txq = ice_get_txq_count(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
1917
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
1918
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lib.c
192
pf->num_lan_tx = vsi->alloc_txq;
drivers/net/ethernet/intel/ice/ice_lib.c
195
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
202
vsi->alloc_rxq = ice_get_rxq_count(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2040
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2041
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lib.c
206
pf->num_lan_rx = vsi->alloc_rxq;
drivers/net/ethernet/intel/ice/ice_lib.c
2082
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2086
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2093
if (!test_bit(ICE_FLAG_LLDP_AQ_FLTR, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
2104
status = ice_lldp_fltr_add_remove(&pf->hw, vsi, create);
drivers/net/ethernet/intel/ice/ice_lib.c
2106
set_bit(ICE_FLAG_LLDP_AQ_FLTR, pf->flags);
drivers/net/ethernet/intel/ice/ice_lib.c
2125
void ice_cfg_sw_rx_lldp(struct ice_pf *pf, bool enable)
drivers/net/ethernet/intel/ice/ice_lib.c
2131
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2134
if (!test_bit(ICE_FLAG_SRIOV_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
2137
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_lib.c
2163
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2172
port_info = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_lib.c
2184
agg_node_iter = &pf->pf_agg_node[0];
drivers/net/ethernet/intel/ice/ice_lib.c
2195
agg_node_iter = &pf->vf_agg_node[0];
drivers/net/ethernet/intel/ice/ice_lib.c
2270
static int ice_vsi_cfg_tc_lan(struct ice_pf *pf, struct ice_vsi *vsi)
drivers/net/ethernet/intel/ice/ice_lib.c
2273
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2285
max_txqs[i] = pf->num_lan_tx;
drivers/net/ethernet/intel/ice/ice_lib.c
2313
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2316
vsi->vsw = pf->first_sw;
drivers/net/ethernet/intel/ice/ice_lib.c
2385
if (test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
2392
if (test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
2421
if (test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
244
dev_warn(ice_pf_to_dev(pf), "Unknown VSI type %d\n", vsi_type);
drivers/net/ethernet/intel/ice/ice_lib.c
2472
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2491
WARN_ON(pf->ctrl_vsi_idx != ICE_NO_VSI);
drivers/net/ethernet/intel/ice/ice_lib.c
2492
pf->ctrl_vsi_idx = vsi->idx;
drivers/net/ethernet/intel/ice/ice_lib.c
2505
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2511
dev_err(ice_pf_to_dev(pf), "Failed to remove RDMA scheduler config for VSI %u, err %d\n",
drivers/net/ethernet/intel/ice/ice_lib.c
2547
ice_vsi_setup(struct ice_pf *pf, struct ice_vsi_cfg_params *params)
drivers/net/ethernet/intel/ice/ice_lib.c
2549
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2560
vsi = ice_vsi_alloc(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
2580
if (!ice_is_safe_mode(pf) && vsi->type == ICE_VSI_PF) {
drivers/net/ethernet/intel/ice/ice_lib.c
2603
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2604
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_lib.c
2640
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2664
devm_free_irq(ice_pf_to_dev(pf), irq_num, vsi->q_vectors[i]);
drivers/net/ethernet/intel/ice/ice_lib.c
286
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2868
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lib.c
2872
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
2874
if (test_bit(ICE_FLAG_RSS_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
2882
if (!ice_is_safe_mode(pf) &&
drivers/net/ethernet/intel/ice/ice_lib.c
2883
!test_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_lib.c
2894
if (!ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_lib.c
301
status = ice_free_vsi(&pf->hw, vsi->idx, ctxt, false, NULL);
drivers/net/ethernet/intel/ice/ice_lib.c
3026
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
303
dev_err(ice_pf_to_dev(pf), "Failed to delete VSI %i in FW - error: %d\n",
drivers/net/ethernet/intel/ice/ice_lib.c
3031
vsi_stat = pf->vsi_stats[vsi->idx];
drivers/net/ethernet/intel/ice/ice_lib.c
3088
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_lib.c
3095
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3118
ret = ice_vsi_cfg_tc_lan(pf, vsi);
drivers/net/ethernet/intel/ice/ice_lib.c
3125
ret = ice_schedule_reset(pf, ICE_RESET_PFR);
drivers/net/ethernet/intel/ice/ice_lib.c
315
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3167
int ice_wait_for_reset(struct ice_pf *pf, unsigned long timeout)
drivers/net/ethernet/intel/ice/ice_lib.c
3171
ret = wait_event_interruptible_timeout(pf->reset_wait_queue,
drivers/net/ethernet/intel/ice/ice_lib.c
3172
!ice_is_reset_in_progress(pf->state),
drivers/net/ethernet/intel/ice/ice_lib.c
318
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
3204
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3222
if (vsi->type == ICE_VSI_PF && ice_is_adq_active(pf))
drivers/net/ethernet/intel/ice/ice_lib.c
3228
dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_lib.c
3247
if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
3362
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3369
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
3384
test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
340
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3400
test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
3412
ret = ice_update_vsi(&pf->hw, vsi->idx, ctx, NULL);
drivers/net/ethernet/intel/ice/ice_lib.c
3419
test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
345
if (!pf->vsi_stats)
drivers/net/ethernet/intel/ice/ice_lib.c
348
vsi_stat = pf->vsi_stats[vsi->idx];
drivers/net/ethernet/intel/ice/ice_lib.c
3646
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3651
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
369
pf->vsi_stats[vsi->idx] = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
3707
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3712
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
381
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
384
vsi_stats = pf->vsi_stats[vsi->idx];
drivers/net/ethernet/intel/ice/ice_lib.c
3844
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
3848
if (pf->lag && pf->lag->primary) {
drivers/net/ethernet/intel/ice/ice_lib.c
3849
dev_dbg(ice_pf_to_dev(pf), "Interface is primary in aggregate - not deleting prune list\n");
drivers/net/ethernet/intel/ice/ice_lib.c
3861
if (pf->lag && pf->lag->primary) {
drivers/net/ethernet/intel/ice/ice_lib.c
3862
dev_dbg(ice_pf_to_dev(pf), "Interface is primary in aggregate - not deleting QinQ prune list\n");
drivers/net/ethernet/intel/ice/ice_lib.c
3928
bool ice_is_feature_supported(struct ice_pf *pf, enum ice_feature f)
drivers/net/ethernet/intel/ice/ice_lib.c
3933
return test_bit(f, pf->features);
drivers/net/ethernet/intel/ice/ice_lib.c
3941
void ice_set_feature_support(struct ice_pf *pf, enum ice_feature f)
drivers/net/ethernet/intel/ice/ice_lib.c
3946
set_bit(f, pf->features);
drivers/net/ethernet/intel/ice/ice_lib.c
3954
void ice_clear_feature_support(struct ice_pf *pf, enum ice_feature f)
drivers/net/ethernet/intel/ice/ice_lib.c
3959
clear_bit(f, pf->features);
drivers/net/ethernet/intel/ice/ice_lib.c
3968
void ice_init_feature_support(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
3970
switch (pf->hw.device_id) {
drivers/net/ethernet/intel/ice/ice_lib.c
3977
ice_set_feature_support(pf, ICE_F_DSCP);
drivers/net/ethernet/intel/ice/ice_lib.c
3978
if (ice_is_phy_rclk_in_netlist(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lib.c
3979
ice_set_feature_support(pf, ICE_F_PHY_RCLK);
drivers/net/ethernet/intel/ice/ice_lib.c
3981
if (!ice_pf_src_tmr_owned(pf))
drivers/net/ethernet/intel/ice/ice_lib.c
3983
if (ice_is_cgu_in_netlist(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lib.c
3984
ice_set_feature_support(pf, ICE_F_CGU);
drivers/net/ethernet/intel/ice/ice_lib.c
3985
if (ice_is_clock_mux_in_netlist(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lib.c
3986
ice_set_feature_support(pf, ICE_F_SMA_CTRL);
drivers/net/ethernet/intel/ice/ice_lib.c
3987
if (ice_gnss_is_module_present(&pf->hw))
drivers/net/ethernet/intel/ice/ice_lib.c
3988
ice_set_feature_support(pf, ICE_F_GNSS);
drivers/net/ethernet/intel/ice/ice_lib.c
3994
if (pf->hw.mac_type == ICE_MAC_GENERIC_3K_E825)
drivers/net/ethernet/intel/ice/ice_lib.c
3995
ice_set_feature_support(pf, ICE_F_PHY_RCLK);
drivers/net/ethernet/intel/ice/ice_lib.c
3997
if (pf->hw.mac_type == ICE_MAC_E830) {
drivers/net/ethernet/intel/ice/ice_lib.c
3998
ice_set_feature_support(pf, ICE_F_MBX_LIMIT);
drivers/net/ethernet/intel/ice/ice_lib.c
3999
ice_set_feature_support(pf, ICE_F_GCS);
drivers/net/ethernet/intel/ice/ice_lib.c
4000
ice_set_feature_support(pf, ICE_F_TXTIME);
drivers/net/ethernet/intel/ice/ice_lib.c
446
struct ice_pf *pf = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
452
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
453
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
455
if (!pf->vsi[vsi->idx] || pf->vsi[vsi->idx] != vsi) {
drivers/net/ethernet/intel/ice/ice_lib.c
460
mutex_lock(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
463
pf->vsi[vsi->idx] = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
464
pf->next_vsi = vsi->idx;
drivers/net/ethernet/intel/ice/ice_lib.c
469
mutex_unlock(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
523
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
527
if (!pf->vsi_stats)
drivers/net/ethernet/intel/ice/ice_lib.c
530
if (pf->vsi_stats[vsi->idx])
drivers/net/ethernet/intel/ice/ice_lib.c
548
pf->vsi_stats[vsi->idx] = vsi_stat;
drivers/net/ethernet/intel/ice/ice_lib.c
557
pf->vsi_stats[vsi->idx] = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
616
struct ice_vsi *ice_vsi_alloc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
618
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
622
mutex_lock(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
628
if (pf->next_vsi == ICE_NO_VSI) {
drivers/net/ethernet/intel/ice/ice_lib.c
637
vsi->back = pf;
drivers/net/ethernet/intel/ice/ice_lib.c
641
vsi->idx = pf->next_vsi;
drivers/net/ethernet/intel/ice/ice_lib.c
642
pf->vsi[pf->next_vsi] = vsi;
drivers/net/ethernet/intel/ice/ice_lib.c
645
pf->next_vsi = ice_get_free_slot(pf->vsi, pf->num_alloc_vsi,
drivers/net/ethernet/intel/ice/ice_lib.c
646
pf->next_vsi);
drivers/net/ethernet/intel/ice/ice_lib.c
651
mutex_unlock(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
665
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
672
if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_lib.c
680
g_val = pf->hw.func_caps.fd_fltr_guar;
drivers/net/ethernet/intel/ice/ice_lib.c
685
b_val = pf->hw.func_caps.fd_fltr_best_effort;
drivers/net/ethernet/intel/ice/ice_lib.c
702
if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
720
main_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
75
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
755
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
757
.qs_mutex = &pf->avail_q_mutex,
drivers/net/ethernet/intel/ice/ice_lib.c
758
.pf_map = pf->avail_txqs,
drivers/net/ethernet/intel/ice/ice_lib.c
759
.pf_map_size = pf->max_pf_txqs,
drivers/net/ethernet/intel/ice/ice_lib.c
767
.qs_mutex = &pf->avail_q_mutex,
drivers/net/ethernet/intel/ice/ice_lib.c
768
.pf_map = pf->avail_rxqs,
drivers/net/ethernet/intel/ice/ice_lib.c
769
.pf_map_size = pf->max_pf_rxqs,
drivers/net/ethernet/intel/ice/ice_lib.c
78
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
800
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
803
mutex_lock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
806
clear_bit(vsi->txq_map[i], pf->avail_txqs);
drivers/net/ethernet/intel/ice/ice_lib.c
811
clear_bit(vsi->rxq_map[i], pf->avail_rxqs);
drivers/net/ethernet/intel/ice/ice_lib.c
815
mutex_unlock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_lib.c
824
bool ice_is_safe_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
826
return !test_bit(ICE_FLAG_ADV_FEATURES, pf->flags);
drivers/net/ethernet/intel/ice/ice_lib.c
835
bool ice_is_rdma_ena(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_lib.c
840
err = devl_param_driverinit_value_get(priv_to_devlink(pf),
drivers/net/ethernet/intel/ice/ice_lib.c
843
return err ? test_bit(ICE_FLAG_RDMA_ENA, pf->flags) : value.vbool;
drivers/net/ethernet/intel/ice/ice_lib.c
855
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
858
if (ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_lib.c
861
status = ice_rem_vsi_rss_cfg(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_lib.c
863
dev_dbg(ice_pf_to_dev(pf), "ice_rem_vsi_rss_cfg failed for vsi = %d, error = %d\n",
drivers/net/ethernet/intel/ice/ice_lib.c
873
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
876
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_lib.c
883
if (!ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_lib.c
884
ice_rem_vsi_rss_list(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_lib.c
894
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_lib.c
897
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_lib.c
902
cap = &pf->hw.func_caps.common_cap;
drivers/net/ethernet/intel/ice/ice_lib.c
934
dev_dbg(ice_pf_to_dev(pf), "Unsupported VSI type %s\n",
drivers/net/ethernet/intel/ice/ice_lib.h
105
bool ice_is_safe_mode(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.h
106
bool ice_is_rdma_ena(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.h
127
bool ice_is_feature_supported(struct ice_pf *pf, enum ice_feature f);
drivers/net/ethernet/intel/ice/ice_lib.h
128
void ice_set_feature_support(struct ice_pf *pf, enum ice_feature f);
drivers/net/ethernet/intel/ice/ice_lib.h
129
void ice_clear_feature_support(struct ice_pf *pf, enum ice_feature f);
drivers/net/ethernet/intel/ice/ice_lib.h
130
void ice_init_feature_support(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.h
23
bool ice_pf_state_is_nominal(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.h
40
void ice_cfg_sw_rx_lldp(struct ice_pf *pf, bool enable);
drivers/net/ethernet/intel/ice/ice_lib.h
53
ice_vsi_setup(struct ice_pf *pf, struct ice_vsi_cfg_params *params);
drivers/net/ethernet/intel/ice/ice_lib.h
71
struct ice_vsi *ice_vsi_alloc(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_lib.h
75
int ice_wait_for_reset(struct ice_pf *pf, unsigned long timeout);
drivers/net/ethernet/intel/ice/ice_main.c
1031
static void ice_check_phy_fw_load(struct ice_pf *pf, u8 link_cfg_err)
drivers/net/ethernet/intel/ice/ice_main.c
1034
clear_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1038
if (test_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
1042
dev_err(ice_pf_to_dev(pf), "Device failed to load the FW for the external PHY. Please download and install the latest NVM for your device and try again\n");
drivers/net/ethernet/intel/ice/ice_main.c
1043
set_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1055
static void ice_check_module_power(struct ice_pf *pf, u8 link_cfg_err)
drivers/net/ethernet/intel/ice/ice_main.c
1060
clear_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1067
if (test_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
1071
dev_err(ice_pf_to_dev(pf), "The installed module is incompatible with the device's NVM image. Cannot start link\n");
drivers/net/ethernet/intel/ice/ice_main.c
1072
set_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1074
dev_err(ice_pf_to_dev(pf), "The module's power requirements exceed the device's power supply. Cannot start link\n");
drivers/net/ethernet/intel/ice/ice_main.c
1075
set_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1087
static void ice_check_link_cfg_err(struct ice_pf *pf, u8 link_cfg_err)
drivers/net/ethernet/intel/ice/ice_main.c
1089
ice_check_module_power(pf, link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
1090
ice_check_phy_fw_load(pf, link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
1103
ice_link_event(struct ice_pf *pf, struct ice_port_info *pi, bool link_up,
drivers/net/ethernet/intel/ice/ice_main.c
1106
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1128
ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
1136
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1141
if (!test_bit(ICE_FLAG_NO_MEDIA, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_main.c
1143
set_bit(ICE_FLAG_NO_MEDIA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
1152
pf->link_down_events++;
drivers/net/ethernet/intel/ice/ice_main.c
1154
ice_ptp_link_change(pf, link_up);
drivers/net/ethernet/intel/ice/ice_main.c
1156
if (ice_is_dcb_active(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
1157
if (test_bit(ICE_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
1158
ice_dcb_rebuild(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1161
ice_set_dflt_mib(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1166
ice_vc_notify_link_state(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1175
static void ice_watchdog_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1180
if (test_bit(ICE_DOWN, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
1181
test_bit(ICE_CFG_BUSY, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1186
pf->serv_tmr_prev + pf->serv_tmr_period))
drivers/net/ethernet/intel/ice/ice_main.c
1189
pf->serv_tmr_prev = jiffies;
drivers/net/ethernet/intel/ice/ice_main.c
119
static void ice_check_for_hang_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1194
ice_update_pf_stats(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1195
ice_for_each_vsi(pf, i)
drivers/net/ethernet/intel/ice/ice_main.c
1196
if (pf->vsi[i] && pf->vsi[i]->netdev)
drivers/net/ethernet/intel/ice/ice_main.c
1197
ice_update_vsi_stats(pf->vsi[i]);
drivers/net/ethernet/intel/ice/ice_main.c
1235
ice_handle_link_event(struct ice_pf *pf, struct ice_rq_event_info *event)
drivers/net/ethernet/intel/ice/ice_main.c
1242
port_info = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_main.c
1246
status = ice_link_event(pf, port_info,
drivers/net/ethernet/intel/ice/ice_main.c
1250
dev_dbg(ice_pf_to_dev(pf), "Could not process link event, error %d\n",
drivers/net/ethernet/intel/ice/ice_main.c
127
ice_for_each_vsi(pf, v)
drivers/net/ethernet/intel/ice/ice_main.c
1272
void ice_aq_prep_for_event(struct ice_pf *pf, struct ice_aq_task *task,
drivers/net/ethernet/intel/ice/ice_main.c
1279
spin_lock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
128
if (pf->vsi[v] && pf->vsi[v]->type == ICE_VSI_PF) {
drivers/net/ethernet/intel/ice/ice_main.c
1280
hlist_add_head(&task->entry, &pf->aq_wait_list);
drivers/net/ethernet/intel/ice/ice_main.c
1281
spin_unlock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
129
vsi = pf->vsi[v];
drivers/net/ethernet/intel/ice/ice_main.c
1296
int ice_aq_wait_for_event(struct ice_pf *pf, struct ice_aq_task *task,
drivers/net/ethernet/intel/ice/ice_main.c
1300
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1305
ret = wait_event_interruptible_timeout(pf->aq_wait_queue,
drivers/net/ethernet/intel/ice/ice_main.c
1333
spin_lock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1335
spin_unlock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1358
static void ice_aq_check_events(struct ice_pf *pf, u16 opcode,
drivers/net/ethernet/intel/ice/ice_main.c
1365
spin_lock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1366
hlist_for_each_entry(task, &pf->aq_wait_list, entry) {
drivers/net/ethernet/intel/ice/ice_main.c
1386
spin_unlock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1389
wake_up(&pf->aq_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
1399
static void ice_aq_cancel_waiting_tasks(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1403
spin_lock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1404
hlist_for_each_entry(task, &pf->aq_wait_list, entry)
drivers/net/ethernet/intel/ice/ice_main.c
1406
spin_unlock_bh(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1408
wake_up(&pf->aq_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
1418
static int __ice_clean_ctrlq(struct ice_pf *pf, enum ice_ctl_q q_type)
drivers/net/ethernet/intel/ice/ice_main.c
1420
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1422
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
1429
if (test_bit(ICE_RESET_FAILED, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1519
ice_aq_check_events(pf, opcode, &event);
drivers/net/ethernet/intel/ice/ice_main.c
1523
if (ice_handle_link_event(pf, &event))
drivers/net/ethernet/intel/ice/ice_main.c
1527
ice_vf_lan_overflow_event(pf, &event);
drivers/net/ethernet/intel/ice/ice_main.c
1530
if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT)) {
drivers/net/ethernet/intel/ice/ice_main.c
1531
ice_vc_process_vf_msg(pf, &event, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
1542
ice_vc_process_vf_msg(pf, &event, &data);
drivers/net/ethernet/intel/ice/ice_main.c
1550
ice_dcb_process_lldp_set_mib_change(pf, &event);
drivers/net/ethernet/intel/ice/ice_main.c
1553
ice_process_health_status_event(pf, &event);
drivers/net/ethernet/intel/ice/ice_main.c
1586
static void ice_clean_adminq_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1588
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
1590
if (!test_bit(ICE_ADMINQ_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1593
if (__ice_clean_ctrlq(pf, ICE_CTL_Q_ADMIN))
drivers/net/ethernet/intel/ice/ice_main.c
1596
clear_bit(ICE_ADMINQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1604
__ice_clean_ctrlq(pf, ICE_CTL_Q_ADMIN);
drivers/net/ethernet/intel/ice/ice_main.c
1613
static void ice_clean_mailboxq_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1615
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
1617
if (!test_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1620
if (__ice_clean_ctrlq(pf, ICE_CTL_Q_MAILBOX))
drivers/net/ethernet/intel/ice/ice_main.c
1623
clear_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1626
__ice_clean_ctrlq(pf, ICE_CTL_Q_MAILBOX);
drivers/net/ethernet/intel/ice/ice_main.c
1635
static void ice_clean_sbq_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1637
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
1643
clear_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1647
if (!test_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1650
if (__ice_clean_ctrlq(pf, ICE_CTL_Q_SB))
drivers/net/ethernet/intel/ice/ice_main.c
1653
clear_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1656
__ice_clean_ctrlq(pf, ICE_CTL_Q_SB);
drivers/net/ethernet/intel/ice/ice_main.c
1667
void ice_service_task_schedule(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1669
if (!test_bit(ICE_SERVICE_DIS, pf->state) &&
drivers/net/ethernet/intel/ice/ice_main.c
1670
!test_and_set_bit(ICE_SERVICE_SCHED, pf->state) &&
drivers/net/ethernet/intel/ice/ice_main.c
1671
!test_bit(ICE_NEEDS_RESTART, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
1672
queue_work(ice_wq, &pf->serv_task);
drivers/net/ethernet/intel/ice/ice_main.c
1679
static void ice_service_task_complete(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1681
WARN_ON(!test_bit(ICE_SERVICE_SCHED, pf->state));
drivers/net/ethernet/intel/ice/ice_main.c
1685
clear_bit(ICE_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1695
static int ice_service_task_stop(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1699
ret = test_and_set_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1701
if (pf->serv_tmr.function)
drivers/net/ethernet/intel/ice/ice_main.c
1702
timer_delete_sync(&pf->serv_tmr);
drivers/net/ethernet/intel/ice/ice_main.c
1703
if (pf->serv_task.func)
drivers/net/ethernet/intel/ice/ice_main.c
1704
cancel_work_sync(&pf->serv_task);
drivers/net/ethernet/intel/ice/ice_main.c
1706
clear_bit(ICE_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1716
static void ice_service_task_restart(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1718
clear_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1719
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1728
struct ice_pf *pf = timer_container_of(pf, t, serv_tmr);
drivers/net/ethernet/intel/ice/ice_main.c
1730
mod_timer(&pf->serv_tmr, round_jiffies(pf->serv_tmr_period + jiffies));
drivers/net/ethernet/intel/ice/ice_main.c
1731
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1745
static void ice_mdd_maybe_reset_vf(struct ice_pf *pf, struct ice_vf *vf,
drivers/net/ethernet/intel/ice/ice_main.c
1748
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1750
if (!test_bit(ICE_FLAG_MDD_AUTO_RESET_VF, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
1763
pf->hw.pf_id, vf->vf_id);
drivers/net/ethernet/intel/ice/ice_main.c
1777
static void ice_handle_mdd_event(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1779
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1780
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
1785
if (!test_and_clear_bit(ICE_MDD_EVENT_PENDING, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
1789
ice_print_vfs_mdd_events(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1801
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1804
ice_report_mdd_event(pf, ICE_MDD_SRC_TX_PQM, pf_num, vf_num,
drivers/net/ethernet/intel/ice/ice_main.c
1816
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1819
ice_report_mdd_event(pf, ICE_MDD_SRC_TX_TCLAN, pf_num, vf_num,
drivers/net/ethernet/intel/ice/ice_main.c
1831
if (netif_msg_rx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1834
ice_report_mdd_event(pf, ICE_MDD_SRC_RX, pf_num, vf_num, event,
drivers/net/ethernet/intel/ice/ice_main.c
1843
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1850
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1857
if (netif_msg_rx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1864
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1865
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_main.c
187
static int ice_init_mac_fltr(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
1872
set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1873
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1884
set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1885
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1896
set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1897
if (netif_msg_tx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1908
set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
1909
if (netif_msg_rx_err(pf))
drivers/net/ethernet/intel/ice/ice_main.c
1917
ice_mdd_maybe_reset_vf(pf, vf, reset_vf_tx,
drivers/net/ethernet/intel/ice/ice_main.c
192
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
1920
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
1922
ice_print_vfs_mdd_events(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2010
struct ice_pf *pf = pi->hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
2021
dev_err(ice_pf_to_dev(pf), "Get PHY capability failed.\n");
drivers/net/ethernet/intel/ice/ice_main.c
2025
pf->nvm_phy_type_hi = pcaps->phy_type_high;
drivers/net/ethernet/intel/ice/ice_main.c
2026
pf->nvm_phy_type_lo = pcaps->phy_type_low;
drivers/net/ethernet/intel/ice/ice_main.c
2042
struct ice_pf *pf = pi->hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
2044
ldo = &pf->link_dflt_override;
drivers/net/ethernet/intel/ice/ice_main.c
2054
set_bit(ICE_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
2055
set_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
2080
struct ice_pf *pf = pi->hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
2082
ldo = &pf->link_dflt_override;
drivers/net/ethernet/intel/ice/ice_main.c
2090
cfg->phy_type_low = pf->nvm_phy_type_lo &
drivers/net/ethernet/intel/ice/ice_main.c
2092
cfg->phy_type_high = pf->nvm_phy_type_hi &
drivers/net/ethernet/intel/ice/ice_main.c
2098
set_bit(ICE_LINK_DEFAULT_OVERRIDE_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2119
struct ice_pf *pf = pi->hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
2136
dev_err(ice_pf_to_dev(pf), "Get PHY capability failed.\n");
drivers/net/ethernet/intel/ice/ice_main.c
2146
set_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
2153
(pf->link_dflt_override.options & ICE_LINK_OVERRIDE_EN)) {
drivers/net/ethernet/intel/ice/ice_main.c
2168
set_bit(ICE_PHY_INIT_COMPLETE, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2189
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
2198
if (!test_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_main.c
2202
if (test_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
2288
err = ice_aq_set_phy_cfg(&pf->hw, pi, cfg, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
2306
static void ice_check_media_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
2313
if (!test_bit(ICE_FLAG_NO_MEDIA, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
2316
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2326
ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
2329
if (!test_bit(ICE_PHY_INIT_COMPLETE, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
2341
clear_bit(ICE_FLAG_NO_MEDIA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
2351
struct ice_pf *pf = container_of(work, struct ice_pf, serv_task);
drivers/net/ethernet/intel/ice/ice_main.c
2353
set_bit(ICE_ADMINQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2354
ice_clean_adminq_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2356
ice_service_task_complete(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2358
mod_timer(&pf->serv_tmr, jiffies + msecs_to_jiffies(100));
drivers/net/ethernet/intel/ice/ice_main.c
2367
struct ice_pf *pf = container_of(work, struct ice_pf, serv_task);
drivers/net/ethernet/intel/ice/ice_main.c
2370
if (pf->health_reporters.tx_hang_buf.tx_ring) {
drivers/net/ethernet/intel/ice/ice_main.c
2371
ice_report_tx_hang(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2372
pf->health_reporters.tx_hang_buf.tx_ring = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
2375
ice_reset_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2378
if (ice_is_reset_in_progress(pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2379
test_bit(ICE_SUSPENDED, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2380
test_bit(ICE_NEEDS_RESTART, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
2381
ice_service_task_complete(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2385
if (test_and_clear_bit(ICE_AUX_ERR_PENDING, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
2392
swap(event->reg, pf->oicr_err_reg);
drivers/net/ethernet/intel/ice/ice_main.c
2393
ice_send_event_to_aux(pf, event);
drivers/net/ethernet/intel/ice/ice_main.c
2401
if (test_and_clear_bit(ICE_FLAG_UNPLUG_AUX_DEV, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
2402
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2405
if (test_and_clear_bit(ICE_FLAG_PLUG_AUX_DEV, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
2406
ice_plug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2408
if (test_and_clear_bit(ICE_FLAG_MTU_CHANGED, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
2414
ice_send_event_to_aux(pf, event);
drivers/net/ethernet/intel/ice/ice_main.c
2419
ice_clean_adminq_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2420
ice_check_media_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2421
ice_check_for_hang_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2422
ice_sync_fltr_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2423
ice_handle_mdd_event(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2424
ice_watchdog_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2426
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
2427
ice_service_task_complete(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2431
ice_process_vflr_event(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2432
ice_clean_mailboxq_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2433
ice_clean_sbq_subtask(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2434
ice_sync_arfs_fltrs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2435
ice_flush_fdir_ctx(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2438
ice_service_task_complete(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2444
if (time_after(jiffies, (start_time + pf->serv_tmr_period)) ||
drivers/net/ethernet/intel/ice/ice_main.c
2445
test_bit(ICE_MDD_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2446
test_bit(ICE_VFLR_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2447
test_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2448
test_bit(ICE_FD_VF_FLUSH_CTX, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2449
test_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
2450
test_bit(ICE_ADMINQ_EVENT_PENDING, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
2451
mod_timer(&pf->serv_tmr, jiffies);
drivers/net/ethernet/intel/ice/ice_main.c
2479
int ice_schedule_reset(struct ice_pf *pf, enum ice_reset_req reset)
drivers/net/ethernet/intel/ice/ice_main.c
2481
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2484
if (test_bit(ICE_RESET_FAILED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
2489
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
2496
set_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2499
set_bit(ICE_CORER_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2502
set_bit(ICE_GLOBR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
2508
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2536
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
2543
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2758
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
2760
.qs_mutex = &pf->avail_q_mutex,
drivers/net/ethernet/intel/ice/ice_main.c
2761
.pf_map = pf->avail_txqs,
drivers/net/ethernet/intel/ice/ice_main.c
2762
.pf_map_size = pf->max_pf_txqs,
drivers/net/ethernet/intel/ice/ice_main.c
2772
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
2835
mutex_lock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
2837
clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
2840
mutex_unlock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
2859
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
2871
mutex_lock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
2873
clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
2876
mutex_unlock(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
2890
devm_kfree(ice_pf_to_dev(pf), vsi->xdp_rings);
drivers/net/ethernet/intel/ice/ice_main.c
3088
static void ice_ena_misc_vector(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3090
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3118
wr32(hw, GLINT_DYN_CTL(pf->oicr_irq.index),
drivers/net/ethernet/intel/ice/ice_main.c
3121
if (!pf->hw.dev_caps.ts_dev_info.ts_ll_int_read)
drivers/net/ethernet/intel/ice/ice_main.c
3124
wr32(hw, GLINT_DYN_CTL(pf->ll_ts_irq.index + pf_intr_start_offset),
drivers/net/ethernet/intel/ice/ice_main.c
3135
struct ice_pf *pf = data;
drivers/net/ethernet/intel/ice/ice_main.c
3143
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3144
tx = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_main.c
3159
wr32(hw, GLINT_DYN_CTL(pf->ll_ts_irq.index + pf_intr_start_offset),
drivers/net/ethernet/intel/ice/ice_main.c
3172
struct ice_pf *pf = (struct ice_pf *)data;
drivers/net/ethernet/intel/ice/ice_main.c
3174
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3178
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3179
set_bit(ICE_ADMINQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3180
set_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3181
set_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3188
pf->sw_int_count++;
drivers/net/ethernet/intel/ice/ice_main.c
3193
set_bit(ICE_MDD_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3197
if (test_bit(ICE_VF_RESETS_DISABLED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
3204
set_bit(ICE_VFLR_EVENT_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3217
pf->corer_count++;
drivers/net/ethernet/intel/ice/ice_main.c
3219
pf->globr_count++;
drivers/net/ethernet/intel/ice/ice_main.c
3221
pf->empr_count++;
drivers/net/ethernet/intel/ice/ice_main.c
3228
if (!test_and_set_bit(ICE_RESET_OICR_RECV, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
3230
set_bit(ICE_CORER_RECV, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3232
set_bit(ICE_GLOBR_RECV, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3234
set_bit(ICE_EMPR_RECV, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3256
ret = ice_ptp_ts_irq(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3265
if (ice_pf_src_tmr_owned(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
3267
pf->ptp.ext_ts_irq |= gltsyn_stat &
drivers/net/ethernet/intel/ice/ice_main.c
3272
ice_ptp_extts_event(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3278
pf->oicr_err_reg |= oicr;
drivers/net/ethernet/intel/ice/ice_main.c
3279
set_bit(ICE_AUX_ERR_PENDING, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3292
set_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
3295
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3309
struct ice_pf *pf = data;
drivers/net/ethernet/intel/ice/ice_main.c
3312
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3314
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
3317
if (test_and_clear_bit(ICE_MISC_THREAD_TX_TSTAMP, pf->misc_thread))
drivers/net/ethernet/intel/ice/ice_main.c
3318
ice_ptp_process_ts(pf);
drivers/net/ethernet/intel/ice/ice_main.c
332
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
3324
if (ice_ptp_tx_tstamps_pending(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
333
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3363
static void ice_free_irq_msix_ll_ts(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3365
int irq_num = pf->ll_ts_irq.virq;
drivers/net/ethernet/intel/ice/ice_main.c
3368
devm_free_irq(ice_pf_to_dev(pf), irq_num, pf);
drivers/net/ethernet/intel/ice/ice_main.c
3370
ice_free_irq(pf, pf->ll_ts_irq);
drivers/net/ethernet/intel/ice/ice_main.c
3377
static void ice_free_irq_msix_misc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3379
int misc_irq_num = pf->oicr_irq.virq;
drivers/net/ethernet/intel/ice/ice_main.c
3380
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3389
devm_free_irq(ice_pf_to_dev(pf), misc_irq_num, pf);
drivers/net/ethernet/intel/ice/ice_main.c
3391
ice_free_irq(pf, pf->oicr_irq);
drivers/net/ethernet/intel/ice/ice_main.c
3392
if (pf->hw.dev_caps.ts_dev_info.ts_ll_int_read)
drivers/net/ethernet/intel/ice/ice_main.c
3393
ice_free_irq_msix_ll_ts(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3437
static int ice_req_irq_msix_misc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3439
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3440
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
3445
if (!pf->int_name[0])
drivers/net/ethernet/intel/ice/ice_main.c
3446
snprintf(pf->int_name, sizeof(pf->int_name) - 1, "%s-%s:misc",
drivers/net/ethernet/intel/ice/ice_main.c
3449
if (!pf->int_name_ll_ts[0])
drivers/net/ethernet/intel/ice/ice_main.c
3450
snprintf(pf->int_name_ll_ts, sizeof(pf->int_name_ll_ts) - 1,
drivers/net/ethernet/intel/ice/ice_main.c
3456
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
3460
irq = ice_alloc_irq(pf, false);
drivers/net/ethernet/intel/ice/ice_main.c
3464
pf->oicr_irq = irq;
drivers/net/ethernet/intel/ice/ice_main.c
3465
err = devm_request_threaded_irq(dev, pf->oicr_irq.virq, ice_misc_intr,
drivers/net/ethernet/intel/ice/ice_main.c
3467
pf->int_name, pf);
drivers/net/ethernet/intel/ice/ice_main.c
3470
pf->int_name, err);
drivers/net/ethernet/intel/ice/ice_main.c
3471
ice_free_irq(pf, pf->oicr_irq);
drivers/net/ethernet/intel/ice/ice_main.c
3476
if (!pf->hw.dev_caps.ts_dev_info.ts_ll_int_read)
drivers/net/ethernet/intel/ice/ice_main.c
3479
irq = ice_alloc_irq(pf, false);
drivers/net/ethernet/intel/ice/ice_main.c
3483
pf->ll_ts_irq = irq;
drivers/net/ethernet/intel/ice/ice_main.c
3484
err = devm_request_irq(dev, pf->ll_ts_irq.virq, ice_ll_ts_intr, 0,
drivers/net/ethernet/intel/ice/ice_main.c
3485
pf->int_name_ll_ts, pf);
drivers/net/ethernet/intel/ice/ice_main.c
3488
pf->int_name_ll_ts, err);
drivers/net/ethernet/intel/ice/ice_main.c
3489
ice_free_irq(pf, pf->ll_ts_irq);
drivers/net/ethernet/intel/ice/ice_main.c
3494
ice_ena_misc_vector(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3496
ice_ena_ctrlq_interrupts(hw, pf->oicr_irq.index);
drivers/net/ethernet/intel/ice/ice_main.c
3499
if (pf->hw.dev_caps.ts_dev_info.ts_ll_int_read)
drivers/net/ethernet/intel/ice/ice_main.c
3501
((pf->ll_ts_irq.index + pf_intr_start_offset) &
drivers/net/ethernet/intel/ice/ice_main.c
3503
wr32(hw, GLINT_ITR(ICE_RX_ITR, pf->oicr_irq.index),
drivers/net/ethernet/intel/ice/ice_main.c
3519
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_main.c
3521
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
3528
netdev->udp_tunnel_nic_info = &pf->hw.udp_tunnel_nic;
drivers/net/ethernet/intel/ice/ice_main.c
3547
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_main.c
3548
bool is_dvm_ena = ice_is_dvm_ena(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
3554
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
3634
if (ice_is_feature_supported(pf, ICE_F_GCS))
drivers/net/ethernet/intel/ice/ice_main.c
3663
ice_pf_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi)
drivers/net/ethernet/intel/ice/ice_main.c
3671
return ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_main.c
3675
ice_chnl_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi,
drivers/net/ethernet/intel/ice/ice_main.c
3685
return ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_main.c
3697
ice_ctrl_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi)
drivers/net/ethernet/intel/ice/ice_main.c
3705
return ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_main.c
3717
ice_lb_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi)
drivers/net/ethernet/intel/ice/ice_main.c
3725
return ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_main.c
3925
u16 ice_get_avail_txq_count(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3927
return ice_get_avail_q_count(pf->avail_txqs, &pf->avail_q_mutex,
drivers/net/ethernet/intel/ice/ice_main.c
3928
pf->max_pf_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
3935
u16 ice_get_avail_rxq_count(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3937
return ice_get_avail_q_count(pf->avail_rxqs, &pf->avail_q_mutex,
drivers/net/ethernet/intel/ice/ice_main.c
3938
pf->max_pf_rxqs);
drivers/net/ethernet/intel/ice/ice_main.c
3945
void ice_deinit_pf(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3949
mutex_destroy(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
3950
mutex_destroy(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
3951
mutex_destroy(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
3952
mutex_destroy(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
3953
mutex_destroy(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
3954
mutex_destroy(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
3956
if (pf->avail_txqs) {
drivers/net/ethernet/intel/ice/ice_main.c
3957
bitmap_free(pf->avail_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
3958
pf->avail_txqs = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
3961
if (pf->avail_rxqs) {
drivers/net/ethernet/intel/ice/ice_main.c
3962
bitmap_free(pf->avail_rxqs);
drivers/net/ethernet/intel/ice/ice_main.c
3963
pf->avail_rxqs = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
3966
if (pf->txtime_txqs) {
drivers/net/ethernet/intel/ice/ice_main.c
3967
bitmap_free(pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
3968
pf->txtime_txqs = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
3971
if (pf->ptp.clock)
drivers/net/ethernet/intel/ice/ice_main.c
3972
ptp_clock_unregister(pf->ptp.clock);
drivers/net/ethernet/intel/ice/ice_main.c
3974
if (!xa_empty(&pf->irq_tracker.entries))
drivers/net/ethernet/intel/ice/ice_main.c
3975
ice_free_irq_msix_misc(pf);
drivers/net/ethernet/intel/ice/ice_main.c
3977
xa_destroy(&pf->dyn_ports);
drivers/net/ethernet/intel/ice/ice_main.c
3978
xa_destroy(&pf->sf_nums);
drivers/net/ethernet/intel/ice/ice_main.c
3985
static void ice_set_pf_caps(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
3987
struct ice_hw_func_caps *func_caps = &pf->hw.func_caps;
drivers/net/ethernet/intel/ice/ice_main.c
3989
clear_bit(ICE_FLAG_RDMA_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3991
set_bit(ICE_FLAG_RDMA_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3992
clear_bit(ICE_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3994
set_bit(ICE_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3995
clear_bit(ICE_FLAG_SRIOV_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3997
set_bit(ICE_FLAG_SRIOV_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
3998
pf->vfs.num_supported = min_t(int, func_caps->num_allocd_vfs,
drivers/net/ethernet/intel/ice/ice_main.c
4001
clear_bit(ICE_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4003
set_bit(ICE_FLAG_RSS_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4005
clear_bit(ICE_FLAG_FD_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4012
pf->ctrl_vsi_idx = ICE_NO_VSI;
drivers/net/ethernet/intel/ice/ice_main.c
4013
set_bit(ICE_FLAG_FD_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4015
ice_alloc_fd_guar_item(&pf->hw, &unused,
drivers/net/ethernet/intel/ice/ice_main.c
4018
ice_alloc_fd_shrd_item(&pf->hw, &unused,
drivers/net/ethernet/intel/ice/ice_main.c
4022
clear_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4024
set_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4026
pf->max_pf_txqs = func_caps->common_cap.num_txq;
drivers/net/ethernet/intel/ice/ice_main.c
4027
pf->max_pf_rxqs = func_caps->common_cap.num_rxq;
drivers/net/ethernet/intel/ice/ice_main.c
4030
void ice_start_service_task(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4032
timer_setup(&pf->serv_tmr, ice_service_timer, 0);
drivers/net/ethernet/intel/ice/ice_main.c
4033
pf->serv_tmr_period = HZ;
drivers/net/ethernet/intel/ice/ice_main.c
4034
INIT_WORK(&pf->serv_task, ice_service_task);
drivers/net/ethernet/intel/ice/ice_main.c
4035
clear_bit(ICE_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
4043
int ice_init_pf(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4045
struct udp_tunnel_nic_info *udp_tunnel_nic = &pf->hw.udp_tunnel_nic;
drivers/net/ethernet/intel/ice/ice_main.c
4046
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4047
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
4050
mutex_init(&pf->sw_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
4051
mutex_init(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
4052
mutex_init(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
4053
mutex_init(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
4055
INIT_HLIST_HEAD(&pf->aq_wait_list);
drivers/net/ethernet/intel/ice/ice_main.c
4056
spin_lock_init(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
4057
init_waitqueue_head(&pf->aq_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
4059
init_waitqueue_head(&pf->reset_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
4061
mutex_init(&pf->avail_q_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
4063
mutex_init(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
4064
hash_init(pf->vfs.table);
drivers/net/ethernet/intel/ice/ice_main.c
4065
if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT))
drivers/net/ethernet/intel/ice/ice_main.c
4066
wr32(&pf->hw, E830_MBX_PF_IN_FLIGHT_VF_MSGS_THRESH,
drivers/net/ethernet/intel/ice/ice_main.c
4069
ice_mbx_init_snapshot(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
4071
xa_init(&pf->dyn_ports);
drivers/net/ethernet/intel/ice/ice_main.c
4072
xa_init(&pf->sf_nums);
drivers/net/ethernet/intel/ice/ice_main.c
4074
pf->avail_txqs = bitmap_zalloc(pf->max_pf_txqs, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_main.c
4075
pf->avail_rxqs = bitmap_zalloc(pf->max_pf_rxqs, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_main.c
4076
pf->txtime_txqs = bitmap_zalloc(pf->max_pf_txqs, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_main.c
4077
if (!pf->avail_txqs || !pf->avail_rxqs || !pf->txtime_txqs)
drivers/net/ethernet/intel/ice/ice_main.c
4093
err = ice_req_irq_msix_misc(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4102
ice_deinit_pf(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4139
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
4145
while (test_and_set_bit(ICE_CFG_BUSY, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
4162
dev_dbg(ice_pf_to_dev(pf), "Link is down, queue count change happens when link is brought up\n");
drivers/net/ethernet/intel/ice/ice_main.c
4178
ice_pf_dcb_recfg(pf, locked);
drivers/net/ethernet/intel/ice/ice_main.c
4183
dev_err(ice_pf_to_dev(pf), "Error during VSI rebuild: %d. Unload and reload the driver.\n",
drivers/net/ethernet/intel/ice/ice_main.c
4186
clear_bit(ICE_CFG_BUSY, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
4197
static void ice_set_safe_mode_vlan_cfg(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4199
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4211
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
4250
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
4253
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4333
ice_load_pkg(const struct firmware *firmware, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4336
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4337
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
4354
clear_bit(ICE_FLAG_ADV_FEATURES, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4361
set_bit(ICE_FLAG_ADV_FEATURES, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4372
static void ice_verify_cacheline_size(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4374
if (rd32(&pf->hw, GLPCI_CNF2) & GLPCI_CNF2_CACHELINE_SIZE_M)
drivers/net/ethernet/intel/ice/ice_main.c
4375
dev_warn(ice_pf_to_dev(pf), "%d Byte cache line assumption is invalid, driver may have Tx timeouts!\n",
drivers/net/ethernet/intel/ice/ice_main.c
4385
static int ice_send_version(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4395
return ice_aq_send_driver_ver(&pf->hw, &dv, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
4404
static int ice_init_fdir(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4406
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4413
ctrl_vsi = ice_ctrl_vsi_setup(pf, pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4425
mutex_init(&pf->hw.fdir_fltr_lock);
drivers/net/ethernet/intel/ice/ice_main.c
4427
err = ice_fdir_create_dflt_rules(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4434
ice_fdir_release_flows(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
4438
if (pf->ctrl_vsi_idx != ICE_NO_VSI) {
drivers/net/ethernet/intel/ice/ice_main.c
4439
pf->vsi[pf->ctrl_vsi_idx] = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
4440
pf->ctrl_vsi_idx = ICE_NO_VSI;
drivers/net/ethernet/intel/ice/ice_main.c
4445
static void ice_deinit_fdir(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4447
struct ice_vsi *vsi = ice_get_ctrl_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4454
if (pf->ctrl_vsi_idx != ICE_NO_VSI) {
drivers/net/ethernet/intel/ice/ice_main.c
4455
pf->vsi[pf->ctrl_vsi_idx] = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
4456
pf->ctrl_vsi_idx = ICE_NO_VSI;
drivers/net/ethernet/intel/ice/ice_main.c
4459
mutex_destroy(&(&pf->hw)->fdir_fltr_lock);
drivers/net/ethernet/intel/ice/ice_main.c
4466
static char *ice_get_opt_fw_name(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4471
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/intel/ice/ice_main.c
4499
static int ice_request_fw(struct ice_pf *pf, const struct firmware **firmware)
drivers/net/ethernet/intel/ice/ice_main.c
4501
char *opt_fw_filename = ice_get_opt_fw_name(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4502
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4533
struct ice_pf *pf = hw->back;
drivers/net/ethernet/intel/ice/ice_main.c
4537
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4578
static void ice_init_supported_rxdids(struct ice_hw *hw, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4580
pf->supported_rxdids = BIT(ICE_RXDID_LEGACY_1);
drivers/net/ethernet/intel/ice/ice_main.c
4588
pf->supported_rxdids |= BIT(i);
drivers/net/ethernet/intel/ice/ice_main.c
4602
static int ice_init_ddp_config(struct ice_hw *hw, struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4604
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4608
err = ice_request_fw(pf, &firmware);
drivers/net/ethernet/intel/ice/ice_main.c
4623
ice_load_pkg(firmware, pf);
drivers/net/ethernet/intel/ice/ice_main.c
4627
ice_init_supported_rxdids(hw, pf);
drivers/net/ethernet/intel/ice/ice_main.c
4636
static void ice_print_wake_reason(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4638
u32 wus = pf->wakeup_reason;
drivers/net/ethernet/intel/ice/ice_main.c
4656
dev_info(ice_pf_to_dev(pf), "Wake reason: %s", wake_str);
drivers/net/ethernet/intel/ice/ice_main.c
4738
void ice_init_dev_hw(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4740
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
4743
ice_init_feature_support(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4745
err = ice_init_ddp_config(hw, pf);
drivers/net/ethernet/intel/ice/ice_main.c
4751
if (err || ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
4761
int ice_init_dev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4763
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4766
ice_set_pf_caps(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4767
err = ice_init_interrupt_scheme(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4773
ice_start_service_task(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4778
void ice_deinit_dev(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4780
ice_service_task_stop(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4783
ice_reset(&pf->hw, ICE_RESET_PFR);
drivers/net/ethernet/intel/ice/ice_main.c
4784
pci_wait_for_pending_transaction(pf->pdev);
drivers/net/ethernet/intel/ice/ice_main.c
4785
ice_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4788
static void ice_init_features(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4790
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4792
if (ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_main.c
4796
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
4797
ice_ptp_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4799
if (ice_is_feature_supported(pf, ICE_F_GNSS))
drivers/net/ethernet/intel/ice/ice_main.c
4800
ice_gnss_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4802
if (ice_is_feature_supported(pf, ICE_F_CGU) ||
drivers/net/ethernet/intel/ice/ice_main.c
4803
ice_is_feature_supported(pf, ICE_F_PHY_RCLK))
drivers/net/ethernet/intel/ice/ice_main.c
4804
ice_dpll_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4807
if (ice_init_fdir(pf))
drivers/net/ethernet/intel/ice/ice_main.c
4811
if (ice_init_pf_dcb(pf, false)) {
drivers/net/ethernet/intel/ice/ice_main.c
4812
clear_bit(ICE_FLAG_DCB_CAPABLE, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4813
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4815
ice_cfg_lldp_mib_change(&pf->hw, true);
drivers/net/ethernet/intel/ice/ice_main.c
4818
if (ice_init_lag(pf))
drivers/net/ethernet/intel/ice/ice_main.c
4821
ice_hwmon_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4824
static void ice_deinit_features(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4826
if (ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_main.c
4829
ice_deinit_lag(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4830
if (test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
4831
ice_cfg_lldp_mib_change(&pf->hw, false);
drivers/net/ethernet/intel/ice/ice_main.c
4832
ice_deinit_fdir(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4833
if (ice_is_feature_supported(pf, ICE_F_GNSS))
drivers/net/ethernet/intel/ice/ice_main.c
4834
ice_gnss_exit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4835
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
4836
ice_ptp_release(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4837
if (test_bit(ICE_FLAG_DPLL, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
4838
ice_dpll_deinit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4839
if (pf->eswitch_mode == DEVLINK_ESWITCH_MODE_SWITCHDEV)
drivers/net/ethernet/intel/ice/ice_main.c
4840
xa_destroy(&pf->eswitch.reprs);
drivers/net/ethernet/intel/ice/ice_main.c
4841
ice_hwmon_exit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4844
static void ice_init_wakeup(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4847
pf->wakeup_reason = rd32(&pf->hw, PFPM_WUS);
drivers/net/ethernet/intel/ice/ice_main.c
4850
ice_print_wake_reason(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4853
wr32(&pf->hw, PFPM_WUS, U32_MAX);
drivers/net/ethernet/intel/ice/ice_main.c
4856
device_set_wakeup_enable(ice_pf_to_dev(pf), false);
drivers/net/ethernet/intel/ice/ice_main.c
4859
static int ice_init_link(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4861
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4864
err = ice_init_link_events(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4871
err = ice_init_nvm_phy_type(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4876
err = ice_update_link_info(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4880
ice_init_link_dflt_override(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4882
ice_check_link_cfg_err(pf,
drivers/net/ethernet/intel/ice/ice_main.c
4883
pf->hw.port_info->phy.link_info.link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
4886
if (pf->hw.port_info->phy.link_info.link_info &
drivers/net/ethernet/intel/ice/ice_main.c
4889
err = ice_init_phy_user_cfg(pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
489
static void ice_sync_fltr_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4893
if (!test_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
4894
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4900
set_bit(ICE_FLAG_NO_MEDIA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4906
static int ice_init_pf_sw(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4908
bool dvm = ice_is_dvm_ena(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
4913
pf->first_sw = kzalloc_obj(*pf->first_sw);
drivers/net/ethernet/intel/ice/ice_main.c
4914
if (!pf->first_sw)
drivers/net/ethernet/intel/ice/ice_main.c
4917
if (pf->hw.evb_veb)
drivers/net/ethernet/intel/ice/ice_main.c
4918
pf->first_sw->bridge_mode = BRIDGE_MODE_VEB;
drivers/net/ethernet/intel/ice/ice_main.c
4920
pf->first_sw->bridge_mode = BRIDGE_MODE_VEPA;
drivers/net/ethernet/intel/ice/ice_main.c
4922
pf->first_sw->pf = pf;
drivers/net/ethernet/intel/ice/ice_main.c
4925
pf->first_sw->sw_id = pf->hw.port_info->sw_id;
drivers/net/ethernet/intel/ice/ice_main.c
4927
err = ice_aq_set_port_params(pf->hw.port_info, dvm, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
493
if (!pf || !(test_bit(ICE_FLAG_FLTR_SYNC, pf->flags)))
drivers/net/ethernet/intel/ice/ice_main.c
4931
vsi = ice_pf_vsi_setup(pf, pf->hw.port_info);
drivers/net/ethernet/intel/ice/ice_main.c
4941
kfree(pf->first_sw);
drivers/net/ethernet/intel/ice/ice_main.c
4945
static void ice_deinit_pf_sw(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4947
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
4953
kfree(pf->first_sw);
drivers/net/ethernet/intel/ice/ice_main.c
4956
static int ice_alloc_vsis(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4958
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
496
clear_bit(ICE_FLAG_FLTR_SYNC, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
4960
pf->num_alloc_vsi = pf->hw.func_caps.guar_num_vsi;
drivers/net/ethernet/intel/ice/ice_main.c
4961
if (!pf->num_alloc_vsi)
drivers/net/ethernet/intel/ice/ice_main.c
4964
if (pf->num_alloc_vsi > UDP_TUNNEL_NIC_MAX_SHARING_DEVICES) {
drivers/net/ethernet/intel/ice/ice_main.c
4967
pf->num_alloc_vsi, UDP_TUNNEL_NIC_MAX_SHARING_DEVICES);
drivers/net/ethernet/intel/ice/ice_main.c
4968
pf->num_alloc_vsi = UDP_TUNNEL_NIC_MAX_SHARING_DEVICES;
drivers/net/ethernet/intel/ice/ice_main.c
4971
pf->vsi = devm_kcalloc(dev, pf->num_alloc_vsi, sizeof(*pf->vsi),
drivers/net/ethernet/intel/ice/ice_main.c
4973
if (!pf->vsi)
drivers/net/ethernet/intel/ice/ice_main.c
4976
pf->vsi_stats = devm_kcalloc(dev, pf->num_alloc_vsi,
drivers/net/ethernet/intel/ice/ice_main.c
4977
sizeof(*pf->vsi_stats), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_main.c
4978
if (!pf->vsi_stats) {
drivers/net/ethernet/intel/ice/ice_main.c
4979
devm_kfree(dev, pf->vsi);
drivers/net/ethernet/intel/ice/ice_main.c
498
ice_for_each_vsi(pf, v)
drivers/net/ethernet/intel/ice/ice_main.c
4986
static void ice_dealloc_vsis(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
4988
devm_kfree(ice_pf_to_dev(pf), pf->vsi_stats);
drivers/net/ethernet/intel/ice/ice_main.c
4989
pf->vsi_stats = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
499
if (pf->vsi[v] && ice_vsi_fltr_changed(pf->vsi[v]) &&
drivers/net/ethernet/intel/ice/ice_main.c
4991
pf->num_alloc_vsi = 0;
drivers/net/ethernet/intel/ice/ice_main.c
4992
devm_kfree(ice_pf_to_dev(pf), pf->vsi);
drivers/net/ethernet/intel/ice/ice_main.c
4993
pf->vsi = NULL;
drivers/net/ethernet/intel/ice/ice_main.c
4996
static int ice_init_devlink(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
500
ice_vsi_sync_fltr(pf->vsi[v])) {
drivers/net/ethernet/intel/ice/ice_main.c
5000
err = ice_devlink_register_params(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5004
ice_devlink_init_regions(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5005
ice_devlink_register(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5006
ice_health_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5011
static void ice_deinit_devlink(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5013
ice_health_deinit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5014
ice_devlink_unregister(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5015
ice_devlink_destroy_regions(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5016
ice_devlink_unregister_params(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5019
static int ice_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
502
set_bit(ICE_FLAG_FLTR_SYNC, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
5021
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5024
err = ice_init_pf(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5030
if (pf->hw.mac_type == ICE_MAC_E830) {
drivers/net/ethernet/intel/ice/ice_main.c
5031
err = pci_enable_ptm(pf->pdev, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
5036
err = ice_alloc_vsis(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5040
err = ice_init_pf_sw(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5044
ice_init_wakeup(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5046
err = ice_init_link(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5050
err = ice_send_version(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5054
ice_verify_cacheline_size(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5056
if (ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_main.c
5057
ice_set_safe_mode_vlan_cfg(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5060
pcie_print_link_status(pf->pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5063
clear_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5064
clear_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5067
mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period));
drivers/net/ethernet/intel/ice/ice_main.c
5072
ice_deinit_pf_sw(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5074
ice_dealloc_vsis(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5076
ice_deinit_pf(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5080
static void ice_deinit(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5082
set_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5083
set_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5085
ice_deinit_pf_sw(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5086
ice_dealloc_vsis(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5087
ice_deinit_pf(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5096
int ice_load(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5101
devl_assert_locked(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5103
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5115
err = ice_init_mac_fltr(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5119
err = ice_devlink_create_pf_port(pf);
drivers/net/ethernet/intel/ice/ice_main.c
512
static void ice_pf_dis_all_vsi(struct ice_pf *pf, bool locked)
drivers/net/ethernet/intel/ice/ice_main.c
5123
SET_NETDEV_DEVLINK_PORT(vsi->netdev, &pf->devlink_port);
drivers/net/ethernet/intel/ice/ice_main.c
5135
ice_init_features(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5137
err = ice_init_rdma(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5142
ice_rdma_finalize_setup(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5144
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5146
clear_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5151
ice_deinit_features(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5156
ice_devlink_destroy_pf_port(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5169
void ice_unload(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
517
ice_for_each_vsi(pf, v)
drivers/net/ethernet/intel/ice/ice_main.c
5171
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5173
devl_assert_locked(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5175
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5176
ice_deinit_rdma(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5177
ice_deinit_features(pf);
drivers/net/ethernet/intel/ice/ice_main.c
518
if (pf->vsi[v])
drivers/net/ethernet/intel/ice/ice_main.c
5180
ice_devlink_destroy_pf_port(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5184
static int ice_probe_recovery_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5186
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
519
ice_dis_vsi(pf->vsi[v], locked);
drivers/net/ethernet/intel/ice/ice_main.c
5191
INIT_HLIST_HEAD(&pf->aq_wait_list);
drivers/net/ethernet/intel/ice/ice_main.c
5192
spin_lock_init(&pf->aq_wait_lock);
drivers/net/ethernet/intel/ice/ice_main.c
5193
init_waitqueue_head(&pf->aq_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
5195
timer_setup(&pf->serv_tmr, ice_service_timer, 0);
drivers/net/ethernet/intel/ice/ice_main.c
5196
pf->serv_tmr_period = HZ;
drivers/net/ethernet/intel/ice/ice_main.c
5197
INIT_WORK(&pf->serv_task, ice_service_task_recovery_mode);
drivers/net/ethernet/intel/ice/ice_main.c
5198
clear_bit(ICE_SERVICE_SCHED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5199
err = ice_create_all_ctrlq(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
5203
scoped_guard(devl, priv_to_devlink(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
5204
err = ice_init_devlink(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5209
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
522
pf->pf_agg_node[node].num_vsis = 0;
drivers/net/ethernet/intel/ice/ice_main.c
5227
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_main.c
525
pf->vf_agg_node[node].num_vsis = 0;
drivers/net/ethernet/intel/ice/ice_main.c
5263
pf = ice_allocate_pf(dev);
drivers/net/ethernet/intel/ice/ice_main.c
5264
if (!pf)
drivers/net/ethernet/intel/ice/ice_main.c
5268
pf->aux_idx = -1;
drivers/net/ethernet/intel/ice/ice_main.c
5278
pf->pdev = pdev;
drivers/net/ethernet/intel/ice/ice_main.c
5279
pci_set_drvdata(pdev, pf);
drivers/net/ethernet/intel/ice/ice_main.c
5280
set_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5282
set_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5284
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5288
hw->back = pf;
drivers/net/ethernet/intel/ice/ice_main.c
5299
pf->msg_enable = netif_msg_init(debug, ICE_DFLT_NETIF_M);
drivers/net/ethernet/intel/ice/ice_main.c
5307
return ice_probe_recovery_mode(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5320
pf->adapter = adapter;
drivers/net/ethernet/intel/ice/ice_main.c
5322
err = ice_init_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5326
err = ice_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5330
devl_lock(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5331
err = ice_load(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5335
err = ice_init_devlink(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5338
devl_unlock(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5343
ice_unload(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5345
devl_unlock(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5346
ice_deinit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5354
ice_deinit_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
536
ice_prepare_for_reset(struct ice_pf *pf, enum ice_reset_req reset_type)
drivers/net/ethernet/intel/ice/ice_main.c
5364
static void ice_set_wake(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5366
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5367
bool wol = pf->wol_ena;
drivers/net/ethernet/intel/ice/ice_main.c
538
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5387
static void ice_setup_mc_magic_wake(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5389
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5390
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5396
if (!pf->wol_ena)
drivers/net/ethernet/intel/ice/ice_main.c
5399
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5425
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5429
if (!ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
543
dev_dbg(ice_pf_to_dev(pf), "reset_type=%d\n", reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
5434
if (ice_is_recovery_mode(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_main.c
5435
ice_service_task_stop(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5436
scoped_guard(devl, priv_to_devlink(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
5437
ice_deinit_devlink(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5442
if (test_bit(ICE_FLAG_SRIOV_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
5443
set_bit(ICE_VF_RESETS_DISABLED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5444
ice_free_vfs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5447
if (!ice_is_safe_mode(pf))
drivers/net/ethernet/intel/ice/ice_main.c
5448
ice_remove_arfs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5450
devl_lock(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5451
ice_dealloc_all_dynamic_ports(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5452
ice_deinit_devlink(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5454
ice_unload(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5455
devl_unlock(priv_to_devlink(pf));
drivers/net/ethernet/intel/ice/ice_main.c
5457
ice_deinit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5458
ice_vsi_release_all(pf);
drivers/net/ethernet/intel/ice/ice_main.c
546
if (test_bit(ICE_PREPARED_FOR_RESET, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
5460
ice_setup_mc_magic_wake(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5461
ice_set_wake(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5464
ice_deinit_hw(&pf->hw);
drivers/net/ethernet/intel/ice/ice_main.c
5466
ice_deinit_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5467
ice_aq_cancel_waiting_tasks(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5468
set_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5477
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5482
pci_wake_from_d3(pdev, pf->wol_ena);
drivers/net/ethernet/intel/ice/ice_main.c
549
synchronize_irq(pf->oicr_irq.virq);
drivers/net/ethernet/intel/ice/ice_main.c
5493
static void ice_prepare_for_shutdown(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5495
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5500
ice_vc_notify_reset(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5502
dev_dbg(ice_pf_to_dev(pf), "Tearing down internal switch for shutdown\n");
drivers/net/ethernet/intel/ice/ice_main.c
5505
ice_pf_dis_all_vsi(pf, false);
drivers/net/ethernet/intel/ice/ice_main.c
5507
ice_for_each_vsi(pf, v)
drivers/net/ethernet/intel/ice/ice_main.c
5508
if (pf->vsi[v])
drivers/net/ethernet/intel/ice/ice_main.c
5509
pf->vsi[v]->vsi_num = 0;
drivers/net/ethernet/intel/ice/ice_main.c
551
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5524
static int ice_reinit_interrupt_scheme(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
5526
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5533
ret = ice_init_interrupt_scheme(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5540
ice_for_each_vsi(pf, v) {
drivers/net/ethernet/intel/ice/ice_main.c
5541
if (!pf->vsi[v])
drivers/net/ethernet/intel/ice/ice_main.c
5544
ret = ice_vsi_alloc_q_vectors(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
5547
ice_vsi_map_rings_to_vectors(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
5549
ice_vsi_set_napi_queues(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
555
ice_vc_notify_reset(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5553
ret = ice_req_irq_msix_misc(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5564
if (pf->vsi[v]) {
drivers/net/ethernet/intel/ice/ice_main.c
5566
ice_vsi_clear_napi_queues(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
5568
ice_vsi_free_q_vectors(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
558
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
5584
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_main.c
5587
pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5589
if (!ice_pf_state_is_nominal(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
559
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_main.c
5600
disabled = ice_service_task_stop(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5602
ice_unplug_aux_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5603
ice_deinit_rdma(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5606
if (test_and_set_bit(ICE_SUSPENDED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5608
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
561
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_main.c
5612
if (test_bit(ICE_DOWN, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
5613
ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5616
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5620
ice_setup_mc_magic_wake(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5622
ice_prepare_for_shutdown(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5624
ice_set_wake(pf);
drivers/net/ethernet/intel/ice/ice_main.c
563
if (ice_is_eswitch_mode_switchdev(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
5631
ice_free_irq_msix_misc(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5632
ice_for_each_vsi(pf, v) {
drivers/net/ethernet/intel/ice/ice_main.c
5633
if (!pf->vsi[v])
drivers/net/ethernet/intel/ice/ice_main.c
5636
ice_vsi_clear_napi_queues(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
5638
ice_vsi_free_q_vectors(pf->vsi[v]);
drivers/net/ethernet/intel/ice/ice_main.c
5640
ice_clear_interrupt_scheme(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5643
pci_wake_from_d3(pdev, pf->wol_ena);
drivers/net/ethernet/intel/ice/ice_main.c
565
ice_eswitch_br_fdb_flush(pf->eswitch.br_offloads->bridge);
drivers/net/ethernet/intel/ice/ice_main.c
5656
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_main.c
5672
pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5673
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
5675
pf->wakeup_reason = rd32(hw, PFPM_WUS);
drivers/net/ethernet/intel/ice/ice_main.c
5676
ice_print_wake_reason(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5681
ret = ice_reinit_interrupt_scheme(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5685
ret = ice_init_rdma(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5690
clear_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5694
clear_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5696
if (ice_schedule_reset(pf, reset_type))
drivers/net/ethernet/intel/ice/ice_main.c
5699
clear_bit(ICE_SUSPENDED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
570
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5700
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5703
mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period));
drivers/net/ethernet/intel/ice/ice_main.c
5719
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5721
if (!pf) {
drivers/net/ethernet/intel/ice/ice_main.c
5727
if (!test_bit(ICE_SUSPENDED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5728
ice_service_task_stop(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5730
if (!test_bit(ICE_PREPARED_FOR_RESET, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5731
set_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5732
ice_prepare_for_reset(pf, ICE_RESET_PFR);
drivers/net/ethernet/intel/ice/ice_main.c
5748
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5764
reg = rd32(&pf->hw, GLGEN_RTRIG);
drivers/net/ethernet/intel/ice/ice_main.c
5783
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5785
if (!pf) {
drivers/net/ethernet/intel/ice/ice_main.c
579
if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
5791
if (test_bit(ICE_SUSPENDED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5797
ice_restore_all_vfs_msi_state(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5799
ice_do_reset(pf, ICE_RESET_PFR);
drivers/net/ethernet/intel/ice/ice_main.c
5800
ice_service_task_restart(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5801
mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period));
drivers/net/ethernet/intel/ice/ice_main.c
5810
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_main.c
5812
if (!test_bit(ICE_SUSPENDED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5813
ice_service_task_stop(pf);
drivers/net/ethernet/intel/ice/ice_main.c
5815
if (!test_bit(ICE_PREPARED_FOR_RESET, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
5816
set_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
5817
ice_prepare_for_reset(pf, ICE_RESET_PFR);
drivers/net/ethernet/intel/ice/ice_main.c
595
clear_bit(ICE_FLAG_TC_MQPRIO, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
5999
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
6000
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
6012
if (test_bit(ICE_DOWN, pf->state) ||
drivers/net/ethernet/intel/ice/ice_main.c
6013
ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
6019
if (ice_chnl_dmac_fltr_cnt(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
607
set_bit(ICE_VSI_REBUILD_PENDING, ice_get_main_vsi(pf)->state);
drivers/net/ethernet/intel/ice/ice_main.c
608
ice_pf_dis_all_vsi(pf, false);
drivers/net/ethernet/intel/ice/ice_main.c
610
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
611
ice_ptp_prepare_for_reset(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
613
if (ice_is_feature_supported(pf, ICE_F_GNSS))
drivers/net/ethernet/intel/ice/ice_main.c
614
ice_gnss_exit(pf);
drivers/net/ethernet/intel/ice/ice_main.c
621
set_bit(ICE_PREPARED_FOR_RESET, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
629
static void ice_do_reset(struct ice_pf *pf, enum ice_reset_req reset_type)
drivers/net/ethernet/intel/ice/ice_main.c
631
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
632
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
636
if (pf->lag && pf->lag->bonded && reset_type == ICE_RESET_PFR) {
drivers/net/ethernet/intel/ice/ice_main.c
641
ice_prepare_for_reset(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
646
set_bit(ICE_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
647
clear_bit(ICE_RESET_OICR_RECV, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
648
clear_bit(ICE_PREPARED_FOR_RESET, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
649
clear_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
650
clear_bit(ICE_CORER_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
6502
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
6506
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
6507
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_main.c
651
clear_bit(ICE_GLOBR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
6513
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
6514
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_main.c
652
wake_up(&pf->reset_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
6554
if (!(features & NETIF_F_HW_TC) && ice_is_adq_active(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
6555
dev_err(ice_pf_to_dev(pf), "ADQ is active, can't turn hw_tc_offload off\n");
drivers/net/ethernet/intel/ice/ice_main.c
6562
assign_bit(ICE_FLAG_CLS_FLOWER, pf->flags, ena);
drivers/net/ethernet/intel/ice/ice_main.c
6571
if (ice_is_feature_supported(pf, ICE_F_GCS) &&
drivers/net/ethernet/intel/ice/ice_main.c
6574
dev_err(ice_pf_to_dev(pf), "To enable TSO, you must first disable HW checksum.\n");
drivers/net/ethernet/intel/ice/ice_main.c
6576
dev_err(ice_pf_to_dev(pf), "To enable HW checksum, you must first disable TSO.\n");
drivers/net/ethernet/intel/ice/ice_main.c
661
pf->pfr_count++;
drivers/net/ethernet/intel/ice/ice_main.c
662
ice_rebuild(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
663
clear_bit(ICE_PREPARED_FOR_RESET, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
664
clear_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
665
wake_up(&pf->reset_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
666
ice_reset_all_vfs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
674
static void ice_reset_subtask(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
6779
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
6803
ice_ptp_link_change(pf, true);
drivers/net/ethernet/intel/ice/ice_main.c
6812
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
688
if (test_bit(ICE_RESET_OICR_RECV, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
69
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_main.c
690
if (test_and_clear_bit(ICE_CORER_RECV, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
692
if (test_and_clear_bit(ICE_GLOBR_RECV, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
694
if (test_and_clear_bit(ICE_EMPR_RECV, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
6970
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
699
ice_prepare_for_reset(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
7004
if (likely(pf->stat_prev_loaded)) {
drivers/net/ethernet/intel/ice/ice_main.c
702
if (ice_check_reset(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_main.c
7025
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
7028
test_bit(ICE_CFG_BUSY, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
703
set_bit(ICE_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7044
cur_ns->rx_crc_errors = pf->stats.crc_errors;
drivers/net/ethernet/intel/ice/ice_main.c
7045
cur_ns->rx_errors = pf->stats.crc_errors +
drivers/net/ethernet/intel/ice/ice_main.c
7046
pf->stats.illegal_bytes +
drivers/net/ethernet/intel/ice/ice_main.c
7047
pf->stats.rx_undersize +
drivers/net/ethernet/intel/ice/ice_main.c
7048
pf->stats.rx_jabber +
drivers/net/ethernet/intel/ice/ice_main.c
7049
pf->stats.rx_fragments +
drivers/net/ethernet/intel/ice/ice_main.c
7050
pf->stats.rx_oversize;
drivers/net/ethernet/intel/ice/ice_main.c
7052
cur_ns->rx_missed_errors = pf->stats.eth.rx_discards;
drivers/net/ethernet/intel/ice/ice_main.c
706
pf->hw.reset_ongoing = false;
drivers/net/ethernet/intel/ice/ice_main.c
7060
void ice_update_pf_stats(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
7063
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
7068
prev_ps = &pf->stats_prev;
drivers/net/ethernet/intel/ice/ice_main.c
7069
cur_ps = &pf->stats;
drivers/net/ethernet/intel/ice/ice_main.c
707
ice_rebuild(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
7071
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
7072
pf->stat_prev_loaded = false;
drivers/net/ethernet/intel/ice/ice_main.c
7074
ice_stat_update40(hw, GLPRT_GORCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7078
ice_stat_update40(hw, GLPRT_UPRCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7082
ice_stat_update40(hw, GLPRT_MPRCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7086
ice_stat_update40(hw, GLPRT_BPRCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7090
ice_stat_update32(hw, PRTRPB_RDPC, pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7094
ice_stat_update40(hw, GLPRT_GOTCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7098
ice_stat_update40(hw, GLPRT_UPTCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
71
return &pf->pdev->dev;
drivers/net/ethernet/intel/ice/ice_main.c
7102
ice_stat_update40(hw, GLPRT_MPTCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7106
ice_stat_update40(hw, GLPRT_BPTCL(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
711
clear_bit(ICE_RESET_OICR_RECV, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7110
ice_stat_update32(hw, GLPRT_TDOLD(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7114
ice_stat_update40(hw, GLPRT_PRC64L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7117
ice_stat_update40(hw, GLPRT_PRC127L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
712
clear_bit(ICE_PREPARED_FOR_RESET, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7120
ice_stat_update40(hw, GLPRT_PRC255L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7123
ice_stat_update40(hw, GLPRT_PRC511L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7126
ice_stat_update40(hw, GLPRT_PRC1023L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7129
ice_stat_update40(hw, GLPRT_PRC1522L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
713
clear_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7132
ice_stat_update40(hw, GLPRT_PRC9522L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7135
ice_stat_update40(hw, GLPRT_PTC64L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7138
ice_stat_update40(hw, GLPRT_PTC127L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
714
clear_bit(ICE_CORER_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7141
ice_stat_update40(hw, GLPRT_PTC255L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7144
ice_stat_update40(hw, GLPRT_PTC511L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7147
ice_stat_update40(hw, GLPRT_PTC1023L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
715
clear_bit(ICE_GLOBR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7150
ice_stat_update40(hw, GLPRT_PTC1522L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7153
ice_stat_update40(hw, GLPRT_PTC9522L(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
716
wake_up(&pf->reset_wait_queue);
drivers/net/ethernet/intel/ice/ice_main.c
7160
pf->stat_prev_loaded, &prev_ps->fd_sb_match,
drivers/net/ethernet/intel/ice/ice_main.c
7162
ice_stat_update32(hw, GLPRT_LXONRXC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7165
ice_stat_update32(hw, GLPRT_LXOFFRXC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7168
ice_stat_update32(hw, GLPRT_LXONTXC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
717
ice_reset_all_vfs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7171
ice_stat_update32(hw, GLPRT_LXOFFTXC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7174
ice_update_dcb_stats(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7176
ice_stat_update32(hw, GLPRT_CRCERRS(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7179
ice_stat_update32(hw, GLPRT_ILLERRC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7182
ice_stat_update32(hw, GLPRT_MLFC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7186
ice_stat_update32(hw, GLPRT_MRFC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7190
ice_stat_update32(hw, GLPRT_RLEC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7193
ice_stat_update32(hw, GLPRT_RUC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7196
ice_stat_update32(hw, GLPRT_RFC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7199
ice_stat_update32(hw, GLPRT_ROC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7202
ice_stat_update32(hw, GLPRT_RJC(port), pf->stat_prev_loaded,
drivers/net/ethernet/intel/ice/ice_main.c
7205
cur_ps->fd_sb_status = test_bit(ICE_FLAG_FD_ENA, pf->flags) ? 1 : 0;
drivers/net/ethernet/intel/ice/ice_main.c
7207
pf->stat_prev_loaded = true;
drivers/net/ethernet/intel/ice/ice_main.c
724
if (test_bit(ICE_PFR_REQ, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
726
if (pf->lag && pf->lag->bonded) {
drivers/net/ethernet/intel/ice/ice_main.c
727
dev_dbg(ice_pf_to_dev(pf), "PFR on a bonded interface, promoting to CORER\n");
drivers/net/ethernet/intel/ice/ice_main.c
7279
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
7280
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
731
if (test_bit(ICE_CORER_REQ, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
733
if (test_bit(ICE_GLOBR_REQ, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
740
if (!test_bit(ICE_DOWN, pf->state) &&
drivers/net/ethernet/intel/ice/ice_main.c
741
!test_bit(ICE_CFG_BUSY, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
742
ice_do_reset(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
7475
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
7479
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7531
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
7548
dev_driver_string(ice_pf_to_dev(pf)), vsi->netdev->name);
drivers/net/ethernet/intel/ice/ice_main.c
7553
if (bitmap_empty(pf->txtime_txqs, pf->max_pf_txqs))
drivers/net/ethernet/intel/ice/ice_main.c
7591
static void ice_vsi_release_all(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
7595
if (!pf->vsi)
drivers/net/ethernet/intel/ice/ice_main.c
7598
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_main.c
7599
if (!pf->vsi[i])
drivers/net/ethernet/intel/ice/ice_main.c
7602
if (pf->vsi[i]->type == ICE_VSI_CHNL)
drivers/net/ethernet/intel/ice/ice_main.c
7605
err = ice_vsi_release(pf->vsi[i]);
drivers/net/ethernet/intel/ice/ice_main.c
7607
dev_dbg(ice_pf_to_dev(pf), "Failed to release pf->vsi[%d], err %d, vsi_num = %d\n",
drivers/net/ethernet/intel/ice/ice_main.c
7608
i, err, pf->vsi[i]->vsi_num);
drivers/net/ethernet/intel/ice/ice_main.c
7619
static int ice_vsi_rebuild_by_type(struct ice_pf *pf, enum ice_vsi_type type)
drivers/net/ethernet/intel/ice/ice_main.c
7621
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7624
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_main.c
7625
struct ice_vsi *vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_main.c
7639
err = ice_replay_vsi(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_main.c
7649
vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_main.c
7670
static void ice_update_pf_netdev_link(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
7675
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_main.c
7676
struct ice_vsi *vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_main.c
7681
ice_get_link_status(pf->vsi[i]->port_info, &link_up);
drivers/net/ethernet/intel/ice/ice_main.c
7683
netif_carrier_on(pf->vsi[i]->netdev);
drivers/net/ethernet/intel/ice/ice_main.c
7684
netif_tx_wake_all_queues(pf->vsi[i]->netdev);
drivers/net/ethernet/intel/ice/ice_main.c
7686
netif_carrier_off(pf->vsi[i]->netdev);
drivers/net/ethernet/intel/ice/ice_main.c
7687
netif_tx_stop_all_queues(pf->vsi[i]->netdev);
drivers/net/ethernet/intel/ice/ice_main.c
7702
static void ice_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type)
drivers/net/ethernet/intel/ice/ice_main.c
7704
struct ice_vsi *vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7705
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7706
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
7710
if (test_bit(ICE_DOWN, pf->state))
drivers/net/ethernet/intel/ice/ice_main.c
7721
pf->fw_emp_reset_disabled = false;
drivers/net/ethernet/intel/ice/ice_main.c
7733
if (!ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
7739
ice_load_pkg(NULL, pf);
drivers/net/ethernet/intel/ice/ice_main.c
7770
err = ice_aq_set_port_params(pf->hw.port_info, dvm, NULL);
drivers/net/ethernet/intel/ice/ice_main.c
7779
err = ice_req_irq_msix_misc(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7785
if (test_bit(ICE_FLAG_FD_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
7800
if (test_bit(ICE_FLAG_DCB_ENA, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
7801
ice_dcb_rebuild(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7807
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
7808
ice_ptp_rebuild(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
7810
if (ice_is_feature_supported(pf, ICE_F_GNSS))
drivers/net/ethernet/intel/ice/ice_main.c
7811
ice_gnss_init(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7814
err = ice_vsi_rebuild_by_type(pf, ICE_VSI_PF);
drivers/net/ethernet/intel/ice/ice_main.c
7821
err = ice_rebuild_channels(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7830
if (test_bit(ICE_FLAG_FD_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
7831
err = ice_vsi_rebuild_by_type(pf, ICE_VSI_CTRL);
drivers/net/ethernet/intel/ice/ice_main.c
7842
ice_fdir_replay_fltrs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7844
ice_rebuild_arfs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7850
ice_update_pf_netdev_link(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7853
err = ice_send_version(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7863
clear_bit(ICE_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7865
ice_health_clear(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7867
ice_rdma_finalize_setup(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7868
if (ice_is_feature_supported(pf, ICE_F_SRIOV_LAG))
drivers/net/ethernet/intel/ice/ice_main.c
7869
ice_lag_rebuild(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7872
ice_ptp_restore_timestamp_mode(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7875
ice_ptp_queue_work(pf);
drivers/net/ethernet/intel/ice/ice_main.c
7883
set_bit(ICE_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
7886
set_bit(ICE_NEEDS_RESTART, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
79
static void ice_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type);
drivers/net/ethernet/intel/ice/ice_main.c
7901
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
7924
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
7944
set_bit(ICE_FLAG_MTU_CHANGED, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
81
static void ice_vsi_release_all(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_main.c
8151
struct ice_pf *pf = ice_netdev_to_pf(dev);
drivers/net/ethernet/intel/ice/ice_main.c
8154
bmode = pf->first_sw->bridge_mode;
drivers/net/ethernet/intel/ice/ice_main.c
8221
struct ice_pf *pf = ice_netdev_to_pf(dev);
drivers/net/ethernet/intel/ice/ice_main.c
8223
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
8227
pf_sw = pf->first_sw;
drivers/net/ethernet/intel/ice/ice_main.c
8244
ice_for_each_vsi(pf, v) {
drivers/net/ethernet/intel/ice/ice_main.c
8245
if (!pf->vsi[v])
drivers/net/ethernet/intel/ice/ice_main.c
8247
err = ice_vsi_update_bridge_mode(pf->vsi[v], mode);
drivers/net/ethernet/intel/ice/ice_main.c
8282
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
8285
pf->tx_timeout_count++;
drivers/net/ethernet/intel/ice/ice_main.c
8291
if (ice_is_pfc_causing_hung_q(pf, txqueue)) {
drivers/net/ethernet/intel/ice/ice_main.c
8292
dev_info(ice_pf_to_dev(pf), "Fake Tx hang detected on queue %u, timeout caused by PFC storm\n",
drivers/net/ethernet/intel/ice/ice_main.c
83
static int ice_rebuild_channels(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_main.c
8308
if (time_after(jiffies, (pf->tx_timeout_last_recovery + HZ * 20)))
drivers/net/ethernet/intel/ice/ice_main.c
8309
pf->tx_timeout_recovery_level = 1;
drivers/net/ethernet/intel/ice/ice_main.c
8310
else if (time_before(jiffies, (pf->tx_timeout_last_recovery +
drivers/net/ethernet/intel/ice/ice_main.c
8315
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
8327
ice_prep_tx_hang_report(pf, tx_ring, vsi->vsi_num, head, intr);
drivers/net/ethernet/intel/ice/ice_main.c
8330
pf->tx_timeout_last_recovery = jiffies;
drivers/net/ethernet/intel/ice/ice_main.c
8332
pf->tx_timeout_recovery_level, txqueue);
drivers/net/ethernet/intel/ice/ice_main.c
8334
switch (pf->tx_timeout_recovery_level) {
drivers/net/ethernet/intel/ice/ice_main.c
8336
set_bit(ICE_PFR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
8339
set_bit(ICE_CORER_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
8342
set_bit(ICE_GLOBR_REQ, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
8346
set_bit(ICE_DOWN, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
8348
set_bit(ICE_SERVICE_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
8352
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8353
pf->tx_timeout_recovery_level++;
drivers/net/ethernet/intel/ice/ice_main.c
8447
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
8462
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8578
static int ice_add_vsi_to_fdir(struct ice_pf *pf, struct ice_vsi *vsi)
drivers/net/ethernet/intel/ice/ice_main.c
8580
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8588
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
8641
static int ice_add_channel(struct ice_pf *pf, u16 sw_id, struct ice_channel *ch)
drivers/net/ethernet/intel/ice/ice_main.c
8643
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8651
vsi = ice_chnl_vsi_setup(pf, pf->hw.port_info, ch);
drivers/net/ethernet/intel/ice/ice_main.c
8657
ice_add_vsi_to_fdir(pf, vsi);
drivers/net/ethernet/intel/ice/ice_main.c
8759
ice_setup_hw_channel(struct ice_pf *pf, struct ice_vsi *vsi,
drivers/net/ethernet/intel/ice/ice_main.c
8762
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8768
ret = ice_add_channel(pf, sw_id, ch);
drivers/net/ethernet/intel/ice/ice_main.c
8797
ice_setup_channel(struct ice_pf *pf, struct ice_vsi *vsi,
drivers/net/ethernet/intel/ice/ice_main.c
8800
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8809
sw_id = pf->first_sw->sw_id;
drivers/net/ethernet/intel/ice/ice_main.c
8812
ret = ice_setup_hw_channel(pf, vsi, ch, sw_id, ICE_VSI_CHNL);
drivers/net/ethernet/intel/ice/ice_main.c
8850
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
8856
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8868
if (!ice_setup_channel(pf, vsi, ch)) {
drivers/net/ethernet/intel/ice/ice_main.c
8898
static void ice_rem_all_chnl_fltrs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
8905
&pf->tc_flower_fltr_list,
drivers/net/ethernet/intel/ice/ice_main.c
8917
status = ice_rem_adv_rule_by_id(&pf->hw, &rule);
drivers/net/ethernet/intel/ice/ice_main.c
8920
dev_dbg(ice_pf_to_dev(pf), "TC flower filter (rule_id %u) does not exist\n",
drivers/net/ethernet/intel/ice/ice_main.c
8923
dev_err(ice_pf_to_dev(pf), "failed to delete TC flower filter, status %d\n",
drivers/net/ethernet/intel/ice/ice_main.c
8933
pf->num_dmac_chnl_fltrs--;
drivers/net/ethernet/intel/ice/ice_main.c
8952
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
8957
ice_rem_all_chnl_fltrs(pf);
drivers/net/ethernet/intel/ice/ice_main.c
8961
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
8999
ice_fdir_rem_adq_chnl(&pf->hw, ch->ch_vsi->idx);
drivers/net/ethernet/intel/ice/ice_main.c
9026
static int ice_rebuild_channels(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
9028
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9036
main_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9040
if (!test_bit(ICE_FLAG_TC_MQPRIO, pf->flags) ||
drivers/net/ethernet/intel/ice/ice_main.c
9055
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_main.c
9058
vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_main.c
9075
vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_main.c
9078
err = ice_replay_vsi(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_main.c
9142
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9172
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_main.c
9179
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_main.c
9200
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9207
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9212
clear_bit(ICE_FLAG_TC_MQPRIO, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
9225
if (pf->hw.port_info->is_custom_tx_enabled) {
drivers/net/ethernet/intel/ice/ice_main.c
9229
ice_tear_down_devlink_rate_tree(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9238
set_bit(ICE_FLAG_TC_MQPRIO, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
9244
set_bit(ICE_FLAG_CLS_FLOWER, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
9260
if (!hw && !test_bit(ICE_FLAG_TC_MQPRIO, pf->flags))
drivers/net/ethernet/intel/ice/ice_main.c
9263
if (!hw && !test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
9264
vsi->req_txq = min_t(int, ice_get_avail_txq_count(pf),
drivers/net/ethernet/intel/ice/ice_main.c
9266
vsi->req_rxq = min_t(int, ice_get_avail_rxq_count(pf),
drivers/net/ethernet/intel/ice/ice_main.c
9303
clear_bit(ICE_RESET_FAILED, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
9319
if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_main.c
9376
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_main.c
9383
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9384
while (test_and_set_bit(ICE_CFG_BUSY, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
9392
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9402
clear_bit(ICE_CFG_BUSY, pf->state);
drivers/net/ethernet/intel/ice/ice_main.c
9417
struct ice_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9423
if (!ice_is_feature_supported(pf, ICE_F_TXTIME))
drivers/net/ethernet/intel/ice/ice_main.c
9431
set_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
9433
clear_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
9451
clear_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
9464
struct ice_pf *pf = np->vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
947
static void ice_set_dflt_mib(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_main.c
949
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_main.c
9490
if (ice_is_eswitch_mode_switchdev(pf)) {
drivers/net/ethernet/intel/ice/ice_main.c
9495
cdev = pf->cdev_info;
drivers/net/ethernet/intel/ice/ice_main.c
9497
mutex_lock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
9508
mutex_lock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
9510
mutex_unlock(&pf->tc_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
9515
mutex_unlock(&pf->adev_mutex);
drivers/net/ethernet/intel/ice/ice_main.c
953
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_main.c
9655
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_main.c
9657
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
9678
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9682
if (test_bit(ICE_NEEDS_RESTART, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
9696
ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err);
drivers/net/ethernet/intel/ice/ice_main.c
9700
clear_bit(ICE_FLAG_NO_MEDIA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
9701
if (!test_bit(ICE_PHY_INIT_COMPLETE, pf->state)) {
drivers/net/ethernet/intel/ice/ice_main.c
9717
set_bit(ICE_FLAG_NO_MEDIA, pf->flags);
drivers/net/ethernet/intel/ice/ice_main.c
9743
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_main.c
9745
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
1007
ice_ptp_mark_tx_tracker_stale(&pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
101
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1017
static int ice_ptp_write_init(struct ice_pf *pf, struct timespec64 *ts)
drivers/net/ethernet/intel/ice/ice_ptp.c
1020
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1033
static int ice_ptp_write_adj(struct ice_pf *pf, s32 adj)
drivers/net/ethernet/intel/ice/ice_ptp.c
1035
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1049
static u64 ice_base_incval(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
105
switch (pf->ptp.tx_interrupt_mode) {
drivers/net/ethernet/intel/ice/ice_ptp.c
1051
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1056
dev_dbg(ice_pf_to_dev(pf), "PTP: using base increment value of 0x%016llx\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1070
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
1075
pf = ptp_port_to_pf(port);
drivers/net/ethernet/intel/ice/ice_ptp.c
1076
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1090
dev_err(ice_pf_to_dev(pf), "PTP failed to check port %d Tx FIFO, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1107
dev_dbg(ice_pf_to_dev(pf), "Try %d, port %d FIFO not empty\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1111
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_ptp.c
1140
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
1146
pf = ptp_port_to_pf(port);
drivers/net/ethernet/intel/ice/ice_ptp.c
1147
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1149
if (ice_is_reset_in_progress(pf->state)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
1151
kthread_queue_delayed_work(pf->ptp.kworker,
drivers/net/ethernet/intel/ice/ice_ptp.c
1163
kthread_queue_delayed_work(pf->ptp.kworker,
drivers/net/ethernet/intel/ice/ice_ptp.c
1177
struct ice_pf *pf = ptp_port_to_pf(ptp_port);
drivers/net/ethernet/intel/ice/ice_ptp.c
1179
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
118
enable = pf->ptp.tstamp_config.tx_type == HWTSTAMP_TX_ON;
drivers/net/ethernet/intel/ice/ice_ptp.c
1201
dev_err(ice_pf_to_dev(pf), "PTP failed to set PHY port %d down, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1220
struct ice_pf *pf = ptp_port_to_pf(ptp_port);
drivers/net/ethernet/intel/ice/ice_ptp.c
1222
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1258
kthread_queue_delayed_work(pf->ptp.kworker, &ptp_port->ov_work,
drivers/net/ethernet/intel/ice/ice_ptp.c
1269
dev_err(ice_pf_to_dev(pf), "PTP failed to set PHY port %d up, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1282
void ice_ptp_link_change(struct ice_pf *pf, bool linkup)
drivers/net/ethernet/intel/ice/ice_ptp.c
1285
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1287
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
1290
ptp_port = &pf->ptp.port;
drivers/net/ethernet/intel/ice/ice_ptp.c
1296
if (pf->hw.reset_ongoing)
drivers/net/ethernet/intel/ice/ice_ptp.c
1300
test_bit(ICE_FLAG_DPLL, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
1303
mutex_lock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
1316
dev_err_once(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_ptp.c
1324
dev_err_once(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_ptp.c
1330
mutex_unlock(&pf->dplls.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
1346
dev_warn(ice_pf_to_dev(pf), "%s: Unknown PHY type\n", __func__);
drivers/net/ethernet/intel/ice/ice_ptp.c
136
static void ice_set_rx_tstamp(struct ice_pf *pf, bool on)
drivers/net/ethernet/intel/ice/ice_ptp.c
1363
static int ice_ptp_cfg_phy_interrupt(struct ice_pf *pf, bool ena, u32 threshold)
drivers/net/ethernet/intel/ice/ice_ptp.c
1365
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1366
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
141
vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1417
static void ice_ptp_reset_phy_timestamping(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1419
ice_ptp_port_phy_restart(&pf->ptp.port);
drivers/net/ethernet/intel/ice/ice_ptp.c
1426
static void ice_ptp_restart_all_phy(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1430
list_for_each(entry, &pf->adapter->ports.ports) {
drivers/net/ethernet/intel/ice/ice_ptp.c
1450
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1451
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1455
incval = adjust_by_scaled_ppm(ice_base_incval(pf), scaled_ppm);
drivers/net/ethernet/intel/ice/ice_ptp.c
1458
dev_err(ice_pf_to_dev(pf), "PTP failed to set incval, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
1470
void ice_ptp_extts_event(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1473
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1478
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
1491
if (!(pf->ptp.ext_ts_irq & (1 << chan)))
drivers/net/ethernet/intel/ice/ice_ptp.c
1499
pin_desc_idx = ice_ptp_find_pin_idx(pf, PTP_PF_EXTTS, chan);
drivers/net/ethernet/intel/ice/ice_ptp.c
1503
desc = &pf->ptp.ice_pin_desc[pin_desc_idx];
drivers/net/ethernet/intel/ice/ice_ptp.c
1509
pf->ptp.ext_ts_irq &= ~(1 << chan);
drivers/net/ethernet/intel/ice/ice_ptp.c
1510
ptp_clock_event(pf->ptp.clock, &event);
drivers/net/ethernet/intel/ice/ice_ptp.c
1524
static int ice_ptp_cfg_extts(struct ice_pf *pf, struct ptp_extts_request *rq,
drivers/net/ethernet/intel/ice/ice_ptp.c
1528
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1536
pin_desc_idx = ice_ptp_find_pin_idx(pf, PTP_PF_EXTTS, chan);
drivers/net/ethernet/intel/ice/ice_ptp.c
1540
gpio_pin = pf->ptp.ice_pin_desc[pin_desc_idx].gpio[0];
drivers/net/ethernet/intel/ice/ice_ptp.c
1570
for (unsigned int i = 0; i < pf->ptp.info.n_ext_ts; i++)
drivers/net/ethernet/intel/ice/ice_ptp.c
1571
if ((pf->ptp.extts_rqs[i].flags &
drivers/net/ethernet/intel/ice/ice_ptp.c
1592
static void ice_ptp_disable_all_extts(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1594
for (unsigned int i = 0; i < pf->ptp.info.n_ext_ts ; i++)
drivers/net/ethernet/intel/ice/ice_ptp.c
1595
if (pf->ptp.extts_rqs[i].flags & PTP_ENABLE_FEATURE)
drivers/net/ethernet/intel/ice/ice_ptp.c
1596
ice_ptp_cfg_extts(pf, &pf->ptp.extts_rqs[i],
drivers/net/ethernet/intel/ice/ice_ptp.c
1599
synchronize_irq(pf->oicr_irq.virq);
drivers/net/ethernet/intel/ice/ice_ptp.c
1608
static void ice_ptp_enable_all_extts(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
161
static void ice_ptp_disable_timestamp_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1610
for (unsigned int i = 0; i < pf->ptp.info.n_ext_ts ; i++)
drivers/net/ethernet/intel/ice/ice_ptp.c
1611
if (pf->ptp.extts_rqs[i].flags & PTP_ENABLE_FEATURE)
drivers/net/ethernet/intel/ice/ice_ptp.c
1612
ice_ptp_cfg_extts(pf, &pf->ptp.extts_rqs[i],
drivers/net/ethernet/intel/ice/ice_ptp.c
163
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1695
static int ice_ptp_cfg_perout(struct ice_pf *pf, struct ptp_perout_request *rq,
drivers/net/ethernet/intel/ice/ice_ptp.c
170
ice_set_rx_tstamp(pf, false);
drivers/net/ethernet/intel/ice/ice_ptp.c
1700
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1703
pin_desc_idx = ice_ptp_find_pin_idx(pf, PTP_PF_PEROUT, rq->index);
drivers/net/ethernet/intel/ice/ice_ptp.c
1707
gpio_pin = pf->ptp.ice_pin_desc[pin_desc_idx].gpio[1];
drivers/net/ethernet/intel/ice/ice_ptp.c
1708
prop_delay_ns = pf->ptp.ice_pin_desc[pin_desc_idx].delay[1];
drivers/net/ethernet/intel/ice/ice_ptp.c
1717
if (strncmp(pf->ptp.pin_desc[pin_desc_idx].name, "1PPS", 64) == 0 &&
drivers/net/ethernet/intel/ice/ice_ptp.c
1719
dev_err(ice_pf_to_dev(pf), "1PPS pin supports only 1 s period\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
1724
dev_err(ice_pf_to_dev(pf), "CLK Period must be an even value\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
1740
clk = ice_ptp_read_src_clk_reg(pf, NULL) + NSEC_PER_MSEC * 500;
drivers/net/ethernet/intel/ice/ice_ptp.c
1758
static void ice_ptp_disable_all_perout(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1760
for (unsigned int i = 0; i < pf->ptp.info.n_per_out; i++)
drivers/net/ethernet/intel/ice/ice_ptp.c
1761
if (pf->ptp.perout_rqs[i].period.sec ||
drivers/net/ethernet/intel/ice/ice_ptp.c
1762
pf->ptp.perout_rqs[i].period.nsec)
drivers/net/ethernet/intel/ice/ice_ptp.c
1763
ice_ptp_cfg_perout(pf, &pf->ptp.perout_rqs[i],
drivers/net/ethernet/intel/ice/ice_ptp.c
1775
static void ice_ptp_enable_all_perout(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
1777
for (unsigned int i = 0; i < pf->ptp.info.n_per_out; i++)
drivers/net/ethernet/intel/ice/ice_ptp.c
1778
if (pf->ptp.perout_rqs[i].period.sec ||
drivers/net/ethernet/intel/ice/ice_ptp.c
1779
pf->ptp.perout_rqs[i].period.nsec)
drivers/net/ethernet/intel/ice/ice_ptp.c
1780
ice_ptp_cfg_perout(pf, &pf->ptp.perout_rqs[i],
drivers/net/ethernet/intel/ice/ice_ptp.c
1796
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1799
pin_desc = &pf->ptp.ice_pin_desc[pin];
drivers/net/ethernet/intel/ice/ice_ptp.c
180
void ice_ptp_restore_timestamp_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
182
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1832
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1839
&pf->ptp.perout_rqs[rq->perout.index];
drivers/net/ethernet/intel/ice/ice_ptp.c
1841
err = ice_ptp_cfg_perout(pf, &rq->perout, on);
drivers/net/ethernet/intel/ice/ice_ptp.c
185
ice_ptp_cfg_tx_interrupt(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1853
&pf->ptp.extts_rqs[rq->extts.index];
drivers/net/ethernet/intel/ice/ice_ptp.c
1855
err = ice_ptp_cfg_extts(pf, &rq->extts, on);
drivers/net/ethernet/intel/ice/ice_ptp.c
187
enable_rx = pf->ptp.tstamp_config.rx_filter == HWTSTAMP_FILTER_ALL;
drivers/net/ethernet/intel/ice/ice_ptp.c
188
ice_set_rx_tstamp(pf, enable_rx);
drivers/net/ethernet/intel/ice/ice_ptp.c
1881
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1884
time_ns = ice_ptp_read_src_clk_reg(pf, sts);
drivers/net/ethernet/intel/ice/ice_ptp.c
1900
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1902
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1911
dev_warn(ice_pf_to_dev(pf), "Failed to mark timestamps as invalid before settime\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
1920
ice_ptp_disable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1922
err = ice_ptp_write_init(pf, &ts64);
drivers/net/ethernet/intel/ice/ice_ptp.c
1926
ice_ptp_reset_cached_phctime(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1929
ice_ptp_enable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1933
ice_ptp_restart_all_phy(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1936
dev_err(ice_pf_to_dev(pf), "PTP failed to set time %d\n", err);
drivers/net/ethernet/intel/ice/ice_ptp.c
1969
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
1970
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
1974
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1991
ice_ptp_disable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
1993
err = ice_ptp_write_adj(pf, delta);
drivers/net/ethernet/intel/ice/ice_ptp.c
1996
ice_ptp_enable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2005
ice_ptp_reset_cached_phctime(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
203
u64 ice_ptp_read_src_clk_reg(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.c
206
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2073
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
2095
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
2101
pf = ctx->pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
2102
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
211
hw = ice_get_primary_hw(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2113
dev_err(ice_pf_to_dev(pf), "PTP failed to get cross timestamp lock\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
214
guard(spinlock)(&pf->adapter->ptp_gltsyn_time_lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
2177
struct ice_pf *pf = ptp_info_to_pf(info);
drivers/net/ethernet/intel/ice/ice_ptp.c
2179
.pf = pf,
drivers/net/ethernet/intel/ice/ice_ptp.c
2182
switch (pf->hw.mac_type) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2210
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ptp.c
2212
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
2215
*config = pf->ptp.tstamp_config;
drivers/net/ethernet/intel/ice/ice_ptp.c
2225
static int ice_ptp_set_timestamp_mode(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.c
2230
pf->ptp.tstamp_config.tx_type = HWTSTAMP_TX_OFF;
drivers/net/ethernet/intel/ice/ice_ptp.c
2233
pf->ptp.tstamp_config.tx_type = HWTSTAMP_TX_ON;
drivers/net/ethernet/intel/ice/ice_ptp.c
2241
pf->ptp.tstamp_config.rx_filter = HWTSTAMP_FILTER_NONE;
drivers/net/ethernet/intel/ice/ice_ptp.c
2257
pf->ptp.tstamp_config.rx_filter = HWTSTAMP_FILTER_ALL;
drivers/net/ethernet/intel/ice/ice_ptp.c
2264
ice_ptp_restore_timestamp_mode(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2281
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_ptp.c
2284
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
2287
err = ice_ptp_set_timestamp_mode(pf, config);
drivers/net/ethernet/intel/ice/ice_ptp.c
2292
*config = pf->ptp.tstamp_config;
drivers/net/ethernet/intel/ice/ice_ptp.c
2335
static void ice_ptp_setup_pin_cfg(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2337
for (unsigned int i = 0; i < pf->ptp.info.n_pins; i++) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2338
const struct ice_ptp_pin_desc *desc = &pf->ptp.ice_pin_desc[i];
drivers/net/ethernet/intel/ice/ice_ptp.c
2339
struct ptp_pin_desc *pin = &pf->ptp.pin_desc[i];
drivers/net/ethernet/intel/ice/ice_ptp.c
2342
if (!ice_is_feature_supported(pf, ICE_F_SMA_CTRL))
drivers/net/ethernet/intel/ice/ice_ptp.c
2352
pf->ptp.info.pin_config = pf->ptp.pin_desc;
drivers/net/ethernet/intel/ice/ice_ptp.c
2362
static void ice_ptp_disable_pins(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2364
struct ptp_clock_info *info = &pf->ptp.info;
drivers/net/ethernet/intel/ice/ice_ptp.c
2366
dev_warn(ice_pf_to_dev(pf), "Failed to configure PTP pin control\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
2384
static int ice_ptp_parse_sdp_entries(struct ice_pf *pf, __le16 *entries,
drivers/net/ethernet/intel/ice/ice_ptp.c
2428
dev_dbg(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_ptp.c
2433
pf->ptp.info.n_pins = n_pins;
drivers/net/ethernet/intel/ice/ice_ptp.c
2446
static void ice_ptp_set_funcs_e82x(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2448
pf->ptp.info.getcrosststamp = ice_ptp_getcrosststamp;
drivers/net/ethernet/intel/ice/ice_ptp.c
2450
if (pf->hw.mac_type == ICE_MAC_GENERIC_3K_E825) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2451
pf->ptp.ice_pin_desc = ice_pin_desc_e825c;
drivers/net/ethernet/intel/ice/ice_ptp.c
2452
pf->ptp.info.n_pins = ARRAY_SIZE(ice_pin_desc_e825c);
drivers/net/ethernet/intel/ice/ice_ptp.c
2454
pf->ptp.ice_pin_desc = ice_pin_desc_e82x;
drivers/net/ethernet/intel/ice/ice_ptp.c
2455
pf->ptp.info.n_pins = ARRAY_SIZE(ice_pin_desc_e82x);
drivers/net/ethernet/intel/ice/ice_ptp.c
2457
ice_ptp_setup_pin_cfg(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2469
static void ice_ptp_set_funcs_e810(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2473
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
2477
err = ice_ptp_read_sdp_ac(&pf->hw, entries, &num_entries);
drivers/net/ethernet/intel/ice/ice_ptp.c
2480
if (ice_is_feature_supported(pf, ICE_F_SMA_CTRL)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2484
pf->ptp.ice_pin_desc = ice_pin_desc_e810;
drivers/net/ethernet/intel/ice/ice_ptp.c
2485
pf->ptp.info.n_pins = ARRAY_SIZE(ice_pin_desc_e810);
drivers/net/ethernet/intel/ice/ice_ptp.c
2489
desc = devm_kcalloc(ice_pf_to_dev(pf), ICE_N_PINS_MAX,
drivers/net/ethernet/intel/ice/ice_ptp.c
2495
err = ice_ptp_parse_sdp_entries(pf, entries, num_entries, desc);
drivers/net/ethernet/intel/ice/ice_ptp.c
2503
ice_ptp_setup_pin_cfg(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2507
devm_kfree(ice_pf_to_dev(pf), desc);
drivers/net/ethernet/intel/ice/ice_ptp.c
2508
ice_ptp_disable_pins(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2521
static void ice_ptp_set_funcs_e830(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2524
if (pcie_ptm_enabled(pf->pdev) && boot_cpu_has(X86_FEATURE_ART))
drivers/net/ethernet/intel/ice/ice_ptp.c
2525
pf->ptp.info.getcrosststamp = ice_ptp_getcrosststamp;
drivers/net/ethernet/intel/ice/ice_ptp.c
2529
pf->ptp.ice_pin_desc = ice_pin_desc_e810;
drivers/net/ethernet/intel/ice/ice_ptp.c
2530
pf->ptp.info.n_pins = ARRAY_SIZE(ice_pin_desc_e810);
drivers/net/ethernet/intel/ice/ice_ptp.c
2531
ice_ptp_setup_pin_cfg(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2538
static void ice_ptp_set_caps(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2540
struct ptp_clock_info *info = &pf->ptp.info;
drivers/net/ethernet/intel/ice/ice_ptp.c
2541
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2561
switch (pf->hw.mac_type) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2563
ice_ptp_set_funcs_e810(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2566
ice_ptp_set_funcs_e830(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2570
ice_ptp_set_funcs_e82x(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2586
static long ice_ptp_create_clock(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2592
if (pf->ptp.clock)
drivers/net/ethernet/intel/ice/ice_ptp.c
2595
ice_ptp_set_caps(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2597
info = &pf->ptp.info;
drivers/net/ethernet/intel/ice/ice_ptp.c
2598
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2601
pf->ptp.clock = ptp_clock_register(info, dev);
drivers/net/ethernet/intel/ice/ice_ptp.c
2602
if (IS_ERR(pf->ptp.clock)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2603
dev_err(ice_pf_to_dev(pf), "Failed to register PTP clock device");
drivers/net/ethernet/intel/ice/ice_ptp.c
2604
return PTR_ERR(pf->ptp.clock);
drivers/net/ethernet/intel/ice/ice_ptp.c
2658
void ice_ptp_process_ts(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2660
switch (pf->ptp.tx_interrupt_mode) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2666
ice_ptp_process_tx_tstamp(&pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
2670
ice_ptp_tx_tstamp_owner(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2674
pf->ptp.tx_interrupt_mode);
drivers/net/ethernet/intel/ice/ice_ptp.c
2693
static bool ice_any_port_has_timestamps(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2697
scoped_guard(mutex, &pf->adapter->ports.lock) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2698
list_for_each_entry(port, &pf->adapter->ports.ports,
drivers/net/ethernet/intel/ice/ice_ptp.c
2710
bool ice_ptp_tx_tstamps_pending(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2712
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2716
switch (pf->ptp.tx_interrupt_mode) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2720
if (ice_port_has_timestamps(&pf->ptp.port.tx))
drivers/net/ethernet/intel/ice/ice_ptp.c
2724
if (ice_any_port_has_timestamps(pf))
drivers/net/ethernet/intel/ice/ice_ptp.c
2729
pf->ptp.tx_interrupt_mode);
drivers/net/ethernet/intel/ice/ice_ptp.c
2738
err = ice_get_phy_tx_tstamp_ready(&pf->hw, i, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
2753
irqreturn_t ice_ptp_ts_irq(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2755
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2764
struct ice_ptp_tx *tx = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_ptp.c
2767
if (!ice_pf_state_is_nominal(pf))
drivers/net/ethernet/intel/ice/ice_ptp.c
2788
if (!ice_ptp_pf_handles_tx_interrupt(pf))
drivers/net/ethernet/intel/ice/ice_ptp.c
2791
set_bit(ICE_MISC_THREAD_TX_TSTAMP, pf->misc_thread);
drivers/net/ethernet/intel/ice/ice_ptp.c
2795
ice_ptp_process_ts(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2797
if (ice_ptp_tx_tstamps_pending(pf)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2823
static void ice_ptp_maybe_trigger_tx_interrupt(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2825
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2826
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2830
if (!pf->ptp.port.tx.has_ready_bitmap)
drivers/net/ethernet/intel/ice/ice_ptp.c
2833
if (!ice_pf_src_tmr_owned(pf))
drivers/net/ethernet/intel/ice/ice_ptp.c
2840
err = ice_get_phy_tx_tstamp_ready(&pf->hw, i, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
2861
struct ice_pf *pf = container_of(ptp, struct ice_pf, ptp);
drivers/net/ethernet/intel/ice/ice_ptp.c
2864
if (pf->ptp.state != ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
2867
err = ice_ptp_update_cached_phctime(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2869
ice_ptp_maybe_trigger_tx_interrupt(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2883
void ice_ptp_queue_work(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2885
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_ptp.c
2886
pf->ptp.state == ICE_PTP_READY)
drivers/net/ethernet/intel/ice/ice_ptp.c
2887
kthread_queue_delayed_work(pf->ptp.kworker, &pf->ptp.work, 0);
drivers/net/ethernet/intel/ice/ice_ptp.c
2896
static void ice_ptp_prepare_rebuild_sec(struct ice_pf *pf, bool rebuild,
drivers/net/ethernet/intel/ice/ice_ptp.c
2901
list_for_each(entry, &pf->adapter->ports.ports) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2926
void ice_ptp_prepare_for_reset(struct ice_pf *pf, enum ice_reset_req reset_type)
drivers/net/ethernet/intel/ice/ice_ptp.c
2928
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
2929
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2938
ice_ptp_disable_timestamp_mode(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2945
if (ice_pf_src_tmr_owned(pf) && hw->mac_type == ICE_MAC_GENERIC_3K_E825)
drivers/net/ethernet/intel/ice/ice_ptp.c
2946
ice_ptp_prepare_rebuild_sec(pf, false, reset_type);
drivers/net/ethernet/intel/ice/ice_ptp.c
2948
ice_ptp_release_tx_tracker(pf, &pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
2951
ice_ptp_disable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
2953
src_tmr = ice_get_ptp_src_clock_index(&pf->hw);
drivers/net/ethernet/intel/ice/ice_ptp.c
2956
wr32(&pf->hw, GLTSYN_ENA(src_tmr), (u32)~GLTSYN_ENA_TSYN_ENA_M);
drivers/net/ethernet/intel/ice/ice_ptp.c
2969
static int ice_ptp_rebuild_owner(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
2971
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
2972
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
2992
err = ice_ptp_write_incval(hw, ice_base_incval(pf));
drivers/net/ethernet/intel/ice/ice_ptp.c
3006
err = ice_ptp_write_init(pf, &ts);
drivers/net/ethernet/intel/ice/ice_ptp.c
3016
ice_ptp_flush_all_tx_tracker(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3019
err = ice_ptp_cfg_phy_interrupt(pf, true, 1);
drivers/net/ethernet/intel/ice/ice_ptp.c
3023
ice_ptp_restart_all_phy(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3026
ice_ptp_enable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3027
ice_ptp_enable_all_extts(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3041
void ice_ptp_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type)
drivers/net/ethernet/intel/ice/ice_ptp.c
3043
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
3047
ice_ptp_prepare_for_reset(pf, reset_type);
drivers/net/ethernet/intel/ice/ice_ptp.c
3050
dev_err(ice_pf_to_dev(pf), "PTP was not initialized\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
3054
if (ice_pf_src_tmr_owned(pf) && reset_type != ICE_RESET_PFR) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3055
err = ice_ptp_rebuild_owner(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3062
dev_info(ice_pf_to_dev(pf), "PTP reset successful\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
3067
dev_err(ice_pf_to_dev(pf), "PTP reset failed %d\n", err);
drivers/net/ethernet/intel/ice/ice_ptp.c
3070
static int ice_ptp_setup_adapter(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3072
if (!ice_pf_src_tmr_owned(pf) || !ice_is_primary(&pf->hw))
drivers/net/ethernet/intel/ice/ice_ptp.c
3075
pf->adapter->ctrl_pf = pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
3080
static int ice_ptp_setup_pf(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3082
struct ice_ptp *ctrl_ptp = ice_get_ctrl_ptp(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3083
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
3086
dev_info(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_ptp.c
3091
if (pf->hw.mac_type == ICE_MAC_UNKNOWN)
drivers/net/ethernet/intel/ice/ice_ptp.c
3095
mutex_lock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3098
&pf->adapter->ports.ports);
drivers/net/ethernet/intel/ice/ice_ptp.c
3099
mutex_unlock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3104
static void ice_ptp_cleanup_pf(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3106
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
3108
if (pf->hw.mac_type != ICE_MAC_UNKNOWN) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3109
mutex_lock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3111
mutex_unlock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3122
int ice_ptp_clock_index(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3124
struct ice_ptp *ctrl_ptp = ice_get_ctrl_ptp(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3142
static int ice_ptp_init_owner(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3144
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
3150
dev_err(ice_pf_to_dev(pf), "Failed to initialize PHC, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
3157
dev_err(ice_pf_to_dev(pf), "Failed to initialize CGU, status %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
3169
err = ice_ptp_write_incval(hw, ice_base_incval(pf));
drivers/net/ethernet/intel/ice/ice_ptp.c
3175
err = ice_ptp_write_init(pf, &ts);
drivers/net/ethernet/intel/ice/ice_ptp.c
3183
err = ice_ptp_cfg_phy_interrupt(pf, true, 1);
drivers/net/ethernet/intel/ice/ice_ptp.c
3188
err = ice_ptp_create_clock(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3194
pf->ptp.clock = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
3208
static int ice_ptp_init_work(struct ice_pf *pf, struct ice_ptp *ptp)
drivers/net/ethernet/intel/ice/ice_ptp.c
3219
dev_name(ice_pf_to_dev(pf)));
drivers/net/ethernet/intel/ice/ice_ptp.c
3238
static int ice_ptp_init_port(struct ice_pf *pf, struct ice_ptp_port *ptp_port)
drivers/net/ethernet/intel/ice/ice_ptp.c
3240
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
3248
return ice_ptp_init_tx(pf, &ptp_port->tx, ptp_port->port_num);
drivers/net/ethernet/intel/ice/ice_ptp.c
3252
return ice_ptp_init_tx_e82x(pf, &ptp_port->tx,
drivers/net/ethernet/intel/ice/ice_ptp.c
3268
static void ice_ptp_init_tx_interrupt_mode(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3270
switch (pf->hw.mac_type) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3276
if (ice_pf_src_tmr_owned(pf))
drivers/net/ethernet/intel/ice/ice_ptp.c
3277
pf->ptp.tx_interrupt_mode = ICE_PTP_TX_INTERRUPT_ALL;
drivers/net/ethernet/intel/ice/ice_ptp.c
3279
pf->ptp.tx_interrupt_mode = ICE_PTP_TX_INTERRUPT_NONE;
drivers/net/ethernet/intel/ice/ice_ptp.c
3283
pf->ptp.tx_interrupt_mode = ICE_PTP_TX_INTERRUPT_SELF;
drivers/net/ethernet/intel/ice/ice_ptp.c
3299
void ice_ptp_init(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3301
struct ice_ptp *ptp = &pf->ptp;
drivers/net/ethernet/intel/ice/ice_ptp.c
3302
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
3315
ice_ptp_init_tx_interrupt_mode(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3320
if (ice_pf_src_tmr_owned(pf) && ice_is_primary(hw)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3321
err = ice_ptp_setup_adapter(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3324
err = ice_ptp_init_owner(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3329
err = ice_ptp_setup_pf(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3333
err = ice_ptp_init_port(pf, &ptp->port);
drivers/net/ethernet/intel/ice/ice_ptp.c
3338
ice_ptp_reset_phy_timestamping(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3341
ice_ptp_cfg_tx_interrupt(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3345
err = ice_ptp_init_work(pf, ptp);
drivers/net/ethernet/intel/ice/ice_ptp.c
3349
dev_info(ice_pf_to_dev(pf), "PTP init successful\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
3354
ice_ptp_cleanup_pf(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3357
if (pf->ptp.clock) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3359
pf->ptp.clock = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
3365
dev_err(ice_pf_to_dev(pf), "PTP failed %d\n", err);
drivers/net/ethernet/intel/ice/ice_ptp.c
3375
void ice_ptp_release(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
3377
if (pf->ptp.state == ICE_PTP_UNINIT)
drivers/net/ethernet/intel/ice/ice_ptp.c
3380
if (pf->ptp.state != ICE_PTP_READY) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3381
mutex_destroy(&pf->ptp.port.ps_lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3382
ice_ptp_cleanup_pf(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3383
if (pf->ptp.clock) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3384
ptp_clock_unregister(pf->ptp.clock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3385
pf->ptp.clock = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
3390
pf->ptp.state = ICE_PTP_UNINIT;
drivers/net/ethernet/intel/ice/ice_ptp.c
3393
ice_ptp_disable_timestamp_mode(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3395
ice_ptp_cleanup_pf(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3397
ice_ptp_release_tx_tracker(pf, &pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
3399
ice_ptp_disable_all_extts(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3401
kthread_cancel_delayed_work_sync(&pf->ptp.work);
drivers/net/ethernet/intel/ice/ice_ptp.c
3403
ice_ptp_port_phy_stop(&pf->ptp.port);
drivers/net/ethernet/intel/ice/ice_ptp.c
3404
mutex_destroy(&pf->ptp.port.ps_lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3405
if (pf->ptp.kworker) {
drivers/net/ethernet/intel/ice/ice_ptp.c
3406
kthread_destroy_worker(pf->ptp.kworker);
drivers/net/ethernet/intel/ice/ice_ptp.c
3407
pf->ptp.kworker = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
341
static u64 ice_ptp_extend_40b_ts(struct ice_pf *pf, u64 in_tstamp)
drivers/net/ethernet/intel/ice/ice_ptp.c
3410
if (!pf->ptp.clock)
drivers/net/ethernet/intel/ice/ice_ptp.c
3414
ice_ptp_disable_all_perout(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
3416
ptp_clock_unregister(pf->ptp.clock);
drivers/net/ethernet/intel/ice/ice_ptp.c
3417
pf->ptp.clock = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
3419
dev_info(ice_pf_to_dev(pf), "Removed PTP clock\n");
drivers/net/ethernet/intel/ice/ice_ptp.c
347
discard_time = pf->ptp.cached_phc_jiffies + msecs_to_jiffies(2000);
drivers/net/ethernet/intel/ice/ice_ptp.c
349
pf->ptp.tx_hwtstamp_discarded++;
drivers/net/ethernet/intel/ice/ice_ptp.c
353
return ice_ptp_extend_32b_ts(pf->ptp.cached_phc_time,
drivers/net/ethernet/intel/ice/ice_ptp.c
385
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
391
pf = ptp_port_to_pf(ptp_port);
drivers/net/ethernet/intel/ice/ice_ptp.c
392
params = &pf->hw.ptp.phy.e810;
drivers/net/ethernet/intel/ice/ice_ptp.c
397
pf->ptp.tx_hwtstamp_timeouts++;
drivers/net/ethernet/intel/ice/ice_ptp.c
414
wr32(&pf->hw, REG_LL_PROXY_H,
drivers/net/ethernet/intel/ice/ice_ptp.c
437
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
444
pf = ptp_port_to_pf(ptp_port);
drivers/net/ethernet/intel/ice/ice_ptp.c
445
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
446
params = &pf->hw.ptp.phy.e810;
drivers/net/ethernet/intel/ice/ice_ptp.c
457
raw_tstamp = rd32(&pf->hw, REG_LL_PROXY_L);
drivers/net/ethernet/intel/ice/ice_ptp.c
459
reg_ll_high = rd32(&pf->hw, REG_LL_PROXY_H);
drivers/net/ethernet/intel/ice/ice_ptp.c
470
dev_err(ice_pf_to_dev(pf), "Failed to get the Tx tstamp - FW not ready");
drivers/net/ethernet/intel/ice/ice_ptp.c
499
tstamp = ice_ptp_extend_40b_ts(pf, raw_tstamp);
drivers/net/ethernet/intel/ice/ice_ptp.c
505
pf->ptp.tx_hwtstamp_good++;
drivers/net/ethernet/intel/ice/ice_ptp.c
565
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
57
static struct ice_pf *ice_get_ctrl_pf(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
573
pf = ptp_port_to_pf(ptp_port);
drivers/net/ethernet/intel/ice/ice_ptp.c
574
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
59
return !pf->adapter ? NULL : pf->adapter->ctrl_pf;
drivers/net/ethernet/intel/ice/ice_ptp.c
601
pf->ptp.tx_hwtstamp_timeouts++;
drivers/net/ethernet/intel/ice/ice_ptp.c
62
static struct ice_ptp *ice_get_ctrl_ptp(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
64
struct ice_pf *ctrl_pf = ice_get_ctrl_pf(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
664
tstamp = ice_ptp_extend_40b_ts(pf, raw_tstamp);
drivers/net/ethernet/intel/ice/ice_ptp.c
677
pf->ptp.tx_hwtstamp_good += tstamp_good;
drivers/net/ethernet/intel/ice/ice_ptp.c
680
static void ice_ptp_tx_tstamp_owner(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
684
mutex_lock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
685
list_for_each_entry(port, &pf->adapter->ports.ports, list_node) {
drivers/net/ethernet/intel/ice/ice_ptp.c
693
mutex_unlock(&pf->adapter->ports.lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
740
ice_ptp_flush_tx_tracker(struct ice_pf *pf, struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
742
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_ptp.c
750
dev_dbg(ice_pf_to_dev(pf), "Failed to get the Tx tstamp ready bitmap for block %u, err %d\n",
drivers/net/ethernet/intel/ice/ice_ptp.c
77
static int ice_ptp_find_pin_idx(struct ice_pf *pf, enum ptp_pin_function func,
drivers/net/ethernet/intel/ice/ice_ptp.c
775
pf->ptp.tx_hwtstamp_flushed++;
drivers/net/ethernet/intel/ice/ice_ptp.c
80
const struct ptp_clock_info *info = &pf->ptp.info;
drivers/net/ethernet/intel/ice/ice_ptp.c
811
ice_ptp_flush_all_tx_tracker(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
815
list_for_each_entry(port, &pf->adapter->ports.ports, list_node)
drivers/net/ethernet/intel/ice/ice_ptp.c
827
ice_ptp_release_tx_tracker(struct ice_pf *pf, struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
836
synchronize_irq(pf->oicr_irq.virq);
drivers/net/ethernet/intel/ice/ice_ptp.c
838
ice_ptp_flush_tx_tracker(pf, tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
865
static int ice_ptp_init_tx_e82x(struct ice_pf *pf, struct ice_ptp_tx *tx,
drivers/net/ethernet/intel/ice/ice_ptp.c
887
static int ice_ptp_init_tx(struct ice_pf *pf, struct ice_ptp_tx *tx, u8 port)
drivers/net/ethernet/intel/ice/ice_ptp.c
897
tx->has_ready_bitmap = pf->hw.mac_type != ICE_MAC_E810;
drivers/net/ethernet/intel/ice/ice_ptp.c
919
static int ice_ptp_update_cached_phctime(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
921
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
926
update_before = pf->ptp.cached_phc_jiffies + msecs_to_jiffies(2000);
drivers/net/ethernet/intel/ice/ice_ptp.c
927
if (pf->ptp.cached_phc_time &&
drivers/net/ethernet/intel/ice/ice_ptp.c
929
unsigned long time_taken = jiffies - pf->ptp.cached_phc_jiffies;
drivers/net/ethernet/intel/ice/ice_ptp.c
933
pf->ptp.late_cached_phc_updates++;
drivers/net/ethernet/intel/ice/ice_ptp.c
937
systime = ice_ptp_read_src_clk_reg(pf, NULL);
drivers/net/ethernet/intel/ice/ice_ptp.c
940
WRITE_ONCE(pf->ptp.cached_phc_time, systime);
drivers/net/ethernet/intel/ice/ice_ptp.c
941
WRITE_ONCE(pf->ptp.cached_phc_jiffies, jiffies);
drivers/net/ethernet/intel/ice/ice_ptp.c
943
if (test_and_set_bit(ICE_CFG_BUSY, pf->state))
drivers/net/ethernet/intel/ice/ice_ptp.c
946
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_ptp.c
947
struct ice_vsi *vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_ptp.c
962
clear_bit(ICE_CFG_BUSY, pf->state);
drivers/net/ethernet/intel/ice/ice_ptp.c
980
static void ice_ptp_reset_cached_phctime(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
982
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
988
err = ice_ptp_update_cached_phctime(pf);
drivers/net/ethernet/intel/ice/ice_ptp.c
99
static void ice_ptp_cfg_tx_interrupt(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.c
998
kthread_queue_delayed_work(pf->ptp.kworker, &pf->ptp.work,
drivers/net/ethernet/intel/ice/ice_ptp.h
294
int ice_ptp_clock_index(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
301
void ice_ptp_restore_timestamp_mode(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
303
void ice_ptp_extts_event(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
307
void ice_ptp_process_ts(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
308
irqreturn_t ice_ptp_ts_irq(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
309
bool ice_ptp_tx_tstamps_pending(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
310
u64 ice_ptp_read_src_clk_reg(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.h
315
void ice_ptp_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type);
drivers/net/ethernet/intel/ice/ice_ptp.h
316
void ice_ptp_prepare_for_reset(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.h
318
void ice_ptp_init(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
319
void ice_ptp_release(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
320
void ice_ptp_link_change(struct ice_pf *pf, bool linkup);
drivers/net/ethernet/intel/ice/ice_ptp.h
321
void ice_ptp_queue_work(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_ptp.h
337
static inline void ice_ptp_restore_timestamp_mode(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_ptp.h
338
static inline void ice_ptp_extts_event(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_ptp.h
350
static inline void ice_ptp_process_ts(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_ptp.h
352
static inline irqreturn_t ice_ptp_ts_irq(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.h
357
static inline bool ice_ptp_tx_tstamps_pending(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.h
362
static inline u64 ice_ptp_read_src_clk_reg(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.h
375
static inline void ice_ptp_rebuild(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.h
380
static inline void ice_ptp_prepare_for_reset(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_ptp.h
384
static inline void ice_ptp_init(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_ptp.h
385
static inline void ice_ptp_release(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_ptp.h
386
static inline void ice_ptp_link_change(struct ice_pf *pf, bool linkup)
drivers/net/ethernet/intel/ice/ice_ptp.h
390
static inline void ice_ptp_queue_work(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp.h
394
static inline int ice_ptp_clock_index(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
1893
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
1917
zo = rd32(ice_get_primary_hw(pf), GLTSYN_SHTIME_0(tmr_idx));
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
1918
lo = rd32(ice_get_primary_hw(pf), GLTSYN_SHTIME_L(tmr_idx));
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
2056
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
2086
lo = rd32(ice_get_primary_hw(pf), GLTSYN_INCVAL_L(tmr_idx));
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
2087
hi = rd32(ice_get_primary_hw(pf), GLTSYN_INCVAL_H(tmr_idx));
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
335
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
339
hw = ice_get_primary_hw(pf);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
354
struct ice_pf *pf = container_of(hw, struct ice_pf, hw);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
357
hw = ice_get_primary_hw(pf);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
359
guard(spinlock)(&pf->adapter->ptp_gltsyn_time_lock);
drivers/net/ethernet/intel/ice/ice_repr.c
353
static void ice_repr_set_tx_topology(struct ice_pf *pf, struct devlink *devlink)
drivers/net/ethernet/intel/ice/ice_repr.c
356
if (ice_is_adq_active(pf) || ice_is_dcb_active(pf) ||
drivers/net/ethernet/intel/ice/ice_repr.c
357
!ice_is_switchdev_running(pf))
drivers/net/ethernet/intel/ice/ice_repr.c
360
ice_devlink_rate_init_tx_topology(devlink, ice_get_main_vsi(pf));
drivers/net/ethernet/intel/ice/ice_repr.c
433
devlink = priv_to_devlink(vf->pf);
drivers/net/ethernet/intel/ice/ice_repr.c
434
ice_repr_set_tx_topology(vf->pf, devlink);
drivers/net/ethernet/intel/ice/ice_repr.c
527
struct ice_repr *ice_repr_get(struct ice_pf *pf, u32 id)
drivers/net/ethernet/intel/ice/ice_repr.c
529
return xa_load(&pf->eswitch.reprs, id);
drivers/net/ethernet/intel/ice/ice_repr.h
57
struct ice_repr *ice_repr_get(struct ice_pf *pf, u32 id);
drivers/net/ethernet/intel/ice/ice_sf_eth.c
100
struct ice_pf *pf = dyn_port->pf;
drivers/net/ethernet/intel/ice/ice_sf_eth.c
107
vsi->port_info = pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_sf_eth.c
110
priv = ice_allocate_sf(&adev->dev, pf);
drivers/net/ethernet/intel/ice/ice_sf_eth.c
263
struct ice_pf *pf = dyn_port->pf;
drivers/net/ethernet/intel/ice/ice_sf_eth.c
282
pdev = pf->pdev;
drivers/net/ethernet/intel/ice/ice_sriov.c
1004
dev_info(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_sriov.c
1011
vf->first_vector_idx = ice_virt_get_irqs(pf, vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
102
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1041
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1042
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1045
err = ice_check_sriov_allowed(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1051
ice_free_vfs(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1059
err = ice_pci_sriov_ena(pf, num_vfs);
drivers/net/ethernet/intel/ice/ice_sriov.c
1073
void ice_process_vflr_event(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
1075
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
1080
if (!test_and_clear_bit(ICE_VFLR_EVENT_PENDING, pf->state) ||
drivers/net/ethernet/intel/ice/ice_sriov.c
1081
!ice_has_vfs(pf))
drivers/net/ethernet/intel/ice/ice_sriov.c
1084
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
1085
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
1096
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
1111
static struct ice_vf *ice_get_vf_from_pfq(struct ice_pf *pf, u16 pfq)
drivers/net/ethernet/intel/ice/ice_sriov.c
1117
ice_for_each_vf_rcu(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
1147
static u32 ice_globalq_to_pfq(struct ice_pf *pf, u32 globalq)
drivers/net/ethernet/intel/ice/ice_sriov.c
1149
return globalq - pf->hw.func_caps.common_cap.rxq_first_id;
drivers/net/ethernet/intel/ice/ice_sriov.c
1162
ice_vf_lan_overflow_event(struct ice_pf *pf, struct ice_rq_event_info *event)
drivers/net/ethernet/intel/ice/ice_sriov.c
1170
dev_dbg(ice_pf_to_dev(pf), "GLDCB_RTCTQ: 0x%08x\n", gldcb_rtctq);
drivers/net/ethernet/intel/ice/ice_sriov.c
1175
vf = ice_get_vf_from_pfq(pf, ice_globalq_to_pfq(pf, queue));
drivers/net/ethernet/intel/ice/ice_sriov.c
1193
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1199
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1201
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1253
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1257
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1299
int __ice_set_vf_mac(struct ice_pf *pf, u16 vf_id, const u8 *mac)
drivers/net/ethernet/intel/ice/ice_sriov.c
1305
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
131
void ice_free_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
1311
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
133
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
134
struct ice_vfs *vfs = &pf->vfs;
drivers/net/ethernet/intel/ice/ice_sriov.c
135
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
1377
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1381
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1385
if (ice_is_eswitch_mode_switchdev(pf)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
1386
dev_info(ice_pf_to_dev(pf), "Trusted VF is forbidden in switchdev mode\n");
drivers/net/ethernet/intel/ice/ice_sriov.c
139
if (!ice_has_vfs(pf))
drivers/net/ethernet/intel/ice/ice_sriov.c
1407
dev_info(ice_pf_to_dev(pf), "VF %u is now %strusted\n",
drivers/net/ethernet/intel/ice/ice_sriov.c
142
while (test_and_set_bit(ICE_VF_DIS, pf->state))
drivers/net/ethernet/intel/ice/ice_sriov.c
1427
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1431
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1467
static int ice_calc_all_vfs_min_tx_rate(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
1474
ice_for_each_vf_rcu(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_sriov.c
149
if (!pci_vfs_assigned(pf->pdev))
drivers/net/ethernet/intel/ice/ice_sriov.c
150
pci_disable_sriov(pf->pdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1504
all_vfs_min_tx_rate = ice_calc_all_vfs_min_tx_rate(vf->pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1510
dev_err(ice_pf_to_dev(vf->pf), "min_tx_rate of %d Mbps on VF %u would cause oversubscription of %d Mbps based on the current link speed %d Mbps\n",
drivers/net/ethernet/intel/ice/ice_sriov.c
1531
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1537
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1539
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1553
if (min_tx_rate && ice_is_dcb_active(pf)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
156
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
159
ice_eswitch_detach_vf(pf, vf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1600
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1606
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
161
ice_virt_free_irqs(pf, vf->first_vector_idx, vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
1682
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
1688
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1696
if (!ice_is_supported_port_vlan_proto(&pf->hw, local_vlan_proto)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
170
if (!pci_vfs_assigned(pf->pdev)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
1702
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
1743
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
1746
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1749
vf->mdd_rx_events.count, pf->hw.pf_id, vf->vf_id,
drivers/net/ethernet/intel/ice/ice_sriov.c
1751
test_bit(ICE_FLAG_MDD_AUTO_RESET_VF, pf->flags)
drivers/net/ethernet/intel/ice/ice_sriov.c
1761
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
1764
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1767
vf->mdd_tx_events.count, pf->hw.pf_id, vf->vf_id,
drivers/net/ethernet/intel/ice/ice_sriov.c
1769
test_bit(ICE_FLAG_MDD_AUTO_RESET_VF, pf->flags)
drivers/net/ethernet/intel/ice/ice_sriov.c
1779
void ice_print_vfs_mdd_events(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
1785
if (!test_and_clear_bit(ICE_MDD_VF_PRINT_PENDING, pf->state))
drivers/net/ethernet/intel/ice/ice_sriov.c
1789
if (time_is_after_jiffies(pf->vfs.last_printed_mdd_jiffies + HZ * 1))
drivers/net/ethernet/intel/ice/ice_sriov.c
1792
pf->vfs.last_printed_mdd_jiffies = jiffies;
drivers/net/ethernet/intel/ice/ice_sriov.c
1794
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
1795
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
1810
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
182
ice_free_vf_entries(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
1820
void ice_restore_all_vfs_msi_state(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
1825
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_sriov.c
186
clear_bit(ICE_VF_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_sriov.c
187
clear_bit(ICE_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_sriov.c
200
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
208
vsi = ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_sriov.c
211
dev_err(ice_pf_to_dev(pf), "Failed to create VF VSI\n");
drivers/net/ethernet/intel/ice/ice_sriov.c
234
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
239
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
24
static void ice_free_vf_entries(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
244
pf->hw.func_caps.common_cap.msix_vector_first_id;
drivers/net/ethernet/intel/ice/ice_sriov.c
26
struct ice_vfs *vfs = &pf->vfs;
drivers/net/ethernet/intel/ice/ice_sriov.c
278
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
280
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
371
static int ice_set_per_vf_res(struct ice_pf *pf, u16 num_vfs)
drivers/net/ethernet/intel/ice/ice_sriov.c
375
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
377
lockdep_assert_held(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
383
msix_avail_for_sriov = pf->virt_irq_tracker.num_entries;
drivers/net/ethernet/intel/ice/ice_sriov.c
402
avail_qs = ice_get_avail_txq_count(pf) / num_vfs;
drivers/net/ethernet/intel/ice/ice_sriov.c
410
avail_qs = ice_get_avail_rxq_count(pf) / num_vfs;
drivers/net/ethernet/intel/ice/ice_sriov.c
423
pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq);
drivers/net/ethernet/intel/ice/ice_sriov.c
424
pf->vfs.num_msix_per = num_msix_per_vf;
drivers/net/ethernet/intel/ice/ice_sriov.c
426
num_vfs, pf->vfs.num_msix_per, pf->vfs.num_qps_per);
drivers/net/ethernet/intel/ice/ice_sriov.c
440
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
444
vf->first_vector_idx = ice_virt_get_irqs(pf, vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
467
static int ice_start_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
469
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
474
lockdep_assert_held(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
477
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
482
dev_err(ice_pf_to_dev(pf), "Failed to initialize VSI resources for VF %d, error %d\n",
drivers/net/ethernet/intel/ice/ice_sriov.c
487
retval = ice_eswitch_attach_vf(pf, vf);
drivers/net/ethernet/intel/ice/ice_sriov.c
489
dev_err(ice_pf_to_dev(pf), "Failed to attach VF %d to eswitch, error %d",
drivers/net/ethernet/intel/ice/ice_sriov.c
50
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
505
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
537
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
552
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
554
wr32(&pf->hw, VF_MBX_ARQLEN(vf->vf_id), 0);
drivers/net/ethernet/intel/ice/ice_sriov.c
555
wr32(&pf->hw, VF_MBX_ATQLEN(vf->vf_id), 0);
drivers/net/ethernet/intel/ice/ice_sriov.c
567
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
573
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
574
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
614
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
623
reg = rd32(&pf->hw, VPGEN_VFRSTAT(vf->vf_id));
drivers/net/ethernet/intel/ice/ice_sriov.c
639
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
655
wr32(&vf->pf->hw, VFGEN_RSTAT(vf->vf_id), VIRTCHNL_VFR_VFACTIVE);
drivers/net/ethernet/intel/ice/ice_sriov.c
683
static int ice_create_vf_entries(struct ice_pf *pf, u16 num_vfs)
drivers/net/ethernet/intel/ice/ice_sriov.c
685
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/intel/ice/ice_sriov.c
686
struct ice_vfs *vfs = &pf->vfs;
drivers/net/ethernet/intel/ice/ice_sriov.c
705
vf->pf = pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
717
vf->vf_sw_id = pf->first_sw;
drivers/net/ethernet/intel/ice/ice_sriov.c
733
ice_free_vf_entries(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
742
static int ice_ena_vfs(struct ice_pf *pf, u16 num_vfs)
drivers/net/ethernet/intel/ice/ice_sriov.c
744
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
745
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
749
wr32(hw, GLINT_DYN_CTL(pf->oicr_irq.index),
drivers/net/ethernet/intel/ice/ice_sriov.c
751
set_bit(ICE_OICR_INTR_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_sriov.c
754
ret = pci_enable_sriov(pf->pdev, num_vfs);
drivers/net/ethernet/intel/ice/ice_sriov.c
758
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
760
ret = ice_set_per_vf_res(pf, num_vfs);
drivers/net/ethernet/intel/ice/ice_sriov.c
767
ret = ice_create_vf_entries(pf, num_vfs);
drivers/net/ethernet/intel/ice/ice_sriov.c
77
wr32(&pf->hw, GLINT_DYN_CTL(i), GLINT_DYN_CTL_CLEARPBA_M);
drivers/net/ethernet/intel/ice/ice_sriov.c
774
ret = ice_start_vfs(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
78
ice_flush(&pf->hw);
drivers/net/ethernet/intel/ice/ice_sriov.c
781
clear_bit(ICE_VF_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_sriov.c
784
if (test_and_clear_bit(ICE_OICR_INTR_DIS, pf->state))
drivers/net/ethernet/intel/ice/ice_sriov.c
787
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
792
ice_free_vf_entries(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
794
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_sriov.c
795
pci_disable_sriov(pf->pdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
799
clear_bit(ICE_OICR_INTR_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_sriov.c
810
static int ice_pci_sriov_ena(struct ice_pf *pf, int num_vfs)
drivers/net/ethernet/intel/ice/ice_sriov.c
812
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
816
ice_free_vfs(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
820
if (num_vfs > pf->vfs.num_supported) {
drivers/net/ethernet/intel/ice/ice_sriov.c
822
num_vfs, pf->vfs.num_supported);
drivers/net/ethernet/intel/ice/ice_sriov.c
827
err = ice_ena_vfs(pf, num_vfs);
drivers/net/ethernet/intel/ice/ice_sriov.c
833
set_bit(ICE_FLAG_SRIOV_ENA, pf->flags);
drivers/net/ethernet/intel/ice/ice_sriov.c
841
static int ice_check_sriov_allowed(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_sriov.c
843
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_sriov.c
845
if (!test_bit(ICE_FLAG_SRIOV_CAPABLE, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
850
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
855
if (!ice_pf_state_is_nominal(pf)) {
drivers/net/ethernet/intel/ice/ice_sriov.c
871
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
873
return pf->virt_irq_tracker.num_entries;
drivers/net/ethernet/intel/ice/ice_sriov.c
876
static void ice_sriov_remap_vectors(struct ice_pf *pf, u16 restricted_id)
drivers/net/ethernet/intel/ice/ice_sriov.c
885
ice_for_each_vf(pf, bkt, tmp_vf) {
drivers/net/ethernet/intel/ice/ice_sriov.c
892
ice_virt_free_irqs(pf, tmp_vf->first_vector_idx,
drivers/net/ethernet/intel/ice/ice_sriov.c
900
tmp_vf = ice_get_vf_by_id(pf, vf_ids[i]);
drivers/net/ethernet/intel/ice/ice_sriov.c
905
ice_virt_get_irqs(pf, tmp_vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
91
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_sriov.c
931
struct ice_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/intel/ice/ice_sriov.c
937
if (!ice_get_num_vfs(pf))
drivers/net/ethernet/intel/ice/ice_sriov.c
947
if (queues > min(ice_get_avail_txq_count(pf),
drivers/net/ethernet/intel/ice/ice_sriov.c
948
ice_get_avail_rxq_count(pf)))
drivers/net/ethernet/intel/ice/ice_sriov.c
954
vf = ice_get_vf_by_dev(pf, vf_dev);
drivers/net/ethernet/intel/ice/ice_sriov.c
97
hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_sriov.c
974
ice_virt_free_irqs(pf, vf->first_vector_idx, vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
977
ice_sriov_remap_vectors(pf, vf->vf_id);
drivers/net/ethernet/intel/ice/ice_sriov.c
981
vf->first_vector_idx = ice_virt_get_irqs(pf, vf->num_msix);
drivers/net/ethernet/intel/ice/ice_sriov.c
994
dev_info(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_sriov.h
29
void ice_process_vflr_event(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_sriov.h
31
int __ice_set_vf_mac(struct ice_pf *pf, u16 vf_id, const u8 *mac);
drivers/net/ethernet/intel/ice/ice_sriov.h
36
void ice_free_vfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_sriov.h
37
void ice_restore_all_vfs_msi_state(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_sriov.h
59
ice_vf_lan_overflow_event(struct ice_pf *pf, struct ice_rq_event_info *event);
drivers/net/ethernet/intel/ice/ice_sriov.h
60
void ice_print_vfs_mdd_events(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_sriov.h
69
static inline void ice_process_vflr_event(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_sriov.h
70
static inline void ice_free_vfs(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_sriov.h
72
void ice_vf_lan_overflow_event(struct ice_pf *pf, struct ice_rq_event_info *event) { }
drivers/net/ethernet/intel/ice/ice_sriov.h
73
static inline void ice_print_vfs_mdd_events(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_sriov.h
76
static inline void ice_restore_all_vfs_msi_state(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/ice_sriov.h
86
__ice_set_vf_mac(struct ice_pf __always_unused *pf,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1082
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1087
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1147
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1148
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1157
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1158
if (ice_is_safe_mode(pf)) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1262
pf->num_dmac_chnl_fltrs++;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2130
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2134
ice_handle_del_pf_lldp_drop_rule(pf);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2142
err = ice_rem_adv_rule_by_id(&pf->hw, &rule_rem);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2163
pf->num_dmac_chnl_fltrs--;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2230
ice_find_tc_flower_fltr(struct ice_pf *pf, unsigned long cookie)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2234
hlist_for_each_entry(fltr, &pf->tc_flower_fltr_list, tc_flower_node)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2257
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2260
if (ice_is_reset_in_progress(pf->state))
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2262
if (test_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags))
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2269
!test_bit(ICE_FLAG_CLS_FLOWER, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2280
fltr = ice_find_tc_flower_fltr(pf, cls_flower->cookie);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2292
hlist_add_head(&fltr->tc_flower_node, &pf->tc_flower_fltr_list);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2305
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2309
fltr = ice_find_tc_flower_fltr(pf, cls_flower->cookie);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2311
if (!test_bit(ICE_FLAG_TC_MQPRIO, pf->flags) &&
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2312
hlist_empty(&pf->tc_flower_fltr_list))
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2338
void ice_replay_tc_fltrs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2344
&pf->tc_flower_fltr_list,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
795
ice_find_pf_tx_lldp_fltr(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
799
hlist_for_each_entry(fltr, &pf->tc_flower_fltr_list, tc_flower_node)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
806
static bool ice_any_vf_lldp_tx_ena(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
811
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
825
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
834
if (!deinit && ice_any_vf_lldp_tx_ena(pf))
drivers/net/ethernet/intel/ice/ice_tc_lib.c
837
err = ice_rem_adv_rule_by_id(&pf->hw, &remove_entry);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
859
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
872
err = ice_add_adv_rule(&pf->hw, list, ARRAY_SIZE(list), &rinfo,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
875
dev_err(&pf->pdev->dev,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
890
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
892
hlist_for_each_entry(fltr, &pf->tc_flower_fltr_list, tc_flower_node) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
900
static void ice_handle_del_pf_lldp_drop_rule(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
905
ice_for_each_vsi(pf, i) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
906
struct ice_vsi *vf_vsi = pf->vsi[i];
drivers/net/ethernet/intel/ice/ice_tc_lib.h
208
static inline int ice_chnl_dmac_fltr_cnt(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_tc_lib.h
210
return pf->num_dmac_chnl_fltrs;
drivers/net/ethernet/intel/ice/ice_tc_lib.h
218
void ice_replay_tc_fltrs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_txrx.c
2302
struct ice_pf *pf = ice_netdev_to_pf(netdev);
drivers/net/ethernet/intel/ice/ice_txrx.c
2305
dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1001
vfs = &pf->vfs;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1022
if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT))
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1023
ice_mbx_vf_clear_cnt_e830(&pf->hw, vf->vf_id);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1025
ice_mbx_init_vf_info(&pf->hw, &vf->mbx_info);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1032
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1034
if (!ice_is_feature_supported(pf, ICE_F_MBX_LIMIT))
drivers/net/ethernet/intel/ice/ice_vf_lib.c
105
u16 ice_get_num_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1084
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1087
dev_err(ice_pf_to_dev(pf), "VF ID: %u in reset. Try again.\n",
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1100
return vf->pf->hw.port_info;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
112
ice_for_each_vf_rcu(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1252
ice_vsi_release(vf->pf->vsi[vf->ctrl_vsi_idx]);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1266
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1274
vsi = ice_vsi_setup(pf, &params);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1276
dev_err(ice_pf_to_dev(pf), "Failed to create VF control VSI\n");
drivers/net/ethernet/intel/ice/ice_vf_lib.c
128
return vf->pf->vsi[vf->lan_vsi_idx];
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1295
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1300
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1378
struct ice_vsi *ice_get_vf_ctrl_vsi(struct ice_pf *pf, struct ice_vsi *vsi)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1385
ice_for_each_vf_rcu(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
1387
ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/ice_vf_lib.c
145
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
147
return (test_bit(ICE_VF_DIS, pf->state) ||
drivers/net/ethernet/intel/ice/ice_vf_lib.c
25
struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
263
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
276
dev_err(ice_pf_to_dev(pf),
drivers/net/ethernet/intel/ice/ice_vf_lib.c
297
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
30
hash_for_each_possible_rcu(pf->vfs.table, vf, entry, vf_id) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
303
dev_err(ice_pf_to_dev(pf), "failed to rebuild VF %d VSI\n",
drivers/net/ethernet/intel/ice/ice_vf_lib.c
310
vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
326
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
373
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
419
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
427
if (ice_is_eswitch_mode_switchdev(vf->pf))
drivers/net/ethernet/intel/ice/ice_vf_lib.c
466
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
473
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
481
status = ice_move_vsi_to_agg(pf->hw.port_info, vsi->agg_node->agg_id,
drivers/net/ethernet/intel/ice/ice_vf_lib.c
496
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
576
bool ice_is_any_vf_in_unicast_promisc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
583
ice_for_each_vf_rcu(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
629
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
635
if (!test_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
643
dev_err(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode failed\n");
drivers/net/ethernet/intel/ice/ice_vf_lib.c
646
dev_info(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode succeeded\n");
drivers/net/ethernet/intel/ice/ice_vf_lib.c
653
dev_err(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode failed\n");
drivers/net/ethernet/intel/ice/ice_vf_lib.c
656
dev_info(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode succeeded\n");
drivers/net/ethernet/intel/ice/ice_vf_lib.c
729
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
731
if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT))
drivers/net/ethernet/intel/ice/ice_vf_lib.c
732
ice_mbx_vf_clear_cnt_e830(&pf->hw, vf->vf_id);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
748
void ice_reset_all_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
750
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
751
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
756
if (!ice_has_vfs(pf))
drivers/net/ethernet/intel/ice/ice_vf_lib.c
759
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
762
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
766
if (test_and_set_bit(ICE_VF_DIS, pf->state)) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
767
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
772
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
779
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
791
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
794
ice_eswitch_detach_vf(pf, vf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
810
ice_eswitch_attach_vf(pf, vf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
816
clear_bit(ICE_VF_DIS, pf->state);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
818
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
827
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
861
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.c
868
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
873
if (test_bit(ICE_VF_RESETS_DISABLED, pf->state)) {
drivers/net/ethernet/intel/ice/ice_vf_lib.c
884
mutex_lock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
885
act_prt = ice_lag_prepare_vf_reset(pf->lag);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
89
bool ice_has_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
94
return !hash_empty(pf->vfs.table);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
971
ice_lag_complete_vf_reset(pf->lag, act_prt);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
972
mutex_unlock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/ice_vf_lib.c
998
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.h
143
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_vf_lib.h
268
#define ice_for_each_vf(pf, bkt, vf) \
drivers/net/ethernet/intel/ice/ice_vf_lib.h
269
hash_for_each((pf)->vfs.table, (bkt), (vf), entry)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
285
#define ice_for_each_vf_rcu(pf, bkt, vf) \
drivers/net/ethernet/intel/ice/ice_vf_lib.h
286
hash_for_each_rcu((pf)->vfs.table, (bkt), (vf), entry)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
289
struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
291
static inline struct ice_vf *ice_get_vf_by_dev(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_vf_lib.h
299
return ice_get_vf_by_id(pf, pci_iov_vf_id(vf_dev));
drivers/net/ethernet/intel/ice/ice_vf_lib.h
303
bool ice_has_vfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
304
u16 ice_get_num_vfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
309
bool ice_is_any_vf_in_unicast_promisc(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
318
void ice_reset_all_vfs(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
319
struct ice_vsi *ice_get_vf_ctrl_vsi(struct ice_pf *pf, struct ice_vsi *vsi);
drivers/net/ethernet/intel/ice/ice_vf_lib.h
323
static inline struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
328
static inline struct ice_vf *ice_get_vf_by_dev(struct ice_pf *pf,
drivers/net/ethernet/intel/ice/ice_vf_lib.h
338
static inline bool ice_has_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
343
static inline u16 ice_get_num_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
367
static inline bool ice_is_any_vf_in_unicast_promisc(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
389
static inline void ice_reset_all_vfs(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/ice_vf_lib.h
394
ice_get_vf_ctrl_vsi(struct ice_pf *pf, struct ice_vsi *vsi)
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
133
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
144
vlan_ops = ice_is_dvm_ena(&pf->hw) ?
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
174
dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
247
dev_dbg(ice_pf_to_dev(vf->pf), "Failed to disable Rx VLAN filtering for old VF with VIRTCHNL_VF_OFFLOAD_VLAN support\n");
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
27
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
32
if (ice_is_dvm_ena(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
60
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
70
if (ice_is_dvm_ena(&pf->hw)) {
drivers/net/ethernet/intel/ice/ice_vf_vsi_vlan_ops.c
84
if (!test_bit(ICE_FLAG_VF_VLAN_PRUNING, pf->flags))
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
325
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
338
pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
352
status = ice_update_vsi(&pf->hw, vsi->idx, ctxt, NULL);
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
356
libie_aq_str(pf->hw.adminq.sq_last_status));
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
66
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_vsi_vlan_lib.c
73
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/ice_xsk.c
109
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_xsk.c
110
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_xsk.c
51
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_xsk.c
52
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/ice_xsk.c
81
struct ice_pf *pf = vsi->back;
drivers/net/ethernet/intel/ice/ice_xsk.c
82
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
100
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
102
if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
drivers/net/ethernet/intel/ice/virt/fdir.c
1153
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
130
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
135
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1376
devm_kfree(ice_pf_to_dev(vf->pf), conf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1397
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1402
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1403
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1404
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
1414
ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/virt/fdir.c
1459
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1464
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1480
set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
drivers/net/ethernet/intel/ice/virt/fdir.c
1481
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1493
struct ice_pf *pf = ctrl_vsi->back;
drivers/net/ethernet/intel/ice/virt/fdir.c
1508
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1529
set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
drivers/net/ethernet/intel/ice/virt/fdir.c
1530
ice_service_task_schedule(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1542
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1546
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1547
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
1548
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
158
pf->vsi[vf->ctrl_vsi_idx] = NULL;
drivers/net/ethernet/intel/ice/virt/fdir.c
1592
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1685
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1750
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
177
fdir->fdir_prof = devm_kcalloc(ice_pf_to_dev(vf->pf),
drivers/net/ethernet/intel/ice/virt/fdir.c
1803
void ice_flush_fdir_ctx(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/virt/fdir.c
1808
if (!test_and_clear_bit(ICE_FD_VF_FLUSH_CTX, pf->state))
drivers/net/ethernet/intel/ice/virt/fdir.c
1811
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/fdir.c
1812
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/virt/fdir.c
1813
struct device *dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
186
fdir->fdir_prof[flow] = devm_kzalloc(ice_pf_to_dev(vf->pf),
drivers/net/ethernet/intel/ice/virt/fdir.c
1868
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/fdir.c
1883
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1981
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
1987
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
1988
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
2001
ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/virt/fdir.c
2093
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
2098
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
2099
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
2108
if (!ice_fdir_num_avail_fltr(&pf->hw, vf_vsi) ||
drivers/net/ethernet/intel/ice/virt/fdir.c
212
devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof[flow]);
drivers/net/ethernet/intel/ice/virt/fdir.c
2251
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
2259
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
2260
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
2280
ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/virt/fdir.c
231
devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof);
drivers/net/ethernet/intel/ice/virt/fdir.c
2321
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
2326
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
2327
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
294
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
338
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
481
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
488
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
489
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
644
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
650
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
651
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
652
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
657
ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
drivers/net/ethernet/intel/ice/virt/fdir.c
752
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.c
837
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/fdir.c
857
hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/fdir.c
927
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/fdir.h
51
void ice_flush_fdir_ctx(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/virt/fdir.h
55
static inline void ice_flush_fdir_ctx(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/virt/queues.c
133
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/queues.c
135
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/queues.c
148
n_used = pf->num_quanta_prof_used;
drivers/net/ethernet/intel/ice/virt/queues.c
151
pf->num_quanta_prof_used++;
drivers/net/ethernet/intel/ice/virt/queues.c
615
dev_err(ice_pf_to_dev(vf->pf), "VF-%d trying to configure more than allocated number of queues: %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
624
dev_warn(ice_pf_to_dev(vf->pf), "The maximum queue %d rate limit configuration may not take effect because the maximum TX rate for VF-%d is %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
632
dev_warn(ice_pf_to_dev(vf->pf), "The minimum queue %d rate limit configuration may not take effect because the minimum TX rate for VF-%d is %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
639
dev_warn(ice_pf_to_dev(vf->pf), "VF-%d trying to configure invalid queue_id\n",
drivers/net/ethernet/intel/ice/virt/queues.c
645
dev_warn(ice_pf_to_dev(vf->pf), "VF-%d trying to configure a traffic class higher than allowed\n",
drivers/net/ethernet/intel/ice/virt/queues.c
707
dev_err(ice_pf_to_dev(vf->pf), "VF-%d trying to configure more than allocated number of queues: %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
72
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/queues.c
721
dev_err(ice_pf_to_dev(vf->pf), "quanta size should be the product of 64\n");
drivers/net/ethernet/intel/ice/virt/queues.c
754
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/queues.c
760
mutex_lock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/virt/queues.c
761
act_prt = ice_lag_prepare_vf_reset(pf->lag);
drivers/net/ethernet/intel/ice/virt/queues.c
775
dev_err(ice_pf_to_dev(pf), "VF-%d requesting more than supported number of queues: %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
821
dev_warn(ice_pf_to_dev(pf), "VF-%d failed to configure TX queue %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
861
dev_warn(ice_pf_to_dev(pf), "VF-%d failed to configure RX queue %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
874
if (!(BIT(rxdid) & pf->supported_rxdids))
drivers/net/ethernet/intel/ice/virt/queues.c
891
ice_lag_complete_vf_reset(pf->lag, act_prt);
drivers/net/ethernet/intel/ice/virt/queues.c
892
mutex_unlock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/virt/queues.c
901
dev_err(ice_pf_to_dev(pf), "VF-%d could not disable RX queue %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
904
dev_err(ice_pf_to_dev(pf), "VF-%d could not disable TX queue %d\n",
drivers/net/ethernet/intel/ice/virt/queues.c
908
ice_lag_complete_vf_reset(pf->lag, act_prt);
drivers/net/ethernet/intel/ice/virt/queues.c
909
mutex_unlock(&pf->lag_mutex);
drivers/net/ethernet/intel/ice/virt/queues.c
932
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/queues.c
938
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/queues.c
945
tx_rx_queue_left = min_t(u16, ice_get_avail_txq_count(pf),
drivers/net/ethernet/intel/ice/virt/queues.c
946
ice_get_avail_rxq_count(pf));
drivers/net/ethernet/intel/ice/virt/rss.c
1304
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1306
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1339
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1341
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1377
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1441
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1443
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1518
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1519
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1569
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1570
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
1576
if (!test_bit(ICE_FLAG_RSS_ENA, vf->pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1708
if (!test_bit(ICE_FLAG_RSS_ENA, vf->pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1754
if (!test_bit(ICE_FLAG_RSS_ENA, vf->pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1796
if (!test_bit(ICE_FLAG_RSS_ENA, vf->pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1832
if (!test_bit(ICE_FLAG_RSS_ENA, vf->pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1833
dev_err(ice_pf_to_dev(vf->pf), "RSS not supported by PF\n");
drivers/net/ethernet/intel/ice/virt/rss.c
1864
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/rss.c
1869
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/rss.c
1876
if (!test_bit(ICE_FLAG_RSS_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/virt/rss.c
1892
status = ice_rem_vsi_rss_cfg(&pf->hw, vsi->idx);
drivers/net/ethernet/intel/ice/virt/rss.c
1909
status = ice_add_avf_rss_cfg(&pf->hw, vsi, vrh->hashcfg);
drivers/net/ethernet/intel/ice/virt/rss.c
505
return ice_hw_ptype_ena(&vf->pf->hw, ptype);
drivers/net/ethernet/intel/ice/virt/rss.c
740
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
742
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/rss.c
773
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/rss.c
775
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
104
void ice_vc_notify_link_state(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1043
test_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, vf->pf->flags))
drivers/net/ethernet/intel/ice/virt/virtchnl.c
109
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
110
ice_for_each_vf(pf, bkt, vf)
drivers/net/ethernet/intel/ice/virt/virtchnl.c
112
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1135
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1142
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1185
if (ice_vf_is_port_vlan_ena(vf) && !ice_is_dvm_ena(&pf->hw)) {
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1195
!ice_is_dvm_ena(&pf->hw) &&
drivers/net/ethernet/intel/ice/virt/virtchnl.c
121
void ice_vc_notify_reset(struct ice_pf *pf)
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1228
if (!vlan_promisc && vid && !ice_is_dvm_ena(&pf->hw)) {
drivers/net/ethernet/intel/ice/virt/virtchnl.c
125
if (!ice_has_vfs(pf))
drivers/net/ethernet/intel/ice/virt/virtchnl.c
130
ice_vc_vf_broadcast(pf, VIRTCHNL_OP_EVENT, VIRTCHNL_STATUS_SUCCESS,
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1426
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1439
rxdid = pf->supported_rxdids;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
149
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
152
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
153
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
155
aq_ret = ice_aq_send_msg_to_vf(&pf->hw, vf->vf_id, v_opcode, v_retval,
drivers/net/ethernet/intel/ice/virt/virtchnl.c
157
if (aq_ret && pf->hw.mailboxq.sq_last_status != LIBIE_AQ_RC_ENOSYS) {
drivers/net/ethernet/intel/ice/virt/virtchnl.c
160
libie_aq_str(pf->hw.mailboxq.sq_last_status));
drivers/net/ethernet/intel/ice/virt/virtchnl.c
1668
if (ice_is_dvm_ena(&vf->pf->hw))
drivers/net/ethernet/intel/ice/virt/virtchnl.c
203
if (ice_is_eswitch_mode_switchdev(vf->pf))
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2471
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2488
phc_time->time = ice_ptp_read_src_clk_reg(pf, NULL);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
249
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2565
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2574
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2590
dev_err(ice_pf_to_dev(pf), "VF attempting to override administratively set MAC address\n");
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2627
dev_dbg(ice_pf_to_dev(vf->pf),
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2697
struct ice_pf *pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
27
ice_vc_vf_broadcast(struct ice_pf *pf, enum virtchnl_ops v_opcode,
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2700
pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2701
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2707
status = ice_mbx_vf_state_handler(&pf->hw, mbxdata, &vf->mbx_info,
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2714
struct ice_vsi *pf_vsi = ice_get_main_vsi(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2736
void ice_vc_process_vf_msg(struct ice_pf *pf, struct ice_rq_event_info *event,
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2748
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
2750
vf = ice_get_vf_by_id(pf, vf_id);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
30
struct ice_hw *hw = &pf->hw;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
34
mutex_lock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
35
ice_for_each_vf(pf, bkt, vf) {
drivers/net/ethernet/intel/ice/virt/virtchnl.c
405
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
423
pi = pf->hw.port_info;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
47
mutex_unlock(&pf->vfs.table_lock);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
485
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
507
dev = ice_pf_to_dev(pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
536
if (!test_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, pf->flags)) {
drivers/net/ethernet/intel/ice/virt/virtchnl.c
769
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
798
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
84
struct ice_hw *hw = &vf->pf->hw;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
896
struct device *dev = ice_pf_to_dev(vf->pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.c
941
struct ice_pf *pf = vf->pf;
drivers/net/ethernet/intel/ice/virt/virtchnl.c
966
dev_err(ice_pf_to_dev(pf), "Can't add more MAC addresses, because VF-%d is not trusted, switch the VF to trusted mode in order to add more functionalities\n",
drivers/net/ethernet/intel/ice/virt/virtchnl.h
104
static inline void ice_vc_notify_link_state(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/virt/virtchnl.h
105
static inline void ice_vc_notify_reset(struct ice_pf *pf) { }
drivers/net/ethernet/intel/ice/virt/virtchnl.h
134
ice_vc_process_vf_msg(struct ice_pf *pf, struct ice_rq_event_info *event,
drivers/net/ethernet/intel/ice/virt/virtchnl.h
87
void ice_vc_notify_link_state(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.h
88
void ice_vc_notify_reset(struct ice_pf *pf);
drivers/net/ethernet/intel/ice/virt/virtchnl.h
93
void ice_vc_process_vf_msg(struct ice_pf *pf, struct ice_rq_event_info *event,
drivers/net/ethernet/intel/igb/e1000_82575.c
2068
void igb_vmdq_set_anti_spoofing_pf(struct e1000_hw *hw, bool enable, int pf)
drivers/net/ethernet/intel/igb/e1000_82575.c
2091
reg_val ^= (BIT(pf) | BIT(pf + MAX_NUM_VFS));
drivers/net/ethernet/marvell/octeon_ep/octep_regs_cn9k_pf.h
398
static inline u64 cn9k_pemx_pfx_csx_pfcfgx(u64 pem, u32 pf, u32 offset)
drivers/net/ethernet/marvell/octeon_ep/octep_regs_cn9k_pf.h
404
pf_addr_bits = FIELD_PREP(CN9K_PF_GENMASK, pf);
drivers/net/ethernet/marvell/octeon_ep/octep_regs_cnxk_pf.h
401
#define CNXK_PEMX_PFX_CSX_PFCFGX(pem, pf, offset) ({ typeof(offset) _off = (offset); \
drivers/net/ethernet/marvell/octeon_ep/octep_regs_cnxk_pf.h
404
| (pf) << 18 \
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
273
int pf, err;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
296
for (pf = 0; pf < ndevs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
297
rvu_write64(rvu, BLKADDR_RVUM, RVU_MBOX_AF_PFX_ADDR(pf),
drivers/net/ethernet/marvell/octeontx2/af/mcs.h
147
struct mcs_pfvf *pf;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
100
pfvf = &mcs->pf[rvu_get_pf(rvu->pdev, pcifunc)];
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
124
int pf;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
126
pf = rvu_get_pf(rvu->pdev, event->pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
130
req = otx2_mbox_alloc_msg_mcs_intr_notify(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
142
otx2_mbox_wait_for_zero(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
144
otx2_mbox_msg_send_up(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
146
otx2_mbox_wait_for_rsp(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
196
pfvf = &mcs->pf[rvu_get_pf(rvu->pdev, pcifunc)];
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
902
mcs->pf = devm_kcalloc(mcs->dev, hw->total_pfs,
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
904
if (!mcs->pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1073
rvu->pf = devm_kcalloc(rvu->dev, hw->total_pfs,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1075
if (!rvu->pf) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1488
struct rvu_pfvf *pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1490
pf = rvu_get_pfvf(rvu, pcifunc & ~RVU_PFVF_FUNC_MASK);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1494
blkaddr = pf->nix_blkaddr;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1507
if (is_sdp_pfvf(rvu, pcifunc) && pf->sdp_info->node_id == 1)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2869
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2871
pf = flrwork - rvu->flr_wrk;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2872
if (pf >= rvu->hw->total_pfs) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2873
rvu_afvf_flr_handler(rvu, pf - rvu->hw->total_pfs);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2877
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2879
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2887
rvu_write64(rvu, BLKADDR_RVUM, RVU_AF_PFTRPEND, BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2890
rvu_write64(rvu, BLKADDR_RVUM, RVU_AF_PFFLR_INT_ENA_W1S, BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2921
u8 pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2927
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2928
if (intr & (1ULL << pf)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2931
BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2934
BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2936
queue_work(rvu->flr_wq, &rvu->flr_wrk[pf].work);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2988
u8 pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2995
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2996
if (intr & (1ULL << pf)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2999
BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3002
BIT_ULL(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3043
struct rvu_pfvf *pfvf = &rvu->pf[0];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3046
pfvf = &rvu->pf[0];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3273
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3276
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3277
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3278
rvu_write64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf),
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3576
rvu_alloc_cint_qint_mem(rvu, &rvu->pf[RVU_AFPF], BLKADDR_NIX0,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
403
void rvu_get_pf_numvfs(struct rvu *rvu, int pf, int *numvfs, int *hwvf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
408
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
417
int pf, func;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
420
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
424
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
435
return &rvu->pf[rvu_get_pf(rvu->pdev, pcifunc)];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
440
int pf, vf, nvfs;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
443
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
444
if (pf >= rvu->hw->total_pfs)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
452
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
573
static void rvu_check_min_msix_vec(struct rvu *rvu, int nvecs, int pf, int vf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
583
pf, vf - 1, nvecs);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
588
if (pf == 0)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
597
pf, nvecs, min_vecs);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
603
int pf, vf, numvfs, hwvf, err;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
609
for (pf = 0; pf < hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
610
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
615
rvu_get_pf_numvfs(rvu, pf, &numvfs, &hwvf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
617
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
619
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_MSIX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
621
rvu_check_min_msix_vec(rvu, pfvf->msix.max, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
637
if (!pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
644
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_INT_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
649
RVU_PRIV_PFX_INT_CFG(pf), cfg | offset);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
656
RVU_PRIV_PFX_MSIX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
658
rvu_check_min_msix_vec(rvu, pfvf->msix.max, pf, vf + 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
738
pfvf = &rvu->pf[id];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
760
pfvf = &rvu->pf[RVU_AFPF];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
768
int pf, vf, numvfs, hwvf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
772
for (pf = 0; pf < hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
774
if (!pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
777
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
780
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
781
if (rvu->fwdata && pf < PF_MACNUM_MAX) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
782
mac = &rvu->fwdata->pf_macs[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
794
rvu_get_pf_numvfs(rvu, pf, &numvfs, &hwvf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
1050
int rvu_npc_get_pkind(struct rvu *rvu, u16 pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
1052
int npc_config_ts_kpuaction(struct rvu *rvu, int pf, u16 pcifunc, bool en);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
1108
bool is_mac_feature_supported(struct rvu *rvu, int pf, int feature);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
54
static inline u16 rvu_make_pcifunc(struct pci_dev *pdev, int pf, int func)
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
57
return ((pf & RVU_CN20K_PFVF_PF_MASK) <<
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
583
struct rvu_pfvf *pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
62
return ((pf & RVU_OTX2_PFVF_PF_MASK) <<
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
911
void rvu_get_pf_numvfs(struct rvu *rvu, int pf, int *numvfs, int *hwvf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
961
static inline bool is_pf_cgxmapped(struct rvu *rvu, u8 pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
963
return (pf >= PF_CGXMAP_BASE && pf <= rvu->cgx_mapped_pfs) &&
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
964
!is_sdp_pf(rvu, rvu_make_pcifunc(rvu->pdev, pf, 0));
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
992
void rvu_cgx_enadis_rx_bp(struct rvu *rvu, int pf, bool enable);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1003
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1009
if (!is_mac_feature_supported(rvu, pf, RVU_LMAC_FEAT_FC))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1015
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1018
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1044
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1053
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1056
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1071
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1074
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1077
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
108
static void rvu_map_cgx_nix_block(struct rvu *rvu, int pf,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1090
int pf, lf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1097
pf = cgxlmac_to_pf(rvu, cgx_get_cgxid(cgxd), lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1098
if (pf < 0)
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1099
return pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1104
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
111
struct rvu_pfvf *pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1134
parent_pf = &rvu->pf[rvu_get_pf(rvu->pdev, pcifunc)];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1177
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1180
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1185
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1193
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1199
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1202
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1220
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1231
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_idx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1246
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
125
int pf = PF_CGXMAP_BASE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1252
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1264
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1273
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1280
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1289
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1292
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1318
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1327
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1330
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1343
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1348
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1351
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
167
rvu->pf2cgxlmac_map[pf] = cgxlmac_id_to_bmap(cgx, lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
168
rvu->cgxlmac2pf_map[CGX_OFFSET(cgx) + lmac] = 1 << pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
170
pkind->pfchan_map[free_pkind] = ((pf) & 0x3F) << 16;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
171
rvu_map_cgx_nix_block(rvu, pf, cgx, lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
173
rvu_get_pf_numvfs(rvu, pf, &numvfs, &hwvfs);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
175
pf++;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
44
bool is_mac_feature_supported(struct rvu *rvu, int pf, int feature)
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
465
void rvu_cgx_enadis_rx_bp(struct rvu *rvu, int pf, bool enable)
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
471
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
474
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
487
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
49
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
495
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
504
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
512
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
52
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
529
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
544
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
580
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
590
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_idx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
636
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
645
parent_pf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
655
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_idx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
666
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
673
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_idx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
684
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
688
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
694
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
696
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
707
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
717
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
731
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
740
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
749
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
766
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
788
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
798
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
807
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
817
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
826
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
831
if (!is_mac_feature_supported(rvu, pf, RVU_LMAC_FEAT_PTP))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
837
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
840
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
849
if (npc_config_ts_kpuaction(rvu, pf, pcifunc, enable))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
876
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
882
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
885
set_bit(pf, &rvu->pf_notify_bmap);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
889
clear_bit(pf, &rvu->pf_notify_bmap);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
913
int pf, err;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
915
pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
917
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
920
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
931
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
935
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
938
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_idx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
973
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
980
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cn10k.c
76
u64 pa, val, pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cn10k.c
86
pf = rvu_get_pf(rvu->pdev, pcifunc) & RVU_OTX2_PFVF_PF_MASK;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cn10k.c
87
val = BIT_ULL(63) | BIT_ULL(14) | BIT_ULL(13) | pf << 8 |
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2975
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2982
pf = cgxlmac_to_pf(rvu, cgx_get_cgxid(cgxd), lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2985
pdev = pci_get_domain_bus_and_slot(domain, pf + 1, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2996
dev_name(&pdev->dev), pf, bcast, mcast);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3199
int pf, vf, numvfs, blkaddr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3245
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3246
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3249
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3252
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, (vf + 1));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3509
int pf, vf = -1;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3523
pf = rvu_get_pf(rvu->pdev, iter->owner);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3524
seq_printf(s, "\n\tInstalled by: PF%d ", pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3541
pf = rvu_get_pf(rvu->pdev, target);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3542
seq_printf(s, "\tForward to: PF%d ", pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
556
int pf, vf, num_vfs, hw_vfs;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
597
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
599
pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
601
index = pf * apr_vfs * LMT_MAPTBL_ENTRY_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
612
rvu_get_pf_numvfs(rvu, pf, &num_vfs, &hw_vfs);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
614
index = (pf * apr_vfs * LMT_MAPTBL_ENTRY_SIZE) +
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
617
"PF%d:VF%d \t\t", pf, vf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
682
int index, pf, vf, lf_str_size = 12, buf_size = 256;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
691
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
693
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, vf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
722
int pf, vf, pcifunc;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
760
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
764
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, vf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
769
sprintf(lfs, "PF%d:VF%d", pf, vf - 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
774
sprintf(lfs, "PF%d", pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
826
int pf, domain, blkid;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
837
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
838
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
841
pdev = pci_get_domain_bus_and_slot(domain, pf + 1, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
847
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
855
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2061
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2066
} else if (is_pf_cgxmapped(rvu, pf)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2067
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2079
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2085
} else if (is_pf_cgxmapped(rvu, pf)) { /* CGX links */
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2437
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2451
if (is_pf_cgxmapped(rvu, pf)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2452
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2468
rvu_cgx_enadis_rx_bp(rvu, pf, false);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2521
rvu_cgx_enadis_rx_bp(rvu, pf, true);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2831
u8 pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2835
if (!is_pf_cgxmapped(rvu, pf) && !is_rep_dev(rvu, pcifunc))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
338
int pkind, pf, vf, lbkid, vfid;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
343
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
344
if (!is_pf_cgxmapped(rvu, pf) && type != NIX_INTF_TYPE_LBK &&
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
350
pfvf->cgx_lmac = rvu->pf2cgxlmac_map[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
353
pkind = rvu_npc_get_pkind(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3554
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3563
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3564
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3598
int err, pf, numvfs, idx;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3604
for (pf = 1; pf < (rvu->cgx_mapped_pfs + 1); pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3605
cfg = rvu_read64(rvu, BLKADDR_RVUM, RVU_PRIV_PFX_CFG(pf));
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3612
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3632
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
420
parent_pf = &rvu->pf[rvu_get_pf(rvu->pdev, pcifunc)];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4567
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4583
rvu_get_pf_numvfs(rvu, pf, &numvfs, &hwvf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4596
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4614
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4650
if (is_pf_cgxmapped(rvu, pf)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4652
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx, &lmac);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4654
} else if (pf == 0) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5252
int nixlf, err, pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5270
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5271
if (is_pf_cgxmapped(rvu, pf) && rvu->rep_mode)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5282
int nixlf, err, pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5303
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5304
if (is_pf_cgxmapped(rvu, pf) && rvu->rep_mode)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5315
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5333
if (is_pf_cgxmapped(rvu, pf) && rvu->rep_mode)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5365
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5370
if (npc_config_ts_kpuaction(rvu, pf, pcifunc, false))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5400
int blkaddr, pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5404
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5405
if (!is_mac_feature_supported(rvu, pf, RVU_LMAC_FEAT_PTP))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
586
int blkaddr, pf, type, err;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
594
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
596
if (!is_pf_cgxmapped(rvu, pf) && type != NIX_INTF_TYPE_LBK)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
732
int blkaddr, pf, type, chan_id = 0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
740
pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
746
if (!is_pf_cgxmapped(rvu, pf) && type != NIX_INTF_TYPE_LBK &&
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
152
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
156
if (pf && !(pcifunc & RVU_PFVF_FUNC_MASK)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
158
pf--;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
159
index = mcam->pf_offset + (pf * RSVD_MCAM_ENTRIES_PER_PF);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
3442
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
3447
rxpkind = rvu_npc_get_pkind(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
3468
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
78
int rvu_npc_get_pkind(struct rvu *rvu, u16 pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
86
if (((map >> 16) & 0x3F) == pf)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
94
int npc_config_ts_kpuaction(struct rvu *rvu, int pf, u16 pcifunc, bool enable)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
99
pkind = rvu_npc_get_pkind(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1468
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1476
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1515
int pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1523
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1563
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1569
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1576
dev_err(rvu->dev, "%s MAC (%pM) del PF=%d failed\n", __func__, pfvf->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1581
__func__, pfvf->mac_addr, pf, seq_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1596
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1611
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1636
__func__, req->mac_addr, pfvf->mac_addr, pfvf->default_mac, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1646
pfvf->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1654
req->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1661
__func__, req->mac_addr, pfvf->mac_addr, pfvf->default_mac, pf, seq_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1678
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1684
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1694
__func__, req->mac_addr, pf, seq_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1699
req->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1714
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1720
__func__, pf, req->index);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1725
__func__, pf, req->index);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1739
int pf = rvu_get_pf(rvu->pdev, req->hdr.pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1746
rvu_get_cgx_lmac_id(rvu->pf2cgxlmac_map[pf], &cgx_id, &lmac_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1748
pfvf = &rvu->pf[pf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1760
__func__, req->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1768
__func__, pfvf->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1783
__func__, req->mac_addr, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.c
1792
__func__, req->mac_addr, pf, seq_id);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_hash.h
142
#define RVU_PFFUNC(pdev, pf, func) rvu_make_pcifunc(pdev, pf, func)
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
115
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
120
pf = rvu_get_pf(rvu->pdev, rvu->rep_pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
123
req = otx2_mbox_alloc_msg_rep_event_up_notify(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
134
otx2_mbox_wait_for_zero(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
135
otx2_mbox_msg_send_up(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
320
int pf, vf, numvfs;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
324
for (pf = 1; pf < hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
325
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
328
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
346
rvu_get_pf_numvfs(rvu, pf, &numvfs, NULL);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
348
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, vf + 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
40
int pf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
42
pf = rvu_get_pf(rvu->pdev, event->pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
441
int pf, vf, numvfs, hwvf, rep = 0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
453
for (pf = 0; pf < rvu->hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
454
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
456
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
460
rvu_get_pf_numvfs(rvu, pf, &numvfs, &hwvf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
48
msg = otx2_mbox_alloc_msg_rep_event_up_notify(rvu, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
59
otx2_mbox_wait_for_zero(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
61
otx2_mbox_msg_send_up(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
63
otx2_mbox_wait_for_rsp(&rvu->afpf_wq_info.mbox_up, pf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_sdp.c
22
u16 pf = rvu_get_pf(rvu->pdev, pcifunc);
drivers/net/ethernet/marvell/octeontx2/af/rvu_sdp.c
26
if (pf == sdp_pf_num[i])
drivers/net/ethernet/marvell/octeontx2/af/rvu_sdp.c
61
pfvf = &rvu->pf[sdp_pf_num[0]];
drivers/net/ethernet/marvell/octeontx2/af/rvu_sdp.c
72
pfvf = &rvu->pf[sdp_pf_num[i]];
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
114
pf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
121
pf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
127
rvu_get_pf_numvfs(rvu, pf, &numvfs, NULL);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
129
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, (vf + 1));
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
136
pf, vf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
145
pf, vf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
224
int pf, vf, numvfs;
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
235
for (pf = 1; pf < hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
236
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
239
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
244
pf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
249
rvu_get_pf_numvfs(rvu, pf, &numvfs, NULL);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
251
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, (vf + 1));
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
256
pf, vf, err);
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
89
int pf, vf, numvfs;
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
92
for (pf = 1; pf < hw->total_pfs; pf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
93
if (!is_pf_cgxmapped(rvu, pf))
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
96
pcifunc = rvu_make_pcifunc(rvu->pdev, pf, 0);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
103
req->nix_pf_func = pf->pcifunc;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
1037
cn10k_cpt_inst_flush(pf, &inst, sizeof(struct cpt_inst_s));
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
107
ret = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
110
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
114
static void cn10k_outb_cptlf_free(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
116
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
117
otx2_mbox_alloc_msg_cpt_lf_free(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
118
otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
119
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
122
static int cn10k_outb_cptlf_config(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
127
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
128
req = otx2_mbox_alloc_msg_cpt_inline_ipsec_cfg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
134
req->nix_pf_func = pf->pcifunc;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
135
ret = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
137
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
141
static void cn10k_outb_cptlf_iq_enable(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
146
reg_val = otx2_read64(pf, CN10K_CPT_LF_INPROG);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
148
otx2_write64(pf, CN10K_CPT_LF_INPROG, reg_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
151
reg_val = otx2_read64(pf, CN10K_CPT_LF_CTL);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
153
otx2_write64(pf, CN10K_CPT_LF_CTL, reg_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
156
static void cn10k_outb_cptlf_iq_disable(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
165
otx2_write64(pf, CN10K_CPT_LF_CTL, 0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
171
reg_val = otx2_read64(pf, CN10K_CPT_LF_INPROG);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
178
netdev_err(pf->netdev, "Timeout to cleanup CPT IQ\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
187
otx2_write64(pf, CN10K_CPT_LF_INPROG, reg_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
192
reg_val = otx2_read64(pf, CN10K_CPT_LF_INPROG);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
197
reg_val = otx2_read64(pf, CN10K_CPT_LF_Q_GRP_PTR);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
204
reg_val = otx2_read64(pf, CN10K_CPT_LF_INPROG);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
22
static bool cn10k_cpt_device_set_inuse(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
233
static int cn10k_outb_cptlf_iq_alloc(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
235
struct cn10k_cpt_inst_queue *iq = &pf->ipsec.iq;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
240
iq->real_vaddr = dma_alloc_coherent(pf->dev, iq->size,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
255
static void cn10k_outb_cptlf_iq_free(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
257
struct cn10k_cpt_inst_queue *iq = &pf->ipsec.iq;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
260
dma_free_coherent(pf->dev, iq->size, iq->real_vaddr,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
267
static int cn10k_outb_cptlf_iq_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
27
state = atomic_cmpxchg(&pf->ipsec.cpt_state,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
273
ret = cn10k_outb_cptlf_iq_alloc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
278
cn10k_outb_cptlf_iq_disable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
281
otx2_write64(pf, CN10K_CPT_LF_Q_BASE, pf->ipsec.iq.dma_addr);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
286
otx2_write64(pf, CN10K_CPT_LF_Q_SIZE, reg_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
291
static int cn10k_outb_cptlf_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
296
ret = cn10k_outb_cptlf_iq_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
301
ret = cn10k_outb_cptlf_config(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
306
cn10k_outb_cptlf_iq_enable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
309
cn10k_outb_cptlf_iq_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
315
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
319
ret = cn10k_outb_cptlf_attach(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
324
ret = cn10k_outb_cptlf_alloc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
329
ret = cn10k_outb_cptlf_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
333
pf->ipsec.io_addr = (__force u64)otx2_get_regaddr(pf,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
337
pf->flags |= OTX2_FLAG_IPSEC_OFFLOAD_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
339
cn10k_cpt_device_set_available(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
343
cn10k_outb_cptlf_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
345
cn10k_outb_cptlf_detach(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
349
static int cn10k_outb_cpt_clean(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
353
if (!cn10k_cpt_device_set_inuse(pf)) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
354
netdev_err(pf->netdev, "CPT LF device unavailable\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
359
pf->flags &= ~OTX2_FLAG_IPSEC_OFFLOAD_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
362
cn10k_outb_cptlf_iq_disable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
365
otx2_write64(pf, CN10K_CPT_LF_Q_BASE, 0);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
366
otx2_write64(pf, CN10K_CPT_LF_Q_SIZE, 0);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
369
cn10k_outb_cptlf_iq_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
372
cn10k_outb_cptlf_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
373
ret = cn10k_outb_cptlf_detach(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
375
netdev_err(pf->netdev, "Failed to detach CPT LF\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
377
cn10k_cpt_device_set_unavailable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
381
static void cn10k_cpt_inst_flush(struct otx2_nic *pf, struct cpt_inst_s *inst,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
387
lmt_info = per_cpu_ptr(pf->hw.lmt_info, smp_processor_id());
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
39
static void cn10k_cpt_device_set_available(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
397
tar_addr |= pf->ipsec.io_addr | (((size / 16) - 1) & 0x7) << 4;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
403
static int cn10k_wait_for_cpt_respose(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
41
atomic_set(&pf->ipsec.cpt_state, CN10K_CPT_HW_AVAILABLE);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
411
netdev_err(pf->netdev, "CPT response timeout\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
419
netdev_err(pf->netdev, "compcode=%x doneint=%x\n",
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
421
netdev_err(pf->netdev, "uc_compcode=%x uc_info=%llx esn=%llx\n",
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
427
static int cn10k_outb_write_sa(struct otx2_nic *pf, struct qmem *sa_info)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
44
static void cn10k_cpt_device_set_unavailable(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
442
res = dma_alloc_coherent(pf->dev, sizeof(struct cpt_res_s),
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
448
sa_dptr = dma_alloc_coherent(pf->dev, sa_size, &dptr_iova, GFP_ATOMIC);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
450
dma_free_coherent(pf->dev, sizeof(struct cpt_res_s), res,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
46
atomic_set(&pf->ipsec.cpt_state, CN10K_CPT_HW_UNAVAILABLE);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
472
if (!cn10k_cpt_device_set_inuse(pf)) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
477
cn10k_cpt_inst_flush(pf, &inst, sizeof(struct cpt_inst_s));
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
479
ret = cn10k_wait_for_cpt_respose(pf, res);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
485
otx2_write64(pf, CN10K_CPT_LF_CTX_FLUSH, reg_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
488
cn10k_cpt_device_set_available(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
49
static int cn10k_outb_cptlf_attach(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
490
dma_free_coherent(pf->dev, sa_size, sa_dptr, dptr_iova);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
491
dma_free_coherent(pf->dev, sizeof(struct cpt_res_s), res, res_iova);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
526
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
534
pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
535
sa_entry->ctx_size = (pf->ipsec.sa_size / OTX2_ALIGN) & 0xF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
54
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
56
attach = otx2_mbox_alloc_msg_attach_resources(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
64
ret = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
67
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
672
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
679
pf = netdev_priv(dev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
681
err = qmem_alloc(pf->dev, &sa_info, pf->ipsec.sa_size, OTX2_ALIGN);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
688
err = cn10k_outb_write_sa(pf, sa_info);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
691
qmem_free(pf->dev, sa_info);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
697
if (!pf->ipsec.outb_sa_count)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
699
pf->ipsec.outb_sa_count++;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
71
static int cn10k_outb_cptlf_detach(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
717
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
723
pf = netdev_priv(dev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
730
sa_entry->ctx_size = (pf->ipsec.sa_size / OTX2_ALIGN) & 0xF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
733
err = cn10k_outb_write_sa(pf, sa_info);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
738
qmem_free(pf->dev, sa_info);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
743
if (!--pf->ipsec.outb_sa_count)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
744
queue_work(pf->ipsec.sa_workq, &pf->ipsec.sa_work);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
756
struct otx2_nic *pf = container_of(ipsec, struct otx2_nic, ipsec);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
76
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
761
netdev_update_features(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
767
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
77
detach = otx2_mbox_alloc_msg_detach_resources(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
770
if (!is_dev_support_ipsec_offload(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
778
if (pf->ipsec.outb_sa_count) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
779
netdev_err(pf->netdev, "SA installed on this device\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
783
return cn10k_outb_cpt_clean(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
788
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
791
if (!is_dev_support_ipsec_offload(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
798
pf->ipsec.sa_size = sa_size;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
800
INIT_WORK(&pf->ipsec.sa_work, cn10k_ipsec_sa_wq_handler);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
801
pf->ipsec.sa_workq = alloc_workqueue("cn10k_ipsec_sa_workq",
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
803
if (!pf->ipsec.sa_workq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
804
netdev_err(pf->netdev, "SA alloc workqueue failed\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
813
cn10k_cpt_device_set_unavailable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
818
void cn10k_ipsec_clean(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
820
if (!is_dev_support_ipsec_offload(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
823
if (!(pf->flags & OTX2_FLAG_IPSEC_OFFLOAD_ENABLED))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
826
if (pf->ipsec.sa_workq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
827
destroy_workqueue(pf->ipsec.sa_workq);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
828
pf->ipsec.sa_workq = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
831
cn10k_outb_cpt_clean(pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
85
ret = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
88
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
92
static int cn10k_outb_cptlf_alloc(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
931
bool cn10k_ipsec_transmit(struct otx2_nic *pf, struct netdev_queue *txq,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
948
if (!(pf->flags & OTX2_FLAG_IPSEC_OFFLOAD_ENABLED))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
964
if (dlen == 0 && netif_msg_tx_err(pf)) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
965
netdev_err(pf->netdev, "Invalid IP header, ip-length zero\n");
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
97
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
98
req = otx2_mbox_alloc_msg_cpt_lf_alloc(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
993
inst.rvu_pf_func = pf->pcifunc;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.h
227
void cn10k_ipsec_clean(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.h
231
bool cn10k_ipsec_transmit(struct otx2_nic *pf, struct netdev_queue *txq,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.h
240
static inline __maybe_unused void cn10k_ipsec_clean(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.h
258
cn10k_ipsec_transmit(struct otx2_nic *pf, struct netdev_queue *txq,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
102
void cn20k_enable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
105
otx2_write64(pf, RVU_MBOX_PF_VFPF_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
106
otx2_write64(pf, RVU_MBOX_PF_VFPF_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
107
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
108
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
111
otx2_write64(pf, RVU_MBOX_PF_VFPF_INT_ENA_W1SX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
112
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INT_ENA_W1SX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
115
otx2_write64(pf, RVU_MBOX_PF_VFPF_INT_ENA_W1SX(1),
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
117
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INT_ENA_W1SX(1),
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
122
void cn20k_disable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
127
otx2_write64(pf, RVU_MBOX_PF_VFPF_INT_ENA_W1CX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
128
otx2_write64(pf, RVU_MBOX_PF_VFPF_INT_ENA_W1CX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
129
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INT_ENA_W1CX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
130
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INT_ENA_W1CX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
132
otx2_write64(pf, RVU_MBOX_PF_VFPF_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
133
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
136
otx2_write64(pf, RVU_MBOX_PF_VFPF_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
137
otx2_write64(pf, RVU_MBOX_PF_VFPF1_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
142
vector = pci_irq_vector(pf->pdev, intr_vec);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
143
free_irq(vector, pf->hw.pfvf_irq_devid[vec]);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
150
struct otx2_nic *pf = irq_data->pf;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
158
intr = otx2_read64(pf, irq_data->intr_status);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
159
otx2_write64(pf, irq_data->intr_status, intr);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
16
struct otx2_nic *pf = pf_irq;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
160
mbox = pf->mbox_pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
163
trace_otx2_msg_interrupt(pf->pdev, "VF(s) to PF", intr);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
165
irq_data->pf_queue_work_hdlr(mbox, pf->mbox_pfvf_wq, irq_data->start,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
17
struct mbox *mw = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
171
int cn20k_register_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
173
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
179
irq_data = devm_kcalloc(pf->dev, 4,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
214
irq_data[vec].pf = pf;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
218
if (pf->pcifunc)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
220
"RVUPF%d_VF%d Mbox%d", rvu_get_pf(pf->pdev,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
221
pf->pcifunc), vec / 2, vec % 2);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
227
ret = request_irq(pci_irq_vector(pf->pdev, intr_vec),
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
228
pf->hw_ops->pfvf_mbox_intr_handler, 0,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
23
pf_trig_val = otx2_read64(pf, RVU_PF_INT) & 0x3ULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
232
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
238
cn20k_enable_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
26
otx2_write64(pf, RVU_PF_INT, pf_trig_val);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
35
queue_work(pf->mbox_wq, &mw->mbox_up_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
37
trace_otx2_msg_interrupt(pf->pdev, "UP message from AF to PF",
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
48
queue_work(pf->mbox_wq, &mw->mbox_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
49
trace_otx2_msg_interrupt(pf->pdev, "DOWN reply from AF to PF",
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.h
14
int cn20k_register_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.h
15
void cn20k_disable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.h
16
void cn20k_enable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1162
pfvf = wrk->pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1212
pfvf->refill_wrk[qidx].pf = pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1053
int otx2_init_rsrc(struct pci_dev *pdev, struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1055
int otx2_alloc_queue_mem(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1057
void otx2_free_hw_resources(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1058
int otx2_wq_init(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1059
int otx2_check_pf_usable(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1060
int otx2_pfaf_mbox_init(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1061
int otx2_register_mbox_intr(struct otx2_nic *pf, bool probe_af);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1062
int otx2_realloc_msix_vectors(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1063
void otx2_pfaf_mbox_destroy(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1064
void otx2_disable_mbox_intr(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1065
void otx2_disable_napi(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1088
void mbox_handler_nix_txsch_alloc(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1120
int otx2_mcam_flow_init(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1123
void otx2_mcam_flow_del(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1124
int otx2_destroy_ntuple_flows(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1137
int otx2_enable_rxvlan(struct otx2_nic *pf, bool enable);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1159
int otx2_dmacflt_get_max_cnt(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1160
int otx2_dmacflt_add(struct otx2_nic *pf, const u8 *mac, u32 bit_pos);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1161
int otx2_dmacflt_remove(struct otx2_nic *pf, const u8 *mac, u32 bit_pos);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1162
int otx2_dmacflt_update(struct otx2_nic *pf, u8 *mac, u32 bit_pos);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1163
void otx2_dmacflt_reinstall_flows(struct otx2_nic *pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1211
int rvu_event_up_notify(struct otx2_nic *pf, struct rep_event *info);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
287
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
298
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
303
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
455
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
10
static int otx2_dmacflt_do_add(struct otx2_nic *pf, const u8 *mac,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
115
static int otx2_dmacflt_remove_pfmac(struct otx2_nic *pf, u32 dmac_index)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
120
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
121
req = otx2_mbox_alloc_msg_cgx_mac_addr_reset(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
123
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
128
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
130
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
134
int otx2_dmacflt_remove(struct otx2_nic *pf, const u8 *mac,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
137
u32 dmacindex = pf->flow_cfg->bmap_to_dmacindex[bit_pos];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
139
if (ether_addr_equal(mac, pf->netdev->dev_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
140
return otx2_dmacflt_remove_pfmac(pf, dmacindex);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
142
return otx2_dmacflt_do_remove(pf, mac, dmacindex);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
149
int otx2_dmacflt_get_max_cnt(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
155
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
156
msg = otx2_mbox_alloc_msg_cgx_mac_max_entries_get(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
159
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
163
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
168
otx2_mbox_get_rsp(&pf->mbox.mbox, 0, &msg->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
17
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
175
pf->flow_cfg->dmacflt_max_flows = rsp->max_dmac_filters;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
178
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
182
int otx2_dmacflt_update(struct otx2_nic *pf, u8 *mac, u32 bit_pos)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
188
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
19
req = otx2_mbox_alloc_msg_cgx_mac_addr_add(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
190
req = otx2_mbox_alloc_msg_cgx_mac_addr_update(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
193
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
198
req->index = pf->flow_cfg->bmap_to_dmacindex[bit_pos];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
202
rc = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
207
otx2_mbox_get_rsp(&pf->mbox.mbox, 0, &req->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
21
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
213
pf->flow_cfg->bmap_to_dmacindex[bit_pos] = rsp->index;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
216
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
26
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
30
otx2_mbox_get_rsp(&pf->mbox.mbox, 0, &req->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
32
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
39
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
43
static int otx2_dmacflt_add_pfmac(struct otx2_nic *pf, u32 *dmac_index)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
49
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
51
req = otx2_mbox_alloc_msg_cgx_mac_addr_set(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
53
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
59
ether_addr_copy(req->mac_addr, pf->netdev->dev_addr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
60
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
66
otx2_mbox_get_rsp(&pf->mbox.mbox, 0, &req->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
75
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
79
int otx2_dmacflt_add(struct otx2_nic *pf, const u8 *mac, u32 bit_pos)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
86
dmacindex = &pf->flow_cfg->bmap_to_dmacindex[bit_pos];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
88
if (ether_addr_equal(mac, pf->netdev->dev_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
89
return otx2_dmacflt_add_pfmac(pf, dmacindex);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dmac_flt.c
91
return otx2_dmacflt_do_add(pf, mac, dmacindex);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
1240
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
1243
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
1264
req = otx2_mbox_alloc_msg_cgx_set_link_mode(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
1291
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1459
int otx2_enable_rxvlan(struct otx2_nic *pf, bool enable)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1466
if (!(pf->flags & OTX2_FLAG_RX_VLAN_SUPPORT))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1470
err = otx2_install_rxvlan_offload_flow(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1474
err = otx2_delete_rxvlan_offload_flow(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1479
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1480
req = otx2_mbox_alloc_msg_nix_vtag_cfg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1482
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1493
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1495
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1499
rsp_hdr = otx2_mbox_get_rsp(&pf->mbox.mbox, 0, &req->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1501
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1505
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1509
void otx2_dmacflt_reinstall_flows(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1514
list_for_each_entry(iter, &pf->flow_cfg->flow_list, list) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1517
otx2_dmacflt_add(pf, eth_hdr->h_dest,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
291
int otx2_mcam_flow_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
295
pf->flow_cfg = devm_kzalloc(pf->dev, sizeof(struct otx2_flow_config),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
297
if (!pf->flow_cfg)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
300
pf->flow_cfg->dmacflt_bmap = devm_kcalloc(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
303
if (!pf->flow_cfg->dmacflt_bmap)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
306
INIT_LIST_HEAD(&pf->flow_cfg->flow_list);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
307
INIT_LIST_HEAD(&pf->flow_cfg->flow_list_tc);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
309
pf->flow_cfg->ucast_flt_cnt = OTX2_DEFAULT_UNICAST_FLOWS;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
310
pf->flow_cfg->ntuple_cnt = OTX2_DEFAULT_FLOWCOUNT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
315
err = otx2_mcam_entry_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
320
if (!(pf->flags & OTX2_FLAG_UCAST_FLTR_SUPPORT))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
323
pf->mac_table = devm_kzalloc(pf->dev, sizeof(struct otx2_mac_table)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
324
* pf->flow_cfg->ucast_flt_cnt, GFP_KERNEL);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
325
if (!pf->mac_table)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
328
otx2_dmacflt_get_max_cnt(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
331
if (!pf->flow_cfg->dmacflt_max_flows)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
334
pf->flow_cfg->bmap_to_dmacindex =
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
335
devm_kzalloc(pf->dev, sizeof(u32) *
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
336
pf->flow_cfg->dmacflt_max_flows,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
339
if (!pf->flow_cfg->bmap_to_dmacindex)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
342
pf->flags |= OTX2_FLAG_DMACFLTR_SUPPORT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
347
void otx2_mcam_flow_del(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
349
otx2_destroy_mcam_flows(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
356
static int otx2_do_add_macfilter(struct otx2_nic *pf, const u8 *mac)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
358
struct otx2_flow_config *flow_cfg = pf->flow_cfg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
362
if (!(pf->flags & OTX2_FLAG_UCAST_FLTR_SUPPORT))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
366
if (netdev_uc_count(pf->netdev) > pf->flow_cfg->ucast_flt_cnt)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
369
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
370
req = otx2_mbox_alloc_msg_npc_install_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
372
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
377
for (i = 0; i < pf->flow_cfg->ucast_flt_cnt; i++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
378
if (pf->mac_table[i].inuse)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
380
ether_addr_copy(pf->mac_table[i].addr, mac);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
381
pf->mac_table[i].inuse = true;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
382
pf->mac_table[i].mcam_entry =
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
384
req->entry = pf->mac_table[i].mcam_entry;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
391
req->channel = pf->hw.rx_chan_base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
396
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
397
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
404
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
406
if (!bitmap_empty(pf->flow_cfg->dmacflt_bmap,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
407
pf->flow_cfg->dmacflt_max_flows))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
412
return otx2_do_add_macfilter(pf, mac);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
415
static bool otx2_get_mcamentry_for_mac(struct otx2_nic *pf, const u8 *mac,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
420
for (i = 0; i < pf->flow_cfg->ucast_flt_cnt; i++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
421
if (!pf->mac_table[i].inuse)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
424
if (ether_addr_equal(pf->mac_table[i].addr, mac)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
425
*mcam_entry = pf->mac_table[i].mcam_entry;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
426
pf->mac_table[i].inuse = false;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
435
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
440
if (!otx2_get_mcamentry_for_mac(pf, mac, &mcam_entry))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
443
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
444
req = otx2_mbox_alloc_msg_npc_delete_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
446
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
451
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
452
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
101
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1CX(1), INTR_MASK(vfs - 64));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1010
struct otx2_nic *pf = af_mbox->pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
102
irq = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFFLR1);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1021
trace_otx2_msg_status(pf->pdev, "PF-AF up queue handler(notification)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
103
free_irq(irq, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1030
otx2_process_mbox_msg_up(pf, msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1034
if (devid && pci_num_vf(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1035
otx2_forward_vf_mbox_msgs(pf, &pf->mbox.mbox_up,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1046
struct otx2_nic *pf = (struct otx2_nic *)pf_irq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1047
struct mbox *mw = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1054
otx2_write64(pf, RVU_PF_INT, BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1057
mbox_data = otx2_read64(pf, RVU_PF_PFAF_MBOX0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
106
static void otx2_flr_wq_destroy(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1061
otx2_write64(pf, RVU_PF_PFAF_MBOX0, mbox_data);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1069
queue_work(pf->mbox_wq, &mw->mbox_up_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1071
trace_otx2_msg_interrupt(pf->pdev, "UP message from AF to PF",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1074
trace_otx2_msg_status(pf->pdev, "PF-AF up work queued(interrupt)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
108
if (!pf->flr_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1080
otx2_write64(pf, RVU_PF_PFAF_MBOX0, mbox_data);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1088
queue_work(pf->mbox_wq, &mw->mbox_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1090
trace_otx2_msg_interrupt(pf->pdev, "DOWN reply from AF to PF",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1093
trace_otx2_msg_status(pf->pdev, "PF-AF down work queued(interrupt)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
110
destroy_workqueue(pf->flr_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1100
void otx2_disable_mbox_intr(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1105
if (!is_cn20k(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1106
vector = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_AFPF_MBOX);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1107
otx2_write64(pf, RVU_PF_INT_ENA_W1C, BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1109
vector = pci_irq_vector(pf->pdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
111
pf->flr_wq = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1111
otx2_write64(pf, RVU_PF_INT_ENA_W1C,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1114
free_irq(vector, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1118
int otx2_register_mbox_intr(struct otx2_nic *pf, bool probe_af)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
112
devm_kfree(pf->dev, pf->flr_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1120
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1126
if (!is_cn20k(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1129
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1131
(pf->pdev, RVU_PF_INT_VEC_AFPF_MBOX),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1132
pf->hw_ops->pfaf_mbox_intr_handler,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1133
0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1138
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1140
(pf->pdev, RVU_MBOX_PF_INT_VEC_AFPF_MBOX),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1141
pf->hw_ops->pfaf_mbox_intr_handler,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1142
0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1145
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1153
if (!is_cn20k(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1154
otx2_write64(pf, RVU_PF_INT, BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1155
otx2_write64(pf, RVU_PF_INT_ENA_W1S, BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1157
otx2_write64(pf, RVU_PF_INT, BIT_ULL(0) | BIT_ULL(1));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1158
otx2_write64(pf, RVU_PF_INT_ENA_W1S, BIT_ULL(0) |
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1166
req = otx2_mbox_alloc_msg_ready(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1168
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1171
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1173
dev_warn(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1175
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
118
struct otx2_nic *pf = flrwork->pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1182
void otx2_pfaf_mbox_destroy(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1184
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1186
if (pf->mbox_wq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1187
destroy_workqueue(pf->mbox_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1188
pf->mbox_wq = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
119
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1191
if (mbox->mbox.hwbase && !is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1199
int otx2_pfaf_mbox_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1201
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1205
mbox->pfvf = pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1206
pf->mbox_wq = alloc_ordered_workqueue("otx2_pfaf_mailbox",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1208
if (!pf->mbox_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1215
if (is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1216
hwbase = pf->reg_base + RVU_PFX_FUNC_PFAF_MBOX +
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1224
(pf->pdev, PCI_MBOX_BAR_NUM), MBOX_SIZE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1226
dev_err(pf->dev, "Unable to map PFAF mailbox region\n");
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
123
vf = flrwork - pf->flr_wrk;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1231
err = otx2_mbox_init(&mbox->mbox, hwbase, pf->pdev, pf->reg_base,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1236
err = otx2_mbox_init(&mbox->mbox_up, hwbase, pf->pdev, pf->reg_base,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1241
err = otx2_mbox_bbuf_init(mbox, pf->pdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1251
otx2_pfaf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1255
static int otx2_cgx_config_linkevents(struct otx2_nic *pf, bool enable)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1260
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1262
msg = otx2_mbox_alloc_msg_cgx_start_linkevents(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1264
msg = otx2_mbox_alloc_msg_cgx_stop_linkevents(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1267
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1271
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1272
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1293
static int otx2_cgx_config_loopback(struct otx2_nic *pf, bool enable)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1298
if (enable && !bitmap_empty(pf->flow_cfg->dmacflt_bmap,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1299
pf->flow_cfg->dmacflt_max_flows))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1300
netdev_warn(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1303
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1305
msg = otx2_mbox_alloc_msg_cgx_intlbk_enable(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1307
msg = otx2_mbox_alloc_msg_cgx_intlbk_disable(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1310
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1314
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1315
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
134
if (!otx2_sync_mbox_msg(&pf->mbox)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1395
struct otx2_nic *pf = data;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
140
otx2_write64(pf, RVU_PF_VFTRPENDX(reg), BIT_ULL(vf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1401
for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1402
ptr = otx2_get_regaddr(pf, NIX_LF_CQ_OP_INT);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1405
otx2_write64(pf, NIX_LF_CQ_OP_INT, (qidx << 44) |
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
141
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1SX(reg), BIT_ULL(vf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1411
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1413
qidx, otx2_read64(pf, NIX_LF_ERR_INT));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1416
netdev_err(pf->netdev, "CQ%lld: Doorbell error",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1419
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1424
schedule_work(&pf->reset_task);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1428
for (qidx = 0; qidx < otx2_get_total_tx_queues(pf); qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1432
sq = &pf->qset.sq[qidx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1441
ptr = otx2_get_regaddr(pf, NIX_LF_SQ_OP_INT);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1443
otx2_write64(pf, NIX_LF_SQ_OP_INT, (qidx << 44) |
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1447
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1449
qidx, otx2_read64(pf, NIX_LF_ERR_INT));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1453
sq_op_err_dbg = otx2_read64(pf, NIX_LF_SQ_OP_ERR_DBG);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1458
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1464
otx2_write64(pf, NIX_LF_SQ_OP_ERR_DBG, BIT_ULL(44));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1474
mnq_err_dbg = otx2_read64(pf, NIX_LF_MNQ_ERR_DBG);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1479
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1483
otx2_write64(pf, NIX_LF_MNQ_ERR_DBG, BIT_ULL(44));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1486
snd_err_dbg = otx2_read64(pf, NIX_LF_SEND_ERR_DBG);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1489
netdev_err(pf->netdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
149
struct otx2_nic *pf = (struct otx2_nic *)pf_irq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1494
otx2_write64(pf, NIX_LF_SEND_ERR_DBG, BIT_ULL(44));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1500
netdev_err(pf->netdev, "SQ%lld: SQB allocation failed",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1503
schedule_work(&pf->reset_task);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1512
struct otx2_nic *pf = (struct otx2_nic *)cq_poll->dev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1520
otx2_write64(pf, NIX_LF_CINTX_ENA_W1C(qidx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1523
pf->napi_events++;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
153
if (pf->total_vfs > 64)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1530
void otx2_disable_napi(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1532
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1537
for (qidx = 0; qidx < pf->hw.cint_cnt; qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1548
static void otx2_free_cq_res(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1550
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1555
otx2_ctx_disable(&pf->mbox, NIX_AQ_CTYPE_CQ, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1558
qmem_free(pf->dev, cq->cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1562
static void otx2_free_sq_res(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1564
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1569
otx2_ctx_disable(&pf->mbox, NIX_AQ_CTYPE_SQ, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
157
intr = otx2_read64(pf, RVU_PF_VFFLR_INTX(reg));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1571
otx2_sq_free_sqbs(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1572
for (qidx = 0; qidx < otx2_get_total_tx_queues(pf); qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1577
qmem_free(pf->dev, sq->sqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1578
qmem_free(pf->dev, sq->sqe_ring);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1579
qmem_free(pf->dev, sq->cpt_resp);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1580
qmem_free(pf->dev, sq->tso_hdrs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1586
static int otx2_get_rbuf_size(struct otx2_nic *pf, int mtu)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1592
if (pf->hw.rbuf_len)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1593
return ALIGN(pf->hw.rbuf_len, OTX2_ALIGN) + OTX2_HEAD_ROOM;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1614
int otx2_init_hw_resources(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1617
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1618
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1627
hw->sqpool_cnt = otx2_get_total_tx_queues(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1630
if (!otx2_rep_dev(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1632
pf->tx_max_pktlen = pf->netdev->max_mtu + OTX2_ETH_HLEN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1633
pf->rbsize = otx2_get_rbuf_size(pf, pf->netdev->mtu);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1638
err = otx2_config_npa(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1643
err = otx2_config_nix(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1648
otx2_nix_cpt_config_bp(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
165
queue_work(pf->flr_wq, &pf->flr_wrk[dev].work);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1651
if (!is_otx2_lbkvf(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1652
otx2_nix_config_bp(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1655
err = otx2_rq_aura_pool_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1661
err = otx2_sq_aura_pool_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1667
err = otx2_txsch_alloc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
167
otx2_write64(pf, RVU_PF_VFFLR_INTX(reg), BIT_ULL(vf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1674
if (pf->pfc_en) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1675
err = otx2_pfc_txschq_alloc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1683
err = otx2_config_nix_queues(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
169
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1CX(reg),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1692
for (idx = 0; idx < pf->hw.txschq_cnt[lvl]; idx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1693
err = otx2_txschq_config(pf, lvl, idx, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1695
dev_err(pf->dev, "Failed to config TXSCH\n");
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1703
if (pf->pfc_en) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1704
err = otx2_pfc_txschq_config(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1716
otx2_free_sq_res(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1717
otx2_free_cq_res(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1720
otx2_txschq_stop(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1722
otx2_sq_free_sqbs(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1724
otx2_free_aura_ptr(pf, AURA_NIX_RQ);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1727
otx2_aura_pool_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1734
dev_err(pf->dev, "%s failed to free nixlf\n", __func__);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1741
dev_err(pf->dev, "%s failed to free npalf\n", __func__);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1749
void otx2_free_hw_resources(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1751
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1753
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1759
otx2_sqb_flush(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1762
otx2_txschq_stop(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1765
if (pf->pfc_en)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1766
otx2_pfc_txschq_stop(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1769
if (!otx2_rep_dev(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1770
otx2_clean_qos_queues(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1774
if (!is_otx2_lbkvf(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1775
otx2_nix_config_bp(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
178
struct otx2_nic *pf = (struct otx2_nic *)pf_irq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1785
otx2_cleanup_rx_cqes(pf, cq, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1787
otx2_cleanup_tx_cqes(pf, cq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1789
otx2_free_pending_sqe(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1791
otx2_free_sq_res(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1794
otx2_free_aura_ptr(pf, AURA_NIX_RQ);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1796
otx2_free_cq_res(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1799
if (!otx2_rep_dev(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1800
cn10k_free_all_ipolicers(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1807
if (!(pf->flags & OTX2_FLAG_PF_SHUTDOWN))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1810
dev_err(pf->dev, "%s failed to free nixlf\n", __func__);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1817
otx2_aura_pool_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
182
if (pf->total_vfs > 64)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1824
dev_err(pf->dev, "%s failed to free npalf\n", __func__);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1845
static void otx2_do_set_rx_mode(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1847
struct net_device *netdev = pf->netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1855
(netdev_uc_count(netdev) > pf->flow_cfg->ucast_flt_cnt)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
186
intr = otx2_read64(pf, RVU_PF_VFME_INTX(reg));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1863
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1864
req = otx2_mbox_alloc_msg_nix_set_rx_mode(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1866
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1877
if (otx2_promisc_use_mce_list(pf))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1880
otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1881
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1924
int otx2_alloc_queue_mem(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1926
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
193
otx2_write64(pf, RVU_PF_VFTRPENDX(reg), BIT_ULL(vf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1933
pf->hw.non_qos_queues = pf->hw.tx_queues + pf->hw.xdp_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1934
pf->hw.cint_cnt = max3(pf->hw.rx_queues, pf->hw.tx_queues,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1935
pf->hw.tc_tx_queues);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1937
pf->qset.cq_cnt = pf->hw.rx_queues + otx2_get_total_tx_queues(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1939
qset->napi = kzalloc_objs(*cq_poll, pf->hw.cint_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1948
qset->cq = kzalloc_objs(struct otx2_cq_queue, pf->qset.cq_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
195
otx2_write64(pf, RVU_PF_VFME_INTX(reg), BIT_ULL(vf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1953
otx2_get_total_tx_queues(pf));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1957
qset->rq = kzalloc_objs(struct otx2_rcv_queue, pf->hw.rx_queues);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1971
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1973
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1979
err = otx2_alloc_queue_mem(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1983
err = otx2_init_hw_resources(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1988
for (qidx = 0; qidx < pf->hw.cint_cnt; qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1997
(qidx < pf->hw.rx_queues) ? qidx : CINT_INVALID_CQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1998
cq_poll->cq_ids[CQ_TX] = (qidx < pf->hw.tx_queues) ?
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1999
qidx + pf->hw.rx_queues : CINT_INVALID_CQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2000
if (pf->xdp_prog)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2001
cq_poll->cq_ids[CQ_XDP] = (qidx < pf->hw.xdp_queues) ?
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2002
(qidx + pf->hw.rx_queues +
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2003
pf->hw.tx_queues) :
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2008
cq_poll->cq_ids[CQ_QOS] = (qidx < pf->hw.tc_tx_queues) ?
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2009
(qidx + pf->hw.rx_queues +
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
201
static int otx2_register_flr_me_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2010
pf->hw.non_qos_queues) :
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2013
cq_poll->dev = (void *)pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2021
err = otx2_hw_set_mtu(pf, netdev->mtu);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2026
otx2_setup_segmentation(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2029
err = otx2_rss_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
203
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2034
vec = pf->hw.nix_msixoff + NIX_LF_QINT_VEC_START;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2035
irq_name = &pf->hw.irq_name[vec * NAME_SIZE];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2037
snprintf(irq_name, NAME_SIZE, "%s-qerr", pf->netdev->name);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2039
err = request_irq(pci_irq_vector(pf->pdev, vec),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2040
otx2_q_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2042
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2044
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2049
otx2_write64(pf, NIX_LF_QINTX_ENA_W1S(0), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2052
vec = pf->hw.nix_msixoff + NIX_LF_CINT_VEC_START;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2053
for (qidx = 0; qidx < pf->hw.cint_cnt; qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2054
irq_name = &pf->hw.irq_name[vec * NAME_SIZE];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2058
pf->netdev->name, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2060
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2062
rvu_get_pf(pf->pdev, pf->pcifunc), qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2067
err = request_irq(pci_irq_vector(pf->pdev, vec),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2071
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2073
rvu_get_pf(pf->pdev, pf->pcifunc), qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2078
otx2_config_irq_coalescing(pf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2081
otx2_write64(pf, NIX_LF_CINTX_INT(qidx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2082
otx2_write64(pf, NIX_LF_CINTX_ENA_W1S(qidx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2085
otx2_set_cints_affinity(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2087
if (pf->flags & OTX2_FLAG_RX_VLAN_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2088
otx2_enable_rxvlan(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2091
if (pf->flags & OTX2_FLAG_TX_TSTAMP_ENABLED) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2092
pf->flags &= ~OTX2_FLAG_TX_TSTAMP_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2093
otx2_config_hw_tx_tstamp(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2095
if (pf->flags & OTX2_FLAG_RX_TSTAMP_ENABLED) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2096
pf->flags &= ~OTX2_FLAG_RX_TSTAMP_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2097
otx2_config_hw_rx_tstamp(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
210
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2100
pf->flags &= ~OTX2_FLAG_INTF_DOWN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2101
pf->flags &= ~OTX2_FLAG_PORT_UP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2106
otx2_qos_config_txschq(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2109
if (pf->linfo.link_up && !(pf->pcifunc & RVU_PFVF_FUNC_MASK))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
211
ret = request_irq(pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFME0),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2110
otx2_handle_link_event(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2113
if (pf->flags & OTX2_FLAG_DMACFLTR_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2114
otx2_dmacflt_reinstall_flows(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2116
otx2_tc_apply_ingress_police_rules(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2118
err = otx2_rxtx_enable(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
212
otx2_pf_me_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2129
otx2_do_set_rx_mode(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2134
otx2_rxtx_enable(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2138
pf->flags |= OTX2_FLAG_INTF_DOWN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
214
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2140
otx2_free_cints(pf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2141
vec = pci_irq_vector(pf->pdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2142
pf->hw.nix_msixoff + NIX_LF_QINT_VEC_START);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2143
otx2_write64(pf, NIX_LF_QINTX_ENA_W1C(0), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2144
free_irq(vec, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2146
otx2_disable_napi(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2147
otx2_free_hw_resources(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2156
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2158
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2162
if (pf->flags & OTX2_FLAG_INTF_DOWN)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2168
pf->flags |= OTX2_FLAG_INTF_DOWN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2173
otx2_rxtx_enable(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2176
pf->hw.rss_info.enable = false;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2179
vec = pci_irq_vector(pf->pdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2180
pf->hw.nix_msixoff + NIX_LF_QINT_VEC_START);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2181
otx2_write64(pf, NIX_LF_QINTX_ENA_W1C(0), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2182
free_irq(vec, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2185
vec = pf->hw.nix_msixoff + NIX_LF_CINT_VEC_START;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2186
for (qidx = 0; qidx < pf->hw.cint_cnt; qidx++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2188
otx2_write64(pf, NIX_LF_CINTX_ENA_W1C(qidx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2190
synchronize_irq(pci_irq_vector(pf->pdev, vec));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2199
for (wrk = 0; wrk < pf->qset.cq_cnt; wrk++)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2200
cancel_delayed_work_sync(&pf->refill_wrk[wrk].pool_refill_work);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2201
devm_kfree(pf->dev, pf->refill_wrk);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2203
otx2_free_hw_resources(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2204
otx2_free_cints(pf, pf->hw.cint_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2205
otx2_disable_napi(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
221
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2219
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
222
ret = request_irq(pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFFLR0),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2229
sq_idx = (qidx >= pf->hw.tx_queues) ? (qidx + pf->hw.xdp_queues) : qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
223
otx2_pf_flr_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2233
(!skb_shinfo(skb)->gso_size && skb->len > pf->tx_max_pktlen)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2234
dev_stats = &pf->hw.dev_stats;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2240
sq = &pf->qset.sq[sq_idx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2243
if (!otx2_sq_append_skb(pf, txq, sq, skb, qidx)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
225
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2258
static int otx2_qos_select_htb_queue(struct otx2_nic *pf, struct sk_buff *skb,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2266
classid = READ_ONCE(pf->qos.defcls);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2271
return otx2_get_txq_by_classid(pf, classid);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2277
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2284
qos_enabled = netdev->real_num_tx_queues > pf->hw.tx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2289
u16 htb_maj_id = smp_load_acquire(&pf->qos.maj_id);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2292
txq = otx2_qos_select_htb_queue(pf, skb, htb_maj_id);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2305
if ((vlan_prio > pf->hw.tx_queues - 1) ||
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2306
!pf->pfc_alloc_status[vlan_prio])
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2315
return txq % pf->hw.tx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
233
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2334
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2336
queue_work(pf->otx2_wq, &pf->rx_mode_work);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2341
struct otx2_nic *pf = container_of(work, struct otx2_nic, rx_mode_work);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2343
otx2_do_set_rx_mode(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
235
(pf->pdev, RVU_PF_INT_VEC_VFME1),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2350
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2353
return otx2_cgx_config_loopback(pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2357
return otx2_enable_rxvlan(pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
236
otx2_pf_me_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2369
struct otx2_nic *pf = container_of(work, struct otx2_nic, reset_task);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2371
if (!netif_running(pf->netdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2375
otx2_stop(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2376
pf->reset_count++;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2377
otx2_open(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2378
netif_trans_update(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
238
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
243
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
245
(pf->pdev, RVU_PF_INT_VEC_VFFLR1),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
246
otx2_pf_flr_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
248
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2521
static int otx2_do_set_vf_mac(struct otx2_nic *pf, int vf, const u8 *mac)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2526
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2527
req = otx2_mbox_alloc_msg_npc_install_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2536
req->channel = pf->hw.rx_chan_base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2543
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2545
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
255
otx2_write64(pf, RVU_PF_VFME_INTX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2551
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2552
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2559
if (vf >= pf->total_vfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
256
otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1SX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2565
config = &pf->vf_configs[vf];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2568
ret = otx2_do_set_vf_mac(pf, vf, mac);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2576
static int otx2_do_set_vf_vlan(struct otx2_nic *pf, int vf, u16 vlan, u8 qos,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2579
struct otx2_flow_config *flow_cfg = pf->flow_cfg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2588
config = &pf->vf_configs[vf];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
259
otx2_write64(pf, RVU_PF_VFFLR_INTX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2593
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2597
vtag_req = otx2_mbox_alloc_msg_nix_vtag_cfg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
260
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1SX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2606
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2613
del_req = otx2_mbox_alloc_msg_npc_delete_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2621
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2626
del_req = otx2_mbox_alloc_msg_npc_delete_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2634
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2640
req = otx2_mbox_alloc_msg_npc_install_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
265
otx2_write64(pf, RVU_PF_VFME_INTX(1), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2653
req->channel = pf->hw.rx_chan_base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
266
otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1SX(1),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2661
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2666
vtag_req = otx2_mbox_alloc_msg_nix_vtag_cfg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2678
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2683
(&pf->mbox.mbox, 0, &vtag_req->hdr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
269
otx2_write64(pf, RVU_PF_VFFLR_INTX(1), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2690
req = otx2_mbox_alloc_msg_npc_install_flow(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
270
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1SX(1),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2700
req->channel = pf->hw.tx_chan_base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2708
err = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2711
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2718
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2719
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2734
if (!(pf->flags & OTX2_FLAG_VF_VLAN_SUPPORT))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2737
return otx2_do_set_vf_vlan(pf, vf, vlan, qos, proto);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2743
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2744
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2753
config = &pf->vf_configs[vf];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
276
static int otx2_pf_flr_init(struct otx2_nic *pf, int num_vfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2762
static int otx2_xdp_xmit_tx(struct otx2_nic *pf, struct xdp_frame *xdpf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2768
dma_addr = otx2_dma_map_page(pf, virt_to_page(xdpf->data),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2771
if (dma_mapping_error(pf->dev, dma_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2774
err = otx2_xdp_sq_append_pkt(pf, xdpf, dma_addr, xdpf->len,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2777
otx2_dma_unmap_page(pf, dma_addr, xdpf->len, DMA_TO_DEVICE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2787
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2795
qidx += pf->hw.tx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2796
sq = pf->xdp_prog ? &pf->qset.sq[qidx] : NULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
280
pf->flr_wq = alloc_ordered_workqueue("otx2_pf_flr_wq", WQ_HIGHPRI);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2809
err = otx2_xdp_xmit_tx(pf, xdpf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
281
if (!pf->flr_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2816
static int otx2_xdp_setup(struct otx2_nic *pf, struct bpf_prog *prog)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2818
struct net_device *dev = pf->netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2819
bool if_up = netif_running(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2828
otx2_stop(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2830
old_prog = xchg(&pf->xdp_prog, prog);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2835
if (pf->xdp_prog)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2836
bpf_prog_add(pf->xdp_prog, pf->hw.rx_queues - 1);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
284
pf->flr_wrk = devm_kcalloc(pf->dev, num_vfs,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2841
if (pf->xdp_prog) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2842
pf->hw.xdp_queues = pf->hw.rx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2845
pf->hw.xdp_queues = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2850
otx2_open(pf->netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2857
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
286
if (!pf->flr_wrk) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2861
return otx2_xdp_setup(pf, xdp->prog);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2863
return otx2_xsk_pool_setup(pf, xdp->xsk.pool, xdp->xsk.queue_id);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2869
static int otx2_set_vf_permissions(struct otx2_nic *pf, int vf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
287
destroy_workqueue(pf->flr_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2875
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2876
req = otx2_mbox_alloc_msg_set_vf_perm(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2886
if (pf->vf_configs[vf].trusted)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2891
rc = otx2_sync_mbox_msg(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2893
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2900
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2901
struct pci_dev *pdev = pf->pdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2907
if (pf->vf_configs[vf].trusted == enable)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2910
pf->vf_configs[vf].trusted = enable;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2911
rc = otx2_set_vf_permissions(pf, vf, OTX2_TRUSTED_VF);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2914
pf->vf_configs[vf].trusted = !enable;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2916
netdev_info(pf->netdev, "VF %d is %strusted\n",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
292
pf->flr_wrk[vf].pf = pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
293
INIT_WORK(&pf->flr_wrk[vf].work, otx2_flr_handler);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2948
int otx2_wq_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2950
pf->otx2_wq = create_singlethread_workqueue("otx2_wq");
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2951
if (!pf->otx2_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2954
INIT_WORK(&pf->rx_mode_work, otx2_rx_mode_wrk_handler);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2955
INIT_WORK(&pf->reset_task, otx2_reset_task);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2977
int otx2_realloc_msix_vectors(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2979
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2988
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2992
dev_err(pf->dev, "%s: Failed to realloc %d IRQ vectors\n",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2997
return otx2_register_mbox_intr(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3001
static int otx2_sriov_vfcfg_init(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3005
pf->vf_configs = devm_kcalloc(pf->dev, pf->total_vfs,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3008
if (!pf->vf_configs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3011
for (i = 0; i < pf->total_vfs; i++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3012
pf->vf_configs[i].pf = pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3013
pf->vf_configs[i].intf_down = true;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3014
pf->vf_configs[i].trusted = false;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3015
INIT_DELAYED_WORK(&pf->vf_configs[i].link_event_work,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3022
static void otx2_sriov_vfcfg_cleanup(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3026
if (!pf->vf_configs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3029
for (i = 0; i < pf->total_vfs; i++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3030
cancel_delayed_work_sync(&pf->vf_configs[i].link_event_work);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3031
otx2_set_vf_permissions(pf, i, OTX2_RESET_VF_PERM);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3035
int otx2_init_rsrc(struct pci_dev *pdev, struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3038
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3053
pf->reg_base = pcim_iomap(pdev, PCI_CFG_REG_BAR_NUM, 0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3054
if (!pf->reg_base) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3059
err = otx2_check_pf_usable(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3063
if (!is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3076
otx2_setup_dev_hw_settings(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3078
if (is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3079
cn20k_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3081
otx2_init_hw_ops(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3084
err = otx2_pfaf_mbox_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3089
err = otx2_register_mbox_intr(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3096
err = otx2_attach_npa_nix(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3100
err = otx2_realloc_msix_vectors(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3104
err = cn10k_lmtst_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3111
if (pf->hw.lmt_info)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3112
free_percpu(pf->hw.lmt_info);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3113
if (test_bit(CN10K_LMTST, &pf->hw.cap_flag))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3114
qmem_free(pf->dev, pf->dync_lmt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3115
otx2_detach_resources(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3117
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3119
otx2_pfaf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3132
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3159
netdev = alloc_etherdev_mqs(sizeof(*pf), qcount + qos_txqs, qcount);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3165
pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3166
pf->netdev = netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3167
pf->pdev = pdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3168
pf->dev = dev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3169
pf->total_vfs = pci_sriov_get_totalvfs(pdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3170
pf->flags |= OTX2_FLAG_INTF_DOWN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3172
hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3182
err = otx2_init_rsrc(pdev, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3194
otx2_ptp_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3207
pf->iommu_domain = iommu_get_domain_for_dev(dev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3215
err = otx2_mcam_flow_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3219
otx2_set_hw_capabilities(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3221
err = cn10k_mcs_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3225
if (pf->flags & OTX2_FLAG_NTUPLE_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3228
if (pf->flags & OTX2_FLAG_UCAST_FLTR_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3235
if (pf->flags & OTX2_FLAG_RX_VLAN_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3241
if (pf->flags & OTX2_FLAG_TC_FLOWER_SUPPORT)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3255
netdev->max_mtu = otx2_get_max_mtu(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3259
otx2_reset_mac_stats(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3271
err = otx2_wq_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3277
err = otx2_init_tc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3281
err = otx2_register_dl(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3286
err = otx2_sriov_vfcfg_init(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3291
otx2_cgx_config_linkevents(pf, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3293
pf->af_xdp_zc_qidx = bitmap_zalloc(qcount, GFP_KERNEL);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3294
if (!pf->af_xdp_zc_qidx) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3305
otx2_qos_init(pf, qos_txqs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3311
bitmap_free(pf->af_xdp_zc_qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3314
otx2_sriov_vfcfg_cleanup(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3316
otx2_unregister_dl(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3317
otx2_shutdown_tc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3319
otx2_mcam_flow_del(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3323
cn10k_ipsec_clean(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3325
cn10k_mcs_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3327
otx2_mcam_flow_del(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3329
otx2_ptp_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3331
if (pf->hw.lmt_info)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3332
free_percpu(pf->hw.lmt_info);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3333
if (test_bit(CN10K_LMTST, &pf->hw.cap_flag))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3334
qmem_free(pf->dev, pf->dync_lmt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3335
otx2_detach_resources(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3336
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3337
otx2_pfaf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3351
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3356
vf_idx = config - config->pf->vf_configs;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3357
pf = config->pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3362
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3366
if (!otx2_mbox_wait_for_zero(&pf->mbox_pfvf[0].mbox_up, vf_idx)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3368
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3372
msghdr = otx2_mbox_alloc_msg_rsp(&pf->mbox_pfvf[0].mbox_up, vf_idx,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3375
dev_err(pf->dev, "Failed to create VF%d link event\n", vf_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3376
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3383
req->hdr.pcifunc = pf->pcifunc;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3384
memcpy(&req->link_info, &pf->linfo, sizeof(req->link_info));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3386
otx2_mbox_wait_for_zero(&pf->mbox_pfvf[0].mbox_up, vf_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3388
otx2_sync_mbox_up_msg(&pf->mbox_pfvf[0], vf_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3390
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3396
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3400
ret = otx2_pfvf_mbox_init(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3404
ret = otx2_register_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3408
ret = otx2_pf_flr_init(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3412
ret = otx2_register_flr_me_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3422
otx2_disable_flr_me_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3424
otx2_flr_wq_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3426
otx2_disable_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3428
otx2_pfvf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3435
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3443
otx2_disable_flr_me_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3444
otx2_flr_wq_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3445
otx2_disable_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3446
otx2_pfvf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3459
static void otx2_ndc_sync(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3461
struct mbox *mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3477
dev_err(pf->dev, "NDC sync operation failed\n");
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3485
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3490
pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3492
pf->flags |= OTX2_FLAG_PF_SHUTDOWN;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3494
if (pf->flags & OTX2_FLAG_TX_TSTAMP_ENABLED)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3495
otx2_config_hw_tx_tstamp(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3496
if (pf->flags & OTX2_FLAG_RX_TSTAMP_ENABLED)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3497
otx2_config_hw_rx_tstamp(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3500
if (pf->flags & OTX2_FLAG_RX_PAUSE_ENABLED ||
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3501
(pf->flags & OTX2_FLAG_TX_PAUSE_ENABLED)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3502
pf->flags &= ~OTX2_FLAG_RX_PAUSE_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3503
pf->flags &= ~OTX2_FLAG_TX_PAUSE_ENABLED;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3504
otx2_config_pause_frm(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3509
if (pf->pfc_en) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3510
pf->pfc_en = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3511
otx2_config_priority_flow_ctrl(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3514
cancel_work_sync(&pf->reset_task);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3516
otx2_cgx_config_linkevents(pf, false);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3518
otx2_unregister_dl(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3520
cn10k_ipsec_clean(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3521
cn10k_mcs_free(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3522
otx2_sriov_disable(pf->pdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3523
otx2_sriov_vfcfg_cleanup(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3524
if (pf->otx2_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3525
destroy_workqueue(pf->otx2_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3527
otx2_ptp_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3528
otx2_mcam_flow_del(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3529
otx2_shutdown_tc(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3530
otx2_shutdown_qos(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3531
otx2_ndc_sync(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3532
otx2_detach_resources(&pf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3533
if (pf->hw.lmt_info)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3534
free_percpu(pf->hw.lmt_info);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3535
if (test_bit(CN10K_LMTST, &pf->hw.cap_flag))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3536
qmem_free(pf->dev, pf->dync_lmt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3537
otx2_disable_mbox_intr(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3538
otx2_pfaf_mbox_destroy(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3539
pci_free_irq_vectors(pf->pdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3540
bitmap_free(pf->af_xdp_zc_qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
358
static int otx2_forward_vf_mbox_msgs(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
377
dst_mbox = &pf->mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
386
mutex_lock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
397
dev_warn(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
400
dst_mdev->mbase = pf->mbox.bbuf_base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
401
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
412
otx2_forward_msg_pfvf(dst_mdev, &pf->mbox_pfvf[0].mbox,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
413
pf->mbox.bbuf_base, vf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
414
mutex_unlock(&pf->mbox.lock);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
422
dst_mbox = &pf->mbox_pfvf[0];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
435
dev_warn(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
443
otx2_forward_msg_pfvf(&pf->mbox_pfvf->mbox_up.dev[vf],
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
444
&pf->mbox.mbox_up,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
445
pf->mbox_pfvf[vf].bbuf_base,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
459
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
462
pf = vf_mbox->pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
463
vf_idx = vf_mbox - pf->mbox_pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
465
mbox = &pf->mbox_pfvf[0].mbox;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
470
trace_otx2_msg_status(pf->pdev, "PF-VF down queue handler(forwarding)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
485
err = otx2_forward_vf_mbox_msgs(pf, mbox, MBOX_DIR_PFAF, vf_idx,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
499
struct otx2_nic *pf = vf_mbox->pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
505
vf_idx = vf_mbox - pf->mbox_pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
506
mbox = &pf->mbox_pfvf[0].mbox_up;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
511
trace_otx2_msg_status(pf->pdev, "PF-VF up queue handler(response)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
518
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
524
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
536
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
552
struct otx2_nic *pf = (struct otx2_nic *)(pf_irq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
553
int vfs = pf->total_vfs;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
557
mbox = pf->mbox_pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
560
intr = otx2_read64(pf, RVU_PF_VFPF_MBOX_INTX(1));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
561
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(1), intr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
562
otx2_queue_vf_work(mbox, pf->mbox_pfvf_wq, 64, vfs, intr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
568
intr = otx2_read64(pf, RVU_PF_VFPF_MBOX_INTX(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
569
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(0), intr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
571
otx2_queue_vf_work(mbox, pf->mbox_pfvf_wq, 0, vfs, intr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
579
static void *cn20k_pfvf_mbox_alloc(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
58
struct otx2_nic *pf = netdev_priv(netdev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
584
err = qmem_alloc(&pf->pdev->dev, &mbox_addr, numvfs, MBOX_SIZE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
586
dev_err(pf->dev, "qmem alloc fail\n");
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
590
otx2_write64(pf, RVU_PF_VF_MBOX_ADDR, (u64)mbox_addr->iova);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
591
pf->pfvf_mbox_addr = mbox_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
596
static int otx2_pfvf_mbox_init(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
606
pf->mbox_pfvf = devm_kcalloc(&pf->pdev->dev, numvfs,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
608
if (!pf->mbox_pfvf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
611
pf->mbox_pfvf_wq = alloc_workqueue("otx2_pfvf_mailbox",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
614
if (!pf->mbox_pfvf_wq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
62
if (pf->xdp_prog && new_mtu > MAX_XDP_MTU) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
621
if (is_cn20k(pf->pdev)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
622
hwbase = (void __iomem *)cn20k_pfvf_mbox_alloc(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
627
if (test_bit(CN10K_MBOX, &pf->hw.cap_flag))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
628
base = pci_resource_start(pf->pdev, PCI_MBOX_BAR_NUM) +
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
631
base = readq(pf->reg_base + RVU_PF_VF_BAR4_ADDR);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
633
hwbase = ioremap_wc(base, MBOX_SIZE * pf->total_vfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
640
mbox = &pf->mbox_pfvf[0];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
641
err = otx2_mbox_init(&mbox->mbox, hwbase, pf->pdev, pf->reg_base,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
646
err = otx2_mbox_init(&mbox->mbox_up, hwbase, pf->pdev, pf->reg_base,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
652
mbox->pfvf = pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
661
if (hwbase && !(is_cn20k(pf->pdev)))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
664
destroy_workqueue(pf->mbox_pfvf_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
668
static void otx2_pfvf_mbox_destroy(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
670
struct mbox *mbox = &pf->mbox_pfvf[0];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
675
if (pf->mbox_pfvf_wq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
676
destroy_workqueue(pf->mbox_pfvf_wq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
677
pf->mbox_pfvf_wq = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
680
if (mbox->mbox.hwbase && !is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
683
qmem_free(&pf->pdev->dev, pf->pfvf_mbox_addr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
688
static void otx2_enable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
691
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
692
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
695
otx2_write64(pf, RVU_PF_VFPF_MBOX_INT_ENA_W1SX(0), INTR_MASK(numvfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
698
otx2_write64(pf, RVU_PF_VFPF_MBOX_INT_ENA_W1SX(1),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
703
static void otx2_disable_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
707
if (is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
708
return cn20k_disable_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
711
otx2_write64(pf, RVU_PF_VFPF_MBOX_INT_ENA_W1CX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
712
otx2_write64(pf, RVU_PF_VFPF_MBOX_INT_ENA_W1CX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
714
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(0), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
715
vector = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFPF_MBOX0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
716
free_irq(vector, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
719
otx2_write64(pf, RVU_PF_VFPF_MBOX_INTX(1), ~0ull);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
720
vector = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFPF_MBOX1);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
721
free_irq(vector, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
725
static int otx2_register_pfvf_mbox_intr(struct otx2_nic *pf, int numvfs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
727
struct otx2_hw *hw = &pf->hw;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
731
if (is_cn20k(pf->pdev))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
732
return cn20k_register_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
736
if (pf->pcifunc)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
738
"RVUPF%d_VF Mbox0", rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
741
err = request_irq(pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFPF_MBOX0),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
742
otx2_pfvf_mbox_intr_handler, 0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
744
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
752
if (pf->pcifunc)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
755
rvu_get_pf(pf->pdev, pf->pcifunc));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
758
err = request_irq(pci_irq_vector(pf->pdev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
761
0, irq_name, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
763
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
769
otx2_enable_pfvf_mbox_intr(pf, numvfs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
774
static void otx2_process_pfaf_mbox_msg(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
780
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
786
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
795
struct otx2_vf_config *config = &pf->vf_configs[devid - 1];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
80
static void otx2_disable_flr_me_intr(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
814
pf->pcifunc = msg->pcifunc;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
817
mbox_handler_msix_offset(pf, (struct msix_offset_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
82
int irq, vfs = pf->total_vfs;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
820
mbox_handler_npa_lf_alloc(pf, (struct npa_lf_alloc_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
823
mbox_handler_nix_lf_alloc(pf, (struct nix_lf_alloc_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
826
mbox_handler_nix_bp_enable(pf, (struct nix_bp_cfg_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
829
mbox_handler_cgx_stats(pf, (struct cgx_stats_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
832
mbox_handler_cgx_fec_stats(pf, (struct cgx_fec_stats_rsp *)msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
836
dev_err(pf->dev,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
85
otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1CX(0), INTR_MASK(vfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
850
struct otx2_nic *pf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
86
irq = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFME0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
861
pf = af_mbox->pfvf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
863
trace_otx2_msg_status(pf->pdev, "PF-AF down queue handler(response)",
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
868
otx2_process_pfaf_mbox_msg(pf, msg);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
87
free_irq(irq, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
877
static void otx2_handle_link_event(struct otx2_nic *pf)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
879
struct cgx_link_user_info *linfo = &pf->linfo;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
880
struct net_device *netdev = pf->netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
882
if (pf->flags & OTX2_FLAG_PORT_UP)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
897
static int otx2_mbox_up_handler_rep_event_up_notify(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
90
otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1CX(0), INTR_MASK(vfs));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
901
struct net_device *netdev = pf->netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
91
irq = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFFLR0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
910
pf->flags |= OTX2_FLAG_PORT_UP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
914
pf->flags &= ~OTX2_FLAG_PORT_UP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
92
free_irq(irq, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
921
rvu_event_up_notify(pf, info);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
926
int otx2_mbox_up_handler_mcs_intr_notify(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
930
cn10k_handle_mcs_event(pf, event);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
935
int otx2_mbox_up_handler_cgx_link_event(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
942
pf->linfo = msg->link_info;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
945
for (i = 0; i < pci_num_vf(pf->pdev); i++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
946
struct otx2_vf_config *config = &pf->vf_configs[i];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
956
if (pf->flags & OTX2_FLAG_INTF_DOWN)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
959
otx2_handle_link_event(pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
963
static int otx2_process_mbox_msg_up(struct otx2_nic *pf,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
968
otx2_reply_invalid_msg(&pf->mbox.mbox_up, 0, 0, req->id);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
97
otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1CX(1), INTR_MASK(vfs - 64));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
979
&pf->mbox.mbox_up, 0, \
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
98
irq = pci_irq_vector(pf->pdev, RVU_PF_INT_VEC_VFME1);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
99
free_irq(irq, pf);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
990
pf, (struct _req_type *)req, rsp); \
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
999
otx2_reply_invalid_msg(&pf->mbox.mbox_up, 0, 0, req->id);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
119
int otx2_xsk_pool_enable(struct otx2_nic *pf, struct xsk_buff_pool *pool, u16 qidx)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
121
u16 rx_queues = pf->hw.rx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
122
u16 tx_queues = pf->hw.tx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
128
err = xsk_pool_dma_map(pool, pf->dev, DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
132
set_bit(qidx, pf->af_xdp_zc_qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
133
otx2_clean_up_rq(pf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
135
otx2_set_rss_table(pf, DEFAULT_RSS_CONTEXT_GROUP, NULL);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
137
return otx2_xsk_wakeup(pf->netdev, qidx, XDP_WAKEUP_RX);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
140
int otx2_xsk_pool_disable(struct otx2_nic *pf, u16 qidx)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
142
struct net_device *netdev = pf->netdev;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
150
sq = &pf->qset.sq[qidx + pf->hw.tx_queues];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
152
otx2_clean_up_rq(pf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
153
clear_bit(qidx, pf->af_xdp_zc_qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
156
otx2_set_rss_table(pf, DEFAULT_RSS_CONTEXT_GROUP, NULL);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
161
int otx2_xsk_pool_setup(struct otx2_nic *pf, struct xsk_buff_pool *pool, u16 qidx)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
164
return otx2_xsk_pool_enable(pf, pool, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
166
return otx2_xsk_pool_disable(pf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
171
struct otx2_nic *pf = netdev_priv(dev);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
173
struct otx2_qset *qset = &pf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
175
if (pf->flags & OTX2_FLAG_INTF_DOWN)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
178
if (queue_id >= pf->hw.rx_queues || queue_id >= pf->hw.tx_queues)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
187
otx2_write64(pf, NIX_LF_CINTX_ENA_W1S(cq_poll->cint_idx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
188
otx2_write64(pf, NIX_LF_CINTX_INT_W1S(cq_poll->cint_idx), BIT_ULL(0));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.h
14
int otx2_xsk_pool_setup(struct otx2_nic *pf, struct xsk_buff_pool *pool, u16 qid);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.h
15
int otx2_xsk_pool_enable(struct otx2_nic *pf, struct xsk_buff_pool *pool, u16 qid);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.h
16
int otx2_xsk_pool_disable(struct otx2_nic *pf, u16 qid);
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
250
attrs.pci_vf.pf = rvu_get_pf(priv->pdev, rep->pcifunc);
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
311
int rvu_event_up_notify(struct otx2_nic *pf, struct rep_event *info)
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
314
rvu_rep_state_evt_handler(pf, info);
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
419
struct otx2_nic *pf = rep->mdev;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
426
(!skb_shinfo(skb)->gso_size && skb->len > pf->tx_max_pktlen)) {
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
433
sq = &pf->qset.sq[rep->rep_id];
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
436
if (!otx2_sq_append_skb(pf, txq, sq, skb, rep->rep_id)) {
drivers/net/ethernet/marvell/octeontx2/nic/rep.h
54
int rvu_event_up_notify(struct otx2_nic *pf, struct rep_event *info);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
445
int pf;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
461
pf = mlx4_master_func_num(dev);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
463
priv->mfunc.master.res_tracker.res_alloc[RES_QP].quota[pf];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
465
priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[pf];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
467
priv->mfunc.master.res_tracker.res_alloc[RES_SRQ].quota[pf];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
469
priv->mfunc.master.res_tracker.res_alloc[RES_MTT].quota[pf];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
471
priv->mfunc.master.res_tracker.res_alloc[RES_MPT].quota[pf];
drivers/net/ethernet/mellanox/mlx5/core/lag/debugfs.c
148
seq_printf(file, "%s\n", dev_name(ldev->pf[i].dev->device));
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1023
pf_mdev = ldev->pf[pf_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1108
mdev = ldev->pf[pf_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1124
mdev = ldev->pf[pf_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1155
dev0 = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1185
mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1193
if (mlx5_get_roce_state(ldev->pf[i].dev))
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1194
mlx5_nic_vport_enable_roce(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1203
err = mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1214
mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1251
devcom = ldev->pf[i].dev->priv.hca_devcom_comp;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1320
if (ldev->pf[i].netdev == ndev_tmp) {
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1544
ldev->pf[fn].netdev = netdev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1558
if (ldev->pf[i].netdev == netdev) {
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1559
ldev->pf[i].netdev = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1571
ldev->pf[fn].dev = dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1574
MLX5_NB_INIT(&ldev->pf[fn].port_change_nb,
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1576
mlx5_eq_notifier_register(dev, &ldev->pf[fn].port_change_nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1585
if (ldev->pf[fn].dev != dev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1588
if (ldev->pf[fn].port_change_nb.nb.notifier_call)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1589
mlx5_eq_notifier_unregister(dev, &ldev->pf[fn].port_change_nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1591
ldev->pf[fn].dev = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1752
if (ldev->pf[i].dev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1762
if (ldev->pf[i].dev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1823
res = ldev && __mlx5_lag_is_active(ldev) && idx >= 0 && dev == ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1911
if (ldev->pf[i].netdev == slave) {
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1952
if (ldev->pf[idx].dev != dev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1961
peer_dev = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
1999
mdev[idx++] = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
244
if (ldev->pf[i].dev &&
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
245
ldev->pf[i].port_change_nb.nb.notifier_call) {
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
246
struct mlx5_nb *nb = &ldev->pf[i].port_change_nb;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
248
mlx5_eq_notifier_unregister(ldev->pf[i].dev, nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
315
if (ldev->pf[i].netdev == ndev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
358
if (ldev->pf[i].netdev)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
430
if (ldev->pf[i].has_drop)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
440
if (!ldev->pf[i].has_drop)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
443
mlx5_esw_acl_ingress_vport_drop_rule_destroy(ldev->pf[i].dev->priv.eswitch,
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
445
ldev->pf[i].has_drop = false;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
471
dev = ldev->pf[disabled_index].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
475
ldev->pf[disabled_index].has_drop = true;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
507
dev0 = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
536
ndev = ldev->pf[i].netdev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
541
ndev = ldev->pf[last_idx].netdev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
566
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
618
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
673
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
676
struct mlx5_eswitch *slave_esw = ldev->pf[i].dev->priv.eswitch;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
687
ldev->pf[j].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
705
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
752
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
808
dev0 = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
817
ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
860
dev = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
865
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
868
if (mlx5_eswitch_mode(ldev->pf[i].dev) != mode)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
873
if (mlx5_sriov_is_enabled(ldev->pf[i].dev))
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
876
roce_support = mlx5_get_roce_state(ldev->pf[first_idx].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
878
if (mlx5_get_roce_state(ldev->pf[i].dev) != roce_support)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
889
if (ldev->pf[i].dev->priv.flags &
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
893
ldev->pf[i].dev->priv.flags &= ~MLX5_PRIV_FLAGS_DISABLE_IB_ADEV;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
894
mlx5_rescan_drivers_locked(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
903
if (ldev->pf[i].dev->priv.flags &
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
907
ldev->pf[i].dev->priv.flags |= MLX5_PRIV_FLAGS_DISABLE_IB_ADEV;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
908
mlx5_rescan_drivers_locked(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
924
dev0 = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
935
mlx5_nic_vport_disable_roce(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
947
if (!(ldev->pf[i].dev->priv.flags & MLX5_PRIV_FLAGS_DISABLE_ALL_ADEV))
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
948
mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
961
dev = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
972
dev = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
989
roce_lag = roce_lag && !mlx5_sriov_is_enabled(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
993
roce_lag = roce_lag && is_mdev_legacy_mode(ldev->pf[i].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.h
67
struct lag_func pf[MLX5_MAX_PORTS];
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
149
return ldev->pf[ldev_idx].netdev;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
181
mlx5_core_dbg(ldev->pf[idx].dev,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
197
mlx5_core_warn(ldev->pf[idx].dev,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
206
if (ldev->pf[i].netdev == nh_dev0)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
243
if (ldev->pf[i].netdev == fib_nh->fib_nh_dev)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
32
return mlx5_esw_multipath_prereq(ldev->pf[idx0].dev,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
33
ldev->pf[idx1].dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
83
mlx5_core_warn(ldev->pf[idx0].dev,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
89
mlx5_notifier_call_chain(ldev->pf[idx0].dev->priv.events,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
94
mlx5_notifier_call_chain(ldev->pf[idx1].dev->priv.events,
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
108
err = mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
124
mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch);
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
19
dev = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
40
dev = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
56
dev = ldev->pf[i].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
85
dev0 = ldev->pf[idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
310
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
359
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
523
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
53
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
539
dev = ldev->pf[first_idx].dev;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
597
dest.vport.vhca_id = MLX5_CAP_GEN(ldev->pf[ports[idx] - 1].dev,
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
87
dest.vport.vhca_id = MLX5_CAP_GEN(ldev->pf[affinity - 1].dev,
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
235
struct pci_dev *pf = pci_physfn(vf);
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
240
dev = pci_get_drvdata(pf);
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
247
mlx5_get_default_msix_vec_count(dev, pci_num_vf(pf));
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
252
return nfp_mbox_cmd(abm->app->pf, NFP_MBOX_PCIE_ABM_ENABLE,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
258
return nfp_mbox_cmd(abm->app->pf, NFP_MBOX_PCIE_ABM_DISABLE,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
299
nfp_err(abm->app->pf->cpp, "vNIC mailbox too small for prio offload: %u, need: %u\n",
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
327
nfp_abm_ctrl_find_rtsym(struct nfp_pf *pf, const char *name, unsigned int size)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
331
sym = nfp_rtsym_lookup(pf->rtbl, name);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
333
nfp_err(pf->cpp, "Symbol '%s' not found\n", name);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
337
nfp_err(pf->cpp,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
356
return nfp_abm_ctrl_find_rtsym(abm->app->pf, pf_symbol, size);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
361
struct nfp_pf *pf = abm->app->pf;
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
365
abm->pf_id = nfp_cppcore_pcie_unit(pf->cpp);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
368
res = nfp_pf_rtsym_read_optional(pf, NFP_RED_SUPPORT_SYM_NAME, 1);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
374
res = nfp_pf_rtsym_read_optional(pf, NFP_NUM_BANDS_SYM_NAME, 1);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
379
res = nfp_pf_rtsym_read_optional(pf, NFP_NUM_PRIOS_SYM_NAME, 1);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
385
res = nfp_pf_rtsym_read_optional(pf, NFP_ACT_MASK_SYM_NAME,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
398
nfp_err(pf->cpp,
drivers/net/ethernet/netronome/nfp/abm/main.c
110
err = nfp_port_init_phy_port(app->pf, app, port, alink->id);
drivers/net/ethernet/netronome/nfp/abm/main.c
115
port->pf_split = app->pf->max_data_vnics > 1;
drivers/net/ethernet/netronome/nfp/abm/main.c
175
struct nfp_pf *pf = abm->app->pf;
drivers/net/ethernet/netronome/nfp/abm/main.c
178
list_for_each_entry(nn, &pf->vnics, vnic_list)
drivers/net/ethernet/netronome/nfp/abm/main.c
207
struct nfp_pf *pf = app->pf;
drivers/net/ethernet/netronome/nfp/abm/main.c
218
list_for_each_entry(nn, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/abm/main.c
257
nfp_abm_vnic_set_mac(struct nfp_pf *pf, struct nfp_abm *abm, struct nfp_net *nn,
drivers/net/ethernet/netronome/nfp/abm/main.c
260
struct nfp_eth_table_port *eth_port = &pf->eth_tbl->ports[id];
drivers/net/ethernet/netronome/nfp/abm/main.c
266
if (id > pf->eth_tbl->count) {
drivers/net/ethernet/netronome/nfp/abm/main.c
267
nfp_warn(pf->cpp, "No entry for persistent MAC address\n");
drivers/net/ethernet/netronome/nfp/abm/main.c
275
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/abm/main.c
277
nfp_warn(pf->cpp, "Failed to access the NSP for persistent MAC address: %ld\n",
drivers/net/ethernet/netronome/nfp/abm/main.c
284
nfp_warn(pf->cpp, "NSP doesn't support PF MAC generation\n");
drivers/net/ethernet/netronome/nfp/abm/main.c
293
nfp_warn(pf->cpp, "Reading persistent MAC address failed: %d\n",
drivers/net/ethernet/netronome/nfp/abm/main.c
302
nfp_warn(pf->cpp, "Can't parse persistent MAC address (%s)\n",
drivers/net/ethernet/netronome/nfp/abm/main.c
315
struct nfp_eth_table_port *eth_port = &app->pf->eth_tbl->ports[id];
drivers/net/ethernet/netronome/nfp/abm/main.c
350
nfp_abm_vnic_set_mac(app->pf, abm, nn, id);
drivers/net/ethernet/netronome/nfp/abm/main.c
445
struct nfp_pf *pf = app->pf;
drivers/net/ethernet/netronome/nfp/abm/main.c
450
if (!pf->eth_tbl) {
drivers/net/ethernet/netronome/nfp/abm/main.c
451
nfp_err(pf->cpp, "ABM NIC requires ETH table\n");
drivers/net/ethernet/netronome/nfp/abm/main.c
454
if (pf->max_data_vnics != pf->eth_tbl->count) {
drivers/net/ethernet/netronome/nfp/abm/main.c
455
nfp_err(pf->cpp, "ETH entries don't match vNICs (%d vs %d)\n",
drivers/net/ethernet/netronome/nfp/abm/main.c
456
pf->max_data_vnics, pf->eth_tbl->count);
drivers/net/ethernet/netronome/nfp/abm/main.c
459
if (!pf->mac_stats_bar) {
drivers/net/ethernet/netronome/nfp/abm/main.c
496
reprs = nfp_reprs_alloc(pf->max_data_vnics);
drivers/net/ethernet/netronome/nfp/abm/main.c
501
reprs = nfp_reprs_alloc(pf->max_data_vnics);
drivers/net/ethernet/netronome/nfp/bpf/main.c
210
struct nfp_cpp *cpp = bpf->app->pf->cpp;
drivers/net/ethernet/netronome/nfp/bpf/main.c
338
struct nfp_cpp *cpp = app->pf->cpp;
drivers/net/ethernet/netronome/nfp/bpf/main.c
342
mem = nfp_rtsym_map(app->pf->rtbl, "_abi_bpf_capabilities", "bpf.cap",
drivers/net/ethernet/netronome/nfp/bpf/main.c
65
struct nfp_pf *pf = app->pf;
drivers/net/ethernet/netronome/nfp/bpf/main.c
69
if (!pf->eth_tbl) {
drivers/net/ethernet/netronome/nfp/bpf/main.c
70
nfp_err(pf->cpp, "No ETH table\n");
drivers/net/ethernet/netronome/nfp/bpf/main.c
73
if (pf->max_data_vnics != pf->eth_tbl->count) {
drivers/net/ethernet/netronome/nfp/bpf/main.c
74
nfp_err(pf->cpp, "ETH entries don't match vNICs (%d vs %d)\n",
drivers/net/ethernet/netronome/nfp/bpf/main.c
75
pf->max_data_vnics, pf->eth_tbl->count);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
284
dev_warn_once(&nfp_prog->bpf->app->pf->pdev->dev,
drivers/net/ethernet/netronome/nfp/devlink_param.c
102
nfp_warn(pf->cpp, "can't access NSP: %d\n", err);
drivers/net/ethernet/netronome/nfp/devlink_param.c
110
nfp_warn(pf->cpp, "HWinfo lookup failed: %d\n", err);
drivers/net/ethernet/netronome/nfp/devlink_param.c
116
nfp_warn(pf->cpp, "HWinfo '%s' value %li invalid\n",
drivers/net/ethernet/netronome/nfp/devlink_param.c
140
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/devlink_param.c
150
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/devlink_param.c
153
nfp_warn(pf->cpp, "can't access NSP: %d\n", err);
drivers/net/ethernet/netronome/nfp/devlink_param.c
162
nfp_warn(pf->cpp, "HWinfo set failed: %d\n", err);
drivers/net/ethernet/netronome/nfp/devlink_param.c
209
static int nfp_devlink_supports_params(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/devlink_param.c
215
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/devlink_param.c
218
dev_err(&pf->pdev->dev, "Failed to access the NSP: %d\n", err);
drivers/net/ethernet/netronome/nfp/devlink_param.c
229
int nfp_devlink_params_register(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/devlink_param.c
231
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/devlink_param.c
234
err = nfp_devlink_supports_params(pf);
drivers/net/ethernet/netronome/nfp/devlink_param.c
242
void nfp_devlink_params_unregister(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/devlink_param.c
246
err = nfp_devlink_supports_params(pf);
drivers/net/ethernet/netronome/nfp/devlink_param.c
250
devl_params_unregister(priv_to_devlink(pf), nfp_devlink_params,
drivers/net/ethernet/netronome/nfp/devlink_param.c
88
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/devlink_param.c
99
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/flower/main.c
381
u8 nfp_pcie = nfp_cppcore_pcie_unit(app->pf->cpp);
drivers/net/ethernet/netronome/nfp/flower/main.c
437
app->pf->vf_cfg_mem + i * NFP_NET_CFG_BAR_SZ;
drivers/net/ethernet/netronome/nfp/flower/main.c
496
struct nfp_eth_table *eth_tbl = app->pf->eth_tbl;
drivers/net/ethernet/netronome/nfp/flower/main.c
545
err = nfp_port_init_phy_port(app->pf, app, port, i);
drivers/net/ethernet/netronome/nfp/flower/main.c
554
nfp_net_get_mac_addr(app->pf, repr, port);
drivers/net/ethernet/netronome/nfp/flower/main.c
632
if (app->pf->num_vfs)
drivers/net/ethernet/netronome/nfp/flower/main.c
657
if (app->pf->num_vfs) {
drivers/net/ethernet/netronome/nfp/flower/main.c
661
app->pf->num_vfs);
drivers/net/ethernet/netronome/nfp/flower/main.c
686
feat = nfp_rtsym_read_le(app->pf->rtbl,
drivers/net/ethernet/netronome/nfp/flower/main.c
708
err = nfp_rtsym_write_le(app->pf->rtbl, "_abi_flower_host_mask",
drivers/net/ethernet/netronome/nfp/flower/main.c
717
err = nfp_rtsym_write_le(app->pf->rtbl,
drivers/net/ethernet/netronome/nfp/flower/main.c
730
err = nfp_rtsym_write_le(app->pf->rtbl,
drivers/net/ethernet/netronome/nfp/flower/main.c
751
const struct nfp_pf *pf = app->pf;
drivers/net/ethernet/netronome/nfp/flower/main.c
755
if (!pf->eth_tbl) {
drivers/net/ethernet/netronome/nfp/flower/main.c
760
if (!pf->mac_stats_bar) {
drivers/net/ethernet/netronome/nfp/flower/main.c
765
if (!pf->vf_cfg_bar) {
drivers/net/ethernet/netronome/nfp/flower/main.c
770
version = nfp_rtsym_read_le(app->pf->rtbl, "hw_flower_version", &err);
drivers/net/ethernet/netronome/nfp/flower/main.c
776
num_mems = nfp_rtsym_read_le(app->pf->rtbl, "CONFIG_FC_HOST_CTX_SPLIT",
drivers/net/ethernet/netronome/nfp/flower/main.c
793
ctx_count = nfp_rtsym_read_le(app->pf->rtbl, "CONFIG_FC_HOST_CTX_COUNT",
drivers/net/ethernet/netronome/nfp/flower/main.c
831
features = nfp_rtsym_read_le(app->pf->rtbl,
drivers/net/ethernet/netronome/nfp/nfp_app.c
228
struct nfp_app *nfp_app_alloc(struct nfp_pf *pf, enum nfp_app_id id)
drivers/net/ethernet/netronome/nfp/nfp_app.c
233
nfp_err(pf->cpp, "unknown FW app ID 0x%02x, driver too old or support for FW not built in\n", id);
drivers/net/ethernet/netronome/nfp/nfp_app.c
246
app->pf = pf;
drivers/net/ethernet/netronome/nfp/nfp_app.c
247
app->cpp = pf->cpp;
drivers/net/ethernet/netronome/nfp/nfp_app.c
248
app->pdev = pf->pdev;
drivers/net/ethernet/netronome/nfp/nfp_app.c
59
if (!app || !app->pf->mip)
drivers/net/ethernet/netronome/nfp/nfp_app.c
61
return nfp_mip_name(app->pf->mip);
drivers/net/ethernet/netronome/nfp/nfp_app.h
163
struct nfp_pf *pf;
drivers/net/ethernet/netronome/nfp/nfp_app.h
179
devl_assert_locked(priv_to_devlink(app->pf));
drivers/net/ethernet/netronome/nfp/nfp_app.h
184
return devl_lock_is_held(priv_to_devlink(app->pf));
drivers/net/ethernet/netronome/nfp/nfp_app.h
348
trace_devlink_hwmsg(priv_to_devlink(app->pf), false, 0,
drivers/net/ethernet/netronome/nfp/nfp_app.h
356
trace_devlink_hwmsg(priv_to_devlink(app->pf), false, 0,
drivers/net/ethernet/netronome/nfp/nfp_app.h
364
trace_devlink_hwmsg(priv_to_devlink(app->pf), true, 0,
drivers/net/ethernet/netronome/nfp/nfp_app.h
376
trace_devlink_hwmsg(priv_to_devlink(app->pf), true, 0, data, len);
drivers/net/ethernet/netronome/nfp/nfp_app.h
436
struct nfp_app *nfp_app_alloc(struct nfp_pf *pf, enum nfp_app_id id);
drivers/net/ethernet/netronome/nfp/nfp_app.h
445
int nfp_app_nic_vnic_init_phy_port(struct nfp_pf *pf, struct nfp_app *app,
drivers/net/ethernet/netronome/nfp/nfp_app_nic.c
11
int nfp_app_nic_vnic_init_phy_port(struct nfp_pf *pf, struct nfp_app *app,
drivers/net/ethernet/netronome/nfp/nfp_app_nic.c
16
if (!pf->eth_tbl)
drivers/net/ethernet/netronome/nfp/nfp_app_nic.c
23
err = nfp_port_init_phy_port(pf, app, nn->port, id);
drivers/net/ethernet/netronome/nfp/nfp_app_nic.c
37
err = nfp_app_nic_vnic_init_phy_port(app->pf, app, nn, id);
drivers/net/ethernet/netronome/nfp/nfp_app_nic.c
41
nfp_net_get_mac_addr(app->pf, nn->dp.netdev, nn->port);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
101
ret = nfp_devlink_fill_eth_port_from_id(pf, port, &eth_port);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
116
return nfp_devlink_set_lanes(pf, eth_port.index, lanes);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
123
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
125
return nfp_shared_buf_pool_get(pf, sb_index, pool_index, pool_info);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
134
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
136
return nfp_shared_buf_pool_set(pf, sb_index, pool_index,
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
142
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
144
return nfp_app_eswitch_mode_get(pf->app, mode);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
150
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
152
return nfp_app_eswitch_mode_set(pf->app, mode);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
167
nfp_devlink_versions_get_hwinfo(struct nfp_pf *pf, struct devlink_info_req *req)
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
178
val = nfp_hwinfo_lookup(pf->hwinfo, info->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
241
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
247
vendor = nfp_hwinfo_lookup(pf->hwinfo, "assembly.vendor");
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
248
part = nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno");
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
249
sn = nfp_hwinfo_lookup(pf->hwinfo, "assembly.serial");
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
269
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
29
nfp_devlink_fill_eth_port_from_id(struct nfp_pf *pf,
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
301
return nfp_devlink_versions_get_hwinfo(pf, req);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
360
devlink = priv_to_devlink(app->pf);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
39
nfp_devlink_set_lanes(struct nfp_pf *pf, unsigned int idx, unsigned int lanes)
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
44
nsp = nfp_eth_config_start(pf->cpp, idx);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
60
return nfp_net_refresh_port_table_sync(pf);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
67
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
73
ret = nfp_devlink_fill_eth_port_from_id(pf, port, &eth_port);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
88
return nfp_devlink_set_lanes(pf, eth_port.index, lanes);
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
95
struct nfp_pf *pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
105
int nfp_hwmon_register(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
110
if (!pf->nspi) {
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
111
nfp_warn(pf->cpp, "not registering HWMON (no NSP info)\n");
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
114
if (!pf->nspi->sensor_mask) {
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
115
nfp_info(pf->cpp,
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
120
pf->hwmon_dev = hwmon_device_register_with_info(&pf->pdev->dev, "nfp",
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
121
pf, &nfp_chip_info,
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
123
return PTR_ERR_OR_ZERO(pf->hwmon_dev);
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
126
void nfp_hwmon_unregister(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
128
if (!IS_REACHABLE(CONFIG_HWMON) || !pf->hwmon_dev)
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
131
hwmon_device_unregister(pf->hwmon_dev);
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
39
struct nfp_pf *pf = dev_get_drvdata(dev);
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
54
if (!(pf->nspi->sensor_mask & BIT(id)))
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
58
return nfp_hwmon_read_sensor(pf->cpp, id, val);
drivers/net/ethernet/netronome/nfp/nfp_hwmon.c
60
return nfp_hwmon_read_sensor(pf->cpp, id, val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
101
return nfp_rtsym_map(pf->rtbl, pf_symbol, name, min_size, area);
drivers/net/ethernet/netronome/nfp/nfp_main.c
105
int nfp_mbox_cmd(struct nfp_pf *pf, u32 cmd, void *in_data, u64 in_length,
drivers/net/ethernet/netronome/nfp/nfp_main.c
113
if (!pf->mbox)
drivers/net/ethernet/netronome/nfp/nfp_main.c
116
max_data_sz = nfp_rtsym_size(pf->mbox) - NFP_MBOX_SYM_MIN_SIZE;
drivers/net/ethernet/netronome/nfp/nfp_main.c
119
err = nfp_rtsym_readl(pf->cpp, pf->mbox, NFP_MBOX_CMD, &val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
121
nfp_warn(pf->cpp, "failed to issue command (%u): %u, err: %d\n",
drivers/net/ethernet/netronome/nfp/nfp_main.c
127
n = nfp_rtsym_write(pf->cpp, pf->mbox, NFP_MBOX_DATA, in_data,
drivers/net/ethernet/netronome/nfp/nfp_main.c
132
err = nfp_rtsym_writeq(pf->cpp, pf->mbox, NFP_MBOX_DATA_LEN, in_length);
drivers/net/ethernet/netronome/nfp/nfp_main.c
137
err = nfp_rtsym_readl(pf->cpp, pf->mbox, NFP_MBOX_DATA_LEN, &val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
142
err = nfp_rtsym_writeq(pf->cpp, pf->mbox, NFP_MBOX_CMD, cmd);
drivers/net/ethernet/netronome/nfp/nfp_main.c
149
err = nfp_rtsym_readl(pf->cpp, pf->mbox, NFP_MBOX_CMD, &val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
162
err = nfp_rtsym_readl(pf->cpp, pf->mbox, NFP_MBOX_DATA_LEN, &val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
167
n = nfp_rtsym_read(pf->cpp, pf->mbox, NFP_MBOX_DATA,
drivers/net/ethernet/netronome/nfp/nfp_main.c
173
err = nfp_rtsym_readl(pf->cpp, pf->mbox, NFP_MBOX_RET, &val);
drivers/net/ethernet/netronome/nfp/nfp_main.c
182
static bool nfp_board_ready(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
188
cp = nfp_hwinfo_lookup(pf->hwinfo, "board.state");
drivers/net/ethernet/netronome/nfp/nfp_main.c
199
static int nfp_pf_board_state_wait(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
203
while (!nfp_board_ready(pf)) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
205
nfp_err(pf->cpp, "NFP board initialization timeout\n");
drivers/net/ethernet/netronome/nfp/nfp_main.c
209
nfp_info(pf->cpp, "waiting for board initialization\n");
drivers/net/ethernet/netronome/nfp/nfp_main.c
214
kfree(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_main.c
215
pf->hwinfo = nfp_hwinfo_read(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
221
static int nfp_pcie_sriov_read_nfd_limit(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
225
pf->limit_vfs = nfp_rtsym_read_le(pf->rtbl, "nfd_vf_cfg_max_vfs", &err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
228
pf->limit_vfs = ~0;
drivers/net/ethernet/netronome/nfp/nfp_main.c
232
nfp_warn(pf->cpp, "Warning: VF limit read failed: %d\n", err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
236
err = pci_sriov_set_totalvfs(pf->pdev, pf->limit_vfs);
drivers/net/ethernet/netronome/nfp/nfp_main.c
238
nfp_warn(pf->cpp, "Failed to set VF count in sysfs: %d\n", err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
245
struct nfp_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/netronome/nfp/nfp_main.c
249
if (num_vfs > pf->limit_vfs) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
250
nfp_info(pf->cpp, "Firmware limits number of VFs to %u\n",
drivers/net/ethernet/netronome/nfp/nfp_main.c
251
pf->limit_vfs);
drivers/net/ethernet/netronome/nfp/nfp_main.c
261
devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
264
err = nfp_app_sriov_enable(pf->app, num_vfs);
drivers/net/ethernet/netronome/nfp/nfp_main.c
272
pf->num_vfs = num_vfs;
drivers/net/ethernet/netronome/nfp/nfp_main.c
274
dev_dbg(&pdev->dev, "Created %d VFs.\n", pf->num_vfs);
drivers/net/ethernet/netronome/nfp/nfp_main.c
290
struct nfp_pf *pf = pci_get_drvdata(pdev);
drivers/net/ethernet/netronome/nfp/nfp_main.c
293
devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
306
nfp_app_sriov_disable(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_main.c
308
pf->num_vfs = 0;
drivers/net/ethernet/netronome/nfp/nfp_main.c
329
int nfp_flash_update_common(struct nfp_pf *pf, const struct firmware *fw,
drivers/net/ethernet/netronome/nfp/nfp_main.c
332
struct device *dev = &pf->pdev->dev;
drivers/net/ethernet/netronome/nfp/nfp_main.c
336
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
358
nfp_net_fw_request(struct pci_dev *pdev, struct nfp_pf *pf, const char *name)
drivers/net/ethernet/netronome/nfp/nfp_main.c
364
nfp_info(pf->cpp, " %s: %s\n",
drivers/net/ethernet/netronome/nfp/nfp_main.c
380
nfp_net_fw_find(struct pci_dev *pdev, struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
390
nfp_info(pf->cpp, "Looking for firmware file in order of priority:\n");
drivers/net/ethernet/netronome/nfp/nfp_main.c
393
interface = nfp_cpp_interface(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
394
nfp_cpp_serial(pf->cpp, &serial);
drivers/net/ethernet/netronome/nfp/nfp_main.c
397
fw = nfp_net_fw_request(pdev, pf, fw_name);
drivers/net/ethernet/netronome/nfp/nfp_main.c
403
fw = nfp_net_fw_request(pdev, pf, fw_name);
drivers/net/ethernet/netronome/nfp/nfp_main.c
408
if (!pf->eth_tbl) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
413
fw_model = nfp_hwinfo_lookup(pf->hwinfo, "nffw.partno");
drivers/net/ethernet/netronome/nfp/nfp_main.c
415
fw_model = nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno");
drivers/net/ethernet/netronome/nfp/nfp_main.c
424
for (i = 0; spc > 0 && i < pf->eth_tbl->count; i += j) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
425
port = &pf->eth_tbl->ports[i];
drivers/net/ethernet/netronome/nfp/nfp_main.c
427
while (i + j < pf->eth_tbl->count &&
drivers/net/ethernet/netronome/nfp/nfp_main.c
442
return nfp_net_fw_request(pdev, pf, fw_name);
drivers/net/ethernet/netronome/nfp/nfp_main.c
481
nfp_fw_load(struct pci_dev *pdev, struct nfp_pf *pf, struct nfp_nsp *nsp)
drivers/net/ethernet/netronome/nfp/nfp_main.c
496
interface = nfp_cpp_interface(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
531
fw = nfp_net_fw_find(pdev, pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
581
pf->unload_fw_on_remove = true;
drivers/net/ethernet/netronome/nfp/nfp_main.c
587
nfp_nsp_init_ports(struct pci_dev *pdev, struct nfp_pf *pf,
drivers/net/ethernet/netronome/nfp/nfp_main.c
593
pf->eth_tbl = __nfp_eth_read_ports(pf->cpp, nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
594
if (!pf->eth_tbl)
drivers/net/ethernet/netronome/nfp/nfp_main.c
600
for (i = 0; i < pf->eth_tbl->count; i++)
drivers/net/ethernet/netronome/nfp/nfp_main.c
601
needs_reinit |= pf->eth_tbl->ports[i].override_changed;
drivers/net/ethernet/netronome/nfp/nfp_main.c
605
kfree(pf->eth_tbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
609
pf->eth_tbl = __nfp_eth_read_ports(pf->cpp, nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
612
static int nfp_nsp_init(struct pci_dev *pdev, struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
617
err = nfp_resource_wait(pf->cpp, NFP_RESOURCE_NSP, 30);
drivers/net/ethernet/netronome/nfp/nfp_main.c
621
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
632
nfp_nsp_init_ports(pdev, pf, nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
634
pf->nspi = __nfp_nsp_identify(nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
635
if (pf->nspi)
drivers/net/ethernet/netronome/nfp/nfp_main.c
636
dev_info(&pdev->dev, "BSP: %s\n", pf->nspi->version);
drivers/net/ethernet/netronome/nfp/nfp_main.c
638
err = nfp_fw_load(pdev, pf, nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
640
kfree(pf->nspi);
drivers/net/ethernet/netronome/nfp/nfp_main.c
641
kfree(pf->eth_tbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
646
pf->fw_loaded = !!err;
drivers/net/ethernet/netronome/nfp/nfp_main.c
655
static void nfp_fw_unload(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
660
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
662
nfp_err(pf->cpp, "Reset failed, can't open NSP\n");
drivers/net/ethernet/netronome/nfp/nfp_main.c
668
dev_warn(&pf->pdev->dev, "Couldn't unload firmware: %d\n", err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
670
dev_info(&pf->pdev->dev, "Firmware safely unloaded\n");
drivers/net/ethernet/netronome/nfp/nfp_main.c
675
static int nfp_pf_find_rtsyms(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
680
pf_id = nfp_cppcore_pcie_unit(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
684
pf->mbox = nfp_rtsym_lookup(pf->rtbl, pf_symbol);
drivers/net/ethernet/netronome/nfp/nfp_main.c
685
if (pf->mbox && nfp_rtsym_size(pf->mbox) < NFP_MBOX_SYM_MIN_SIZE) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
686
nfp_err(pf->cpp, "PF mailbox symbol too small: %llu < %d\n",
drivers/net/ethernet/netronome/nfp/nfp_main.c
687
nfp_rtsym_size(pf->mbox), NFP_MBOX_SYM_MIN_SIZE);
drivers/net/ethernet/netronome/nfp/nfp_main.c
694
int nfp_net_pf_get_app_id(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
696
return nfp_pf_rtsym_read_optional(pf, "_pf%u_net_app_id",
drivers/net/ethernet/netronome/nfp/nfp_main.c
700
static u64 nfp_net_pf_get_app_cap(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
706
snprintf(name, sizeof(name), "_pf%u_net_app_cap", nfp_cppcore_pcie_unit(pf->cpp));
drivers/net/ethernet/netronome/nfp/nfp_main.c
708
val = nfp_rtsym_read_le(pf->rtbl, name, &err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
711
nfp_err(pf->cpp, "Unable to read symbol %s\n", name);
drivers/net/ethernet/netronome/nfp/nfp_main.c
719
static void nfp_pf_cfg_hwinfo(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
72
int nfp_pf_rtsym_read_optional(struct nfp_pf *pf, const char *format,
drivers/net/ethernet/netronome/nfp/nfp_main.c
726
nsp = nfp_nsp_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
733
sp_indiff = (nfp_net_pf_get_app_id(pf) == NFP_APP_FLOWER_NIC) ||
drivers/net/ethernet/netronome/nfp/nfp_main.c
734
(nfp_net_pf_get_app_cap(pf) & NFP_NET_APP_CAP_SP_INDIFF);
drivers/net/ethernet/netronome/nfp/nfp_main.c
743
nfp_warn(pf->cpp, "HWinfo(sp_indiff=%d) set failed: %d\n", sp_indiff, err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
748
kfree(pf->eth_tbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
749
pf->eth_tbl = __nfp_eth_read_ports(pf->cpp, nsp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
761
struct nfp_pf *pf;
drivers/net/ethernet/netronome/nfp/nfp_main.c
788
devlink = devlink_alloc(&nfp_devlink_ops, sizeof(*pf), &pdev->dev);
drivers/net/ethernet/netronome/nfp/nfp_main.c
79
snprintf(name, sizeof(name), format, nfp_cppcore_pcie_unit(pf->cpp));
drivers/net/ethernet/netronome/nfp/nfp_main.c
793
pf = devlink_priv(devlink);
drivers/net/ethernet/netronome/nfp/nfp_main.c
794
INIT_LIST_HEAD(&pf->vnics);
drivers/net/ethernet/netronome/nfp/nfp_main.c
795
INIT_LIST_HEAD(&pf->ports);
drivers/net/ethernet/netronome/nfp/nfp_main.c
796
pci_set_drvdata(pdev, pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
797
pf->pdev = pdev;
drivers/net/ethernet/netronome/nfp/nfp_main.c
798
pf->dev_info = dev_info;
drivers/net/ethernet/netronome/nfp/nfp_main.c
800
pf->wq = alloc_workqueue("nfp-%s", WQ_PERCPU, 2, pci_name(pdev));
drivers/net/ethernet/netronome/nfp/nfp_main.c
801
if (!pf->wq) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
806
pf->cpp = nfp_cpp_from_nfp6000_pcie(pdev, dev_info);
drivers/net/ethernet/netronome/nfp/nfp_main.c
807
if (IS_ERR(pf->cpp)) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
808
err = PTR_ERR(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
81
val = nfp_rtsym_read_le(pf->rtbl, name, &err);
drivers/net/ethernet/netronome/nfp/nfp_main.c
812
err = nfp_resource_table_init(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
816
pf->hwinfo = nfp_hwinfo_read(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
819
nfp_hwinfo_lookup(pf->hwinfo, "assembly.vendor"),
drivers/net/ethernet/netronome/nfp/nfp_main.c
820
nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno"),
drivers/net/ethernet/netronome/nfp/nfp_main.c
821
nfp_hwinfo_lookup(pf->hwinfo, "assembly.serial"),
drivers/net/ethernet/netronome/nfp/nfp_main.c
822
nfp_hwinfo_lookup(pf->hwinfo, "assembly.revision"),
drivers/net/ethernet/netronome/nfp/nfp_main.c
823
nfp_hwinfo_lookup(pf->hwinfo, "cpld.version"));
drivers/net/ethernet/netronome/nfp/nfp_main.c
825
err = nfp_pf_board_state_wait(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
829
err = nfp_nsp_init(pdev, pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
833
pf->mip = nfp_mip_open(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
834
pf->rtbl = __nfp_rtsym_table_read(pf->cpp, pf->mip);
drivers/net/ethernet/netronome/nfp/nfp_main.c
836
err = nfp_pf_find_rtsyms(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
840
pf->dump_flag = NFP_DUMP_NSP_DIAG;
drivers/net/ethernet/netronome/nfp/nfp_main.c
841
pf->dumpspec = nfp_net_dump_load_dumpspec(pf->cpp, pf->rtbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
843
err = nfp_pcie_sriov_read_nfd_limit(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
847
pf->num_vfs = pci_num_vf(pdev);
drivers/net/ethernet/netronome/nfp/nfp_main.c
848
if (pf->num_vfs > pf->limit_vfs) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
85
nfp_err(pf->cpp, "Unable to read symbol %s\n", name);
drivers/net/ethernet/netronome/nfp/nfp_main.c
851
pf->num_vfs, pf->limit_vfs);
drivers/net/ethernet/netronome/nfp/nfp_main.c
856
nfp_pf_cfg_hwinfo(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
858
err = nfp_net_pci_probe(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
862
err = nfp_hwmon_register(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
871
nfp_net_pci_remove(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
873
kfree(pf->rtbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
874
nfp_mip_close(pf->mip);
drivers/net/ethernet/netronome/nfp/nfp_main.c
875
if (pf->unload_fw_on_remove)
drivers/net/ethernet/netronome/nfp/nfp_main.c
876
nfp_fw_unload(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
877
kfree(pf->eth_tbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
878
kfree(pf->nspi);
drivers/net/ethernet/netronome/nfp/nfp_main.c
879
vfree(pf->dumpspec);
drivers/net/ethernet/netronome/nfp/nfp_main.c
881
kfree(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_main.c
883
nfp_cpp_free(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
885
destroy_workqueue(pf->wq);
drivers/net/ethernet/netronome/nfp/nfp_main.c
899
struct nfp_pf *pf;
drivers/net/ethernet/netronome/nfp/nfp_main.c
901
pf = pci_get_drvdata(pdev);
drivers/net/ethernet/netronome/nfp/nfp_main.c
902
if (!pf)
drivers/net/ethernet/netronome/nfp/nfp_main.c
905
nfp_hwmon_unregister(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
909
nfp_net_pci_remove(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
911
vfree(pf->dumpspec);
drivers/net/ethernet/netronome/nfp/nfp_main.c
912
kfree(pf->rtbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
913
nfp_mip_close(pf->mip);
drivers/net/ethernet/netronome/nfp/nfp_main.c
914
if (unload_fw && pf->unload_fw_on_remove)
drivers/net/ethernet/netronome/nfp/nfp_main.c
915
nfp_fw_unload(pf);
drivers/net/ethernet/netronome/nfp/nfp_main.c
917
destroy_workqueue(pf->wq);
drivers/net/ethernet/netronome/nfp/nfp_main.c
919
kfree(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_main.c
920
nfp_cpp_free(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_main.c
922
kfree(pf->eth_tbl);
drivers/net/ethernet/netronome/nfp/nfp_main.c
923
kfree(pf->nspi);
drivers/net/ethernet/netronome/nfp/nfp_main.c
924
devlink_free(priv_to_devlink(pf));
drivers/net/ethernet/netronome/nfp/nfp_main.c
93
nfp_pf_map_rtsym(struct nfp_pf *pf, const char *name, const char *sym_fmt,
drivers/net/ethernet/netronome/nfp/nfp_main.c
99
nfp_cppcore_pcie_unit(pf->cpp));
drivers/net/ethernet/netronome/nfp/nfp_main.h
150
int nfp_net_pci_probe(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
151
void nfp_net_pci_remove(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
153
int nfp_hwmon_register(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
154
void nfp_hwmon_unregister(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
157
nfp_net_get_mac_addr(struct nfp_pf *pf, struct net_device *netdev,
drivers/net/ethernet/netronome/nfp/nfp_main.h
162
int nfp_pf_rtsym_read_optional(struct nfp_pf *pf, const char *format,
drivers/net/ethernet/netronome/nfp/nfp_main.h
164
int nfp_net_pf_get_app_id(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
166
nfp_pf_map_rtsym(struct nfp_pf *pf, const char *name, const char *sym_fmt,
drivers/net/ethernet/netronome/nfp/nfp_main.h
168
int nfp_mbox_cmd(struct nfp_pf *pf, u32 cmd, void *in_data, u64 in_length,
drivers/net/ethernet/netronome/nfp/nfp_main.h
170
int nfp_flash_update_common(struct nfp_pf *pf, const struct firmware *fw,
drivers/net/ethernet/netronome/nfp/nfp_main.h
179
s64 nfp_net_dump_calculate_size(struct nfp_pf *pf, struct nfp_dumpspec *spec,
drivers/net/ethernet/netronome/nfp/nfp_main.h
181
int nfp_net_dump_populate_buffer(struct nfp_pf *pf, struct nfp_dumpspec *spec,
drivers/net/ethernet/netronome/nfp/nfp_main.h
184
int nfp_shared_buf_register(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
185
void nfp_shared_buf_unregister(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
186
int nfp_shared_buf_pool_get(struct nfp_pf *pf, unsigned int sb, u16 pool_index,
drivers/net/ethernet/netronome/nfp/nfp_main.h
188
int nfp_shared_buf_pool_set(struct nfp_pf *pf, unsigned int sb,
drivers/net/ethernet/netronome/nfp/nfp_main.h
192
int nfp_devlink_params_register(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_main.h
193
void nfp_devlink_params_unregister(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
117
typedef int (*nfp_tlv_visit)(struct nfp_pf *pf, struct nfp_dump_tl *tl,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
121
nfp_traverse_tlvs(struct nfp_pf *pf, void *data, u32 data_length, void *param,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
146
err = tlv_visit(pf, tl, param);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
195
static int nfp_calc_fwname_tlv_size(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
197
u32 fwname_len = strlen(nfp_mip_name(pf->mip));
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
202
static int nfp_calc_hwinfo_field_sz(struct nfp_pf *pf, struct nfp_dump_tl *spec)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
212
value = nfp_hwinfo_lookup(pf->hwinfo, spec->data);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
234
nfp_calc_rtsym_dump_sz(struct nfp_pf *pf, struct nfp_dump_tl *spec)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
236
struct nfp_rtsym_table *rtbl = pf->rtbl;
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
256
nfp_add_tlv_size(struct nfp_pf *pf, struct nfp_dump_tl *tl, void *param)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
264
*size += nfp_calc_fwname_tlv_size(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
286
*size += nfp_calc_rtsym_dump_sz(pf, tl);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
289
hwinfo_size = nfp_hwinfo_get_packed_str_size(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
293
*size += nfp_calc_hwinfo_field_sz(pf, tl);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
304
nfp_calc_specific_level_size(struct nfp_pf *pf, struct nfp_dump_tl *dump_level,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
312
return nfp_traverse_tlvs(pf, dump_level->data,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
317
s64 nfp_net_dump_calculate_size(struct nfp_pf *pf, struct nfp_dumpspec *spec,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
326
err = nfp_traverse_tlvs(pf, spec->data, spec->size, &lev_sz,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
374
static int nfp_dump_fwname(struct nfp_pf *pf, struct nfp_dump_state *dump)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
381
fwname = nfp_mip_name(pf->mip);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
395
nfp_dump_hwinfo(struct nfp_pf *pf, struct nfp_dump_tl *spec,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
403
hwinfo = nfp_hwinfo_get_packed_strings(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
404
hwinfo_size = nfp_hwinfo_get_packed_str_size(pf->hwinfo);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
416
static int nfp_dump_hwinfo_field(struct nfp_pf *pf, struct nfp_dump_tl *spec,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
431
value = nfp_hwinfo_lookup(pf->hwinfo, key);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
454
nfp_dump_csr_range(struct nfp_pf *pf, struct nfp_dumpspec_csr *spec_csr,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
489
err = nfp_xpb_readl(pf->cpp, cpp_rd_addr, (u32 *)dest);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
491
bytes_read = nfp_cpp_read(pf->cpp, cpp_id, cpp_rd_addr,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
558
nfp_dump_indirect_csr_range(struct nfp_pf *pf,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
591
err = nfp_read_all_indirect_csr_ctx(pf->cpp, spec_csr,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
606
nfp_dump_single_rtsym(struct nfp_pf *pf, struct nfp_dumpspec_rtsym *spec,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
613
struct nfp_rtsym_table *rtbl = pf->rtbl;
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
654
bytes_read = nfp_rtsym_read(pf->cpp, sym, 0, dest, sym_size);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
665
nfp_dump_for_tlv(struct nfp_pf *pf, struct nfp_dump_tl *tl, void *param)
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
674
err = nfp_dump_fwname(pf, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
682
err = nfp_dump_csr_range(pf, spec_csr, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
688
err = nfp_dump_indirect_csr_range(pf, spec_csr, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
694
err = nfp_dump_single_rtsym(pf, spec_rtsym, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
699
err = nfp_dump_hwinfo(pf, tl, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
704
err = nfp_dump_hwinfo_field(pf, tl, dump);
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
718
nfp_dump_specific_level(struct nfp_pf *pf, struct nfp_dump_tl *dump_level,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
726
return nfp_traverse_tlvs(pf, dump_level->data,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
748
int nfp_net_dump_populate_buffer(struct nfp_pf *pf, struct nfp_dumpspec *spec,
drivers/net/ethernet/netronome/nfp/nfp_net_debugdump.c
763
err = nfp_traverse_tlvs(pf, spec->data, spec->size, &dump,
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1947
app->pf->dump_flag = val->flag;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1951
if (!app->pf->dumpspec)
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1954
len = nfp_net_dump_calculate_size(app->pf, app->pf->dumpspec,
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1959
app->pf->dump_flag = val->flag;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1960
app->pf->dump_len = len;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1973
if (app->pf->dump_flag == NFP_DUMP_NSP_DIAG)
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1976
dump->flag = app->pf->dump_flag;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1977
dump->len = app->pf->dump_len;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1991
if (app->pf->dump_flag == NFP_DUMP_NSP_DIAG)
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1994
dump->flag = app->pf->dump_flag;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1995
dump->len = app->pf->dump_len;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1997
return nfp_net_dump_populate_buffer(app->pf, app->pf->dumpspec, dump,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
113
nn = nfp_net_alloc(pf->pdev, pf->dev_info, ctrl_bar, needs_netdev,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
118
nn->app = pf->app;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
126
err = nfp_app_vnic_alloc(pf->app, nn, id);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
133
pf->num_vnics++;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
134
list_add_tail(&nn->vnic_list, &pf->vnics);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
140
nfp_net_pf_init_vnic(struct nfp_pf *pf, struct nfp_net *nn, unsigned int id)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
147
err = nfp_devlink_port_register(pf->app, nn->port);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
156
nfp_net_debugfs_vnic_add(nn, pf->ddir);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
161
err = nfp_app_vnic_init(pf->app, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
178
nfp_net_pf_alloc_vnics(struct nfp_pf *pf, void __iomem *ctrl_bar,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
185
for (i = 0; i < pf->max_data_vnics; i++) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
186
nn = nfp_net_pf_alloc_vnic(pf, true, ctrl_bar, qc_bar,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
200
nfp_net_pf_free_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
203
if (list_empty(&pf->vnics))
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
209
nfp_net_pf_free_vnics(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
213
static void nfp_net_pf_clean_vnic(struct nfp_pf *pf, struct nfp_net *nn)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
216
nfp_app_vnic_clean(pf->app, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
223
static int nfp_net_pf_alloc_irqs(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
230
list_for_each_entry(nn, &pf->vnics, vnic_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
232
pf->irq_entries = kzalloc_objs(*pf->irq_entries, wanted_irqs);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
233
if (!pf->irq_entries)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
236
num_irqs = nfp_net_irqs_alloc(pf->pdev, pf->irq_entries,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
237
NFP_NET_MIN_VNIC_IRQS * pf->num_vnics,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
240
nfp_warn(pf->cpp, "Unable to allocate MSI-X vectors\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
241
kfree(pf->irq_entries);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
247
vnics_left = pf->num_vnics;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
248
list_for_each_entry(nn, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
253
nfp_net_irqs_assign(nn, &pf->irq_entries[num_irqs - irqs_left],
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
262
static void nfp_net_pf_free_irqs(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
264
nfp_net_irqs_disable(pf->pdev);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
265
kfree(pf->irq_entries);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
268
static int nfp_net_pf_init_vnics(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
276
list_for_each_entry(nn, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
279
err = nfp_net_pf_init_vnic(pf, nn, id);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
289
list_for_each_entry_continue_reverse(nn, &pf->vnics, vnic_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
291
nfp_net_pf_clean_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
296
nfp_net_pf_app_init(struct nfp_pf *pf, u8 __iomem *qc_bar, unsigned int stride)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
298
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
302
pf->app = nfp_app_alloc(pf, nfp_net_pf_get_app_id(pf));
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
303
if (IS_ERR(pf->app))
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
304
return PTR_ERR(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
307
err = nfp_app_init(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
312
if (!nfp_app_needs_ctrl_vnic(pf->app))
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
315
ctrl_bar = nfp_pf_map_rtsym(pf, "net.ctrl", "_pf%u_net_ctrl_bar",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
316
NFP_PF_CSR_SLICE_SIZE, &pf->ctrl_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
318
nfp_err(pf->cpp, "Failed to find ctrl vNIC memory symbol\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
323
pf->ctrl_vnic = nfp_net_pf_alloc_vnic(pf, false, ctrl_bar, qc_bar,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
325
if (IS_ERR(pf->ctrl_vnic)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
326
err = PTR_ERR(pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
333
nfp_cpp_area_release_free(pf->ctrl_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
336
nfp_app_clean(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
339
nfp_app_free(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
340
pf->app = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
344
static void nfp_net_pf_app_clean(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
346
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
348
if (pf->ctrl_vnic) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
349
nfp_net_pf_free_vnic(pf, pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
350
nfp_cpp_area_release_free(pf->ctrl_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
354
nfp_app_clean(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
357
nfp_app_free(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
358
pf->app = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
361
static int nfp_net_pf_app_start_ctrl(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
365
if (!pf->ctrl_vnic)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
367
err = nfp_net_pf_init_vnic(pf, pf->ctrl_vnic, 0);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
371
err = nfp_ctrl_open(pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
378
nfp_net_pf_clean_vnic(pf, pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
382
static void nfp_net_pf_app_stop_ctrl(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
384
if (!pf->ctrl_vnic)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
386
nfp_ctrl_close(pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
387
nfp_net_pf_clean_vnic(pf, pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
390
static int nfp_net_pf_app_start(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
394
err = nfp_net_pf_app_start_ctrl(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
398
err = nfp_app_start(pf->app, pf->ctrl_vnic);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
402
if (pf->num_vfs) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
403
err = nfp_app_sriov_enable(pf->app, pf->num_vfs);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
411
nfp_app_stop(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
413
nfp_net_pf_app_stop_ctrl(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
417
static void nfp_net_pf_app_stop(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
419
if (pf->num_vfs)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
420
nfp_app_sriov_disable(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
421
nfp_app_stop(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
422
nfp_net_pf_app_stop_ctrl(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
425
static void nfp_net_pci_unmap_mem(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
427
if (pf->vfcfg_tbl2_area)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
428
nfp_cpp_area_release_free(pf->vfcfg_tbl2_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
429
if (pf->vf_cfg_bar)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
430
nfp_cpp_area_release_free(pf->vf_cfg_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
431
if (pf->mac_stats_bar)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
432
nfp_cpp_area_release_free(pf->mac_stats_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
433
nfp_cpp_area_release_free(pf->qc_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
434
nfp_cpp_area_release_free(pf->data_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
437
static int nfp_net_pci_map_mem(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
443
min_size = pf->max_data_vnics * NFP_PF_CSR_SLICE_SIZE;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
444
mem = nfp_pf_map_rtsym(pf, "net.bar0", "_pf%d_net_bar0",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
445
min_size, &pf->data_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
447
nfp_err(pf->cpp, "Failed to find data vNIC memory symbol\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
451
if (pf->eth_tbl) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
452
min_size = NFP_MAC_STATS_SIZE * (pf->eth_tbl->max_index + 1);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
453
pf->mac_stats_mem = nfp_rtsym_map(pf->rtbl, "_mac_stats",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
455
&pf->mac_stats_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
456
if (IS_ERR(pf->mac_stats_mem)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
457
if (PTR_ERR(pf->mac_stats_mem) != -ENOENT) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
458
err = PTR_ERR(pf->mac_stats_mem);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
461
pf->mac_stats_mem = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
465
pf->vf_cfg_mem = nfp_pf_map_rtsym(pf, "net.vfcfg", "_pf%d_net_vf_bar",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
466
NFP_NET_CFG_BAR_SZ * pf->limit_vfs,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
467
&pf->vf_cfg_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
468
if (IS_ERR(pf->vf_cfg_mem)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
469
if (PTR_ERR(pf->vf_cfg_mem) != -ENOENT) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
47
nfp_net_get_mac_addr(struct nfp_pf *pf, struct net_device *netdev,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
470
err = PTR_ERR(pf->vf_cfg_mem);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
473
pf->vf_cfg_mem = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
476
min_size = NFP_NET_VF_CFG_SZ * pf->limit_vfs + NFP_NET_VF_CFG_MB_SZ;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
477
pf->vfcfg_tbl2 = nfp_pf_map_rtsym(pf, "net.vfcfg_tbl2",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
479
min_size, &pf->vfcfg_tbl2_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
480
if (IS_ERR(pf->vfcfg_tbl2)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
481
if (PTR_ERR(pf->vfcfg_tbl2) != -ENOENT) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
482
err = PTR_ERR(pf->vfcfg_tbl2);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
485
pf->vfcfg_tbl2 = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
489
mem = nfp_cpp_map_area(pf->cpp, "net.qc", cpp_id,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
490
nfp_qcp_queue_offset(pf->dev_info, 0),
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
491
pf->dev_info->qc_area_sz, &pf->qc_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
493
nfp_err(pf->cpp, "Failed to map Queue Controller area.\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
501
if (pf->vfcfg_tbl2_area)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
502
nfp_cpp_area_release_free(pf->vfcfg_tbl2_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
504
if (pf->vf_cfg_bar)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
505
nfp_cpp_area_release_free(pf->vf_cfg_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
507
if (pf->mac_stats_bar)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
508
nfp_cpp_area_release_free(pf->mac_stats_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
510
nfp_cpp_area_release_free(pf->data_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
591
int nfp_net_refresh_port_table_sync(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
593
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
602
if (list_empty(&pf->vnics))
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
607
list_for_each_entry(port, &pf->ports, port_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
610
eth_table = nfp_eth_read_ports(pf->cpp);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
612
list_for_each_entry(port, &pf->ports, port_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
616
nfp_err(pf->cpp, "Error refreshing port config!\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
620
list_for_each_entry(port, &pf->ports, port_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
622
nfp_net_eth_port_update(pf->cpp, port, eth_table);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
628
err = nfp_reprs_resync_phys_ports(pf->app);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
633
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
637
nfp_net_pf_clean_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
638
nfp_net_pf_free_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
646
struct nfp_pf *pf = container_of(work, struct nfp_pf,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
648
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
651
nfp_net_refresh_port_table_sync(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
657
struct nfp_pf *pf = port->app->pf;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
661
queue_work(pf->wq, &pf->port_refresh_work);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
689
int nfp_net_pci_probe(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
691
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
697
INIT_WORK(&pf->port_refresh_work, nfp_net_refresh_vnics);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
699
if (!pf->rtbl) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
700
nfp_err(pf->cpp, "No %s, giving up.\n",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
701
pf->fw_loaded ? "symbol table" : "firmware found");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
705
pf->max_data_vnics = nfp_net_pf_get_num_ports(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
706
if ((int)pf->max_data_vnics < 0)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
707
return pf->max_data_vnics;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
709
err = nfp_net_pci_map_mem(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
713
ctrl_bar = nfp_cpp_area_iomem(pf->data_vnic_bar);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
714
qc_bar = nfp_cpp_area_iomem(pf->qc_area);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
723
nfp_err(pf->cpp, "Unknown Firmware ABI %d.%d.%d.%d\n",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
733
nfp_warn(pf->cpp, "OBSOLETE Firmware detected - VF isolation not available\n");
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
74
static int nfp_net_pf_get_num_ports(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
740
nfp_err(pf->cpp, "Unsupported Firmware ABI %d.%d.%d.%d\n",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
748
err = nfp_net_pf_app_init(pf, qc_bar, stride);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
752
err = nfp_shared_buf_register(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
757
err = nfp_devlink_params_register(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
76
return nfp_pf_rtsym_read_optional(pf, "nfd_cfg_pf%u_num_ports", 1);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
761
pf->ddir = nfp_net_debugfs_device_add(pf->pdev);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
764
err = nfp_net_pf_alloc_vnics(pf, ctrl_bar, qc_bar, stride);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
768
err = nfp_net_pf_alloc_irqs(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
772
err = nfp_net_pf_app_start(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
776
err = nfp_net_pf_init_vnics(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
786
nfp_net_pf_app_stop(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
788
nfp_net_pf_free_irqs(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
79
static void nfp_net_pf_free_vnic(struct nfp_pf *pf, struct nfp_net *nn)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
790
nfp_net_pf_free_vnics(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
792
nfp_net_debugfs_dir_clean(&pf->ddir);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
793
nfp_devlink_params_unregister(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
796
nfp_shared_buf_unregister(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
798
cancel_work_sync(&pf->port_refresh_work);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
799
nfp_net_pf_app_clean(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
801
nfp_net_pci_unmap_mem(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
805
void nfp_net_pci_remove(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
807
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
810
devlink_unregister(priv_to_devlink(pf));
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
812
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
815
nfp_net_pf_clean_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
816
nfp_net_pf_free_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
819
nfp_net_pf_app_stop(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
82
nfp_app_vnic_free(pf->app, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
821
nfp_net_debugfs_dir_clean(&pf->ddir);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
823
nfp_devlink_params_unregister(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
827
nfp_shared_buf_unregister(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
829
nfp_net_pf_free_irqs(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
830
nfp_net_pf_app_clean(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
831
nfp_net_pci_unmap_mem(pf);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
833
cancel_work_sync(&pf->port_refresh_work);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
85
pf->num_vnics--;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
89
static void nfp_net_pf_free_vnics(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
93
list_for_each_entry_safe(nn, next, &pf->vnics, vnic_list)
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
95
nfp_net_pf_free_vnic(pf, nn);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
99
nfp_net_pf_alloc_vnic(struct nfp_pf *pf, bool needs_netdev,
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
114
nfp_warn(app->pf->cpp,
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
142
writel(vlan_tag, app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_VLAN);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
172
app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_RATE);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
193
vf_ctrl = readb(app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
196
writeb(vf_ctrl, app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
217
vf_ctrl = readb(app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
22
if (!app || !app->pf->vfcfg_tbl2)
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
220
writeb(vf_ctrl, app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
25
cap_vf = readw(app->pf->vfcfg_tbl2 + NFP_NET_VF_CFG_MB_CAP);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
251
vf_ctrl = readb(app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
254
writeb(vf_ctrl, app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
276
mac_hi = readl(app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
277
mac_lo = readw(app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_MAC_LO);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
279
flags = readb(app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_CTRL);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
28
nfp_warn(app->pf->cpp, "ndo_set_vf_%s not supported\n", msg);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
280
vlan_tag = readl(app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_VLAN);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
298
rate = readl(app->pf->vfcfg_tbl2 + vf_offset +
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
32
if (vf < 0 || vf >= app->pf->num_vfs) {
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
34
nfp_warn(app->pf->cpp, "invalid VF id %d\n", vf);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
48
writeb(vf, app->pf->vfcfg_tbl2 + NFP_NET_VF_CFG_MB_VF_NUM);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
49
writew(update, app->pf->vfcfg_tbl2 + NFP_NET_VF_CFG_MB_UPD);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
51
nn = list_first_entry(&app->pf->vnics, struct nfp_net, vnic_list);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
57
ret = readw(app->pf->vfcfg_tbl2 + NFP_NET_VF_CFG_MB_RET);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
59
nfp_warn(app->pf->cpp,
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
75
nfp_warn(app->pf->cpp,
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
83
writel(get_unaligned_be32(mac), app->pf->vfcfg_tbl2 + vf_offset);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
85
app->pf->vfcfg_tbl2 + vf_offset + NFP_NET_VF_CFG_MAC_LO);
drivers/net/ethernet/netronome/nfp/nfp_net_sriov.c
89
nfp_info(app->pf->cpp,
drivers/net/ethernet/netronome/nfp/nfp_port.c
172
int nfp_port_init_phy_port(struct nfp_pf *pf, struct nfp_app *app,
drivers/net/ethernet/netronome/nfp/nfp_port.c
176
if (!pf->eth_tbl || id >= pf->eth_tbl->count) {
drivers/net/ethernet/netronome/nfp/nfp_port.c
182
if (pf->eth_tbl->ports[id].override_changed) {
drivers/net/ethernet/netronome/nfp/nfp_port.c
185
pf->eth_tbl->ports[id].index);
drivers/net/ethernet/netronome/nfp/nfp_port.c
190
port->eth_port = &pf->eth_tbl->ports[id];
drivers/net/ethernet/netronome/nfp/nfp_port.c
191
port->eth_id = pf->eth_tbl->ports[id].index;
drivers/net/ethernet/netronome/nfp/nfp_port.c
193
if (pf->mac_stats_mem)
drivers/net/ethernet/netronome/nfp/nfp_port.c
195
pf->mac_stats_mem + port->eth_id * NFP_MAC_STATS_SIZE;
drivers/net/ethernet/netronome/nfp/nfp_port.c
214
list_add_tail(&port->port_list, &app->pf->ports);
drivers/net/ethernet/netronome/nfp/nfp_port.h
135
int nfp_port_init_phy_port(struct nfp_pf *pf, struct nfp_app *app,
drivers/net/ethernet/netronome/nfp/nfp_port.h
140
int nfp_net_refresh_port_table_sync(struct nfp_pf *pf);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
102
pf->shared_bufs = kmalloc_objs(pf->shared_bufs[0], num_entries);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
103
if (!pf->shared_bufs) {
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
109
struct nfp_shared_buf *sb = &pf->shared_bufs[i];
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
124
pf->num_shared_bufs = num_entries;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
13
static u32 nfp_shared_buf_pool_unit(struct nfp_pf *pf, unsigned int sb)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
133
le32_to_cpu(pf->shared_bufs[i].id));
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
134
kfree(pf->shared_bufs);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
140
void nfp_shared_buf_unregister(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
142
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
145
for (i = 0; i < pf->num_shared_bufs; i++)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
147
le32_to_cpu(pf->shared_bufs[i].id));
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
148
kfree(pf->shared_bufs);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
18
for (i = 0; i < pf->num_shared_bufs; i++)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
19
if (pf->shared_bufs[i].id == sb_id)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
20
return le32_to_cpu(pf->shared_bufs[i].pool_size_unit);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
26
int nfp_shared_buf_pool_get(struct nfp_pf *pf, unsigned int sb, u16 pool_index,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
37
unit_size = nfp_shared_buf_pool_unit(pf, sb);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
41
n = nfp_mbox_cmd(pf, NFP_MBOX_POOL_GET, &id, sizeof(id),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
56
int nfp_shared_buf_pool_set(struct nfp_pf *pf, unsigned int sb,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
69
unit_size = nfp_shared_buf_pool_unit(pf, sb);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
74
return nfp_mbox_cmd(pf, NFP_MBOX_POOL_SET, &set_data, sizeof(set_data),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
78
int nfp_shared_buf_register(struct nfp_pf *pf)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
80
struct devlink *devlink = priv_to_devlink(pf);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
86
if (!pf->mbox)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
89
n = nfp_pf_rtsym_read_optional(pf, NFP_SHARED_BUF_COUNT_SYM_NAME, 0);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
94
sb_desc = nfp_pf_map_rtsym(pf, "sb_tbl", NFP_SHARED_BUF_TABLE_SYM_NAME,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
95
num_entries * sizeof(pf->shared_bufs[0]),
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
614
int pf;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
627
pf = nfp->pdev->devfn & 7;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
628
nfp->iomem.csr = bar->iomem + NFP_PCIE_BAR(pf);
drivers/net/ethernet/netronome/nfp/nic/dcb.c
531
dcb->dcbcfg_tbl = nfp_pf_map_rtsym(app->pf, "net.dcbcfg_tbl",
drivers/net/ethernet/netronome/nfp/nic/main.c
13
struct nfp_pf *pf = app->pf;
drivers/net/ethernet/netronome/nfp/nic/main.c
15
if (pf->eth_tbl && pf->max_data_vnics != pf->eth_tbl->count) {
drivers/net/ethernet/netronome/nfp/nic/main.c
16
nfp_err(pf->cpp, "ETH entries don't match vNICs (%d vs %d)\n",
drivers/net/ethernet/netronome/nfp/nic/main.c
17
pf->max_data_vnics, pf->eth_tbl->count);
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
194
#define PQ_INFO_ELEMENT(vp, pf, tc, port, rl_valid, rl) \
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
195
(((vp) << 0) | ((pf) << 12) | ((tc) << 16) | ((port) << 20) | \
drivers/net/ethernet/qlogic/qed/qed_mcp.c
1922
mdump_retain.pf, mdump_retain.status);
drivers/net/ethernet/qlogic/qed/qed_mfw_hsi.h
1040
u32 pf;
drivers/net/ethernet/sfc/mcdi.c
2068
u16 pf, vf;
drivers/net/ethernet/sfc/mcdi.c
2083
pf = MCDI_DWORD(fi_outbuf, GET_FUNCTION_INFO_OUT_PF);
drivers/net/ethernet/sfc/mcdi.c
2087
PRIVILEGE_MASK_IN_FUNCTION_PF, pf,
drivers/net/ethernet/sfc/tc.c
2512
rule->fallback = &efx->tc->facts.pf;
drivers/net/ethernet/sfc/tc.c
2705
struct efx_tc_flow_rule *rule = &efx->tc->dflt.pf;
drivers/net/ethernet/sfc/tc.c
2772
struct efx_tc_action_set_list *acts = &efx->tc->facts.pf;
drivers/net/ethernet/sfc/tc.c
2913
efx_tc_deconfigure_default_rule(efx, &efx->tc->dflt.pf);
drivers/net/ethernet/sfc/tc.c
2915
efx_tc_deconfigure_fallback_acts(efx, &efx->tc->facts.pf);
drivers/net/ethernet/sfc/tc.c
3030
INIT_LIST_HEAD(&efx->tc->dflt.pf.acts.list);
drivers/net/ethernet/sfc/tc.c
3031
efx->tc->dflt.pf.fw_id = MC_CMD_MAE_ACTION_RULE_INSERT_OUT_ACTION_RULE_ID_NULL;
drivers/net/ethernet/sfc/tc.c
3034
INIT_LIST_HEAD(&efx->tc->facts.pf.list);
drivers/net/ethernet/sfc/tc.c
3035
efx->tc->facts.pf.fw_id = MC_CMD_MAE_ACTION_SET_ALLOC_OUT_ACTION_SET_ID_NULL;
drivers/net/ethernet/sfc/tc.c
3069
EFX_WARN_ON_PARANOID(efx->tc->dflt.pf.fw_id !=
drivers/net/ethernet/sfc/tc.c
3073
EFX_WARN_ON_PARANOID(efx->tc->facts.pf.fw_id !=
drivers/net/ethernet/sfc/tc.h
328
struct efx_tc_flow_rule pf;
drivers/net/ethernet/sfc/tc.h
332
struct efx_tc_action_set_list pf;
drivers/net/ethernet/sfc/tc_encap_actions.c
448
fallback = &efx->tc->facts.pf;
drivers/net/ipvlan/ipvlan_l3s.c
120
.pf = NFPROTO_IPV4,
drivers/net/ipvlan/ipvlan_l3s.c
127
.pf = NFPROTO_IPV6,
drivers/net/netdevsim/dev.c
1478
attrs.pci_vf.pf = 0;
drivers/net/ppp/ppp_generic.c
1051
static int ppp_unattached_ioctl(struct net *net, struct ppp_file *pf,
drivers/net/ppp/ppp_generic.c
106
#define PF_TO_X(pf, X) container_of(pf, X, file)
drivers/net/ppp/ppp_generic.c
108
#define PF_TO_PPP(pf) PF_TO_X(pf, struct ppp)
drivers/net/ppp/ppp_generic.c
109
#define PF_TO_CHANNEL(pf) PF_TO_X(pf, struct channel)
drivers/net/ppp/ppp_generic.c
264
static int ppp_unattached_ioctl(struct net *net, struct ppp_file *pf,
drivers/net/ppp/ppp_generic.c
289
static void init_ppp_file(struct ppp_file *pf, int kind);
drivers/net/ppp/ppp_generic.c
3415
init_ppp_file(struct ppp_file *pf, int kind)
drivers/net/ppp/ppp_generic.c
3417
pf->kind = kind;
drivers/net/ppp/ppp_generic.c
3418
skb_queue_head_init(&pf->xq);
drivers/net/ppp/ppp_generic.c
3419
skb_queue_head_init(&pf->rq);
drivers/net/ppp/ppp_generic.c
3420
refcount_set(&pf->refcnt, 1);
drivers/net/ppp/ppp_generic.c
3421
init_waitqueue_head(&pf->rwait);
drivers/net/ppp/ppp_generic.c
406
struct ppp_file *pf = file->private_data;
drivers/net/ppp/ppp_generic.c
409
if (pf) {
drivers/net/ppp/ppp_generic.c
411
if (pf->kind == INTERFACE) {
drivers/net/ppp/ppp_generic.c
412
ppp = PF_TO_PPP(pf);
drivers/net/ppp/ppp_generic.c
418
if (refcount_dec_and_test(&pf->refcnt)) {
drivers/net/ppp/ppp_generic.c
419
switch (pf->kind) {
drivers/net/ppp/ppp_generic.c
421
ppp_destroy_interface(PF_TO_PPP(pf));
drivers/net/ppp/ppp_generic.c
424
ppp_destroy_channel(PF_TO_CHANNEL(pf));
drivers/net/ppp/ppp_generic.c
435
struct ppp_file *pf = file->private_data;
drivers/net/ppp/ppp_generic.c
444
if (!pf)
drivers/net/ppp/ppp_generic.c
446
add_wait_queue(&pf->rwait, &wait);
drivers/net/ppp/ppp_generic.c
449
skb = skb_dequeue(&pf->rq);
drivers/net/ppp/ppp_generic.c
453
if (pf->dead)
drivers/net/ppp/ppp_generic.c
455
if (pf->kind == INTERFACE) {
drivers/net/ppp/ppp_generic.c
461
struct ppp *ppp = PF_TO_PPP(pf);
drivers/net/ppp/ppp_generic.c
480
remove_wait_queue(&pf->rwait, &wait);
drivers/net/ppp/ppp_generic.c
514
struct ppp_file *pf = file->private_data;
drivers/net/ppp/ppp_generic.c
518
if (!pf)
drivers/net/ppp/ppp_generic.c
524
skb = alloc_skb(count + pf->hdrlen, GFP_KERNEL);
drivers/net/ppp/ppp_generic.c
527
skb_reserve(skb, pf->hdrlen);
drivers/net/ppp/ppp_generic.c
539
switch (pf->kind) {
drivers/net/ppp/ppp_generic.c
541
ppp_xmit_process(PF_TO_PPP(pf), skb);
drivers/net/ppp/ppp_generic.c
544
skb_queue_tail(&pf->xq, skb);
drivers/net/ppp/ppp_generic.c
545
ppp_channel_push(PF_TO_CHANNEL(pf));
drivers/net/ppp/ppp_generic.c
558
struct ppp_file *pf = file->private_data;
drivers/net/ppp/ppp_generic.c
561
if (!pf)
drivers/net/ppp/ppp_generic.c
563
poll_wait(file, &pf->rwait, wait);
drivers/net/ppp/ppp_generic.c
565
if (skb_peek(&pf->rq))
drivers/net/ppp/ppp_generic.c
567
if (pf->dead)
drivers/net/ppp/ppp_generic.c
569
else if (pf->kind == INTERFACE) {
drivers/net/ppp/ppp_generic.c
571
struct ppp *ppp = PF_TO_PPP(pf);
drivers/net/ppp/ppp_generic.c
723
struct ppp_file *pf;
drivers/net/ppp/ppp_generic.c
736
pf = file->private_data;
drivers/net/ppp/ppp_generic.c
737
if (!pf) {
drivers/net/ppp/ppp_generic.c
739
pf, file, cmd, arg);
drivers/net/ppp/ppp_generic.c
755
if (pf->kind == CHANNEL) {
drivers/net/ppp/ppp_generic.c
760
pch = PF_TO_CHANNEL(pf);
drivers/net/ppp/ppp_generic.c
809
if (pf->kind != INTERFACE) {
drivers/net/ppp/ppp_generic.c
816
ppp = PF_TO_PPP(pf);
drivers/net/ppp/ppp_generic.c
990
struct ppp_file *pf;
drivers/net/ppp/ppp_generic.c
996
pf = file->private_data;
drivers/net/ppp/ppp_generic.c
997
if (pf && pf->kind == INTERFACE) {
drivers/net/ppp/ppp_generic.c
998
struct ppp *ppp = PF_TO_PPP(pf);
drivers/net/vrf.c
1175
static struct sk_buff *vrf_rcv_nfhook(u8 pf, unsigned int hook,
drivers/net/vrf.c
1181
if (nf_hook(pf, hook, net, NULL, skb, dev, NULL, vrf_rcv_finish) != 1)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9364
u32 p0 = 0, p1 = 0, pf = 0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9378
p0 = pf;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9394
p0, p1, pf, idx0, idx1, ibit);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9396
if (bidx != 5 && pf <= p0 && pf < p1) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9399
pf = p0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9402
pf = p1;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9406
iorq, iorq, idxf[iorq], pf);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9418
u32 p0 = 0, p1 = 0, pf = 0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9444
p0 = pf;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9470
p0, p1, pf, idx0, idx1, iq_err, gop, ibit);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9472
if (!(!first_search && pf <= p0 && pf < p1)) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9474
pf = p0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9477
pf = p1;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9494
pf, gerr & 0x0F, perr & 0x3F);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9524
pf = p1;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9527
} else if (pf > p1) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9528
pf = p1;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
9533
p1, pf, gef & 0x0F, pef & 0x3F);
drivers/pci/controller/dwc/pcie-designware.h
177
#define PCIE_ATU_FUNC_NUM(pf) ((pf) << 20)
drivers/pci/iov.c
43
struct pci_dev *pf;
drivers/pci/iov.c
48
pf = pci_physfn(dev);
drivers/pci/iov.c
49
return (pci_dev_id(dev) - (pci_dev_id(pf) + pf->sriov->offset)) /
drivers/pci/iov.c
50
pf->sriov->stride;
drivers/pci/tsm.c
100
pci_iov_virtfn_bus(pf, j),
drivers/pci/tsm.c
101
pci_iov_virtfn_devfn(pf, j));
drivers/pci/tsm.c
129
struct pci_dev *pf __free(pci_dev_put) = pci_get_slot(
drivers/pci/tsm.c
132
if (!pf)
drivers/pci/tsm.c
136
for (int j = pci_num_vf(pf) - 1; j >= 0; j--) {
drivers/pci/tsm.c
139
pci_domain_nr(pf->bus),
drivers/pci/tsm.c
140
pci_iov_virtfn_bus(pf, j),
drivers/pci/tsm.c
141
pci_iov_virtfn_devfn(pf, j));
drivers/pci/tsm.c
150
cb(pf, data);
drivers/pci/tsm.c
85
struct pci_dev *pf __free(pci_dev_put) = pci_get_slot(
drivers/pci/tsm.c
88
if (!pf)
drivers/pci/tsm.c
93
cb(pf, data);
drivers/pci/tsm.c
96
for (int j = 0; j < pci_num_vf(pf); j++) {
drivers/pci/tsm.c
99
pci_domain_nr(pf->bus),
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
158
static void mtk_hw_bits_part(struct mtk_pin_field *pf, int *h, int *l)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
160
*l = 32 - pf->bitpos;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
161
*h = get_count_order(pf->mask) - *l;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
165
struct mtk_pin_field *pf, int value)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
169
mtk_hw_bits_part(pf, &nbits_h, &nbits_l);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
171
mtk_rmw(hw, pf->index, pf->offset, pf->mask << pf->bitpos,
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
172
(value & pf->mask) << pf->bitpos);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
174
mtk_rmw(hw, pf->index, pf->offset + pf->next, BIT(nbits_h) - 1,
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
175
(value & pf->mask) >> nbits_l);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
179
struct mtk_pin_field *pf, int *value)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
183
mtk_hw_bits_part(pf, &nbits_h, &nbits_l);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
185
l = (mtk_r32(hw, pf->index, pf->offset)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
186
>> pf->bitpos) & (BIT(nbits_l) - 1);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
187
h = (mtk_r32(hw, pf->index, pf->offset + pf->next))
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
196
struct mtk_pin_field pf;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
199
err = mtk_hw_pin_field_get(hw, desc, field, &pf);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
203
if (value < 0 || value > pf.mask)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
206
if (!pf.next)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
207
mtk_rmw(hw, pf.index, pf.offset, pf.mask << pf.bitpos,
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
208
(value & pf.mask) << pf.bitpos);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
210
mtk_hw_write_cross_field(hw, &pf, value);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
219
struct mtk_pin_field pf;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
222
err = mtk_hw_pin_field_get(hw, desc, field, &pf);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
226
if (!pf.next)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
227
*value = (mtk_r32(hw, pf.index, pf.offset)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
228
>> pf.bitpos) & pf.mask;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
230
mtk_hw_read_cross_field(hw, &pf, value);
drivers/pinctrl/pinctrl-pic32.c
1773
const struct pic32_function *pf = &pctl->functions[func];
drivers/pinctrl/pinctrl-pic32.c
1774
const char *fname = pf->name;
drivers/pinctrl/pinctrl-pistachio.c
949
const struct pistachio_function *pf = &pctl->functions[func];
drivers/pinctrl/pinctrl-pistachio.c
970
if (pf->scenarios) {
drivers/pinctrl/pinctrl-pistachio.c
971
for (i = 0; i < pf->nscenarios; i++) {
drivers/pinctrl/pinctrl-pistachio.c
972
if (pf->scenarios[i] == group)
drivers/pinctrl/pinctrl-pistachio.c
975
if (WARN_ON(i == pf->nscenarios))
drivers/pinctrl/pinctrl-pistachio.c
978
val = pctl_readl(pctl, pf->scenario_reg);
drivers/pinctrl/pinctrl-pistachio.c
979
val &= ~(pf->scenario_mask << pf->scenario_shift);
drivers/pinctrl/pinctrl-pistachio.c
980
val |= i << pf->scenario_shift;
drivers/pinctrl/pinctrl-pistachio.c
981
pctl_writel(pctl, val, pf->scenario_reg);
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
112
struct pinfunction *pf = pctl->functions + function;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
114
return pf->name;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
130
struct pinfunction *pf = pctl->functions + function;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
132
*groups = pf->groups;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
133
*num_groups = pf->ngroups;
drivers/regulator/qcom_rpm-regulator.c
111
.pf = { 0, 0xF0000000, 28 },
drivers/regulator/qcom_rpm-regulator.c
122
.pf = { 0, 0xF0000000, 28 },
drivers/regulator/qcom_rpm-regulator.c
136
.pf = { 0, 0x000003C0, 6 },
drivers/regulator/qcom_rpm-regulator.c
35
struct request_member pf; /* pin function */
drivers/regulator/qcom_rpm-regulator.c
71
.pf = { 0, 0xC0000000, 30 },
drivers/regulator/qcom_rpm-regulator.c
82
.pf = { 0, 0xC0000000, 30 },
drivers/regulator/qcom_rpm-regulator.c
94
.pf = { 0, 0x000000C0, 6 },
drivers/scsi/csiostor/csio_hw.c
2682
u32 pf = 0;
drivers/scsi/csiostor/csio_hw.c
2686
pf = SOURCEPF_G(csio_rd_reg32(hw, PL_WHOAMI_A));
drivers/scsi/csiostor/csio_hw.c
2688
pf = T6_SOURCEPF_G(csio_rd_reg32(hw, PL_WHOAMI_A));
drivers/scsi/csiostor/csio_hw.c
2723
csio_set_reg_field(hw, PL_INT_MAP0_A, 0, 1 << pf);
drivers/scsi/csiostor/csio_hw.c
2739
u32 pf = 0;
drivers/scsi/csiostor/csio_hw.c
2742
pf = SOURCEPF_G(csio_rd_reg32(hw, PL_WHOAMI_A));
drivers/scsi/csiostor/csio_hw.c
2744
pf = T6_SOURCEPF_G(csio_rd_reg32(hw, PL_WHOAMI_A));
drivers/scsi/csiostor/csio_hw.c
2753
csio_set_reg_field(hw, PL_INT_MAP0_A, 1 << pf, 0);
drivers/scsi/csiostor/csio_mb.c
194
unsigned int pf, unsigned int vf, unsigned int nparams,
drivers/scsi/csiostor/csio_mb.c
208
FW_PARAMS_CMD_PFN_V(pf) |
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2250
cdev->pfvf = FW_PFVF_CMD_PFN_V(lldi->pf);
drivers/scsi/elx/libefc/efc_node.c
902
void (*pf)(struct efc_sm_ctx *ctx,
drivers/scsi/elx/libefc/efc_node.c
905
pf = node->nodedb_state;
drivers/scsi/elx/libefc/efc_node.c
908
efc_node_transition(node, pf, NULL);
drivers/scsi/qedi/qedi_main.c
2016
u8 pf;
drivers/scsi/qedi/qedi_main.c
2020
pf = qedi->dev_info.common.abs_pf_id;
drivers/scsi/qedi/qedi_main.c
2027
(pf == (block->id & NVM_ISCSI_CFG_BLK_MAPPED_PF_ID_MASK)
drivers/scsi/scsi_debug.c
3166
int pf, sp, ps, md_len, bd_len, off, spf, pg_len;
drivers/scsi/scsi_debug.c
3173
pf = cmd[1] & 0x10;
drivers/scsi/scsi_debug.c
3176
if ((0 == pf) || sp || (param_len > SDEBUG_MAX_MSELECT_SZ)) {
drivers/scsi/scsi_lib.c
2265
int scsi_mode_select(struct scsi_device *sdev, int pf, int sp,
drivers/scsi/scsi_lib.c
2277
cmd[1] = (pf ? 0x10 : 0) | (sp ? 0x01 : 0);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1439
(csk->com.cdev->lldi.pf));
drivers/target/target_core_spc.c
1169
bool pf = !!(cdb[1] & 0x10);
drivers/target/target_core_spc.c
1189
if (!pf) {
drivers/vdpa/octeon_ep/octep_vdpa_main.c
705
struct octep_pf *pf = pci_get_drvdata(pf_dev);
drivers/vdpa/octeon_ep/octep_vdpa_main.c
712
bus_region.start = pf->vf_base + idx * pf->vf_stride;
drivers/vdpa/octeon_ep/octep_vdpa_main.c
713
bus_region.end = bus_region.start + pf->vf_stride - 1;
drivers/vdpa/octeon_ep/octep_vdpa_main.c
719
struct octep_pf *pf = pci_get_drvdata(pdev);
drivers/vdpa/octeon_ep/octep_vdpa_main.c
720
u8 __iomem *addr = pf->base[OCTEP_HW_MBOX_BAR];
drivers/vdpa/octeon_ep/octep_vdpa_main.c
730
pf->enabled_vfs = num_vfs;
drivers/vdpa/octeon_ep/octep_vdpa_main.c
733
if (vf_pdev->device != pf->vf_devid)
drivers/vdpa/octeon_ep/octep_vdpa_main.c
745
for (i = 0; i < pf->enabled_vfs; i++)
drivers/vdpa/octeon_ep/octep_vdpa_main.c
754
struct octep_pf *pf = pci_get_drvdata(pdev);
drivers/vdpa/octeon_ep/octep_vdpa_main.c
760
pf->enabled_vfs = 0;
fs/affs/symlink.c
36
char *pf;
fs/affs/symlink.c
38
pf = sbi->s_prefix ? sbi->s_prefix : "/";
fs/affs/symlink.c
39
while (i < 1023 && (c = pf[i]))
fs/erofs/xattr.c
151
struct erofs_xattr_prefix_item *pf = sbi->xattr_prefixes +
fs/erofs/xattr.c
154
if (pf >= sbi->xattr_prefixes + sbi->xattr_prefix_count)
fs/erofs/xattr.c
156
infix = pf->prefix->infix;
fs/erofs/xattr.c
157
infix_len = pf->infix_len;
fs/erofs/xattr.c
158
base_index = pf->prefix->base_index;
fs/erofs/xattr.c
202
struct erofs_xattr_prefix_item *pf = sbi->xattr_prefixes +
fs/erofs/xattr.c
205
if (pf >= sbi->xattr_prefixes + sbi->xattr_prefix_count)
fs/erofs/xattr.c
208
if (it->index != pf->prefix->base_index ||
fs/erofs/xattr.c
209
it->name.len != entry.e_name_len + pf->infix_len)
fs/erofs/xattr.c
212
if (memcmp(it->name.name, pf->prefix->infix, pf->infix_len))
fs/erofs/xattr.c
215
it->infix_len = pf->infix_len;
fs/erofs/xattr.c
533
struct erofs_xattr_prefix_item *pf = pfs + sbi->ishare_xattr_prefix_id;
fs/erofs/xattr.c
536
newpfx = krealloc(pf->prefix,
fs/erofs/xattr.c
537
sizeof(*newpfx) + pf->infix_len + 1, GFP_KERNEL);
fs/erofs/xattr.c
539
newpfx->infix[pf->infix_len] = '\0';
fs/erofs/xattr.c
540
pf->prefix = newpfx;
fs/nfsd/nfs4xdr.c
5066
nfsd4_encode_secinfo4(struct xdr_stream *xdr, rpc_authflavor_t pf,
fs/nfsd/nfs4xdr.c
5072
if (rpcauth_get_gssinfo(pf, &info) == 0) {
fs/nfsd/nfs4xdr.c
5083
} else if (pf < RPC_AUTH_MAXFLAVOR) {
fs/nfsd/nfs4xdr.c
5087
status = nfsd4_encode_uint32_t(xdr, pf);
include/linux/iio/trigger_consumer.h
48
void iio_dealloc_pollfunc(struct iio_poll_func *pf);
include/linux/net.h
362
#define MODULE_ALIAS_NET_PF_PROTO(pf, proto) \
include/linux/net.h
363
MODULE_ALIAS("net-pf-" __stringify(pf) "-proto-" __stringify(proto))
include/linux/net.h
365
#define MODULE_ALIAS_NET_PF_PROTO_TYPE(pf, proto, type) \
include/linux/net.h
366
MODULE_ALIAS("net-pf-" __stringify(pf) "-proto-" __stringify(proto) \
include/linux/net.h
369
#define MODULE_ALIAS_NET_PF_PROTO_NAME(pf, proto, name) \
include/linux/net.h
370
MODULE_ALIAS("net-pf-" __stringify(pf) "-proto-" __stringify(proto) \
include/linux/net/intel/i40e_client.h
83
void *pf;
include/linux/netfilter.h
106
u8 pf;
include/linux/netfilter.h
163
u_int8_t pf,
include/linux/netfilter.h
171
p->pf = pf;
include/linux/netfilter.h
184
u_int8_t pf;
include/linux/netfilter.h
227
static inline int nf_hook(u_int8_t pf, unsigned int hook, struct net *net,
include/linux/netfilter.h
236
if (__builtin_constant_p(pf) &&
include/linux/netfilter.h
238
!static_key_false(&nf_hooks_needed[pf][hook]))
include/linux/netfilter.h
243
switch (pf) {
include/linux/netfilter.h
270
nf_hook_state_init(&state, hook, pf, indev, outdev,
include/linux/netfilter.h
298
NF_HOOK_COND(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk,
include/linux/netfilter.h
306
((ret = nf_hook(pf, hook, net, sk, skb, in, out, okfn)) == 1))
include/linux/netfilter.h
312
NF_HOOK(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk, struct sk_buff *skb,
include/linux/netfilter.h
316
int ret = nf_hook(pf, hook, net, sk, skb, in, out, okfn);
include/linux/netfilter.h
323
NF_HOOK_LIST(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk,
include/linux/netfilter.h
330
if (__builtin_constant_p(pf) &&
include/linux/netfilter.h
332
!static_key_false(&nf_hooks_needed[pf][hook]))
include/linux/netfilter.h
337
switch (pf) {
include/linux/netfilter.h
352
nf_hook_state_init(&state, hook, pf, in, out, sk, net, okfn);
include/linux/netfilter.h
360
int nf_setsockopt(struct sock *sk, u_int8_t pf, int optval, sockptr_t opt,
include/linux/netfilter.h
362
int nf_getsockopt(struct sock *sk, u_int8_t pf, int optval, char __user *opt,
include/linux/netfilter.h
409
NF_HOOK_COND(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk,
include/linux/netfilter.h
418
NF_HOOK(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk,
include/linux/netfilter.h
426
NF_HOOK_LIST(uint8_t pf, unsigned int hook, struct net *net, struct sock *sk,
include/linux/netfilter.h
433
static inline int nf_hook(u_int8_t pf, unsigned int hook, struct net *net,
include/linux/netfilter.h
80
u8 pf;
include/linux/netfilter/x_tables.h
66
return par->state->pf;
include/linux/pci.h
1029
u32 (*sriov_get_vf_total_msix)(struct pci_dev *pf);
include/linux/pds/pds_common.h
54
int pds_client_register(struct pdsc *pf, char *devname);
include/linux/pds/pds_common.h
55
int pds_client_unregister(struct pdsc *pf, u16 client_id);
include/net/devlink.h
1809
u16 pf, bool external);
include/net/devlink.h
1811
u16 pf, u16 vf, bool external);
include/net/devlink.h
1813
u32 controller, u16 pf, u32 sf,
include/net/devlink.h
43
u16 pf;
include/net/devlink.h
57
u16 pf;
include/net/devlink.h
72
u16 pf;
include/net/lapb.h
76
unsigned char pf; /* Poll/Final */
include/net/netfilter/nf_log.h
40
u_int8_t pf,
include/net/netfilter/nf_log.h
59
int nf_log_register(u_int8_t pf, struct nf_logger *logger);
include/net/netfilter/nf_log.h
63
bool nf_log_is_registered(u_int8_t pf);
include/net/netfilter/nf_log.h
65
int nf_log_set(struct net *net, u_int8_t pf, const struct nf_logger *logger);
include/net/netfilter/nf_log.h
68
int nf_log_bind_pf(struct net *net, u_int8_t pf,
include/net/netfilter/nf_log.h
70
void nf_log_unbind_pf(struct net *net, u_int8_t pf);
include/net/netfilter/nf_log.h
72
int nf_logger_find_get(int pf, enum nf_log_type type);
include/net/netfilter/nf_log.h
73
void nf_logger_put(int pf, enum nf_log_type type);
include/net/netfilter/nf_log.h
81
u_int8_t pf,
include/net/netfilter/nf_log.h
91
u_int8_t pf,
include/net/netfilter/nf_nat.h
71
int nf_nat_register_fn(struct net *net, u8 pf, const struct nf_hook_ops *ops,
include/net/netfilter/nf_nat.h
73
void nf_nat_unregister_fn(struct net *net, u8 pf, const struct nf_hook_ops *ops,
include/net/netfilter/nf_tables.h
61
return pkt->state->pf;
include/net/sctp/sctp.h
576
overhead += sp->pf->af->net_header_len;
include/net/sctp/structs.h
156
struct sctp_pf *pf;
include/scsi/scsi_device.h
470
extern int scsi_mode_select(struct scsi_device *sdev, int pf, int sp,
include/uapi/linux/bpf.h
1829
__u32 pf;
include/uapi/linux/bpf.h
6788
__u32 pf;
include/uapi/linux/netfilter/nfnetlink_queue.h
88
__be16 pf; /* AF_xxx for PF_[UN]BIND */
kernel/locking/lockdep.c
1073
struct pending_free *pf;
kernel/locking/lockdep.c
1076
for (i = 0, pf = delayed_free.pf; i < ARRAY_SIZE(delayed_free.pf); i++, pf++) {
kernel/locking/lockdep.c
1077
if (in_list(&class->lock_entry, &pf->zapped))
kernel/locking/lockdep.c
1202
INIT_LIST_HEAD(&delayed_free.pf[0].zapped);
kernel/locking/lockdep.c
1203
INIT_LIST_HEAD(&delayed_free.pf[1].zapped);
kernel/locking/lockdep.c
402
struct pending_free pf[2];
kernel/locking/lockdep.c
6190
static void remove_class_from_lock_chain(struct pending_free *pf,
kernel/locking/lockdep.c
6220
__set_bit(chain - lock_chains, pf->lock_chains_being_freed);
kernel/locking/lockdep.c
6226
static void remove_class_from_lock_chains(struct pending_free *pf,
kernel/locking/lockdep.c
6236
remove_class_from_lock_chain(pf, chain, class);
kernel/locking/lockdep.c
6244
static void zap_class(struct pending_free *pf, struct lock_class *class)
kernel/locking/lockdep.c
6265
list_move_tail(&class->lock_entry, &pf->zapped);
kernel/locking/lockdep.c
6281
remove_class_from_lock_chains(pf, class);
kernel/locking/lockdep.c
6309
return delayed_free.pf + delayed_free.index;
kernel/locking/lockdep.c
6320
static bool prepare_call_rcu_zapped(struct pending_free *pf)
kernel/locking/lockdep.c
6324
if (list_empty(&pf->zapped))
kernel/locking/lockdep.c
6332
WARN_ON_ONCE(delayed_free.pf + delayed_free.index != pf);
kernel/locking/lockdep.c
6339
static void __free_zapped_classes(struct pending_free *pf)
kernel/locking/lockdep.c
6345
list_for_each_entry(class, &pf->zapped, lock_entry)
kernel/locking/lockdep.c
6348
list_splice_init(&pf->zapped, &free_lock_classes);
kernel/locking/lockdep.c
6352
pf->lock_chains_being_freed, ARRAY_SIZE(lock_chains));
kernel/locking/lockdep.c
6353
bitmap_clear(pf->lock_chains_being_freed, 0, ARRAY_SIZE(lock_chains));
kernel/locking/lockdep.c
6359
struct pending_free *pf;
kernel/locking/lockdep.c
6370
pf = delayed_free.pf + (delayed_free.index ^ 1);
kernel/locking/lockdep.c
6371
__free_zapped_classes(pf);
kernel/locking/lockdep.c
6374
prepare_call_rcu_zapped(delayed_free.pf + delayed_free.index);
kernel/locking/lockdep.c
6393
static void __lockdep_free_key_range(struct pending_free *pf, void *start,
kernel/locking/lockdep.c
6407
zap_class(pf, class);
kernel/locking/lockdep.c
6422
struct pending_free *pf;
kernel/locking/lockdep.c
6430
pf = get_pending_free();
kernel/locking/lockdep.c
6431
__lockdep_free_key_range(pf, start, size);
kernel/locking/lockdep.c
6432
need_callback = prepare_call_rcu_zapped(pf);
kernel/locking/lockdep.c
6450
struct pending_free *pf = delayed_free.pf;
kernel/locking/lockdep.c
6457
__lockdep_free_key_range(pf, start, size);
kernel/locking/lockdep.c
6458
__free_zapped_classes(pf);
kernel/locking/lockdep.c
6496
static void __lockdep_reset_lock(struct pending_free *pf,
kernel/locking/lockdep.c
6511
zap_class(pf, class);
kernel/locking/lockdep.c
6527
struct pending_free *pf;
kernel/locking/lockdep.c
6537
pf = get_pending_free();
kernel/locking/lockdep.c
6538
__lockdep_reset_lock(pf, lock);
kernel/locking/lockdep.c
6539
need_callback = prepare_call_rcu_zapped(pf);
kernel/locking/lockdep.c
6554
struct pending_free *pf = delayed_free.pf;
kernel/locking/lockdep.c
6559
__lockdep_reset_lock(pf, lock);
kernel/locking/lockdep.c
6560
__free_zapped_classes(pf);
kernel/locking/lockdep.c
6586
struct pending_free *pf;
kernel/locking/lockdep.c
6608
pf = get_pending_free();
kernel/locking/lockdep.c
6609
__lockdep_free_key_range(pf, key, 1);
kernel/locking/lockdep.c
6610
need_callback = prepare_call_rcu_zapped(pf);
kernel/trace/trace_events_inject.c
37
struct ftrace_event_field **pf, u64 *pv)
kernel/trace/trace_events_inject.c
65
*pf = field;
net/ax25/ax25_ds_in.c
100
static int ax25_ds_state2_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int pf, int type)
net/ax25/ax25_ds_in.c
110
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
117
if (pf) {
net/ax25/ax25_ds_in.c
127
if (pf) {
net/ax25/ax25_ds_in.c
145
static int ax25_ds_state3_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int ns, int nr, int pf, int type)
net/ax25/ax25_ds_in.c
159
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
172
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
192
if (type == AX25_COMMAND && pf)
net/ax25/ax25_ds_in.c
213
if (type == AX25_COMMAND && pf)
net/ax25/ax25_ds_in.c
235
if (pf) ax25_ds_enquiry_response(ax25);
net/ax25/ax25_ds_in.c
244
if (pf) {
net/ax25/ax25_ds_in.c
254
if (pf) ax25_ds_enquiry_response(ax25);
net/ax25/ax25_ds_in.c
281
int queued = 0, frametype, ns, nr, pf;
net/ax25/ax25_ds_in.c
283
frametype = ax25_decode(ax25, skb, &ns, &nr, &pf);
net/ax25/ax25_ds_in.c
287
queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type);
net/ax25/ax25_ds_in.c
290
queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type);
net/ax25/ax25_ds_in.c
293
queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type);
net/ax25/ax25_ds_in.c
32
static int ax25_ds_state1_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int pf, int type)
net/ax25/ax25_ds_in.c
38
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
44
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
48
ax25_send_control(ax25, AX25_DM, pf, AX25_RESPONSE);
net/ax25/ax25_ds_in.c
82
if (pf)
net/ax25/ax25_ds_in.c
87
if (pf)
net/ax25/ax25_std_in.c
103
static int ax25_std_state2_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int pf, int type)
net/ax25/ax25_std_in.c
108
ax25_send_control(ax25, AX25_DM, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
112
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
118
if (pf)
net/ax25/ax25_std_in.c
126
if (pf) ax25_send_control(ax25, AX25_DM, AX25_POLLON, AX25_RESPONSE);
net/ax25/ax25_std_in.c
141
static int ax25_std_state3_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int ns, int nr, int pf, int type)
net/ax25/ax25_std_in.c
155
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
168
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
182
if (type == AX25_COMMAND && pf)
net/ax25/ax25_std_in.c
194
if (type == AX25_COMMAND && pf)
net/ax25/ax25_std_in.c
220
if (pf) ax25_std_enquiry_response(ax25);
net/ax25/ax25_std_in.c
229
if (pf) {
net/ax25/ax25_std_in.c
239
if (pf) ax25_std_enquiry_response(ax25);
net/ax25/ax25_std_in.c
242
ax25_send_control(ax25, AX25_REJ, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
266
static int ax25_std_state4_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int ns, int nr, int pf, int type)
net/ax25/ax25_std_in.c
280
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
295
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
309
if (type == AX25_RESPONSE && pf) {
net/ax25/ax25_std_in.c
326
if (type == AX25_COMMAND && pf)
net/ax25/ax25_std_in.c
338
if (pf && type == AX25_RESPONSE) {
net/ax25/ax25_std_in.c
355
if (type == AX25_COMMAND && pf)
net/ax25/ax25_std_in.c
374
if (pf)
net/ax25/ax25_std_in.c
384
if (pf) {
net/ax25/ax25_std_in.c
39
static int ax25_std_state1_machine(ax25_cb *ax25, struct sk_buff *skb, int frametype, int pf, int type)
net/ax25/ax25_std_in.c
394
if (pf) ax25_std_enquiry_response(ax25);
net/ax25/ax25_std_in.c
397
ax25_send_control(ax25, AX25_REJ, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
421
int queued = 0, frametype, ns, nr, pf;
net/ax25/ax25_std_in.c
423
frametype = ax25_decode(ax25, skb, &ns, &nr, &pf);
net/ax25/ax25_std_in.c
427
queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type);
net/ax25/ax25_std_in.c
430
queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type);
net/ax25/ax25_std_in.c
433
queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type);
net/ax25/ax25_std_in.c
436
queued = ax25_std_state4_machine(ax25, skb, frametype, ns, nr, pf, type);
net/ax25/ax25_std_in.c
45
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
51
ax25_send_control(ax25, AX25_UA, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
55
ax25_send_control(ax25, AX25_DM, pf, AX25_RESPONSE);
net/ax25/ax25_std_in.c
59
if (pf) {
net/ax25/ax25_std_in.c
81
if (pf) {
net/ax25/ax25_subr.c
103
*ns = *nr = *pf = 0;
net/ax25/ax25_subr.c
110
*pf = frame[0] & AX25_PF;
net/ax25/ax25_subr.c
114
*pf = frame[0] & AX25_PF;
net/ax25/ax25_subr.c
117
*pf = frame[0] & AX25_PF;
net/ax25/ax25_subr.c
125
*pf = frame[1] & AX25_EPF;
net/ax25/ax25_subr.c
130
*pf = frame[1] & AX25_EPF;
net/ax25/ax25_subr.c
134
*pf = frame[0] & AX25_PF;
net/ax25/ax25_subr.c
97
int ax25_decode(ax25_cb *ax25, struct sk_buff *skb, int *ns, int *nr, int *pf)
net/bluetooth/rfcomm/core.c
1705
static int rfcomm_recv_data(struct rfcomm_session *s, u8 dlci, int pf, struct sk_buff *skb)
net/bluetooth/rfcomm/core.c
1709
BT_DBG("session %p state %ld dlci %d pf %d", s, s->state, dlci, pf);
net/bluetooth/rfcomm/core.c
1717
if (pf && d->cfc) {
net/bluetooth/rfcomm/core.c
86
#define __ctrl(type, pf) (((type & 0xef) | (pf << 4)))
net/bpf/test_run.c
1665
switch (user->pf) {
net/bpf/test_run.c
1692
state->pf = user->pf;
net/bpf/test_run.c
1717
.pf = NFPROTO_IPV4,
net/bpf/test_run.c
1777
if (hook_state.pf == NFPROTO_IPV4)
net/bpf/test_run.c
1783
if (hook_state.pf == NFPROTO_IPV6)
net/bpf/test_run.c
1793
skb->protocol = nfproto_eth(hook_state.pf);
net/bridge/br_netfilter_hooks.c
1052
.pf = NFPROTO_BRIDGE,
net/bridge/br_netfilter_hooks.c
1059
.pf = NFPROTO_BRIDGE,
net/bridge/br_netfilter_hooks.c
1066
.pf = NFPROTO_BRIDGE,
net/bridge/br_netfilter_hooks.c
1072
.pf = NFPROTO_BRIDGE,
net/bridge/br_netfilter_hooks.c
1078
.pf = NFPROTO_IPV4,
net/bridge/br_netfilter_hooks.c
1084
.pf = NFPROTO_IPV6,
net/bridge/br_netfilter_hooks.c
670
u8 pf)
net/bridge/br_netfilter_hooks.c
699
if (pf == NFPROTO_IPV4) {
net/bridge/br_netfilter_hooks.c
704
} else if (pf == NFPROTO_IPV6) {
net/bridge/br_netfilter_hooks.c
716
NF_HOOK(pf, NF_INET_FORWARD, state->net, NULL, skb,
net/bridge/br_netfilter_hooks.c
933
u_int8_t pf;
net/bridge/br_netfilter_hooks.c
948
pf = NFPROTO_IPV4;
net/bridge/br_netfilter_hooks.c
951
pf = NFPROTO_IPV6;
net/bridge/br_netfilter_hooks.c
961
if (pf == NFPROTO_IPV4)
net/bridge/br_netfilter_hooks.c
966
NF_HOOK(pf, NF_INET_POST_ROUTING, state->net, state->sk, skb,
net/bridge/netfilter/ebt_log.c
74
ebt_log_packet(struct net *net, u_int8_t pf, unsigned int hooknum,
net/bridge/netfilter/ebtable_broute.c
88
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_filter.c
56
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_filter.c
62
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_filter.c
68
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_nat.c
56
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_nat.c
62
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtable_nat.c
68
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/ebtables.c
2544
.pf = PF_INET,
net/bridge/netfilter/nf_conntrack_bridge.c
265
bridge_state.pf = NFPROTO_IPV4;
net/bridge/netfilter/nf_conntrack_bridge.c
279
bridge_state.pf = NFPROTO_IPV6;
net/bridge/netfilter/nf_conntrack_bridge.c
413
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/nf_conntrack_bridge.c
419
.pf = NFPROTO_BRIDGE,
net/bridge/netfilter/nf_conntrack_bridge.c
425
.pf = NFPROTO_BRIDGE,
net/core/pktgen.c
185
pf(IPV6) /* Interface in IPV6 Mode */ \
net/core/pktgen.c
186
pf(IPSRC_RND) /* IP-Src Random */ \
net/core/pktgen.c
187
pf(IPDST_RND) /* IP-Dst Random */ \
net/core/pktgen.c
188
pf(TXSIZE_RND) /* Transmit size is random */ \
net/core/pktgen.c
189
pf(UDPSRC_RND) /* UDP-Src Random */ \
net/core/pktgen.c
190
pf(UDPDST_RND) /* UDP-Dst Random */ \
net/core/pktgen.c
191
pf(UDPCSUM) /* Include UDP checksum */ \
net/core/pktgen.c
192
pf(NO_TIMESTAMP) /* Don't timestamp packets (default TS) */ \
net/core/pktgen.c
193
pf(MPLS_RND) /* Random MPLS labels */ \
net/core/pktgen.c
194
pf(QUEUE_MAP_RND) /* queue map Random */ \
net/core/pktgen.c
195
pf(QUEUE_MAP_CPU) /* queue map mirrors smp_processor_id() */ \
net/core/pktgen.c
196
pf(FLOW_SEQ) /* Sequential flows */ \
net/core/pktgen.c
197
pf(IPSEC) /* ipsec on for flows */ \
net/core/pktgen.c
198
pf(MACSRC_RND) /* MAC-Src Random */ \
net/core/pktgen.c
199
pf(MACDST_RND) /* MAC-Dst Random */ \
net/core/pktgen.c
200
pf(VID_RND) /* Random VLAN ID */ \
net/core/pktgen.c
201
pf(SVID_RND) /* Random SVLAN ID */ \
net/core/pktgen.c
202
pf(NODE) /* Node memory alloc*/ \
net/core/pktgen.c
203
pf(SHARED) /* Shared SKB */ \
net/devlink/port.c
1378
u16 pf, bool external)
net/devlink/port.c
1386
attrs->pci_pf.pf = pf;
net/devlink/port.c
1402
u16 pf, u16 vf, bool external)
net/devlink/port.c
1410
attrs->pci_vf.pf = pf;
net/devlink/port.c
1426
u16 pf, u32 sf, bool external)
net/devlink/port.c
1434
attrs->pci_sf.pf = pf;
net/devlink/port.c
1536
n = snprintf(name, len, "pf%u", attrs->pci_pf.pf);
net/devlink/port.c
1547
attrs->pci_vf.pf, attrs->pci_vf.vf);
net/devlink/port.c
1557
n = snprintf(name, len, "pf%usf%u", attrs->pci_sf.pf,
net/devlink/port.c
247
nla_put_u16(msg, DEVLINK_ATTR_PORT_PCI_PF_NUMBER, attrs->pci_pf.pf))
net/devlink/port.c
255
nla_put_u16(msg, DEVLINK_ATTR_PORT_PCI_PF_NUMBER, attrs->pci_vf.pf) ||
net/devlink/port.c
265
attrs->pci_sf.pf) ||
net/ipv4/netfilter/arp_tables.c
1610
.pf = PF_INET,
net/ipv4/netfilter/ip_tables.c
1828
.pf = PF_INET,
net/ipv4/netfilter/iptable_nat.c
35
.pf = NFPROTO_IPV4,
net/ipv4/netfilter/iptable_nat.c
41
.pf = NFPROTO_IPV4,
net/ipv4/netfilter/iptable_nat.c
47
.pf = NFPROTO_IPV4,
net/ipv4/netfilter/iptable_nat.c
53
.pf = NFPROTO_IPV4,
net/ipv4/netfilter/nf_defrag_ipv4.c
102
.pf = NFPROTO_IPV4,
net/ipv4/netfilter/nf_defrag_ipv4.c
96
.pf = NFPROTO_IPV4,
net/ipv6/ila/ila_xlat.c
192
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/ip6_tables.c
1835
.pf = PF_INET6,
net/ipv6/netfilter/ip6table_nat.c
37
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/ip6table_nat.c
43
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/ip6table_nat.c
49
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/ip6table_nat.c
55
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/nf_defrag_ipv6_hooks.c
79
.pf = NFPROTO_IPV6,
net/ipv6/netfilter/nf_defrag_ipv6_hooks.c
85
.pf = NFPROTO_IPV6,
net/ipv6/seg6_local.c
556
#define flv8986_act_tbl_idx(pf, fm) \
net/ipv6/seg6_local.c
557
((((pf) << bits_per(SEG6_LOCAL_FLV8986_SUPP_OPS)) | \
net/ipv6/seg6_local.c
576
#define tbl_cfg(act, pf, fm) \
net/ipv6/seg6_local.c
577
[flv8986_act_tbl_idx(SEG6_LOCAL_PKTINFO_##pf, \
net/lapb/lapb_in.c
115
lapb_dbg(1, "(%p) S1 RX SABM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
118
lapb->dev, frame->pf);
net/lapb/lapb_in.c
119
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
123
lapb->dev, frame->pf);
net/lapb/lapb_in.c
124
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
130
lapb_dbg(1, "(%p) S1 RX SABME(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
133
lapb->dev, frame->pf);
net/lapb/lapb_in.c
134
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
138
lapb->dev, frame->pf);
net/lapb/lapb_in.c
139
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
145
lapb_dbg(1, "(%p) S1 RX DISC(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
146
lapb_dbg(1, "(%p) S1 TX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
147
lapb_send_control(lapb, LAPB_DM, frame->pf, LAPB_RESPONSE);
net/lapb/lapb_in.c
151
lapb_dbg(1, "(%p) S1 RX UA(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
152
if (frame->pf) {
net/lapb/lapb_in.c
167
lapb_dbg(1, "(%p) S1 RX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
168
if (frame->pf) {
net/lapb/lapb_in.c
193
lapb->dev, frame->pf);
net/lapb/lapb_in.c
194
lapb_dbg(1, "(%p) S2 TX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
195
lapb_send_control(lapb, LAPB_DM, frame->pf, LAPB_RESPONSE);
net/lapb/lapb_in.c
199
lapb_dbg(1, "(%p) S2 RX DISC(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
200
lapb_dbg(1, "(%p) S2 TX UA(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
201
lapb_send_control(lapb, LAPB_UA, frame->pf, LAPB_RESPONSE);
net/lapb/lapb_in.c
205
lapb_dbg(1, "(%p) S2 RX UA(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
206
if (frame->pf) {
net/lapb/lapb_in.c
216
lapb_dbg(1, "(%p) S2 RX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
217
if (frame->pf) {
net/lapb/lapb_in.c
231
lapb->dev, frame->pf);
net/lapb/lapb_in.c
232
lapb_dbg(1, "(%p) S2 RX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
233
if (frame->pf)
net/lapb/lapb_in.c
234
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
255
lapb_dbg(1, "(%p) S3 RX SABM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
258
lapb->dev, frame->pf);
net/lapb/lapb_in.c
259
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
263
lapb->dev, frame->pf);
net/lapb/lapb_in.c
264
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
278
lapb_dbg(1, "(%p) S3 RX SABME(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
281
lapb->dev, frame->pf);
net/lapb/lapb_in.c
282
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
294
lapb->dev, frame->pf);
net/lapb/lapb_in.c
295
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
301
lapb_dbg(1, "(%p) S3 RX DISC(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
304
lapb_send_control(lapb, LAPB_UA, frame->pf, LAPB_RESPONSE);
net/lapb/lapb_in.c
312
lapb_dbg(1, "(%p) S3 RX DM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
323
lapb->dev, frame->pf, frame->nr);
net/lapb/lapb_in.c
325
lapb_check_need_response(lapb, frame->cr, frame->pf);
net/lapb/lapb_in.c
342
lapb->dev, frame->pf, frame->nr);
net/lapb/lapb_in.c
344
lapb_check_need_response(lapb, frame->cr, frame->pf);
net/lapb/lapb_in.c
361
lapb->dev, frame->pf, frame->nr);
net/lapb/lapb_in.c
363
lapb_check_need_response(lapb, frame->cr, frame->pf);
net/lapb/lapb_in.c
383
lapb->dev, frame->pf, frame->ns, frame->nr);
net/lapb/lapb_in.c
417
if (frame->pf)
net/lapb/lapb_in.c
428
if (frame->pf)
net/lapb/lapb_in.c
432
lapb->dev, frame->pf, lapb->vr);
net/lapb/lapb_in.c
434
lapb_send_control(lapb, LAPB_REJ, frame->pf,
net/lapb/lapb_in.c
44
lapb_dbg(1, "(%p) S0 RX SABM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
443
lapb->dev, frame->pf,
net/lapb/lapb_in.c
452
lapb_dbg(1, "(%p) S3 RX ILLEGAL(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
47
lapb->dev, frame->pf);
net/lapb/lapb_in.c
477
lapb_dbg(1, "(%p) S4 RX SABM(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
48
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
480
lapb->dev, frame->pf);
net/lapb/lapb_in.c
481
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
485
lapb->dev, frame->pf);
net/lapb/lapb_in.c
487
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
502
lapb_dbg(1, "(%p) S4 RX SABME(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
505
lapb->dev, frame->pf);
net/lapb/lapb_in.c
507
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
52
lapb->dev, frame->pf);
net/lapb/lapb_in.c
520
lapb->dev, frame->pf);
net/lapb/lapb_in.c
521
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
54
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
69
lapb_dbg(1, "(%p) S0 RX SABME(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
72
lapb->dev, frame->pf);
net/lapb/lapb_in.c
74
lapb_send_control(lapb, LAPB_UA, frame->pf,
net/lapb/lapb_in.c
87
lapb->dev, frame->pf);
net/lapb/lapb_in.c
88
lapb_send_control(lapb, LAPB_DM, frame->pf,
net/lapb/lapb_in.c
94
lapb_dbg(1, "(%p) S0 RX DISC(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
95
lapb_dbg(1, "(%p) S0 TX UA(%d)\n", lapb->dev, frame->pf);
net/lapb/lapb_in.c
96
lapb_send_control(lapb, LAPB_UA, frame->pf, LAPB_RESPONSE);
net/lapb/lapb_out.c
201
void lapb_check_need_response(struct lapb_cb *lapb, int type, int pf)
net/lapb/lapb_out.c
203
if (type == LAPB_COMMAND && pf)
net/lapb/lapb_subr.c
157
frame->pf = skb->data[1] & LAPB_EPF;
net/lapb/lapb_subr.c
169
frame->pf = skb->data[1] & LAPB_EPF;
net/lapb/lapb_subr.c
178
frame->pf = skb->data[0] & LAPB_SPF;
net/lapb/lapb_subr.c
191
frame->pf = skb->data[0] & LAPB_SPF;
net/lapb/lapb_subr.c
198
frame->pf = skb->data[0] & LAPB_SPF;
net/lapb/lapb_subr.c
204
frame->pf = skb->data[0] & LAPB_SPF;
net/netfilter/core.c
276
nf_hook_entry_head(struct net *net, int pf, unsigned int hooknum,
net/netfilter/core.c
279
switch (pf) {
net/netfilter/core.c
348
int pf)
net/netfilter/core.c
350
if ((pf == NFPROTO_NETDEV && reg->hooknum == NF_NETDEV_INGRESS) ||
net/netfilter/core.c
351
(pf == NFPROTO_INET && reg->hooknum == NF_INET_INGRESS))
net/netfilter/core.c
358
int pf)
net/netfilter/core.c
360
return pf == NFPROTO_NETDEV && reg->hooknum == NF_NETDEV_EGRESS;
net/netfilter/core.c
363
static void nf_static_key_inc(const struct nf_hook_ops *reg, int pf)
net/netfilter/core.c
368
if (pf == NFPROTO_INET && reg->hooknum == NF_INET_INGRESS) {
net/netfilter/core.c
369
pf = NFPROTO_NETDEV;
net/netfilter/core.c
374
static_key_slow_inc(&nf_hooks_needed[pf][hooknum]);
net/netfilter/core.c
378
static void nf_static_key_dec(const struct nf_hook_ops *reg, int pf)
net/netfilter/core.c
383
if (pf == NFPROTO_INET && reg->hooknum == NF_INET_INGRESS) {
net/netfilter/core.c
384
pf = NFPROTO_NETDEV;
net/netfilter/core.c
389
static_key_slow_dec(&nf_hooks_needed[pf][hooknum]);
net/netfilter/core.c
393
static int __nf_register_net_hook(struct net *net, int pf,
net/netfilter/core.c
400
switch (pf) {
net/netfilter/core.c
425
pp = nf_hook_entry_head(net, pf, reg->hooknum, reg->dev);
net/netfilter/core.c
444
if (nf_ingress_hook(reg, pf))
net/netfilter/core.c
448
if (nf_egress_hook(reg, pf))
net/netfilter/core.c
451
nf_static_key_inc(reg, pf);
net/netfilter/core.c
485
static void __nf_unregister_net_hook(struct net *net, int pf,
net/netfilter/core.c
491
pp = nf_hook_entry_head(net, pf, reg->hooknum, reg->dev);
net/netfilter/core.c
505
if (nf_ingress_hook(reg, pf))
net/netfilter/core.c
509
if (nf_egress_hook(reg, pf))
net/netfilter/core.c
512
nf_static_key_dec(reg, pf);
net/netfilter/core.c
514
WARN_ONCE(1, "hook not found, pf %d num %d", pf, reg->hooknum);
net/netfilter/core.c
528
if (reg->pf == NFPROTO_INET) {
net/netfilter/core.c
536
__nf_unregister_net_hook(net, reg->pf, reg);
net/netfilter/core.c
558
if (reg->pf == NFPROTO_INET) {
net/netfilter/core.c
575
err = __nf_register_net_hook(net, reg->pf, reg);
net/netfilter/ipset/ip_set_bitmap_port.c
100
ip_set_get_ip_port(const struct sk_buff *skb, u8 pf, bool src, __be16 *port)
net/netfilter/ipset/ip_set_bitmap_port.c
105
switch (pf) {
net/netfilter/ipset/ip_set_core.c
2363
.pf = PF_INET,
net/netfilter/ipvs/ip_vs_core.c
1336
int af = state->pf;
net/netfilter/ipvs/ip_vs_core.c
1917
int af = state->pf;
net/netfilter/ipvs/ip_vs_core.c
2110
if (state->pf == NFPROTO_IPV4) {
net/netfilter/ipvs/ip_vs_core.c
2133
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2142
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2149
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2156
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2164
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2171
.pf = NFPROTO_IPV4,
net/netfilter/ipvs/ip_vs_core.c
2182
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_core.c
2191
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_core.c
2198
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_core.c
2205
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_core.c
2213
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_core.c
2220
.pf = NFPROTO_IPV6,
net/netfilter/ipvs/ip_vs_ctl.c
3168
.pf = PF_INET,
net/netfilter/ipvs/ip_vs_xmit.c
651
static inline int ip_vs_nat_send_or_cont(int pf, struct sk_buff *skb,
net/netfilter/ipvs/ip_vs_xmit.c
673
NF_HOOK(pf, NF_INET_LOCAL_OUT, cp->ipvs->net, NULL, skb,
net/netfilter/ipvs/ip_vs_xmit.c
682
static inline int ip_vs_send_or_cont(int pf, struct sk_buff *skb,
net/netfilter/ipvs/ip_vs_xmit.c
695
NF_HOOK(pf, NF_INET_LOCAL_OUT, cp->ipvs->net, NULL, skb,
net/netfilter/nf_bpf_link.c
147
nf_link->hook_ops.pf, nf_link->hook_ops.hooknum,
net/netfilter/nf_bpf_link.c
157
info->netfilter.pf = nf_link->hook_ops.pf;
net/netfilter/nf_bpf_link.c
184
switch (attr->link_create.netfilter.pf) {
net/netfilter/nf_bpf_link.c
235
link->hook_ops.pf = attr->link_create.netfilter.pf;
net/netfilter/nf_bpf_link.c
79
switch (link->hook_ops.pf) {
net/netfilter/nf_conntrack_core.c
1879
dataoff, state->pf, protonum, state->net,
net/netfilter/nf_conntrack_core.c
1943
if (state->pf == NFPROTO_IPV4 && protonum == IPPROTO_ICMP)
net/netfilter/nf_conntrack_core.c
1946
else if (state->pf == NFPROTO_IPV6 && protonum == IPPROTO_ICMPV6)
net/netfilter/nf_conntrack_core.c
2028
dataoff = get_l4proto(skb, skb_network_offset(skb), state->pf, &protonum);
net/netfilter/nf_conntrack_core.c
400
unsigned int nhoff, u8 pf, u8 *l4num)
net/netfilter/nf_conntrack_core.c
402
switch (pf) {
net/netfilter/nf_conntrack_proto.c
235
.pf = NFPROTO_IPV4,
net/netfilter/nf_conntrack_proto.c
241
.pf = NFPROTO_IPV4,
net/netfilter/nf_conntrack_proto.c
247
.pf = NFPROTO_IPV4,
net/netfilter/nf_conntrack_proto.c
253
.pf = NFPROTO_IPV4,
net/netfilter/nf_conntrack_proto.c
312
.pf = PF_INET,
net/netfilter/nf_conntrack_proto.c
368
.pf = NFPROTO_IPV6,
net/netfilter/nf_conntrack_proto.c
392
.pf = NFPROTO_IPV6,
net/netfilter/nf_conntrack_proto.c
398
.pf = NFPROTO_IPV6,
net/netfilter/nf_conntrack_proto.c
404
.pf = NFPROTO_IPV6,
net/netfilter/nf_conntrack_proto.c
410
.pf = NFPROTO_IPV6,
net/netfilter/nf_conntrack_proto.c
66
nf_log_packet(net, state->pf, 0, skb, state->in, state->out,
net/netfilter/nf_conntrack_proto_icmp.c
125
state->pf, state->net, &origtuple))
net/netfilter/nf_conntrack_proto_icmp.c
172
if (state->pf == AF_INET) {
net/netfilter/nf_conntrack_proto_icmp.c
177
} else if (state->pf == AF_INET6) {
net/netfilter/nf_conntrack_proto_icmp.c
86
if (state->pf != NFPROTO_IPV4)
net/netfilter/nf_conntrack_proto_icmpv6.c
100
if (state->pf != NFPROTO_IPV6)
net/netfilter/nf_conntrack_proto_tcp.c
823
nf_checksum(skb, state->hook, dataoff, IPPROTO_TCP, state->pf)) {
net/netfilter/nf_conntrack_proto_udp.c
174
state->pf)) {
net/netfilter/nf_conntrack_proto_udp.c
75
nf_checksum(skb, state->hook, dataoff, IPPROTO_UDP, state->pf)) {
net/netfilter/nf_log.c
103
rcu_assign_pointer(loggers[pf][logger->type], logger);
net/netfilter/nf_log.c
136
bool nf_log_is_registered(u_int8_t pf)
net/netfilter/nf_log.c
140
if (pf >= NFPROTO_NUMPROTO) {
net/netfilter/nf_log.c
146
if (rcu_access_pointer(loggers[pf][i]))
net/netfilter/nf_log.c
154
int nf_log_bind_pf(struct net *net, u_int8_t pf,
net/netfilter/nf_log.c
157
if (pf >= ARRAY_SIZE(net->nf.nf_loggers))
net/netfilter/nf_log.c
160
if (__find_logger(pf, logger->name) == NULL) {
net/netfilter/nf_log.c
164
rcu_assign_pointer(net->nf.nf_loggers[pf], logger);
net/netfilter/nf_log.c
170
void nf_log_unbind_pf(struct net *net, u_int8_t pf)
net/netfilter/nf_log.c
172
if (pf >= ARRAY_SIZE(net->nf.nf_loggers))
net/netfilter/nf_log.c
175
RCU_INIT_POINTER(net->nf.nf_loggers[pf], NULL);
net/netfilter/nf_log.c
180
int nf_logger_find_get(int pf, enum nf_log_type type)
net/netfilter/nf_log.c
185
if (pf >= ARRAY_SIZE(loggers))
net/netfilter/nf_log.c
190
if (pf == NFPROTO_INET) {
net/netfilter/nf_log.c
205
logger = rcu_dereference(loggers[pf][type]);
net/netfilter/nf_log.c
217
void nf_logger_put(int pf, enum nf_log_type type)
net/netfilter/nf_log.c
221
if (pf == NFPROTO_INET) {
net/netfilter/nf_log.c
228
logger = rcu_dereference(loggers[pf][type]);
net/netfilter/nf_log.c
238
u_int8_t pf,
net/netfilter/nf_log.c
252
logger = rcu_dereference(loggers[pf][loginfo->type]);
net/netfilter/nf_log.c
254
logger = rcu_dereference(net->nf.nf_loggers[pf]);
net/netfilter/nf_log.c
260
logger->logfn(net, pf, hooknum, skb, in, out, loginfo, prefix);
net/netfilter/nf_log.c
267
u_int8_t pf,
net/netfilter/nf_log.c
279
logger = rcu_dereference(net->nf.nf_loggers[pf]);
net/netfilter/nf_log.c
28
static struct nf_logger *__find_logger(int pf, const char *str_logger)
net/netfilter/nf_log.c
284
logger->logfn(net, pf, hooknum, skb, in, out, loginfo, prefix);
net/netfilter/nf_log.c
34
log = nft_log_dereference(loggers[pf][i]);
net/netfilter/nf_log.c
45
int nf_log_set(struct net *net, u_int8_t pf, const struct nf_logger *logger)
net/netfilter/nf_log.c
49
if (pf == NFPROTO_UNSPEC || pf >= ARRAY_SIZE(net->nf.nf_loggers))
net/netfilter/nf_log.c
53
log = nft_log_dereference(net->nf.nf_loggers[pf]);
net/netfilter/nf_log.c
55
rcu_assign_pointer(net->nf.nf_loggers[pf], logger);
net/netfilter/nf_log.c
79
int nf_log_register(u_int8_t pf, struct nf_logger *logger)
net/netfilter/nf_log.c
84
if (pf >= ARRAY_SIZE(init_net.nf.nf_loggers))
net/netfilter/nf_log.c
89
if (pf == NFPROTO_UNSPEC) {
net/netfilter/nf_log.c
99
if (rcu_access_pointer(loggers[pf][logger->type])) {
net/netfilter/nf_log_syslog.c
110
nf_log_dump_packet_common(struct nf_log_buf *m, u8 pf,
net/netfilter/nf_log_syslog.c
134
static void nf_log_arp_packet(struct net *net, u_int8_t pf,
net/netfilter/nf_log_syslog.c
151
nf_log_dump_packet_common(m, pf, hooknum, skb, in, out, loginfo,
net/netfilter/nf_log_syslog.c
833
static void nf_log_ip_packet(struct net *net, u_int8_t pf,
net/netfilter/nf_log_syslog.c
850
nf_log_dump_packet_common(m, pf, hooknum, skb, in,
net/netfilter/nf_log_syslog.c
868
static void nf_log_ip6_packet(struct net *net, u_int8_t pf,
net/netfilter/nf_log_syslog.c
885
nf_log_dump_packet_common(m, pf, hooknum, skb, in, out,
net/netfilter/nf_log_syslog.c
903
static void nf_log_unknown_packet(struct net *net, u_int8_t pf,
net/netfilter/nf_log_syslog.c
921
nf_log_dump_packet_common(m, pf, hooknum, skb, in, out, loginfo,
net/netfilter/nf_log_syslog.c
929
static void nf_log_netdev_packet(struct net *net, u_int8_t pf,
net/netfilter/nf_log_syslog.c
939
nf_log_ip_packet(net, pf, hooknum, skb, in, out, loginfo, prefix);
net/netfilter/nf_log_syslog.c
942
nf_log_ip6_packet(net, pf, hooknum, skb, in, out, loginfo, prefix);
net/netfilter/nf_log_syslog.c
946
nf_log_arp_packet(net, pf, hooknum, skb, in, out, loginfo, prefix);
net/netfilter/nf_log_syslog.c
949
nf_log_unknown_packet(net, pf, hooknum, skb,
net/netfilter/nf_nat_core.c
1180
int nf_nat_register_fn(struct net *net, u8 pf, const struct nf_hook_ops *ops,
net/netfilter/nf_nat_core.c
1190
if (WARN_ON_ONCE(pf >= ARRAY_SIZE(nat_net->nat_proto_net)))
net/netfilter/nf_nat_core.c
1193
nat_proto_net = &nat_net->nat_proto_net[pf];
net/netfilter/nf_nat_core.c
1255
void nf_nat_unregister_fn(struct net *net, u8 pf, const struct nf_hook_ops *ops,
net/netfilter/nf_nat_core.c
1265
if (pf >= ARRAY_SIZE(nat_net->nat_proto_net))
net/netfilter/nf_nat_core.c
1268
nat_proto_net = &nat_net->nat_proto_net[pf];
net/netfilter/nf_nat_proto.c
1051
.pf = NFPROTO_IPV6,
net/netfilter/nf_nat_proto.c
1058
.pf = NFPROTO_IPV6,
net/netfilter/nf_nat_proto.c
1065
.pf = NFPROTO_IPV6,
net/netfilter/nf_nat_proto.c
1072
.pf = NFPROTO_IPV6,
net/netfilter/nf_nat_proto.c
1080
return nf_nat_register_fn(net, ops->pf, ops, nf_nat_ipv6_ops,
net/netfilter/nf_nat_proto.c
1087
nf_nat_unregister_fn(net, ops->pf, ops, ARRAY_SIZE(nf_nat_ipv6_ops));
net/netfilter/nf_nat_proto.c
1097
if (WARN_ON_ONCE(ops->pf != NFPROTO_INET))
net/netfilter/nf_nat_proto.c
790
.pf = NFPROTO_IPV4,
net/netfilter/nf_nat_proto.c
797
.pf = NFPROTO_IPV4,
net/netfilter/nf_nat_proto.c
804
.pf = NFPROTO_IPV4,
net/netfilter/nf_nat_proto.c
811
.pf = NFPROTO_IPV4,
net/netfilter/nf_nat_proto.c
819
return nf_nat_register_fn(net, ops->pf, ops, nf_nat_ipv4_ops,
net/netfilter/nf_nat_proto.c
826
nf_nat_unregister_fn(net, ops->pf, ops, ARRAY_SIZE(nf_nat_ipv4_ops));
net/netfilter/nf_queue.c
169
switch (state->pf) {
net/netfilter/nf_queue.c
216
switch (entry->state.pf) {
net/netfilter/nf_sockopt.c
107
int nf_getsockopt(struct sock *sk, u_int8_t pf, int val, char __user *opt,
net/netfilter/nf_sockopt.c
113
ops = nf_sockopt_find(sk, pf, val, 1);
net/netfilter/nf_sockopt.c
32
if (ops->pf == reg->pf
net/netfilter/nf_sockopt.c
62
static struct nf_sockopt_ops *nf_sockopt_find(struct sock *sk, u_int8_t pf,
net/netfilter/nf_sockopt.c
69
if (ops->pf == pf) {
net/netfilter/nf_sockopt.c
92
int nf_setsockopt(struct sock *sk, u_int8_t pf, int val, sockptr_t opt,
net/netfilter/nf_sockopt.c
98
ops = nf_sockopt_find(sk, pf, val, 0);
net/netfilter/nf_synproxy_core.c
1179
.pf = NFPROTO_IPV6,
net/netfilter/nf_synproxy_core.c
1185
.pf = NFPROTO_IPV6,
net/netfilter/nf_synproxy_core.c
756
.pf = NFPROTO_IPV4,
net/netfilter/nf_synproxy_core.c
762
.pf = NFPROTO_IPV4,
net/netfilter/nf_tables_api.c
2638
ops->pf = family;
net/netfilter/nf_tables_api.c
2884
ops->pf = basechain->ops.pf;
net/netfilter/nf_tables_api.c
8838
ops->pf = NFPROTO_NETDEV;
net/netfilter/nf_tables_api.c
9685
ops->pf = NFPROTO_NETDEV;
net/netfilter/nf_tables_offload.c
232
if (ops->pf != NFPROTO_NETDEV ||
net/netfilter/nfnetlink_hook.c
238
if (ops->pf == NFPROTO_INET && ops->hooknum == NF_INET_INGRESS)
net/netfilter/nfnetlink_hook.c
279
nfnl_hook_entries_head(u8 pf, unsigned int hook, struct net *net, const char *dev)
net/netfilter/nfnetlink_hook.c
286
switch (pf) {
net/netfilter/nfnetlink_log.c
460
u_int8_t pf,
net/netfilter/nfnetlink_log.c
476
0, pf, NFNETLINK_V0, htons(inst->group_num));
net/netfilter/nfnetlink_log.c
497
if (pf == PF_BRIDGE) {
net/netfilter/nfnetlink_log.c
534
if (pf == PF_BRIDGE) {
net/netfilter/nfnetlink_log.c
644
if ((pf == NFPROTO_NETDEV || pf == NFPROTO_BRIDGE) &&
net/netfilter/nfnetlink_log.c
681
u_int8_t pf,
net/netfilter/nfnetlink_log.c
753
if (pf == NFPROTO_NETDEV || pf == NFPROTO_BRIDGE)
net/netfilter/nfnetlink_log.c
801
__build_packet_message(log, inst, skb, data_len, pf,
net/netfilter/nfnetlink_log.c
889
u_int8_t pf = info->nfmsg->nfgen_family;
net/netfilter/nfnetlink_log.c
895
return nf_log_bind_pf(info->net, pf, &nfulnl_logger);
net/netfilter/nfnetlink_log.c
897
nf_log_unbind_pf(info->net, pf);
net/netfilter/nfnetlink_queue.c
1057
switch (entry->state.pf) {
net/netfilter/nfnetlink_queue.c
1510
if (entry->state.pf == PF_BRIDGE) {
net/netfilter/nfnetlink_queue.c
284
static struct nf_hook_entries *nf_hook_entries_head(const struct net *net, u8 pf, u8 hooknum)
net/netfilter/nfnetlink_queue.c
286
switch (pf) {
net/netfilter/nfnetlink_queue.c
327
switch (entry->state.pf) {
net/netfilter/nfnetlink_queue.c
349
u8 pf;
net/netfilter/nfnetlink_queue.c
352
pf = entry->state.pf;
net/netfilter/nfnetlink_queue.c
354
hooks = nf_hook_entries_head(net, pf, entry->state.hook);
net/netfilter/nfnetlink_queue.c
569
if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb))
net/netfilter/nfnetlink_queue.c
587
if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb))
net/netfilter/nfnetlink_queue.c
734
0, entry->state.pf, NFNETLINK_V0,
net/netfilter/nfnetlink_queue.c
754
if (entry->state.pf == PF_BRIDGE) {
net/netfilter/nfnetlink_queue.c
788
if (entry->state.pf == PF_BRIDGE) {
net/netfilter/nft_chain_filter.c
150
switch (state->pf) {
net/netfilter/nft_chain_filter.c
173
ingress_state.pf = NFPROTO_IPV4;
net/netfilter/nft_chain_filter.c
181
ingress_state.pf = NFPROTO_IPV6;
net/netfilter/nft_chain_nat.c
17
switch (state->pf) {
net/netfilter/nft_chain_route.c
121
switch (state->pf) {
net/netfilter/x_tables.c
1777
ops[i].pf = table->af;
net/openvswitch/conntrack.c
740
.pf = info->family,
net/sched/act_ct.c
1039
state.pf = family;
net/sched/cls_basic.c
215
fnew->pf = alloc_percpu(struct tc_basic_pcnt);
net/sched/cls_basic.c
216
if (!fnew->pf) {
net/sched/cls_basic.c
243
free_percpu(fnew->pf);
net/sched/cls_basic.c
291
struct tc_basic_pcnt *pf = per_cpu_ptr(f->pf, cpu);
net/sched/cls_basic.c
293
gpf.rcnt += pf->rcnt;
net/sched/cls_basic.c
294
gpf.rhit += pf->rhit;
net/sched/cls_basic.c
36
struct tc_basic_pcnt __percpu *pf;
net/sched/cls_basic.c
49
__this_cpu_inc(f->pf->rcnt);
net/sched/cls_basic.c
52
__this_cpu_inc(f->pf->rhit);
net/sched/cls_basic.c
94
free_percpu(f->pf);
net/sched/cls_matchall.c
204
new->pf = alloc_percpu(struct tc_matchall_pcnt);
net/sched/cls_matchall.c
205
if (!new->pf) {
net/sched/cls_matchall.c
23
struct tc_matchall_pcnt __percpu *pf;
net/sched/cls_matchall.c
241
free_percpu(new->pf);
net/sched/cls_matchall.c
356
struct tc_matchall_pcnt *pf = per_cpu_ptr(head->pf, cpu);
net/sched/cls_matchall.c
358
gpf.rhit += pf->rhit;
net/sched/cls_matchall.c
41
__this_cpu_inc(head->pf->rhit);
net/sched/cls_matchall.c
54
free_percpu(head->pf);
net/sched/cls_u32.c
1114
n->pf = __alloc_percpu(struct_size(n->pf, kcnts, s->nkeys),
net/sched/cls_u32.c
1116
if (!n->pf) {
net/sched/cls_u32.c
1194
free_percpu(n->pf);
net/sched/cls_u32.c
142
__this_cpu_inc(n->pf->rcnt);
net/sched/cls_u32.c
1426
struct tc_u32_pcnt *pf = per_cpu_ptr(n->pf, cpu);
net/sched/cls_u32.c
1428
gpf->rcnt += pf->rcnt;
net/sched/cls_u32.c
1429
gpf->rhit += pf->rhit;
net/sched/cls_u32.c
1431
gpf->kcnts[i] += pf->kcnts[i];
net/sched/cls_u32.c
173
__this_cpu_inc(n->pf->kcnts[j]);
net/sched/cls_u32.c
189
__this_cpu_inc(n->pf->rhit);
net/sched/cls_u32.c
417
free_percpu(n->pf);
net/sched/cls_u32.c
54
struct tc_u32_pcnt __percpu *pf;
net/sched/cls_u32.c
846
new->pf = n->pf;
net/sched/em_ipset.c
64
state.pf = NFPROTO_IPV4;
net/sched/em_ipset.c
70
state.pf = NFPROTO_IPV6;
net/sched/em_ipset.c
80
opt.family = state.pf;
net/sctp/bind_addr.c
324
if (opt->pf->cmp_addr(&laddr->a, addr, opt)) {
net/sctp/bind_addr.c
346
if (sp->pf->af->cmp_addr(&laddr->a, &laddr2->a) &&
net/sctp/bind_addr.c
390
conflict = sp->pf->cmp_addr(&laddr->a, addr, sp);
net/sctp/bind_addr.c
446
if (opt->pf->cmp_addr(&laddr->a, addr, opt))
net/sctp/ipv6.c
1001
af = opt->pf->af;
net/sctp/ipv6.c
972
af = opt->pf->af;
net/sctp/protocol.c
1196
int sctp_register_pf(struct sctp_pf *pf, sa_family_t family)
net/sctp/protocol.c
1202
sctp_pf_inet_specific = pf;
net/sctp/protocol.c
1207
sctp_pf_inet6_specific = pf;
net/sctp/sm_make_chunk.c
235
num_types = sp->pf->supported_addrs(sp, types);
net/sctp/socket.c
1241
sp->pf->to_sk_daddr(daddr, sk);
net/sctp/socket.c
2144
sp->pf->event_msgname(event, msg->msg_name, addr_len);
net/sctp/socket.c
2146
sp->pf->skb_msgname(head_skb, msg->msg_name, addr_len);
net/sctp/socket.c
224
if (!sctp_sk(sk)->pf->send_verify(sctp_sk(sk), (addr)))
net/sctp/socket.c
343
if (!opt->pf->af_supported(addr->sa.sa_family, opt))
net/sctp/socket.c
351
!opt->pf->af_supported(AF_INET, opt))
net/sctp/socket.c
401
if (!sp->pf->bind_verify(sp, addr))
net/sctp/socket.c
454
sp->pf->to_sk_saddr(addr, sk);
net/sctp/socket.c
4584
struct sctp_af *af = sctp_sk(sk)->pf->af;
net/sctp/socket.c
4864
newsp->pf->to_sk_daddr(&asoc->peer.primary_addr, newsk);
net/sctp/socket.c
4887
newsp->pf->copy_ip_options(sk, newsk);
net/sctp/socket.c
5121
sp->pf = sctp_get_pf_specific(sk->sk_family);
net/sctp/socket.c
8120
struct sctp_af *af = sctp_sk(sk)->pf->af;
net/sctp/socket.c
8556
struct sctp_af *af = sctp_sk(sk)->pf->af;
net/sctp/socket.c
8809
af = sctp_sk(sk)->pf->af;
net/sctp/transport.c
412
struct sctp_pf *pf = sctp_get_pf_specific(dst->ops->family);
net/sctp/transport.c
415
pf->af->from_sk(&addr, sk);
net/sctp/transport.c
416
pf->to_sk_daddr(&t->ipaddr, sk);
net/sctp/transport.c
418
pf->to_sk_daddr(&addr, sk);
net/sctp/transport.c
462
opt->pf->to_sk_saddr(&transport->saddr, asoc->base.sk);
net/socket.c
1539
const struct net_proto_family *pf;
net/socket.c
1590
pf = rcu_dereference(net_families[family]);
net/socket.c
1592
if (!pf)
net/socket.c
1599
if (!try_module_get(pf->owner))
net/socket.c
1605
err = pf->create(net, sock, protocol, kern);
net/socket.c
1612
pf->create, family, type, protocol);
net/socket.c
1627
module_put(pf->owner);
net/socket.c
1639
module_put(pf->owner);
net/sunrpc/auth_gss/gss_mech_switch.c
34
struct pf_desc *pf;
net/sunrpc/auth_gss/gss_mech_switch.c
38
pf = &gm->gm_pfs[i];
net/sunrpc/auth_gss/gss_mech_switch.c
39
if (pf->domain)
net/sunrpc/auth_gss/gss_mech_switch.c
40
auth_domain_put(pf->domain);
net/sunrpc/auth_gss/gss_mech_switch.c
41
kfree(pf->auth_domain_name);
net/sunrpc/auth_gss/gss_mech_switch.c
42
pf->auth_domain_name = NULL;
net/sunrpc/auth_gss/gss_mech_switch.c
64
struct pf_desc *pf;
net/sunrpc/auth_gss/gss_mech_switch.c
68
pf = &gm->gm_pfs[i];
net/sunrpc/auth_gss/gss_mech_switch.c
69
pf->auth_domain_name = make_auth_domain_name(pf->name);
net/sunrpc/auth_gss/gss_mech_switch.c
71
if (pf->auth_domain_name == NULL)
net/sunrpc/auth_gss/gss_mech_switch.c
74
pf->pseudoflavor, pf->auth_domain_name);
net/sunrpc/auth_gss/gss_mech_switch.c
79
pf->domain = dom;
security/apparmor/lsm.c
2398
.pf = NFPROTO_IPV4,
security/apparmor/lsm.c
2405
.pf = NFPROTO_IPV6,
security/selinux/hooks.c
5871
family = state->pf;
security/selinux/hooks.c
5943
if (selinux_netlbl_skbuff_setsid(skb, state->pf, sid) != 0)
security/selinux/hooks.c
5964
ad_net_init_from_iif(&ad, &net, state->out->ifindex, state->pf);
security/selinux/hooks.c
6024
family = state->pf;
security/selinux/hooks.c
7796
.pf = NFPROTO_IPV4,
security/selinux/hooks.c
7802
.pf = NFPROTO_IPV4,
security/selinux/hooks.c
7808
.pf = NFPROTO_IPV4,
security/selinux/hooks.c
7815
.pf = NFPROTO_IPV6,
security/selinux/hooks.c
7821
.pf = NFPROTO_IPV6,
security/selinux/hooks.c
7827
.pf = NFPROTO_IPV6,
security/smack/smack_netfilter.c
41
.pf = NFPROTO_IPV4,
security/smack/smack_netfilter.c
48
.pf = NFPROTO_IPV6,
tools/bpf/bpftool/link.c
226
info->netfilter.pf);
tools/bpf/bpftool/link.c
725
unsigned int pf = info->netfilter.pf;
tools/bpf/bpftool/link.c
727
if (pf < ARRAY_SIZE(pf2name))
tools/bpf/bpftool/link.c
728
pfname = pf2name[pf];
tools/bpf/bpftool/link.c
730
switch (pf) {
tools/bpf/bpftool/link.c
748
printf("\n\tpf: %u", pf);
tools/bpf/bpftool/net.c
832
delta = nfa->netfilter.pf - nfb->netfilter.pf;
tools/include/uapi/linux/bpf.h
1829
__u32 pf;
tools/include/uapi/linux/bpf.h
6788
__u32 pf;
tools/lib/bpf/bpf.c
804
attr.link_create.netfilter.pf = OPTS_GET(opts, netfilter.pf, 0);
tools/lib/bpf/bpf.h
437
__u32 pf;
tools/lib/bpf/libbpf.c
13408
lopts.netfilter.pf = OPTS_GET(opts, pf, 0);
tools/lib/bpf/libbpf.h
864
__u32 pf;
tools/perf/util/probe-finder.c
1002
probe_point_inline_cb, (void *)pf);
tools/perf/util/probe-finder.c
1017
static int find_probe_point_by_func(struct probe_finder *pf)
tools/perf/util/probe-finder.c
1019
struct dwarf_callback_param _param = {.data = (void *)pf,
tools/perf/util/probe-finder.c
1021
dwarf_getfuncs(&pf->cu_die, probe_point_search_cb, &_param, 0);
tools/perf/util/probe-finder.c
1061
struct probe_finder *pf)
tools/perf/util/probe-finder.c
1063
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
1070
pf->lcache = intlist__new(NULL);
tools/perf/util/probe-finder.c
1071
if (!pf->lcache)
tools/perf/util/probe-finder.c
1079
.cu_die = &pf->cu_die,
tools/perf/util/probe-finder.c
1080
.sp_die = &pf->sp_die,
tools/perf/util/probe-finder.c
1084
.data = pf,
tools/perf/util/probe-finder.c
1090
ret = probe_point_search_cb(&pf->sp_die, &probe_param);
tools/perf/util/probe-finder.c
1099
diep = dwarf_offdie(dbg->dbg, off + cuhl, &pf->cu_die);
tools/perf/util/probe-finder.c
1107
pf->fname = cu_find_realpath(&pf->cu_die, pp->file);
tools/perf/util/probe-finder.c
1109
pf->fname = NULL;
tools/perf/util/probe-finder.c
1111
if (!pp->file || pf->fname) {
tools/perf/util/probe-finder.c
1113
ret = find_probe_point_by_func(pf);
tools/perf/util/probe-finder.c
1115
ret = find_probe_point_lazy(&pf->cu_die, pf);
tools/perf/util/probe-finder.c
1117
pf->lno = pp->line;
tools/perf/util/probe-finder.c
1118
ret = find_probe_point_by_line(pf);
tools/perf/util/probe-finder.c
1127
intlist__delete(pf->lcache);
tools/perf/util/probe-finder.c
1128
pf->lcache = NULL;
tools/perf/util/probe-finder.c
1135
struct probe_finder *pf)
tools/perf/util/probe-finder.c
1141
if (pf->cfi_eh || pf->cfi_dbg)
tools/perf/util/probe-finder.c
1142
return debuginfo__find_probe_location(dbg, pf);
tools/perf/util/probe-finder.c
1152
pf->e_machine = ehdr.e_machine;
tools/perf/util/probe-finder.c
1153
pf->e_flags = ehdr.e_flags;
tools/perf/util/probe-finder.c
1160
pf->cfi_eh = dwarf_getcfi_elf(elf);
tools/perf/util/probe-finder.c
1162
pf->cfi_dbg = dwarf_getcfi(dbg->dbg);
tools/perf/util/probe-finder.c
1165
ret = debuginfo__find_probe_location(dbg, pf);
tools/perf/util/probe-finder.c
1170
struct probe_finder *pf;
tools/perf/util/probe-finder.c
1182
struct probe_finder *pf = vf->pf;
tools/perf/util/probe-finder.c
1190
if (convert_variable_location(die_mem, vf->pf->addr,
tools/perf/util/probe-finder.c
1191
vf->pf->fb_ops, &pf->sp_die,
tools/perf/util/probe-finder.c
1192
pf, /*tvar=*/NULL) == 0) {
tools/perf/util/probe-finder.c
1203
if (dwarf_haspc(die_mem, vf->pf->addr)) {
tools/perf/util/probe-finder.c
1211
if (pf->abstrace_dieoffset != dwarf_dieoffset(&var_die))
tools/perf/util/probe-finder.c
1221
static int expand_probe_args(Dwarf_Die *sc_die, struct probe_finder *pf,
tools/perf/util/probe-finder.c
1227
struct local_vars_finder vf = {.pf = pf, .args = args, .vars = false,
tools/perf/util/probe-finder.c
1230
for (i = 0; i < pf->pev->nargs; i++) {
tools/perf/util/probe-finder.c
1232
if (strcmp(pf->pev->args[i].var, PROBE_ARG_VARS) == 0)
tools/perf/util/probe-finder.c
1234
else if (strcmp(pf->pev->args[i].var, PROBE_ARG_PARAMS) != 0) {
tools/perf/util/probe-finder.c
1236
args[n] = pf->pev->args[i];
tools/perf/util/probe-finder.c
1240
pr_debug("Expanding %s into:", pf->pev->args[i].var);
tools/perf/util/probe-finder.c
1258
if (tf->pf.addr == tf->tevs[i].point.address)
tools/perf/util/probe-finder.c
1265
static int add_probe_trace_event(Dwarf_Die *sc_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
1268
container_of(pf, struct trace_event_finder, pf);
tools/perf/util/probe-finder.c
1269
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
1291
ret = convert_to_trace_point(&pf->sp_die, tf->mod, pf->addr,
tools/perf/util/probe-finder.c
1302
tev->lang = dwarf_srclang(dwarf_diecu(sc_die, &pf->cu_die, NULL, NULL));
tools/perf/util/probe-finder.c
1314
ret = expand_probe_args(sc_die, pf, args);
tools/perf/util/probe-finder.c
1327
pf->pvar = &args[i];
tools/perf/util/probe-finder.c
1328
pf->tvar = &tev->args[i];
tools/perf/util/probe-finder.c
1330
ret = find_variable(sc_die, pf);
tools/perf/util/probe-finder.c
1391
.pf = {.pev = pev, .dbg = dbg, .callback = add_probe_trace_event},
tools/perf/util/probe-finder.c
1404
tf.pf.skip_empty_arg = true;
tools/perf/util/probe-finder.c
1406
ret = debuginfo__find_probes(dbg, &tf.pf);
tools/perf/util/probe-finder.c
1407
if (ret >= 0 && tf.pf.skip_empty_arg)
tools/perf/util/probe-finder.c
1410
dwarf_cfi_end(tf.pf.cfi_eh);
tools/perf/util/probe-finder.c
1435
ret = convert_variable_location(die_mem, af->pf.addr,
tools/perf/util/probe-finder.c
1436
af->pf.fb_ops, &af->pf.sp_die,
tools/perf/util/probe-finder.c
1437
&af->pf, /*tvar=*/NULL);
tools/perf/util/probe-finder.c
1461
ret2 = die_get_var_range(&af->pf.sp_die,
tools/perf/util/probe-finder.c
1474
if (af->child && dwarf_haspc(die_mem, af->pf.addr))
tools/perf/util/probe-finder.c
1489
if (af->pf.addr == af->vls[i].point.address)
tools/perf/util/probe-finder.c
1497
static int add_available_vars(Dwarf_Die *sc_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
1500
container_of(pf, struct available_var_finder, pf);
tools/perf/util/probe-finder.c
1501
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
1522
ret = convert_to_trace_point(&pf->sp_die, af->mod, pf->addr,
tools/perf/util/probe-finder.c
1542
die_find_child(&pf->cu_die, collect_variables_cb, (void *)af, &die_mem);
tools/perf/util/probe-finder.c
1563
.pf = {.pev = pev, .dbg = dbg, .callback = add_available_vars},
tools/perf/util/probe-finder.c
1576
ret = debuginfo__find_probes(dbg, &af.pf);
tools/perf/util/probe-finder.c
182
regs = get_dwarf_regstr(regn, pf->e_machine, pf->e_flags);
tools/perf/util/probe-finder.c
458
static int convert_variable(Dwarf_Die *vr_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
466
ret = convert_variable_location(vr_die, pf->addr, pf->fb_ops,
tools/perf/util/probe-finder.c
467
&pf->sp_die, pf, pf->tvar);
tools/perf/util/probe-finder.c
468
if (ret == -ENOENT && pf->skip_empty_arg)
tools/perf/util/probe-finder.c
472
print_var_not_found(pf->pvar->var);
tools/perf/util/probe-finder.c
475
else if (ret == 0 && pf->pvar->field) {
tools/perf/util/probe-finder.c
476
ret = convert_variable_fields(vr_die, pf->pvar->var,
tools/perf/util/probe-finder.c
477
pf->pvar->field, &pf->tvar->ref,
tools/perf/util/probe-finder.c
478
&die_mem, pf->pvar->user_access);
tools/perf/util/probe-finder.c
482
ret = convert_variable_type(vr_die, pf->tvar, pf->pvar->type,
tools/perf/util/probe-finder.c
483
pf->pvar->user_access);
tools/perf/util/probe-finder.c
489
static int find_variable(Dwarf_Die *sc_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
496
if (!is_c_varname(pf->pvar->var))
tools/perf/util/probe-finder.c
497
return copy_to_probe_trace_arg(pf->tvar, pf->pvar);
tools/perf/util/probe-finder.c
499
if (pf->pvar->name)
tools/perf/util/probe-finder.c
500
pf->tvar->name = strdup(pf->pvar->name);
tools/perf/util/probe-finder.c
502
buf = synthesize_perf_probe_arg(pf->pvar);
tools/perf/util/probe-finder.c
508
pf->tvar->name = buf;
tools/perf/util/probe-finder.c
510
if (pf->tvar->name == NULL)
tools/perf/util/probe-finder.c
513
pr_debug("Searching '%s' variable in context.\n", pf->pvar->var);
tools/perf/util/probe-finder.c
515
if (!die_find_variable_at(sc_die, pf->pvar->var, pf->addr, &vr_die)) {
tools/perf/util/probe-finder.c
517
if (!die_find_variable_at(&pf->cu_die, pf->pvar->var,
tools/perf/util/probe-finder.c
519
if (pf->skip_empty_arg)
tools/perf/util/probe-finder.c
522
pf->pvar->var);
tools/perf/util/probe-finder.c
527
ret = convert_variable(&vr_die, pf);
tools/perf/util/probe-finder.c
585
static int call_probe_finder(Dwarf_Die *sc_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
599
if (!die_find_realfunc(&pf->cu_die, pf->addr, &pf->sp_die)) {
tools/perf/util/probe-finder.c
600
if (die_find_tailfunc(&pf->cu_die, pf->addr, &pf->sp_die)) {
tools/perf/util/probe-finder.c
602
dwarf_diename(&pf->sp_die));
tools/perf/util/probe-finder.c
611
memcpy(&pf->sp_die, sc_die, sizeof(Dwarf_Die));
tools/perf/util/probe-finder.c
614
dwarf_attr(&pf->sp_die, DW_AT_frame_base, &fb_attr);
tools/perf/util/probe-finder.c
615
ret = dwarf_getlocation_addr(&fb_attr, pf->addr, &pf->fb_ops, &nops, 1);
tools/perf/util/probe-finder.c
617
pf->fb_ops = NULL;
tools/perf/util/probe-finder.c
618
} else if (nops == 1 && pf->fb_ops[0].atom == DW_OP_call_frame_cfa &&
tools/perf/util/probe-finder.c
619
(pf->cfi_eh != NULL || pf->cfi_dbg != NULL)) {
tools/perf/util/probe-finder.c
620
if ((dwarf_cfi_addrframe(pf->cfi_eh, pf->addr, &frame) != 0 &&
tools/perf/util/probe-finder.c
621
(dwarf_cfi_addrframe(pf->cfi_dbg, pf->addr, &frame) != 0)) ||
tools/perf/util/probe-finder.c
622
dwarf_frame_cfa(frame, &pf->fb_ops, &nops) != 0) {
tools/perf/util/probe-finder.c
624
(uintmax_t)pf->addr);
tools/perf/util/probe-finder.c
631
ret = pf->callback(sc_die, pf);
tools/perf/util/probe-finder.c
635
pf->fb_ops = NULL;
tools/perf/util/probe-finder.c
692
static Dwarf_Die *find_best_scope(struct probe_finder *pf, Dwarf_Die *die_mem)
tools/perf/util/probe-finder.c
695
.function = pf->pev->point.function,
tools/perf/util/probe-finder.c
696
.file = pf->fname,
tools/perf/util/probe-finder.c
697
.line = pf->lno,
tools/perf/util/probe-finder.c
704
ret = cu_walk_functions_at(&pf->cu_die, pf->addr, find_best_scope_cb,
tools/perf/util/probe-finder.c
707
cu_walk_functions_at(&pf->cu_die, pf->addr,
tools/perf/util/probe-finder.c
713
static int verify_representive_line(struct probe_finder *pf, const char *fname,
tools/perf/util/probe-finder.c
72
const struct probe_finder *pf,
tools/perf/util/probe-finder.c
721
if (cu_find_lineinfo(&pf->cu_die, addr, &__fname, &__lineno) < 0)
tools/perf/util/probe-finder.c
730
if (pf->pev->point.function) {
tools/perf/util/probe-finder.c
732
pf->addr = addr;
tools/perf/util/probe-finder.c
733
if (find_best_scope(pf, &die_mem)
tools/perf/util/probe-finder.c
734
&& die_match_name(&die_mem, pf->pev->point.function)
tools/perf/util/probe-finder.c
749
struct probe_finder *pf = data;
tools/perf/util/probe-finder.c
753
if (lineno != pf->lno || strtailcmp(fname, pf->fname) != 0)
tools/perf/util/probe-finder.c
756
if (verify_representive_line(pf, fname, lineno, addr))
tools/perf/util/probe-finder.c
759
pf->addr = addr;
tools/perf/util/probe-finder.c
760
sc_die = find_best_scope(pf, &die_mem);
tools/perf/util/probe-finder.c
766
ret = call_probe_finder(sc_die, pf);
tools/perf/util/probe-finder.c
773
static int find_probe_point_by_line(struct probe_finder *pf)
tools/perf/util/probe-finder.c
775
return die_walk_lines(&pf->cu_die, probe_point_line_walker, pf);
tools/perf/util/probe-finder.c
821
struct probe_finder *pf = data;
tools/perf/util/probe-finder.c
825
if (!intlist__has_entry(pf->lcache, lineno) ||
tools/perf/util/probe-finder.c
826
strtailcmp(fname, pf->fname) != 0)
tools/perf/util/probe-finder.c
831
pf->addr = addr;
tools/perf/util/probe-finder.c
832
pf->lno = lineno;
tools/perf/util/probe-finder.c
833
sc_die = find_best_scope(pf, &die_mem);
tools/perf/util/probe-finder.c
839
ret = call_probe_finder(sc_die, pf);
tools/perf/util/probe-finder.c
849
static int find_probe_point_lazy(Dwarf_Die *sp_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
855
if (intlist__empty(pf->lcache)) {
tools/perf/util/probe-finder.c
858
comp_dir = cu_get_comp_dir(&pf->cu_die);
tools/perf/util/probe-finder.c
859
if (pf->dbg->build_id) {
tools/perf/util/probe-finder.c
862
build_id__init(&bid, pf->dbg->build_id, BUILD_ID_SIZE);
tools/perf/util/probe-finder.c
865
ret = find_source_path(pf->fname, sbuild_id, comp_dir, &fpath);
tools/perf/util/probe-finder.c
872
ret = find_lazy_match_lines(pf->lcache, fpath,
tools/perf/util/probe-finder.c
873
pf->pev->point.lazy_line);
tools/perf/util/probe-finder.c
879
return die_walk_lines(sp_die, probe_point_lazy_walker, pf);
tools/perf/util/probe-finder.c
882
static void skip_prologue(Dwarf_Die *sp_die, struct probe_finder *pf)
tools/perf/util/probe-finder.c
884
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
887
if (!pf->pev->uprobes)
tools/perf/util/probe-finder.c
891
if (die_is_optimized_target(&pf->cu_die))
tools/perf/util/probe-finder.c
895
if (!pf->addr)
tools/perf/util/probe-finder.c
904
if (!perf_probe_with_var(pf->pev))
tools/perf/util/probe-finder.c
909
pf->addr);
tools/perf/util/probe-finder.c
911
die_skip_prologue(sp_die, &pf->cu_die, &pf->addr);
tools/perf/util/probe-finder.c
916
struct probe_finder *pf = data;
tools/perf/util/probe-finder.c
917
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
922
ret = find_probe_point_lazy(in_die, pf);
tools/perf/util/probe-finder.c
935
pf->addr = addr;
tools/perf/util/probe-finder.c
936
pf->addr += pp->offset;
tools/perf/util/probe-finder.c
938
(uintmax_t)pf->addr);
tools/perf/util/probe-finder.c
940
ret = call_probe_finder(in_die, pf);
tools/perf/util/probe-finder.c
956
struct probe_finder *pf = param->data;
tools/perf/util/probe-finder.c
957
struct perf_probe_point *pp = &pf->pev->point;
tools/perf/util/probe-finder.c
976
pf->fname = fname;
tools/perf/util/probe-finder.c
977
pf->abstrace_dieoffset = dwarf_dieoffset(sp_die);
tools/perf/util/probe-finder.c
979
dwarf_decl_line(sp_die, &pf->lno);
tools/perf/util/probe-finder.c
980
pf->lno += pp->line;
tools/perf/util/probe-finder.c
981
param->retval = find_probe_point_by_line(pf);
tools/perf/util/probe-finder.c
984
die_entrypc(sp_die, &pf->addr);
tools/perf/util/probe-finder.c
986
if (pf->addr == 0) {
tools/perf/util/probe-finder.c
992
param->retval = find_probe_point_lazy(sp_die, pf);
tools/perf/util/probe-finder.c
994
skip_prologue(sp_die, pf);
tools/perf/util/probe-finder.c
995
pf->addr += pp->offset;
tools/perf/util/probe-finder.c
997
param->retval = call_probe_finder(sp_die, pf);
tools/perf/util/probe-finder.h
58
int (*callback)(Dwarf_Die *sc_die, struct probe_finder *pf);
tools/perf/util/probe-finder.h
83
struct probe_finder pf;
tools/perf/util/probe-finder.h
91
struct probe_finder pf;
tools/testing/selftests/alsa/test-pcmtest-driver.c
37
char pf[64];
tools/testing/selftests/alsa/test-pcmtest-driver.c
48
sprintf(pf, "/sys/kernel/debug/pcmtest/fill_pattern%d", i);
tools/testing/selftests/alsa/test-pcmtest-driver.c
49
fp = fopen(pf, "r");
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
543
static int set_port_drop(int pf, bool multi_port)
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
553
"protocol", pf == PF_INET6 ? "ipv6" : "ip",
tools/testing/selftests/bpf/prog_tests/ip_check_defrag.c
84
.pf = ipv6 ? NFPROTO_IPV6 : NFPROTO_IPV4,
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
10
__u32 pf;
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
21
{ .pf = NFPROTO_NUMPROTO, .name = "invalid-pf", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
22
{ .pf = NFPROTO_IPV4, .hooknum = 42, .name = "invalid-hooknum", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
23
{ .pf = NFPROTO_IPV4, .priority = INT_MIN, .name = "invalid-priority-min", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
24
{ .pf = NFPROTO_IPV4, .priority = INT_MAX, .name = "invalid-priority-max", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
25
{ .pf = NFPROTO_IPV4, .flags = UINT_MAX, .name = "invalid-flags", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
27
{ .pf = NFPROTO_INET, .priority = 1, .name = "invalid-inet-not-supported", },
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
30
.pf = NFPROTO_IPV4,
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
38
.pf = NFPROTO_IPV6,
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
60
ASSERT_EQ(info.netfilter.pf, nf_expected.pf, "info nf protocol family");
tools/testing/selftests/bpf/prog_tests/netfilter_link_attach.c
88
X(opts, pf, i);
tools/testing/selftests/bpf/progs/linked_list.c
112
struct foo *f[200], *pf;
tools/testing/selftests/bpf/progs/linked_list.c
146
pf = container_of(n, struct foo, node2);
tools/testing/selftests/bpf/progs/linked_list.c
147
if (pf->data != (ARRAY_SIZE(f) - i - 1)) {
tools/testing/selftests/bpf/progs/linked_list.c
148
bpf_obj_drop(pf);
tools/testing/selftests/bpf/progs/linked_list.c
152
bpf_list_push_back(head, &pf->node2);
tools/testing/selftests/bpf/progs/linked_list.c
165
pf = container_of(n, struct foo, node2);
tools/testing/selftests/bpf/progs/linked_list.c
166
if (pf->data != i) {
tools/testing/selftests/bpf/progs/linked_list.c
167
bpf_obj_drop(pf);
tools/testing/selftests/bpf/progs/linked_list.c
170
bpf_obj_drop(pf);
tools/testing/selftests/bpf/progs/verifier_netfilter_ctx.c
104
if (state->pf != 2)
tools/testing/selftests/net/mptcp/mptcp_connect.c
1075
static void check_sockaddr(int pf, struct sockaddr_storage *ss,
tools/testing/selftests/net/mptcp/mptcp_connect.c
1082
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_connect.c
1096
fprintf(stderr, "accept: Unknown pf %d, salen %u\n", pf, salen);
tools/testing/selftests/net/mptcp/mptcp_connect.c
1104
if (ss->ss_family != pf)
tools/testing/selftests/net/mptcp/mptcp_connect.c
1106
(int)ss->ss_family, pf);
tools/testing/selftests/net/mptcp/mptcp_connect.c
1205
check_sockaddr(pf, &ss, salen);
tools/testing/selftests/net/mptcp/mptcp_connect.c
1561
pf = AF_INET6;
tools/testing/selftests/net/mptcp/mptcp_connect.c
1603
pf = AF_INET6;
tools/testing/selftests/net/mptcp/mptcp_connect.c
246
static void set_transparent(int fd, int pf)
tools/testing/selftests/net/mptcp/mptcp_connect.c
250
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_connect.c
316
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_connect.c
322
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_connect.c
336
set_transparent(sock, pf);
tools/testing/selftests/net/mptcp/mptcp_connect.c
382
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_connect.c
69
static int pf = AF_INET;
tools/testing/selftests/net/mptcp/mptcp_inq.c
109
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_inq.c
115
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_inq.c
155
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_inq.c
197
pf = AF_INET6;
tools/testing/selftests/net/mptcp/mptcp_inq.c
40
static int pf = AF_INET;
tools/testing/selftests/net/mptcp/mptcp_inq.c
483
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_inq.c
491
xerror("Unknown pf %d\n", pf);
tools/testing/selftests/net/mptcp/mptcp_inq.c
516
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_inq.c
524
xerror("Unknown pf %d\n", pf);
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
196
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
202
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
242
hints.ai_family = pf;
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
273
pf = AF_INET6;
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
30
static int pf = AF_INET;
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
495
assert(addrs.addr[0].sa_family == pf);
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
710
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
718
xerror("Unknown pf %d\n", pf);
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
778
switch (pf) {
tools/testing/selftests/net/mptcp/mptcp_sockopt.c
786
xerror("Unknown pf %d\n", pf);
tools/testing/selftests/net/netfilter/nf_queue.c
129
.pf = htons(AF_INET),