#include "lm5710.h"
unsigned long log2_align(unsigned long n);
u64_t lm_get_timestamp_of_recent_cid_recycling(struct _lm_device_t *pdev)
{
return pdev->vars.last_recycling_timestamp;
}
u32_t lm_get_max_supported_toe_cons(struct _lm_device_t *pdev)
{
if ( CHK_NULL(pdev) )
{
return 0;
}
return pdev->params.max_supported_toe_cons;
}
u8_t lm_get_toe_rss_possibility(struct _lm_device_t *pdev)
{
if ( CHK_NULL(pdev) )
{
return 0;
}
return (pdev->params.l4_rss_is_possible != L4_RSS_DISABLED);
}
lm_status_t lm_get_iscsi_boot_info_block( struct _lm_device_t *pdev, struct _iscsi_info_block_hdr_t* iscsi_info_block_hdr_ptr )
{
u32_t val = 0;
u32_t offset = 0;
const u8_t func_mb_id = FUNC_MAILBOX_ID(pdev);
shmem_region_t* shmem_region_dummy = NULL;
shmem2_region_t* shmem2_region_dummy = NULL;
shared_hw_cfg_t* shared_hw_cfg_dummy = NULL;
port_hw_cfg_t* port_hw_cfg_dummy = NULL;
shared_feat_cfg_t* shared_feat_cfg_dummy = NULL;
port_feat_cfg_t* port_feat_cfg_dummy = NULL;
mf_cfg_t* mf_cfg_dummy = NULL;
UNREFERENCED_PARAMETER_(shmem_region_dummy);
UNREFERENCED_PARAMETER_(shmem2_region_dummy);
UNREFERENCED_PARAMETER_(shared_hw_cfg_dummy);
UNREFERENCED_PARAMETER_(port_hw_cfg_dummy);
UNREFERENCED_PARAMETER_(shared_feat_cfg_dummy);
UNREFERENCED_PARAMETER_(port_feat_cfg_dummy);
UNREFERENCED_PARAMETER_(mf_cfg_dummy);
if ( CHK_NULL( iscsi_info_block_hdr_ptr ) )
{
return LM_STATUS_INVALID_PARAMETER ;
}
if (pdev->hw_info.mcp_detected == 1)
{
offset = OFFSETOF(shmem_region_t,func_mb[func_mb_id].iscsi_boot_signature);
LM_SHMEM_READ(pdev, offset, &val );
iscsi_info_block_hdr_ptr->signature = val ;
offset = OFFSETOF(shmem_region_t,func_mb[func_mb_id].iscsi_boot_block_offset);
LM_SHMEM_READ(pdev, offset, &val );
if (val == UEFI_BOOT_SIGNATURE)
{
SET_FLAGS(iscsi_info_block_hdr_ptr->boot_flags, BOOT_INFO_FLAGS_UEFI_BOOT );
}
else
{
RESET_FLAGS(iscsi_info_block_hdr_ptr->boot_flags, BOOT_INFO_FLAGS_UEFI_BOOT );
}
}
else
{
iscsi_info_block_hdr_ptr->signature = 0;
}
return LM_STATUS_SUCCESS ;
}
lm_status_t
lm_get_ibft_physical_addr_for_efi(
struct _lm_device_t *pdev, u32_t *phy_hi, u32_t *phy_lo
)
{
u32_t offset = 0;
u32_t val = 0;
const u8_t func_mb_id = FUNC_MAILBOX_ID(pdev);
if (pdev->hw_info.mcp_detected == 1)
{
offset = OFFSETOF(shmem_region_t,func_mb[func_mb_id].iscsi_boot_signature);
LM_SHMEM_READ(pdev, offset, &val );
offset = OFFSETOF(shmem_region_t,func_mb[func_mb_id].iscsi_boot_block_offset);
LM_SHMEM_READ(pdev, offset, &val );
if (val == UEFI_BOOT_SIGNATURE)
{
offset = OFFSETOF(shmem2_region_t,ibft_host_addr);
LM_SHMEM2_READ(pdev, offset , &val);
*phy_lo = val;
*phy_hi = 0;
return LM_STATUS_SUCCESS;
}
}
return LM_STATUS_FAILURE;
}
lm_status_t
lm_get_sriov_info(lm_device_t *pdev)
{
lm_status_t rc = LM_STATUS_SUCCESS;
u32_t val;
if (!CHIP_IS_E1x(pdev)) {
#ifdef VF_INVOLVED
rc = mm_get_sriov_info(pdev, &pdev->hw_info.sriov_info);
if (rc != LM_STATUS_SUCCESS) {
return rc;
}
#endif
#ifdef __LINUX
lm_set_virt_mode(pdev, DEVICE_TYPE_PF, (pdev->hw_info.sriov_info.total_vfs? VT_BASIC_VF : VT_NONE));
#elif defined(_VBD_CMD_)
lm_set_virt_mode(pdev, DEVICE_TYPE_PF, (pdev->hw_info.sriov_info.total_vfs? VT_CHANNEL_VF : VT_NONE));
#endif
val = REG_RD(pdev, PCICFG_OFFSET + GRC_CONFIG_REG_PF_INIT_VF);
pdev->hw_info.sriov_info.first_vf_in_pf = ((val & GRC_CR_PF_INIT_VF_PF_FIRST_VF_NUM_MASK) * 8) - E2_MAX_NUM_OF_VFS*PATH_ID(pdev);
DbgMessage(pdev, WARN, "First VF in PF = %d\n", pdev->hw_info.sriov_info.first_vf_in_pf);
}
return rc;
}
static void lm_print_func_info(lm_device_t *pdev)
{
DbgMessage(pdev, WARN, "lm_get_shmem_info: FUNC_ID: %d\n", FUNC_ID(pdev));
DbgMessage(pdev, WARN, "lm_get_shmem_info: PCI_FUNC_ID: %d\n", ABS_FUNC_ID(pdev));
DbgMessage(pdev, WARN, "lm_get_shmem_info: PORT_ID: %d\n", PORT_ID(pdev));
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
DbgMessage(pdev, WARN, "lm_get_shmem_info: ETH_PORT_ID: %d\n", PATH_ID(pdev) + 2*PORT_ID(pdev));
}
else
{
DbgMessage(pdev, WARN, "lm_get_shmem_info: ETH_PORT_ID: %d\n", PATH_ID(pdev) + PORT_ID(pdev));
}
DbgMessage(pdev, WARN, "lm_get_shmem_info: PATH_ID: %d\n", PATH_ID(pdev));
DbgMessage(pdev, WARN, "lm_get_shmem_info: VNIC_ID: %d\n", VNIC_ID(pdev));
DbgMessage(pdev, WARN, "lm_get_shmem_info: FUNC_MAILBOX_ID: %d\n", FUNC_MAILBOX_ID(pdev));
}
lm_status_t
lm_get_function_num(lm_device_t *pdev)
{
u32_t val = 0;
val = REG_RD(pdev, BAR_ME_REGISTER);
pdev->params.pfunc_rel = (u8_t)((val & ME_REG_PF_NUM) >> ME_REG_PF_NUM_SHIFT);
pdev->params.path_id = (u8_t)((val & ME_REG_ABS_PF_NUM) >> ME_REG_ABS_PF_NUM_SHIFT) & 1;
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
pdev->params.pfunc_abs = (pdev->params.pfunc_rel << 1) | pdev->params.path_id;
}
else
{
pdev->params.pfunc_abs = pdev->params.pfunc_rel | pdev->params.path_id;
}
pdev->params.pfunc_mb_id = FUNC_MAILBOX_ID(pdev);
DbgMessage(pdev, INFORM , "relative function %d absolute function %d\n", pdev->params.pfunc_rel, pdev->params.pfunc_abs);
lm_print_func_info(pdev);
return LM_STATUS_SUCCESS;
}
lm_status_t lm_get_pcicfg_mps_mrrs(lm_device_t * pdev)
{
lm_status_t lm_status = LM_STATUS_SUCCESS;
u32_t val = 0;
lm_status = mm_read_pci(pdev, PCICFG_DEVICE_CONTROL, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.max_payload_size = (val & 0xe0)>>5;
pdev->hw_info.max_read_req_size = (val & 0x7000)>>12;
DbgMessage(pdev, INFORMi, "reg 0xd8 0x%x \n max_payload %d max_read_req %d \n",
val,pdev->hw_info.max_payload_size,pdev->hw_info.max_read_req_size);
return lm_status ;
}
lm_status_t lm_get_pcicfg_info(lm_device_t *pdev)
{
lm_status_t lm_status;
u32_t val;
if (IS_PFDEV(pdev) || IS_CHANNEL_VFDEV(pdev))
{
lm_status = mm_read_pci(pdev, PCICFG_VENDOR_ID_OFFSET, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
if (val != 0xFFFFFFFF)
{
pdev->hw_info.vid = (u16_t) val;
pdev->hw_info.did = (u16_t) (val >> 16);
}
else if (IS_SW_CHANNEL_VIRT_MODE(pdev))
{
pdev->hw_info.vid = 0x14E4;
pdev->hw_info.did = 0x166F;
}
DbgMessage(pdev, INFORMi, "vid 0x%x\n", pdev->hw_info.vid);
DbgMessage(pdev, INFORMi, "did 0x%x\n", pdev->hw_info.did);
}
else
{
DbgMessage(pdev, WARN, "vid&did for VBD VF will be known later\n");
}
lm_status = mm_read_pci(pdev, PCICFG_SUBSYSTEM_VENDOR_ID_OFFSET, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.svid = (u16_t) val;
DbgMessage(pdev, INFORMi, "svid 0x%x\n", pdev->hw_info.svid);
pdev->hw_info.ssid = (u16_t) (val >> 16);
DbgMessage(pdev, INFORMi, "ssid 0x%x\n", pdev->hw_info.ssid);
lm_status = mm_read_pci(pdev, PCICFG_INT_LINE, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.irq = (u8_t) val;
DbgMessage(pdev, INFORMi, "IRQ 0x%x\n", pdev->hw_info.irq);
pdev->hw_info.int_pin = (u8_t) (val >> 8);
DbgMessage(pdev, INFORMi, "Int pin 0x%x\n", pdev->hw_info.int_pin);
lm_status = mm_read_pci(pdev, PCICFG_CACHE_LINE_SIZE, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.cache_line_size = (u8_t) val;
DbgMessage(pdev, INFORMi, "Cache line size 0x%x\n", (u8_t) val);
pdev->hw_info.latency_timer = (u8_t) (val >> 8);
DbgMessage(pdev, INFORMi, "Latency timer 0x%x\n", (u8_t) (val >> 8));
lm_status = mm_read_pci(pdev, PCICFG_REVISION_ID_OFFSET, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.rev_id = (u8_t) val;
DbgMessage(pdev, INFORMi, "Revision id 0x%x\n", pdev->hw_info.rev_id);
if (IS_PFDEV(pdev))
{
lm_status = mm_read_pci(pdev, PCICFG_LINK_CONTROL, &val);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
pdev->hw_info.pcie_lane_width = (val & 0x3f00000) >> 20;
DbgMessage(pdev, INFORMi, "pcie_lane_width 0x%x\n", pdev->hw_info.pcie_lane_width);
pdev->hw_info.pcie_lane_speed = (val & 0xf0000) >> 16;
DbgMessage(pdev, INFORMi, "pcie_lane_speed 0x%x\n", pdev->hw_info.pcie_lane_speed);
lm_status = lm_get_pcicfg_mps_mrrs(pdev);
}
if (GET_FLAGS(pdev->hw_info.rev_id,PCICFG_REVESION_ID_MASK) == PCICFG_REVESION_ID_ERROR_VAL)
{
return LM_STATUS_FAILURE;
}
return lm_status;
}
static __inline lm_status_t lm_get_bar_offset_direct(
IN struct _lm_device_t * pdev,
IN u8_t bar_num,
OUT lm_address_t * bar_addr )
{
u32_t pci_reg, val;
lm_status_t lm_status;
switch (bar_num) {
case BAR_0:
pci_reg = PCICFG_BAR_1_LOW;
break;
case BAR_1:
pci_reg = PCICFG_BAR_1_LOW + 8;
break;
case BAR_2:
pci_reg = PCICFG_BAR_1_LOW + 16;
break;
default:
DbgMessage(pdev, FATAL, "Unsupported bar index: %d\n", bar_num);
DbgBreakIfAll(1);
return LM_STATUS_INVALID_PARAMETER;
}
lm_status = mm_read_pci(pdev, pci_reg, &val);
if(lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
bar_addr->as_u32.low = val & 0xfffffff0;;
DbgMessage(pdev, INFORMi, "BAR %d low 0x%x\n", bar_num,
bar_addr->as_u32.low);
pci_reg += 4;
lm_status = mm_read_pci(pdev, pci_reg, &val);
if(lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
bar_addr->as_u32.high = val;
DbgMessage(pdev, INFORMi, "BAR %d high 0x%x\n", bar_num,
bar_addr->as_u32.high);
return LM_STATUS_SUCCESS;
}
static __inline lm_status_t lm_get_bar_size_direct (
IN lm_device_t *pdev,
IN u8_t bar_num,
OUT u32_t * val_p)
{
u32_t bar_address = 0;
u32_t bar_size;
switch (bar_num) {
case BAR_0:
bar_address = GRC_CONFIG_2_SIZE_REG;
break;
case BAR_1:
bar_address = GRC_BAR2_CONFIG;
break;
case BAR_2:
bar_address = GRC_BAR3_CONFIG;
break;
default:
DbgMessage(pdev, FATAL, "Invalid Bar Num\n");
return LM_STATUS_INVALID_PARAMETER;
}
lm_reg_rd_ind(pdev,PCICFG_OFFSET + bar_address,&bar_size);
ASSERT_STATIC(PCI_CONFIG_2_BAR1_SIZE == PCI_CONFIG_2_BAR2_SIZE);
ASSERT_STATIC(PCI_CONFIG_2_BAR2_SIZE == PCI_CONFIG_2_BAR3_SIZE);
bar_size = (bar_size & PCI_CONFIG_2_BAR1_SIZE);
if (bar_size == 0)
{
return LM_STATUS_FAILURE;
}
else
{
*val_p = (0x40 << ((bar_size - 1)))*0x400;
}
return LM_STATUS_SUCCESS;
}
lm_status_t lm_get_bars_info(lm_device_t *pdev)
{
lm_status_t lm_status;
u32_t bar_map_size = 0;
u8_t i;
for (i = 0; i < ARRSIZE(pdev->hw_info.mem_base); i++)
{
lm_status = mm_get_bar_offset(pdev, i, &pdev->hw_info.mem_base[i]);
DbgMessage(pdev, INFORMi, "Bar_Offset=0x%x\n", pdev->hw_info.mem_base[i]);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
if(pdev->hw_info.mem_base[i].as_u64 == 0)
{
DbgMessage(pdev, WARNi, "BAR %d IS NOT PRESENT\n", i);
if(i==0)
{
DbgBreakMsg("BAR 0 must be present\n");
}
}
}
for (i = 0; i < MAX_NUM_BAR; i++)
{
if(pdev->hw_info.mem_base[i].as_u64 == 0)
{
continue;
}
lm_status = mm_get_bar_size(pdev, i, &(pdev->hw_info.bar_size[i]));
if ( lm_status != LM_STATUS_SUCCESS )
{
return lm_status;
}
DbgMessage(pdev, INFORMi, "bar %d size 0x%x\n", i, pdev->hw_info.bar_size[i]);
if (BAR_1 == i )
{
if (IS_PFDEV(pdev))
{
#ifdef VF_INVOLVED
bar_map_size = pdev->hw_info.bar_size[i];
#else
bar_map_size = LM_DQ_CID_SIZE * MAX_ETH_CONS;
#endif
}
else
{
bar_map_size = LM_DQ_CID_SIZE;
}
#ifndef VF_INVOLVED
DbgBreakIf(bar_map_size >= pdev->hw_info.bar_size[i]);
#endif
}
else
{
bar_map_size = pdev->hw_info.bar_size[i];
}
if(lm_is_function_after_flr(pdev) ||
#ifdef VF_INVOLVED
lm_is_function_after_flr(PFDEV(pdev)) ||
#endif
(pdev->vars.mapped_bar_addr[i] == NULL))
{
pdev->vars.mapped_bar_addr[i] = NULL;
pdev->vars.mapped_bar_addr[i] = mm_map_io_base(
pdev,
pdev->hw_info.mem_base[i],
bar_map_size,
i);
if(pdev->vars.mapped_bar_addr[i] == NULL)
{
DbgMessage(pdev, FATAL, "bar %d map io failed\n", i);
return LM_STATUS_FAILURE;
}
else
{
DbgMessage(pdev, INFORMi, "mem_base[%d]=%p size=0x%x\n", i, pdev->vars.mapped_bar_addr[i], pdev->hw_info.bar_size[i]);
}
}
}
#if defined(__LINUX) || defined(_VBD_)
if (IS_PFDEV(pdev))
{
pdev->hw_info.pcie_caps_offset = mm_get_cap_offset(pdev, PCI_CAP_PCIE);
if (pdev->hw_info.pcie_caps_offset != 0 && pdev->hw_info.pcie_caps_offset != 0xFFFFFFFF)
{
mm_read_pci(pdev, pdev->hw_info.pcie_caps_offset + PCIE_DEV_CAPS, &pdev->hw_info.pcie_dev_capabilities);
DbgMessage(pdev, WARN,"Device Capability of PCIe caps is %x\n",pdev->hw_info.pcie_dev_capabilities);
if (pdev->hw_info.pcie_dev_capabilities)
{
if (pdev->hw_info.pcie_dev_capabilities & PCIE_DEV_CAPS_FLR_CAPABILITY)
{
pdev->hw_info.flr_capable = TRUE;
}
else
{
pdev->hw_info.flr_capable = FALSE;
}
}
else
{
pdev->hw_info.pci_cfg_trust = PCI_CFG_NOT_TRUSTED;
}
}
else
{
pdev->hw_info.pci_cfg_trust = PCI_CFG_NOT_TRUSTED;
}
if (!lm_is_function_after_flr(pdev))
{
pdev->hw_info.grc_didvid = REG_RD(pdev, (PCICFG_OFFSET + PCICFG_VENDOR_ID_OFFSET));
lm_status = mm_read_pci(pdev, PCICFG_VENDOR_ID_OFFSET, &pdev->hw_info.pci_cfg_didvid);
if (lm_status == LM_STATUS_SUCCESS)
{
if (pdev->hw_info.grc_didvid != pdev->hw_info.pci_cfg_didvid)
{
pdev->hw_info.flr_capable = TRUE;
pdev->params.is_flr = TRUE;
}
}
}
}
#endif
if (lm_is_function_after_flr(pdev))
{
u32_t m_e,tr_e,tw_e;
u32_t i_cycles;
REG_WR(pdev, PGLUE_B_REG_INTERNAL_PFID_ENABLE_TARGET_READ, 1);
for (i_cycles = 0; i_cycles < 1000; i_cycles++)
{
mm_wait(pdev,999);
}
tr_e = REG_RD(pdev, PGLUE_B_REG_INTERNAL_PFID_ENABLE_TARGET_READ);
tw_e = REG_RD(pdev, PGLUE_B_REG_INTERNAL_PFID_ENABLE_TARGET_WRITE);
m_e = REG_RD(pdev, PGLUE_B_REG_INTERNAL_PFID_ENABLE_MASTER);
DbgMessage(pdev, INFORM, "M:0x%x, TR:0x%x, TW:0x%x\n",m_e,tr_e,tw_e);
if (tw_e != 0x1)
{
DbgBreakMsg("BAR 0 must be present\n");
return LM_STATUS_FAILURE;
}
}
return LM_STATUS_SUCCESS;
}
lm_status_t lm_get_chip_id_and_mode(lm_device_t *pdev)
{
u32_t val;
u32_t chip_rev;
val=REG_RD(PFDEV(pdev),MISC_REG_CHIP_NUM);
CHIP_NUM_SET(pdev->hw_info.chip_id,val);
val=REG_RD(PFDEV(pdev),MISC_REG_CHIP_TYPE);
if (val & CHIP_OPT_MISC_DO_BIT)
{
switch (pdev->hw_info.chip_id)
{
case CHIP_NUM_57810:
pdev->hw_info.chip_id = CHIP_NUM_57811;
break;
case CHIP_NUM_57810_MF:
pdev->hw_info.chip_id = CHIP_NUM_57811_MF;
break;
default:
DbgMessage(pdev, FATAL, "Un-supported chip id for OTP: %d\n", pdev->hw_info.chip_id);
DbgBreakIfAll(1);
return LM_STATUS_FAILURE;
}
}
val=REG_RD(PFDEV(pdev),MISC_REG_CHIP_REV);
chip_rev = (val & 0xF)<<CHIP_REV_SHIFT;
pdev->hw_info.chip_id |= chip_rev;
if(chip_rev <= CHIP_REV_ASIC_MAX)
{
pdev->vars.clk_factor = 1;
}
else if(chip_rev & CHIP_REV_SIM_IS_FPGA)
{
pdev->vars.clk_factor = LM_FPGA_FACTOR;
DbgMessage(pdev, INFORMi, "FPGA: forcing MPS from %d to 0.\n", pdev->hw_info.max_payload_size);
pdev->hw_info.max_payload_size = 0;
}
else
{
pdev->vars.clk_factor = LM_EMUL_FACTOR;
}
val=REG_RD(PFDEV(pdev),MISC_REG_CHIP_METAL);
pdev->hw_info.chip_id |= (val & 0xff) << 4;
val=REG_RD(PFDEV(pdev),MISC_REG_BOND_ID);
pdev->hw_info.chip_id |= (val & 0xf);
DbgMessage(pdev, INFORMi , "chip id 0x%x\n", pdev->hw_info.chip_id);
val=REG_RD(PFDEV(pdev),MISC_REG_CHIP_TEST_REG);
pdev->hw_info.silent_chip_rev = (val & 0xff);
DbgMessage(pdev, INFORMi , "silent chip rev 0x%x\n", pdev->hw_info.silent_chip_rev);
if (!CHIP_IS_E1x(pdev))
{
val = REG_RD(PFDEV(pdev), MISC_REG_PORT4MODE_EN_OVWR);
DbgMessage(pdev, WARN, "MISC_REG_PORT4MODE_EN_OVWR = %d\n", val);
if ((val & 1) == 0)
{
val = REG_RD(PFDEV(pdev), MISC_REG_PORT4MODE_EN);
}
else
{
val = (val >> 1) & 1;
}
pdev->hw_info.chip_port_mode = val? LM_CHIP_PORT_MODE_4 : LM_CHIP_PORT_MODE_2;
DbgMessage(pdev, WARN, "chip_port_mode %s\n", (pdev->hw_info.chip_port_mode == LM_CHIP_PORT_MODE_4 )? "4_PORT" : "2_PORT");
}
else
{
pdev->hw_info.chip_port_mode = LM_CHIP_PORT_MODE_NONE;
DbgMessage(pdev, WARN, "chip_port_mode NONE\n");
}
return LM_STATUS_SUCCESS;
}
static void lm_get_igu_cam_info(lm_device_t *pdev)
{
lm_intr_blk_info_t *blk_info = &pdev->hw_info.intr_blk_info;
u8_t igu_test_vectors = FALSE;
#define IGU_CAM_VFID_MATCH(pdev, igu_fid) (!(igu_fid & IGU_FID_ENCODE_IS_PF) && ((igu_fid & IGU_FID_VF_NUM_MASK) == ABS_VFID(pdev)))
#define IGU_CAM_PFID_MATCH(pdev, igu_fid) ((igu_fid & IGU_FID_ENCODE_IS_PF) && ((igu_fid & IGU_FID_PF_NUM_MASK) == FUNC_ID(pdev)))
if (INTR_BLK_MODE(pdev) == INTR_BLK_MODE_BC)
{
blk_info->igu_info.igu_sb_cnt = MAX_RSS_CHAINS;
blk_info->igu_info.igu_u_sb_offset = 0;
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_2)
{
blk_info->igu_info.igu_base_sb = VNIC_ID(pdev) * MAX_RSS_CHAINS;
blk_info->igu_info.igu_dsb_id = MAX_VNIC_NUM * MAX_RSS_CHAINS + VNIC_ID(pdev);
}
else
{
blk_info->igu_info.igu_base_sb = FUNC_ID(pdev) * MAX_RSS_CHAINS;
blk_info->igu_info.igu_dsb_id = MAX_VNIC_NUM * MAX_RSS_CHAINS + FUNC_ID(pdev);
}
}
else
{
u8_t igu_sb_id;
u8_t fid;
u8_t vec;
u8_t vf_id;
u32_t val;
u8_t current_pf_id = 0;
u8_t recent_vf_id = 0xFF;
blk_info->igu_info.igu_sb_cnt = 0;
blk_info->igu_info.igu_test_sb_cnt = 0;
blk_info->igu_info.igu_base_sb = 0xff;
for (vf_id = 0; vf_id < E2_MAX_NUM_OF_VFS; vf_id++)
{
blk_info->igu_info.vf_igu_info[vf_id].igu_base_sb = 0xFF;
blk_info->igu_info.vf_igu_info[vf_id].igu_sb_cnt = 0;
blk_info->igu_info.vf_igu_info[vf_id].igu_test_sb_cnt = 0;
blk_info->igu_info.vf_igu_info[vf_id].igu_test_mode = FALSE;
}
for (igu_sb_id = 0; igu_sb_id < IGU_REG_MAPPING_MEMORY_SIZE; igu_sb_id++ )
{
lm_igu_block_t * lm_igu_sb = &IGU_SB(pdev,igu_sb_id);
lm_igu_sb->block_dump = val = REG_RD(PFDEV(pdev), IGU_REG_MAPPING_MEMORY + 4*igu_sb_id);
DbgMessage(pdev, WARN, "addr:0x%x IGU_CAM[%d]=%x\n",IGU_REG_MAPPING_MEMORY + 4*igu_sb_id, igu_sb_id, val);
if (!(val & IGU_REG_MAPPING_MEMORY_VALID))
{
if (!IS_MULTI_VNIC(pdev) && (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_2))
{
lm_igu_sb->status = LM_IGU_STATUS_AVAILABLE;
}
else if (current_pf_id == FUNC_ID(pdev))
{
lm_igu_sb->status = LM_IGU_STATUS_AVAILABLE;
}
else
{
lm_igu_sb->status = 0;
}
continue;
}
else
{
lm_igu_sb->status = LM_IGU_STATUS_VALID;
}
fid = (val & IGU_REG_MAPPING_MEMORY_FID_MASK) >> IGU_REG_MAPPING_MEMORY_FID_SHIFT;
if (fid & IGU_FID_ENCODE_IS_PF)
{
current_pf_id = lm_igu_sb->pf_number = fid & IGU_FID_PF_NUM_MASK;
if (lm_igu_sb->pf_number == FUNC_ID(pdev))
{
lm_igu_sb->status |= (LM_IGU_STATUS_AVAILABLE | LM_IGU_STATUS_PF);
}
else
{
lm_igu_sb->status |= LM_IGU_STATUS_PF;
}
}
else
{
lm_igu_sb->vf_number = fid & IGU_FID_VF_NUM_MASK;
if ((lm_igu_sb->vf_number >= pdev->hw_info.sriov_info.first_vf_in_pf)
&& (lm_igu_sb->vf_number < (pdev->hw_info.sriov_info.first_vf_in_pf + pdev->hw_info.sriov_info.total_vfs)))
{
lm_igu_sb->status |= LM_IGU_STATUS_AVAILABLE;
}
}
lm_igu_sb->vector_number = (val & IGU_REG_MAPPING_MEMORY_VECTOR_MASK) >> IGU_REG_MAPPING_MEMORY_VECTOR_SHIFT;
DbgMessage(pdev, VERBOSEi, "FID[%d]=%d\n", igu_sb_id, fid);
if ((IS_PFDEV(pdev) && IGU_CAM_PFID_MATCH(pdev, fid)) ||
(IS_VFDEV(pdev) && IGU_CAM_VFID_MATCH(pdev, fid)))
{
vec = (val & IGU_REG_MAPPING_MEMORY_VECTOR_MASK) >> IGU_REG_MAPPING_MEMORY_VECTOR_SHIFT;
DbgMessage(pdev, INFORMi, "VEC[%d]=%d\n", igu_sb_id, vec);
if (igu_test_vectors)
{
blk_info->igu_info.igu_test_sb_cnt++;
}
else
{
if (vec == 0 && IS_PFDEV(pdev))
{
blk_info->igu_info.igu_dsb_id = igu_sb_id;
}
else
{
if (blk_info->igu_info.igu_base_sb == 0xff)
{
blk_info->igu_info.igu_base_sb = igu_sb_id;
}
blk_info->igu_info.igu_sb_cnt++;
}
}
if (recent_vf_id != 0xFF)
{
if (!blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode)
{
DbgMessage(pdev, WARN, "Consecutiveness of IGU for VF%d is broken. My be it's IGU test mode\n",recent_vf_id);
}
blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode = TRUE;
}
}
else if (IS_CHANNEL_VIRT_MODE_MASTER_PFDEV(pdev))
{
if (!(fid & IGU_FID_ENCODE_IS_PF))
{
vf_id = fid & IGU_FID_VF_NUM_MASK;
if (blk_info->igu_info.vf_igu_info[vf_id].igu_base_sb == 0xff)
{
blk_info->igu_info.vf_igu_info[vf_id].igu_base_sb = igu_sb_id;
}
if (recent_vf_id != vf_id)
{
if (recent_vf_id != 0xFF)
{
if (!blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode)
{
DbgMessage(pdev, WARN, "Consecutiveness of IGU for VF%d is broken. My be it's IGU test mode\n",recent_vf_id);
}
blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode = TRUE;
}
}
recent_vf_id = vf_id;
if (blk_info->igu_info.vf_igu_info[vf_id].igu_test_mode)
{
blk_info->igu_info.vf_igu_info[vf_id].igu_test_sb_cnt++;
}
else
{
blk_info->igu_info.vf_igu_info[vf_id].igu_sb_cnt++;
}
}
else
{
if (recent_vf_id != 0xFF)
{
if (!blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode)
{
DbgMessage(pdev, WARN, "Consecutiveness of IGU for VF%d is broken. My be it's IGU test mode\n",recent_vf_id);
}
blk_info->igu_info.vf_igu_info[recent_vf_id].igu_test_mode = TRUE;
}
}
if (blk_info->igu_info.igu_base_sb != 0xff)
{
if (!igu_test_vectors)
{
DbgMessage(pdev, WARN, "Consecutiveness of IGU is broken. My be it's IGU test mode\n");
}
igu_test_vectors = TRUE;
}
}
else
{
if (blk_info->igu_info.igu_base_sb != 0xff)
{
if (!igu_test_vectors) {
DbgMessage(pdev, WARN, "Consecutiveness of IGU is broken. My be it's IGU test mode\n");
}
igu_test_vectors = TRUE;
}
}
}
#ifndef _VBD_
blk_info->igu_info.igu_sb_cnt = min(blk_info->igu_info.igu_sb_cnt, (u8_t)16);
#endif
DbgMessage(pdev, WARN, "igu_sb_cnt=%d igu_dsb_id=%d igu_base_sb = %d igu_us_sb_offset = %d igu_test_cnt=%d\n",
blk_info->igu_info.igu_sb_cnt, blk_info->igu_info.igu_dsb_id, blk_info->igu_info.igu_base_sb, blk_info->igu_info.igu_u_sb_offset,
blk_info->igu_info.igu_test_sb_cnt);
if ((FUNC_ID(pdev) < 2) && (pdev->hw_info.mf_info.mf_mode != SINGLE_FUNCTION))
{
if (blk_info->igu_info.igu_sb_cnt < 1)
{
DbgMessage(pdev, FATAL, "Igu sb cnt is not valid value=%d\n", blk_info->igu_info.igu_sb_cnt);
}
if (blk_info->igu_info.igu_base_sb == 0xff)
{
DbgMessage(pdev, FATAL, "Igu base sb is not valid value=%d\n", blk_info->igu_info.igu_base_sb);
}
}
#define IGU_MAX_INTA_SB_CNT 31
if (pdev->params.b_inta_mode_prvided_by_os && (blk_info->igu_info.igu_sb_cnt > IGU_MAX_INTA_SB_CNT ))
{
blk_info->igu_info.igu_sb_cnt = IGU_MAX_INTA_SB_CNT ;
}
}
DbgMessage(pdev, WARN, "IGU CAM INFO: BASE_SB: %d DSB: %d IGU_SB_CNT: %d\n", blk_info->igu_info.igu_base_sb, blk_info->igu_info.igu_dsb_id, blk_info->igu_info.igu_sb_cnt);
}
lm_status_t lm_get_intr_blk_info(lm_device_t *pdev)
{
lm_intr_blk_info_t *blk_info = &pdev->hw_info.intr_blk_info;
u32_t bar_base;
u8_t igu_func_id = 0;
if (CHIP_IS_E1x(pdev))
{
blk_info->blk_type = INTR_BLK_HC;
blk_info->access_type = INTR_BLK_ACCESS_GRC;
blk_info->blk_mode = INTR_BLK_MODE_NORM;
blk_info->simd_addr_womask = HC_REG_COMMAND_REG + PORT_ID(pdev)*32 + COMMAND_REG_SIMD_NOMASK;
if (CHIP_REV_IS_EMUL(pdev))
{
blk_info->simd_addr_wmask = HC_REG_COMMAND_REG + PORT_ID(pdev)*32 + COMMAND_REG_SIMD_NOMASK;
}
else
{
blk_info->simd_addr_wmask = HC_REG_COMMAND_REG + PORT_ID(pdev)*32 + COMMAND_REG_SIMD_MASK;
}
}
else
{
ASSERT_STATIC(MAX_RSS_CHAINS <= 32);
pdev->hw_info.intr_blk_info.blk_type = INTR_BLK_IGU;
if (REG_RD(PFDEV(pdev), IGU_REG_BLOCK_CONFIGURATION) & IGU_BLOCK_CONFIGURATION_REG_BACKWARD_COMP_EN)
{
DbgMessage(pdev, FATAL, "IGU Backward Compatible Mode\n");
blk_info->blk_mode = INTR_BLK_MODE_BC;
}
else
{
DbgMessage(pdev, WARN, "IGU Normal Mode\n");
blk_info->blk_mode = INTR_BLK_MODE_NORM;
}
lm_get_igu_cam_info(pdev);
igu_func_id = (1 << IGU_FID_ENCODE_IS_PF_SHIFT) | FUNC_ID(pdev);
blk_info->igu_info.igu_func_id = igu_func_id;
if (pdev->params.igu_access_mode == INTR_BLK_ACCESS_GRC)
{
DbgMessage(pdev, FATAL, "IGU - GRC\n");
if (IS_VFDEV(pdev))
{
DbgBreakMsg("VF Can't work in GRC Access mode!\n");
return LM_STATUS_FAILURE;
}
blk_info->access_type = INTR_BLK_ACCESS_GRC;
blk_info->cmd_ctrl_rd_womask =
((IGU_REG_SISR_MDPC_WOMASK_UPPER << IGU_CTRL_REG_ADDRESS_SHIFT) |
(igu_func_id << IGU_CTRL_REG_FID_SHIFT) |
(IGU_CTRL_CMD_TYPE_RD << IGU_CTRL_REG_TYPE_SHIFT));
blk_info->simd_addr_womask = IGU_REG_COMMAND_REG_32LSB_DATA;
if (CHIP_REV_IS_EMUL(pdev))
{
blk_info->cmd_ctrl_rd_wmask =
((IGU_REG_SISR_MDPC_WOMASK_UPPER << IGU_CTRL_REG_ADDRESS_SHIFT) |
(igu_func_id << IGU_CTRL_REG_FID_SHIFT) |
(IGU_CTRL_CMD_TYPE_RD << IGU_CTRL_REG_TYPE_SHIFT));
}
else
{
blk_info->cmd_ctrl_rd_wmask =
((IGU_REG_SISR_MDPC_WMASK_LSB_UPPER << IGU_CTRL_REG_ADDRESS_SHIFT) |
(igu_func_id << IGU_CTRL_REG_FID_SHIFT) |
(IGU_CTRL_CMD_TYPE_RD << IGU_CTRL_REG_TYPE_SHIFT));
}
blk_info->simd_addr_wmask = IGU_REG_COMMAND_REG_32LSB_DATA;
}
else
{
DbgMessage(pdev, WARN, "IGU - IGUMEM\n");
blk_info->access_type = INTR_BLK_ACCESS_IGUMEM;
bar_base = IS_PFDEV(pdev)? BAR_IGU_INTMEM : VF_BAR0_IGU_OFFSET;
blk_info->simd_addr_womask = bar_base + IGU_REG_SISR_MDPC_WOMASK_UPPER*8;
if (CHIP_REV_IS_EMUL(pdev))
{
blk_info->simd_addr_wmask = bar_base + IGU_REG_SISR_MDPC_WOMASK_UPPER*8;
}
else
{
blk_info->simd_addr_wmask = bar_base + IGU_REG_SISR_MDPC_WMASK_LSB_UPPER*8;
}
}
}
return LM_STATUS_SUCCESS;
}
lm_status_t lm_get_nvm_info(lm_device_t *pdev)
{
u32_t val = REG_RD(pdev,MCP_REG_MCPR_NVM_CFG4);
pdev->hw_info.flash_spec.total_size = NVRAM_1MB_SIZE << (val & MCPR_NVM_CFG4_FLASH_SIZE);
pdev->hw_info.flash_spec.page_size = NVRAM_PAGE_SIZE;
return LM_STATUS_SUCCESS;
}
#if defined(DOS) || defined(__LINUX)
#define DEFAULT_CONNECTIONS_TOE 1880
#define MAX_CONNECTIONS 2048
#define MAX_CONNECTIONS_ISCSI 128
#define MAX_CONNECTIONS_RDMA 10
#define MAX_CONNECTIONS_TOE 1880
#define MAX_CONNECTIONS_FCOE 0
#define MAX_CONNECTIONS_VF 128
#else
#define MAX_CONNECTIONS (min(16384,(32768 / (log2_align(pdev->hw_info.mf_info.vnics_per_port)))))
#define DEFAULT_CONNECTIONS_TOE 1880
#define MAX_CONNECTIONS_ISCSI 128
#define MAX_CONNECTIONS_RDMA 10
#define MAX_CONNECTIONS_FCOE 1024
#define MAX_CONNECTIONS_VF (1 << (LM_VF_MAX_RVFID_SIZE + LM_MAX_VF_CID_WND_SIZE + 1))
#define MAX_CONNECTIONS_TOE (min(8192,MAX_CONNECTIONS - MAX_CONNECTIONS_ISCSI - MAX_CONNECTIONS_RDMA - MAX_CONNECTIONS_FCOE - MAX_ETH_CONS - MAX_CONNECTIONS_VF))
#endif
#define MAX_CONNECTIONS_TOE_NO_LICENSE 0
#define MAX_CONNECTIONS_ISCSI_NO_LICENSE 0
#define MAX_CONNECTIONS_RDMA_NO_LICENSE 0
#define MAX_CONNECTIONS_FCOE_NO_LICENSE 0
#define MAX_CONNECTIONS_FCOE_NO_MCP 128
static u32_t lm_parse_license_info(u32 val, u8_t is_high)
{
if (is_high)
{
val &=0xFFFF0000;
if(val)
{
val ^= FW_ENCODE_32BIT_PATTERN;
}
val >>= 16;
}
else
{
val &= 0xffff;
if(val)
{
val ^= FW_ENCODE_16BIT_PATTERN;
}
}
return val;
}
static u32_t lm_parse_license_info_bounded(u32 val, u32_t max_cons, u8_t is_high)
{
u32_t license_from_shmem =0;
license_from_shmem = lm_parse_license_info(val, is_high);
val = min(license_from_shmem, max_cons);
return val;
}
#define NO_MCP_WA_CFG_SET_ADDR (0xA0000)
#define NO_MCP_WA_CFG_SET_MAGIC (0x88AA55FF)
#define NO_MCP_WA_MULTI_VNIC_MODE (0xA0004)
#define NO_MCP_WA_VNICS_PER_PORT(port) (0xA0008 + 4*(port))
#define NO_MCP_WA_OVLAN(func) (0xA0010 + 4*(func))
#define NO_MCP_WA_FORCE_5710 (0xA0030)
#define NO_MCP_WA_VALID_LIC_ADDR (0xA0040)
#define NO_MCP_WA_VALID_LIC_MAGIC (0xCCAAFFEE)
#define NO_MCP_WA_TOE_LIC (0xA0048)
#define NO_MCP_WA_ISCSI_LIC (0xA0050)
#define NO_MCP_WA_RDMA_LIC (0xA0058)
#define NO_MCP_WA_CLC_SHMEM (0xAF900)
static lm_status_t lm_get_shmem_license_info(lm_device_t *pdev)
{
u32_t max_toe_cons[PORT_MAX] = {0,0};
u32_t max_rdma_cons[PORT_MAX] = {0,0};
u32_t max_iscsi_cons[PORT_MAX] = {0,0};
u32_t max_fcoe_cons[PORT_MAX] = {0,0};
u32_t max_eth_cons[PORT_MAX] = {0,0};
u32_t max_bar_supported_cons[PORT_MAX] = {0};
u32_t max_supported_cons[PORT_MAX] = {0};
u32_t val = 0;
u8_t port = 0;
u32_t offset = 0;
pdev->hw_info.max_common_conns = 0;
for (port = 0; port < PORT_MAX; port++)
{
if (pdev->hw_info.mcp_detected == 1)
{
LM_SHMEM_READ(pdev, OFFSETOF(shmem_region_t, validity_map[port]),&val);
if(GET_FLAGS(val, SHR_MEM_VALIDITY_LIC_MANUF_KEY_IN_EFFECT | SHR_MEM_VALIDITY_LIC_UPGRADE_KEY_IN_EFFECT))
{
offset = OFFSETOF(shmem_region_t, drv_lic_key[port].max_toe_conn) & 0xfffffffc;
LM_SHMEM_READ(pdev, offset, &val);
max_toe_cons[port] = lm_parse_license_info_bounded(val, MAX_CONNECTIONS_TOE,FALSE);
DbgMessage(pdev, INFORMi, "max_toe_conn from shmem %d for port %d\n",val, port);
offset = OFFSETOF(shmem_region_t, drv_lic_key[port].max_um_rdma_conn) & 0xfffffffc;
LM_SHMEM_READ(pdev, offset, &val);
max_rdma_cons[port] = lm_parse_license_info_bounded(val, MAX_CONNECTIONS_RDMA,FALSE);
DbgMessage(pdev, INFORMi, "max_rdma_conn from shmem %d for port %d\n",val, port);
offset = OFFSETOF(shmem_region_t, drv_lic_key[port].max_iscsi_trgt_conn) & 0xfffffffc;
LM_SHMEM_READ(pdev, offset, &val);
max_iscsi_cons[port] = lm_parse_license_info_bounded(val, MAX_CONNECTIONS_ISCSI,TRUE);
DbgMessage(pdev, INFORMi, "max_iscsi_conn from shmem %d for port %d\n",val, port);
offset = OFFSETOF(shmem_region_t, drv_lic_key[port].max_fcoe_init_conn) & 0xfffffffc;
LM_SHMEM_READ(pdev, offset, &val);
if(0 == lm_parse_license_info(val,TRUE))
{
max_fcoe_cons[port] = 0;
}
else
{
max_fcoe_cons[port] = MAX_CONNECTIONS_FCOE;
}
DbgMessage(pdev, INFORMi, "max_fcoe_conn from shmem %d for port %d\n",val, port);
}
else
{
max_toe_cons[port] = MAX_CONNECTIONS_TOE_NO_LICENSE;
max_rdma_cons[port] = MAX_CONNECTIONS_ISCSI_NO_LICENSE;
max_iscsi_cons[port] = MAX_CONNECTIONS_RDMA_NO_LICENSE;
max_fcoe_cons[port] = MAX_CONNECTIONS_FCOE_NO_LICENSE;
}
if (CHIP_IS_E1x(pdev))
{
max_eth_cons[port] = MAX_ETH_REG_CONS;
}
else
{
max_eth_cons[port] = MAX_CONNECTIONS_VF;
}
if (PORT_ID(pdev) == port)
{
max_bar_supported_cons[port] = pdev->hw_info.bar_size[BAR_1] / LM_DQ_CID_SIZE;
}
else
{
LM_SHMEM_READ(pdev, OFFSETOF(shmem_region_t, dev_info.port_feature_config[port].config), &val);
val = (val & PORT_FEAT_CFG_BAR2_SIZE_MASK) >> PORT_FEAT_CFG_BAR2_SIZE_SHIFT;
if (val != 0)
{
val = (0x40 << ((val - 1)))*0x400;
}
max_bar_supported_cons[port] = val / LM_DQ_CID_SIZE;
}
}
else
{
LM_SHMEM_READ(pdev, NO_MCP_WA_VALID_LIC_ADDR+4*port, &val);
if (val == NO_MCP_WA_VALID_LIC_MAGIC)
{
LM_SHMEM_READ(pdev, NO_MCP_WA_TOE_LIC+4*port, &val);
max_toe_cons[port] = val;
LM_SHMEM_READ(pdev, NO_MCP_WA_ISCSI_LIC+4*port, &val);
max_iscsi_cons[port] = val;
LM_SHMEM_READ(pdev, NO_MCP_WA_RDMA_LIC+4*port, &val);
max_rdma_cons[port] = val;
max_fcoe_cons[port] = MAX_CONNECTIONS_FCOE_NO_MCP;
}
else
{
#ifdef VF_INVOLVED
max_toe_cons[port] = DEFAULT_CONNECTIONS_TOE - 100;
#else
max_toe_cons[port] = DEFAULT_CONNECTIONS_TOE;
#endif
max_iscsi_cons[port] = MAX_CONNECTIONS_ISCSI;
max_rdma_cons[port] = MAX_CONNECTIONS_RDMA;
max_fcoe_cons[port] = MAX_CONNECTIONS_FCOE_NO_MCP;
}
if (CHIP_IS_E1x(pdev))
{
max_eth_cons[port] = MAX_ETH_REG_CONS;
}
else
{
max_eth_cons[port] = MAX_CONNECTIONS_VF;
}
max_bar_supported_cons[port] = pdev->hw_info.bar_size[BAR_1] / LM_DQ_CID_SIZE;
}
max_supported_cons[port] = log2_align(max_toe_cons[port] + max_rdma_cons[port] + max_iscsi_cons[port] + max_fcoe_cons[port] + max_eth_cons[port]);
max_supported_cons[port] = min(max_supported_cons[port], max_bar_supported_cons[port]);
#if defined(DOS) || defined(__LINUX)
max_supported_cons[port] = min(max_supported_cons[port], (u32_t)1024);
#endif
if (max_supported_cons[port] > pdev->hw_info.max_common_conns)
{
pdev->hw_info.max_common_conns = max_supported_cons[port];
}
}
port = PORT_ID(pdev);
if (ERR_IF(max_supported_cons[port] < max_eth_cons[port]))
{
return LM_STATUS_INVALID_PARAMETER;
}
if ((max_iscsi_cons[port] + max_rdma_cons[port] + max_toe_cons[port] + max_fcoe_cons[port] + max_eth_cons[port]) > max_supported_cons[port])
{
if ((max_iscsi_cons[port] + max_rdma_cons[port] + max_fcoe_cons[port] + max_eth_cons[port]) > max_supported_cons[port])
{
u32_t s = max_iscsi_cons[port] + max_rdma_cons[port] + max_toe_cons[port] + max_fcoe_cons[port];
u32_t t = max_supported_cons[port] - pdev->params.max_eth_including_vfs_conns;
max_iscsi_cons[port] *=t;
max_iscsi_cons[port] /=s;
max_rdma_cons[port] *=t;
max_rdma_cons[port] /=s;
max_toe_cons[port] *=t;
max_toe_cons[port] /=s;
max_fcoe_cons[port] *=t;
max_fcoe_cons[port] /=s;
}
else
{
max_toe_cons[port] = max_supported_cons[port] - (max_iscsi_cons[port] + max_rdma_cons[port] + max_fcoe_cons[port] + max_eth_cons[port]);
}
}
if (ERR_IF((max_iscsi_cons[port] + max_rdma_cons[port] + max_fcoe_cons[port] + max_toe_cons[port] + max_eth_cons[port]) > max_supported_cons[port]))
{
return LM_STATUS_INVALID_PARAMETER;
}
pdev->hw_info.max_port_toe_conn = max_toe_cons[port];
DbgMessage(pdev, INFORMi, "max_toe_conn from shmem %d\n",pdev->hw_info.max_port_toe_conn);
pdev->hw_info.max_port_rdma_conn = max_rdma_cons[port];
DbgMessage(pdev, INFORMi, "max_rdma_conn from shmem %d\n",pdev->hw_info.max_port_rdma_conn);
pdev->hw_info.max_port_iscsi_conn = max_iscsi_cons[port];
DbgMessage(pdev, INFORMi, "max_iscsi_conn from shmem %d\n",pdev->hw_info.max_port_iscsi_conn);
pdev->hw_info.max_port_fcoe_conn = max_fcoe_cons[port];
DbgMessage(pdev, INFORMi, "max_fcoe_conn from shmem %d\n",pdev->hw_info.max_port_fcoe_conn);
pdev->hw_info.max_port_conns = log2_align(pdev->hw_info.max_port_toe_conn +
pdev->hw_info.max_port_rdma_conn + pdev->hw_info.max_port_iscsi_conn
+ pdev->hw_info.max_port_fcoe_conn + pdev->params.max_eth_including_vfs_conns);
if (ERR_IF(pdev->hw_info.max_port_conns > max_bar_supported_cons[port]))
{
return LM_STATUS_INVALID_PARAMETER;
}
return LM_STATUS_SUCCESS;
}
static lm_status_t lm_check_valid_mf_cfg(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
lm_status_t lm_status = LM_STATUS_SUCCESS;
const u8_t func_id = FUNC_ID(pdev);
u8_t i = 0;
u8_t j = 0;
u32_t mf_cfg1 = 0;
u32_t mf_cfg2 = 0;
u32_t ovlan1 = 0;
u32_t ovlan2 = 0;
u32_t dynamic_cfg = 0;
ASSERT_STATIC(OFFSETOF(mf_cfg_t,shared_mf_config.clp_mb) == 0);
ASSERT_STATIC(OFFSETOF(mf_cfg_t,func_mf_config) == 36);
ASSERT_STATIC(OFFSETOF(func_mf_cfg_t,config) == 0);
ASSERT_STATIC(FUNC_MF_CFG_FUNC_HIDE == 0x1);
ASSERT_STATIC(FUNC_MF_CFG_PROTOCOL_ETHERNET_WITH_RDMA == 0x4);
ASSERT_STATIC(FUNC_MF_CFG_FUNC_DISABLED == 0x8);
ASSERT_STATIC(OFFSETOF(func_mf_cfg_t,mac_upper) == 4);
ASSERT_STATIC(OFFSETOF(func_mf_cfg_t,mac_lower) == 8);
ASSERT_STATIC(FUNC_MF_CFG_UPPERMAC_DEFAULT == 0x0000ffff);
ASSERT_STATIC(FUNC_MF_CFG_LOWERMAC_DEFAULT == 0xffffffff);
ASSERT_STATIC(OFFSETOF(func_mf_cfg_t,e1hov_tag) == 12);
ASSERT_STATIC(FUNC_MF_CFG_E1HOV_TAG_DEFAULT == 0x0000ffff);
ASSERT_STATIC(sizeof(func_mf_cfg_t) == 24);
DbgMessage(pdev, INFORMi, "MF cfg parameters for function %d:\n", func_id);
DbgMessage(pdev, INFORMi, "\t func_mf_cfg=0x%x\n\t multi_vnics_mode=%d\n\t vnics_per_port=%d\n\t ovlan/vifid=%d\n\t min_bw=%d\n\t max_bw=%d\n",
mf_info->func_mf_cfg,
mf_info->vnics_per_port,
mf_info->multi_vnics_mode,
mf_info->ext_id,
mf_info->min_bw,
mf_info->max_bw);
DbgMessage(pdev, INFORMi, "\t mac addr (overiding main and iscsi): %02x %02x %02x %02x %02x %02x\n",
pdev->hw_info.mac_addr[0],
pdev->hw_info.mac_addr[1],
pdev->hw_info.mac_addr[2],
pdev->hw_info.mac_addr[3],
pdev->hw_info.mac_addr[4],
pdev->hw_info.mac_addr[5]);
if (GET_FLAGS(mf_info->func_mf_cfg, FUNC_MF_CFG_FUNC_HIDE))
{
DbgMessage(pdev, FATAL, "Enumerated function %d, is marked as hidden\n", func_id);
lm_status = LM_STATUS_FAILURE;
goto _end;
}
if (mf_info->vnics_per_port > 1 && !mf_info->multi_vnics_mode)
{
DbgMessage(pdev, FATAL, "invalid mf mode configuration: vnics_per_port=%d, multi_vnics_mode=%d\n",
mf_info->vnics_per_port,
mf_info->multi_vnics_mode);
lm_status = LM_STATUS_FAILURE;
goto _end;
}
if (mf_info->mf_mode == MULTI_FUNCTION_SD)
{
if ((VNIC_ID(pdev) > 0) && !VALID_OVLAN(OVLAN(pdev)))
{
DbgMessage(pdev, WARNi, "invalid mf mode configuration: VNICID=%d, Function is enumerated, ovlan (%d) is invalid\n",
VNIC_ID(pdev), OVLAN(pdev));
#ifdef EDIAG
if (mf_info->sd_mode != SD_UFP_MODE)
{
lm_status = LM_STATUS_FAILURE;
}
#else
lm_status = LM_STATUS_FAILURE;
#endif
goto _end;
}
if (!VALID_OVLAN(OVLAN(pdev)) && mf_info->multi_vnics_mode)
{
DbgMessage(pdev, FATAL, "invalid mf mode configuration: multi_vnics_mode=%d, ovlan=%d\n",
mf_info->multi_vnics_mode,
OVLAN(pdev));
#ifdef EDIAG
if (mf_info->sd_mode != SD_UFP_MODE)
{
lm_status = LM_STATUS_FAILURE;
}
#else
lm_status = LM_STATUS_FAILURE;
#endif
goto _end;
}
LM_FOREACH_ABS_FUNC_IN_PORT(pdev, i)
{
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[i].config),&mf_cfg1);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[i].e1hov_tag), &ovlan1);
if (!GET_FLAGS(mf_cfg1, FUNC_MF_CFG_FUNC_HIDE) &&
(((mf_info->multi_vnics_mode) && !VALID_OVLAN(ovlan1)) ||
((!mf_info->multi_vnics_mode) && VALID_OVLAN(ovlan1))))
{
#ifdef EDIAG
if (mf_info->sd_mode != SD_UFP_MODE)
{
lm_status = LM_STATUS_FAILURE;
}
#else
lm_status= LM_STATUS_FAILURE;
#endif
goto _end;
}
}
LM_FOREACH_ABS_FUNC_IN_PORT(pdev, i)
{
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[i].config),&mf_cfg1);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[i].e1hov_tag), &ovlan1);
for (j = i + 2; j < E1H_FUNC_MAX; j += 2)
{
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[j].config),&mf_cfg2);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[j].e1hov_tag), &ovlan2);
if (!GET_FLAGS(mf_cfg1, FUNC_MF_CFG_FUNC_HIDE) && VALID_OVLAN(ovlan1) &&
!GET_FLAGS(mf_cfg2, FUNC_MF_CFG_FUNC_HIDE) && VALID_OVLAN(ovlan2) &&
(ovlan1 == ovlan2) )
{
lm_status = LM_STATUS_FAILURE;
DbgBreakIf(1);
goto _end;
}
}
}
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, port_mf_config[PATH_ID(pdev)][PORT_ID(pdev)].dynamic_cfg),&dynamic_cfg );
if( PORT_MF_CFG_E1HOV_TAG_DEFAULT == ( dynamic_cfg & PORT_MF_CFG_E1HOV_TAG_MASK ) )
{
pdev->hw_info.is_dcc_active = FALSE;
}
else
{
pdev->hw_info.is_dcc_active = TRUE;
}
}
_end:
return lm_status;
}
void lm_cmng_get_shmem_info( lm_device_t* pdev )
{
u32_t val = 0;
u8_t i = 0;
u8_t vnic = 0;
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;;
if( !IS_MF_MODE_CAPABLE(pdev) )
{
DbgBreakIf(1) ;
return;
}
LM_FOREACH_ABS_FUNC_IN_PORT(pdev, i)
{
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[i].config),&val);
mf_info->min_bw[vnic] = (GET_FLAGS(val, FUNC_MF_CFG_MIN_BW_MASK) >> FUNC_MF_CFG_MIN_BW_SHIFT);
mf_info->max_bw[vnic] = (GET_FLAGS(val, FUNC_MF_CFG_MAX_BW_MASK) >> FUNC_MF_CFG_MAX_BW_SHIFT);
vnic++;
}
}
static u8_t lm_get_vnics_per_port(lm_device_t* pdev)
{
if (pdev->hw_info.mf_info.multi_vnics_mode)
{
return LM_PFS_PER_PORT(pdev);
}
else
{
return 1;
}
}
static lm_status_t lm_get_shmem_mf_cfg_info_sd(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t val = 0;
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].e1hov_tag),&val);
mf_info->ext_id = (u16_t)val;
mf_info->multi_vnics_mode = 1;
if(!VALID_OVLAN(OVLAN(pdev)))
{
DbgMessage(pdev, FATAL, "Invalid mf mode configuration: VNICID=%d, Function is enumerated, ovlan (%d) is invalid\n",
VNIC_ID(pdev), OVLAN(pdev));
#ifdef EDIAG
if (mf_info->sd_mode != SD_UFP_MODE)
{
return LM_STATUS_FAILURE;
}
#else
return LM_STATUS_FAILURE;
#endif
}
if (GET_FLAGS(mf_info->func_mf_cfg, FUNC_MF_CFG_PROTOCOL_MASK) == FUNC_MF_CFG_PROTOCOL_ISCSI)
{
pdev->params.mf_proto_support_flags |= LM_PROTO_SUPPORT_ISCSI;
}
else if (GET_FLAGS(mf_info->func_mf_cfg, FUNC_MF_CFG_PROTOCOL_MASK) == FUNC_MF_CFG_PROTOCOL_FCOE)
{
pdev->params.mf_proto_support_flags |= LM_PROTO_SUPPORT_FCOE;
}
else
{
pdev->params.mf_proto_support_flags |= LM_PROTO_SUPPORT_ETHERNET;
}
mf_info->vnics_per_port = lm_get_vnics_per_port(pdev);
return LM_STATUS_SUCCESS;
}
static lm_status_t lm_get_shmem_mf_cfg_info_sd_bd(lm_device_t *pdev)
{
lm_status_t lm_status = lm_get_shmem_mf_cfg_info_sd(pdev);
return lm_status;
}
static lm_status_t lm_get_shmem_mf_cfg_info_sd_ufp(lm_device_t *pdev)
{
lm_status_t lm_status = lm_get_shmem_mf_cfg_info_sd(pdev);
return lm_status;
}
static void _copy_mac_upper_lower_to_arr(IN u32_t mac_upper, IN u32_t mac_lower, OUT u8_t* mac_addr)
{
if(mac_addr)
{
mac_addr[0] = (u8_t) (mac_upper >> 8);
mac_addr[1] = (u8_t) mac_upper;
mac_addr[2] = (u8_t) (mac_lower >> 24);
mac_addr[3] = (u8_t) (mac_lower >> 16);
mac_addr[4] = (u8_t) (mac_lower >> 8);
mac_addr[5] = (u8_t) mac_lower;
}
}
static void lm_get_shmem_ext_mac_addresses(lm_device_t *pdev)
{
u32_t mac_upper = 0;
u32_t mac_lower = 0;
u32_t offset = 0;
const u8_t abs_func_id = ABS_FUNC_ID(pdev);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].iscsi_mac_addr_upper);
LM_MFCFG_READ(pdev, offset, &mac_upper);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].iscsi_mac_addr_lower);
LM_MFCFG_READ(pdev, offset, &mac_lower);
_copy_mac_upper_lower_to_arr(mac_upper, mac_lower, pdev->hw_info.iscsi_mac_addr);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_mac_addr_upper);
LM_MFCFG_READ(pdev, offset, &mac_upper);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_mac_addr_lower);
LM_MFCFG_READ(pdev, offset, &mac_lower);
_copy_mac_upper_lower_to_arr(mac_upper, mac_lower, pdev->hw_info.fcoe_mac_addr);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_wwn_port_name_upper);
LM_MFCFG_READ(pdev, offset, &mac_upper);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_wwn_port_name_lower);
LM_MFCFG_READ(pdev, offset, &mac_lower);
_copy_mac_upper_lower_to_arr(mac_upper, mac_lower, &(pdev->hw_info.fcoe_wwn_port_name[2]));
pdev->hw_info.fcoe_wwn_port_name[0] = (u8_t) (mac_upper >> 24);
pdev->hw_info.fcoe_wwn_port_name[1] = (u8_t) (mac_upper >> 16);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_wwn_node_name_upper);
LM_MFCFG_READ(pdev, offset, &mac_upper);
offset = OFFSETOF(mf_cfg_t, func_ext_config[abs_func_id].fcoe_wwn_node_name_lower);
LM_MFCFG_READ(pdev, offset, &mac_lower);
_copy_mac_upper_lower_to_arr(mac_upper, mac_lower, &(pdev->hw_info.fcoe_wwn_node_name[2]));
pdev->hw_info.fcoe_wwn_node_name[0] = (u8_t) (mac_upper >> 24);
pdev->hw_info.fcoe_wwn_node_name[1] = (u8_t) (mac_upper >> 16);
}
static u32_t
lm_get_shmem_ext_proto_support_flags(lm_device_t *pdev)
{
u32_t func_ext_cfg = 0;
u32_t proto_support_flags = 0;
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_ext_config[ABS_FUNC_ID(pdev)].func_cfg),&func_ext_cfg);
if (GET_FLAGS(func_ext_cfg, MACP_FUNC_CFG_FLAGS_ENABLED ))
{
if (GET_FLAGS(func_ext_cfg, MACP_FUNC_CFG_FLAGS_ETHERNET))
{
proto_support_flags |= LM_PROTO_SUPPORT_ETHERNET;
}
if (GET_FLAGS(func_ext_cfg, MACP_FUNC_CFG_FLAGS_ISCSI_OFFLOAD))
{
proto_support_flags |= LM_PROTO_SUPPORT_ISCSI;
}
if (GET_FLAGS(func_ext_cfg, MACP_FUNC_CFG_FLAGS_FCOE_OFFLOAD))
{
proto_support_flags |= LM_PROTO_SUPPORT_FCOE;
}
}
return proto_support_flags;
}
static lm_status_t lm_get_shmem_mf_cfg_info_si(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t val = 0;
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_ext_config[ABS_FUNC_ID(pdev)].func_cfg),&val);
val = val & MACP_FUNC_CFG_FLAGS_MASK;
mf_info->multi_vnics_mode = (val != 0);
mf_info->path_has_ovlan = FALSE;
pdev->params.mf_proto_support_flags = lm_get_shmem_ext_proto_support_flags(pdev);
mf_info->vnics_per_port = lm_get_vnics_per_port(pdev);
return LM_STATUS_SUCCESS;
}
lm_status_t lm_get_shmem_mf_cfg_info_niv(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t func_config = 0;
u32_t niv_config = 0;
u32_t e1hov_tag = 0;
mf_info->multi_vnics_mode = TRUE;
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].e1hov_tag),&e1hov_tag);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].config), &func_config);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].afex_config), &niv_config);
mf_info->ext_id = (u16_t)(GET_FLAGS(e1hov_tag, FUNC_MF_CFG_E1HOV_TAG_MASK)>>FUNC_MF_CFG_E1HOV_TAG_SHIFT);
mf_info->default_vlan = (u16_t)(GET_FLAGS(e1hov_tag, FUNC_MF_CFG_AFEX_VLAN_MASK)>>FUNC_MF_CFG_AFEX_VLAN_SHIFT);
mf_info->niv_allowed_priorities = (u8_t)(GET_FLAGS(niv_config, FUNC_MF_CFG_AFEX_COS_FILTER_MASK)>>FUNC_MF_CFG_AFEX_COS_FILTER_SHIFT);
mf_info->niv_default_cos = (u8_t)(GET_FLAGS(func_config, FUNC_MF_CFG_TRANSMIT_PRIORITY_MASK)>>FUNC_MF_CFG_TRANSMIT_PRIORITY_SHIFT);
mf_info->afex_vlan_mode = GET_FLAGS(niv_config, FUNC_MF_CFG_AFEX_VLAN_MODE_MASK)>>FUNC_MF_CFG_AFEX_VLAN_MODE_SHIFT;
mf_info->niv_mba_enabled = GET_FLAGS(niv_config, FUNC_MF_CFG_AFEX_MBA_ENABLED_MASK)>>FUNC_MF_CFG_AFEX_MBA_ENABLED_SHIFT;
pdev->params.mf_proto_support_flags = lm_get_shmem_ext_proto_support_flags(pdev);
mf_info->vnics_per_port = lm_get_vnics_per_port(pdev);
return LM_STATUS_SUCCESS;
}
static lm_status_t lm_shmem_set_default(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u8_t i;
mf_info->multi_vnics_mode = 0;
mf_info->vnics_per_port = 1;
mf_info->ext_id = 0xffff;
ASSERT_STATIC( ARRSIZE(mf_info->min_bw) == ARRSIZE(mf_info->max_bw) )
for (i = 0; i < ARRSIZE(mf_info->min_bw); i++)
{
mf_info->min_bw[i] = 0;
mf_info->max_bw[i] = 200;
}
pdev->hw_info.shmem_base = 0;
pdev->hw_info.max_port_toe_conn = MAX_CONNECTIONS_TOE;
pdev->hw_info.max_port_rdma_conn = MAX_CONNECTIONS_RDMA;
pdev->hw_info.max_port_iscsi_conn = MAX_CONNECTIONS_ISCSI;
pdev->hw_info.max_port_fcoe_conn = MAX_CONNECTIONS_FCOE;
pdev->hw_info.max_port_conns = MAX_CONNECTIONS;
pdev->hw_info.max_common_conns = MAX_CONNECTIONS;
return LM_STATUS_SUCCESS;
}
static u32_t lm_get_shmem_base_addr(lm_device_t *pdev)
{
u32_t val = 0;
u32_t min_shmem_addr = 0;
u32_t max_shmem_addr = 0;
val = REG_RD(pdev,MISC_REG_SHARED_MEM_ADDR);
if (CHIP_IS_E1(pdev))
{
min_shmem_addr = 0xa0000;
max_shmem_addr = 0xb0000;
}
else if (CHIP_IS_E1H(pdev))
{
min_shmem_addr = 0xa0000;
max_shmem_addr = 0xc0000;
}
else if (CHIP_IS_E2E3(pdev))
{
min_shmem_addr = 0x3a0000;
max_shmem_addr = 0x3c8000;
}
else
{
u32 pcicfg_chip;
mm_read_pci(pdev, 0, &pcicfg_chip);
DbgMessage(pdev, FATAL , "Unknown chip 0x%x, pcicfg[0]=0x%x, GRC[0x2000]=0x%x\n",
CHIP_NUM(pdev), pcicfg_chip, REG_RD(pdev, 0x2000));
DbgBreakMsg("Unknown chip version");
}
if (val < min_shmem_addr || val >= max_shmem_addr)
{
val = 0;
}
return val;
}
static lm_status_t lm_get_shmem_info_no_mcp_bypass(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
lm_status_t lm_status = LM_STATUS_SUCCESS;
u32_t val = 0;
DbgMessage(pdev, WARN, "MCP Down Detected\n");
#ifndef _VBD_CMD_
val = REG_RD(pdev,MISC_REG_SHARED_MEM_ADDR);
DbgMessage(pdev, FATAL, "FW ShMem addr: 0x%x\n", val);
#endif
pdev->hw_info.mcp_detected = 0;
LM_SHMEM_READ(pdev, NO_MCP_WA_CFG_SET_ADDR, &val);
if (val == NO_MCP_WA_CFG_SET_MAGIC)
{
LM_SHMEM_READ(pdev, NO_MCP_WA_FORCE_5710, &val);
LM_SHMEM_READ(pdev, NO_MCP_WA_MULTI_VNIC_MODE, &val);
mf_info->multi_vnics_mode = (u8_t)val;
if (mf_info->multi_vnics_mode)
{
LM_SHMEM_READ(pdev, NO_MCP_WA_OVLAN(ABS_FUNC_ID(pdev)), &val);
mf_info->ext_id = (u16_t)val;
mf_info->multi_vnics_mode = VALID_OVLAN(mf_info->ext_id)? 1 : 0;
mf_info->path_has_ovlan = mf_info->multi_vnics_mode;
if ((CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4) && !mf_info->multi_vnics_mode)
{
u8_t other_port = !PORT_ID(pdev);
u8_t abs_func_on_other_port = PATH_ID(pdev) + 2*other_port;
LM_SHMEM_READ(pdev, NO_MCP_WA_OVLAN(abs_func_on_other_port), &val);
mf_info->path_has_ovlan = VALID_OVLAN((u16_t)val) ? 1 : 0;
}
if (mf_info->path_has_ovlan)
{
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
mf_info->vnics_per_port = 2;
}
else
{
mf_info->vnics_per_port = 4;
}
}
mf_info->mf_mode = MULTI_FUNCTION_SD;
}
lm_status = lm_get_shmem_license_info(pdev);
if (lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
}
if (mf_info->multi_vnics_mode)
{
if (!VALID_OVLAN(mf_info->ext_id))
{
DbgMessage(pdev, FATAL, "Invalid ovlan (0x%x) configured for Func %d. Can't load the function.\n",
mf_info->ext_id, ABS_FUNC_ID(pdev));
lm_status = LM_STATUS_FAILURE;
}
}
if ((mf_info->vnics_per_port - 1 < VNIC_ID(pdev)) || ( !mf_info->multi_vnics_mode && (VNIC_ID(pdev) > 0)))
{
DbgMessage(pdev, FATAL, "Invalid vnics_per_port (%d) configured for Func %d. Can't load the function.\n",
mf_info->vnics_per_port, ABS_FUNC_ID(pdev));
lm_status = LM_STATUS_FAILURE;
}
return lm_status;
}
static lm_status_t lm_get_shmem_shared_hw_config(lm_device_t *pdev)
{
u32_t val = 0;
u8_t i = 0;
LM_SHMEM_READ(pdev, OFFSETOF(shmem_region_t, dev_info.shared_hw_config.config),&val);
pdev->hw_info.nvm_hw_config = val;
pdev->params.link.hw_led_mode = ((pdev->hw_info.nvm_hw_config & SHARED_HW_CFG_LED_MODE_MASK) >> SHARED_HW_CFG_LED_MODE_SHIFT);
DbgMessage(pdev, INFORMi, "nvm_hw_config %d\n",val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.shared_hw_config.config2),&val);
pdev->hw_info.nvm_hw_config2 = val;
DbgMessage(pdev, INFORMi, "nvm_hw_configs %d\n",val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.shared_hw_config.part_num),&val);
pdev->hw_info.board_num[0] = (u8_t) val;
pdev->hw_info.board_num[1] = (u8_t) (val >> 8);
pdev->hw_info.board_num[2] = (u8_t) (val >> 16);
pdev->hw_info.board_num[3] = (u8_t) (val >> 24);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.shared_hw_config.part_num)+4,&val);
pdev->hw_info.board_num[4] = (u8_t) val;
pdev->hw_info.board_num[5] = (u8_t) (val >> 8);
pdev->hw_info.board_num[6] = (u8_t) (val >> 16);
pdev->hw_info.board_num[7] = (u8_t) (val >> 24);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.shared_hw_config.part_num)+8,&val);
pdev->hw_info.board_num[8] = (u8_t) val;
pdev->hw_info.board_num[9] = (u8_t) (val >> 8);
pdev->hw_info.board_num[10] =(u8_t) (val >> 16);
pdev->hw_info.board_num[11] =(u8_t) (val >> 24);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.shared_hw_config.part_num)+12,&val);
pdev->hw_info.board_num[12] = (u8_t) val;
pdev->hw_info.board_num[13] = (u8_t) (val >> 8);
pdev->hw_info.board_num[14] = (u8_t) (val >> 16);
pdev->hw_info.board_num[15] = (u8_t) (val >> 24);
DbgMessage(pdev, INFORMi, "board_sn: ");
for (i = 0 ; i < 16 ; i++ )
{
DbgMessage(pdev, INFORMi, "%02x",pdev->hw_info.board_num[i]);
}
DbgMessage(pdev, INFORMi, "\n");
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.shared_feature_config.config),&val);
if GET_FLAGS(val, SHARED_FEAT_CFG_OVERRIDE_PREEMPHASIS_CFG_ENABLED)
{
SET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_OVERRIDE_PREEMPHASIS_ENABLED);
}
else
{
RESET_FLAGS(pdev->params.link.feature_config_flags,ELINK_FEATURE_CONFIG_OVERRIDE_PREEMPHASIS_ENABLED);
}
#ifdef EDIAG
SET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_DISABLE_REMOTE_FAULT_DET);
SET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_IEEE_PHY_TEST);
#endif
return LM_STATUS_SUCCESS;
}
static u32_t lm_get_shmem_mf_cfg_base(lm_device_t *pdev)
{
u32_t shmem2_size;
u32_t offset;
u32_t mf_cfg_offset_value;
offset = pdev->hw_info.shmem_base + OFFSETOF(shmem_region_t, func_mb) + E1H_FUNC_MAX * sizeof(struct drv_func_mb);
if (pdev->hw_info.shmem_base2 != 0)
{
LM_SHMEM2_READ(pdev, OFFSETOF(shmem2_region_t,size), &shmem2_size);
if (shmem2_size > OFFSETOF(shmem2_region_t,mf_cfg_addr))
{
LM_SHMEM2_READ(pdev, OFFSETOF(shmem2_region_t,mf_cfg_addr), &mf_cfg_offset_value);
if (SHMEM_MF_CFG_ADDR_NONE != mf_cfg_offset_value)
{
offset = mf_cfg_offset_value;
}
}
}
return offset;
}
static lm_status_t lm_get_shmem_port_hw_config(lm_device_t *pdev)
{
u32_t val;
const u8_t port = PORT_ID(pdev);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].mba_config),
&val);
pdev->hw_info.mba_features = (val & PORT_FEATURE_MBA_BOOT_AGENT_TYPE_MASK);
DbgMessage(pdev, INFORMi, "mba_features %d\n",pdev->hw_info.mba_features);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].mba_vlan_cfg),
&val);
pdev->hw_info.mba_vlan_cfg = val ;
DbgMessage(pdev, INFORMi, "mba_vlan_cfg 0x%x\n",pdev->hw_info.mba_vlan_cfg);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].config),
&val);
pdev->hw_info.port_feature_config = val;
DbgMessage(pdev, INFORMi, "port_feature_config 0x%x\n",pdev->hw_info.port_feature_config);
#ifndef DOS
if(val & PORT_FEAT_CFG_AUTOGREEEN_ENABLED) {
SET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_AUTOGREEEN_ENABLED);
} else {
RESET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_AUTOGREEEN_ENABLED);
}
#endif
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_hw_config[port].speed_capability_mask),&val);
pdev->params.link.speed_cap_mask[0] = val & PORT_HW_CFG_SPEED_CAPABILITY_D0_MASK;
DbgMessage(pdev, INFORMi, "speed_cap_mask1 %d\n",val);
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_hw_config[port].speed_capability_mask2),&val);
pdev->params.link.speed_cap_mask[1] = val & PORT_HW_CFG_SPEED_CAPABILITY_D0_MASK;
DbgMessage(pdev, INFORMi, "speed_cap_mask2 %d\n",val);
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_hw_config[port].lane_config),&val);
pdev->params.link.lane_config = val;
DbgMessage(pdev, INFORMi, "lane_config %d\n",val);
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].link_config),&val);
pdev->hw_info.link_config[ELINK_INT_PHY] = val;
pdev->params.link.switch_cfg = val & PORT_FEATURE_CONNECTED_SWITCH_MASK;
DbgMessage(pdev, INFORMi, "link config %d\n",val);
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].link_config2),&val);
pdev->hw_info.link_config[ELINK_EXT_PHY1] = val;
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_hw_config[port].multi_phy_config),&val);
pdev->params.link.multi_phy_config = val;
pdev->hw_info.multi_phy_config = val;
pdev->hw_info.no_10g_kr = FALSE ;
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_hw_config[port].default_cfg),&val);
pdev->hw_info.phy_force_kr_enabler = (val & PORT_HW_CFG_FORCE_KR_ENABLER_MASK) ;
if(( PORT_HW_CFG_FORCE_KR_ENABLER_NOT_FORCED != pdev->hw_info.phy_force_kr_enabler ) &&
( FALSE == ( pdev->params.link.speed_cap_mask[0] & (PORT_HW_CFG_SPEED_CAPABILITY_D0_10G | PORT_HW_CFG_SPEED_CAPABILITY_D0_20G))) )
{
pdev->hw_info.no_10g_kr = TRUE ;
}
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.port_feature_config[port].eee_power_mode),&val);
pdev->params.link.eee_mode = val & PORT_FEAT_CFG_EEE_POWER_MODE_MASK;
DbgMessage(pdev, INFORMi, "eee_power_mode 0x%x\n", pdev->params.link.eee_mode);
if ((pdev->params.link.eee_mode & PORT_FEAT_CFG_EEE_POWER_MODE_MASK) != PORT_FEAT_CFG_EEE_POWER_MODE_DISABLED)
{
SET_FLAGS(pdev->params.link.eee_mode,
ELINK_EEE_MODE_ENABLE_LPI |
ELINK_EEE_MODE_ADV_LPI);
}
return LM_STATUS_SUCCESS;
}
static void lm_set_path_has_ovlan(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t val = 0;
mf_info->path_has_ovlan = FALSE;
if (mf_info->mf_mode == MULTI_FUNCTION_SD)
{
mf_info->path_has_ovlan = TRUE;
}
else if (mf_info->mf_mode == SINGLE_FUNCTION)
{
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
u8_t other_port = !PORT_ID(pdev);
u8_t abs_func_on_other_port = PATH_ID(pdev) + 2*other_port;
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[abs_func_on_other_port].e1hov_tag),&val);
mf_info->path_has_ovlan = VALID_OVLAN((u16_t)val) ? 1 : 0;
}
}
}
lm_status_t lm_get_shmem_mf_cfg_info(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t val = 0;
u32_t val2 = 0;
u32_t mac_upper = 0;
lm_status_t status = LM_STATUS_SUCCESS;
mf_info->vnics_per_port = 1;
mf_info->multi_vnics_mode = FALSE;
mf_info->path_has_ovlan = FALSE;
mf_info->mf_mode = SINGLE_FUNCTION;
pdev->params.mf_proto_support_flags = 0;
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.shared_feature_config.config),&val);
val &= SHARED_FEAT_CFG_FORCE_SF_MODE_MASK;
switch (val)
{
case SHARED_FEAT_CFG_FORCE_SF_MODE_SWITCH_INDEPT:
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].mac_upper),&mac_upper);
if (mac_upper != FUNC_MF_CFG_UPPERMAC_DEFAULT)
{
mf_info->mf_mode = MULTI_FUNCTION_SI;
}
else
{
DbgMessage(pdev, WARNi, "Illegal configuration for switch independent mode\n");
}
DbgBreakIf(CHIP_IS_E1x(pdev));
break;
case SHARED_FEAT_CFG_FORCE_SF_MODE_MF_ALLOWED:
case SHARED_FEAT_CFG_FORCE_SF_MODE_SPIO4:
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].e1hov_tag),&val);
val &= FUNC_MF_CFG_E1HOV_TAG_MASK;
if (val != FUNC_MF_CFG_E1HOV_TAG_DEFAULT)
{
mf_info->mf_mode = MULTI_FUNCTION_SD;
mf_info->sd_mode = SD_REGULAR_MODE;
}
else
{
DbgMessage(pdev, WARNi, "Illegal configuration for switch dependent mode\n");
}
break;
case SHARED_FEAT_CFG_FORCE_SF_MODE_FORCED_SF:
return LM_STATUS_SUCCESS;
case SHARED_FEAT_CFG_FORCE_SF_MODE_AFEX_MODE:
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].mac_upper),&mac_upper);
if ((LM_SHMEM2_HAS(pdev, afex_driver_support)) &&
(mac_upper != FUNC_MF_CFG_UPPERMAC_DEFAULT) )
{
mf_info->mf_mode = MULTI_FUNCTION_AFEX;
}
else
{
DbgMessage(pdev, WARNi, "Illegal configuration for NPAR-SD mode\n");
}
DbgBreakIf(CHIP_IS_E1x(pdev));
break;
case SHARED_FEAT_CFG_FORCE_SF_MODE_BD_MODE:
mf_info->mf_mode = MULTI_FUNCTION_SD;
mf_info->sd_mode = SD_BD_MODE;
DbgMessage(pdev, WARN, "lm_get_shmem_info: SF_MODE_BD_MODE is detected.\n");
break;
case SHARED_FEAT_CFG_FORCE_SF_MODE_UFP_MODE:
mf_info->mf_mode = MULTI_FUNCTION_SD;
mf_info->sd_mode = SD_UFP_MODE;
DbgMessage(pdev, WARN, "lm_get_shmem_info: SF_MODE_UFP_MODE is detected.\n");
break;
case SHARED_FEAT_CFG_FORCE_SF_MODE_EXTENDED_MODE:
LM_SHMEM_READ(pdev, OFFSETOF(shmem_region_t, dev_info.shared_hw_config.config_3),&val);
val2 &= SHARED_HW_CFG_EXTENDED_MF_MODE_MASK;
switch (val2)
{
case SHARED_HW_CFG_EXTENDED_MF_MODE_NPAR1_DOT_5:
mf_info->mf_mode = MULTI_FUNCTION_SI;
break;
default:
DbgBreakMsg(" Unknown extended mf mode\n");
return LM_STATUS_FAILURE;
}
break;
default:
DbgBreakMsg(" Unknown mf mode\n");
return LM_STATUS_FAILURE;
}
lm_set_path_has_ovlan(pdev);
if (mf_info->mf_mode == SINGLE_FUNCTION)
{
if (VNIC_ID(pdev) >= 1)
{
return LM_STATUS_FAILURE;
}
return LM_STATUS_SUCCESS;
}
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].config),&val);
mf_info->func_mf_cfg = val;
switch(mf_info->mf_mode)
{
case MULTI_FUNCTION_SD:
{
switch (mf_info->sd_mode)
{
case SD_REGULAR_MODE:
status = lm_get_shmem_mf_cfg_info_sd(pdev);
break;
case SD_UFP_MODE:
status = lm_get_shmem_mf_cfg_info_sd_ufp(pdev);
break;
case SD_BD_MODE:
status = lm_get_shmem_mf_cfg_info_sd_bd(pdev);
break;
default:
DbgBreak();
}
if(status != LM_STATUS_SUCCESS)
return status;
}
break;
case MULTI_FUNCTION_SI:
{
lm_get_shmem_mf_cfg_info_si(pdev);
}
break;
case MULTI_FUNCTION_AFEX:
{
lm_get_shmem_mf_cfg_info_niv(pdev);
}
break;
default:
{
DbgBreakIfAll(TRUE);
return LM_STATUS_FAILURE;
}
}
lm_cmng_get_shmem_info(pdev);
return lm_check_valid_mf_cfg(pdev);
}
static void lm_fcoe_set_default_wwns(lm_device_t *pdev)
{
mm_memcpy(&(pdev->hw_info.fcoe_wwn_port_name[2]), pdev->hw_info.fcoe_mac_addr, 6);
pdev->hw_info.fcoe_wwn_port_name[0] = 0x20;
pdev->hw_info.fcoe_wwn_port_name[1] = 0;
mm_memcpy(&(pdev->hw_info.fcoe_wwn_node_name[2]), pdev->hw_info.fcoe_mac_addr, 6);
pdev->hw_info.fcoe_wwn_node_name[0] = 0x10;
pdev->hw_info.fcoe_wwn_node_name[1] = 0;
}
static lm_status_t lm_get_shmem_mf_mac_info(lm_device_t *pdev)
{
lm_hardware_mf_info_t *mf_info = &pdev->hw_info.mf_info;
u32_t mac_upper = 0;
u32_t mac_lower = 0;
if (mf_info->mf_mode == SINGLE_FUNCTION)
{
return LM_STATUS_FAILURE;
}
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].mac_upper),&mac_upper);
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_mf_config[ABS_FUNC_ID(pdev)].mac_lower),&mac_lower);
SET_FLAGS(mf_info->flags,MF_INFO_VALID_MAC);
_copy_mac_upper_lower_to_arr(mac_upper, mac_lower, pdev->hw_info.mac_addr);
switch (mf_info->mf_mode)
{
case MULTI_FUNCTION_SD:
{
if ( CHIP_IS_E1x(pdev) || IS_SD_UFP_MODE(pdev) || IS_SD_BD_MODE(pdev))
{
mm_memcpy(pdev->hw_info.iscsi_mac_addr, pdev->hw_info.mac_addr, 6);
mm_memcpy(pdev->hw_info.fcoe_mac_addr, pdev->hw_info.mac_addr, 6);
break;
}
}
case MULTI_FUNCTION_SI:
case MULTI_FUNCTION_AFEX:
lm_get_shmem_ext_mac_addresses(pdev);
break;
}
return LM_STATUS_SUCCESS;
}
static lm_status_t lm_get_shmem_sf_mac_info(lm_device_t *pdev)
{
u32_t val = 0;
u32_t val2 = 0;
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.port_hw_config[PORT_ID(pdev)].mac_upper),&val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t, dev_info.port_hw_config[PORT_ID(pdev)].mac_lower),&val2);
_copy_mac_upper_lower_to_arr(val, val2, pdev->hw_info.mac_addr);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].iscsi_mac_upper),&val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].iscsi_mac_lower),&val2);
_copy_mac_upper_lower_to_arr(val, val2, pdev->hw_info.iscsi_mac_addr);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_fip_mac_upper),&val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_fip_mac_lower),&val2);
_copy_mac_upper_lower_to_arr(val, val2, pdev->hw_info.fcoe_mac_addr);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_wwn_port_name_upper),&val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_wwn_port_name_lower),&val2);
_copy_mac_upper_lower_to_arr(val, val2, &(pdev->hw_info.fcoe_wwn_port_name[2]));
pdev->hw_info.fcoe_wwn_port_name[0] = (u8_t) (val >> 24);
pdev->hw_info.fcoe_wwn_port_name[1] = (u8_t) (val >> 16);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_wwn_node_name_upper),&val);
LM_SHMEM_READ(pdev,
OFFSETOF(shmem_region_t,dev_info.port_hw_config[PORT_ID(pdev)].fcoe_wwn_node_name_lower),&val2);
_copy_mac_upper_lower_to_arr(val, val2, &(pdev->hw_info.fcoe_wwn_node_name[2]));
pdev->hw_info.fcoe_wwn_node_name[0] = (u8_t) (val >> 24);
pdev->hw_info.fcoe_wwn_node_name[1] = (u8_t) (val >> 16);
DbgMessage(pdev, INFORMi, "main mac addr: %02x %02x %02x %02x %02x %02x\n",
pdev->hw_info.mac_addr[0],
pdev->hw_info.mac_addr[1],
pdev->hw_info.mac_addr[2],
pdev->hw_info.mac_addr[3],
pdev->hw_info.mac_addr[4],
pdev->hw_info.mac_addr[5]);
DbgMessage(pdev, INFORMi, "iSCSI mac addr: %02x %02x %02x %02x %02x %02x\n",
pdev->hw_info.iscsi_mac_addr[0],
pdev->hw_info.iscsi_mac_addr[1],
pdev->hw_info.iscsi_mac_addr[2],
pdev->hw_info.iscsi_mac_addr[3],
pdev->hw_info.iscsi_mac_addr[4],
pdev->hw_info.iscsi_mac_addr[5]);
return LM_STATUS_SUCCESS;
}
static void lm_get_shmem_sf_sriov_info(lm_device_t *pdev)
{
const lm_chip_port_mode_t port_mode = CHIP_PORT_MODE(pdev);
u32_t offset = 0;
u32_t val = 0;
u8_t port_max = (port_mode == LM_CHIP_PORT_MODE_2)? 1 : PORT_MAX;
const u8_t port = PORT_ID(pdev);
u8_t port_idx = 0;
u8_t sriov_enabled = 0xff;
u8_t sriov_disabled = 0xff;
ASSERT_STATIC((FIELD_SIZE(struct shm_dev_info, port_hw_config)/FIELD_SIZE(struct shm_dev_info, port_hw_config[0])) >= max(PORT_MAX,1));
if (CHIP_IS_E1x(pdev))
{
pdev->hw_info.sriov_info.shmem_num_vfs_in_pf = 0;
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = FALSE;
return;
}
for (port_idx = 0; port_idx < port_max; port_idx++)
{
offset = OFFSETOF(shmem_region_t,dev_info.port_hw_config[port_idx].pf_allocation);
LM_SHMEM_READ(pdev, offset, &val);
val = (val & PORT_HW_CFG_NUMBER_OF_VFS_MASK) >> PORT_HW_CFG_NUMBER_OF_VFS_SHIFT;
if (0 == val)
{
sriov_disabled = 1;
}
else
{
sriov_enabled = 1;
}
if (port_idx == port)
{
pdev->hw_info.sriov_info.shmem_num_vfs_in_pf = val;
}
}
if (sriov_disabled == sriov_enabled)
{
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = TRUE;
}
else
{
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = FALSE;
}
}
static void lm_get_shmem_mf_sriov_info(lm_device_t *pdev)
{
u32_t offset = 0;
u32_t val = 0;
u8_t func = 0;
const u8_t abs_func = ABS_FUNC_ID(pdev);
u8_t abs_func_idx = 0;
u8_t sriov_enabled = 0xff;
u8_t sriov_disabled = 0xff;
ASSERT_STATIC((FIELD_SIZE(struct mf_cfg, func_mf_config) / FIELD_SIZE(struct mf_cfg, func_mf_config[0])) == E2_FUNC_MAX*2);
if (CHIP_IS_E1x(pdev))
{
pdev->hw_info.sriov_info.shmem_num_vfs_in_pf = 0;
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = FALSE;
return;
}
for (func = 0; func < E2_FUNC_MAX; func++)
{
abs_func_idx = PATH_ID(pdev) + func*2;
offset = OFFSETOF(mf_cfg_t, func_mf_config[abs_func_idx].pf_allocation);
LM_MFCFG_READ(pdev, offset,&val);
val = (val & FUNC_MF_CFG_NUMBER_OF_VFS_MASK) >> FUNC_MF_CFG_NUMBER_OF_VFS_SHIFT;
if (0 == val)
{
sriov_disabled = 1;
}
else
{
sriov_enabled = 1;
}
if (abs_func_idx == abs_func)
{
pdev->hw_info.sriov_info.shmem_num_vfs_in_pf = val;
}
}
if (sriov_disabled == sriov_enabled)
{
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = TRUE;
}
else
{
pdev->hw_info.sriov_info.b_pf_asymetric_configuration = FALSE;
}
}
static lm_status_t lm_get_shmem_mac_info(lm_device_t *pdev)
{
lm_status_t lm_status = LM_STATUS_SUCCESS;
if (pdev->hw_info.mf_info.mf_mode == SINGLE_FUNCTION)
{
lm_status = lm_get_shmem_sf_mac_info(pdev);
}
else
{
lm_status = lm_get_shmem_mf_mac_info(pdev);
}
return lm_status;
}
static void lm_get_shmem_sriov_info(lm_device_t *pdev)
{
const u32_t bc_rev = LM_GET_BC_REV_MAJOR(pdev);
if (CHIP_IS_E1x(pdev) || (bc_rev < BC_REV_IE_SRIOV_SUPPORTED))
{
return;
}
if (pdev->hw_info.mf_info.mf_mode == SINGLE_FUNCTION)
{
lm_get_shmem_sf_sriov_info(pdev);
}
else
{
lm_get_shmem_mf_sriov_info(pdev);
}
}
static void lm_get_shmem_fw_flow_control(lm_device_t *pdev)
{
u32_t func_ext_cfg = 0;
ASSERT_STATIC(MAX_MF_MODE == 4);
if (((pdev->hw_info.mf_info.mf_mode == MULTI_FUNCTION_SI) ||
(pdev->hw_info.mf_info.mf_mode == MULTI_FUNCTION_AFEX)) &&
(!CHIP_IS_E1x(pdev)))
{
LM_MFCFG_READ(pdev, OFFSETOF(mf_cfg_t, func_ext_config[ABS_FUNC_ID(pdev)].func_cfg), &func_ext_cfg);
if (GET_FLAGS(func_ext_cfg, MACP_FUNC_CFG_PAUSE_ON_HOST_RING))
{
pdev->params.l2_fw_flow_ctrl = 1;
}
else
{
pdev->params.l2_fw_flow_ctrl = 0;
}
}
}
lm_status_t lm_get_shmem_info(lm_device_t *pdev)
{
lm_status_t lm_status = LM_STATUS_SUCCESS;
u32_t val = 0;
lm_shmem_set_default(pdev);
val = lm_get_shmem_base_addr(pdev);
if (!val)
{
DbgMessage(pdev, WARNi, "NO MCP\n");
return lm_get_shmem_info_no_mcp_bypass(pdev);
}
pdev->hw_info.mcp_detected = 1;
pdev->hw_info.shmem_base = val;
pdev->hw_info.shmem_base2 = REG_RD(pdev, PATH_ID(pdev) ? MISC_REG_GENERIC_CR_1 : MISC_REG_GENERIC_CR_0);
pdev->hw_info.mf_cfg_base = lm_get_shmem_mf_cfg_base(pdev);
DbgMessage(pdev, WARNi, "MCP Up Detected. shmem_base=0x%x shmem_base2=0x%x mf_cfg_offset=0x%x\n",
pdev->hw_info.shmem_base, pdev->hw_info.shmem_base2, pdev->hw_info.mf_cfg_base);
lm_status = lm_verify_validity_map( pdev );
if(LM_STATUS_SUCCESS != lm_status )
{
DbgMessage(pdev, FATAL, "lm_get_shmem_info: Shmem signature not present.\n");
pdev->hw_info.mcp_detected = 0;
return LM_STATUS_SUCCESS;
}
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t,dev_info.bc_rev),&val);
pdev->hw_info.bc_rev = val;
DbgMessage(pdev, INFORMi, "bc_rev %d\n",val);
lm_status = lm_get_shmem_shared_hw_config(pdev);
if (lm_status != LM_STATUS_SUCCESS)
{
DbgMessage(pdev, WARNi, "lm_get_shmem_shared_hw_config returned lm_status=%d\n", lm_status);
return lm_status;
}
lm_status = lm_get_shmem_port_hw_config(pdev);
if (lm_status != LM_STATUS_SUCCESS)
{
DbgMessage(pdev, WARNi, "lm_get_shmem_port_hw_config returned lm_status=%d\n", lm_status);
return lm_status;
}
#ifdef _LICENSE_H
lm_status = lm_get_shmem_license_info(pdev);
if (lm_status != LM_STATUS_SUCCESS)
{
DbgMessage(pdev, WARNi, "lm_get_shmem_license_info returned lm_status=%d\n", lm_status);
return lm_status;
}
#endif
if (IS_MF_MODE_CAPABLE(pdev) && (pdev->hw_info.mf_cfg_base != SHMEM_MF_CFG_ADDR_NONE))
{
lm_status = lm_get_shmem_mf_cfg_info(pdev);
if (lm_status != LM_STATUS_SUCCESS)
{
DbgMessage(pdev, WARNi, "lm_get_shmem_mf_cfg_info returned lm_status=%d\n", lm_status);
return lm_status;
}
}
else if (FUNC_ID(pdev) != PORT_ID(pdev))
{
DbgMessage(pdev, WARNi, "Illegal to load func %d of port %d on non MF mode capable device\n");
return LM_STATUS_FAILURE;
}
lm_get_shmem_sriov_info(pdev);
lm_status = lm_get_shmem_mac_info(pdev);
lm_get_shmem_fw_flow_control(pdev);
return lm_status;
}
void init_link_params(lm_device_t *pdev)
{
u32_t val = 0;
u32_t feat_val = 0;
const u8_t port = PORT_ID(pdev);
pdev->params.link.port = port;
pdev->params.link.lfa_base = 0;
pdev->params.link.shmem_base = NO_MCP_WA_CLC_SHMEM;
pdev->params.link.shmem2_base= NO_MCP_WA_CLC_SHMEM;
if (pdev->hw_info.mcp_detected)
{
pdev->params.link.shmem_base = pdev->hw_info.shmem_base;
pdev->params.link.shmem2_base= pdev->hw_info.shmem_base2;
if (LM_SHMEM2_HAS(pdev,lfa_host_addr[port]))
{
LM_SHMEM2_READ(pdev, OFFSETOF(shmem2_region_t, lfa_host_addr[port]), &pdev->params.link.lfa_base);
}
}
pdev->params.link.chip_id = pdev->hw_info.chip_id;
pdev->params.link.cb = pdev;
if(pdev->hw_info.mf_info.mf_mode == MULTI_FUNCTION_AFEX)
{
SET_FLAGS( pdev->params.link.feature_config_flags, ELINK_FEATURE_CONFIG_BC_SUPPORTS_AFEX );
}
if (CHIP_REV_IS_SLOW(pdev))
{
val = CHIP_BONDING(pdev);
DbgMessage(pdev, WARN, "init_link_params: chip bond id is 0x%x\n",val);
if (pdev->hw_info.chip_port_mode == LM_CHIP_PORT_MODE_4)
{
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_BMAC;
}
else if (val & 0x4)
{
if (CHIP_IS_E3(pdev))
{
pdev->params.link.req_line_speed[0] = ELINK_SPEED_1000;
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_XMAC;
}
else
{
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_BMAC;
}
}
else if (val & 0x8)
{
if (CHIP_IS_E3(pdev))
{
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_UMAC;
}
else
{
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_EMAC;
}
}
if (val & 2)
{
feat_val |= ELINK_FEATURE_CONFIG_EMUL_DISABLE_EMAC;
}
SET_FLAGS(pdev->params.link.feature_config_flags, feat_val);
}
}
static void lm_init_cam_params(lm_device_t *pdev)
{
#define LM_CAM_SIZE_EMUL (5)
#define LM_MC_TABLE_SIZE_EMUL (1)
#define LM_CAM_SIZE_EMUL_E2 (40)
u16_t mc_credit;
u16_t uc_credit;
u8_t b_is_asic = CHIP_REV_IS_ASIC(pdev);
u8_t num_ports = 2;
u8_t num_funcs;
mm_mem_zero(pdev->params.uc_table_size, sizeof(pdev->params.uc_table_size));
mm_mem_zero(pdev->params.mc_table_size, sizeof(pdev->params.mc_table_size));
if (CHIP_IS_E1(pdev))
{
pdev->params.cam_size = b_is_asic? MAX_MAC_CREDIT_E1 / num_ports : LM_CAM_SIZE_EMUL;
mc_credit = b_is_asic? LM_MC_NDIS_TABLE_SIZE : LM_MC_TABLE_SIZE_EMUL;
uc_credit = pdev->params.cam_size - mc_credit;
pdev->params.uc_table_size[LM_CLI_IDX_ISCSI] = 1;
pdev->params.uc_table_size[LM_CLI_IDX_NDIS] = uc_credit - 1;
pdev->params.mc_table_size[LM_CLI_IDX_NDIS] = mc_credit;
DbgMessage(pdev, INFORMi, "uc_table_size[ndis]=%d, uc_table_size[ndis]=%d, mc_table_size[ndis]=%d\n",
pdev->params.uc_table_size[LM_CLI_IDX_NDIS], pdev->params.uc_table_size[LM_CLI_IDX_ISCSI],
pdev->params.mc_table_size[LM_CLI_IDX_NDIS]);
}
else if (CHIP_IS_E1H(pdev))
{
pdev->params.cam_size = b_is_asic? MAX_MAC_CREDIT_E1H / num_ports: LM_CAM_SIZE_EMUL;
pdev->params.cam_size = pdev->params.cam_size / pdev->params.vnics_per_port;
uc_credit = pdev->params.cam_size;
pdev->params.uc_table_size[LM_CLI_IDX_ISCSI] = 1;
pdev->params.uc_table_size[LM_CLI_IDX_NDIS] = uc_credit - 1;
pdev->params.mc_table_size[LM_CLI_IDX_NDIS] = LM_MC_NDIS_TABLE_SIZE;
DbgMessage(pdev, INFORMi, "uc_table_size[ndis]=%d, uc_table_size[ndis]=%d, mc_table_size[ndis]=%d\n",
pdev->params.uc_table_size[LM_CLI_IDX_NDIS], pdev->params.uc_table_size[LM_CLI_IDX_ISCSI],
pdev->params.mc_table_size[LM_CLI_IDX_NDIS]);
}
else if (CHIP_IS_E2E3(pdev))
{
num_ports = (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)? 2 : 1;
num_funcs = VNICS_PER_PATH(pdev);
if (num_funcs > 1)
{
pdev->params.cam_size = b_is_asic? ((MAX_MAC_CREDIT_E2 - GET_NUM_VFS_PER_PATH(pdev))/ num_funcs + GET_NUM_VFS_PER_PF(pdev)): LM_CAM_SIZE_EMUL_E2;
}
else
{
pdev->params.cam_size = b_is_asic? MAX_MAC_CREDIT_E2 : LM_CAM_SIZE_EMUL_E2;
}
uc_credit = pdev->params.cam_size;
pdev->params.uc_table_size[LM_CLI_IDX_ISCSI] = 1;
pdev->params.uc_table_size[LM_CLI_IDX_FCOE] = 1;
pdev->params.uc_table_size[LM_CLI_IDX_NDIS] = uc_credit - 2;
pdev->params.mc_table_size[LM_CLI_IDX_NDIS] = LM_MC_NDIS_TABLE_SIZE;
pdev->params.mc_table_size[LM_CLI_IDX_FCOE] = LM_MC_FCOE_TABLE_SIZE;
DbgMessage(pdev, INFORMi, "uc_table_size[ndis]=%d, uc_table_size[ndis]=%d, uc_table_size[fcoe]=%d, mc_table_size[ndis]=%d, mc_table_size[fcoe]=%d\n",
pdev->params.uc_table_size[LM_CLI_IDX_NDIS], pdev->params.uc_table_size[LM_CLI_IDX_ISCSI],
pdev->params.uc_table_size[LM_CLI_IDX_FCOE],
pdev->params.mc_table_size[LM_CLI_IDX_NDIS], pdev->params.mc_table_size[LM_CLI_IDX_FCOE]);
}
else
{
DbgBreakIfAll("New Chip?? initialize cam params!\n");
}
if (IS_PFDEV(pdev))
{
pdev->params.base_offset_in_cam_table = ((num_ports == 2)? FUNC_ID(pdev) : VNIC_ID(pdev)) * LM_CAM_SIZE(pdev);
}
else if (IS_CHANNEL_VFDEV(pdev))
{
pdev->params.base_offset_in_cam_table = 0;
pdev->params.mc_table_size[LM_CLI_IDX_NDIS] = 0;
}
}
static lm_status_t lm_init_params(lm_device_t *pdev, u8_t validate)
{
typedef struct _param_entry_t
{
u32_t offset;
u32_t asic_default;
u32_t fpga_default;
u32_t emulation_default;
u32_t min;
u32_t max;
} param_entry_t;
#define _OFFSET(_name) (OFFSETOF(lm_device_t, params._name))
#define PARAM_VAL(_pdev, _entry) \
(*((u32_t *) ((u8_t *) (_pdev) + (_entry)->offset)))
#define SET_PARAM_VAL(_pdev, _entry, _val) \
*((u32_t *) ((u8_t *) (_pdev) + (_entry)->offset)) = (_val)
static param_entry_t param_list[] =
{
{ _OFFSET(mtu[LM_CLI_IDX_NDIS]), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(mtu[LM_CLI_IDX_ISCSI]), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(mtu[LM_CLI_IDX_FCOE]), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(mtu[LM_CLI_IDX_OOO]), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(mtu[LM_CLI_IDX_FWD]), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(mtu_max), 9216, 9216, 9216, 1500, 9216 },
{ _OFFSET(rcv_buffer_offset), 0, 0, 0, 0, 9000 },
{ _OFFSET(l2_rx_desc_cnt[LM_CLI_IDX_NDIS]), 200, 200, 200, 0, 32767 },
{ _OFFSET(l2_rx_desc_cnt[LM_CLI_IDX_FCOE]), 200, 200, 200, 0, 32767 },
{ _OFFSET(l2_rx_desc_cnt[LM_CLI_IDX_OOO]), 500, 500, 500, 0, 32767 },
{ _OFFSET(l2_tx_bd_page_cnt[LM_CLI_IDX_NDIS]), 2, 2, 2, 1, 127 },
{ _OFFSET(l2_tx_bd_page_cnt[LM_CLI_IDX_FCOE]), 2, 2, 2, 1, 127 },
{ _OFFSET(l2_tx_coal_buf_cnt[LM_CLI_IDX_NDIS]), 0, 0, 0, 0, 20 },
{ _OFFSET(l2_tx_coal_buf_cnt[LM_CLI_IDX_FCOE]), 0, 0, 0, 0, 20 },
{ _OFFSET(l2_tx_bd_page_cnt[LM_CLI_IDX_FWD]) , 2, 2, 2, 1, 127 },
{ _OFFSET(test_mode), 0, 0, 0, 0, 0 },
{ _OFFSET(ofld_cap), 0, 0, 0, 0, 0 },
{ _OFFSET(wol_cap), 0, 0, 0, 0, 0 },
{ _OFFSET(i2c_interval_sec), 0, 0, 0, 0, 1000 },
{ _OFFSET(flow_ctrl_cap), 0, 0, 0, 0, 0x80000000 },
{ _OFFSET(eee_policy), LM_EEE_CONTROL_NVRAM, LM_EEE_CONTROL_NVRAM, LM_EEE_CONTROL_NVRAM, LM_EEE_CONTROL_HIGH, LM_EEE_CONTROL_NVRAM },
{ _OFFSET(req_medium), 0xff00, 0x00ff, 0x00ff, 0, 0xfffff },
{ _OFFSET(interrupt_mode), LM_INT_MODE_INTA, LM_INT_MODE_INTA, LM_INT_MODE_INTA, LM_INT_MODE_INTA, LM_INT_MODE_MIMD},
{ _OFFSET(igu_access_mode), INTR_BLK_ACCESS_IGUMEM, INTR_BLK_ACCESS_IGUMEM, INTR_BLK_ACCESS_IGUMEM, INTR_BLK_ACCESS_GRC, INTR_BLK_ACCESS_IGUMEM},
{ _OFFSET(sw_config), 4, 4, 4, 0, 4},
{ _OFFSET(selective_autoneg), 0, 0, 0, 0, 0 },
{ _OFFSET(autogreeen), LM_AUTOGREEEN_NVRAM, LM_AUTOGREEEN_NVRAM, LM_AUTOGREEEN_NVRAM, LM_AUTOGREEEN_DISABLED, LM_AUTOGREEEN_NVRAM },
{ _OFFSET(wire_speed), 1, 0, 0, 0, 0 },
{ _OFFSET(phy_int_mode), 2, 2, 2, 0, 0 },
{ _OFFSET(link_chng_mode), 2, 2, 2, 0, 0 },
{ _OFFSET(max_func_connections), 1024, 1024, 1024, 0, 500000},
#ifdef VF_INVOLVED
{ _OFFSET(max_func_toe_cons), 310, 310, 310, 0, 500000},
#else
{ _OFFSET(max_func_toe_cons), 750, 750, 750, 0, 500000},
#endif
{ _OFFSET(max_func_rdma_cons), 10, 10, 10, 0, 500000},
{ _OFFSET(max_func_iscsi_cons), 128, 128, 128, 0, 500000},
{ _OFFSET(max_func_fcoe_cons), 64, 64, 20, 0, 500000},
{ _OFFSET(context_line_size), LM_CONTEXT_SIZE, LM_CONTEXT_SIZE, LM_CONTEXT_SIZE, 0, LM_CONTEXT_SIZE },
{ _OFFSET(context_waste_size), 0, 0, 0, 0, 1024 },
{ _OFFSET(num_context_in_page), 4, 4, 4, 0, 128},
{ _OFFSET(client_page_size), 0x1000, 0x1000, 0x1000,0x1000, 0x20000 },
{ _OFFSET(elt_page_size), 0x1000, 0x1000, 0x1000,0x1000, 0x20000 },
{ _OFFSET(ilt_client_page_size), 0x1000, 0x1000, 0x1000,0x1000, 0x20000 },
{ _OFFSET(cfc_last_lcid), 0xff, 0xff, 0xff, 0x1, 0xff },
{ _OFFSET(override_rss_chain_cnt), 0, 0, 0, 0, 16 },
{ _OFFSET(network_type), LM_NETOWRK_TYPE_WAN, LM_NETOWRK_TYPE_WAN, LM_NETOWRK_TYPE_WAN,LM_NETOWRK_TYPE_LAN, LM_NETOWRK_TYPE_WAN },
{ _OFFSET(max_cwnd_wan), 12500000, 12500000, 12500000,12500000, 12500000 },
{ _OFFSET(max_cwnd_lan), 1250000 , 1250000, 1250000, 1250000, 1250000 },
{ _OFFSET(cid_allocation_mode), LM_CID_ALLOC_DELAY , LM_CID_ALLOC_DELAY, LM_CID_ALLOC_DELAY,LM_CID_ALLOC_DELAY, LM_CID_ALLOC_NUM_MODES},
{ _OFFSET(int_coalesing_mode), LM_INT_COAL_PERIODIC_SYNC, LM_INT_COAL_NONE, LM_INT_COAL_NONE, 1, LM_INT_COAL_NUM_MODES },
{ _OFFSET(int_per_sec_rx[0]), 5000, 5000, 5000, 1, 200000 },
{ _OFFSET(int_per_sec_rx[1]), 5000, 5000, 5000, 1, 200000 },
{ _OFFSET(int_per_sec_rx[2]), 5000, 5000, 5000, 1, 200000 },
{ _OFFSET(int_per_sec_rx[3]), 5000, 5000, 5000, 1, 200000 },
{ _OFFSET(int_per_sec_tx[0]), 7500, 7500, 7500, 1, 200000 },
{ _OFFSET(int_per_sec_tx[1]), 3800, 3800, 3800, 1, 200000 },
{ _OFFSET(int_per_sec_tx[2]), 3800, 3800, 3800, 1, 200000 },
{ _OFFSET(int_per_sec_tx[3]), 3800, 3800, 3800, 1, 200000 },
{ _OFFSET(vf_int_per_sec_rx[LM_VF_INT_LOW_IDX]), 5000, 5000, 5000, 1, 200000 },
{ _OFFSET(vf_int_per_sec_rx[LM_VF_INT_MEDIUM_IDX]), 10000, 5000, 5000, 1, 200000 },
{ _OFFSET(vf_int_per_sec_rx[LM_VF_INT_HIGH_IDX]), 20000, 5000, 5000, 1, 200000 },
{ _OFFSET(vf_int_per_sec_tx[LM_VF_INT_LOW_IDX]), 3800, 3800, 3800, 1, 200000 },
{ _OFFSET(vf_int_per_sec_tx[LM_VF_INT_MEDIUM_IDX]), 8000, 3800, 3800, 1, 200000 },
{ _OFFSET(vf_int_per_sec_tx[LM_VF_INT_HIGH_IDX]), 16000, 3800, 3800, 1, 200000 },
{ _OFFSET(enable_dynamic_hc[0]), 1, 1, 1, 0, 1 },
{ _OFFSET(enable_dynamic_hc[1]), 1, 1, 1, 0, 1 },
{ _OFFSET(enable_dynamic_hc[2]), 1, 1, 1, 0, 1 },
{ _OFFSET(enable_dynamic_hc[3]), 0, 0, 0, 0, 1 },
{ _OFFSET(hc_timeout0[SM_RX_ID][0]), 12, 12, 12, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_RX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_RX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_RX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_RX_ID][1]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_RX_ID][1]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_RX_ID][1]), 120, 120, 120, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_RX_ID][1]), 240, 240, 240, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_RX_ID][2]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_RX_ID][2]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_RX_ID][2]), 120, 120, 120, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_RX_ID][2]), 240, 240, 240, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_RX_ID][3]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_RX_ID][3]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_RX_ID][3]), 120, 120, 120, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_RX_ID][3]), 240, 240, 240, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_TX_ID][0]), 12, 12, 12, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_TX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_TX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_TX_ID][0]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_TX_ID][1]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_TX_ID][1]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_TX_ID][1]), 120, 120, 120, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_TX_ID][1]), 240, 240, 240, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_TX_ID][2]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_TX_ID][2]), 12, 12, 12, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_TX_ID][2]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_TX_ID][2]), 64, 64, 64, 1, 0xff },
{ _OFFSET(hc_timeout0[SM_TX_ID][3]), 6, 6, 6, 1, 0xff },
{ _OFFSET(hc_timeout1[SM_TX_ID][3]), 48, 48, 48, 1, 0xff },
{ _OFFSET(hc_timeout2[SM_TX_ID][3]), 120, 120, 120, 1, 0xff },
{ _OFFSET(hc_timeout3[SM_TX_ID][3]), 240, 240, 240, 1, 0xff },
{ _OFFSET(hc_threshold0[SM_RX_ID]), 0x2000, 0x2000, 0x2000,1, 0xffffffff },
{ _OFFSET(hc_threshold1[SM_RX_ID]), 0x10000, 0x10000, 0x10000,1, 0xffffffff },
{ _OFFSET(hc_threshold2[SM_RX_ID]), 0x50000, 0x50000, 0x50000,1, 0xffffffff },
{ _OFFSET(hc_threshold0[SM_TX_ID]), 0x2000, 0x2000, 0x2000,1, 0xffffffff },
{ _OFFSET(hc_threshold1[SM_TX_ID]), 0x10000, 0x10000, 0x10000,1, 0xffffffff },
{ _OFFSET(hc_threshold2[SM_TX_ID]), 0x20000, 0x20000, 0x20000,1, 0xffffffff },
{ _OFFSET(l2_dynamic_hc_min_bytes_per_packet), 0, 0, 0, 0, 0xffff },
{ _OFFSET(l4_hc_ustorm_thresh), 12, 12, 12, 12, 0xffffffff },
{ _OFFSET(l4_scq_page_cnt), 2, 2, 2, 2, 127 },
{ _OFFSET(l4_rcq_page_cnt), 3, 3, 3, 3, 127 },
{ _OFFSET(l4_grq_page_cnt), 2, 2, 2, 2, 127 },
{ _OFFSET(l4_tx_chain_page_cnt), 2, 2, 2, 2, 127 },
{ _OFFSET(l4_rx_chain_page_cnt), 2, 2, 2, 2, 127 },
{ _OFFSET(l4_gen_buf_size), LM_PAGE_SIZE,LM_PAGE_SIZE,LM_PAGE_SIZE,LM_PAGE_SIZE,16*LM_PAGE_SIZE },
{ _OFFSET(l4_history_cqe_cnt), 20, 20, 20, 1, 20 },
{ _OFFSET(l4_ignore_grq_push_enabled), 0, 0, 0, 0, 1 },
{ _OFFSET(l4cli_flags), 0, 0, 0, 0, 1 },
{ _OFFSET(l4cli_ticks_per_second), 1000, 1000, 1000, 500, 10000 },
{ _OFFSET(l4cli_ack_frequency), 2, 2, 2, 1, 255 },
{ _OFFSET(l4cli_delayed_ack_ticks), 200, 200, 200, 1, 255 },
{ _OFFSET(l4cli_max_retx), 6, 6, 6, 1, 255 },
{ _OFFSET(l4cli_doubt_reachability_retx),3, 3, 3, 1, 255 },
{ _OFFSET(l4cli_sws_prevention_ticks), 1000, 1000, 1000, 200, 0xffffffff },
{ _OFFSET(l4cli_dup_ack_threshold), 3, 3, 3, 1, 255 },
{ _OFFSET(l4cli_push_ticks), 100, 100, 100, 1, 0xffffffff },
{ _OFFSET(l4cli_nce_stale_ticks), 0xffffff,0xffffff,0xffffff, 1, 0xffffffff },
{ _OFFSET(l4cli_starting_ip_id), 0, 0, 0, 0, 0xffff },
{ _OFFSET(keep_vlan_tag), 1 , 1, 1, 0, 1 },
{ _OFFSET(cmng_enable), 0, 0, 0, 0, 1},
{ _OFFSET(cmng_rate_shaping_enable),1, 1, 1, 0, 1},
{ _OFFSET(cmng_fairness_enable), 1, 1, 1, 0, 1},
{ _OFFSET(cmng_safc_rate_thresh), 3, 3, 3, 0, 10},
{ _OFFSET(cmng_activate_safc), 0, 0, 0, 0, 1},
{ _OFFSET(cmng_fair_port0_rate), 10, 10, 10, 1, 10},
{ _OFFSET(cmng_eth_weight), 8, 8, 8, 0, 10},
{ _OFFSET(cmng_toe_weight), 8, 8, 8, 0, 10},
{ _OFFSET(cmng_rdma_weight), 8, 8, 8, 0, 10},
{ _OFFSET(cmng_iscsi_weight), 8, 8, 8, 0, 10},
{ _OFFSET(cmng_eth_rate), 10, 10, 10, 0, 10},
{ _OFFSET(cmng_toe_rate), 10, 10, 10, 0, 10},
{ _OFFSET(cmng_rdma_rate), 2, 2, 2, 0, 10},
{ _OFFSET(cmng_iscsi_rate), 4, 2, 2, 0, 10},
{ _OFFSET(cmng_toe_con_number), 20, 20, 20, 0, 1024},
{ _OFFSET(cmng_rdma_con_number), 2, 2, 2, 0, 1024},
{ _OFFSET(cmng_iscsi_con_number), 40, 40, 40, 0, 1024},
{ _OFFSET(l5sc_max_pending_tasks), 64, 64, 64, 64, 2048},
{ _OFFSET(max_fcoe_task), 64, 64, 64, 0, 4096},
#if 0
{ _OFFSET(disable_patent_using), 1, 1, 1, 0, 1},
#else
{ _OFFSET(disable_patent_using), 0, 0, 0, 0, 1},
#endif
{ _OFFSET(l4_grq_filling_threshold_divider), 64, 64, 64, 2, 2048},
{ _OFFSET(l4_free_cid_delay_time), 2000, 10000, 10000, 0, 10000},
{ _OFFSET(preemphasis_enable), 0, 0, 0, 0, 1},
{ _OFFSET(preemphasis_rx_0), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_rx_1), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_rx_2), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_rx_3), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_tx_0), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_tx_1), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_tx_2), 0, 0, 0, 0, 0xffff},
{ _OFFSET(preemphasis_tx_3), 0, 0, 0, 0, 0xffff},
{ _OFFSET(disable_pcie_nfr), 0, 0, 0, 0, 1},
{ _OFFSET(debug_cap_flags), 0xffffffff, 0xffffffff, 0xffffffff, 0, 0xffffffff},
{ _OFFSET(try_not_align_page_multiplied_memory), 1, 1, 1, 0, 1},
{ _OFFSET(l4_limit_isles), 0, 0, 0, 0, 1},
{ _OFFSET(l4_max_rcv_wnd_size), 0x100000,0x100000,0x100000, 0, 0x1000000},
{ _OFFSET(ndsb_type), 1, 1, 1, 0, 2},
{ _OFFSET(l4_dominance_threshold), 10, 10, 10, 0, 0xFF},
{ _OFFSET(l4_max_dominance_value), 20, 20, 20, 0, 0xFF},
{ _OFFSET(l4_data_integrity), 0x0, 0x0, 0x0, 0x0, 0x3},
{ _OFFSET(l4_start_port), 5001, 5001, 5001, 0, 0xFFFFFFFF},
{ _OFFSET(l4_num_of_ports), 50, 50, 50, 0, 0xFFFF},
{ _OFFSET(l4_skip_start_bytes), 4, 4, 4, 0, 0xFFFFFFFF},
{ _OFFSET(phy_priority_mode), PHY_PRIORITY_MODE_HW_DEF, PHY_PRIORITY_MODE_HW_DEF, PHY_PRIORITY_MODE_HW_DEF, PHY_PRIORITY_MODE_HW_DEF, PHY_PRIORITY_MODE_HW_PIN},
{ _OFFSET(grc_timeout_max_ignore), 0, 0, 0, 0, 0xFFFFFFFF},
{ _OFFSET(enable_error_recovery), 0, 0, 0, 0, 1},
{ _OFFSET(validate_sq_complete), 0, 0, 0, 0, 1},
{ _OFFSET(npar_vm_switching_enable),0, 0, 0, 0, 1},
{ _OFFSET(flow_control_reporting_mode),LM_FLOW_CONTROL_REPORTING_MODE_DISABLED,LM_FLOW_CONTROL_REPORTING_MODE_DISABLED,LM_FLOW_CONTROL_REPORTING_MODE_DISABLED,LM_FLOW_CONTROL_REPORTING_MODE_DISABLED,LM_FLOW_CONTROL_REPORTING_MODE_ENABLED},
{ _OFFSET(tpa_desc_cnt_per_chain), 0, 0, 0, 0, 0x10000},
{ _OFFSET(sriov_inc_mac), 0, 0, 0, 0, 64},
{ _OFFSET(e3_cos_modes), 0, 0, 0, 0, 1},
{ _OFFSET(e3_network_cos_mode), 0, 0, 0, 0, 1},
{ _OFFSET(fw_valid_mask), 0xFFFFFFFF,0xFFFFFFFF,0xFFFFFFFF,0, 0xFFFFFFFF},
{ _OFFSET(record_sp), 0x0, 0x0, 0x0, 0, 0xf},
{ 0, 0, 0, 0, 0, 0}
};
param_entry_t *param = NULL;
size_t csize = 0;
u32_t flow_control = 0;
u8_t i = 0;
u8_t port_base_aux_qzone = 0;
u8_t base_fw_qzone_id = 0;
DbgMessage(pdev, INFORMi , "### lm_init_param\n");
if (!validate)
{
param = param_list;
while(param->offset)
{
if(CHIP_REV_IS_FPGA(pdev))
{
SET_PARAM_VAL(pdev, param, param->fpga_default);
}
else if(CHIP_REV_IS_EMUL(pdev))
{
SET_PARAM_VAL(pdev, param, param->emulation_default);
}
else
{
SET_PARAM_VAL(pdev, param, param->asic_default);
}
param++;
}
pdev->params.rss_caps = (LM_RSS_CAP_IPV4 | LM_RSS_CAP_IPV6);
pdev->params.rss_chain_cnt = 1;
pdev->params.tss_chain_cnt = 1;
if (IS_PFDEV(pdev))
{
pdev->params.sb_cnt = MAX_RSS_CHAINS / pdev->params.vnics_per_port;
if (CHIP_IS_E1x(pdev))
{
pdev->params.max_pf_sb_cnt = pdev->params.fw_sb_cnt = HC_SB_MAX_SB_E1X / 2 / pdev->params.vnics_per_port;
pdev->params.base_fw_ndsb = FUNC_ID(pdev) * pdev->params.fw_sb_cnt;
if (CHIP_IS_E1(pdev)) {
pdev->params.fw_client_cnt = pdev->params.max_pf_fw_client_cnt = ETH_MAX_RX_CLIENTS_E1;
} else {
pdev->params.fw_client_cnt = pdev->params.max_pf_fw_client_cnt = ETH_MAX_RX_CLIENTS_E1H / pdev->params.vnics_per_port;
}
pdev->params.base_fw_client_id = VNIC_ID(pdev) * pdev->params.fw_client_cnt;
}
else
{
#ifdef _VBD_
pdev->params.sb_cnt = LM_IGU_SB_CNT(pdev);
#endif
if (pdev->params.sb_cnt > LM_IGU_SB_CNT(pdev)) {
pdev->params.sb_cnt = LM_IGU_SB_CNT(pdev);
}
#ifndef LM_NUM_DSBS
#define LM_NUM_DSBS 1
#endif
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
pdev->params.base_fw_ndsb = IGU_BASE_NDSB(pdev) - (FUNC_ID(pdev) + 1)* LM_NUM_DSBS;
pdev->params.fw_aux_qzone_cnt = (ETH_MAX_RX_CLIENTS_E2 - PXP_REG_HST_ZONE_PERMISSION_TABLE_SIZE) / pdev->params.vnics_per_port / 2;
port_base_aux_qzone = PORT_ID(pdev)* ((ETH_MAX_RX_CLIENTS_E2 - PXP_REG_HST_ZONE_PERMISSION_TABLE_SIZE)/PORT_MAX);
pdev->params.aux_fw_qzone_id = PXP_REG_HST_ZONE_PERMISSION_TABLE_SIZE + port_base_aux_qzone + VNIC_ID(pdev) * pdev->params.fw_aux_qzone_cnt;
pdev->params.base_fw_client_id = pdev->params.base_fw_ndsb + FUNC_ID(pdev) * MAX_NON_RSS_FW_CLIENTS;
}
else
{
pdev->params.base_fw_ndsb = IGU_BASE_NDSB(pdev) - (VNIC_ID(pdev) + 1) * LM_NUM_DSBS;
pdev->params.fw_aux_qzone_cnt = (ETH_MAX_RX_CLIENTS_E2 - PXP_REG_HST_ZONE_PERMISSION_TABLE_SIZE) / pdev->params.vnics_per_port;
pdev->params.aux_fw_qzone_id = PXP_REG_HST_ZONE_PERMISSION_TABLE_SIZE + VNIC_ID(pdev) * pdev->params.fw_aux_qzone_cnt;
pdev->params.base_fw_client_id = pdev->params.base_fw_ndsb + VNIC_ID(pdev) * MAX_NON_RSS_FW_CLIENTS;
}
pdev->params.fw_sb_cnt = LM_IGU_SB_CNT(pdev);
#ifdef VF_INVOLVED
pdev->params.fw_sb_cnt = pdev->params.fw_sb_cnt + lm_pf_get_vf_available_igu_blocks(pdev);
if ((VNICS_PER_PORT(pdev) == 1) && (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_2))
{
pdev->params.fw_client_cnt = ETH_MAX_RX_CLIENTS_E2;
}
else
{
pdev->params.fw_client_cnt = pdev->params.fw_sb_cnt + MAX_NON_RSS_FW_CLIENTS;
}
#else
pdev->params.fw_client_cnt = pdev->params.fw_sb_cnt + MAX_NON_RSS_FW_CLIENTS;
#endif
pdev->params.fw_base_qzone_cnt = pdev->params.fw_sb_cnt;
base_fw_qzone_id = pdev->params.base_fw_ndsb;
pdev->params.max_pf_sb_cnt = LM_IGU_SB_CNT(pdev);
pdev->params.max_pf_fw_client_cnt = pdev->params.max_pf_sb_cnt + MAX_NON_RSS_FW_CLIENTS;
}
DbgMessage(pdev, WARN, "SB counts(from %d): %d rss, %d max(pf), %d fw ndsbs accessible\n",
pdev->params.base_fw_ndsb, pdev->params.sb_cnt, pdev->params.max_pf_sb_cnt, pdev->params.fw_sb_cnt);
DbgBreakIf(pdev->params.sb_cnt > pdev->params.max_pf_sb_cnt);
DbgBreakIf(pdev->params.max_pf_sb_cnt > pdev->params.fw_sb_cnt);
DbgMessage(pdev, WARN, "FW clients (from %d): %d max(pf), %d fw cliens accessible\n",
pdev->params.base_fw_client_id, pdev->params.max_pf_fw_client_cnt, pdev->params.fw_client_cnt);
if (CHIP_IS_E2E3(pdev)) {
u8_t qz_idx;
for (qz_idx = 0; qz_idx < pdev->params.fw_base_qzone_cnt; qz_idx++)
{
pdev->params.fw_qzone_id[qz_idx] = base_fw_qzone_id + qz_idx;
}
DbgMessage(pdev, WARN, "%d base FW Q zone IDs from %d\n", pdev->params.fw_base_qzone_cnt, base_fw_qzone_id);
DbgMessage(pdev, WARN, "%d aux FW Q zone IDs from %d\n", pdev->params.fw_aux_qzone_cnt, pdev->params.aux_fw_qzone_id);
}
} else if (IS_CHANNEL_VFDEV(pdev)) {
pdev->params.sb_cnt = 16;
} else {
pdev->params.sb_cnt = 1;
}
pdev->params.max_rss_chains = ((IS_PFDEV(pdev) && IGU_U_NDSB_OFFSET(pdev)) ? min(IGU_U_NDSB_OFFSET(pdev),LM_SB_CNT(pdev)) : LM_SB_CNT(pdev));
if (pdev->params.max_rss_chains > MAX_RSS_CHAINS)
{
pdev->params.max_rss_chains = MAX_RSS_CHAINS;
}
#ifndef EDIAG
if(0 == pdev->params.max_rss_chains)
{
DbgBreakMsg("Zero isn't a valid value for pdev->params.max_rss_chains ");
return LM_STATUS_FAILURE;
}
#endif
pdev->params.base_cam_offset = 0;
pdev->params.map_client_to_cid[LM_CLI_IDX_NDIS] = 0;
pdev->params.map_client_to_cid[LM_CLI_IDX_ISCSI] = i = LM_MAX_RSS_CHAINS(pdev);
pdev->params.map_client_to_cid[LM_CLI_IDX_OOO] = ++i;
pdev->params.map_client_to_cid[LM_CLI_IDX_FCOE] = ++i;
pdev->params.map_client_to_cid[LM_CLI_IDX_FWD] = ++i;
pdev->params.start_mp_chain = ++i;
DbgBreakIf(pdev->params.map_client_to_cid[LM_CLI_IDX_OOO] > pdev->params.map_client_to_cid[LM_CLI_IDX_FCOE]);
pdev->params.l4_rss_chain_cnt = 1;
pdev->params.l4_tss_chain_cnt = 1;
pdev->params.l4_rss_base_chain_idx = 0;
if (CHIP_IS_E1x(pdev))
{
pdev->params.l4_base_fw_rss_id = VNIC_ID(pdev) * pdev->params.sb_cnt;
}
else
{
pdev->params.l4_base_fw_rss_id = VNIC_ID(pdev);
}
pdev->params.eth_align_enable = 0;
lm_init_cam_params(pdev);
if((CHIP_REV_IS_SLOW(pdev)
#ifdef DUMMY_MAC_FOR_VF
|| IS_VFDEV(pdev)
#endif
)
&&
(!(GET_FLAGS(pdev->hw_info.mf_info.flags,MF_INFO_VALID_MAC))))
{
pdev->params.mac_addr[0] = pdev->hw_info.mac_addr[0] = 0x00;
pdev->params.mac_addr[1] = pdev->hw_info.mac_addr[1] = 0x50;
pdev->params.mac_addr[2] = pdev->hw_info.mac_addr[2] = 0xc2;
pdev->params.mac_addr[3] = pdev->hw_info.mac_addr[3] = 0x2c;
pdev->params.mac_addr[4] = pdev->hw_info.mac_addr[4] = 0x70 + (IS_PFDEV(pdev) ? 0 : (1 + 64*PATH_ID(pdev) + ABS_VFID(pdev)));
if (CHIP_IS_E1x(pdev))
{
pdev->params.mac_addr[5] = pdev->hw_info.mac_addr[5] = 0x9a + 2 * FUNC_ID(pdev);
}
else
{
pdev->params.mac_addr[5] = pdev->hw_info.mac_addr[5] = 0x9a + PATH_ID(pdev)*8 + PORT_ID(pdev)*4 + VNIC_ID(pdev)*2;
}
mm_memcpy(pdev->hw_info.iscsi_mac_addr, pdev->hw_info.mac_addr, 6);
pdev->hw_info.iscsi_mac_addr[5]++;
mm_memcpy(pdev->hw_info.fcoe_mac_addr, pdev->hw_info.iscsi_mac_addr, 6);
pdev->hw_info.fcoe_mac_addr[5]++;
lm_fcoe_set_default_wwns(pdev);
}
else
{
pdev->params.mac_addr[0] = pdev->hw_info.mac_addr[0];
pdev->params.mac_addr[1] = pdev->hw_info.mac_addr[1];
pdev->params.mac_addr[2] = pdev->hw_info.mac_addr[2];
pdev->params.mac_addr[3] = pdev->hw_info.mac_addr[3];
pdev->params.mac_addr[4] = pdev->hw_info.mac_addr[4];
pdev->params.mac_addr[5] = pdev->hw_info.mac_addr[5];
}
if(CHIP_REV_IS_EMUL(pdev))
{
DbgMessage(pdev, INFORMi, "Emulation is detected.\n");
pdev->params.test_mode |= TEST_MODE_IGNORE_SHMEM_SIGNATURE;
pdev->params.test_mode |= TEST_MODE_LOG_REG_ACCESS;
DbgMessage(pdev, INFORMi , "test mode is 0x%x \n",pdev->params.test_mode);
}
else
{
DbgMessage(pdev, INFORMi, "ASIC is detected.\n");
}
if (!pdev->hw_info.mcp_detected)
{
pdev->params.test_mode |= TEST_MODE_NO_MCP;
}
flow_control = (pdev->hw_info.link_config[ELINK_INT_PHY] & PORT_FEATURE_FLOW_CONTROL_MASK);
switch (flow_control)
{
case PORT_FEATURE_FLOW_CONTROL_AUTO:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_AUTO_PAUSE;
break;
case PORT_FEATURE_FLOW_CONTROL_TX:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_TRANSMIT_PAUSE;
break;
case PORT_FEATURE_FLOW_CONTROL_RX:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_RECEIVE_PAUSE;
break;
case PORT_FEATURE_FLOW_CONTROL_BOTH:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_TRANSMIT_PAUSE | LM_FLOW_CONTROL_RECEIVE_PAUSE;
break;
case PORT_FEATURE_FLOW_CONTROL_NONE:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_NONE;
break;
default:
pdev->params.flow_ctrl_cap = LM_FLOW_CONTROL_NONE;
break;
}
ASSERT_STATIC(MAX_MF_MODE == 4);
if ((pdev->hw_info.mf_info.mf_mode == SINGLE_FUNCTION) ||
(pdev->hw_info.mf_info.mf_mode == MULTI_FUNCTION_SD) ||
(CHIP_IS_E1x(pdev)))
{
pdev->params.l2_fw_flow_ctrl = 0;
}
pdev->params.l4_fw_flow_ctrl = 0;
pdev->params.fw_stats_init_value = TRUE;
pdev->params.mf_mode = pdev->hw_info.mf_info.mf_mode;
if (pdev->params.mf_mode == MULTI_FUNCTION_SD)
{
pdev->params.sd_mode = pdev->hw_info.mf_info.sd_mode;
}
}
else
{
param = param_list;
while(param->offset)
{
if(param->min != 0 || param->max != 0)
{
if(PARAM_VAL(pdev, param) < param->min ||
PARAM_VAL(pdev, param) > param->max)
{
if(CHIP_REV_IS_FPGA(pdev))
{
SET_PARAM_VAL(pdev, param, param->fpga_default);
}
else if(CHIP_REV_IS_EMUL(pdev))
{
SET_PARAM_VAL(pdev, param, param->emulation_default);
}
else
{
SET_PARAM_VAL(pdev, param, param->asic_default);
}
}
}
param++;
}
csize = max(sizeof(struct eth_context),sizeof(struct toe_context));
csize = max(sizeof(struct iscsi_context),csize);
DbgBreakIf(csize>1024);
if((pdev->params.mac_addr[0] == 0 &&
pdev->params.mac_addr[1] == 0 &&
pdev->params.mac_addr[2] == 0 &&
pdev->params.mac_addr[3] == 0 &&
pdev->params.mac_addr[4] == 0 &&
pdev->params.mac_addr[5] == 0) || (pdev->params.mac_addr[0] & 1))
{
DbgMessage(pdev, WARNi, "invalid MAC number.\n");
pdev->params.mac_addr[0] = pdev->hw_info.mac_addr[0];
pdev->params.mac_addr[1] = pdev->hw_info.mac_addr[1];
pdev->params.mac_addr[2] = pdev->hw_info.mac_addr[2];
pdev->params.mac_addr[3] = pdev->hw_info.mac_addr[3];
pdev->params.mac_addr[4] = pdev->hw_info.mac_addr[4];
pdev->params.mac_addr[5] = pdev->hw_info.mac_addr[5];
}
if (CHIP_IS_E1(pdev))
{
if ((pdev->params.l2_fw_flow_ctrl == 1) || (pdev->params.l4_fw_flow_ctrl == 1))
{
DbgMessage(pdev, WARNi, "L2 FW Flow control not supported on E1\n");
pdev->params.l2_fw_flow_ctrl = 0;
pdev->params.l4_fw_flow_ctrl = 0;
}
}
}
for (i = 0; i < ARRSIZE(pdev->params.l2_cli_con_params); i++)
{
lm_cli_idx_t lm_cli_idx = LM_CHAIN_IDX_CLI(pdev, i);
ASSERT_STATIC( ARRSIZE(pdev->params.l2_rx_desc_cnt) == ARRSIZE(pdev->params.mtu));
if( lm_cli_idx >= ARRSIZE(pdev->params.l2_rx_desc_cnt))
{
continue;
}
pdev->params.l2_cli_con_params[i].mtu = pdev->params.mtu[lm_cli_idx];
if(i < (LM_SB_CNT(pdev) + MAX_NON_RSS_CHAINS))
{
pdev->params.l2_cli_con_params[i].num_rx_desc = pdev->params.l2_rx_desc_cnt[lm_cli_idx];
pdev->params.l2_cli_con_params[i].attributes = LM_CLIENT_ATTRIBUTES_RX | LM_CLIENT_ATTRIBUTES_TX | LM_CLIENT_ATTRIBUTES_REG_CLI;
}
else
{
pdev->params.l2_cli_con_params[i].attributes = LM_CLIENT_ATTRIBUTES_TX;
}
}
return LM_STATUS_SUCCESS;
}
lm_status_t
lm_get_dev_info(
lm_device_t *pdev)
{
lm_status_t lm_status = LM_STATUS_SUCCESS;
DbgMessage(pdev, INFORMi , "### lm_get_dev_info\n");
mm_memset( pdev->product_version, 0xff, sizeof(pdev->product_version) );
lm_status = lm_get_pcicfg_info(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
lm_status = lm_get_bars_info(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
if (!IS_CHANNEL_VFDEV(pdev)) {
lm_status = lm_get_chip_id_and_mode(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
}
if (IS_PFDEV(pdev)) {
lm_status = lm_get_function_num(pdev);
if (lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
if ( lm_set_init_arrs(pdev) != 0 ) {
DbgMessage(pdev, FATAL, "Unknown chip revision\n");
return LM_STATUS_UNKNOWN_ADAPTER;
}
} else {
#ifdef VF_INVOLVED
if (IS_VFDEV(pdev))
{
lm_vf_get_vf_id(pdev);
}
#endif
}
#ifdef __LINUX
if (lm_is_function_after_flr(pdev))
{
if (IS_PFDEV(pdev)) {
lm_status = lm_cleanup_after_flr(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
} else {
#ifdef __LINUX
u32_t pcie_caps_offset = mm_get_cap_offset(pdev, PCI_CAP_PCIE);
if (pcie_caps_offset != 0 && pcie_caps_offset != 0xFFFFFFFF) {
u32_t dev_control_and_status = 0xFFFFFFFF;
mm_read_pci(pdev, pcie_caps_offset + PCIE_DEV_CTRL, &dev_control_and_status);
DbgMessage(pdev, FATAL,"Device Control&Status of PCIe caps is %x\n",dev_control_and_status);
if (dev_control_and_status & (PCIE_DEV_STATUS_PENDING_TRANSACTION << 16)) {
DbgBreak();
}
}
#else
DbgMessage(pdev, FATAL, "Function mm_get_cap_offset is not implemented yet\n");
DbgBreak();
#endif
lm_fl_reset_clear_inprogress(pdev);
}
}
#endif
if (IS_CHANNEL_VIRT_MODE_MASTER_PFDEV(pdev)) {
pdev->params.max_eth_including_vfs_conns = 1 << (LM_VF_MAX_RVFID_SIZE + LM_VF_CID_WND_SIZE(pdev) + 1);
} else if (IS_PFDEV(pdev)) {
pdev->params.max_eth_including_vfs_conns = MAX_VF_ETH_CONS;
if(IS_DCB_SUPPORTED_BY_CHIP(pdev))
{
#ifdef _VBD_
pdev->params.max_eth_including_vfs_conns += 3 * MAX_HW_CHAINS + MAX_NON_RSS_CHAINS;
#else
pdev->params.max_eth_including_vfs_conns += MAX_ETH_CONS;
#endif
}
else
{
#ifdef _VBD_
pdev->params.max_eth_including_vfs_conns += MAX_ETH_REG_CHAINS;
#else
pdev->params.max_eth_including_vfs_conns += MAX_ETH_REG_CONS;
#endif
}
}
else
{
pdev->params.max_eth_including_vfs_conns = MAX_RSS_CHAINS;
}
if (IS_PFDEV(pdev)) {
lm_status = lm_get_sriov_info(pdev);
if (lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
lm_status = lm_get_nvm_info(pdev);
if (lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
lm_status = lm_get_shmem_info(pdev);
if (lm_status != LM_STATUS_SUCCESS) {
return lm_status;
}
} else if (IS_CHANNEL_VFDEV(pdev)) {
pdev->hw_info.mf_info.multi_vnics_mode = 0;
pdev->hw_info.mf_info.vnics_per_port = 1;
pdev->hw_info.mf_info.ext_id = 0xffff;
pdev->hw_info.mcp_detected = FALSE;
pdev->hw_info.chip_id = CHIP_NUM_5712E;
pdev->hw_info.max_port_conns = log2_align(MAX_ETH_CONS);
pdev->debug_info.ack_en[0] = 1;
}
#ifdef VF_INVOLVED
if (IS_VFDEV(pdev)) {
lm_vf_enable_vf(pdev);
}
#endif
pdev->ver_num =
(LM_DRIVER_MAJOR_VER << 24) |
(LM_DRIVER_MINOR_VER << 16) |
(LM_DRIVER_FIX_NUM << 8) |
LM_DRIVER_ENG_NUM ;
mm_build_ver_string(pdev);
pdev->ver_num_fw = (BCM_5710_FW_MAJOR_VERSION << 24) |
(BCM_5710_FW_MINOR_VERSION << 16) |
(BCM_5710_FW_REVISION_VERSION<<8) |
(BCM_5710_FW_ENGINEERING_VERSION) ;
pdev->params.vnics_per_port = pdev->hw_info.mf_info.vnics_per_port;
pdev->params.ovlan = VALID_OVLAN(OVLAN(pdev)) ? OVLAN(pdev) : 0;
pdev->params.multi_vnics_mode = pdev->hw_info.mf_info.multi_vnics_mode;
pdev->params.path_has_ovlan = pdev->hw_info.mf_info.path_has_ovlan;
if IS_MULTI_VNIC(pdev)
{
lm_cmng_calc_params(pdev);
}
if (IS_PFDEV(pdev))
{
init_link_params(pdev);
}
if (IS_CHANNEL_VFDEV(pdev))
{
pdev->hw_info.intr_blk_info.blk_type = INTR_BLK_IGU;
pdev->hw_info.intr_blk_info.blk_mode = INTR_BLK_MODE_NORM;
pdev->hw_info.intr_blk_info.access_type = INTR_BLK_ACCESS_IGUMEM;
}
else
{
lm_status = lm_get_intr_blk_info(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
}
lm_status = lm_init_params(pdev, 0);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
lm_status = lm_mcp_cmd_init(pdev);
if( LM_STATUS_SUCCESS != lm_status )
{
DbgMessage(pdev, FATAL, "lm_get_shmem_info: mcp_cmd_init failed. lm_status=0x%x\n", lm_status);
}
if (CHIP_PORT_MODE(pdev) == LM_CHIP_PORT_MODE_4)
{
if (pdev->params.path_has_ovlan && !pdev->params.multi_vnics_mode)
{
DbgMessage(pdev, WARN, "func_id = %d Setting link speed to 1000MBPS\n", ABS_FUNC_ID(pdev));
SET_MEDIUM_SPEED(pdev->params.req_medium, LM_MEDIUM_SPEED_1000MBPS);
}
}
lm_status = mm_get_user_config(pdev);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
lm_status = lm_init_params(pdev, 1);
if(lm_status != LM_STATUS_SUCCESS)
{
return lm_status;
}
DbgMessage(pdev, INFORMi , "### lm_get_dev_info exit\n");
return LM_STATUS_SUCCESS;
}
u8_t lm_get_port_id_from_func_abs( const u32_t chip_num, const lm_chip_port_mode_t lm_chip_port_mode, const u8_t abs_func )
{
u8_t port_id = 0xff;
u8_t modulus_res = 0;
do
{
if( CHIP_IS_E1x_PARAM( chip_num ) )
{
port_id = abs_func % PORT_MAX;
break;
}
switch( lm_chip_port_mode )
{
case LM_CHIP_PORT_MODE_2:
{
DbgBreakIf( CHIP_IS_E1x_PARAM( chip_num ) );
port_id = 0;
}
break;
case LM_CHIP_PORT_MODE_4:
{
modulus_res = abs_func % 4;
switch (modulus_res)
{
case 0:
case 1:
port_id = 0;
break;
case 2:
case 3:
port_id = 1;
break;
default:
break;
}
}
break;
default:
DbgBreakIf(TRUE);
break;
}
}while(0);
return port_id;
}
u8_t lm_get_abs_func_vector( const u32_t chip_num, const lm_chip_port_mode_t chip_port_mode, const u8_t b_multi_vnics_mode, const u8_t path_id )
{
u8_t abs_func_vector = 0;
do
{
if( CHIP_IS_E1x_PARAM(chip_num) && !b_multi_vnics_mode )
{
SET_BIT( abs_func_vector, 0 );
SET_BIT( abs_func_vector, 1 );
break;
}
if( CHIP_IS_E1H_PARAM(chip_num) && b_multi_vnics_mode )
{
SET_BIT( abs_func_vector, 0 );
SET_BIT( abs_func_vector, 1 );
SET_BIT( abs_func_vector, 2 );
SET_BIT( abs_func_vector, 3 );
SET_BIT( abs_func_vector, 4 );
SET_BIT( abs_func_vector, 5 );
SET_BIT( abs_func_vector, 6 );
SET_BIT( abs_func_vector, 7 );
break;
}
DbgBreakIf( CHIP_IS_E1x_PARAM(chip_num) );
switch ( chip_port_mode )
{
case LM_CHIP_PORT_MODE_2:
{
DbgBreakIf( !CHIP_IS_E2_PARAM(chip_num) && !CHIP_IS_E3_PARAM(chip_num) );
if( b_multi_vnics_mode )
{
SET_BIT( abs_func_vector, (0 + path_id) );
SET_BIT( abs_func_vector, (2 + path_id) );
SET_BIT( abs_func_vector, (4 + path_id) );
SET_BIT( abs_func_vector, (6 + path_id) );
break;
}
else
{
SET_BIT( abs_func_vector, path_id );
break;
}
}
break;
case LM_CHIP_PORT_MODE_4:
{
if( b_multi_vnics_mode )
{
if (chip_num != CHIP_NUM_57800)
{
SET_BIT( abs_func_vector, (0 + path_id) );
SET_BIT( abs_func_vector, (2 + path_id) );
SET_BIT( abs_func_vector, (4 + path_id) );
SET_BIT( abs_func_vector, (6 + path_id) );
}
else
{
SET_BIT( abs_func_vector, (0 + path_id) );
SET_BIT( abs_func_vector, (2 + path_id) );
SET_BIT( abs_func_vector, (4 + path_id) );
}
break;
}
else
{
if (chip_num != CHIP_NUM_57800)
{
SET_BIT( abs_func_vector, (0 + path_id) );
SET_BIT( abs_func_vector, (2 + path_id) );
}
else
{
SET_BIT( abs_func_vector, (0 + path_id) );
SET_BIT( abs_func_vector, (2 + path_id) );
SET_BIT( abs_func_vector, (4 + path_id) );
}
break;
}
}
break;
default:
{
DbgBreakIf(TRUE);
break;
}
}
}while(0);
return abs_func_vector;
}
lm_status_t lm_verify_validity_map(lm_device_t *pdev)
{
u64_t wait_cnt = 0 ;
u64_t wait_cnt_limit = 200000;
u32_t val = 0;
lm_status_t lm_status = LM_STATUS_FAILURE ;
if ( CHK_NULL(pdev) )
{
return LM_STATUS_INVALID_PARAMETER ;
}
wait_cnt_limit*= (u64_t)(pdev->vars.clk_factor) ;
for(wait_cnt = 0; wait_cnt < wait_cnt_limit; wait_cnt++)
{
LM_SHMEM_READ(pdev,OFFSETOF(shmem_region_t, validity_map[PORT_ID(pdev)]),&val);
if((val & (SHR_MEM_VALIDITY_DEV_INFO | SHR_MEM_VALIDITY_MB)) == (SHR_MEM_VALIDITY_DEV_INFO|SHR_MEM_VALIDITY_MB))
{
lm_status = LM_STATUS_SUCCESS ;
break;
}
mm_wait(pdev, 20);
}
DbgMessage(pdev, INFORMi, "lm_verify_validity_map: shmem signature %d\n",val);
return lm_status ;
}
lm_status_t
lm_set_cam_params(struct _lm_device_t * pdev,
u32_t mac_requestors_mask,
u32_t base_offset_in_cam_table,
u32_t cam_size,
u32_t mma_size,
u32_t mc_size)
{
lm_status_t lm_status = LM_STATUS_SUCCESS;
if (IS_VFDEV(pdev)) {
return LM_STATUS_FAILURE;
}
if (base_offset_in_cam_table != LM_KEEP_CURRENT_CAM_VALUE) {
pdev->params.base_offset_in_cam_table = (u8_t)base_offset_in_cam_table;
}
if (cam_size != LM_KEEP_CURRENT_CAM_VALUE) {
pdev->params.cam_size = (u8_t)cam_size;
}
if (mc_size != LM_KEEP_CURRENT_CAM_VALUE) {
if (CHIP_IS_E1(pdev)) {
pdev->params.mc_table_size[LM_CLI_IDX_NDIS] =(u8_t) mc_size;
} else {
pdev->params.mc_table_size[LM_CLI_IDX_FCOE] = (u8_t)mc_size;
}
}
return lm_status;
}
void lm_cmng_calc_params(lm_device_t* pdev )
{
u8_t vnic = 0;
DbgBreakIf(!IS_MULTI_VNIC(pdev));
for (vnic = 0; vnic < MAX_VNIC_NUM; vnic++)
{
if (GET_FLAGS(pdev->hw_info.mf_info.func_mf_cfg , FUNC_MF_CFG_FUNC_HIDE))
{
pdev->params.min_bw[vnic] = 0;
pdev->params.max_bw[vnic] = 0;
}
else
{
pdev->params.min_bw[vnic] = pdev->hw_info.mf_info.min_bw[vnic];
pdev->params.max_bw[vnic] = pdev->hw_info.mf_info.max_bw[vnic];
}
}
}
u16_t
lm_get_max_bw(IN const lm_device_t *pdev,
IN const u32_t link_speed,
IN const u8_t vnic)
{
u16_t max_bw = 0;
DbgBreakIf(0 == IS_MULTI_VNIC(pdev));
if(IS_MF_SD_MODE(pdev) || IS_MF_AFEX_MODE(pdev))
{
max_bw = pdev->params.max_bw[vnic]*100;
}
else
{
DbgBreakIf(FALSE == IS_MF_SI_MODE(pdev));
max_bw = (link_speed * pdev->params.max_bw[vnic])/100;
}
return max_bw;
}
u8_t lm_check_if_pf_assigned_to_vm(struct _lm_device_t *pdev)
{
u8_t b_assigned_to_vm = FALSE;
switch (pdev->hw_info.pci_cfg_trust)
{
case PCI_CFG_NOT_TESTED_FOR_TRUST:
break;
case PCI_CFG_NOT_TRUSTED:
b_assigned_to_vm = TRUE;
break;
case PCI_CFG_TRUSTED:
b_assigned_to_vm = FALSE;
break;
}
return b_assigned_to_vm;
}
u8_t lm_is_fw_version_valid(struct _lm_device_t *pdev)
{
u8_t is_fw_valid = FALSE;
u32_t drv_fw_ver = (BCM_5710_FW_MAJOR_VERSION) |
(BCM_5710_FW_MINOR_VERSION << 8) |
(BCM_5710_FW_REVISION_VERSION << 16) |
(BCM_5710_FW_ENGINEERING_VERSION << 24) ;
u32_t real_fw_ver = REG_RD(pdev,0x2c0000);
u32_t fw_valid_mask;
fw_valid_mask = SWAP_BYTES32(pdev->params.fw_valid_mask);
is_fw_valid = (((drv_fw_ver ^ real_fw_ver) & fw_valid_mask) == 0);
return (is_fw_valid);
}
lm_status_t
lm_set_cli_drv_ver_to_shmem(struct _lm_device_t *pdev)
{
u32_t drv_ver_offset = OFFSETOF(shmem2_region_t,func_os_drv_ver);
u32_t offset = 0;
lm_status_t lm_status = LM_STATUS_SUCCESS;
u32_t shmem2_size = 0;
u32_t index = 0;
if (IS_VFDEV(pdev))
{
return LM_STATUS_SUCCESS;
}
ASSERT_STATIC( sizeof(pdev->lm_cli_drv_ver_to_shmem.cli_drv_ver) == sizeof(struct os_drv_ver) );
offset = drv_ver_offset + (pdev->params.pfunc_mb_id * sizeof(pdev->lm_cli_drv_ver_to_shmem.cli_drv_ver));
DbgMessage(pdev, WARN,"offset= %d \n", offset);
if (pdev->hw_info.shmem_base2 != 0)
{
LM_SHMEM2_READ (pdev, OFFSETOF(shmem2_region_t,size), &shmem2_size);
if (shmem2_size > offset)
{
for (index = 0; index < ARRSIZE(pdev->lm_cli_drv_ver_to_shmem.cli_drv_ver.versions); index++)
{
LM_SHMEM2_WRITE(pdev, offset, pdev->lm_cli_drv_ver_to_shmem.cli_drv_ver.versions[index]);
offset+= sizeof( pdev->lm_cli_drv_ver_to_shmem.cli_drv_ver.versions[index] );
}
}
}
return lm_status;
}
u8_t lm_is_mac_locally_administrated(struct _lm_device_t *pdev, u8_t * mac)
{
u8_t res = FALSE;
if (mac != NULL)
{
res = (mac[0] != pdev->params.mac_addr[0]) ||
(mac[1] != pdev->params.mac_addr[1]) ||
(mac[2] != pdev->params.mac_addr[2]) ||
(mac[3] != pdev->params.mac_addr[3]) ||
(mac[4] != pdev->params.mac_addr[4]) ||
(mac[5] != pdev->params.mac_addr[5]);
}
return res;
}