LM_PAGE_SIZE
cid * LM_PAGE_SIZE;
pdev->hw_info.bin_size) + 256) * LM_PAGE_SIZE;
kq->kwqe_left = (LM_PAGE_SIZE/sizeof(kwqe_t)) * page_cnt - 1;
kq->kwqe_left = (LM_PAGE_SIZE/sizeof(kwqe_t)) * page_cnt - 1;
val = (LM_PAGE_SIZE/sizeof(kwqe_t) - 1) << 16;
val = (LM_PAGE_SIZE/sizeof(kwqe_t)) << 16;
(LM_PAGE_SIZE/sizeof(kcqe_t)) * page_cnt - 1;
val = (LM_PAGE_SIZE/sizeof(kwqe_t) - 1) << 16;
val = (LM_PAGE_SIZE/sizeof(kcqe_t)) << 16;
for(cnt = 0; cnt < LM_PAGE_SIZE; cnt += 4)
mem_virt += LM_PAGE_SIZE;
LM_INC64(&mem_phy, LM_PAGE_SIZE);
mem_size -= LM_PAGE_SIZE;
LM_INC64(&mem_phy, LM_PAGE_SIZE);
*((u8_t **) next_ptr->tx_bd_next_reserved) = mem_virt + LM_PAGE_SIZE;
mem_virt += LM_PAGE_SIZE;
LM_INC64(&page_base_phy, LM_PAGE_SIZE);
if(pdev->params.kcq_history_size > (LM_PAGE_SIZE/sizeof(kcqe_t)) *
pdev->params.kcq_history_size = ((LM_PAGE_SIZE/sizeof(kcqe_t)) *
mem_size = LM_PAGE_SIZE * pdev->params.kwq_page_cnt;
mem_size = LM_PAGE_SIZE * pdev->params.kcq_page_cnt;
mem_size = LM_PAGE_SIZE * pdev->params.gen_bd_page_cnt;
mem_size = LM_PAGE_SIZE * pdev->params.hcopy_bd_page_cnt;
mem_size = LM_PAGE_SIZE * bd_page_cnt;
mem_size = LM_PAGE_SIZE * bd_page_cnt;
page_align_delta = LM_PAGE_SIZE - page_align_delta;
DbgBreakIf(num_wqes != (LM_PAGE_SIZE/sizeof(kwqe_t)) *
#define BD_PER_PAGE (LM_PAGE_SIZE/sizeof(tx_bd_t))
#define LM_PAGE_MASK (LM_PAGE_SIZE - 1)
LM_PAGE_SIZE, /* dma_attr_align */
#define BD_PER_PAGE(bd_size) (LM_PAGE_SIZE/(bd_size))
LM_INC64(&mem_phy, LM_PAGE_SIZE);
*((u8_t **) next_bd->reserved) = mem_virt + LM_PAGE_SIZE;
mem_virt += LM_PAGE_SIZE;
LM_INC64(&buf_base_phy, LM_PAGE_SIZE);
buf_base_virt = (u8_t *)buf_base_virt + LM_PAGE_SIZE;
mm_memset(bd_chain->bd_chain_virt, 0, bd_chain->page_cnt * LM_PAGE_SIZE);
LM_INC64(&mem_phys, LM_PAGE_SIZE);
{ _OFFSET(l4_gen_buf_size), LM_PAGE_SIZE,LM_PAGE_SIZE,LM_PAGE_SIZE,LM_PAGE_SIZE,16*LM_PAGE_SIZE },
REG_WR(pdev,PXP2_REG_RQ_CDU_P_SIZE,LOG2(pdev->params.ilt_client_page_size/LM_PAGE_SIZE));
REG_WR(pdev,PXP2_REG_RQ_TM_P_SIZE,LOG2(pdev->params.ilt_client_page_size/LM_PAGE_SIZE));
REG_WR(pdev,PXP2_REG_RQ_QM_P_SIZE,LOG2(pdev->params.ilt_client_page_size/LM_PAGE_SIZE));
REG_WR(pdev,PXP2_REG_RQ_SRC_P_SIZE,LOG2(pdev->params.ilt_client_page_size/LM_PAGE_SIZE));
LM_PAGE_SIZE,
mm_mem_zero(sq_info->sq_chain.sq_chain_virt, LM_PAGE_SIZE);
mem_size = page_cnt * LM_PAGE_SIZE;
mm_mem_zero(sq_info->sq_chain.sq_chain_virt, LM_PAGE_SIZE);
u32_t const mem_size = page_cnt * LM_PAGE_SIZE;
u32_t const mem_size = page_cnt * LM_PAGE_SIZE;
u32_t const mem_size = page_cnt * LM_PAGE_SIZE;
mem_size = page_cnt * LM_PAGE_SIZE;
#define LM_PAGE_MASK (LM_PAGE_SIZE - 1)
LM_PAGE_SIZE - sizeof(lm_frag_list_t)) / sizeof(lm_frag_t);
gen_buf_size = (gen_buf_size + (LM_PAGE_SIZE-1)) & ~(LM_PAGE_SIZE-1);
#if (LM_PAGE_SIZE != 4096)
mem_size = LM_PAGE_SIZE;
mem_size = ((mem_size / LM_PAGE_SIZE) + 1) * LM_PAGE_SIZE;
mem_size += pdev->params.l4_rx_chain_page_cnt*LM_PAGE_SIZE + /* rx bd-chain */
pdev->params.l4_tx_chain_page_cnt*LM_PAGE_SIZE; /* tx bd-chain */
mem_size = pdev->params.l4_scq_page_cnt * LM_PAGE_SIZE;
mem_size = pdev->params.l4_rcq_page_cnt * LM_PAGE_SIZE;
mem_size = pdev->params.l4_grq_page_cnt * LM_PAGE_SIZE;
#define NEXT_EQ_PAGE_ADDRESS_OFFSET (LM_PAGE_SIZE - sizeof(struct iscsi_kcqe))
mem_size = page_cnt * LM_PAGE_SIZE;
return CEIL_DIV(bufferSize, LM_PAGE_SIZE);
LM_PAGE_SIZE,
pdev->iscsi_info.run_time.hq_size = (u16_t)(hq_pbl_entries * (LM_PAGE_SIZE / sizeof(struct iscsi_hq_bd)));
LM_INTMEM_WRITE16(pdev, TSTORM_ISCSI_PAGE_SIZE_OFFSET(func), LM_PAGE_SIZE, BAR_TSTRORM_INTMEM);
LM_INTMEM_WRITE16(pdev, USTORM_ISCSI_PAGE_SIZE_OFFSET(func), LM_PAGE_SIZE, BAR_USTRORM_INTMEM);
LM_INTMEM_WRITE16(pdev, XSTORM_ISCSI_PAGE_SIZE_OFFSET(func), LM_PAGE_SIZE, BAR_XSTRORM_INTMEM);
LM_INTMEM_WRITE16(pdev, CSTORM_ISCSI_PAGE_SIZE_OFFSET(func), LM_PAGE_SIZE, BAR_CSTRORM_INTMEM);
if (required_page_size < LM_PAGE_SIZE)
required_page_size = LM_PAGE_SIZE;