num_tx_queues
i = m->m_pkthdr.flowid % adapter->num_tx_queues;
i = curcpu % adapter->num_tx_queues;
for (qid = 0; qid < adapter->num_tx_queues; qid++) {
msix_vecs = 1 + adapter->num_rx_queues + adapter->num_tx_queues;
for (i = 0; i < adapter->num_tx_queues; i++) {
for (i = 0; i < adapter->num_tx_queues; i++) {
for (i = 0; i < adapter->num_tx_queues; i++)
for (i = 0; i < adapter->num_tx_queues; i++) {
uint32_t group_c_mask = (1 << adapter->num_tx_queues) - 1;/* bit per Tx q*/
for (i = 0; i < adapter->num_tx_queues; i++)
adapter->num_tx_queues = AL_ETH_NUM_QUEUES;
for (i = 0; i < adapter->num_tx_queues; i++) {
int num_tx_queues;
u16 num_tx_queues;
#define IAVF_NTXQS(_vsi) ((_vsi)->num_tx_queues)
pairs = max(vsi->num_tx_queues, vsi->num_rx_queues);
for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) {
for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) {
for (int q = 0; q < vsi->num_tx_queues; q++) {
for (int i = 0; i < vsi->num_tx_queues; i++) {
vsi->num_tx_queues = vsi->shared->isc_ntxqsets;
for (int i = 0; i < vsi->num_tx_queues; i++, tx_que++)
return (m->m_pkthdr.flowid % sc->pf_vsi.num_tx_queues);
if (bit >= vsi->num_tx_queues) {
for (int i = 0; i < vf->vsi->num_tx_queues; i++)
for (int i = 0; i < vsi->num_tx_queues; i++) {
vsi->num_tx_queues = vsi->num_rx_queues = vf_num_queues;
vsi->num_tx_queues);
malloc(sizeof(struct ice_tx_queue) * vsi->num_tx_queues, M_ICE,
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) {
vf_res->num_queue_pairs = vf->vsi->num_tx_queues;
vsi_res->num_queue_pairs = vf->vsi->num_tx_queues;
if (vqci->num_queue_pairs > vf->vsi->num_tx_queues &&
for (i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0; i < vsi->num_tx_queues; i++)
vsi->num_tx_queues);
vsi->tc_info[0].qcount_tx = vsi->num_tx_queues;
if (vsi->num_tx_queues > 255)
q_teids_size = sizeof(*q_teids) * vsi->num_tx_queues;
q_ids_size = sizeof(*q_ids) * vsi->num_tx_queues;
q_handles_size = sizeof(*q_handles) * vsi->num_tx_queues;
max_txqs[0] = vsi->num_tx_queues;
int num_tx_queues;
vsi->num_tx_queues = ntxqsets;
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) {
vsi->num_tx_queues = 0;
if (vsi->num_rx_queues != vsi->num_tx_queues) {
vsi->num_tx_queues, vsi->num_rx_queues);
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) {
sc->pf_vsi.num_tx_queues);
vsi->num_tx_queues);
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++)
sc->pf_vsi.num_tx_queues);
sc->pf_vsi.num_tx_queues);
vsi->num_tx_queues = vsi->num_rx_queues = ICE_DEFAULT_VF_QUEUES;
vsi->num_tx_queues);
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++)
ice_resmgr_release_map(&sc->tx_qmgr, vsi->tx_qmap, vsi->num_tx_queues);
MPASS(vsi->num_tx_queues == ntxqsets);
if (vsi->num_rx_queues != vsi->num_tx_queues) {
vsi->num_tx_queues, vsi->num_rx_queues);
for (i = 0; i < vsi->num_tx_queues; i++)
for (i = 0, txq = vsi->tx_queues; i < vsi->num_tx_queues; i++, txq++) {
for (i = 0; i < vsi->num_tx_queues; i++)
for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues;
for (int i = 0; i < sc->num_tx_queues; i++) {
for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues;
for (int i = 0; i < sc->num_tx_queues; i++, tx_que++) {
MPASS(sc->num_tx_queues > 0);
MPASS(sc->num_tx_queues == ntxqsets);
sc->num_tx_queues);
for (i = 0; i < sc->num_tx_queues; i++, tx_que++) {
for (i = 0, que = sc->tx_queues; i < sc->num_tx_queues;
for (int i = 0; i < sc->num_tx_queues; i++) {
for (i = 0; i < sc->num_tx_queues; i++, que++) {
MPASS(sc->num_rx_queues == sc->num_tx_queues);
for (int i = 0; i < sc->num_tx_queues; i++, tx_que++) {
MPASS(sc->num_tx_queues == ntxqsets);
sc->num_tx_queues);
for (i = 0; i < sc->num_tx_queues; i++, que++) {
for (i = 0; i < sc->num_tx_queues; i++) {
for (i = 0, que = vsi->tx_queues; i < vsi->num_tx_queues; i++, que++) {
vsi->num_tx_queues = vsi->shared->isc_ntxqsets;
max(vsi->num_rx_queues, vsi->num_tx_queues), &pf->qtag);
int num_tx_queues;
for (int i = 0; i < max(vsi->num_rx_queues, vsi->num_tx_queues); i++, vector++) {
for (int i = 0; i < vsi->num_tx_queues; i++, que++) {
for (int i = 0; i < vsi->num_tx_queues; i++, tx_que++) {
for (int i = 0; i < vsi->num_tx_queues; i++)
for (int i = 0; i < vsi->num_tx_queues; i++)
error = ixl_pf_qmgr_alloc_contiguous(&pf->qmgr, vsi->num_tx_queues, &pf->qtag);
vf->vsi.num_tx_queues = vf->qtag.num_active;
for (i = 0; i < vf->vsi.num_tx_queues; i++) {
for (i = 0; i < vf->vsi.num_tx_queues; i++)
vf->vsi.num_tx_queues = 0;
reply.num_queue_pairs = vf->vsi.num_tx_queues;
reply.vsi_res[0].num_queue_pairs = vf->vsi.num_tx_queues;
if (info->num_queue_pairs == 0 || info->num_queue_pairs > vf->vsi.num_tx_queues) {
vf->vf_num, info->num_queue_pairs, vf->vsi.num_tx_queues);
pair->txq.queue_id >= vf->vsi.num_tx_queues) {
if (largest_txq >= vf->vsi.num_tx_queues) {
if (i >= vf->vsi.num_tx_queues) {
if (i >= vf->vsi.num_tx_queues) {
for (int i = 0; i < vsi->num_tx_queues; i++) {
for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) {
for (i = 0, tx_que = vsi->tx_queues; i < vsi->num_tx_queues; i++, tx_que++) {
for (int q = 0; q < vsi->num_tx_queues; q++) {
ifp->num_tx_queues, ifp->real_num_tx_queues,