bin/kenv/kenv.c
101
eq = strchr(env, '=');
bin/kenv/kenv.c
102
if (eq != NULL) {
bin/kenv/kenv.c
103
*eq++ = '\0';
bin/kenv/kenv.c
104
val = eq;
bin/kenv/kenv.c
65
char *env, *eq, *val;
crypto/heimdal/lib/gssapi/mech/gss_mo.c
406
int eq = 0;
crypto/heimdal/lib/gssapi/mech/gss_mo.c
413
eq = gss_oid_equal(mi->gm_mo[m].option,
crypto/heimdal/lib/gssapi/mech/gss_mo.c
415
if (eq)
crypto/heimdal/lib/gssapi/mech/gss_mo.c
420
eq = gss_oid_equal(&mech_attrs->elements[m],
crypto/heimdal/lib/gssapi/mech/gss_mo.c
422
if (eq)
crypto/heimdal/lib/gssapi/mech/gss_mo.c
426
if (!eq ^ except)
crypto/heimdal/lib/roken/putenv.c
54
const char *eq = (const char *)strchr(string, '=');
crypto/heimdal/lib/roken/putenv.c
57
if (eq == NULL)
crypto/heimdal/lib/roken/putenv.c
59
len = eq - string;
crypto/krb5/src/kdc/t_ndr.c
132
int eq;
crypto/krb5/src/kdc/t_ndr.c
162
eq = data_eq(data_in, data_out);
crypto/krb5/src/kdc/t_ndr.c
164
if (!eq) {
crypto/krb5/src/lib/kdb/kdb5.c
1061
krb5_boolean eq;
crypto/krb5/src/lib/kdb/kdb5.c
1080
eq = krb5_principal_compare(kcontext, entry->princ, source);
crypto/krb5/src/lib/kdb/kdb5.c
1082
if (!eq)
crypto/krb5/src/lib/krb5/ccache/cccursor.c
153
krb5_boolean eq;
crypto/krb5/src/lib/krb5/ccache/cccursor.c
165
eq = krb5_principal_compare(context, princ, client);
crypto/krb5/src/lib/krb5/ccache/cccursor.c
167
if (eq)
crypto/krb5/src/lib/krb5/krb/princ_comp.c
113
krb5_boolean eq;
crypto/krb5/src/lib/krb5/krb/princ_comp.c
117
eq = (krb5int_utf8_normcmp(p1, p2, KRB5_UTF8_CASEFOLD) == 0);
crypto/krb5/src/lib/krb5/krb/princ_comp.c
119
eq = (p1->length == p2->length
crypto/krb5/src/lib/krb5/krb/princ_comp.c
122
eq = data_eq(*p1, *p2);
crypto/krb5/src/lib/krb5/krb/princ_comp.c
124
if (!eq)
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
668
char *eq;
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
670
if ((eq = strchr(s, '=')) == NULL)
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
672
*eq++ = '\0';
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
673
if (!isdigit((int) *eq))
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
679
ib.flags = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
683
ib.cachesize = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
687
ib.maxkeypage = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
691
ib.minkeypage = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
695
ib.lorder = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
699
ib.psize = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
705
ih.bsize = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
709
ih.ffactor = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
713
ih.nelem = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
717
ih.cachesize = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
721
ih.lorder = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
727
rh.flags = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
731
rh.cachesize = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
735
rh.lorder = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
739
rh.reclen = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
743
rh.bval = atoi(eq);
crypto/krb5/src/plugins/kdb/db2/libdb2/test/dbtest.c
747
rh.psize = atoi(eq);
crypto/openssl/crypto/property/property_parse.c
490
const int eq = q[i].type == d[j].type
crypto/openssl/crypto/property/property_parse.c
493
if ((eq && oper == OSSL_PROPERTY_OPER_EQ)
crypto/openssl/crypto/property/property_parse.c
494
|| (!eq && oper == OSSL_PROPERTY_OPER_NE))
crypto/openssl/test/testutil.h
308
DECLARE_COMPARISON(type, name, eq) \
crypto/openssl/test/testutil.h
340
DECLARE_COMPARISON(void *, ptr, eq)
crypto/openssl/test/testutil.h
351
DECLARE_COMPARISON(char *, str, eq)
crypto/openssl/test/testutil/tests.c
223
DEFINE_COMPARISON(type, name, eq, ==, fmt, cast) \
crypto/openssl/test/testutil/tests.c
241
DEFINE_COMPARISON(void *, ptr, eq, ==, "%p", void *)
crypto/openssl/test/testutil/tests.c
381
DEFINE_BN_COMPARISONS(eq, ==, BN_is_zero(a))
crypto/openssl/test/testutil/tests.c
467
DEFINE_TIME_T_COMPARISON(eq, ==)
lib/libc/db/test/dbtest.c
577
char *eq, *index();
lib/libc/db/test/dbtest.c
579
if ((eq = index(s, '=')) == NULL)
lib/libc/db/test/dbtest.c
581
*eq++ = '\0';
lib/libc/db/test/dbtest.c
582
if (!isdigit(*eq))
lib/libc/db/test/dbtest.c
588
ib.flags = atoi(eq);
lib/libc/db/test/dbtest.c
592
ib.cachesize = atoi(eq);
lib/libc/db/test/dbtest.c
596
ib.maxkeypage = atoi(eq);
lib/libc/db/test/dbtest.c
600
ib.minkeypage = atoi(eq);
lib/libc/db/test/dbtest.c
604
ib.lorder = atoi(eq);
lib/libc/db/test/dbtest.c
608
ib.psize = atoi(eq);
lib/libc/db/test/dbtest.c
614
ih.bsize = atoi(eq);
lib/libc/db/test/dbtest.c
618
ih.ffactor = atoi(eq);
lib/libc/db/test/dbtest.c
622
ih.nelem = atoi(eq);
lib/libc/db/test/dbtest.c
626
ih.cachesize = atoi(eq);
lib/libc/db/test/dbtest.c
630
ih.lorder = atoi(eq);
lib/libc/db/test/dbtest.c
636
rh.flags = atoi(eq);
lib/libc/db/test/dbtest.c
640
rh.cachesize = atoi(eq);
lib/libc/db/test/dbtest.c
644
rh.lorder = atoi(eq);
lib/libc/db/test/dbtest.c
648
rh.reclen = atoi(eq);
lib/libc/db/test/dbtest.c
652
rh.bval = atoi(eq);
lib/libc/db/test/dbtest.c
656
rh.psize = atoi(eq);
lib/libc/tests/stdio/scanfloat_test.c
103
ATF_REQUIRE(eq(DBL, d, 12e6));
lib/libc/tests/stdio/scanfloat_test.c
106
ATF_REQUIRE(eq(DBL, d, 1.0));
lib/libc/tests/stdio/scanfloat_test.c
110
ATF_REQUIRE(eq(DBL, d, 0.0));
lib/libc/tests/stdio/scanfloat_test.c
126
ATF_REQUIRE(eq(LDBL, ld, 3.14159265358979323846L));
lib/libc/tests/stdio/scanfloat_test.c
63
ATF_REQUIRE(eq(FLT, f, 3.141592));
lib/libc/tests/stdio/scanfloat_test.c
66
ATF_REQUIRE(eq(DBL, d, 3.141592653589793));
lib/libc/tests/stdio/scanfloat_test.c
69
ATF_REQUIRE(eq(FLT, f, 1.234568e+06));
lib/libc/tests/stdio/scanfloat_test.c
72
ATF_REQUIRE(eq(DBL, d, -1.234568e6));
lib/libc/tests/stdio/scanfloat_test.c
75
ATF_REQUIRE(eq(LDBL, ld, 1.234568e-52L));
lib/libc/tests/stdio/scanfloat_test.c
78
ATF_REQUIRE(eq(DBL, d, 0.1));
lib/libc/tests/stdio/scanfloat_test.c
81
ATF_REQUIRE(eq(DBL, d, 0.2));
lib/libc/tests/stdio/scanfloat_test.c
84
ATF_REQUIRE(eq(DBL, d, 12345.));
lib/libc/tests/stdio/scanfloat_test.c
91
ATF_REQUIRE(eq(FLT, f, -1.23));
lib/libc/tests/stdio/scanfloat_test.c
95
ATF_REQUIRE(eq(DBL, d, 1.25));
lib/libc/tests/stdio/scanfloat_test.c
99
ATF_REQUIRE(eq(DBL, d, 1.23e4));
lib/libc/tests/string/memcmp_test.c
155
ATF_TP_ADD_TC(tp, eq);
lib/libc/tests/string/memcmp_test.c
75
ATF_TC_WITHOUT_HEAD(eq);
lib/libc/tests/string/memcmp_test.c
76
ATF_TC_BODY(eq, tc)
lib/msun/tests/test-utils.h
157
bool eq = fpequal_tol(_x, _y, tol, flags); \
lib/msun/tests/test-utils.h
158
long double _diff = eq ? 0.0L : fabsl(_x - _y); \
lib/msun/tests/test-utils.h
159
ATF_##atf_variant##_MSG(eq, \
sbin/reboot/reboot.c
202
char *eq;
sbin/reboot/reboot.c
205
eq = strchr(raw, '=');
sbin/reboot/reboot.c
206
if (eq == NULL)
sbin/reboot/reboot.c
208
*eq++ = '\0';
sbin/reboot/reboot.c
209
len = strlen(eq);
sbin/reboot/reboot.c
212
if (eq[0] == '"') {
sbin/reboot/reboot.c
213
if (len < 2 || eq[len - 1] != '"')
sbin/reboot/reboot.c
214
errx(1, "Invalid string '%s'", eq);
sbin/reboot/reboot.c
215
eq[len - 1] = '\0';
sbin/reboot/reboot.c
216
return (eq + 1);
sbin/reboot/reboot.c
218
return (eq);
sys/dev/bxe/bxe.c
14678
sc->eq = (union event_ring_elem * )sc->eq_dma.vaddr;
sys/dev/bxe/bxe.c
14688
sc->eq = NULL;
sys/dev/bxe/bxe.c
14707
sc->eq = NULL;
sys/dev/bxe/bxe.c
14728
sc->eq = NULL;
sys/dev/bxe/bxe.c
14747
sc->eq = NULL;
sys/dev/bxe/bxe.c
15234
sc->eq = NULL;
sys/dev/bxe/bxe.c
8484
elem = &sc->eq[EQ_DESC(sw_cons)];
sys/dev/bxe/bxe.c
9876
elem = &sc->eq[EQ_DESC_CNT_PAGE * i - 1];
sys/dev/bxe/bxe.h
1566
union event_ring_elem *eq;
sys/dev/cxgbe/adapter.h
1142
#define EQ_LOCK(eq) mtx_lock(&(eq)->eq_lock)
sys/dev/cxgbe/adapter.h
1143
#define EQ_TRYLOCK(eq) mtx_trylock(&(eq)->eq_lock)
sys/dev/cxgbe/adapter.h
1144
#define EQ_UNLOCK(eq) mtx_unlock(&(eq)->eq_lock)
sys/dev/cxgbe/adapter.h
1145
#define EQ_LOCK_ASSERT_OWNED(eq) mtx_assert(&(eq)->eq_lock, MA_OWNED)
sys/dev/cxgbe/adapter.h
1146
#define EQ_LOCK_ASSERT_NOTOWNED(eq) mtx_assert(&(eq)->eq_lock, MA_NOTOWNED)
sys/dev/cxgbe/adapter.h
1148
#define TXQ_LOCK(txq) EQ_LOCK(&(txq)->eq)
sys/dev/cxgbe/adapter.h
1149
#define TXQ_TRYLOCK(txq) EQ_TRYLOCK(&(txq)->eq)
sys/dev/cxgbe/adapter.h
1150
#define TXQ_UNLOCK(txq) EQ_UNLOCK(&(txq)->eq)
sys/dev/cxgbe/adapter.h
1151
#define TXQ_LOCK_ASSERT_OWNED(txq) EQ_LOCK_ASSERT_OWNED(&(txq)->eq)
sys/dev/cxgbe/adapter.h
1152
#define TXQ_LOCK_ASSERT_NOTOWNED(txq) EQ_LOCK_ASSERT_NOTOWNED(&(txq)->eq)
sys/dev/cxgbe/adapter.h
1379
tx_resume_threshold(struct sge_eq *eq)
sys/dev/cxgbe/adapter.h
1383
return (eq->sidx / 4);
sys/dev/cxgbe/adapter.h
1670
if (__predict_true(wrq->eq.flags & EQ_HW_ALLOCATED))
sys/dev/cxgbe/adapter.h
635
struct sge_eq eq; /* MUST be first */
sys/dev/cxgbe/adapter.h
780
struct sge_eq eq; /* MUST be first */
sys/dev/cxgbe/crypto/t6_kern_tls.c
1085
copy_to_txd(struct sge_eq *eq, caddr_t from, caddr_t *to, int len)
sys/dev/cxgbe/crypto/t6_kern_tls.c
1088
MPASS((uintptr_t)(*to) >= (uintptr_t)&eq->desc[0]);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1089
MPASS((uintptr_t)(*to) < (uintptr_t)&eq->desc[eq->sidx]);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1092
(uintptr_t)&eq->desc[eq->sidx])) {
sys/dev/cxgbe/crypto/t6_kern_tls.c
1095
if ((uintptr_t)(*to) == (uintptr_t)&eq->desc[eq->sidx])
sys/dev/cxgbe/crypto/t6_kern_tls.c
1096
(*to) = (caddr_t)eq->desc;
sys/dev/cxgbe/crypto/t6_kern_tls.c
1098
int portion = (uintptr_t)&eq->desc[eq->sidx] - (uintptr_t)(*to);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1103
bcopy(from, (void *)eq->desc, portion);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1104
(*to) = (caddr_t)eq->desc + portion;
sys/dev/cxgbe/crypto/t6_kern_tls.c
1153
copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1160
copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1162
copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
sys/dev/cxgbe/crypto/t6_kern_tls.c
1172
copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1185
copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1188
copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, pktlen -
sys/dev/cxgbe/crypto/t6_kern_tls.c
1254
copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1261
copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1263
copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
sys/dev/cxgbe/crypto/t6_kern_tls.c
1273
copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1286
copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1289
copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, m->m_len -
sys/dev/cxgbe/crypto/t6_kern_tls.c
1293
copy_to_txd(&txq->eq, (char *)m_tls->m_epg_hdr +
sys/dev/cxgbe/crypto/t6_kern_tls.c
1318
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/crypto/t6_kern_tls.c
1347
using_scratch = (eq->sidx - pidx < SGE_MAX_WR_LEN / EQ_ESIZE);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1588
copy_to_txd(eq, txq->ss, &out, wr_len);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1598
IDXINCR(pidx, ndesc, eq->sidx);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1599
dst = &eq->desc[pidx];
sys/dev/cxgbe/crypto/t6_kern_tls.c
1608
using_scratch = (eq->sidx - pidx <
sys/dev/cxgbe/crypto/t6_kern_tls.c
1634
V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1828
copy_to_txd(eq, txq->ss, &out, wr_len);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1899
copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1906
copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1908
copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
sys/dev/cxgbe/crypto/t6_kern_tls.c
1918
copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1931
copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));
sys/dev/cxgbe/crypto/t6_kern_tls.c
1934
copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, m->m_len -
sys/dev/cxgbe/crypto/t6_kern_tls.c
1953
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/crypto/t6_kern_tls.c
1973
pidx = eq->pidx;
sys/dev/cxgbe/crypto/t6_kern_tls.c
1983
IDXINCR(pidx, ndesc, eq->sidx);
sys/dev/cxgbe/crypto/t6_kern_tls.c
1984
dst = &eq->desc[pidx];
sys/dev/cxgbe/crypto/t6_kern_tls.c
2017
IDXINCR(pidx, ndesc, eq->sidx);
sys/dev/cxgbe/crypto/t6_kern_tls.c
2018
dst = &eq->desc[pidx];
sys/dev/cxgbe/crypto/t6_kern_tls.c
2047
IDXINCR(pidx, ndesc, eq->sidx);
sys/dev/cxgbe/crypto/t6_kern_tls.c
2048
dst = &eq->desc[pidx];
sys/dev/cxgbe/crypto/t6_kern_tls.c
288
V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1124
copy_to_txd(struct sge_eq *eq, const char *from, caddr_t *to, int len)
sys/dev/cxgbe/crypto/t7_kern_tls.c
1127
MPASS((uintptr_t)(*to) >= (uintptr_t)&eq->desc[0]);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1128
MPASS((uintptr_t)(*to) < (uintptr_t)&eq->desc[eq->sidx]);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1131
(uintptr_t)&eq->desc[eq->sidx])) {
sys/dev/cxgbe/crypto/t7_kern_tls.c
1134
if ((uintptr_t)(*to) == (uintptr_t)&eq->desc[eq->sidx])
sys/dev/cxgbe/crypto/t7_kern_tls.c
1135
(*to) = (caddr_t)eq->desc;
sys/dev/cxgbe/crypto/t7_kern_tls.c
1137
int portion = (uintptr_t)&eq->desc[eq->sidx] - (uintptr_t)(*to);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1142
bcopy(from, (void *)eq->desc, portion);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1143
(*to) = (caddr_t)eq->desc + portion;
sys/dev/cxgbe/crypto/t7_kern_tls.c
1193
copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1200
copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1202
copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
sys/dev/cxgbe/crypto/t7_kern_tls.c
1209
copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1218
copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1221
copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, m->m_len -
sys/dev/cxgbe/crypto/t7_kern_tls.c
1225
copy_to_txd(&txq->eq, src, &out, len);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1246
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/crypto/t7_kern_tls.c
1541
using_scratch = (eq->sidx - pidx < ndesc);
sys/dev/cxgbe/crypto/t7_kern_tls.c
1560
V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1952
V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
sys/dev/cxgbe/crypto/t7_kern_tls.c
1997
copy_to_txd(eq, txq->ss, &out, wr_len);
sys/dev/cxgbe/crypto/t7_kern_tls.c
2023
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/crypto/t7_kern_tls.c
2048
pidx = eq->pidx;
sys/dev/cxgbe/crypto/t7_kern_tls.c
2078
IDXINCR(pidx, ndesc, eq->sidx);
sys/dev/cxgbe/crypto/t7_kern_tls.c
2079
dst = &eq->desc[pidx];
sys/dev/cxgbe/t4_main.c
2184
txq->eq.flags &= ~(EQ_ENABLED | EQ_HW_ALLOCATED);
sys/dev/cxgbe/t4_main.c
2190
ofld_txq->wrq.eq.flags &= ~EQ_HW_ALLOCATED;
sys/dev/cxgbe/t4_main.c
2210
wrq->eq.flags &= ~EQ_HW_ALLOCATED;
sys/dev/cxgbe/t4_main.c
2568
txq->eq.flags |= EQ_ENABLED;
sys/dev/cxgbe/t4_main.c
3392
txq->eq.flags |= EQ_QFLUSH;
sys/dev/cxgbe/t4_main.c
3399
txq->eq.flags &= ~EQ_QFLUSH;
sys/dev/cxgbe/t4_main.c
6803
txq->eq.flags |= EQ_ENABLED;
sys/dev/cxgbe/t4_main.c
6876
txq->eq.flags &= ~EQ_ENABLED;
sys/dev/cxgbe/t4_main.c
7355
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_main.c
7356
struct sge_qstat *spg = (void *)&eq->desc[eq->sidx];
sys/dev/cxgbe/t4_main.c
7358
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_main.c
7359
MPASS(!(eq->flags & EQ_ENABLED));
sys/dev/cxgbe/t4_main.c
7368
if (eq->flags & EQ_HW_ALLOCATED) {
sys/dev/cxgbe/t4_main.c
7374
while (spg->cidx != htobe16(eq->pidx))
sys/dev/cxgbe/t4_main.c
7376
while (eq->cidx != eq->pidx)
sys/dev/cxgbe/t4_main.c
7384
while (eq->cidx != eq->pidx) {
sys/dev/cxgbe/t4_main.c
7388
txsd = &txq->sdesc[eq->cidx];
sys/dev/cxgbe/t4_main.c
7394
IDXINCR(eq->cidx, txsd->desc_used, eq->sidx);
sys/dev/cxgbe/t4_main.c
7396
spg->pidx = spg->cidx = htobe16(eq->cidx);
sys/dev/cxgbe/t4_sched.c
324
if (!(txq->eq.flags & EQ_HW_ALLOCATED))
sys/dev/cxgbe/t4_sched.c
337
tc0 = &sc->port[txq->eq.port_id]->sched_params->cl_rl[0];
sys/dev/cxgbe/t4_sched.c
364
V_FW_PARAMS_PARAM_YZ(txq->eq.cntxt_id));
sys/dev/cxgbe/t4_sge.c
2186
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
2188
EQ_LOCK(eq);
sys/dev/cxgbe/t4_sge.c
2191
EQ_UNLOCK(eq);
sys/dev/cxgbe/t4_sge.c
2197
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
2203
EQ_LOCK_ASSERT_OWNED(eq);
sys/dev/cxgbe/t4_sge.c
2207
MPASS(eq->pidx == eq->dbidx);
sys/dev/cxgbe/t4_sge.c
2211
eq->cidx = read_hw_cidx(eq);
sys/dev/cxgbe/t4_sge.c
2212
if (eq->pidx == eq->cidx)
sys/dev/cxgbe/t4_sge.c
2213
available = eq->sidx - 1;
sys/dev/cxgbe/t4_sge.c
2215
available = IDXDIFF(eq->cidx, eq->pidx, eq->sidx) - 1;
sys/dev/cxgbe/t4_sge.c
2222
dst = (void *)&eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
2223
if (__predict_true(eq->sidx - eq->pidx > n)) {
sys/dev/cxgbe/t4_sge.c
2226
eq->pidx += n;
sys/dev/cxgbe/t4_sge.c
2228
int first_portion = (eq->sidx - eq->pidx) * EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
2232
bcopy(&wr->wr[first_portion], &eq->desc[0],
sys/dev/cxgbe/t4_sge.c
2235
eq->pidx = n - (eq->sidx - eq->pidx);
sys/dev/cxgbe/t4_sge.c
2239
if (available < eq->sidx / 4 &&
sys/dev/cxgbe/t4_sge.c
2240
atomic_cmpset_int(&eq->equiq, 0, 1)) {
sys/dev/cxgbe/t4_sge.c
2252
ring_eq_db(sc, eq, dbdiff);
sys/dev/cxgbe/t4_sge.c
2265
ring_eq_db(sc, eq, dbdiff);
sys/dev/cxgbe/t4_sge.c
2275
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
2278
EQ_LOCK_ASSERT_OWNED(eq);
sys/dev/cxgbe/t4_sge.c
2293
MPASS(eq->pidx == eq->dbidx);
sys/dev/cxgbe/t4_sge.c
2964
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
2974
EQ_LOCK(eq);
sys/dev/cxgbe/t4_sge.c
2975
if (__predict_false((eq->flags & EQ_HW_ALLOCATED) == 0)) {
sys/dev/cxgbe/t4_sge.c
2976
EQ_UNLOCK(eq);
sys/dev/cxgbe/t4_sge.c
2985
EQ_UNLOCK(eq);
sys/dev/cxgbe/t4_sge.c
2994
eq->cidx = read_hw_cidx(eq);
sys/dev/cxgbe/t4_sge.c
2995
if (eq->pidx == eq->cidx)
sys/dev/cxgbe/t4_sge.c
2996
available = eq->sidx - 1;
sys/dev/cxgbe/t4_sge.c
2998
available = IDXDIFF(eq->cidx, eq->pidx, eq->sidx) - 1;
sys/dev/cxgbe/t4_sge.c
3002
cookie->pidx = eq->pidx;
sys/dev/cxgbe/t4_sge.c
3006
w = &eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
3007
IDXINCR(eq->pidx, ndesc, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3008
if (__predict_false(cookie->pidx + ndesc > eq->sidx)) {
sys/dev/cxgbe/t4_sge.c
3014
EQ_UNLOCK(eq);
sys/dev/cxgbe/t4_sge.c
3022
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
3035
int n = (eq->sidx - wrq->ss_pidx) * EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
3038
bcopy(&wrq->ss[0], &eq->desc[wrq->ss_pidx], n);
sys/dev/cxgbe/t4_sge.c
3039
bcopy(&wrq->ss[n], &eq->desc[0], wrq->ss_len - n);
sys/dev/cxgbe/t4_sge.c
3044
EQ_LOCK(eq);
sys/dev/cxgbe/t4_sge.c
3047
MPASS(pidx >= 0 && pidx < eq->sidx);
sys/dev/cxgbe/t4_sge.c
3051
MPASS(pidx == eq->dbidx);
sys/dev/cxgbe/t4_sge.c
3061
dst = (void *)&eq->desc[pidx];
sys/dev/cxgbe/t4_sge.c
3062
available = IDXDIFF(eq->cidx, eq->pidx, eq->sidx) - 1;
sys/dev/cxgbe/t4_sge.c
3063
if (available < eq->sidx / 4 &&
sys/dev/cxgbe/t4_sge.c
3064
atomic_cmpset_int(&eq->equiq, 0, 1)) {
sys/dev/cxgbe/t4_sge.c
3074
if (__predict_true(eq->flags & EQ_HW_ALLOCATED))
sys/dev/cxgbe/t4_sge.c
3075
ring_eq_db(wrq->adapter, eq, ndesc);
sys/dev/cxgbe/t4_sge.c
3077
IDXINCR(eq->dbidx, ndesc, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3079
MPASS(IDXDIFF(next->pidx, pidx, eq->sidx) == ndesc);
sys/dev/cxgbe/t4_sge.c
3084
MPASS(IDXDIFF(pidx, prev->pidx, eq->sidx) == prev->ndesc);
sys/dev/cxgbe/t4_sge.c
3095
MPASS(wrq->eq.pidx == wrq->eq.dbidx);
sys/dev/cxgbe/t4_sge.c
3098
EQ_UNLOCK(eq);
sys/dev/cxgbe/t4_sge.c
3104
struct sge_eq *eq = r->cookie;
sys/dev/cxgbe/t4_sge.c
3106
return (total_available_tx_desc(eq) > eq->sidx / 8);
sys/dev/cxgbe/t4_sge.c
3118
discard_tx(struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
3121
return ((eq->flags & (EQ_ENABLED | EQ_QFLUSH)) != EQ_ENABLED);
sys/dev/cxgbe/t4_sge.c
3146
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
3149
if ((txp->npkt > 0 || avail < eq->sidx / 2) &&
sys/dev/cxgbe/t4_sge.c
3150
atomic_cmpset_int(&eq->equiq, 0, 1)) {
sys/dev/cxgbe/t4_sge.c
3152
eq->equeqidx = eq->pidx;
sys/dev/cxgbe/t4_sge.c
3153
} else if (IDXDIFF(eq->pidx, eq->equeqidx, eq->sidx) >= 32) {
sys/dev/cxgbe/t4_sge.c
3155
eq->equeqidx = eq->pidx;
sys/dev/cxgbe/t4_sge.c
3188
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
3203
if (__predict_false(discard_tx(eq))) {
sys/dev/cxgbe/t4_sge.c
3213
reclaim_tx_descs(txq, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3219
if (eq->pidx == eq->cidx)
sys/dev/cxgbe/t4_sge.c
3220
avail = eq->sidx - 1;
sys/dev/cxgbe/t4_sge.c
3222
avail = IDXDIFF(eq->cidx, eq->pidx, eq->sidx) - 1;
sys/dev/cxgbe/t4_sge.c
3252
atomic_load_int(&txq->eq.equiq) != 0) {
sys/dev/cxgbe/t4_sge.c
3284
wr = &eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
3285
IDXINCR(eq->pidx, n, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3313
wr = &eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
3337
IDXINCR(eq->pidx, n, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3342
ring_eq_db(sc, eq, dbdiff);
sys/dev/cxgbe/t4_sge.c
3355
ring_eq_db(sc, eq, dbdiff);
sys/dev/cxgbe/t4_sge.c
3357
} else if (eq->pidx == eq->cidx && txp->npkt > 0 &&
sys/dev/cxgbe/t4_sge.c
3358
atomic_load_int(&txq->eq.equiq) == 0) {
sys/dev/cxgbe/t4_sge.c
3383
wr = &eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
3384
IDXINCR(eq->pidx, n, eq->sidx);
sys/dev/cxgbe/t4_sge.c
3389
ring_eq_db(sc, eq, n);
sys/dev/cxgbe/t4_sge.c
3454
init_eq(struct adapter *sc, struct sge_eq *eq, int eqtype, int qsize,
sys/dev/cxgbe/t4_sge.c
3460
eq->type = eqtype;
sys/dev/cxgbe/t4_sge.c
3461
eq->port_id = port_id;
sys/dev/cxgbe/t4_sge.c
3462
eq->tx_chan = sc->port[port_id]->tx_chan;
sys/dev/cxgbe/t4_sge.c
3463
eq->hw_port = sc->port[port_id]->hw_port;
sys/dev/cxgbe/t4_sge.c
3464
eq->iq = iq;
sys/dev/cxgbe/t4_sge.c
3465
eq->sidx = qsize - sc->params.sge.spg_len / EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
3466
strlcpy(eq->lockname, name, sizeof(eq->lockname));
sys/dev/cxgbe/t4_sge.c
3467
mtx_init(&eq->eq_lock, eq->lockname, NULL, MTX_DEF);
sys/dev/cxgbe/t4_sge.c
3900
if (!(ctrlq->eq.flags & EQ_SW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
3901
MPASS(!(ctrlq->eq.flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
3910
init_eq(sc, &ctrlq->eq, EQ_CTRL, CTRL_EQ_QSIZE,
sys/dev/cxgbe/t4_sge.c
3918
MPASS(ctrlq->eq.flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
3921
if (!(ctrlq->eq.flags & EQ_HW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
3922
MPASS(ctrlq->eq.flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
3926
rc = alloc_eq_hwq(sc, NULL, &ctrlq->eq, idx);
sys/dev/cxgbe/t4_sge.c
3931
MPASS(ctrlq->eq.flags & EQ_HW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
3945
if (ctrlq->eq.flags & EQ_HW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
3946
MPASS(ctrlq->eq.flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
3947
free_eq_hwq(sc, NULL, &ctrlq->eq);
sys/dev/cxgbe/t4_sge.c
3948
MPASS(!(ctrlq->eq.flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
3951
if (ctrlq->eq.flags & EQ_SW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
3952
MPASS(!(ctrlq->eq.flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
3954
MPASS(!(ctrlq->eq.flags & EQ_SW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4373
ctrl_eq_alloc(struct adapter *sc, struct sge_eq *eq, int idx)
sys/dev/cxgbe/t4_sge.c
4377
int qsize = eq->sidx + sc->params.sge.spg_len / EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
4387
c.cmpliqid_eqid = htonl(V_FW_EQ_CTRL_CMD_CMPLIQID(eq->iqid));
sys/dev/cxgbe/t4_sge.c
4391
V_FW_EQ_CTRL_CMD_PCIECHN(eq->hw_port) |
sys/dev/cxgbe/t4_sge.c
4392
F_FW_EQ_CTRL_CMD_FETCHRO | V_FW_EQ_CTRL_CMD_IQID(eq->iqid));
sys/dev/cxgbe/t4_sge.c
4399
c.eqaddr = htobe64(eq->ba);
sys/dev/cxgbe/t4_sge.c
4404
eq->port_id, rc);
sys/dev/cxgbe/t4_sge.c
4408
eq->cntxt_id = G_FW_EQ_CTRL_CMD_EQID(be32toh(c.cmpliqid_eqid));
sys/dev/cxgbe/t4_sge.c
4409
eq->abs_id = G_FW_EQ_CTRL_CMD_PHYSEQID(be32toh(c.physeqid_pkd));
sys/dev/cxgbe/t4_sge.c
4410
cntxt_id = eq->cntxt_id - sc->sge.eq_start;
sys/dev/cxgbe/t4_sge.c
4414
sc->sge.eqmap[cntxt_id] = eq;
sys/dev/cxgbe/t4_sge.c
4420
eth_eq_alloc(struct adapter *sc, struct vi_info *vi, struct sge_eq *eq, int idx)
sys/dev/cxgbe/t4_sge.c
4424
int qsize = eq->sidx + sc->params.sge.spg_len / EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
4438
V_FW_EQ_ETH_CMD_PCIECHN(eq->hw_port) | F_FW_EQ_ETH_CMD_FETCHRO |
sys/dev/cxgbe/t4_sge.c
4439
V_FW_EQ_ETH_CMD_IQID(eq->iqid));
sys/dev/cxgbe/t4_sge.c
4445
c.eqaddr = htobe64(eq->ba);
sys/dev/cxgbe/t4_sge.c
4454
eq->cntxt_id = G_FW_EQ_ETH_CMD_EQID(be32toh(c.eqid_pkd));
sys/dev/cxgbe/t4_sge.c
4455
eq->abs_id = G_FW_EQ_ETH_CMD_PHYSEQID(be32toh(c.physeqid_pkd));
sys/dev/cxgbe/t4_sge.c
4456
cntxt_id = eq->cntxt_id - sc->sge.eq_start;
sys/dev/cxgbe/t4_sge.c
4460
sc->sge.eqmap[cntxt_id] = eq;
sys/dev/cxgbe/t4_sge.c
4481
ofld_eq_alloc(struct adapter *sc, struct vi_info *vi, struct sge_eq *eq,
sys/dev/cxgbe/t4_sge.c
4486
int qsize = eq->sidx + sc->params.sge.spg_len / EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
4503
V_FW_EQ_OFLD_CMD_PCIECHN(eq->hw_port) |
sys/dev/cxgbe/t4_sge.c
4504
F_FW_EQ_OFLD_CMD_FETCHRO | V_FW_EQ_OFLD_CMD_IQID(eq->iqid));
sys/dev/cxgbe/t4_sge.c
4511
c.eqaddr = htobe64(eq->ba);
sys/dev/cxgbe/t4_sge.c
4520
eq->cntxt_id = G_FW_EQ_OFLD_CMD_EQID(be32toh(c.eqid_pkd));
sys/dev/cxgbe/t4_sge.c
4521
eq->abs_id = G_FW_EQ_OFLD_CMD_PHYSEQID(be32toh(c.physeqid_pkd));
sys/dev/cxgbe/t4_sge.c
4522
cntxt_id = eq->cntxt_id - sc->sge.eq_start;
sys/dev/cxgbe/t4_sge.c
4526
sc->sge.eqmap[cntxt_id] = eq;
sys/dev/cxgbe/t4_sge.c
4534
alloc_eq(struct adapter *sc, struct sge_eq *eq, struct sysctl_ctx_list *ctx,
sys/dev/cxgbe/t4_sge.c
4540
MPASS(!(eq->flags & EQ_SW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4542
qsize = eq->sidx + sc->params.sge.spg_len / EQ_ESIZE;
sys/dev/cxgbe/t4_sge.c
4544
rc = alloc_ring(sc, len, &eq->desc_tag, &eq->desc_map, &eq->ba,
sys/dev/cxgbe/t4_sge.c
4545
(void **)&eq->desc);
sys/dev/cxgbe/t4_sge.c
4549
add_eq_sysctls(sc, ctx, oid, eq);
sys/dev/cxgbe/t4_sge.c
4550
eq->flags |= EQ_SW_ALLOCATED;
sys/dev/cxgbe/t4_sge.c
4557
free_eq(struct adapter *sc, struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
4559
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4560
if (eq->type == EQ_ETH)
sys/dev/cxgbe/t4_sge.c
4561
MPASS(eq->pidx == eq->cidx);
sys/dev/cxgbe/t4_sge.c
4563
free_ring(sc, eq->desc_tag, eq->desc_map, eq->ba, eq->desc);
sys/dev/cxgbe/t4_sge.c
4564
mtx_destroy(&eq->eq_lock);
sys/dev/cxgbe/t4_sge.c
4565
bzero(eq, sizeof(*eq));
sys/dev/cxgbe/t4_sge.c
4570
struct sysctl_oid *oid, struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
4574
SYSCTL_ADD_UAUTO(ctx, children, OID_AUTO, "ba", CTLFLAG_RD, &eq->ba,
sys/dev/cxgbe/t4_sge.c
4577
eq->sidx * EQ_ESIZE + sc->params.sge.spg_len,
sys/dev/cxgbe/t4_sge.c
4580
&eq->abs_id, 0, "absolute id of the queue");
sys/dev/cxgbe/t4_sge.c
4582
&eq->cntxt_id, 0, "SGE context id of the queue");
sys/dev/cxgbe/t4_sge.c
4583
SYSCTL_ADD_U16(ctx, children, OID_AUTO, "cidx", CTLFLAG_RD, &eq->cidx,
sys/dev/cxgbe/t4_sge.c
4585
SYSCTL_ADD_U16(ctx, children, OID_AUTO, "pidx", CTLFLAG_RD, &eq->pidx,
sys/dev/cxgbe/t4_sge.c
4588
eq->sidx, "status page index");
sys/dev/cxgbe/t4_sge.c
4592
alloc_eq_hwq(struct adapter *sc, struct vi_info *vi, struct sge_eq *eq, int idx)
sys/dev/cxgbe/t4_sge.c
4596
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4598
eq->iqid = eq->iq->cntxt_id;
sys/dev/cxgbe/t4_sge.c
4599
eq->pidx = eq->cidx = eq->dbidx = 0;
sys/dev/cxgbe/t4_sge.c
4601
eq->equeqidx = 0;
sys/dev/cxgbe/t4_sge.c
4602
eq->doorbells = sc->doorbells;
sys/dev/cxgbe/t4_sge.c
4603
bzero(eq->desc, eq->sidx * EQ_ESIZE + sc->params.sge.spg_len);
sys/dev/cxgbe/t4_sge.c
4605
switch (eq->type) {
sys/dev/cxgbe/t4_sge.c
4607
rc = ctrl_eq_alloc(sc, eq, idx);
sys/dev/cxgbe/t4_sge.c
4611
rc = eth_eq_alloc(sc, vi, eq, idx);
sys/dev/cxgbe/t4_sge.c
4616
rc = ofld_eq_alloc(sc, vi, eq, idx);
sys/dev/cxgbe/t4_sge.c
4621
panic("%s: invalid eq type %d.", __func__, eq->type);
sys/dev/cxgbe/t4_sge.c
4625
eq->type, rc);
sys/dev/cxgbe/t4_sge.c
4629
if (isset(&eq->doorbells, DOORBELL_UDB) ||
sys/dev/cxgbe/t4_sge.c
4630
isset(&eq->doorbells, DOORBELL_UDBWC) ||
sys/dev/cxgbe/t4_sge.c
4631
isset(&eq->doorbells, DOORBELL_WCWR)) {
sys/dev/cxgbe/t4_sge.c
4637
udb += (eq->cntxt_id >> s_qpp) << PAGE_SHIFT; /* pg offset */
sys/dev/cxgbe/t4_sge.c
4638
eq->udb_qid = eq->cntxt_id & mask; /* id in page */
sys/dev/cxgbe/t4_sge.c
4639
if (eq->udb_qid >= PAGE_SIZE / UDBS_SEG_SIZE)
sys/dev/cxgbe/t4_sge.c
4640
clrbit(&eq->doorbells, DOORBELL_WCWR);
sys/dev/cxgbe/t4_sge.c
4642
udb += eq->udb_qid << UDBS_SEG_SHIFT; /* seg offset */
sys/dev/cxgbe/t4_sge.c
4643
eq->udb_qid = 0;
sys/dev/cxgbe/t4_sge.c
4645
eq->udb = (volatile void *)udb;
sys/dev/cxgbe/t4_sge.c
4648
eq->flags |= EQ_HW_ALLOCATED;
sys/dev/cxgbe/t4_sge.c
4653
free_eq_hwq(struct adapter *sc, struct vi_info *vi __unused, struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
4657
MPASS(eq->flags & EQ_HW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4659
switch (eq->type) {
sys/dev/cxgbe/t4_sge.c
4661
rc = -t4_ctrl_eq_free(sc, sc->mbox, sc->pf, 0, eq->cntxt_id);
sys/dev/cxgbe/t4_sge.c
4664
rc = -t4_eth_eq_free(sc, sc->mbox, sc->pf, 0, eq->cntxt_id);
sys/dev/cxgbe/t4_sge.c
4668
rc = -t4_ofld_eq_free(sc, sc->mbox, sc->pf, 0, eq->cntxt_id);
sys/dev/cxgbe/t4_sge.c
4672
panic("%s: invalid eq type %d.", __func__, eq->type);
sys/dev/cxgbe/t4_sge.c
4675
CH_ERR(sc, "failed to free eq (type %d): %d\n", eq->type, rc);
sys/dev/cxgbe/t4_sge.c
4678
eq->flags &= ~EQ_HW_ALLOCATED;
sys/dev/cxgbe/t4_sge.c
4687
struct sge_eq *eq = &wrq->eq;
sys/dev/cxgbe/t4_sge.c
4690
MPASS(!(eq->flags & EQ_SW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4692
rc = alloc_eq(sc, eq, ctx, oid);
sys/dev/cxgbe/t4_sge.c
4695
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4712
free_eq(sc, &wrq->eq);
sys/dev/cxgbe/t4_sge.c
4747
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
4752
if (!(eq->flags & EQ_SW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
4753
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4763
init_eq(sc, &txq->eq, EQ_ETH, vi->qsize_txq, pi->port_id,
sys/dev/cxgbe/t4_sge.c
4766
rc = mp_ring_alloc(&txq->r, eq->sidx, txq, eth_tx,
sys/dev/cxgbe/t4_sge.c
4767
can_resume_eth_tx, M_CXGBE, &eq->eq_lock, M_WAITOK);
sys/dev/cxgbe/t4_sge.c
4776
rc = alloc_eq(sc, eq, &vi->ctx, oid);
sys/dev/cxgbe/t4_sge.c
4782
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4785
TASK_INIT(&txq->tx_reclaim_task, 0, tx_reclaim, eq);
sys/dev/cxgbe/t4_sge.c
4788
txq->sdesc = malloc(eq->sidx * sizeof(struct tx_sdesc), M_CXGBE,
sys/dev/cxgbe/t4_sge.c
4794
if (!(eq->flags & EQ_HW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
4795
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4796
rc = alloc_eq_hwq(sc, vi, eq, idx);
sys/dev/cxgbe/t4_sge.c
4801
MPASS(eq->flags & EQ_HW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4805
sc->sge.eq_base = eq->abs_id - eq->cntxt_id;
sys/dev/cxgbe/t4_sge.c
4807
KASSERT(eq->cntxt_id + sc->sge.eq_base == eq->abs_id,
sys/dev/cxgbe/t4_sge.c
4840
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
4842
if (eq->flags & EQ_HW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
4843
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
4844
free_eq_hwq(sc, NULL, eq);
sys/dev/cxgbe/t4_sge.c
4845
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4848
if (eq->flags & EQ_SW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
4849
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4853
free_eq(sc, eq);
sys/dev/cxgbe/t4_sge.c
4854
MPASS(!(eq->flags & EQ_SW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
4991
struct sge_eq *eq = &ofld_txq->wrq.eq;
sys/dev/cxgbe/t4_sge.c
4998
if (!(eq->flags & EQ_SW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
5008
init_eq(sc, eq, EQ_OFLD, vi->qsize_txq, pi->port_id,
sys/dev/cxgbe/t4_sge.c
5012
init_eq(sc, eq, EQ_OFLD, vi->qsize_txq, pi->port_id,
sys/dev/cxgbe/t4_sge.c
5023
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
5039
if (!(eq->flags & EQ_HW_ALLOCATED)) {
sys/dev/cxgbe/t4_sge.c
5040
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
5043
rc = alloc_eq_hwq(sc, vi, eq, idx);
sys/dev/cxgbe/t4_sge.c
5049
MPASS(eq->flags & EQ_HW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
5062
struct sge_eq *eq = &ofld_txq->wrq.eq;
sys/dev/cxgbe/t4_sge.c
5064
if (eq->flags & EQ_HW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
5065
MPASS(eq->flags & EQ_SW_ALLOCATED);
sys/dev/cxgbe/t4_sge.c
5066
free_eq_hwq(sc, NULL, eq);
sys/dev/cxgbe/t4_sge.c
5067
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
5070
if (eq->flags & EQ_SW_ALLOCATED) {
sys/dev/cxgbe/t4_sge.c
5071
MPASS(!(eq->flags & EQ_HW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
5083
MPASS(!(eq->flags & EQ_SW_ALLOCATED));
sys/dev/cxgbe/t4_sge.c
5637
struct sge_eq *eq;
sys/dev/cxgbe/t4_sge.c
5657
eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
5658
wr = (void *)&eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
5711
if (dst == (void *)&eq->desc[eq->sidx]) {
sys/dev/cxgbe/t4_sge.c
5712
dst = (void *)&eq->desc[0];
sys/dev/cxgbe/t4_sge.c
5715
write_gl_to_txd(txq, m0, &dst, eq->sidx - ndesc < eq->pidx);
sys/dev/cxgbe/t4_sge.c
5719
txsd = &txq->sdesc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
5736
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
5748
copy_to_txd(eq, mtod(m, caddr_t), &dst, m->m_len);
sys/dev/cxgbe/t4_sge.c
5752
txsd = &txq->sdesc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
5770
struct sge_eq *eq;
sys/dev/cxgbe/t4_sge.c
5803
eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
5804
wr = (void *)&eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
5848
if (__predict_false((uintptr_t)dst == (uintptr_t)&eq->desc[eq->sidx]))
sys/dev/cxgbe/t4_sge.c
5849
dst = (caddr_t)&eq->desc[0];
sys/dev/cxgbe/t4_sge.c
5852
write_gl_to_txd(txq, m0, &dst, eq->sidx - ndesc < eq->pidx);
sys/dev/cxgbe/t4_sge.c
5858
copy_to_txd(eq, mtod(m, caddr_t), &dst, m->m_len);
sys/dev/cxgbe/t4_sge.c
5871
txsd = &txq->sdesc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
6035
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
6048
wr = (void *)&eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
6064
checkwrap = eq->sidx - ndesc < eq->pidx;
sys/dev/cxgbe/t4_sge.c
6074
V_ULP_TXPKT_DEST(0) | V_ULP_TXPKT_FID(eq->iqid));
sys/dev/cxgbe/t4_sge.c
6085
(uintptr_t)cpl == (uintptr_t)&eq->desc[eq->sidx])
sys/dev/cxgbe/t4_sge.c
6086
cpl = (void *)&eq->desc[0];
sys/dev/cxgbe/t4_sge.c
6116
(uintptr_t)flitp == (uintptr_t)&eq->desc[eq->sidx])
sys/dev/cxgbe/t4_sge.c
6117
flitp = (void *)&eq->desc[0];
sys/dev/cxgbe/t4_sge.c
6135
txsd = &txq->sdesc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
6146
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
6161
wr = (void *)&eq->desc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
6180
if (i & 1 && (uintptr_t)flitp == (uintptr_t)&eq->desc[eq->sidx])
sys/dev/cxgbe/t4_sge.c
6181
flitp = &eq->desc[0];
sys/dev/cxgbe/t4_sge.c
6216
txsd = &txq->sdesc[eq->pidx];
sys/dev/cxgbe/t4_sge.c
6230
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
6239
MPASS((uintptr_t)(*to) >= (uintptr_t)&eq->desc[0]);
sys/dev/cxgbe/t4_sge.c
6240
MPASS((uintptr_t)(*to) < (uintptr_t)&eq->desc[eq->sidx]);
sys/dev/cxgbe/t4_sge.c
6248
wrap = (__be64 *)(&eq->desc[eq->sidx]);
sys/dev/cxgbe/t4_sge.c
6283
flitp = (void *)eq->desc;
sys/dev/cxgbe/t4_sge.c
6295
*to = (void *)eq->desc;
sys/dev/cxgbe/t4_sge.c
6301
copy_to_txd(struct sge_eq *eq, caddr_t from, caddr_t *to, int len)
sys/dev/cxgbe/t4_sge.c
6304
MPASS((uintptr_t)(*to) >= (uintptr_t)&eq->desc[0]);
sys/dev/cxgbe/t4_sge.c
6305
MPASS((uintptr_t)(*to) < (uintptr_t)&eq->desc[eq->sidx]);
sys/dev/cxgbe/t4_sge.c
6308
(uintptr_t)&eq->desc[eq->sidx])) {
sys/dev/cxgbe/t4_sge.c
6312
int portion = (uintptr_t)&eq->desc[eq->sidx] - (uintptr_t)(*to);
sys/dev/cxgbe/t4_sge.c
6317
bcopy(from, (void *)eq->desc, portion);
sys/dev/cxgbe/t4_sge.c
6318
(*to) = (caddr_t)eq->desc + portion;
sys/dev/cxgbe/t4_sge.c
6323
ring_eq_db(struct adapter *sc, struct sge_eq *eq, u_int n)
sys/dev/cxgbe/t4_sge.c
6329
db = eq->doorbells;
sys/dev/cxgbe/t4_sge.c
6336
*eq->udb = htole32(V_QID(eq->udb_qid) | V_PIDX(n));
sys/dev/cxgbe/t4_sge.c
6348
KASSERT(eq->udb_qid == 0 && n == 1,
sys/dev/cxgbe/t4_sge.c
6350
__func__, eq->doorbells, n, eq->dbidx, eq));
sys/dev/cxgbe/t4_sge.c
6352
dst = (volatile void *)((uintptr_t)eq->udb + UDBS_WR_OFFSET -
sys/dev/cxgbe/t4_sge.c
6354
i = eq->dbidx;
sys/dev/cxgbe/t4_sge.c
6355
src = (void *)&eq->desc[i];
sys/dev/cxgbe/t4_sge.c
6356
while (src != (void *)&eq->desc[i + 1])
sys/dev/cxgbe/t4_sge.c
6363
*eq->udb = htole32(V_QID(eq->udb_qid) | V_PIDX(n));
sys/dev/cxgbe/t4_sge.c
6369
V_QID(eq->cntxt_id) | V_PIDX(n));
sys/dev/cxgbe/t4_sge.c
6373
IDXINCR(eq->dbidx, n, eq->sidx);
sys/dev/cxgbe/t4_sge.c
6377
reclaimable_tx_desc(struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
6381
hw_cidx = read_hw_cidx(eq);
sys/dev/cxgbe/t4_sge.c
6382
return (IDXDIFF(hw_cidx, eq->cidx, eq->sidx));
sys/dev/cxgbe/t4_sge.c
6386
total_available_tx_desc(struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
6390
hw_cidx = read_hw_cidx(eq);
sys/dev/cxgbe/t4_sge.c
6391
pidx = eq->pidx;
sys/dev/cxgbe/t4_sge.c
6394
return (eq->sidx - 1);
sys/dev/cxgbe/t4_sge.c
6396
return (IDXDIFF(hw_cidx, pidx, eq->sidx) - 1);
sys/dev/cxgbe/t4_sge.c
6400
read_hw_cidx(struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
6402
struct sge_qstat *spg = (void *)&eq->desc[eq->sidx];
sys/dev/cxgbe/t4_sge.c
6415
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
6422
can_reclaim = reclaimable_tx_desc(eq);
sys/dev/cxgbe/t4_sge.c
6427
txsd = &txq->sdesc[eq->cidx];
sys/dev/cxgbe/t4_sge.c
6436
__func__, eq->cidx));
sys/dev/cxgbe/t4_sge.c
6445
IDXINCR(eq->cidx, ndesc, eq->sidx);
sys/dev/cxgbe/t4_sge.c
6455
struct sge_eq *eq = &txq->eq;
sys/dev/cxgbe/t4_sge.c
6461
if (eq->cidx == eq->pidx)
sys/dev/cxgbe/t4_sge.c
6462
eq->equeqidx = eq->pidx;
sys/dev/cxgbe/t4_sge.c
6541
handle_wrq_egr_update(struct adapter *sc, struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
6543
struct sge_wrq *wrq = (void *)eq;
sys/dev/cxgbe/t4_sge.c
6545
atomic_readandclear_int(&eq->equiq);
sys/dev/cxgbe/t4_sge.c
6546
taskqueue_enqueue(sc->tq[eq->port_id], &wrq->wrq_tx_task);
sys/dev/cxgbe/t4_sge.c
6550
handle_eth_egr_update(struct adapter *sc, struct sge_eq *eq)
sys/dev/cxgbe/t4_sge.c
6552
struct sge_txq *txq = (void *)eq;
sys/dev/cxgbe/t4_sge.c
6554
MPASS(eq->type == EQ_ETH);
sys/dev/cxgbe/t4_sge.c
6556
atomic_readandclear_int(&eq->equiq);
sys/dev/cxgbe/t4_sge.c
6558
taskqueue_enqueue(sc->tq[eq->port_id], &txq->tx_reclaim_task);
sys/dev/cxgbe/t4_sge.c
6571
struct sge_eq *eq;
sys/dev/cxgbe/t4_sge.c
6579
eq = s->eqmap[qid - s->eq_start - s->eq_base];
sys/dev/cxgbe/t4_sge.c
6580
(*h[eq->type])(sc, eq);
sys/dev/cxgbe/tom/t4_cpl_io.c
2328
if (wrq->eq.type == EQ_OFLD) {
sys/dev/cxgbe/tom/t4_listen.c
583
req->opt0 = htobe64(V_TX_CHAN(lctx->ctrlq->eq.tx_chan));
sys/dev/cxgbe/tom/t4_listen.c
613
req->opt0 = htobe64(V_TX_CHAN(lctx->ctrlq->eq.tx_chan));
sys/dev/cxgbe/tom/t4_tom.c
1459
if ((ofld_txq->wrq.eq.cntxt_id & mask) == (tid & mask))
sys/dev/cxgbe/tom/t4_tom.c
1470
MPASS((ofld_txq->wrq.eq.cntxt_id & mask) == (tid & mask));
sys/dev/mana/gdma.h
154
} eq;
sys/dev/mana/gdma.h
352
} eq;
sys/dev/mana/gdma.h
376
} eq;
sys/dev/mana/gdma.h
389
struct gdma_queue *eq;
sys/dev/mana/gdma.h
460
int mana_gd_test_eq(struct gdma_context *gc, struct gdma_queue *eq);
sys/dev/mana/gdma_main.c
1087
mana_gd_destroy_eq(gc, queue->eq.disable_needed, queue);
sys/dev/mana/gdma_main.c
359
req.log2_throttle_limit = queue->eq.log2_throttle_limit;
sys/dev/mana/gdma_main.c
360
req.eq_pci_msix_index = queue->eq.msix_index;
sys/dev/mana/gdma_main.c
371
queue->eq.disable_needed = true;
sys/dev/mana/gdma_main.c
423
e.eq.id = qid;
sys/dev/mana/gdma_main.c
424
e.eq.tail_ptr = tail_ptr;
sys/dev/mana/gdma_main.c
425
e.eq.arm = num_req;
sys/dev/mana/gdma_main.c
491
mana_gd_process_eqe(struct gdma_queue *eq)
sys/dev/mana/gdma_main.c
493
uint32_t head = eq->head % (eq->queue_size / GDMA_EQE_SIZE);
sys/dev/mana/gdma_main.c
494
struct gdma_context *gc = eq->gdma_dev->gdma_context;
sys/dev/mana/gdma_main.c
495
struct gdma_eqe *eq_eqe_ptr = eq->queue_mem_ptr;
sys/dev/mana/gdma_main.c
530
gc->test_event_eq_id = eq->id;
sys/dev/mana/gdma_main.c
533
"EQE TEST EVENT received for EQ %u\n", eq->id);
sys/dev/mana/gdma_main.c
541
if (!eq->eq.callback)
sys/dev/mana/gdma_main.c
546
eq->eq.callback(eq->eq.context, eq, &event);
sys/dev/mana/gdma_main.c
560
struct gdma_queue *eq = arg;
sys/dev/mana/gdma_main.c
566
gc = eq->gdma_dev->gdma_context;
sys/dev/mana/gdma_main.c
568
num_eqe = eq->queue_size / GDMA_EQE_SIZE;
sys/dev/mana/gdma_main.c
569
eq_eqe_ptr = eq->queue_mem_ptr;
sys/dev/mana/gdma_main.c
571
bus_dmamap_sync(eq->mem_info.dma_tag, eq->mem_info.dma_map,
sys/dev/mana/gdma_main.c
576
eqe = &eq_eqe_ptr[eq->head % num_eqe];
sys/dev/mana/gdma_main.c
580
old_bits = (eq->head / num_eqe - 1) & GDMA_EQE_OWNER_MASK;
sys/dev/mana/gdma_main.c
586
new_bits = (eq->head / num_eqe) & GDMA_EQE_OWNER_MASK;
sys/dev/mana/gdma_main.c
596
eq->id, i, eq->head,
sys/dev/mana/gdma_main.c
614
mana_gd_process_eqe(eq);
sys/dev/mana/gdma_main.c
616
eq->head++;
sys/dev/mana/gdma_main.c
619
bus_dmamap_sync(eq->mem_info.dma_tag, eq->mem_info.dma_map,
sys/dev/mana/gdma_main.c
622
head = eq->head % (num_eqe << GDMA_EQE_OWNER_BITS);
sys/dev/mana/gdma_main.c
624
mana_gd_ring_doorbell(gc, eq->gdma_dev->doorbell, eq->type, eq->id,
sys/dev/mana/gdma_main.c
649
queue->eq.msix_index = msi_index;
sys/dev/mana/gdma_main.c
696
msix_index = queue->eq.msix_index;
sys/dev/mana/gdma_main.c
708
queue->eq.msix_index = INVALID_PCI_MSIX_INDEX;
sys/dev/mana/gdma_main.c
715
mana_gd_test_eq(struct gdma_context *gc, struct gdma_queue *eq)
sys/dev/mana/gdma_main.c
730
req.hdr.dev_id = eq->gdma_dev->dev_id;
sys/dev/mana/gdma_main.c
731
req.queue_index = eq->id;
sys/dev/mana/gdma_main.c
750
eq->id);
sys/dev/mana/gdma_main.c
754
if (eq->id != gc->test_event_eq_id) {
sys/dev/mana/gdma_main.c
757
gc->test_event_eq_id, eq->id);
sys/dev/mana/gdma_main.c
782
if (queue->eq.disable_needed)
sys/dev/mana/gdma_main.c
795
queue->eq.msix_index = INVALID_PCI_MSIX_INDEX;
sys/dev/mana/gdma_main.c
799
if (spec->eq.log2_throttle_limit > log2_num_entries) {
sys/dev/mana/gdma_main.c
802
spec->eq.log2_throttle_limit, log2_num_entries);
sys/dev/mana/gdma_main.c
812
queue->eq.callback = spec->eq.callback;
sys/dev/mana/gdma_main.c
813
queue->eq.context = spec->eq.context;
sys/dev/mana/gdma_main.c
815
queue->eq.log2_throttle_limit = spec->eq.log2_throttle_limit ?: 1;
sys/dev/mana/hw_channel.c
344
spec.eq.context = ctx;
sys/dev/mana/hw_channel.c
345
spec.eq.callback = cb;
sys/dev/mana/hw_channel.c
346
spec.eq.log2_throttle_limit = DEFAULT_LOG2_THROTTLING_FOR_ERROR_EQ;
sys/dev/mana/hw_channel.c
404
struct gdma_queue *eq, *cq;
sys/dev/mana/hw_channel.c
420
err = mana_hwc_create_gdma_eq(hwc, eq_size, ctx, callback, &eq);
sys/dev/mana/hw_channel.c
426
hwc_cq->gdma_eq = eq;
sys/dev/mana/hw_channel.c
429
mana_hwc_comp_event, eq, &cq);
sys/dev/mana/hw_channel.c
679
struct gdma_queue *eq = hwc->cq->gdma_eq;
sys/dev/mana/hw_channel.c
686
eq->mem_info.dma_handle,
sys/dev/mana/hw_channel.c
690
eq->eq.msix_index);
sys/dev/mana/mana_en.c
1298
struct gdma_queue *eq;
sys/dev/mana/mana_en.c
1305
eq = ac->eqs[i].eq;
sys/dev/mana/mana_en.c
1306
if (!eq)
sys/dev/mana/mana_en.c
1309
mana_gd_destroy_queue(gc, eq);
sys/dev/mana/mana_en.c
1331
spec.eq.callback = NULL;
sys/dev/mana/mana_en.c
1332
spec.eq.context = ac->eqs;
sys/dev/mana/mana_en.c
1333
spec.eq.log2_throttle_limit = LOG2_EQ_THROTTLE;
sys/dev/mana/mana_en.c
1336
err = mana_gd_create_mana_eq(gd, &spec, &ac->eqs[i].eq);
sys/dev/mana/mana_en.c
2137
spec.cq.parent_eq = ac->eqs[i].eq;
sys/dev/mana/mana_en.c
2390
struct mana_eq *eq, if_t ndev)
sys/dev/mana/mana_en.c
2474
spec.cq.parent_eq = eq->eq;
sys/dev/mlx4/mlx4_core/mlx4.h
483
struct mlx4_slave_eqe eq[MLX4_MFUNC_MAX_EQES];
sys/dev/mlx4/mlx4_core/mlx4.h
684
struct mlx4_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_cq.c
278
cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn;
sys/dev/mlx4/mlx4_core/mlx4_cq.c
297
cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq;
sys/dev/mlx4/mlx4_core/mlx4_cq.c
322
synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq);
sys/dev/mlx4/mlx4_core/mlx4_cq.c
323
if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq !=
sys/dev/mlx4/mlx4_core/mlx4_cq.c
324
priv->eq_table.eq[MLX4_EQ_ASYNC].irq)
sys/dev/mlx4/mlx4_core/mlx4_cq.c
325
synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1000
eq->page_list[i].map = t;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1002
memset(eq->page_list[i].buf, 0, PAGE_SIZE);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1005
eq->eqn = mlx4_bitmap_alloc(&priv->eq_table.bitmap);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1006
if (eq->eqn == -1)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1009
eq->doorbell = mlx4_get_eq_uar(dev, eq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
101
eq->doorbell);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1010
if (!eq->doorbell) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1015
err = mlx4_mtt_init(dev, npages, PAGE_SHIFT, &eq->mtt);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1019
err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1025
eq_context->log_eq_size = ilog2(eq->nent);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1029
mtt_addr = mlx4_mtt_addr(dev, &eq->mtt);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1033
err = mlx4_SW2HW_EQ(dev, mailbox, eq->eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1042
eq->cons_index = 0;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1047
mlx4_mtt_cleanup(dev, &eq->mtt);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1050
mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1054
if (eq->page_list[i].buf)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1056
eq->page_list[i].buf,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1057
eq->page_list[i].map);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
106
static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1062
kfree(eq->page_list);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1070
struct mlx4_eq *eq)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1078
int npages = PAGE_ALIGN(dev->caps.eqe_size * eq->nent) / PAGE_SIZE;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1080
err = mlx4_HW2SW_EQ(dev, eq->eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1084
synchronize_irq(eq->irq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1086
mlx4_mtt_cleanup(dev, &eq->mtt);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1089
eq->page_list[i].buf,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1090
eq->page_list[i].map);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1092
kfree(eq->page_list);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1093
mlx4_bitmap_free(&priv->eq_table.bitmap, eq->eqn, MLX4_USE_RR);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
110
unsigned long offset = (entry & (eq->nent - 1)) * eqe_size;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1105
if (eq_table->eq[i].have_irq) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1106
eq_table->eq[i].affinity_cpu_id = NOCPU;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1107
free_irq(eq_table->eq[i].irq, eq_table->eq + i);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1108
eq_table->eq[i].have_irq = 0;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1140
priv->eq_table.eq = kcalloc(dev->caps.num_eqs - dev->caps.reserved_eqs,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1141
sizeof *priv->eq_table.eq, GFP_KERNEL);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1142
if (!priv->eq_table.eq)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1150
kfree(mlx4_priv(dev)->eq_table.eq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
118
return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1202
0, &priv->eq_table.eq[MLX4_EQ_ASYNC]);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1204
struct mlx4_eq *eq = &priv->eq_table.eq[i];
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1206
int port = find_first_bit(eq->actv_ports.ports,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
121
static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1224
info->rmap, eq->irq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
123
struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1233
eq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
124
return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1250
err = request_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1252
priv->eq_table.eq + MLX4_EQ_ASYNC);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1256
priv->eq_table.eq[MLX4_EQ_ASYNC].have_irq = 1;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1271
priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1274
priv->eq_table.eq[MLX4_EQ_ASYNC].eqn, err);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1277
eq_set_ci(&priv->eq_table.eq[MLX4_EQ_ASYNC], 1);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1283
mlx4_free_eq(dev, &priv->eq_table.eq[--i]);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1314
priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1327
mlx4_free_eq(dev, &priv->eq_table.eq[i]);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1361
priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1375
priv->eq_table.eq[MLX4_EQ_ASYNC].eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1391
return test_bit(port - 1, priv->eq_table.eq[vector].actv_ports.ports);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1403
priv->eq_table.eq[i].actv_ports.ports);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1417
return !!(bitmap_weight(priv->eq_table.eq[vector].actv_ports.ports,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1436
priv->eq_table.eq[requested_vector].actv_ports.ports)) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1439
struct mlx4_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1445
eq = &priv->eq_table.eq[requested_vector];
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1447
test_bit(port - 1, eq->actv_ports.ports)) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1457
struct mlx4_eq *eq = &priv->eq_table.eq[i];
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1459
if (min_ref_count_val > eq->ref_count &&
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1460
test_bit(port - 1, eq->actv_ports.ports)) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1461
min_ref_count_val = eq->ref_count;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1482
err = request_irq(priv->eq_table.eq[*prequested_vector].irq,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1485
priv->eq_table.eq + *prequested_vector);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1492
eq_set_ci(&priv->eq_table.eq[*prequested_vector], 1);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1493
priv->eq_table.eq[*prequested_vector].have_irq = 1;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1498
priv->eq_table.eq[*prequested_vector].ref_count++;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1516
return priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq_vec)].irq;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1526
priv->eq_table.eq[eq_vec].ref_count--;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1543
disable_irq(priv->eq_table.eq[i].irq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1558
mlx4_msi_x_interrupt(priv->eq_table.eq[i].irq,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
1559
priv->eq_table.eq + i);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
240
struct mlx4_eq *eq = &priv->eq_table.eq[vec];
sys/dev/mlx4/mlx4_core/mlx4_eq.c
242
hint_err = bind_irq_to_cpu(eq->irq, eq->affinity_cpu_id);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
489
static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
506
while ((eqe = next_eqe_sw(eq, dev->caps.eqe_factor, eqe_size))) {
sys/dev/mlx4/mlx4_core/mlx4_eq.c
537
eq->eqn, eq->cons_index, ret);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
565
eq->eqn, eq->cons_index, ret);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
683
eq->eqn, eq->cons_index, ret);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
699
mlx4_warn(dev, "EQ overrun on EQN %d\n", eq->eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
767
eqe->type, eqe->subtype, eq->eqn,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
768
eq->cons_index, eqe->owner, eq->nent,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
771
!!(eq->cons_index & eq->nent) ? "HW" : "SW");
sys/dev/mlx4/mlx4_core/mlx4_eq.c
792
eqe->type, eqe->subtype, eq->eqn,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
793
eq->cons_index, eqe->owner, eq->nent,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
795
!!(eq->cons_index & eq->nent) ? "HW" : "SW");
sys/dev/mlx4/mlx4_core/mlx4_eq.c
804
eqe->type, eqe->subtype, eq->eqn,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
805
eq->cons_index, eqe->owner, eq->nent,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
808
!!(eq->cons_index & eq->nent) ? "HW" : "SW");
sys/dev/mlx4/mlx4_core/mlx4_eq.c
812
++eq->cons_index;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
824
eq_set_ci(eq, 0);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
829
eq_set_ci(eq, 1);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
844
work |= mlx4_eq_int(dev, &priv->eq_table.eq[i]);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
851
struct mlx4_eq *eq = eq_ptr;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
852
struct mlx4_dev *dev = eq->dev;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
854
mlx4_eq_int(dev, eq);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
920
static void __iomem *mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
925
index = eq->eqn / 4 - dev->caps.reserved_eqs / 4;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
930
((eq->eqn / 4) << (dev->uar_page_shift)),
sys/dev/mlx4/mlx4_core/mlx4_eq.c
934
eq->eqn);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
939
return priv->eq_table.uar_map[index] + 0x800 + 8 * (eq->eqn % 4);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
955
u8 intr, struct mlx4_eq *eq)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
967
eq->dev = dev;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
968
eq->nent = roundup_pow_of_two(max(nent, 2));
sys/dev/mlx4/mlx4_core/mlx4_eq.c
97
static void eq_set_ci(struct mlx4_eq *eq, int req_not)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
972
npages = PAGE_ALIGN(eq->nent * dev->caps.eqe_size) / PAGE_SIZE;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
974
eq->page_list = kmalloc(npages * sizeof *eq->page_list,
sys/dev/mlx4/mlx4_core/mlx4_eq.c
976
if (!eq->page_list)
sys/dev/mlx4/mlx4_core/mlx4_eq.c
980
eq->page_list[i].buf = NULL;
sys/dev/mlx4/mlx4_core/mlx4_eq.c
99
__raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) |
sys/dev/mlx4/mlx4_core/mlx4_eq.c
992
eq->page_list[i].buf = dma_alloc_coherent(&dev->persist->
sys/dev/mlx4/mlx4_core/mlx4_eq.c
996
if (!eq->page_list[i].buf)
sys/dev/mlx4/mlx4_core/mlx4_main.c
2685
priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
sys/dev/mlx4/mlx4_core/mlx4_main.c
2689
priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
sys/dev/mlx4/mlx4_core/mlx4_main.c
2818
struct mlx4_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_main.c
2834
eq = &priv->eq_table.eq[eqn];
sys/dev/mlx4/mlx4_core/mlx4_main.c
2836
eq->affinity_cpu_id = requested_cpu % num_online_cpus();
sys/dev/mlx4/mlx4_core/mlx4_main.c
2873
priv->eq_table.eq[MLX4_EQ_ASYNC].irq = entries[0].vector;
sys/dev/mlx4/mlx4_core/mlx4_main.c
2874
bitmap_zero(priv->eq_table.eq[MLX4_EQ_ASYNC].actv_ports.ports,
sys/dev/mlx4/mlx4_core/mlx4_main.c
2881
priv->eq_table.eq[i].irq =
sys/dev/mlx4/mlx4_core/mlx4_main.c
2885
bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
sys/dev/mlx4/mlx4_core/mlx4_main.c
2892
priv->eq_table.eq[i].actv_ports.ports);
sys/dev/mlx4/mlx4_core/mlx4_main.c
2927
priv->eq_table.eq[i].irq = dev->persist->pdev->irq;
sys/dev/mlx4/mlx4_core/mlx4_main.c
2929
bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
1570
enum res_eq_states state, struct res_eq **eq)
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
1612
if (!err && eq)
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
1613
*eq = r;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3048
struct res_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3054
err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_HW, &eq);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3071
eq->mtt = mtt;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3252
struct res_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3255
err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_RESERVED, &eq);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3259
err = get_res(dev, slave, eq->mtt->com.res_id, RES_MTT, NULL);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3267
atomic_dec(&eq->mtt->ref_count);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3268
put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3275
put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3358
struct res_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3361
err = get_res(dev, slave, res_id, RES_EQ, &eq);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
3365
if (eq->com.from_state != RES_EQ_HW) {
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5089
struct res_eq *eq;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5102
list_for_each_entry_safe(eq, tmp, eq_list, com.list) {
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5104
if (eq->com.owner == slave) {
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5105
eqn = eq->com.res_id;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5106
state = eq->com.from_state;
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5111
rb_erase(&eq->com.node,
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5113
list_del(&eq->com.list);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5115
kfree(eq);
sys/dev/mlx4/mlx4_core/mlx4_resource_tracker.c
5127
atomic_dec(&eq->mtt->ref_count);
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2328
int i, j, eq = 0, total_eqs = 0;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2340
ibdev->eq_table[eq] = total_eqs;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2342
&ibdev->eq_table[eq]))
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2343
eq++;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2345
ibdev->eq_table[eq] = -1;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2349
for (i = eq; i < dev->caps.num_comp_vectors;
sys/dev/mlx4/mlx4_ib/mlx4_ib_main.c
2354
ibdev->ib_dev.num_comp_vectors = eq;
sys/dev/mlx5/driver.h
1095
int mlx5_create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, u8 vecidx,
sys/dev/mlx5/driver.h
1097
int mlx5_destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
sys/dev/mlx5/driver.h
1114
int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
sys/dev/mlx5/driver.h
1115
void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq);
sys/dev/mlx5/driver.h
1116
int mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
sys/dev/mlx5/mlx5_core/mlx5_eq.c
100
static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
102
return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
105
static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
107
struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1));
sys/dev/mlx5/mlx5_core/mlx5_eq.c
109
return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
212
static void eq_update_ci(struct mlx5_eq *eq, int arm)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
214
__be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
215
u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
231
static int mlx5_eq_int(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
240
while ((eqe = next_eqe_sw(eq))) {
sys/dev/mlx5/mlx5_core/mlx5_eq.c
247
mlx5_core_dbg(eq->dev, "eqn %d, eqe type %s\n",
sys/dev/mlx5/mlx5_core/mlx5_eq.c
248
eq->eqn, eqe_type_str(eqe->type));
sys/dev/mlx5/mlx5_core/mlx5_eq.c
382
eqe->type, eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
386
++eq->cons_index;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
397
eq_update_ci(eq, 0);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
402
eq_update_ci(eq, 1);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
409
struct mlx5_eq *eq = eq_ptr;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
410
struct mlx5_core_dev *dev = eq->dev;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
414
mlx5_eq_int(dev, eq);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
420
static void init_eq_buf(struct mlx5_eq *eq)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
425
for (i = 0; i < eq->nent; i++) {
sys/dev/mlx5/mlx5_core/mlx5_eq.c
426
eqe = get_eqe(eq, i);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
431
int mlx5_create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, u8 vecidx,
sys/dev/mlx5/mlx5_core/mlx5_eq.c
442
eq->nent = roundup_pow_of_two(nent + MLX5_NUM_SPARE_EQE);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
443
eq->cons_index = 0;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
444
err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, 2 * PAGE_SIZE,
sys/dev/mlx5/mlx5_core/mlx5_eq.c
445
&eq->buf);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
449
init_eq_buf(eq);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
452
MLX5_FLD_SZ_BYTES(create_eq_in, pas[0]) * eq->buf.npages;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
460
mlx5_fill_page_array(&eq->buf, pas);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
466
MLX5_SET(eqc, eqc, log_eq_size, ilog2(eq->nent));
sys/dev/mlx5/mlx5_core/mlx5_eq.c
470
eq->buf.page_shift - MLX5_ADAPTER_PAGE_SHIFT);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
476
eq->eqn = MLX5_GET(create_eq_out, out, eq_number);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
477
eq->irqn = vecidx;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
478
eq->dev = dev;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
479
eq->doorbell = priv->uar->map + MLX5_EQ_DOORBEL_OFFSET;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
481
"mlx5_core", eq);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
500
eq_update_ci(eq, 1);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
506
free_irq(priv->msix_arr[vecidx].vector, eq);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
509
mlx5_cmd_destroy_eq(dev, eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
515
mlx5_buf_free(dev, &eq->buf);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
520
int mlx5_destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
524
free_irq(dev->priv.msix_arr[eq->irqn].vector, eq);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
525
err = mlx5_cmd_destroy_eq(dev, eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
528
eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
529
mlx5_buf_free(dev, &eq->buf);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
640
int mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq,
sys/dev/mlx5/mlx5_core/mlx5_eq.c
647
MLX5_SET(query_eq_in, in, eq_number, eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_eq.c
779
struct mlx5_eq *eq;
sys/dev/mlx5/mlx5_core/mlx5_eq.c
788
list_for_each_entry(eq, &dev->priv.eq_table.comp_eqs_list, list)
sys/dev/mlx5/mlx5_core/mlx5_eq.c
789
mlx5_eq_int(dev, eq);
sys/dev/mlx5/mlx5_core/mlx5_main.c
685
struct mlx5_eq *eq;
sys/dev/mlx5/mlx5_core/mlx5_main.c
689
list_for_each_entry(eq, &table->comp_eqs_list, list) {
sys/dev/mlx5/mlx5_core/mlx5_main.c
690
if (eq->index == vector) {
sys/dev/mlx5/mlx5_core/mlx5_main.c
691
*eqn = eq->eqn;
sys/dev/mlx5/mlx5_core/mlx5_main.c
692
*irqn = eq->irqn;
sys/dev/mlx5/mlx5_core/mlx5_main.c
706
struct mlx5_eq *eq, *n;
sys/dev/mlx5/mlx5_core/mlx5_main.c
709
list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) {
sys/dev/mlx5/mlx5_core/mlx5_main.c
710
list_del(&eq->list);
sys/dev/mlx5/mlx5_core/mlx5_main.c
712
if (mlx5_destroy_unmap_eq(dev, eq))
sys/dev/mlx5/mlx5_core/mlx5_main.c
714
eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_main.c
715
kfree(eq);
sys/dev/mlx5/mlx5_core/mlx5_main.c
724
struct mlx5_eq *eq;
sys/dev/mlx5/mlx5_core/mlx5_main.c
734
eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, dev->priv.numa_node);
sys/dev/mlx5/mlx5_core/mlx5_main.c
736
err = mlx5_create_map_eq(dev, eq,
sys/dev/mlx5/mlx5_core/mlx5_main.c
739
kfree(eq);
sys/dev/mlx5/mlx5_core/mlx5_main.c
742
mlx5_core_dbg(dev, "allocated completion EQN %d\n", eq->eqn);
sys/dev/mlx5/mlx5_core/mlx5_main.c
743
eq->index = i;
sys/dev/mlx5/mlx5_core/mlx5_main.c
745
list_add_tail(&eq->list, &table->comp_eqs_list);
sys/dev/mthca/mthca_cq.c
836
cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
sys/dev/mthca/mthca_cq.c
837
cq_context->comp_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn);
sys/dev/mthca/mthca_cq.c
937
synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector);
sys/dev/mthca/mthca_dev.h
233
struct mthca_eq eq[MTHCA_NUM_EQ];
sys/dev/mthca/mthca_eq.c
173
static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
sys/dev/mthca/mthca_eq.c
184
mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1),
sys/dev/mthca/mthca_eq.c
189
static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
sys/dev/mthca/mthca_eq.c
194
dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8);
sys/dev/mthca/mthca_eq.c
199
static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci)
sys/dev/mthca/mthca_eq.c
202
arbel_set_eq_ci(dev, eq, ci);
sys/dev/mthca/mthca_eq.c
204
tavor_set_eq_ci(dev, eq, ci);
sys/dev/mthca/mthca_eq.c
228
static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry)
sys/dev/mthca/mthca_eq.c
230
unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE;
sys/dev/mthca/mthca_eq.c
231
return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE;
sys/dev/mthca/mthca_eq.c
234
static inline struct mthca_eqe *next_eqe_sw(struct mthca_eq *eq)
sys/dev/mthca/mthca_eq.c
237
eqe = get_eqe(eq, eq->cons_index);
sys/dev/mthca/mthca_eq.c
260
static int mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq)
sys/dev/mthca/mthca_eq.c
267
while ((eqe = next_eqe_sw(eq))) {
sys/dev/mthca/mthca_eq.c
277
disarm_cq(dev, eq->eqn, disarm_cqn);
sys/dev/mthca/mthca_eq.c
349
mthca_warn(dev, "EQ overrun on EQN %d\n", eq->eqn);
sys/dev/mthca/mthca_eq.c
358
eqe->type, eqe->subtype, eq->eqn);
sys/dev/mthca/mthca_eq.c
363
++eq->cons_index;
sys/dev/mthca/mthca_eq.c
379
set_eq_ci(dev, eq, eq->cons_index);
sys/dev/mthca/mthca_eq.c
408
if (ecr & dev->eq_table.eq[i].eqn_mask) {
sys/dev/mthca/mthca_eq.c
409
if (mthca_eq_int(dev, &dev->eq_table.eq[i]))
sys/dev/mthca/mthca_eq.c
410
tavor_set_eq_ci(dev, &dev->eq_table.eq[i],
sys/dev/mthca/mthca_eq.c
411
dev->eq_table.eq[i].cons_index);
sys/dev/mthca/mthca_eq.c
412
tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn);
sys/dev/mthca/mthca_eq.c
420
struct mthca_eq *eq = eq_ptr;
sys/dev/mthca/mthca_eq.c
421
struct mthca_dev *dev = eq->dev;
sys/dev/mthca/mthca_eq.c
423
mthca_eq_int(dev, eq);
sys/dev/mthca/mthca_eq.c
424
tavor_set_eq_ci(dev, eq, eq->cons_index);
sys/dev/mthca/mthca_eq.c
425
tavor_eq_req_not(dev, eq->eqn);
sys/dev/mthca/mthca_eq.c
441
if (mthca_eq_int(dev, &dev->eq_table.eq[i])) {
sys/dev/mthca/mthca_eq.c
443
arbel_set_eq_ci(dev, &dev->eq_table.eq[i],
sys/dev/mthca/mthca_eq.c
444
dev->eq_table.eq[i].cons_index);
sys/dev/mthca/mthca_eq.c
454
struct mthca_eq *eq = eq_ptr;
sys/dev/mthca/mthca_eq.c
455
struct mthca_dev *dev = eq->dev;
sys/dev/mthca/mthca_eq.c
457
mthca_eq_int(dev, eq);
sys/dev/mthca/mthca_eq.c
458
arbel_set_eq_ci(dev, eq, eq->cons_index);
sys/dev/mthca/mthca_eq.c
459
arbel_eq_req_not(dev, eq->eqn_mask);
sys/dev/mthca/mthca_eq.c
468
struct mthca_eq *eq)
sys/dev/mthca/mthca_eq.c
478
eq->dev = dev;
sys/dev/mthca/mthca_eq.c
479
eq->nent = roundup_pow_of_two(max(nent, 2));
sys/dev/mthca/mthca_eq.c
480
npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE;
sys/dev/mthca/mthca_eq.c
482
eq->page_list = kmalloc(npages * sizeof *eq->page_list,
sys/dev/mthca/mthca_eq.c
484
if (!eq->page_list)
sys/dev/mthca/mthca_eq.c
488
eq->page_list[i].buf = NULL;
sys/dev/mthca/mthca_eq.c
500
eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev,
sys/dev/mthca/mthca_eq.c
502
if (!eq->page_list[i].buf)
sys/dev/mthca/mthca_eq.c
506
dma_unmap_addr_set(&eq->page_list[i], mapping, t);
sys/dev/mthca/mthca_eq.c
508
clear_page(eq->page_list[i].buf);
sys/dev/mthca/mthca_eq.c
511
for (i = 0; i < eq->nent; ++i)
sys/dev/mthca/mthca_eq.c
512
set_eqe_hw(get_eqe(eq, i));
sys/dev/mthca/mthca_eq.c
514
eq->eqn = mthca_alloc(&dev->eq_table.alloc);
sys/dev/mthca/mthca_eq.c
515
if (eq->eqn == -1)
sys/dev/mthca/mthca_eq.c
523
&eq->mr);
sys/dev/mthca/mthca_eq.c
535
eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24);
sys/dev/mthca/mthca_eq.c
543
eq_context->lkey = cpu_to_be32(eq->mr.ibmr.lkey);
sys/dev/mthca/mthca_eq.c
545
err = mthca_SW2HW_EQ(dev, mailbox, eq->eqn);
sys/dev/mthca/mthca_eq.c
554
eq->eqn_mask = swab32(1 << eq->eqn);
sys/dev/mthca/mthca_eq.c
555
eq->cons_index = 0;
sys/dev/mthca/mthca_eq.c
557
dev->eq_table.arm_mask |= eq->eqn_mask;
sys/dev/mthca/mthca_eq.c
560
eq->eqn, eq->nent);
sys/dev/mthca/mthca_eq.c
565
mthca_free_mr(dev, &eq->mr);
sys/dev/mthca/mthca_eq.c
568
mthca_free(&dev->eq_table.alloc, eq->eqn);
sys/dev/mthca/mthca_eq.c
572
if (eq->page_list[i].buf)
sys/dev/mthca/mthca_eq.c
574
eq->page_list[i].buf,
sys/dev/mthca/mthca_eq.c
575
dma_unmap_addr(&eq->page_list[i],
sys/dev/mthca/mthca_eq.c
581
kfree(eq->page_list);
sys/dev/mthca/mthca_eq.c
589
struct mthca_eq *eq)
sys/dev/mthca/mthca_eq.c
593
int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) /
sys/dev/mthca/mthca_eq.c
601
err = mthca_HW2SW_EQ(dev, mailbox, eq->eqn);
sys/dev/mthca/mthca_eq.c
605
dev->eq_table.arm_mask &= ~eq->eqn_mask;
sys/dev/mthca/mthca_eq.c
608
mthca_dbg(dev, "Dumping EQ context %02x:\n", eq->eqn);
sys/dev/mthca/mthca_eq.c
618
mthca_free_mr(dev, &eq->mr);
sys/dev/mthca/mthca_eq.c
621
eq->page_list[i].buf,
sys/dev/mthca/mthca_eq.c
622
dma_unmap_addr(&eq->page_list[i], mapping));
sys/dev/mthca/mthca_eq.c
624
kfree(eq->page_list);
sys/dev/mthca/mthca_eq.c
635
if (dev->eq_table.eq[i].have_irq) {
sys/dev/mthca/mthca_eq.c
636
free_irq(dev->eq_table.eq[i].msi_x_vector,
sys/dev/mthca/mthca_eq.c
637
dev->eq_table.eq + i);
sys/dev/mthca/mthca_eq.c
638
dev->eq_table.eq[i].have_irq = 0;
sys/dev/mthca/mthca_eq.c
799
&dev->eq_table.eq[MTHCA_EQ_COMP]);
sys/dev/mthca/mthca_eq.c
805
&dev->eq_table.eq[MTHCA_EQ_ASYNC]);
sys/dev/mthca/mthca_eq.c
811
&dev->eq_table.eq[MTHCA_EQ_CMD]);
sys/dev/mthca/mthca_eq.c
823
snprintf(dev->eq_table.eq[i].irq_name,
sys/dev/mthca/mthca_eq.c
827
err = request_irq(dev->eq_table.eq[i].msi_x_vector,
sys/dev/mthca/mthca_eq.c
831
0, dev->eq_table.eq[i].irq_name,
sys/dev/mthca/mthca_eq.c
832
dev->eq_table.eq + i);
sys/dev/mthca/mthca_eq.c
835
dev->eq_table.eq[i].have_irq = 1;
sys/dev/mthca/mthca_eq.c
838
snprintf(dev->eq_table.eq[0].irq_name, IB_DEVICE_NAME_MAX,
sys/dev/mthca/mthca_eq.c
844
IRQF_SHARED, dev->eq_table.eq[0].irq_name, dev);
sys/dev/mthca/mthca_eq.c
851
0, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
sys/dev/mthca/mthca_eq.c
854
dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn, err);
sys/dev/mthca/mthca_eq.c
857
0, dev->eq_table.eq[MTHCA_EQ_CMD].eqn);
sys/dev/mthca/mthca_eq.c
860
dev->eq_table.eq[MTHCA_EQ_CMD].eqn, err);
sys/dev/mthca/mthca_eq.c
864
arbel_eq_req_not(dev, dev->eq_table.eq[i].eqn_mask);
sys/dev/mthca/mthca_eq.c
866
tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn);
sys/dev/mthca/mthca_eq.c
872
mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_CMD]);
sys/dev/mthca/mthca_eq.c
875
mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_ASYNC]);
sys/dev/mthca/mthca_eq.c
878
mthca_free_eq(dev, &dev->eq_table.eq[MTHCA_EQ_COMP]);
sys/dev/mthca/mthca_eq.c
895
1, dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
sys/dev/mthca/mthca_eq.c
897
1, dev->eq_table.eq[MTHCA_EQ_CMD].eqn);
sys/dev/mthca/mthca_eq.c
900
mthca_free_eq(dev, &dev->eq_table.eq[i]);
sys/dev/mthca/mthca_main.c
763
dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector);
sys/dev/mthca/mthca_main.c
865
mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector = entries[0].vector;
sys/dev/mthca/mthca_main.c
866
mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector = entries[1].vector;
sys/dev/mthca/mthca_main.c
867
mdev->eq_table.eq[MTHCA_EQ_CMD ].msi_x_vector = entries[2].vector;
sys/dev/oce/oce_hw.c
489
oce_arm_eq(sc, sc->eq[0]->eq_id, 0, TRUE, FALSE);
sys/dev/oce/oce_if.c
2313
eqo = sc->eq[i];
sys/dev/oce/oce_if.c
2514
struct oce_eq *eq;
sys/dev/oce/oce_if.c
2539
for_all_evnt_queues(sc, eq, i)
sys/dev/oce/oce_if.c
2540
oce_drain_eq(eq);
sys/dev/oce/oce_if.c
2550
oce_arm_eq(sc, sc->eq[0]->eq_id, 0, TRUE, FALSE);
sys/dev/oce/oce_if.c
2558
struct oce_eq *eq;
sys/dev/oce/oce_if.c
2581
for_all_evnt_queues(sc, eq, i)
sys/dev/oce/oce_if.c
2582
oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
sys/dev/oce/oce_if.c
684
struct oce_eq *eq = ii->eq;
sys/dev/oce/oce_if.c
689
bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
sys/dev/oce/oce_if.c
692
eqe = RING_GET_CONSUMER_ITEM_VA(eq->ring, struct oce_eqe);
sys/dev/oce/oce_if.c
696
bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
sys/dev/oce/oce_if.c
698
RING_GET(eq->ring, 1);
sys/dev/oce/oce_if.c
707
oce_arm_eq(sc, eq->eq_id, num_eqes, FALSE, FALSE);
sys/dev/oce/oce_if.c
710
for (i = 0; i < eq->cq_valid; i++) {
sys/dev/oce/oce_if.c
711
cq = eq->cq[i];
sys/dev/oce/oce_if.c
716
for (i = 0; i < eq->cq_valid; i++) {
sys/dev/oce/oce_if.c
717
cq = eq->cq[i];
sys/dev/oce/oce_if.c
722
oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
sys/dev/oce/oce_if.c
801
if (ii->eq == NULL)
sys/dev/oce/oce_if.c
804
oce_arm_eq(sc, ii->eq->eq_id, 0, FALSE, TRUE);
sys/dev/oce/oce_if.c
808
ii->eq->intr++;
sys/dev/oce/oce_if.h
1060
int oce_mbox_create_eq(struct oce_eq *eq);
sys/dev/oce/oce_if.h
218
#define for_all_evnt_queues(sc, eq, i) \
sys/dev/oce/oce_if.h
219
for (i = 0, eq = sc->eq[0]; i < sc->neqs; i++, eq = sc->eq[i])
sys/dev/oce/oce_if.h
269
struct oce_eq *eq; /* pointer to the connected EQ */
sys/dev/oce/oce_if.h
631
struct oce_eq *eq;
sys/dev/oce/oce_if.h
847
struct oce_eq *eq[OCE_MAX_EQ]; /* Event queues */
sys/dev/oce/oce_if.h
982
void oce_drain_eq(struct oce_eq *eq);
sys/dev/oce/oce_mbox.c
1755
oce_mbox_create_eq(struct oce_eq *eq)
sys/dev/oce/oce_mbox.c
1759
POCE_SOFTC sc = eq->parent;
sys/dev/oce/oce_mbox.c
1773
num_pages = oce_page_list(eq->ring, &fwcmd->params.req.pages[0]);
sys/dev/oce/oce_mbox.c
1776
fwcmd->params.req.ctx.size = (eq->eq_cfg.item_size == 4) ? 0 : 1;
sys/dev/oce/oce_mbox.c
1777
fwcmd->params.req.ctx.count = OCE_LOG2(eq->eq_cfg.q_len / 256);
sys/dev/oce/oce_mbox.c
1779
fwcmd->params.req.ctx.delay_mult = eq->eq_cfg.cur_eqd;
sys/dev/oce/oce_mbox.c
1794
eq->eq_id = HOST_16(fwcmd->params.rsp.eq_id);
sys/dev/oce/oce_mbox.c
1840
ctx->v2.eq_id = cq->eq->eq_id;
sys/dev/oce/oce_mbox.c
1855
ctx->v0.eq_id = cq->eq->eq_id;
sys/dev/oce/oce_queue.c
1001
POCE_SOFTC sc = eq->parent;
sys/dev/oce/oce_queue.c
1004
eqe = RING_GET_CONSUMER_ITEM_VA(eq->ring, struct oce_eqe);
sys/dev/oce/oce_queue.c
1008
bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
sys/dev/oce/oce_queue.c
1011
RING_GET(eq->ring, 1);
sys/dev/oce/oce_queue.c
1015
oce_arm_eq(sc, eq->eq_id, num_eqe, FALSE, TRUE);
sys/dev/oce/oce_queue.c
123
sc->eq[vector] = oce_eq_create(sc, sc->enable_hwlro ? EQ_LEN_2048 : EQ_LEN_1024,
sys/dev/oce/oce_queue.c
126
if (!sc->eq[vector])
sys/dev/oce/oce_queue.c
132
rc = oce_wq_create(wq, sc->eq[i]);
sys/dev/oce/oce_queue.c
141
sc->eq[(i == 0) ? 0:(i-1)]);
sys/dev/oce/oce_queue.c
147
sc->mq = oce_mq_create(sc, sc->eq[0], 64);
sys/dev/oce/oce_queue.c
168
struct oce_eq *eq;
sys/dev/oce/oce_queue.c
191
for_all_evnt_queues(sc, eq, i) {
sys/dev/oce/oce_queue.c
192
if (eq)
sys/dev/oce/oce_queue.c
193
oce_eq_del(sc->eq[i]);
sys/dev/oce/oce_queue.c
311
oce_wq_create(struct oce_wq *wq, struct oce_eq *eq)
sys/dev/oce/oce_queue.c
319
eq,
sys/dev/oce/oce_queue.c
336
eq->cq[eq->cq_valid] = cq;
sys/dev/oce/oce_queue.c
337
eq->cq_valid++;
sys/dev/oce/oce_queue.c
487
oce_rq_create(struct oce_rq *rq, uint32_t if_id, struct oce_eq *eq)
sys/dev/oce/oce_queue.c
492
cq = oce_cq_create(sc, eq,
sys/dev/oce/oce_queue.c
50
static int oce_wq_create(struct oce_wq *wq, struct oce_eq *eq);
sys/dev/oce/oce_queue.c
506
eq->cq[eq->cq_valid] = cq;
sys/dev/oce/oce_queue.c
507
eq->cq_valid++;
sys/dev/oce/oce_queue.c
563
struct oce_eq *eq;
sys/dev/oce/oce_queue.c
567
eq = malloc(sizeof(struct oce_eq), M_DEVBUF, M_NOWAIT | M_ZERO);
sys/dev/oce/oce_queue.c
568
if (eq == NULL)
sys/dev/oce/oce_queue.c
57
static int oce_rq_create(struct oce_rq *rq, uint32_t if_id, struct oce_eq *eq);
sys/dev/oce/oce_queue.c
571
eq->parent = (void *)sc;
sys/dev/oce/oce_queue.c
572
eq->eq_id = 0xffff;
sys/dev/oce/oce_queue.c
573
eq->ring = oce_create_ring_buffer(sc, q_len, item_size);
sys/dev/oce/oce_queue.c
574
if (!eq->ring)
sys/dev/oce/oce_queue.c
577
eq->eq_cfg.q_len = q_len;
sys/dev/oce/oce_queue.c
578
eq->eq_cfg.item_size = item_size;
sys/dev/oce/oce_queue.c
579
eq->eq_cfg.cur_eqd = (uint8_t) eq_delay;
sys/dev/oce/oce_queue.c
581
rc = oce_mbox_create_eq(eq);
sys/dev/oce/oce_queue.c
585
sc->intrs[sc->neqs++].eq = eq;
sys/dev/oce/oce_queue.c
587
return eq;
sys/dev/oce/oce_queue.c
590
oce_eq_del(eq);
sys/dev/oce/oce_queue.c
599
oce_eq_del(struct oce_eq *eq)
sys/dev/oce/oce_queue.c
603
POCE_SOFTC sc = (POCE_SOFTC) eq->parent;
sys/dev/oce/oce_queue.c
605
if (eq->eq_id != 0xffff) {
sys/dev/oce/oce_queue.c
608
fwcmd->params.req.id = eq->eq_id;
sys/dev/oce/oce_queue.c
613
if (eq->ring != NULL) {
sys/dev/oce/oce_queue.c
614
oce_destroy_ring_buffer(sc, eq->ring);
sys/dev/oce/oce_queue.c
615
eq->ring = NULL;
sys/dev/oce/oce_queue.c
618
free(eq, M_DEVBUF);
sys/dev/oce/oce_queue.c
630
oce_mq_create(POCE_SOFTC sc, struct oce_eq *eq, uint32_t q_len)
sys/dev/oce/oce_queue.c
641
cq = oce_cq_create(sc, eq, CQ_LEN_256,
sys/dev/oce/oce_queue.c
65
static void oce_eq_del(struct oce_eq *eq);
sys/dev/oce/oce_queue.c
67
struct oce_eq *eq, uint32_t q_len);
sys/dev/oce/oce_queue.c
710
eq->cq[eq->cq_valid] = cq;
sys/dev/oce/oce_queue.c
711
eq->cq_valid++;
sys/dev/oce/oce_queue.c
712
mq->cq->eq = eq;
sys/dev/oce/oce_queue.c
72
struct oce_eq *eq,
sys/dev/oce/oce_queue.c
839
oce_cq_create(POCE_SOFTC sc, struct oce_eq *eq,
sys/dev/oce/oce_queue.c
858
cq->eq = eq;
sys/dev/oce/oce_queue.c
996
oce_drain_eq(struct oce_eq *eq)
sys/dev/ocs_fc/ocs_ddump.c
466
ocs_ddump_sli4_queue(textbuf, "eq", hw, hw->eq, hw->eq_count,
sys/dev/ocs_fc/ocs_ddump.c
554
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_ddump.c
562
ocs_list_foreach(&hw->eq_list, eq) {
sys/dev/ocs_fc/ocs_ddump.c
563
ocs_ddump_section(textbuf, "eq", eq->instance);
sys/dev/ocs_fc/ocs_ddump.c
564
ocs_ddump_value(textbuf, "queue-id", "%d", eq->queue->id);
sys/dev/ocs_fc/ocs_ddump.c
565
OCS_STAT(ocs_ddump_value(textbuf, "use_count", "%d", eq->use_count));
sys/dev/ocs_fc/ocs_ddump.c
566
ocs_list_foreach(&eq->cq_list, cq) {
sys/dev/ocs_fc/ocs_ddump.c
615
ocs_ddump_endsection(textbuf, "eq", eq->instance);
sys/dev/ocs_fc/ocs_hw.c
1277
sli_queue_free(&hw->sli, &hw->eq[i], destroy_queues, free_memory);
sys/dev/ocs_fc/ocs_hw.c
1442
sli_queue_reset(&hw->sli, &hw->eq[i]);
sys/dev/ocs_fc/ocs_hw.c
2102
rc = sli_queue_is_empty(&hw->sli, &hw->eq[vector]);
sys/dev/ocs_fc/ocs_hw.c
2106
sli_queue_arm(&hw->sli, &hw->eq[vector], TRUE);
sys/dev/ocs_fc/ocs_hw.c
2127
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_hw.c
2148
eq = hw->hw_eq[vector];
sys/dev/ocs_fc/ocs_hw.c
2150
OCS_STAT(eq->use_count++);
sys/dev/ocs_fc/ocs_hw.c
2152
rc = ocs_hw_eq_process(hw, eq, max_isr_time_msec);
sys/dev/ocs_fc/ocs_hw.c
2179
ocs_hw_eq_process(ocs_hw_t *hw, hw_eq_t *eq, uint32_t max_isr_time_msec)
sys/dev/ocs_fc/ocs_hw.c
2192
while (!done && !sli_queue_read(&hw->sli, eq->queue, eqe)) {
sys/dev/ocs_fc/ocs_hw.c
2221
if (eq->queue->n_posted > (eq->queue->posted_limit)) {
sys/dev/ocs_fc/ocs_hw.c
2222
sli_queue_arm(&hw->sli, eq->queue, FALSE);
sys/dev/ocs_fc/ocs_hw.c
2233
sli_queue_eq_arm(&hw->sli, eq->queue, TRUE);
sys/dev/ocs_fc/ocs_hw.c
946
sli_queue_arm(&hw->sli, &hw->eq[i], TRUE);
sys/dev/ocs_fc/ocs_hw.h
1119
hw_eq_t *eq; /*<< parent EQ */
sys/dev/ocs_fc/ocs_hw.h
1233
extern hw_cq_t *hw_new_cq(hw_eq_t *eq, uint32_t entry_count);
sys/dev/ocs_fc/ocs_hw.h
1239
extern void hw_del_eq(hw_eq_t *eq);
sys/dev/ocs_fc/ocs_hw.h
1359
extern int32_t ocs_hw_eq_process(ocs_hw_t *hw, hw_eq_t *eq, uint32_t max_isr_time_msec);
sys/dev/ocs_fc/ocs_hw.h
593
hw_eq_t *eq; /**< EQ that this HIO came up on */
sys/dev/ocs_fc/ocs_hw.h
939
sli4_queue_t eq[OCS_HW_MAX_NUM_EQ];
sys/dev/ocs_fc/ocs_hw_queues.c
135
eq = hw_new_eq(hw, len);
sys/dev/ocs_fc/ocs_hw_queues.c
136
if (eq == NULL) {
sys/dev/ocs_fc/ocs_hw_queues.c
150
if (!eq || !next_qt) {
sys/dev/ocs_fc/ocs_hw_queues.c
156
cq = hw_new_cq(eq, len);
sys/dev/ocs_fc/ocs_hw_queues.c
1763
seq->hw_priv = cq->eq;
sys/dev/ocs_fc/ocs_hw_queues.c
185
if (ocs_varray_add(eq->wq_array, wq)) {
sys/dev/ocs_fc/ocs_hw_queues.c
1870
seq->hw_priv = cq->eq;
sys/dev/ocs_fc/ocs_hw_queues.c
1989
seq->hw_priv = cq->eq;
sys/dev/ocs_fc/ocs_hw_queues.c
208
if (ocs_varray_add(hw->wq_cpu_array[eq->instance % ocs_get_num_cpus()], wq)) {
sys/dev/ocs_fc/ocs_hw_queues.c
233
mrq.rq_cfg[mrq.num_pairs].eq = eq;
sys/dev/ocs_fc/ocs_hw_queues.c
283
eqs[rqset_count] = mrq.rq_cfg[i].eq;
sys/dev/ocs_fc/ocs_hw_queues.c
290
cq = hw_new_cq(mrq.rq_cfg[i].eq, default_lengths[QTOP_CQ]);
sys/dev/ocs_fc/ocs_hw_queues.c
357
hw_eq_t *eq = ocs_malloc(hw->os, sizeof(*eq), OCS_M_ZERO | OCS_M_NOWAIT);
sys/dev/ocs_fc/ocs_hw_queues.c
359
if (eq != NULL) {
sys/dev/ocs_fc/ocs_hw_queues.c
360
eq->type = SLI_QTYPE_EQ;
sys/dev/ocs_fc/ocs_hw_queues.c
361
eq->hw = hw;
sys/dev/ocs_fc/ocs_hw_queues.c
362
eq->entry_count = entry_count;
sys/dev/ocs_fc/ocs_hw_queues.c
363
eq->instance = hw->eq_count++;
sys/dev/ocs_fc/ocs_hw_queues.c
364
eq->queue = &hw->eq[eq->instance];
sys/dev/ocs_fc/ocs_hw_queues.c
365
ocs_list_init(&eq->cq_list, hw_cq_t, link);
sys/dev/ocs_fc/ocs_hw_queues.c
367
eq->wq_array = ocs_varray_alloc(hw->os, OCS_HW_MAX_NUM_WQ);
sys/dev/ocs_fc/ocs_hw_queues.c
368
if (eq->wq_array == NULL) {
sys/dev/ocs_fc/ocs_hw_queues.c
369
ocs_free(hw->os, eq, sizeof(*eq));
sys/dev/ocs_fc/ocs_hw_queues.c
370
eq = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
372
if (sli_queue_alloc(&hw->sli, SLI_QTYPE_EQ, eq->queue, entry_count, NULL, 0)) {
sys/dev/ocs_fc/ocs_hw_queues.c
373
ocs_log_err(hw->os, "EQ[%d] allocation failure\n", eq->instance);
sys/dev/ocs_fc/ocs_hw_queues.c
374
ocs_free(hw->os, eq, sizeof(*eq));
sys/dev/ocs_fc/ocs_hw_queues.c
375
eq = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
377
sli_eq_modify_delay(&hw->sli, eq->queue, 1, 0, 8);
sys/dev/ocs_fc/ocs_hw_queues.c
378
hw->hw_eq[eq->instance] = eq;
sys/dev/ocs_fc/ocs_hw_queues.c
379
ocs_list_add_tail(&hw->eq_list, eq);
sys/dev/ocs_fc/ocs_hw_queues.c
380
ocs_log_debug(hw->os, "create eq[%2d] id %3d len %4d\n", eq->instance, eq->queue->id,
sys/dev/ocs_fc/ocs_hw_queues.c
381
eq->entry_count);
sys/dev/ocs_fc/ocs_hw_queues.c
385
return eq;
sys/dev/ocs_fc/ocs_hw_queues.c
399
hw_new_cq(hw_eq_t *eq, uint32_t entry_count)
sys/dev/ocs_fc/ocs_hw_queues.c
401
ocs_hw_t *hw = eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
405
cq->eq = eq;
sys/dev/ocs_fc/ocs_hw_queues.c
407
cq->instance = eq->hw->cq_count++;
sys/dev/ocs_fc/ocs_hw_queues.c
413
if (sli_queue_alloc(&hw->sli, SLI_QTYPE_CQ, cq->queue, cq->entry_count, eq->queue, 0)) {
sys/dev/ocs_fc/ocs_hw_queues.c
415
eq->instance,
sys/dev/ocs_fc/ocs_hw_queues.c
416
eq->entry_count);
sys/dev/ocs_fc/ocs_hw_queues.c
421
ocs_list_add_tail(&eq->cq_list, cq);
sys/dev/ocs_fc/ocs_hw_queues.c
459
cq->eq = eqs[i];
sys/dev/ocs_fc/ocs_hw_queues.c
476
ocs_list_add_tail(&cqs[i]->eq->cq_list, cqs[i]);
sys/dev/ocs_fc/ocs_hw_queues.c
504
ocs_hw_t *hw = cq->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
510
mq->instance = cq->eq->hw->mq_count++;
sys/dev/ocs_fc/ocs_hw_queues.c
547
ocs_hw_t *hw = cq->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
551
wq->hw = cq->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
554
wq->instance = cq->eq->hw->wq_count++;
sys/dev/ocs_fc/ocs_hw_queues.c
592
ocs_hw_t *hw = cq->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
674
ocs_hw_t *hw = cqs[0]->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
68
hw_eq_t *eq = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
763
hw_del_eq(hw_eq_t *eq)
sys/dev/ocs_fc/ocs_hw_queues.c
765
if (eq != NULL) {
sys/dev/ocs_fc/ocs_hw_queues.c
769
ocs_list_foreach_safe(&eq->cq_list, cq, cq_next) {
sys/dev/ocs_fc/ocs_hw_queues.c
772
ocs_varray_free(eq->wq_array);
sys/dev/ocs_fc/ocs_hw_queues.c
773
ocs_list_remove(&eq->hw->eq_list, eq);
sys/dev/ocs_fc/ocs_hw_queues.c
774
eq->hw->hw_eq[eq->instance] = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
775
ocs_free(eq->hw->os, eq, sizeof(*eq));
sys/dev/ocs_fc/ocs_hw_queues.c
810
ocs_list_remove(&cq->eq->cq_list, cq);
sys/dev/ocs_fc/ocs_hw_queues.c
811
cq->eq->hw->hw_cq[cq->instance] = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
812
ocs_free(cq->eq->hw->os, cq, sizeof(*cq));
sys/dev/ocs_fc/ocs_hw_queues.c
830
mq->cq->eq->hw->hw_mq[mq->instance] = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
831
ocs_free(mq->cq->eq->hw->os, mq, sizeof(*mq));
sys/dev/ocs_fc/ocs_hw_queues.c
849
wq->cq->eq->hw->hw_wq[wq->instance] = NULL;
sys/dev/ocs_fc/ocs_hw_queues.c
850
ocs_free(wq->cq->eq->hw->os, wq, sizeof(*wq));
sys/dev/ocs_fc/ocs_hw_queues.c
868
ocs_hw_t *hw = rq->cq->eq->hw;
sys/dev/ocs_fc/ocs_hw_queues.c
892
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_hw_queues.c
899
ocs_list_foreach(&hw->eq_list, eq) {
sys/dev/ocs_fc/ocs_hw_queues.c
900
ocs_printf("eq[%d] id %2d\n", eq->instance, eq->queue->id);
sys/dev/ocs_fc/ocs_hw_queues.c
901
ocs_list_foreach(&eq->cq_list, cq) {
sys/dev/ocs_fc/ocs_hw_queues.c
938
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_hw_queues.c
942
ocs_list_foreach_safe(&hw->eq_list, eq, eq_next) {
sys/dev/ocs_fc/ocs_hw_queues.c
943
hw_del_eq(eq);
sys/dev/ocs_fc/ocs_hw_queues.c
978
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_hw_queues.c
988
eq = io->eq;
sys/dev/ocs_fc/ocs_hw_queues.c
989
if (likely(eq != NULL)) {
sys/dev/ocs_fc/ocs_hw_queues.c
990
wq = ocs_varray_iter_next(eq->wq_array);
sys/dev/ocs_fc/ocs_hw_queues.h
62
hw_eq_t *eq;
sys/dev/ocs_fc/ocs_hw_queues.h
90
extern void hw_thread_eq_handler(ocs_hw_t *hw, hw_eq_t *eq, uint32_t max_isr_time_msec);
sys/dev/ocs_fc/ocs_scsi.c
1023
hio->eq = io->hw_priv;
sys/dev/ocs_fc/ocs_scsi.c
1257
hio->eq = io->hw_priv;
sys/dev/ocs_fc/sli4.c
2071
sli4_req_common_create_eq_t *eq = NULL;
sys/dev/ocs_fc/sli4.c
2086
eq = (sli4_req_common_create_eq_t *)((uint8_t *)buf + sli_config_off);
sys/dev/ocs_fc/sli4.c
2088
eq->hdr.opcode = SLI4_OPC_COMMON_CREATE_EQ;
sys/dev/ocs_fc/sli4.c
2089
eq->hdr.subsystem = SLI4_SUBSYSTEM_COMMON;
sys/dev/ocs_fc/sli4.c
2090
eq->hdr.request_length = sizeof(sli4_req_common_create_eq_t) -
sys/dev/ocs_fc/sli4.c
2093
eq->hdr.version = 2;
sys/dev/ocs_fc/sli4.c
2094
eq->autovalid = TRUE;
sys/dev/ocs_fc/sli4.c
2097
eq->num_pages = qmem->size / SLI_PAGE_SIZE;
sys/dev/ocs_fc/sli4.c
2098
switch (eq->num_pages) {
sys/dev/ocs_fc/sli4.c
2100
eq->eqesz = SLI4_EQE_SIZE_4;
sys/dev/ocs_fc/sli4.c
2101
eq->count = SLI4_EQ_CNT_1024;
sys/dev/ocs_fc/sli4.c
2104
eq->eqesz = SLI4_EQE_SIZE_4;
sys/dev/ocs_fc/sli4.c
2105
eq->count = SLI4_EQ_CNT_2048;
sys/dev/ocs_fc/sli4.c
2108
eq->eqesz = SLI4_EQE_SIZE_4;
sys/dev/ocs_fc/sli4.c
2109
eq->count = SLI4_EQ_CNT_4096;
sys/dev/ocs_fc/sli4.c
2112
ocs_log_test(sli4->os, "num_pages %d not valid\n", eq->num_pages);
sys/dev/ocs_fc/sli4.c
2115
eq->valid = TRUE;
sys/dev/ocs_fc/sli4.c
2116
eq->arm = FALSE;
sys/dev/ocs_fc/sli4.c
2117
eq->delay_multiplier = 32;
sys/dev/ocs_fc/sli4.c
2120
p < eq->num_pages;
sys/dev/ocs_fc/sli4.c
2122
eq->page_address[p].low = ocs_addr32_lo(addr);
sys/dev/ocs_fc/sli4.c
2123
eq->page_address[p].high = ocs_addr32_hi(addr);
sys/dev/ocs_fc/sli4.c
2148
sli4_req_common_destroy_eq_t *eq = NULL;
sys/dev/ocs_fc/sli4.c
2158
eq = (sli4_req_common_destroy_eq_t *)((uint8_t *)buf + sli_config_off);
sys/dev/ocs_fc/sli4.c
2160
eq->hdr.opcode = SLI4_OPC_COMMON_DESTROY_EQ;
sys/dev/ocs_fc/sli4.c
2161
eq->hdr.subsystem = SLI4_SUBSYSTEM_COMMON;
sys/dev/ocs_fc/sli4.c
2162
eq->hdr.request_length = sizeof(sli4_req_common_destroy_eq_t) -
sys/dev/ocs_fc/sli4.c
2165
eq->eq_id = eq_id;
sys/dev/ocs_fc/sli4.c
4616
sli_eq_modify_delay(sli4_t *sli4, sli4_queue_t *eq, uint32_t num_eq, uint32_t shift, uint32_t delay_mult)
sys/dev/ocs_fc/sli4.c
4619
sli_cmd_common_modify_eq_delay(sli4, sli4->bmbx.virt, SLI4_BMBX_SIZE, eq, num_eq, shift, delay_mult);
sys/dev/ocs_fc/sli4.h
3736
extern int32_t sli_eq_modify_delay(sli4_t *sli4, sli4_queue_t *eq, uint32_t num_eq, uint32_t shift, uint32_t delay_mult);
sys/dev/pms/RefTisa/sallsdk/api/sa.h
3309
bit32 eq; /* error qualifier */
sys/dev/pms/RefTisa/sallsdk/api/saosapi.h
700
bit32 eq
sys/dev/pms/RefTisa/sallsdk/api/saosapi.h
735
bit32 eq
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7077
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7093
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7108
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7124
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7222
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7328
agEvent.eq = errorQualifier;
sys/dev/pms/RefTisa/sallsdk/spc/sampirsp.c
7410
agEvent.eq = subop;
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
3177
encryptEventData.subEvent = pEncryptCBData->eq;
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
3189
TI_DBG1(("ossaHwCB: encrypt response status 0x%x error qualifier 0x%x\n", pEncryptCBData->status, pEncryptCBData->eq));
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8279
bit32 eq
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8286
TI_DBG1(("ossaOperatorManagementCB: status 0x%x eq 0x%x\n", status, eq));
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8290
encryptEventData.subEvent = eq;
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8389
bit32 eq
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8396
TI_DBG1(("ossaSetOperatorCB: agContext %p status 0x%x eq 0x%x\n",agContext, status, eq));
sys/dev/pms/RefTisa/tisa/sassata/common/ossacmnapi.c
8411
TI_DBG1(("ossaSetOperatorCB: OSSA_MPI_ENC_OPERATOR_AUTH_FAILURE error qualifier 0x%x\n",eq));
sys/dev/sound/pcm/feeder_chain.c
789
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
796
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
804
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
808
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
815
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
826
FEEDER_BUILD(eq);
sys/dev/sound/pcm/feeder_chain.c
830
FEEDER_BUILD(eq);
sys/kern/kern_clocksource.c
381
int eq;
sys/kern/kern_clocksource.c
406
eq = (new == *next);
sys/kern/kern_clocksource.c
408
(int)(new >> 32), (u_int)(new & 0xffffffff), eq);
sys/kern/kern_clocksource.c
409
if (!eq) {
sys/kern/subr_hints.c
61
char *line, *eq;
sys/kern/subr_hints.c
66
eq = strchr(cp, '=');
sys/kern/subr_hints.c
67
if (eq == NULL)
sys/kern/subr_hints.c
70
eqidx = eq - cp;
sys/sys/refcount.h
192
_refcount_release_if_cond(old == n, eq)
usr.bin/calendar/sunpos.c
443
int eq[2], sol[2];
usr.bin/calendar/sunpos.c
444
equinoxsolstice(strtol(argv[1], NULL, 10), 0.0, eq, sol);
usr.bin/calendar/sunpos.c
445
printf("%d - %d - %d - %d\n", eq[0], sol[0], eq[1], sol[1]);
usr.bin/sed/compile.c
680
char wfile[_POSIX2_LINE_MAX + 1], *q, *eq;
usr.bin/sed/compile.c
733
eq = wfile + sizeof(wfile) - 1;
usr.bin/sed/compile.c
737
if (q >= eq)
usr.sbin/config/config.y
154
if (machinename != NULL && !eq($2, machinename))
usr.sbin/config/config.y
165
if (machinename != NULL && !eq($2, machinename))
usr.sbin/config/config.y
182
if (eq(cp->cpu_name, $2)) {
usr.sbin/config/config.y
382
if (eq(dp->d_name, name))
usr.sbin/config/config.y
438
if (eq(op->op_name, name))
usr.sbin/config/config.y
455
if (eq(name, OPT_AUTOGEN)) {
usr.sbin/config/main.cc
675
if (eq(dp->d_name, hl->h_name)) {
usr.sbin/config/main.cc
706
if (eq(s, hl->h_name)) {
usr.sbin/config/mkmakefile.cc
162
if (eq(line, "%BEFORE_DEPEND\n"))
usr.sbin/config/mkmakefile.cc
164
else if (eq(line, "%OBJS\n"))
usr.sbin/config/mkmakefile.cc
168
else if (eq(line, "%RULES\n"))
usr.sbin/config/mkmakefile.cc
170
else if (eq(line, "%CLEAN\n"))
usr.sbin/config/mkmakefile.cc
187
const char *eq;
usr.sbin/config/mkmakefile.cc
192
if ((eq = strchr(src, '=')) == NULL) {
usr.sbin/config/mkmakefile.cc
200
while (src < eq) {
usr.sbin/config/mkmakefile.cc
432
if (eq(wd, "include")) {
usr.sbin/config/mkmakefile.cc
466
if (eq(wd, "standard"))
usr.sbin/config/mkmakefile.cc
468
else if (!eq(wd, "optional"))
usr.sbin/config/mkmakefile.cc
474
if (eq(wd, "!")) {
usr.sbin/config/mkmakefile.cc
478
if (eq(wd, "|")) {
usr.sbin/config/mkmakefile.cc
487
if (eq(wd, "no-ctfconvert")) {
usr.sbin/config/mkmakefile.cc
491
if (eq(wd, "no-obj")) {
usr.sbin/config/mkmakefile.cc
495
if (eq(wd, "no-implicit-rule")) {
usr.sbin/config/mkmakefile.cc
504
if (eq(wd, "before-depend")) {
usr.sbin/config/mkmakefile.cc
508
if (eq(wd, "dependency")) {
usr.sbin/config/mkmakefile.cc
516
if (eq(wd, "clean")) {
usr.sbin/config/mkmakefile.cc
524
if (eq(wd, "compile-with")) {
usr.sbin/config/mkmakefile.cc
532
if (eq(wd, "warning")) {
usr.sbin/config/mkmakefile.cc
540
if (eq(wd, "obj-prefix")) {
usr.sbin/config/mkmakefile.cc
548
if (eq(wd, "nowerror")) {
usr.sbin/config/mkmakefile.cc
552
if (eq(wd, "local")) {
usr.sbin/config/mkmakefile.cc
556
if (eq(wd, "no-depend")) {
usr.sbin/config/mkmakefile.cc
565
if (eq(dp->d_name, wd)) {
usr.sbin/config/mkmakefile.cc
88
if (eq(fp->f_fn, file))
usr.sbin/config/mkoptions.cc
112
if (eq(op->op_name, ol->o_name) &&
usr.sbin/config/mkoptions.cc
156
if (eq(name, op->op_name)) {
usr.sbin/config/mkoptions.cc
161
if (oldvalue != NULL && !eq(value, oldvalue))
usr.sbin/config/mkoptions.cc
187
if (eq(name, ol->o_name)) {
usr.sbin/config/mkoptions.cc
210
if (eq(inw, name)) {
usr.sbin/config/mkoptions.cc
216
if (eq(inw, ol->o_name))
usr.sbin/config/mkoptions.cc
218
if (!eq(inw, name) && !ol) {
usr.sbin/config/mkoptions.cc
223
} else if (ol != NULL && !eq(basefile, ol->o_file)) {
usr.sbin/config/mkoptions.cc
244
(value && oldvalue && eq(value, oldvalue)))) {
usr.sbin/config/mkoptions.cc
298
if (eq(po->o_name, name)) {
usr.sbin/config/mkoptions.cc
317
if (eq(po->o_name, chkopt)) {
usr.sbin/config/mkoptions.cc
346
if (eq(po->o_name, optname)) {
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
164
equalizer_load(struct Equalizer *eq, const char *config)
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
167
int N = eq->block_size;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
171
memset(eq->fftw_freq, 0, buffer_size);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
175
if (!equalizer_load_freq_amps(eq, config))
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
180
memcpy(requested_freq, eq->fftw_freq, buffer_size);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
182
fftw_execute(eq->inverse);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
188
eq->fftw_time[N / 2 + i] = eq->fftw_time[i] * weight;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
191
eq->fftw_time[i] = eq->fftw_time[N - i];
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
193
eq->fftw_time[0] = 0;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
195
fftw_execute(eq->forward);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
197
eq->fftw_freq[i] /= (double)N;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
202
double f = (eq->rate / N) * i;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
203
double a = sqrt(pow(eq->fftw_freq[i], 2.0) +
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
204
((i > 0 && i < N / 2) ? pow(eq->fftw_freq[N - i], 2.0) : 0));
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
214
if (eq->do_normalize) {
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
218
sum += fabs(eq->fftw_time[i]);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
221
eq->fftw_time[i] /= sum;
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
225
message("%.3lf ms: %.10lf\n", 1000.0 * i / eq->rate, eq->fftw_time[i]);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
238
equalizer_done(struct Equalizer *eq)
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
241
fftw_destroy_plan(eq->forward);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
242
fftw_destroy_plan(eq->inverse);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
243
free(eq->fftw_time);
usr.sbin/virtual_oss/virtual_equalizer/equalizer.c
244
free(eq->fftw_freq);