Symbol: queue
lib/libcrypto/x509/x509_issuer_cache.c
135
TAILQ_REMOVE(&x509_issuer_lru, found, queue);
lib/libcrypto/x509/x509_issuer_cache.c
136
TAILQ_INSERT_HEAD(&x509_issuer_lru, found, queue);
lib/libcrypto/x509/x509_issuer_cache.c
180
TAILQ_INSERT_HEAD(&x509_issuer_lru, new, queue);
lib/libcrypto/x509/x509_issuer_cache.c
89
TAILQ_REMOVE(&x509_issuer_lru, old, queue);
lib/libcrypto/x509/x509_issuer_cache.h
31
TAILQ_ENTRY(x509_issuer) queue; /* LRU of entries */
lib/libcurses/nc_termios.h
157
#define tcflush(fd, queue) _nc_mingw_tcflush(fd, queue)
lib/libevent/event.c
844
event_queue_remove(struct event_base *base, struct event *ev, int queue)
lib/libevent/event.c
846
if (!(ev->ev_flags & queue))
lib/libevent/event.c
848
ev, ev->ev_fd, queue);
lib/libevent/event.c
853
ev->ev_flags &= ~queue;
lib/libevent/event.c
854
switch (queue) {
lib/libevent/event.c
867
event_errx(1, "%s: unknown queue %x", __func__, queue);
lib/libevent/event.c
872
event_queue_insert(struct event_base *base, struct event *ev, int queue)
lib/libevent/event.c
874
if (ev->ev_flags & queue) {
lib/libevent/event.c
876
if (queue & EVLIST_ACTIVE)
lib/libevent/event.c
880
ev, ev->ev_fd, queue);
lib/libevent/event.c
886
ev->ev_flags |= queue;
lib/libevent/event.c
887
switch (queue) {
lib/libevent/event.c
901
event_errx(1, "%s: unknown queue %x", __func__, queue);
lib/libssl/d1_lib.c
108
dtls1_drain_rcontents(pqueue queue)
lib/libssl/d1_lib.c
113
if (queue == NULL)
lib/libssl/d1_lib.c
116
while ((item = pqueue_pop(queue)) != NULL) {
lib/libssl/d1_lib.c
125
dtls1_drain_records(pqueue queue)
lib/libssl/d1_lib.c
130
if (queue == NULL)
lib/libssl/d1_lib.c
133
while ((item = pqueue_pop(queue)) != NULL) {
lib/libssl/d1_lib.c
142
dtls1_drain_fragments(pqueue queue)
lib/libssl/d1_lib.c
146
if (queue == NULL)
lib/libssl/d1_lib.c
149
while ((item = pqueue_pop(queue)) != NULL) {
lib/libssl/d1_pkt.c
207
dtls1_buffer_record(SSL *s, record_pqueue *queue, unsigned char *priority)
lib/libssl/d1_pkt.c
213
if (pqueue_size(queue->q) >= 100)
lib/libssl/d1_pkt.c
237
if (pqueue_insert(queue->q, item) == NULL)
lib/libssl/d1_pkt.c
253
dtls1_buffer_rcontent(SSL *s, rcontent_pqueue *queue, unsigned char *priority)
lib/libssl/d1_pkt.c
259
if (pqueue_size(queue->q) >= 100)
lib/libssl/d1_pkt.c
273
if (pqueue_insert(queue->q, item) == NULL)
lib/libssl/d1_pkt.c
292
dtls1_retrieve_buffered_record(SSL *s, record_pqueue *queue)
lib/libssl/d1_pkt.c
296
item = pqueue_pop(queue->q);
lib/libssl/d1_pkt.c
310
dtls1_retrieve_buffered_rcontent(SSL *s, rcontent_pqueue *queue)
lib/libssl/d1_pkt.c
315
item = pqueue_pop(queue->q);
regress/sbin/iked/parser/common.c
213
ikev2_msg_flushqueue(struct iked *env, struct iked_msgqueue *queue)
regress/sbin/iked/parser/common.c
218
ikev2_msg_lookup(struct iked *env, struct iked_msgqueue *queue,
sbin/iked/ikev2_msg.c
1200
ikev2_msg_enqueue(struct iked *env, struct iked_msgqueue *queue,
sbin/iked/ikev2_msg.c
1205
if ((mr = ikev2_msg_lookup(env, queue, msg, msg->msg_exchange)) ==
sbin/iked/ikev2_msg.c
1217
TAILQ_INSERT_TAIL(queue, mr, mrt_entry);
sbin/iked/ikev2_msg.c
1226
ikev2_msg_prevail(struct iked *env, struct iked_msgqueue *queue,
sbin/iked/ikev2_msg.c
1231
TAILQ_FOREACH_SAFE(mr, queue, mrt_entry, mrtmp) {
sbin/iked/ikev2_msg.c
1233
ikev2_msg_dispose(env, queue, mr);
sbin/iked/ikev2_msg.c
1238
ikev2_msg_dispose(struct iked *env, struct iked_msgqueue *queue,
sbin/iked/ikev2_msg.c
1250
TAILQ_REMOVE(queue, mr, mrt_entry);
sbin/iked/ikev2_msg.c
1255
ikev2_msg_flushqueue(struct iked *env, struct iked_msgqueue *queue)
sbin/iked/ikev2_msg.c
1259
while ((mr = TAILQ_FIRST(queue)) != NULL)
sbin/iked/ikev2_msg.c
1260
ikev2_msg_dispose(env, queue, mr);
sbin/iked/ikev2_msg.c
1264
ikev2_msg_lookup(struct iked *env, struct iked_msgqueue *queue,
sbin/iked/ikev2_msg.c
1269
TAILQ_FOREACH(mr, queue, mrt_entry) {
sbin/pfctl/parse.y
188
char queue[PF_QNAME_SIZE];
sbin/pfctl/parse.y
500
struct node_queue *queue;
sbin/pfctl/pfctl.c
1729
bcopy(&qi->qs, &q.queue, sizeof(q.queue));
sbin/pfctl/pfctl_optimize.c
790
TAILQ_HEAD( , pf_opt_rule) queue;
sbin/pfctl/pfctl_optimize.c
821
TAILQ_INIT(&queue);
sbin/pfctl/pfctl_optimize.c
822
TAILQ_CONCAT(&queue, &block->sb_rules, por_entry);
sbin/pfctl/pfctl_optimize.c
824
while ((por1 = TAILQ_FIRST(&queue)) != NULL) {
sbin/pfctl/pfctl_optimize.c
825
TAILQ_REMOVE(&queue, por1, por_entry);
sbin/pfctl/pfctl_optimize.c
851
struct pf_opt_queue queue;
sbin/pfctl/pfctl_optimize.c
856
TAILQ_INIT(&queue);
sbin/pfctl/pfctl_optimize.c
883
TAILQ_INSERT_TAIL(&queue, por, por_entry);
sbin/pfctl/pfctl_optimize.c
886
if (construct_superblocks(pf, &queue, &prof_superblocks))
sbin/pfctl/pfctl_queue.c
145
if ((node = pfctl_find_queue_node(pqs.queue.qname,
sbin/pfctl/pfctl_queue.c
146
pqs.queue.ifname)) != NULL) {
sbin/pfctl/pfctl_queue.c
151
pfctl_insert_queue_node(pqs.queue, qstats);
sys/arch/arm64/dev/agintc.c
1850
struct gits_cmd *queue = AGINTC_DMA_KVA(sc->sc_cmdq);
sys/arch/arm64/dev/agintc.c
1852
memcpy(&queue[sc->sc_cmdidx], cmd, sizeof(*cmd));
sys/arch/arm64/dev/agintc.c
1855
cpu_dcache_wb_range((vaddr_t)&queue[sc->sc_cmdidx], sizeof(*cmd));
sys/arch/sparc64/include/hypervisor.h
42
int64_t hv_cpu_qconf(uint64_t queue, uint64_t base, uint64_t nentries);
sys/dev/fdt/if_mvpp.c
3190
int val, queue;
sys/dev/fdt/if_mvpp.c
3196
for (queue = 0; queue < MVPP2_MAX_TXQ; queue++)
sys/dev/fdt/if_mvpp.c
3197
mvpp2_write(port->sc, MVPP2_TXQ_SCHED_TOKEN_CNTR_REG(queue), 0);
sys/dev/fdt/if_mvpp.c
3222
int lrxq, queue;
sys/dev/fdt/if_mvpp.c
3225
queue = port->sc_rxqs[lrxq].id;
sys/dev/fdt/if_mvpp.c
3226
val = mvpp2_read(port->sc, MVPP2_RXQ_CONFIG_REG(queue));
sys/dev/fdt/if_mvpp.c
3228
mvpp2_write(port->sc, MVPP2_RXQ_CONFIG_REG(queue), val);
sys/dev/fdt/if_mvpp.c
3236
int lrxq, queue;
sys/dev/fdt/if_mvpp.c
3239
queue = port->sc_rxqs[lrxq].id;
sys/dev/fdt/if_mvpp.c
3240
val = mvpp2_read(port->sc, MVPP2_RXQ_CONFIG_REG(queue));
sys/dev/fdt/if_mvpp.c
3242
mvpp2_write(port->sc, MVPP2_RXQ_CONFIG_REG(queue), val);
sys/dev/fdt/if_mvpp.c
3251
int queue;
sys/dev/fdt/if_mvpp.c
3254
for (queue = 0; queue < port->sc_ntxq; queue++) {
sys/dev/fdt/if_mvpp.c
3255
txq = &port->sc_txqs[queue];
sys/dev/fdt/if_mvpp.c
3258
qmap |= (1 << queue);
sys/dev/ic/aac.c
1592
aac_enqueue_fib(struct aac_softc *sc, int queue, struct aac_command *cm)
sys/dev/ic/aac.c
1603
pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
sys/dev/ic/aac.c
1604
ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
sys/dev/ic/aac.c
1607
if (pi >= aac_qinfo[queue].size)
sys/dev/ic/aac.c
1617
(sc->aac_qentries[queue] + pi)->aq_fib_size = fib_size;
sys/dev/ic/aac.c
1618
(sc->aac_qentries[queue] + pi)->aq_fib_addr = fib_addr;
sys/dev/ic/aac.c
1621
sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX] = pi + 1;
sys/dev/ic/aac.c
1630
if (aac_qinfo[queue].notify != 0)
sys/dev/ic/aac.c
1631
AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
sys/dev/ic/aac.c
1644
aac_dequeue_fib(struct aac_softc *sc, int queue, u_int32_t *fib_size,
sys/dev/ic/aac.c
1653
pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
sys/dev/ic/aac.c
1654
ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
sys/dev/ic/aac.c
1663
if (pi >= aac_qinfo[queue].size)
sys/dev/ic/aac.c
1671
if (ci >= aac_qinfo[queue].size)
sys/dev/ic/aac.c
1675
*fib_size = (sc->aac_qentries[queue] + ci)->aq_fib_size;
sys/dev/ic/aac.c
1677
switch (queue) {
sys/dev/ic/aac.c
1686
fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr /
sys/dev/ic/aac.c
1703
fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr;
sys/dev/ic/aac.c
1724
sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX] = ci + 1;
sys/dev/ic/aac.c
1727
if (notify && (aac_qinfo[queue].notify != 0))
sys/dev/ic/aac.c
1728
AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
sys/dev/ic/aac.c
1739
aac_enqueue_response(struct aac_softc *sc, int queue, struct aac_fib *fib)
sys/dev/ic/aac.c
1752
pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
sys/dev/ic/aac.c
1753
ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
sys/dev/ic/aac.c
1756
if (pi >= aac_qinfo[queue].size)
sys/dev/ic/aac.c
1766
(sc->aac_qentries[queue] + pi)->aq_fib_size = fib_size;
sys/dev/ic/aac.c
1767
(sc->aac_qentries[queue] + pi)->aq_fib_addr = fib_addr;
sys/dev/ic/aac.c
1770
sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX] = pi + 1;
sys/dev/ic/aac.c
1773
if (aac_qinfo[queue].notify != 0)
sys/dev/ic/aac.c
1774
AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
sys/dev/ic/aac.c
87
int aac_enqueue_response(struct aac_softc *sc, int queue,
sys/dev/ic/aic7xxx.c
384
ahc_run_untagged_queue(struct ahc_softc *ahc, struct scb_tailq *queue)
sys/dev/ic/aic7xxx.c
391
if ((scb = TAILQ_FIRST(queue)) != NULL
sys/dev/ic/aic7xxxvar.h
1184
struct scb_tailq *queue);
sys/dev/ic/ar5210.c
725
u_int queue;
sys/dev/ic/ar5210.c
732
queue = 0;
sys/dev/ic/ar5210.c
736
queue = 1;
sys/dev/ic/ar5210.c
745
bzero(&hal->ah_txq[queue], sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5210.c
746
hal->ah_txq[queue].tqi_type = queue_type;
sys/dev/ic/ar5210.c
750
queue, queue_info) != AH_TRUE)
sys/dev/ic/ar5210.c
754
return (queue);
sys/dev/ic/ar5210.c
758
ar5k_ar5210_setup_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5210.c
761
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
763
if (hal->ah_txq[queue].tqi_type == HAL_TX_QUEUE_INACTIVE)
sys/dev/ic/ar5210.c
766
hal->ah_txq[queue].tqi_aifs = queue_info->tqi_aifs;
sys/dev/ic/ar5210.c
767
hal->ah_txq[queue].tqi_cw_max = queue_info->tqi_cw_max;
sys/dev/ic/ar5210.c
768
hal->ah_txq[queue].tqi_cw_min = queue_info->tqi_cw_min;
sys/dev/ic/ar5210.c
769
hal->ah_txq[queue].tqi_flags = queue_info->tqi_flags;
sys/dev/ic/ar5210.c
775
ar5k_ar5210_get_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5210.c
778
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
779
bcopy(&hal->ah_txq[queue], queue_info, sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5210.c
784
ar5k_ar5210_release_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
786
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
789
hal->ah_txq[queue].tqi_type = HAL_TX_QUEUE_INACTIVE;
sys/dev/ic/ar5210.c
812
ar5k_ar5210_reset_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
817
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
819
tq = &hal->ah_txq[queue];
sys/dev/ic/ar5210.c
865
ar5k_ar5210_get_tx_buf(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
869
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
874
switch (hal->ah_txq[queue].tqi_type) {
sys/dev/ic/ar5210.c
890
ar5k_ar5210_put_tx_buf(struct ath_hal *hal, u_int queue, u_int32_t phys_addr)
sys/dev/ic/ar5210.c
894
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
899
switch (hal->ah_txq[queue].tqi_type) {
sys/dev/ic/ar5210.c
918
ar5k_ar5210_num_tx_pending(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
924
ar5k_ar5210_tx_start(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
928
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
935
switch (hal->ah_txq[queue].tqi_type) {
sys/dev/ic/ar5210.c
964
ar5k_ar5210_stop_tx_dma(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5210.c
968
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5210.c
975
switch (hal->ah_txq[queue].tqi_type) {
sys/dev/ic/ar5211.c
1004
AR5K_REG_ENABLE_BITS(AR5K_AR5211_DCU_MISC(queue),
sys/dev/ic/ar5211.c
1010
AR5K_REG_WRITE(AR5K_AR5211_QCU_RDYTIMECFG(queue),
sys/dev/ic/ar5211.c
1018
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
1023
AR5K_REG_ENABLE_BITS(AR5K_AR5211_DCU_MISC(queue),
sys/dev/ic/ar5211.c
1029
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
1053
ar5k_ar5211_get_tx_buf(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
1055
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
1060
return (AR5K_REG_READ(AR5K_AR5211_QCU_TXDP(queue)));
sys/dev/ic/ar5211.c
1064
ar5k_ar5211_put_tx_buf(struct ath_hal *hal, u_int queue, u_int32_t phys_addr)
sys/dev/ic/ar5211.c
1066
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
1072
if (AR5K_REG_READ_Q(AR5K_AR5211_QCU_TXE, queue))
sys/dev/ic/ar5211.c
1075
AR5K_REG_WRITE(AR5K_AR5211_QCU_TXDP(queue), phys_addr);
sys/dev/ic/ar5211.c
1081
ar5k_ar5211_num_tx_pending(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
1083
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
1084
return (AR5K_AR5211_QCU_STS(queue) & AR5K_AR5211_QCU_STS_FRMPENDCNT);
sys/dev/ic/ar5211.c
1088
ar5k_ar5211_tx_start(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
1090
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
1093
if (AR5K_REG_READ_Q(AR5K_AR5211_QCU_TXD, queue))
sys/dev/ic/ar5211.c
1097
AR5K_REG_WRITE_Q(AR5K_AR5211_QCU_TXE, queue);
sys/dev/ic/ar5211.c
1103
ar5k_ar5211_stop_tx_dma(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
1107
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
1112
AR5K_REG_WRITE_Q(AR5K_AR5211_QCU_TXD, queue);
sys/dev/ic/ar5211.c
1115
pending = AR5K_REG_READ(AR5K_AR5211_QCU_STS(queue)) &
sys/dev/ic/ar5211.c
798
u_int queue;
sys/dev/ic/ar5211.c
804
for (queue = HAL_TX_QUEUE_ID_DATA_MIN;
sys/dev/ic/ar5211.c
805
hal->ah_txq[queue].tqi_type != HAL_TX_QUEUE_INACTIVE;
sys/dev/ic/ar5211.c
806
queue++)
sys/dev/ic/ar5211.c
807
if (queue > HAL_TX_QUEUE_ID_DATA_MAX)
sys/dev/ic/ar5211.c
810
queue = HAL_TX_QUEUE_ID_PSPOLL;
sys/dev/ic/ar5211.c
812
queue = HAL_TX_QUEUE_ID_BEACON;
sys/dev/ic/ar5211.c
814
queue = HAL_TX_QUEUE_ID_CAB;
sys/dev/ic/ar5211.c
821
bzero(&hal->ah_txq[queue], sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5211.c
822
hal->ah_txq[queue].tqi_type = queue_type;
sys/dev/ic/ar5211.c
825
if (ar5k_ar5211_setup_tx_queueprops(hal, queue, queue_info)
sys/dev/ic/ar5211.c
830
AR5K_Q_ENABLE_BITS(hal->ah_txq_interrupts, queue);
sys/dev/ic/ar5211.c
832
return (queue);
sys/dev/ic/ar5211.c
836
ar5k_ar5211_setup_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5211.c
839
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
841
if (hal->ah_txq[queue].tqi_type == HAL_TX_QUEUE_INACTIVE)
sys/dev/ic/ar5211.c
844
bcopy(queue_info, &hal->ah_txq[queue], sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5211.c
849
hal->ah_txq[queue].tqi_flags |=
sys/dev/ic/ar5211.c
856
ar5k_ar5211_get_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5211.c
859
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
860
bcopy(&hal->ah_txq[queue], queue_info, sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5211.c
865
ar5k_ar5211_release_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
867
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
870
hal->ah_txq[queue].tqi_type = HAL_TX_QUEUE_INACTIVE;
sys/dev/ic/ar5211.c
871
AR5K_Q_DISABLE_BITS(hal->ah_txq_interrupts, queue);
sys/dev/ic/ar5211.c
877
ar5k_ar5211_reset_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5211.c
884
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5211.c
886
tq = &hal->ah_txq[queue];
sys/dev/ic/ar5211.c
918
AR5K_REG_WRITE(AR5K_AR5211_DCU_RETRY_LMT(queue),
sys/dev/ic/ar5211.c
940
AR5K_REG_WRITE(AR5K_AR5211_DCU_LCL_IFS(queue),
sys/dev/ic/ar5211.c
949
AR5K_REG_WRITE(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
953
AR5K_REG_WRITE(AR5K_AR5211_QCU_CBRCFG(queue),
sys/dev/ic/ar5211.c
958
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
961
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
966
AR5K_REG_WRITE(AR5K_AR5211_QCU_RDYTIMECFG(queue),
sys/dev/ic/ar5211.c
973
AR5K_REG_WRITE(AR5K_AR5211_DCU_CHAN_TIME(queue),
sys/dev/ic/ar5211.c
979
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5211.c
985
AR5K_REG_WRITE(AR5K_AR5211_DCU_MISC(queue),
sys/dev/ic/ar5211.c
990
AR5K_REG_WRITE(AR5K_AR5211_DCU_MISC(queue),
sys/dev/ic/ar5211.c
999
AR5K_REG_ENABLE_BITS(AR5K_AR5211_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1000
queue = HAL_TX_QUEUE_ID_CAB;
sys/dev/ic/ar5212.c
1007
bzero(&hal->ah_txq[queue], sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5212.c
1009
if (ar5k_ar5212_setup_tx_queueprops(hal, queue, queue_info)
sys/dev/ic/ar5212.c
1013
hal->ah_txq[queue].tqi_type = queue_type;
sys/dev/ic/ar5212.c
1015
AR5K_Q_ENABLE_BITS(hal->ah_txq_interrupts, queue);
sys/dev/ic/ar5212.c
1017
return (queue);
sys/dev/ic/ar5212.c
1021
ar5k_ar5212_setup_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5212.c
1024
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1026
if (hal->ah_txq[queue].tqi_type != HAL_TX_QUEUE_INACTIVE)
sys/dev/ic/ar5212.c
1029
bcopy(queue_info, &hal->ah_txq[queue], sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5212.c
1034
hal->ah_txq[queue].tqi_flags |=
sys/dev/ic/ar5212.c
1041
ar5k_ar5212_get_tx_queueprops(struct ath_hal *hal, int queue,
sys/dev/ic/ar5212.c
1044
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1045
bcopy(&hal->ah_txq[queue], queue_info, sizeof(HAL_TXQ_INFO));
sys/dev/ic/ar5212.c
1050
ar5k_ar5212_release_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1052
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1055
hal->ah_txq[queue].tqi_type = HAL_TX_QUEUE_INACTIVE;
sys/dev/ic/ar5212.c
1056
AR5K_Q_DISABLE_BITS(hal->ah_txq_interrupts, queue);
sys/dev/ic/ar5212.c
1062
ar5k_ar5212_reset_tx_queue(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1069
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1071
tq = &hal->ah_txq[queue];
sys/dev/ic/ar5212.c
1106
AR5K_REG_WRITE(AR5K_AR5212_DCU_RETRY_LMT(queue),
sys/dev/ic/ar5212.c
1128
AR5K_REG_WRITE(AR5K_AR5212_DCU_LCL_IFS(queue),
sys/dev/ic/ar5212.c
1137
AR5K_REG_WRITE(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1141
AR5K_REG_WRITE(AR5K_AR5212_QCU_CBRCFG(queue),
sys/dev/ic/ar5212.c
1146
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1149
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1154
AR5K_REG_WRITE(AR5K_AR5212_QCU_RDYTIMECFG(queue),
sys/dev/ic/ar5212.c
1161
AR5K_REG_WRITE(AR5K_AR5212_DCU_CHAN_TIME(queue),
sys/dev/ic/ar5212.c
1167
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1173
AR5K_REG_WRITE(AR5K_AR5212_DCU_MISC(queue),
sys/dev/ic/ar5212.c
1178
AR5K_REG_WRITE(AR5K_AR5212_DCU_MISC(queue),
sys/dev/ic/ar5212.c
1187
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1192
AR5K_REG_ENABLE_BITS(AR5K_AR5212_DCU_MISC(queue),
sys/dev/ic/ar5212.c
1198
AR5K_REG_WRITE(AR5K_AR5212_QCU_RDYTIMECFG(queue),
sys/dev/ic/ar5212.c
1206
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1211
AR5K_REG_ENABLE_BITS(AR5K_AR5212_DCU_MISC(queue),
sys/dev/ic/ar5212.c
1217
AR5K_REG_ENABLE_BITS(AR5K_AR5212_QCU_MISC(queue),
sys/dev/ic/ar5212.c
1241
ar5k_ar5212_get_tx_buf(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1243
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1248
return (AR5K_REG_READ(AR5K_AR5212_QCU_TXDP(queue)));
sys/dev/ic/ar5212.c
1252
ar5k_ar5212_put_tx_buf(struct ath_hal *hal, u_int queue, u_int32_t phys_addr)
sys/dev/ic/ar5212.c
1254
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1260
if (AR5K_REG_READ_Q(AR5K_AR5212_QCU_TXE, queue))
sys/dev/ic/ar5212.c
1263
AR5K_REG_WRITE(AR5K_AR5212_QCU_TXDP(queue), phys_addr);
sys/dev/ic/ar5212.c
1269
ar5k_ar5212_num_tx_pending(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1271
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1272
return (AR5K_AR5212_QCU_STS(queue) & AR5K_AR5212_QCU_STS_FRMPENDCNT);
sys/dev/ic/ar5212.c
1276
ar5k_ar5212_tx_start(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1278
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1281
if (AR5K_REG_READ_Q(AR5K_AR5212_QCU_TXD, queue))
sys/dev/ic/ar5212.c
1285
AR5K_REG_WRITE_Q(AR5K_AR5212_QCU_TXE, queue);
sys/dev/ic/ar5212.c
1291
ar5k_ar5212_stop_tx_dma(struct ath_hal *hal, u_int queue)
sys/dev/ic/ar5212.c
1295
AR5K_ASSERT_ENTRY(queue, hal->ah_capabilities.cap_queues.q_tx_num);
sys/dev/ic/ar5212.c
1300
AR5K_REG_WRITE_Q(AR5K_AR5212_QCU_TXD, queue);
sys/dev/ic/ar5212.c
1303
pending = AR5K_REG_READ(AR5K_AR5212_QCU_STS(queue)) &
sys/dev/ic/ar5212.c
984
u_int queue;
sys/dev/ic/ar5212.c
990
for (queue = HAL_TX_QUEUE_ID_DATA_MIN;
sys/dev/ic/ar5212.c
991
hal->ah_txq[queue].tqi_type != HAL_TX_QUEUE_INACTIVE;
sys/dev/ic/ar5212.c
992
queue++)
sys/dev/ic/ar5212.c
993
if (queue > HAL_TX_QUEUE_ID_DATA_MAX)
sys/dev/ic/ar5212.c
996
queue = HAL_TX_QUEUE_ID_PSPOLL;
sys/dev/ic/ar5212.c
998
queue = HAL_TX_QUEUE_ID_BEACON;
sys/dev/ic/ar5xxx.h
933
_t HAL_BOOL (_a _n##_setup_tx_queueprops)(struct ath_hal *, int queue, \
sys/dev/ic/ar5xxx.h
935
_t HAL_BOOL (_a _n##_release_tx_queue)(struct ath_hal *, u_int queue); \
sys/dev/ic/ar5xxx.h
936
_t HAL_BOOL (_a _n##_reset_tx_queue)(struct ath_hal *, u_int queue); \
sys/dev/ic/ar5xxx.h
937
_t u_int32_t (_a _n##_get_tx_buf)(struct ath_hal *, u_int queue); \
sys/dev/ic/ar5xxx.h
940
_t HAL_BOOL (_a _n##_tx_start)(struct ath_hal *, u_int queue); \
sys/dev/ic/ar5xxx.h
941
_t HAL_BOOL (_a _n##_stop_tx_dma)(struct ath_hal *, u_int queue); \
sys/dev/ic/qla.c
1307
qla_read_queue_2100(struct qla_softc *sc, bus_size_t queue)
sys/dev/ic/qla.c
1312
a = qla_read(sc, queue);
sys/dev/ic/qla.c
1313
b = qla_read(sc, queue);
sys/dev/ic/wdc.c
718
struct channel_queue *queue;
sys/dev/ic/wdc.c
730
queue = malloc(sizeof(*queue), M_DEVBUF, M_NOWAIT);
sys/dev/ic/wdc.c
731
if (queue != NULL) {
sys/dev/ic/wdc.c
732
TAILQ_INIT(&queue->sc_xfer);
sys/dev/ic/wdc.c
734
return (queue);
sys/dev/ic/wdc.c
738
wdc_free_queue(struct channel_queue *queue)
sys/dev/ic/wdc.c
740
free(queue, M_DEVBUF, sizeof(*queue));
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gc_9_4_3.c
513
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gc_9_4_3.c
515
uint32_t reg_offset = get_sdma_rlc_reg_offset(adev, engine, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
1088
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
45
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
48
nv_grbm_select(adev, mec, pipe, queue, vmid);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.h
70
int engine, int queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10_3.c
45
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10_3.c
48
nv_grbm_select(adev, mec, pipe, queue, vmid);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
43
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
46
soc21_grbm_select(adev, mec, pipe, queue, vmid);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
804
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
31
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
34
soc24_grbm_select(adev, mec, pipe, queue, vmid);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
365
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
49
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
51
uint32_t value = PIPEID(pipe) | MEID(mec) | VMID(vmid) | QUEUEID(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
565
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
43
uint32_t queue, uint32_t vmid)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
45
uint32_t value = PIPEID(pipe) | MEID(mec) | VMID(vmid) | QUEUEID(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
586
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
1227
int engine, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
51
uint32_t queue, uint32_t vmid, uint32_t inst)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
54
soc15_grbm_select(adev, mec, pipe, queue, vmid, GET_INST(GC, inst));
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.h
116
int engine, int queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
107
queue = (*pos & GENMASK_ULL(53, 44)) >> 44;
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
142
amdgpu_gfx_select_me_pipe_q(adev, me, pipe, queue, vmid, 0);
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
273
rd->id.srbm.queue, rd->id.srbm.vmid, rd->id.xcc_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
357
rd->id.srbm.queue = v1_data.srbm.queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
79
unsigned int instance_bank, sh_bank, se_bank, me, pipe, queue, vmid;
sys/dev/pci/drm/amd/amdgpu/amdgpu_debugfs.c
82
instance_bank = sh_bank = se_bank = me = pipe = queue = vmid = 0;
sys/dev/pci/drm/amd/amdgpu/amdgpu_drv.c
2804
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_drv.c
2811
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
173
int queue = ring->queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
180
adev->gfx.num_gfx_rings > 1 && pipe == 1 && queue == 0) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
184
bit = amdgpu_gfx_me_queue_to_bit(adev, me, pipe, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
207
int i, j, queue, pipe;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
220
queue = (i / adev->gfx.mec.num_pipe_per_mec) %
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
223
set_bit(pipe * adev->gfx.mec.num_queue_per_pipe + queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
243
int i, queue, pipe;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
253
queue = (i / adev->gfx.me.num_pipe_per_me) %
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
256
set_bit(pipe * num_queue_per_pipe + queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
274
int mec, pipe, queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
284
amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
291
if ((mec == 1 && pipe > 1) || queue != 0)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
296
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
332
(unsigned char)ring->pipe, (unsigned char)ring->queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
48
int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
55
bit += queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
61
int *mec, int *pipe, int *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
610
int mec, pipe, queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
613
amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
615
set_resource_bit = mec * 4 * 8 + pipe * 8 + queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
63
*queue = bit % adev->gfx.mec.num_queue_per_pipe;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
689
kiq_ring->pipe, kiq_ring->queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
72
int xcc_id, int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
74
return test_bit(amdgpu_gfx_mec_queue_to_bit(adev, mec, pipe, queue),
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
79
int me, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
87
bit += queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
93
int me, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
95
return test_bit(amdgpu_gfx_me_queue_to_bit(adev, me, pipe, queue),
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
347
u32 queue, u32 vmid, u32 xcc_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
591
int pipe, int queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
593
int *mec, int *pipe, int *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
595
int mec, int pipe, int queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
601
int pipe, int queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_mes.c
317
queue_input.queue_id = ring->queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_mes.c
342
queue_input.queue_id = ring->queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_mes.c
369
queue_input.queue_id = ring->queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ring.h
374
u32 queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_umr.h
34
__u32 me, pipe, queue, vmid;
sys/dev/pci/drm/amd/amdgpu/amdgpu_umr.h
44
__u32 me, pipe, queue, vmid;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1000
amdgpu_userq_cleanup(userq_mgr, queue, queue_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1018
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
102
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1030
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1032
r = amdgpu_userq_preempt_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1034
r = amdgpu_userq_unmap_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1047
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1058
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
106
adev->userq_funcs[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1060
r = amdgpu_userq_restore_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1062
r = amdgpu_userq_map_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1076
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
109
if (queue->state == AMDGPU_USERQ_STATE_MAPPED) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1092
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1093
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1094
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1095
(queue->xcp_id == idx)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1096
r = amdgpu_userq_preempt_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
110
r = userq_funcs->preempt(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1111
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
112
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1126
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1127
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1128
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1129
(queue->xcp_id == idx)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
1130
r = amdgpu_userq_restore_helper(uqm, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
114
queue->state = AMDGPU_USERQ_STATE_PREEMPTED;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
123
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
127
adev->userq_funcs[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
130
if (queue->state == AMDGPU_USERQ_STATE_PREEMPTED) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
131
r = userq_funcs->restore(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
133
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
135
queue->state = AMDGPU_USERQ_STATE_MAPPED;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
144
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
148
adev->userq_funcs[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
151
if ((queue->state == AMDGPU_USERQ_STATE_MAPPED) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
152
(queue->state == AMDGPU_USERQ_STATE_PREEMPTED)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
153
r = userq_funcs->unmap(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
155
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
157
queue->state = AMDGPU_USERQ_STATE_UNMAPPED;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
164
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
168
adev->userq_funcs[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
171
if (queue->state == AMDGPU_USERQ_STATE_UNMAPPED) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
172
r = userq_funcs->map(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
174
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
176
queue->state = AMDGPU_USERQ_STATE_MAPPED;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
184
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
186
struct dma_fence *f = queue->last_fence;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
199
struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
203
const struct amdgpu_userq_funcs *uq_funcs = adev->userq_funcs[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
205
uq_funcs->mqd_destroy(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
206
amdgpu_userq_fence_driver_free(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
208
list_del(&queue->userq_va_list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
209
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
383
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
389
queue = amdgpu_userq_find(uq_mgr, queue_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
390
if (!queue) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
395
amdgpu_userq_wait_for_last_fence(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
396
r = amdgpu_bo_reserve(queue->db_obj.obj, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
398
amdgpu_bo_unpin(queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
399
amdgpu_bo_unreserve(queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
401
amdgpu_bo_unref(&queue->db_obj.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
404
debugfs_remove_recursive(queue->debugfs_queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
406
r = amdgpu_userq_unmap_helper(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
410
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
412
amdgpu_userq_cleanup(uq_mgr, queue, queue_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
439
struct amdgpu_usermode_queue *queue = m->private;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
443
if (!queue || !queue->mqd.obj)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
446
bo = amdgpu_bo_ref(queue->mqd.obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
453
seq_printf(m, "queue_type: %d\n", queue->queue_type);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
454
seq_printf(m, "mqd_gpu_address: 0x%llx\n", amdgpu_bo_gpu_offset(queue->mqd.obj));
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
47
static int amdgpu_userq_buffer_va_list_add(struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
483
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
522
queue = kzalloc(sizeof(struct amdgpu_usermode_queue), GFP_KERNEL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
523
if (!queue) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
529
INIT_LIST_HEAD(&queue->userq_va_list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
530
queue->doorbell_handle = args->in.doorbell_handle;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
531
queue->queue_type = args->in.ip_type;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
532
queue->vm = &fpriv->vm;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
533
queue->priority = priority;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
535
db_info.queue_type = queue->queue_type;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
536
db_info.doorbell_handle = queue->doorbell_handle;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
537
db_info.db_obj = &queue->db_obj;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
541
if (amdgpu_userq_input_va_validate(queue, args->in.queue_va, args->in.queue_size) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
542
amdgpu_userq_input_va_validate(queue, args->in.rptr_va, AMDGPU_GPU_PAGE_SIZE) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
543
amdgpu_userq_input_va_validate(queue, args->in.wptr_va, AMDGPU_GPU_PAGE_SIZE)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
545
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
553
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
558
queue->doorbell_index = index;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
559
xa_init_flags(&queue->fence_drv_xa, XA_FLAGS_ALLOC);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
560
r = amdgpu_userq_fence_driver_alloc(adev, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
566
r = uq_funcs->mqd_create(uq_mgr, &args->in, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
569
amdgpu_userq_fence_driver_free(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
570
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
574
qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
577
amdgpu_userq_fence_driver_free(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
578
uq_funcs->mqd_destroy(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
579
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
586
((queue->queue_type == AMDGPU_HW_IP_GFX) ||
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
587
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)))
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
592
r = amdgpu_userq_map_helper(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
596
amdgpu_userq_fence_driver_free(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
597
uq_funcs->mqd_destroy(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
598
kfree(queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
60
list_add(&va_cursor->list, &queue->userq_va_list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
611
queue->debugfs_queue = debugfs_create_dir(queue_name, filp->debugfs_client);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
612
debugfs_create_file("mqd_info", 0444, queue->debugfs_queue, queue, &amdgpu_mqd_info_fops);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
65
int amdgpu_userq_input_va_validate(struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
69
struct amdgpu_vm *vm = queue->vm;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
744
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
749
idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
750
r = amdgpu_userq_restore_helper(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
89
amdgpu_userq_buffer_va_list_add(queue, va_map, user_addr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
899
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
904
idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
905
r = amdgpu_userq_preempt_helper(uq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
918
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
921
idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
922
struct dma_fence *f = queue->last_fence;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
989
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
997
idr_for_each_entry(&userq_mgr->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
998
amdgpu_userq_wait_for_last_fence(userq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
999
amdgpu_userq_unmap_helper(userq_mgr, queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
147
int amdgpu_userq_input_va_validate(struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
81
struct amdgpu_usermode_queue *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
85
struct amdgpu_usermode_queue *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
87
struct amdgpu_usermode_queue *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
89
struct amdgpu_usermode_queue *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.h
91
struct amdgpu_usermode_queue *queue);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
390
static int amdgpu_userq_fence_read_wptr(struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
398
r = amdgpu_bo_reserve(queue->vm->root.bo, false);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
402
addr = queue->userq_prop->wptr_gpu_addr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
405
mapping = amdgpu_vm_bo_lookup_mapping(queue->vm, addr >> PAGE_SHIFT);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
407
amdgpu_bo_unreserve(queue->vm->root.bo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
413
amdgpu_bo_unreserve(queue->vm->root.bo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
488
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
575
queue = idr_find(&userq_mgr->userq_idr, args->queue_id);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
576
if (!queue) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
581
r = amdgpu_userq_fence_read_wptr(queue, &wptr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
593
r = amdgpu_userq_fence_create(queue, userq_fence, wptr, &fence);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
604
dma_fence_put(queue->last_fence);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
605
queue->last_fence = dma_fence_get(fence);
sys/dev/pci/drm/amd/amdgpu/cik.c
945
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/amd/amdgpu/cik.c
951
((queue << SRBM_GFX_CNTL__QUEUEID__SHIFT) & SRBM_GFX_CNTL__QUEUEID_MASK));
sys/dev/pci/drm/amd/amdgpu/cik.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
3752
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4650
int me, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4660
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4670
snprintf(ring->name, sizeof(ring->name), "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4680
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4691
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
4699
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
6732
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
6738
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
6854
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
6871
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
7148
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
7157
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
7179
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9219
(ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9367
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9376
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9540
tmp = REG_SET_FIELD(tmp, CP_VMID_RESET, PIPE0_QUEUES, 1 << ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9542
tmp = REG_SET_FIELD(tmp, CP_VMID_RESET, PIPE1_QUEUES, 1 << ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v10_0.c
9612
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1125
int me, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1135
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1149
snprintf(ring->name, sizeof(ring->name), "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1159
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1171
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
1179
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
396
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
4049
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
4192
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
4493
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
4502
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
4524
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
6484
(ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
6634
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
6643
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
6777
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v11_0.c
6869
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
1005
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
1013
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
2944
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
3070
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
335
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
3371
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
3380
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
3402
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
4860
(ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
5010
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
5019
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
5250
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
5341
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
962
int me, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
972
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
982
snprintf(ring->name, sizeof(ring->name), "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v12_0.c
993
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v6_0.c
3095
ring->queue = i;
sys/dev/pci/drm/amd/amdgpu/gfx_v6_0.c
3096
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v7_0.c
2984
cik_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v7_0.c
4278
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v7_0.c
4287
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v7_0.c
4292
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
1847
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
1859
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
1866
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4300
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4351
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4599
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4610
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4634
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
4981
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
5076
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6573
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6601
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
2153
int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
2164
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
2172
snprintf(ring->name, sizeof(ring->name), "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
3528
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
3854
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
3865
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
3896
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
4081
adev->gfx.kiq[0].ring.queue, 0, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
6228
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
6256
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
7211
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_0.c
963
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
1808
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
212
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
2133
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
2144
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
2175
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
2206
ring->queue, 0, GET_INST(GC, xcc_id));
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
2339
adev->gfx.kiq[xcc_id].ring.queue, 0,
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3326
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3359
ring->queue == queue_id)
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3471
uint32_t pipe, uint32_t queue,
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3478
soc15_grbm_select(adev, me, pipe, queue, 0, GET_INST(GC, xcc_id));
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3556
r = gfx_v9_4_3_unmap_done(adev, ring->me, ring->pipe, ring->queue, ring->xcc_id);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
3596
r = gfx_v9_4_3_unmap_done(adev, ring->me, ring->pipe, ring->queue, ring->xcc_id);
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
971
int xcc_id, int mec, int pipe, int queue)
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
985
ring->queue = queue;
sys/dev/pci/drm/amd/amdgpu/gfx_v9_4_3.c
997
ring->xcc_id, ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
116
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
119
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
120
struct amdgpu_mqd_prop *userq_props = queue->userq_prop;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
137
queue_input.gang_global_priority_level = convert_to_mes_priority(queue->priority);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
139
queue_input.process_id = queue->vm->pasid;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
140
queue_input.queue_type = queue->queue_type;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
141
queue_input.mqd_addr = queue->mqd.gpu_addr;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
145
queue_input.page_table_base_addr = amdgpu_gmc_pd_addr(queue->vm->root.bo);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
146
queue_input.wptr_mc_addr = queue->wptr_obj.gpu_addr;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
161
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
165
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
169
queue_input.doorbell_offset = queue->doorbell_index;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
181
struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
184
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
207
struct amdgpu_usermode_queue *queue;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
231
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
232
if (queue->queue_type == queue_type) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
234
if (queue->doorbell_index == db_array[i]) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
235
queue->state = AMDGPU_USERQ_STATE_HUNG;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
237
amdgpu_userq_fence_driver_force_completion(queue);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
251
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
254
struct amdgpu_mqd *mqd_hw_default = &adev->mqds[queue->queue_type];
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
266
r = amdgpu_userq_create_object(uq_mgr, &queue->mqd, mqd_hw_default->mqd_size);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
277
userq_props->mqd_gpu_addr = queue->mqd.gpu_addr;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
279
userq_props->doorbell_index = queue->doorbell_index;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
280
userq_props->fence_address = queue->fence_drv->gpu_addr;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
282
if (queue->queue_type == AMDGPU_HW_IP_COMPUTE) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
298
r = amdgpu_userq_input_va_validate(queue, compute_mqd->eop_va,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
310
} else if (queue->queue_type == AMDGPU_HW_IP_GFX) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
339
r = amdgpu_userq_input_va_validate(queue, mqd_gfx_v11->shadow_va,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
343
r = amdgpu_userq_input_va_validate(queue, mqd_gfx_v11->csa_va,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
349
} else if (queue->queue_type == AMDGPU_HW_IP_DMA) {
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
364
r = amdgpu_userq_input_va_validate(queue, mqd_sdma_v11->csa_va,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
373
queue->userq_prop = userq_props;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
375
r = mqd_hw_default->init_mqd(adev, (void *)queue->mqd.cpu_ptr, userq_props);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
382
r = mes_userq_create_ctx_space(uq_mgr, queue, mqd_user);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
389
r = mes_userq_create_wptr_mapping(uq_mgr, queue, userq_props->wptr_gpu_addr);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
398
amdgpu_userq_destroy_object(uq_mgr, &queue->fw_obj);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
401
amdgpu_userq_destroy_object(uq_mgr, &queue->mqd);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
411
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
413
amdgpu_userq_destroy_object(uq_mgr, &queue->fw_obj);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
414
kfree(queue->userq_prop);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
415
amdgpu_userq_destroy_object(uq_mgr, &queue->mqd);
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
419
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
423
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
430
if (queue->state != AMDGPU_USERQ_STATE_MAPPED)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
465
struct amdgpu_usermode_queue *queue)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
469
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
472
if (queue->state == AMDGPU_USERQ_STATE_HUNG)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
474
if (queue->state != AMDGPU_USERQ_STATE_PREEMPTED)
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
63
struct amdgpu_usermode_queue *queue,
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
68
struct amdgpu_userq_obj *wptr_obj = &queue->wptr_obj;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
71
wptr_vm = queue->vm;
sys/dev/pci/drm/amd/amdgpu/mes_userqueue.c
96
queue->wptr_obj.gpu_addr = amdgpu_bo_gpu_offset_no_check(wptr_obj->obj);
sys/dev/pci/drm/amd/amdgpu/mes_v11_0.c
1335
ring->queue = 0;
sys/dev/pci/drm/amd/amdgpu/mes_v11_0.c
1342
snprintf(ring->name, sizeof(ring->name), "mes_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_v11_0.c
1358
ring->queue = 0;
sys/dev/pci/drm/amd/amdgpu/mes_v11_0.c
1367
ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_v11_0.c
1550
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_v12_0.c
1523
ring->queue = 0;
sys/dev/pci/drm/amd/amdgpu/mes_v12_0.c
1529
snprintf(ring->name, sizeof(ring->name), "mes_%d.%d.%d", ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_v12_0.c
1550
ring->queue = 0;
sys/dev/pci/drm/amd/amdgpu/mes_v12_0.c
1559
ring->me, ring->pipe, ring->queue);
sys/dev/pci/drm/amd/amdgpu/mes_v12_0.c
1732
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
sys/dev/pci/drm/amd/amdgpu/nv.c
318
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/amd/amdgpu/nv.c
324
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
sys/dev/pci/drm/amd/amdgpu/nv.h
32
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdgpu/sdma_v5_0.c
1703
struct amdgpu_mes_queue *queue;
sys/dev/pci/drm/amd/amdgpu/sdma_v5_0.c
1708
queue = idr_find(&adev->mes.queue_id_idr, mes_queue_id);
sys/dev/pci/drm/amd/amdgpu/sdma_v5_0.c
1709
if (queue) {
sys/dev/pci/drm/amd/amdgpu/sdma_v5_0.c
1711
amdgpu_fence_process(queue->ring);
sys/dev/pci/drm/amd/amdgpu/sdma_v5_2.c
1609
struct amdgpu_mes_queue *queue;
sys/dev/pci/drm/amd/amdgpu/sdma_v5_2.c
1614
queue = idr_find(&adev->mes.queue_id_idr, mes_queue_id);
sys/dev/pci/drm/amd/amdgpu/sdma_v5_2.c
1615
if (queue) {
sys/dev/pci/drm/amd/amdgpu/sdma_v5_2.c
1617
amdgpu_fence_process(queue->ring);
sys/dev/pci/drm/amd/amdgpu/sdma_v6_0.c
1614
int instances, queue;
sys/dev/pci/drm/amd/amdgpu/sdma_v6_0.c
1618
queue = entry->ring_id & 0xf;
sys/dev/pci/drm/amd/amdgpu/sdma_v6_0.c
1627
switch (queue) {
sys/dev/pci/drm/amd/amdgpu/sdma_v7_0.c
1547
int instances, queue;
sys/dev/pci/drm/amd/amdgpu/sdma_v7_0.c
1551
queue = entry->ring_id & 0xf;
sys/dev/pci/drm/amd/amdgpu/sdma_v7_0.c
1560
switch (queue) {
sys/dev/pci/drm/amd/amdgpu/si.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdgpu/soc15.c
364
u32 me, u32 pipe, u32 queue, u32 vmid, int xcc_id)
sys/dev/pci/drm/amd/amdgpu/soc15.c
370
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
sys/dev/pci/drm/amd/amdgpu/soc15.h
110
u32 me, u32 pipe, u32 queue, u32 vmid, int xcc_id);
sys/dev/pci/drm/amd/amdgpu/soc21.c
239
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/amd/amdgpu/soc21.c
245
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
sys/dev/pci/drm/amd/amdgpu/soc21.h
29
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdgpu/soc24.c
102
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/amd/amdgpu/soc24.c
108
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
sys/dev/pci/drm/amd/amdgpu/soc24.h
29
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdgpu/vi.c
579
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/amd/amdgpu/vi.c
585
srbm_gfx_cntl = REG_SET_FIELD(srbm_gfx_cntl, SRBM_GFX_CNTL, QUEUEID, queue);
sys/dev/pci/drm/amd/amdgpu/vi.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
sys/dev/pci/drm/amd/amdkfd/kfd_chardev.c
1464
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
299
static int kfd_dbg_set_queue_workaround(struct queue *q, bool enable)
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
950
struct queue *queue = NULL;
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
957
list_for_each_entry(queue, &qpd->queues_list, list) {
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
958
if (!found && queue->properties.queue_id == source_id) {
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
972
if (!(queue->properties.exception_status & KFD_EC_MASK(exception_code))) {
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
976
exception_status_ptr = &queue->properties.exception_status;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1043
q->pipe, q->queue,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1062
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1108
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1141
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1175
KFD_UNMAP_LATENCY_MS, q->pipe, q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1191
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1254
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1314
q->queue, &q->properties, mm);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1334
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1501
int pipe, queue;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1519
for (queue = 0; queue < get_queues_per_pipe(dqm); queue++)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1520
if (test_bit(pipe_offset + queue,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1522
dqm->allocated_queues[pipe] |= 1 << queue;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1576
struct queue *q, const uint32_t *restore_sdma_id)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
162
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1683
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1937
increment_queue_count(dqm, qpd, kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1952
decrement_queue_count(dqm, qpd, kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
1966
static int create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
200
static int add_queue_mes(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2128
static void set_queue_as_reset(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2154
uint32_t mec, pipe, queue;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2163
amdgpu_queue_mask_bit_to_mec_queue(dqm->dev->adev, i, &mec, &pipe, &queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2167
dqm->dev->adev, pipe, queue, xcc_id);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2174
hang_info.queue_id = queue;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2186
static struct queue *find_queue_by_address(struct device_queue_manager *dqm, uint64_t queue_address)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2190
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2216
struct queue *q = find_queue_by_address(dqm, hang_info.queue_address);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2280
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2421
if (mqd_mgr->check_preemption_failed(mqd_mgr, dqm->packet_mgr.priv_queue->queue->mqd) &&
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2481
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2510
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2645
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2657
q = list_first_entry(&qpd->queues_list, struct queue, list);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2691
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
272
static int remove_queue_mes(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2721
const struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2741
const struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2774
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2790
decrement_queue_count(dqm, qpd, kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
2843
q = list_first_entry(&qpd->queues_list, struct queue, list);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3063
struct queue *q = NULL;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
308
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3247
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3326
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
335
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3429
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3542
void set_queue_snapshot_entry(struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3617
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3665
int pipe, queue;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3695
for (queue = 0; queue < get_queues_per_pipe(dqm); queue++) {
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3696
if (!test_bit(pipe_offset + queue,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3701
pipe, queue,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3709
xcc_id, pipe, queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3721
for (queue = 0;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3722
queue < dqm->dev->kfd->device_info.num_sdma_queues_per_engine;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3723
queue++) {
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3725
dqm->dev->adev, pipe, queue, &dump, &n_regs);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3730
pipe, queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
399
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
414
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
432
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
503
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
532
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
601
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
621
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
65
struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
661
q->pipe, q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
68
struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
69
static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
703
q->queue, &q->properties, current->mm);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
71
struct queue *q, const uint32_t *restore_sdma_id);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
741
static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
759
q->queue = bit;
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
768
pr_debug("hqd slot - pipe %d, queue %d\n", q->pipe, q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
776
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
778
dqm->allocated_queues[q->pipe] |= (1 << q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
854
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
885
q->pipe, q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
913
struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
943
static int update_queue(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
996
KFD_UNMAP_LATENCY_MS, q->pipe, q->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
138
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
146
struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
149
struct queue *q, struct mqd_update_info *minfo);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
188
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
196
const struct queue *q, u32 *mqd_size,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
200
const struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
216
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.h
321
void set_queue_snapshot_entry(struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
160
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
48
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v10.c
32
static void init_sdma_vm_v10(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v10.c
85
static void init_sdma_vm_v10(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v11.c
31
static void init_sdma_vm_v11(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v11.c
85
static void init_sdma_vm_v11(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v12.c
31
static void init_sdma_vm_v12(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v12.c
85
static void init_sdma_vm_v12(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v9.c
130
static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_v9.c
31
static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
162
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
48
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
140
if (init_queue(&kq->queue, &prop) != 0)
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
143
kq->queue->device = dev;
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
144
kq->queue->process = kfd_get_process(current);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
146
kq->queue->mqd_mem_obj = kq->mqd_mgr->allocate_mqd(kq->mqd_mgr->dev,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
147
&kq->queue->properties);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
148
if (!kq->queue->mqd_mem_obj)
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
150
kq->mqd_mgr->init_mqd(kq->mqd_mgr, &kq->queue->mqd,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
151
kq->queue->mqd_mem_obj,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
152
&kq->queue->gart_mqd_addr,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
153
&kq->queue->properties);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
157
kq->queue->pipe = KFD_CIK_HIQ_PIPE;
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
158
kq->queue->queue = KFD_CIK_HIQ_QUEUE;
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
159
kq->mqd_mgr->load_mqd(kq->mqd_mgr, kq->queue->mqd,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
160
kq->queue->pipe, kq->queue->queue,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
161
&kq->queue->properties, NULL);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
175
print_queue(kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
179
kq->mqd_mgr->free_mqd(kq->mqd_mgr, kq->queue->mqd, kq->queue->mqd_mem_obj);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
181
uninit_queue(kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
200
if (kq->queue->properties.type == KFD_QUEUE_TYPE_HIQ && down_read_trylock(&kq->dev->adev->reset_domain->sem)) {
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
202
kq->queue->mqd,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
205
kq->queue->pipe,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
206
kq->queue->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
209
else if (kq->queue->properties.type == KFD_QUEUE_TYPE_DIQ)
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
212
kq->mqd_mgr->free_mqd(kq->mqd_mgr, kq->queue->mqd,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
213
kq->queue->mqd_mem_obj);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
225
kq->queue->properties.doorbell_ptr);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
226
uninit_queue(kq->queue);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
247
queue_size_dwords = kq->queue->properties.queue_size / 4;
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
312
write_kernel_doorbell64(kq->queue->properties.doorbell_ptr,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
317
write_kernel_doorbell(kq->queue->properties.doorbell_ptr,
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.c
329
(kq->queue->properties.queue_size / 4);
sys/dev/pci/drm/amd/amdkfd/kfd_kernel_queue.h
58
struct queue *queue;
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager.c
148
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager.c
204
if (!kq->queue->properties.is_active)
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager.c
209
kq->queue->queue, qpd->is_debug);
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager.c
213
kq->queue,
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager.c
229
q->queue, qpd->is_debug);
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager_v9.c
228
struct queue *q, bool is_static)
sys/dev/pci/drm/amd/amdkfd/kfd_packet_manager_vi.c
144
struct queue *q, bool is_static)
sys/dev/pci/drm/amd/amdkfd/kfd_pm4_headers_vi.h
252
uint32_t queue:6;
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1141
int kfd_procfs_add_queue(struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1142
void kfd_procfs_del_queue(struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1318
int init_queue(struct queue **q, const struct queue_properties *properties);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1319
void uninit_queue(struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1321
void print_queue(struct queue *q);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1354
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1378
struct queue *pqm_get_user_queue(struct process_queue_manager *pqm,
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
1443
struct queue *q, bool is_static);
sys/dev/pci/drm/amd/amdkfd/kfd_priv.h
613
uint32_t queue;
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
106
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2223
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
407
struct queue *q = container_of(kobj, struct queue, kobj);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
521
int kfd_procfs_add_queue(struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
681
void kfd_procfs_del_queue(struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
1110
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
1143
q = pqn->kq->queue;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
245
struct kfd_node *dev, struct queue **q,
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
318
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
40
(pqn->kq && pqn->kq->queue->properties.queue_id == qid))
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
435
kq->queue->properties.queue_id = *qid;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
594
struct queue *q = pqn->q;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
676
struct queue *pqm_get_user_queue(struct process_queue_manager *pqm,
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
748
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
769
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
827
struct queue *q,
sys/dev/pci/drm/amd/amdkfd/kfd_process_queue_manager.c
880
struct queue *q;
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
48
void print_queue(struct queue *q)
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
69
int init_queue(struct queue **q, const struct queue_properties *properties)
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
71
struct queue *tmp_q;
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
83
void uninit_queue(struct queue *q)
sys/dev/pci/drm/amd/include/kgd_kfd_interface.h
339
int engine, int queue);
sys/dev/pci/drm/drm_atomic_helper.c
1919
ret = wait_event_timeout(dev->vblank[i].queue,
sys/dev/pci/drm/drm_vblank.c
1320
ret = wait_event_timeout(vblank->queue,
sys/dev/pci/drm/drm_vblank.c
1382
wake_up(&vblank->queue);
sys/dev/pci/drm/drm_vblank.c
1857
wait = wait_event_interruptible_timeout(vblank->queue,
sys/dev/pci/drm/drm_vblank.c
1965
wake_up(&vblank->queue);
sys/dev/pci/drm/drm_vblank.c
545
init_waitqueue_head(&vblank->queue);
sys/dev/pci/drm/drm_vblank.c
599
return &crtc->dev->vblank[drm_crtc_index(crtc)].queue;
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1489
while ((rb = rb_first_cached(&sched_engine->queue))) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
1568
rb_erase_cached(&p->node, &sched_engine->queue);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
287
rb = rb_first_cached(&sched_engine->queue);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3182
while ((rb = rb_first_cached(&sched_engine->queue))) {
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3192
rb_erase_cached(&p->node, &sched_engine->queue);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3226
sched_engine->queue = RB_ROOT_CACHED;
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
4123
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
sys/dev/pci/drm/i915/gt/intel_gt_debugfs.c
36
wait_event(gt->reset.queue,
sys/dev/pci/drm/i915/gt/intel_reset.c
1543
wait_event(gt->reset.queue,
sys/dev/pci/drm/i915/gt/intel_reset.c
1577
wake_up_all(&gt->reset.queue);
sys/dev/pci/drm/i915/gt/intel_reset.c
1596
if (wait_event_interruptible(gt->reset.queue,
sys/dev/pci/drm/i915/gt/intel_reset.c
1636
if (wait_event_interruptible(gt->reset.queue,
sys/dev/pci/drm/i915/gt/intel_reset.c
1666
init_waitqueue_head(&gt->reset.queue);
sys/dev/pci/drm/i915/gt/intel_reset_types.h
55
wait_queue_head_t queue;
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2002
while ((rb = rb_first_cached(&sched_engine->queue))) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2013
rb_erase_cached(&p->node, &sched_engine->queue);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
2020
sched_engine->queue = RB_ROOT_CACHED;
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
5513
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
953
while ((rb = rb_first_cached(&sched_engine->queue))) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
984
rb_erase_cached(&p->node, &sched_engine->queue);
sys/dev/pci/drm/i915/i915_scheduler.c
118
rb_insert_color_cached(&p->node, &sched_engine->queue, first);
sys/dev/pci/drm/i915/i915_scheduler.c
47
GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
sys/dev/pci/drm/i915/i915_scheduler.c
478
sched_engine->queue = RB_ROOT_CACHED;
sys/dev/pci/drm/i915/i915_scheduler.c
48
rb_first(&sched_engine->queue.rb_root));
sys/dev/pci/drm/i915/i915_scheduler.c
51
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
sys/dev/pci/drm/i915/i915_scheduler.c
75
parent = &sched_engine->queue.rb_root.rb_node;
sys/dev/pci/drm/i915/i915_scheduler.h
71
return RB_EMPTY_ROOT(&sched_engine->queue.rb_root);
sys/dev/pci/drm/i915/i915_scheduler_types.h
154
struct rb_root_cached queue;
sys/dev/pci/drm/i915/selftests/igt_reset.c
22
wait_event(gt->reset.queue,
sys/dev/pci/drm/i915/selftests/igt_reset.c
42
wake_up_all(&gt->reset.queue);
sys/dev/pci/drm/include/drm/drm_vblank.h
124
wait_queue_head_t queue;
sys/dev/pci/drm/include/drm/spsc_queue.h
104
WRITE_ONCE(queue->head, next);
sys/dev/pci/drm/include/drm/spsc_queue.h
109
if (atomic_long_cmpxchg(&queue->tail,
sys/dev/pci/drm/include/drm/spsc_queue.h
110
(long)&node->next, (long) &queue->head) != (long)&node->next) {
sys/dev/pci/drm/include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
sys/dev/pci/drm/include/drm/spsc_queue.h
118
atomic_dec(&queue->job_count);
sys/dev/pci/drm/include/drm/spsc_queue.h
48
static inline void spsc_queue_init(struct spsc_queue *queue)
sys/dev/pci/drm/include/drm/spsc_queue.h
50
queue->head = NULL;
sys/dev/pci/drm/include/drm/spsc_queue.h
51
atomic_long_set(&queue->tail, (long)&queue->head);
sys/dev/pci/drm/include/drm/spsc_queue.h
52
atomic_set(&queue->job_count, 0);
sys/dev/pci/drm/include/drm/spsc_queue.h
55
static inline struct spsc_node *spsc_queue_peek(struct spsc_queue *queue)
sys/dev/pci/drm/include/drm/spsc_queue.h
57
return queue->head;
sys/dev/pci/drm/include/drm/spsc_queue.h
60
static inline int spsc_queue_count(struct spsc_queue *queue)
sys/dev/pci/drm/include/drm/spsc_queue.h
62
return atomic_read(&queue->job_count);
sys/dev/pci/drm/include/drm/spsc_queue.h
65
static inline bool spsc_queue_push(struct spsc_queue *queue, struct spsc_node *node)
sys/dev/pci/drm/include/drm/spsc_queue.h
73
atomic_inc(&queue->job_count);
sys/dev/pci/drm/include/drm/spsc_queue.h
76
tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next);
sys/dev/pci/drm/include/drm/spsc_queue.h
87
return tail == &queue->head;
sys/dev/pci/drm/include/drm/spsc_queue.h
91
static inline struct spsc_node *spsc_queue_pop(struct spsc_queue *queue)
sys/dev/pci/drm/include/drm/spsc_queue.h
98
node = READ_ONCE(queue->head);
sys/dev/pci/drm/radeon/cik.c
1842
u32 me, u32 pipe, u32 queue, u32 vmid)
sys/dev/pci/drm/radeon/cik.c
1847
QUEUEID(queue & 0x7));
sys/dev/pci/drm/radeon/cik.c
4148
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/radeon/cik.c
4167
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/radeon/cik.c
4189
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
sys/dev/pci/drm/radeon/cik.c
4607
rdev->ring[idx].queue, 0);
sys/dev/pci/drm/radeon/cik.c
8414
ring->queue = 0; /* first queue */
sys/dev/pci/drm/radeon/cik.c
8426
ring->queue = 1; /* second queue */
sys/dev/pci/drm/radeon/radeon.h
829
u32 queue;
sys/dev/pci/if_aq_pci.c
2846
int bits, queue;
sys/dev/pci/if_aq_pci.c
2889
queue = 0;
sys/dev/pci/if_aq_pci.c
2892
redir |= (queue << bits);
sys/dev/pci/if_aq_pci.c
2894
queue++;
sys/dev/pci/if_aq_pci.c
2895
if (queue == sc->sc_nqueues)
sys/dev/pci/if_aq_pci.c
2896
queue = 0;
sys/dev/pci/if_ice.c
26749
uint16_t queue = (reg & GL_MDET_TX_TCLAN_QNUM_M) >>
sys/dev/pci/if_ice.c
26755
queue, pf_num, vf_num);
sys/dev/pci/if_ice.c
26775
uint16_t queue = (reg & GL_MDET_TX_PQM_QNUM_M) >>
sys/dev/pci/if_ice.c
26780
sc->sc_dev.dv_xname, ice_mdd_tx_pqm_str(event), queue,
sys/dev/pci/if_ice.c
26800
uint16_t queue = (reg & GL_MDET_RX_QNUM_M) >>
sys/dev/pci/if_ice.c
26805
ice_mdd_rx_str(event), queue, pf_num, vf_num);
sys/dev/pci/if_ix.c
1035
uint64_t queue = 1ULL << vector;
sys/dev/pci/if_ix.c
1039
mask = (IXGBE_EIMS_RTX_QUEUE & queue);
sys/dev/pci/if_ix.c
1042
mask = (queue & 0xFFFFFFFF);
sys/dev/pci/if_ix.c
1045
mask = (queue >> 32);
sys/dev/pci/if_ix.c
1064
uint64_t queue = 1ULL << vector;
sys/dev/pci/if_ix.c
1068
mask = (IXGBE_EIMS_RTX_QUEUE & queue);
sys/dev/pci/if_ix.c
1071
mask = (queue & 0xFFFFFFFF);
sys/dev/pci/if_ix.c
1074
mask = (queue >> 32);
sys/dev/pci/if_ixl.c
435
uint16_t queue;
sys/dev/pci/if_ixv.c
537
uint32_t queue = 1 << vector;
sys/dev/pci/if_ixv.c
540
mask = (IXGBE_EIMS_RTX_QUEUE & queue);
sys/dev/pci/if_ixv.c
548
uint64_t queue = (1ULL << vector);
sys/dev/pci/if_ixv.c
551
mask = (IXGBE_EIMS_RTX_QUEUE & queue);
sys/dev/pci/if_ngbe.c
2459
uint64_t queue = 1ULL << vector;
sys/dev/pci/if_ngbe.c
2462
mask = (queue & 0xffffffff);
sys/dev/pci/if_ngbe.c
2572
uint64_t queue = 1ULL << vector;
sys/dev/pci/if_ngbe.c
2575
mask = (queue & 0xffffffff);
sys/dev/pci/if_vmx.c
129
unsigned int queue;
sys/dev/pci/if_vmx.c
1368
int queue;
sys/dev/pci/if_vmx.c
1378
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
1379
intr_barrier(sc->sc_q[queue].ih);
sys/dev/pci/if_vmx.c
1383
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
1384
vmxnet3_txstop(sc, &sc->sc_q[queue].tx);
sys/dev/pci/if_vmx.c
1385
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
1386
vmxnet3_rxstop(sc, &sc->sc_q[queue].rx);
sys/dev/pci/if_vmx.c
1412
int queue;
sys/dev/pci/if_vmx.c
1424
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
1425
vmxnet3_txinit(sc, &sc->sc_q[queue].tx);
sys/dev/pci/if_vmx.c
1426
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
1427
vmxnet3_rxinit(sc, &sc->sc_q[queue].rx);
sys/dev/pci/if_vmx.c
1429
for (queue = 0; queue < sc->sc_nqueues; queue++) {
sys/dev/pci/if_vmx.c
1430
WRITE_BAR0(sc, VMXNET3_BAR0_RXH1(queue), 0);
sys/dev/pci/if_vmx.c
1431
WRITE_BAR0(sc, VMXNET3_BAR0_RXH2(queue), 0);
sys/dev/pci/if_vmx.c
1765
WRITE_BAR0(sc, VMXNET3_BAR0_TXH(tq->queue), prod);
sys/dev/pci/if_vmx.c
491
int i, queue, qs_len, intr;
sys/dev/pci/if_vmx.c
498
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
499
sc->sc_q[queue].tx.ts = ts++;
sys/dev/pci/if_vmx.c
501
for (queue = 0; queue < sc->sc_nqueues; queue++)
sys/dev/pci/if_vmx.c
502
sc->sc_q[queue].rx.rs = rs++;
sys/dev/pci/if_vmx.c
504
for (queue = 0; queue < sc->sc_nqueues; queue++) {
sys/dev/pci/if_vmx.c
505
intr = sc->sc_q[queue].intr;
sys/dev/pci/if_vmx.c
507
if (vmxnet3_alloc_txring(sc, queue, intr))
sys/dev/pci/if_vmx.c
509
if (vmxnet3_alloc_rxring(sc, queue, intr))
sys/dev/pci/if_vmx.c
597
vmxnet3_alloc_txring(struct vmxnet3_softc *sc, int queue, int intr)
sys/dev/pci/if_vmx.c
599
struct vmxnet3_txqueue *tq = &sc->sc_q[queue].tx;
sys/dev/pci/if_vmx.c
605
tq->queue = queue;
sys/dev/pci/if_vmx.c
640
vmxnet3_alloc_rxring(struct vmxnet3_softc *sc, int queue, int intr)
sys/dev/pci/if_vmx.c
642
struct vmxnet3_rxqueue *rq = &sc->sc_q[queue].rx;
sys/dev/pci/if_vmx.c
676
VMXNET3_BAR0_RXH1(queue) : VMXNET3_BAR0_RXH2(queue);
sys/dev/usb/dwc2/dwc2.c
605
return dwc2_root_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2.c
780
return dwc2_root_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2.c
829
return dwc2_device_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2.c
881
return dwc2_device_bulk_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2.c
928
return dwc2_device_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2.c
979
return dwc2_device_isoc_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/dwc2/dwc2_core.h
171
struct list_head queue;
sys/dev/usb/dwc2/dwc2_core.h
216
struct list_head queue;
sys/dev/usb/ehci.c
1790
return (ehci_root_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ehci.c
2210
return (ehci_root_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ehci.c
2858
return (ehci_device_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ehci.c
3011
return (ehci_device_bulk_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ehci.c
3125
return (ehci_device_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ehci.c
747
if (xfer != SIMPLEQ_FIRST(&xfer->pipe->queue))
sys/dev/usb/ohci.c
2263
return (ohci_root_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ohci.c
2606
return (ohci_root_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ohci.c
2652
return (ohci_device_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ohci.c
2720
return (ohci_device_bulk_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ohci.c
2850
return (ohci_device_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/ohci.c
3065
ohci_device_isoc_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/uhci.c
1615
return (uhci_device_bulk_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/uhci.c
1798
return (uhci_device_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/uhci.c
1836
return (uhci_device_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/uhci.c
2108
uhci_device_isoc_start(SIMPLEQ_FIRST(&xfer->pipe->queue));
sys/dev/usb/uhci.c
2874
return (uhci_root_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/uhci.c
3250
return (uhci_root_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/upd.c
225
struct upd_sensor_head *queue)
sys/dev/usb/upd.c
238
entry->nchildren, entry->children, queue);
sys/dev/usb/upd.c
258
SLIST_INSERT_HEAD(queue, sensor, dep_next);
sys/dev/usb/upd.c
308
upd_request_children(struct upd_softc *sc, struct upd_sensor_head *queue)
sys/dev/usb/upd.c
314
SLIST_FOREACH(sensor, queue, dep_next) {
sys/dev/usb/usb_subr.c
806
SIMPLEQ_INIT(&p->queue);
sys/dev/usb/usbdi.c
153
SIMPLEQ_FOREACH(xfer, &pipe->queue, next) {
sys/dev/usb/usbdi.c
263
if (!SIMPLEQ_EMPTY(&pipe->queue))
sys/dev/usb/usbdi.c
569
while ((xfer = SIMPLEQ_FIRST(&pipe->queue)) != NULL) {
sys/dev/usb/usbdi.c
775
KASSERT(xfer == SIMPLEQ_FIRST(&pipe->queue));
sys/dev/usb/usbdi.c
776
SIMPLEQ_REMOVE_HEAD(&pipe->queue, next);
sys/dev/usb/usbdi.c
782
pipe->repeat, SIMPLEQ_FIRST(&pipe->queue)));
sys/dev/usb/usbdi.c
846
SIMPLEQ_INSERT_TAIL(&pipe->queue, xfer, next);
sys/dev/usb/usbdi.c
878
xfer = SIMPLEQ_FIRST(&pipe->queue);
sys/dev/usb/usbdivar.h
193
SIMPLEQ_HEAD(, usbd_xfer) queue;
sys/dev/usb/xhci.c
1088
skipxfer = SIMPLEQ_FIRST(&xp->pipe.queue);
sys/dev/usb/xhci.c
2441
return (xhci_root_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/xhci.c
2822
return (xhci_root_intr_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/xhci.c
2920
return (xhci_device_ctrl_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/xhci.c
3034
return (xhci_device_generic_start(SIMPLEQ_FIRST(&xfer->pipe->queue)));
sys/dev/usb/xhci.c
808
last = SIMPLEQ_FIRST(&xp->pipe.queue);
sys/dev/usb/xhci.c
815
xfer = SIMPLEQ_FIRST(&xp->pipe.queue);
sys/dev/wscons/wsevent.c
106
struct wscons_event *queue;
sys/dev/wscons/wsevent.c
111
queue = mallocarray(WSEVENT_QSIZE, sizeof(struct wscons_event),
sys/dev/wscons/wsevent.c
114
free(queue, M_DEVBUF,
sys/dev/wscons/wsevent.c
122
ev->ws_q = queue;
sys/kern/kern_sched.c
272
int queue = prio >> 2;
sys/kern/kern_sched.c
291
TAILQ_INSERT_TAIL(&spc->spc_qs[queue], p, p_runq);
sys/kern/kern_sched.c
292
spc->spc_whichqs |= (1U << queue);
sys/kern/kern_sched.c
305
int queue = p->p_runpri >> 2;
sys/kern/kern_sched.c
313
TAILQ_REMOVE(&spc->spc_qs[queue], p, p_runq);
sys/kern/kern_sched.c
314
if (TAILQ_EMPTY(&spc->spc_qs[queue])) {
sys/kern/kern_sched.c
315
spc->spc_whichqs &= ~(1U << queue);
sys/kern/kern_sched.c
326
int queue;
sys/kern/kern_sched.c
333
for (queue = 0; queue < SCHED_NQS; queue++) {
sys/kern/kern_sched.c
334
while ((p = TAILQ_FIRST(&spc->spc_qs[queue]))) {
sys/kern/kern_sched.c
356
queue = ffs(spc->spc_whichqs) - 1;
sys/kern/kern_sched.c
357
p = TAILQ_FIRST(&spc->spc_qs[queue]);
sys/kern/kern_sched.c
514
int queue;
sys/kern/kern_sched.c
521
queue = ffs(spc->spc_whichqs) - 1;
sys/kern/kern_sched.c
522
TAILQ_FOREACH(p, &spc->spc_qs[queue], p_runq) {
sys/kern/kern_softintr.c
176
struct softintr_queue *queue = &softintr_queue[sih->sih_level];
sys/kern/kern_softintr.c
183
TAILQ_INSERT_TAIL(queue, sih, sih_q);
sys/kern/kern_softintr.c
65
struct softintr_queue *queue = &softintr_queue[level];
sys/kern/kern_softintr.c
69
while ((sih = TAILQ_FIRST(queue)) != NULL) {
sys/kern/kern_softintr.c
75
TAILQ_REMOVE(queue, sih, sih_q);
sys/kern/kern_softintr.c
91
TAILQ_INSERT_TAIL(queue, sih, sih_q);
sys/kern/vfs_bio.c
1417
bufcache *cache, struct bufqueue *queue, int64_t *queuepages);
sys/kern/vfs_bio.c
1459
struct bufqueue * queue;
sys/kern/vfs_bio.c
1500
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1502
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1505
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1508
TAILQ_REMOVE(queue, bp, b_freelist);
sys/kern/vfs_bio.c
1543
struct bufqueue * queue;
sys/kern/vfs_bio.c
1597
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1599
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1602
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1605
TAILQ_REMOVE(queue, bp, b_freelist);
sys/kern/vfs_bio.c
1657
struct bufqueue *queue;
sys/kern/vfs_bio.c
1672
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1674
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1677
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1683
queue = &dirtyqueue;
sys/kern/vfs_bio.c
1687
TAILQ_REMOVE(queue, bp, b_freelist);
sys/kern/vfs_bio.c
1692
chillbufs(struct bufcache *cache, struct bufqueue *queue, int64_t *queuepages)
sys/kern/vfs_bio.c
1703
if (queue == &cache->hotqueue)
sys/kern/vfs_bio.c
1705
else if (queue == &cache->warmqueue)
sys/kern/vfs_bio.c
1711
bp = TAILQ_FIRST(queue);
sys/kern/vfs_bio.c
1716
TAILQ_REMOVE(queue, bp, b_freelist);
sys/kern/vfs_bio.c
1728
struct bufqueue *queue;
sys/kern/vfs_bio.c
1751
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1754
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1760
chillbufs(cache, queue, queuepages);
sys/kern/vfs_bio.c
1762
queue = &dirtyqueue;
sys/kern/vfs_bio.c
1766
TAILQ_INSERT_TAIL(queue, bp, b_freelist);
sys/net/pf_ioctl.c
2180
memcpy(&pq->queue, qs, sizeof(pq->queue));
sys/net/pf_ioctl.c
2212
memcpy(&pq->queue, qs, sizeof(pq->queue));
sys/net/pf_ioctl.c
2247
memcpy(qs, &q->queue, sizeof(*qs));
sys/net/pfvar.h
1465
struct pf_queuespec queue;
sys/net/pfvar.h
1471
struct pf_queuespec queue;
sys/net/route.c
1654
rt_timer_add(struct rtentry *rt, struct rttimer_queue *queue, u_int rtableid)
sys/net/route.c
1663
rnew->rtt_queue = queue;
sys/net/route.c
1665
rnew->rtt_expire = getuptime() + queue->rtq_timeout;
sys/net/route.c
1674
if (r->rtt_queue == queue) {
sys/net/route.c
1681
TAILQ_INSERT_TAIL(&queue->rtq_head, rnew, rtt_next);
sys/net/route.c
1682
timeout_add_sec(&rnew->rtt_timeout, queue->rtq_timeout);
sys/scsi/scsi_base.c
249
TAILQ_INIT(&iopl->queue);
sys/scsi/scsi_base.c
281
while ((ioh = TAILQ_FIRST(&iopl->queue)) != NULL) {
sys/scsi/scsi_base.c
282
TAILQ_REMOVE(&iopl->queue, ioh, q_entry);
sys/scsi/scsi_base.c
338
TAILQ_INSERT_TAIL(&iopl->queue, ioh, q_entry);
sys/scsi/scsi_base.c
366
TAILQ_REMOVE(&iopl->queue, ioh, q_entry);
sys/scsi/scsi_base.c
392
ioh = TAILQ_FIRST(&iopl->queue);
sys/scsi/scsi_base.c
394
TAILQ_REMOVE(&iopl->queue, ioh, q_entry);
sys/scsi/scsi_base.c
408
rv = !TAILQ_EMPTY(&iopl->queue);
sys/scsi/scsi_base.c
528
TAILQ_INSERT_TAIL(&link->queue, &xsh->ioh, q_entry);
sys/scsi/scsi_base.c
552
TAILQ_REMOVE(&link->queue, &xsh->ioh, q_entry);
sys/scsi/scsi_base.c
555
TAILQ_REMOVE(&link->pool->queue, &xsh->ioh, q_entry);
sys/scsi/scsi_base.c
587
((ioh = TAILQ_FIRST(&link->queue)) != NULL)) {
sys/scsi/scsi_base.c
590
TAILQ_REMOVE(&link->queue, ioh, q_entry);
sys/scsi/scsi_base.c
591
TAILQ_INSERT_TAIL(&link->pool->queue, ioh, q_entry);
sys/scsi/scsi_base.c
687
while ((ioh = TAILQ_FIRST(&link->queue)) != NULL) {
sys/scsi/scsi_base.c
688
TAILQ_REMOVE(&link->queue, ioh, q_entry);
sys/scsi/scsi_base.c
699
ioh = TAILQ_FIRST(&iopl->queue);
sys/scsi/scsi_base.c
712
TAILQ_REMOVE(&iopl->queue, &xsh->ioh, q_entry);
sys/scsi/scsiconf.c
508
TAILQ_INIT(&link->queue);
sys/scsi/scsiconf.h
260
struct scsi_runq queue;
sys/scsi/scsiconf.h
319
struct scsi_runq queue;
usr.bin/at/at.c
1054
writefile(cwd, timer, queue);
usr.bin/at/at.c
1059
queue = toupper((unsigned char)queue);
usr.bin/at/at.c
1061
queue = DEFAULT_BATCH_QUEUE;
usr.bin/at/at.c
1068
writefile(cwd, timer, queue);
usr.bin/at/at.c
159
newjob(time_t runtimer, int queue)
usr.bin/at/at.c
170
(long long)runtimer, queue);
usr.bin/at/at.c
184
writefile(const char *cwd, time_t runtimer, char queue)
usr.bin/at/at.c
216
if ((fd = newjob(runtimer, queue)) == -1)
usr.bin/at/at.c
391
return (j1->queue - j2->queue);
usr.bin/at/at.c
409
job->queue, timestr);
usr.bin/at/at.c
420
(long long)job->runtimer, job->queue, job->queue,
usr.bin/at/at.c
442
char queue, *ep;
usr.bin/at/at.c
517
queue = *(ep + 1);
usr.bin/at/at.c
519
if (atqueue && (queue != atqueue))
usr.bin/at/at.c
555
job->queue = queue;
usr.bin/at/at.c
866
char queue = DEFAULT_AT_QUEUE;
usr.bin/at/at.c
937
atqueue = queue = *optarg;
usr.bin/at/at.c
938
if (!(islower((unsigned char)queue) ||
usr.bin/at/at.c
939
isupper((unsigned char)queue)))
usr.bin/at/at.h
35
char queue;
usr.bin/dig/lib/isc/task.c
468
isc_tasklist_t queue;
usr.bin/dig/lib/isc/task.c
471
queue = manager->ready_tasks;
usr.bin/dig/lib/isc/task.c
473
queue = manager->ready_priority_tasks;
usr.bin/dig/lib/isc/task.c
475
return (EMPTY(queue));
usr.bin/dig/lib/isc/unix/socket.c
2265
goto queue;
usr.bin/dig/lib/isc/unix/socket.c
2312
queue:
usr.bin/grep/queue.c
40
struct queue *next;
usr.bin/ssh/mux.c
1545
struct sshbuf *queue;
usr.bin/ssh/mux.c
1553
if ((queue = sshbuf_new()) == NULL)
usr.bin/ssh/mux.c
1555
if ((r = sshbuf_put_stringb(queue, m)) != 0)
usr.bin/ssh/mux.c
1558
need = sshbuf_len(queue);
usr.bin/ssh/mux.c
1559
ptr = sshbuf_ptr(queue);
usr.bin/ssh/mux.c
1563
sshbuf_free(queue);
usr.bin/ssh/mux.c
1577
sshbuf_free(queue);
usr.bin/ssh/mux.c
1583
sshbuf_free(queue);
usr.bin/ssh/mux.c
1589
sshbuf_free(queue);
usr.bin/ssh/mux.c
1596
struct sshbuf *queue;
usr.bin/ssh/mux.c
1601
if ((queue = sshbuf_new()) == NULL)
usr.bin/ssh/mux.c
1603
if (mux_client_read(fd, queue, 4, timeout_ms) != 0) {
usr.bin/ssh/mux.c
1607
sshbuf_free(queue);
usr.bin/ssh/mux.c
1611
need = PEEK_U32(sshbuf_ptr(queue));
usr.bin/ssh/mux.c
1612
if (mux_client_read(fd, queue, need, timeout_ms) != 0) {
usr.bin/ssh/mux.c
1615
sshbuf_free(queue);
usr.bin/ssh/mux.c
1619
if ((r = sshbuf_get_string_direct(queue, &ptr, &have)) != 0 ||
usr.bin/ssh/mux.c
1622
sshbuf_free(queue);
usr.bin/ssh/sftp-server.c
493
get_handle(struct sshbuf *queue, int *hp)
usr.bin/ssh/sftp-server.c
500
if ((r = sshbuf_get_string(queue, &handle, &hlen)) != 0)
usr.bin/systat/pftop.c
1550
if ((node = pfctl_find_queue_node(pqs.queue.qname,
usr.bin/systat/pftop.c
1551
pqs.queue.ifname)) != NULL) {
usr.bin/systat/pftop.c
1557
pfctl_insert_queue_node(pqs.queue, qstats);
usr.bin/tmux/cmd-queue.c
121
return (c->queue);
usr.bin/tmux/cmd-queue.c
128
struct cmdq_list *queue;
usr.bin/tmux/cmd-queue.c
130
queue = xcalloc(1, sizeof *queue);
usr.bin/tmux/cmd-queue.c
131
TAILQ_INIT (&queue->list);
usr.bin/tmux/cmd-queue.c
132
return (queue);
usr.bin/tmux/cmd-queue.c
137
cmdq_free(struct cmdq_list *queue)
usr.bin/tmux/cmd-queue.c
139
if (!TAILQ_EMPTY(&queue->list))
usr.bin/tmux/cmd-queue.c
141
free(queue);
usr.bin/tmux/cmd-queue.c
304
struct cmdq_list *queue = cmdq_get(c);
usr.bin/tmux/cmd-queue.c
315
item->queue = queue;
usr.bin/tmux/cmd-queue.c
316
TAILQ_INSERT_TAIL(&queue->list, item, entry);
usr.bin/tmux/cmd-queue.c
321
return (TAILQ_LAST(&queue->list, cmdq_item_list));
usr.bin/tmux/cmd-queue.c
329
struct cmdq_list *queue = after->queue;
usr.bin/tmux/cmd-queue.c
341
item->queue = queue;
usr.bin/tmux/cmd-queue.c
342
TAILQ_INSERT_AFTER(&queue->list, after, item, entry);
usr.bin/tmux/cmd-queue.c
44
struct cmdq_list *queue;
usr.bin/tmux/cmd-queue.c
463
TAILQ_REMOVE(&item->queue->list, item, entry);
usr.bin/tmux/cmd-queue.c
733
struct cmdq_list *queue = cmdq_get(c);
usr.bin/tmux/cmd-queue.c
740
if (TAILQ_EMPTY(&queue->list)) {
usr.bin/tmux/cmd-queue.c
744
if (TAILQ_FIRST(&queue->list)->flags & CMDQ_WAITING) {
usr.bin/tmux/cmd-queue.c
751
item = queue->item = TAILQ_FIRST(&queue->list);
usr.bin/tmux/cmd-queue.c
801
queue->item = NULL;
usr.bin/tmux/cmd-queue.c
815
struct cmdq_list *queue = cmdq_get(c);
usr.bin/tmux/cmd-queue.c
817
if (queue->item == NULL)
usr.bin/tmux/cmd-queue.c
819
if (queue->item->flags & CMDQ_WAITING)
usr.bin/tmux/cmd-queue.c
821
return (queue->item);
usr.bin/tmux/server-client.c
309
c->queue = cmdq_new();
usr.bin/tmux/server-client.c
536
cmdq_free(c->queue);
usr.sbin/cron/atrun.c
113
if ((queue = strchr(file->d_name, '.')) == NULL)
usr.sbin/cron/atrun.c
115
*queue++ = '\0';
usr.sbin/cron/atrun.c
119
if (!isalpha((unsigned char)*queue))
usr.sbin/cron/atrun.c
134
job->queue = *queue;
usr.sbin/cron/atrun.c
182
(long long)job->run_time, job->queue);
usr.sbin/cron/atrun.c
207
if (isupper(job->queue)) {
usr.sbin/cron/atrun.c
227
(long long)batch->run_time, batch->queue);
usr.sbin/cron/atrun.c
516
if (job->queue > 'b') {
usr.sbin/cron/atrun.c
517
if (setpriority(PRIO_PROCESS, 0, job->queue - 'b') != 0)
usr.sbin/cron/atrun.c
66
char *queue;
usr.sbin/cron/structs.h
68
int queue; /* name of the at queue */
usr.sbin/ldpd/accept.c
110
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/ldpd/accept.c
118
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/ldpd/accept.c
35
LIST_HEAD(, accept_ev) queue;
usr.sbin/ldpd/accept.c
47
LIST_INIT(&accept_queue.queue);
usr.sbin/ldpd/accept.c
61
LIST_INSERT_HEAD(&accept_queue.queue, av, entry);
usr.sbin/ldpd/accept.c
76
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/lpr/common_source/common.c
227
getq(struct queue ***namelist)
usr.sbin/lpr/common_source/common.c
230
struct queue *q, **queue = NULL;
usr.sbin/lpr/common_source/common.c
248
queue = calloc(arraysz, sizeof(struct queue *));
usr.sbin/lpr/common_source/common.c
249
if (queue == NULL)
usr.sbin/lpr/common_source/common.c
261
q = malloc(sizeof(struct queue));
usr.sbin/lpr/common_source/common.c
272
struct queue **newqueue;
usr.sbin/lpr/common_source/common.c
273
newqueue = reallocarray(queue,
usr.sbin/lpr/common_source/common.c
274
arraysz, 2 * sizeof(struct queue *));
usr.sbin/lpr/common_source/common.c
280
queue = newqueue;
usr.sbin/lpr/common_source/common.c
282
queue[nitems++] = q;
usr.sbin/lpr/common_source/common.c
286
qsort(queue, nitems, sizeof(struct queue *), compar);
usr.sbin/lpr/common_source/common.c
287
*namelist = queue;
usr.sbin/lpr/common_source/common.c
291
if (queue != NULL) {
usr.sbin/lpr/common_source/common.c
293
free(queue[nitems]);
usr.sbin/lpr/common_source/common.c
294
free(queue);
usr.sbin/lpr/common_source/common.c
306
struct queue *p1 = *(struct queue **)v1;
usr.sbin/lpr/common_source/common.c
307
struct queue *p2 = *(struct queue **)v2;
usr.sbin/lpr/common_source/displayq.c
145
if ((nitems = getq(&queue)) < 0)
usr.sbin/lpr/common_source/displayq.c
235
q = queue[i];
usr.sbin/lpr/common_source/displayq.c
241
free(queue);
usr.sbin/lpr/common_source/displayq.c
95
struct queue *q;
usr.sbin/lpr/common_source/displayq.c
98
struct queue **queue;
usr.sbin/lpr/common_source/lp.h
128
int getq(struct queue ***);
usr.sbin/lpr/lpc/cmds.c
1003
free(queue[i]);
usr.sbin/lpr/lpc/cmds.c
1004
free(queue);
usr.sbin/lpr/lpc/cmds.c
1028
touch(struct queue *q)
usr.sbin/lpr/lpc/cmds.c
1048
struct queue **qq;
usr.sbin/lpr/lpc/cmds.c
1073
for (qq = queue + nitems; --qq >= queue; ) {
usr.sbin/lpr/lpc/cmds.c
1093
for (qq = queue + nitems; --qq >= queue; ) {
usr.sbin/lpr/lpc/cmds.c
294
struct dirent **queue;
usr.sbin/lpr/lpc/cmds.c
312
nitems = scandir(SD, &queue, doselect, sortq);
usr.sbin/lpr/lpc/cmds.c
322
cp = queue[i]->d_name;
usr.sbin/lpr/lpc/cmds.c
326
cp1 = queue[i + 1]->d_name;
usr.sbin/lpr/lpc/cmds.c
68
static int touch(struct queue *);
usr.sbin/lpr/lpc/cmds.c
947
struct queue **queue;
usr.sbin/lpr/lpc/cmds.c
990
nitems = getq(&queue);
usr.sbin/lpr/lpc/cmds.c
994
mtime = queue[0]->q_time;
usr.sbin/lpr/lpd/printjob.c
131
struct queue *q, **qp;
usr.sbin/lpr/lpd/printjob.c
132
struct queue **queue;
usr.sbin/lpr/lpd/printjob.c
202
if ((nitems = getq(&queue)) < 0) {
usr.sbin/lpr/lpd/printjob.c
223
for (qp = queue; nitems--; free(q)) {
usr.sbin/lpr/lpd/printjob.c
295
free(queue);
usr.sbin/lpr/lpd/printjob.c
299
if ((nitems = getq(&queue)) < 0) {
usr.sbin/npppd/npppd/accept.c
112
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/npppd/npppd/accept.c
120
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/npppd/npppd/accept.c
37
LIST_HEAD(, accept_ev) queue;
usr.sbin/npppd/npppd/accept.c
49
LIST_INIT(&accept_queue.queue);
usr.sbin/npppd/npppd/accept.c
63
LIST_INSERT_HEAD(&accept_queue.queue, av, entry);
usr.sbin/npppd/npppd/accept.c
78
LIST_FOREACH(av, &accept_queue.queue, entry)
usr.sbin/rpki-client/http.c
164
static struct http_req_queue queue = TAILQ_HEAD_INITIALIZER(queue);
usr.sbin/rpki-client/http.c
2240
TAILQ_FOREACH_SAFE(req, &queue, entry, nr)
usr.sbin/rpki-client/http.c
595
TAILQ_INSERT_TAIL(&queue, req, entry);
usr.sbin/rpki-client/http.c
657
TAILQ_REMOVE(&queue, req, entry);
usr.sbin/rpki-client/http.c
686
TAILQ_INSERT_HEAD(&queue, req, entry);
usr.sbin/rpki-client/repo.c
1196
entityq_flush(&rp->queue, rp);
usr.sbin/rpki-client/repo.c
1208
entityq_flush(&rp->queue, rp);
usr.sbin/rpki-client/repo.c
1257
entityq_flush(&rp->queue, rp);
usr.sbin/rpki-client/repo.c
1273
entityq_flush(&rp->queue, rp);
usr.sbin/rpki-client/repo.c
1427
TAILQ_INSERT_TAIL(&rp->queue, p, entries);
usr.sbin/rpki-client/repo.c
1441
if (TAILQ_EMPTY(&rp->queue))
usr.sbin/rpki-client/repo.c
343
entityq_flush(&rp->queue, rp);
usr.sbin/rpki-client/repo.c
657
TAILQ_INIT(&rp->queue);
usr.sbin/rpki-client/repo.c
99
struct entityq queue; /* files waiting for repo */
usr.sbin/smtpd/expand.c
83
if (expand->queue)
usr.sbin/smtpd/expand.c
84
TAILQ_INSERT_TAIL(expand->queue, xn, tq_entry);
usr.sbin/smtpd/expand.c
95
if (expand->queue)
usr.sbin/smtpd/expand.c
96
while ((xn = TAILQ_FIRST(expand->queue)))
usr.sbin/smtpd/expand.c
97
TAILQ_REMOVE(expand->queue, xn, tq_entry);
usr.sbin/smtpd/lka.c
894
TAILQ_FOREACH(mn, &lk.maddrmap->queue, entries) {
usr.sbin/smtpd/lka_session.c
81
lks->expand.queue = &lks->nodes;
usr.sbin/smtpd/mailaddr.c
103
TAILQ_INIT(&maddrmap->queue);
usr.sbin/smtpd/mailaddr.c
112
TAILQ_INSERT_TAIL(&maddrmap->queue, mn, entries);
usr.sbin/smtpd/mailaddr.c
120
while ((mn = TAILQ_FIRST(&maddrmap->queue))) {
usr.sbin/smtpd/mailaddr.c
121
TAILQ_REMOVE(&maddrmap->queue, mn, entries);
usr.sbin/smtpd/parse.y
212
| grammar queue '\n'
usr.sbin/smtpd/parse.y
519
queue:
usr.sbin/smtpd/smtpd.c
781
return queue();
usr.sbin/smtpd/smtpd.h
1521
int queue(void);
usr.sbin/smtpd/smtpd.h
450
TAILQ_HEAD(xnodes, expandnode) *queue;
usr.sbin/smtpd/smtpd.h
462
TAILQ_HEAD(xmaddr, maddrnode) queue;
usr.sbin/unbound/dnstap/dtstream.c
462
item->queue = mq;
usr.sbin/unbound/dnstap/dtstream.c
476
if(item->queue == mq) {
usr.sbin/unbound/dnstap/dtstream.c
481
lock_basic_lock(&item->queue->lock);
usr.sbin/unbound/dnstap/dtstream.c
482
item->queue->dtio = NULL;
usr.sbin/unbound/dnstap/dtstream.c
483
lock_basic_unlock(&item->queue->lock);
usr.sbin/unbound/dnstap/dtstream.c
548
if(dtio_find_in_queue(dtio, item->queue))
usr.sbin/unbound/dnstap/dtstream.c
555
if(dtio_find_in_queue(dtio, item->queue))
usr.sbin/unbound/dnstap/dtstream.h
240
struct dt_msg_queue* queue;