emit
bool emit = false, reset_basic_block = false;
emit = true;
emit = true;
emit = true;
if (emit) {
ring->emit, ring->size) > ring->size / 2)
rq->ring->emit);
ce->ring->emit);
intel_ring_reset(ce->ring, ce->ring->emit);
u32 *(*emit)(const struct intel_context *, u32 *))
cs = emit(ce, start);
u32 *(*emit)(const struct intel_context *, u32 *))
cs = emit(ce, start);
pkt = min_t(int, pkt, (ring->size - ring->emit) / sizeof(u32) + 5);
ring->emit = (void *)cs - ring->vaddr;
ring->emit = (void *)cs - ring->vaddr;
ring->emit, ring->size))
space = __intel_ring_space(ring->head, ring->emit, ring->size);
const unsigned int remain_usable = ring->effective_size - ring->emit;
const int remain_actual = ring->size - ring->emit;
GEM_BUG_ON(ring->emit + need_wrap > ring->size);
memset64(ring->vaddr + ring->emit, 0, need_wrap / sizeof(u64));
ring->emit = 0;
GEM_BUG_ON(ring->emit > ring->size - bytes);
cs = ring->vaddr + ring->emit;
ring->emit += bytes;
intel_ring_reset(ring, ring->emit);
ring->emit = tail;
GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs);
GEM_BUG_ON(!IS_ALIGNED(rq->ring->emit, 8)); /* RING_TAIL qword align */
intel_ring_reset(ce->ring, ce->ring->emit);
u32 emit; /* updated during request construction */
GEM_BUG_ON(rq[0]->postfix > ce[1]->ring->emit);
ce[0]->ring->emit,
ce[0]->ring->tail, ce[0]->ring->emit,
ce[1]->ring->tail, ce[1]->ring->emit);
ce[0]->ring->emit,
ce[0]->ring->tail, ce[0]->ring->emit,
ce[1]->ring->tail, ce[1]->ring->emit);
ring->emit = ring->size / 2;
ring->tail = ring->emit;
pr_info("%s emit=%u sz=%d\n", __func__, rq->ring->emit, sz);
if (8 * watcher[1].rq->ring->emit >
rq->head = rq->ring->emit;
rq->infix = rq->ring->emit; /* end of header; start of user payload */
ce->ring->emit = rq->head;
sz = rq->ring->emit - rq->head;
emit($1, 0);
static void emit(ssize_t, int);
emit(instructions[i].index, level + 1);
emit(code, 0);
emit(def);
void emit(definition *);