buf_index
int buf_index;
r = copy_to_user(buf, p->buf + p->buf_index, n);
p->buf_index += n;
p->buf_index = n;
int buf_index;
r = copy_to_user(buf, p->buf + p->buf_index, n);
p->buf_index += n;
p->buf_index = n;
static int read_data_helper(struct sahara_context *context, int buf_index)
memcpy(context->tx[buf_index],
context->tx[buf_index], pkt_data_len, mhi_flag);
.index = elem->buf_index,
fcmd->buf_group = READ_ONCE(cmd->sqe->buf_index);
hisi_ptt_update_aux(hisi_ptt, hisi_ptt->trace_ctrl.buf_index, true);
hisi_ptt->trace_ctrl.buf_index = 0;
addr = ctrl->trace_buf[ctrl->buf_index].addr;
hisi_ptt->trace_ctrl.buf_index = (buf_idx + 1) % HISI_PTT_TRACE_BUF_CNT;
u32 buf_index;
size_t buf_index;
bus->buf_index = 0;
if (bus->buf_index < msg->len) {
writel(msg->buf[bus->buf_index++],
msg->buf[bus->buf_index++] = recv_byte;
if (bus->buf_index < msg->len) {
if (bus->buf_index + 1 == msg->len)
pt1->buf_index = 0;
pt1->table_index, pt1->buf_index);
page = pt1->tables[pt1->table_index].bufs[pt1->buf_index].page;
if (++pt1->buf_index >= PT1_NR_BUFS) {
pt1->buf_index = 0;
int buf_index;
cap->buf_index = 0;
fimc_hw_set_output_addr(fimc, &v_buf->addr, cap->buf_index);
v_buf->index = cap->buf_index;
if (++cap->buf_index >= FIMC_MAX_OUT_BUFS)
cap->buf_index = 0;
if (++cap->buf_index >= FIMC_MAX_OUT_BUFS)
cap->buf_index = 0;
vid_cap->buf_index = 0;
vid_cap->buf_index;
buf->index = vid_cap->buf_index;
if (++vid_cap->buf_index >= FIMC_MAX_OUT_BUFS)
vid_cap->buf_index = 0;
int buf_index;
int buf_index = ivb->index * num_planes + i;
is->is_p_region->shared[32 + buf_index] =
buf_index, ivb->index, i, vb->index,
int buf_index;
buf_index = (is->i2h_cmd.args[1] - 1) % video->buf_count;
vbuf = &video->buffers[buf_index]->vb;
video->buf_mask &= ~BIT(buf_index);
fimc->buf_index = 0;
buf->index = fimc->buf_index++;
if (fimc->buf_index >= fimc->reqbufs_count)
fimc->buf_index = 0;
unsigned short buf_index;
buf->index = vp->buf_index;
vp->buf_index = !vp->buf_index;
unsigned int buf_index;
int buf_index,
mlxsw_reg_pbmc_buf_lossy_set(payload, buf_index, 1);
mlxsw_reg_pbmc_buf_epsb_set(payload, buf_index, 0);
mlxsw_reg_pbmc_buf_size_set(payload, buf_index, size);
int buf_index, u16 size,
mlxsw_reg_pbmc_buf_lossy_set(payload, buf_index, 0);
mlxsw_reg_pbmc_buf_epsb_set(payload, buf_index, 0);
mlxsw_reg_pbmc_buf_size_set(payload, buf_index, size);
mlxsw_reg_pbmc_buf_xoff_threshold_set(payload, buf_index, threshold);
mlxsw_reg_pbmc_buf_xon_threshold_set(payload, buf_index, threshold);
curr_index = rxq->buf_index++;
if (rxq->buf_index == rxq->num_rx_buf)
rxq->buf_index = 0;
rxbuf_oob = &rxq->rx_oobs[rxq->buf_index];
curr = rxq->buf_index;
int count, buf_index, tot_bytes, packets;
buf_index = eval-1;
info = &RX_DESC_INFO(rx, buf_index);
RX_BUFF(rx, buf_index) = 0;
u8 buf_index = 0;
cmd_b + buf_index, cmd_len);
cmd_b + buf_index+3, cmd_len-3);
cmd_b + buf_index, 3);
u8 buf_index = 0;
cmdbuffer + buf_index, 1);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index, 3);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index + 2, 2);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index + 2, 3);
u8 buf_index = 0;
cmdbuffer + buf_index, cmd_len);
cmdbuffer + buf_index+3, cmd_len-3);
cmdbuffer + buf_index, 3);
cmdbuffer + buf_index, 1);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index, 3);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index + 2, 2);
cmdbuffer + buf_index, 2);
cmdbuffer + buf_index + 2, 3);
u8 buf_index = 0;
p_cmdbuffer + buf_index, cmd_len);
p_cmdbuffer + buf_index+3, cmd_len-3);
p_cmdbuffer + buf_index, 3);
u8 buf_index = 0;
u8 buf_index = 0;
cmdbuffer + buf_index, cmd_len);
cmdbuffer + buf_index+3, cmd_len-3);
cmdbuffer + buf_index, 3);
rq->buf_index = 0;
rq->buf_index = 0;
rq->buf_index = 0;
return rq->buf_index++;
unsigned int buf_index;
op->upcall.req.readdir.buf_index = bufi;
new_op->upcall.req.io.buf_index = buffer_index;
__s32 buf_index;
__s32 buf_index;
__s32 buf_index;
op->slot_to_free = op->upcall.req.io.buf_index;
u16 buf_index;
u32 buf_index;
__field( u16, buf_index )
__entry->buf_index = sqe->buf_index;
__entry->buf_index, __entry->personality, __entry->file_index,
__u16 buf_index;
__u16 buf_index;
if (sqe->buf_index || sqe->splice_fd_in)
if (sqe->buf_index || sqe->splice_fd_in)
if (sqe->buf_index || sqe->splice_fd_in)
if (sqe->off || sqe->rw_flags || sqe->buf_index || sqe->splice_fd_in)
sqe->buf_index, sqe->user_data);
if (sqe->off || sqe->len || sqe->buf_index || sqe->splice_fd_in)
if (sqe->off || sqe->rw_flags || sqe->buf_index || sqe->splice_fd_in)
if (sqe->len || sqe->rw_flags || sqe->buf_index || sqe->splice_fd_in)
if (sqe->buf_index || sqe->splice_fd_in)
if (sqe->buf_index || sqe->splice_fd_in)
if (unlikely(sqe->len || sqe->futex_flags || sqe->buf_index ||
if (unlikely(sqe->fd || sqe->buf_index || sqe->file_index ||
req->buf_index = READ_ONCE(sqe->buf_group);
BUILD_BUG_SQE_ELEM(40, __u16, buf_index);
req->buf_index = kbuf->bid;
req->buf_index = READ_ONCE(buf->bid);
req->buf_index = READ_ONCE(buf->bid);
ret = IORING_CQE_F_BUFFER | (req->buf_index << IORING_CQE_BUFFER_SHIFT);
if (unlikely(sqe->buf_index || sqe->personality))
sqe->buf_index || sqe->splice_fd_in))
req->buf_index = READ_ONCE(sqe->buf_index);
sr->notif->buf_index = req->buf_index;
sr->notif->buf_index = req->buf_index;
if (sqe->len || sqe->buf_index)
if (sqe->addr || sqe->rw_flags || sqe->buf_index)
if (sqe->len || sqe->buf_index || sqe->rw_flags || sqe->splice_fd_in)
if (sqe->len || sqe->buf_index || sqe->rw_flags || sqe->splice_fd_in)
if (sqe->addr || sqe->buf_index || sqe->rw_flags || sqe->splice_fd_in || sqe->addr2)
sr->buf_group = req->buf_index;
sr->buf_group = req->buf_index;
req->buf_index = READ_ONCE(sqe->buf_index);
if (sqe->off || sqe->addr || sqe->len || sqe->rw_flags || sqe->buf_index)
if (sqe->off || sqe->addr || sqe->len || sqe->buf_index ||
if (unlikely(sqe->buf_index))
if (sqe->buf_index || sqe->splice_fd_in)
if (sqe->buf_index || sqe->off || sqe->addr)
node = io_rsrc_node_lookup(&ctx->buf_table, req->buf_index);
req->buf_index = READ_ONCE(sqe->buf_index);
io->buf_group = req->buf_index;
if (sqe->buf_index || sqe->splice_fd_in)
if (unlikely(sqe->addr || sqe->buf_index || sqe->splice_fd_in))
if (unlikely(sqe->addr || sqe->buf_index || sqe->splice_fd_in))
if (sqe->buf_index || sqe->rw_flags || sqe->splice_fd_in)
if (sqe->buf_index || sqe->len || sqe->splice_fd_in)
if (sqe->buf_index || sqe->len != 1 || sqe->splice_fd_in)
if (sqe->rw_flags || sqe->addr || sqe->len || sqe->buf_index ||
req->buf_index = READ_ONCE(sqe->buf_index);
if (sqe->addr || sqe->buf_index || sqe->addr3 || sqe->waitid_flags)
__u16 buf_index;
sqe->buf_index = buf_idx;
elem->buf_index = ublk_batch_io_buf_idx(t, q, i);
elem->buf_index = ublk_batch_io_buf_idx(t, q, tag);
unsigned short buf_index = ublk_io_buf_idx(t, q, tag);
sqe[0]->buf_index = buf_index;
io_uring_prep_buf_register(sqe[0], q, tag, q->q_id, buf_index);
sqe[1]->buf_index = buf_index;
io_uring_prep_buf_unregister(sqe[2], q, tag, q->q_id, buf_index);
if (q->tgt_ops->buf_index)
buf.index = q->tgt_ops->buf_index(t, q, tag);
io->buf_index = j++;
io->buf_index = i;
unsigned short buf_index;
unsigned short (*buf_index)(const struct ublk_thread *t,
__u16 buf_index;
return q->ios[tag].buf_index;
.buf_index = ublk_null_buf_index,
sqe->buf_index = buf_idx;
sqe[i]->buf_index = buf_idx;