buf_off
size_t buf_off;
buf_off = ublk_pos_to_buf_off(iocb->ki_pos);
if (buf_off > data_len) {
ret = ublk_copy_user_integrity(req, buf_off, iter, dir);
ret = ublk_copy_user_pages(req, buf_off, iter, dir);
buf_off = 0;
unsigned int buf_off = to_bytes(__block_mod(iter->bi_sector, ec->u_bs));
cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len);
if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio))
ba += buf_off;
int buf_off = 0;
buf_off = 4;
meta->buf[buf_off] = div_u64((soe + dev->cap_frame_eof_offset),
meta->buf[buf_off + 4] = (buf->vb.sequence * 30) % 1000;
*(__u32 *)(meta->buf + buf_off),
*(__u16 *)(meta->buf + buf_off + 4));
unsigned int field, buf_field_cnt, buf_off;
buf_off = get_unaligned_le16(buf + id2off[id] + flash * 2);
if (!buf_off)
if (buf_off >= size)
if (strnlen(&buf[buf_off], size - buf_off) == size - buf_off)
return (const char *)&buf[buf_off];
static void bnx2i_get_write_cmd_bd_idx(struct bnx2i_cmd *cmd, u32 buf_off,
if (buf_off) {
while (buf_off >= (cur_offset + bd_tbl->buffer_length)) {
*start_bd_off = buf_off - cur_offset;
uint32_t buf_off = 0;
if (buf_off >= dma_buf->len) {
buf_off = 0;
buf_addr = dma_buf->vaddr + buf_off;
bytes_copy = min((dma_buf->len - buf_off),
buf_off += bytes_copy;
uint32_t buf_off;
buf_off = sg_addr & (ddp_pagesz - 1);
if (i != 0 && buf_off) {
((buf_off + sg_len) & (ddp_pagesz - 1))) {
size_t buf_off;
for (buf_off = 0; buf_off < count - LPFC_REG_WRITE_KEY_SIZE;
buf_off += sizeof(uint32_t))
writel(*((uint32_t *)(buf + buf_off + LPFC_REG_WRITE_KEY_SIZE)),
phba->ctrl_regs_memmap_p + off + buf_off);
size_t buf_off;
for (buf_off = 0; buf_off < count; buf_off += sizeof(uint32_t)) {
tmp_ptr = (uint32_t *)(buf + buf_off);
*tmp_ptr = readl(phba->ctrl_regs_memmap_p + off + buf_off);
__le32 buf_off;
if (bpf_skb_load_bytes(buf->skb, buf_off(buf), dst, len)) {
if (buf_off(buf) + len > buf->skb->len) {