q_size
unsigned long q_size, alloc_size, pages, order;
q_size = pbm->msiq_ent_count * sizeof(struct pci_sun4v_msiq_entry);
alloc_size = (pbm->msiq_num * q_size);
unsigned long err, base = __pa(pages + (i * q_size));
unsigned long q_size, alloc_size, pages, order;
q_size = pbm->msiq_ent_count * sizeof(struct pci_sun4v_msiq_entry);
alloc_size = (pbm->msiq_num * q_size);
static inline u32 fifo_space_avail(u32 head, u32 tail, u32 q_size)
avail += q_size;
size_t q_size;
q_size = qlen * cqinfo->cmd_size;
size_t rem_q_size = q_size;
size_t q_size, c_size, rem_q_size;
q_size = qlen * OTX_CPT_INST_SIZE;
rem_q_size = q_size;
uint64_t q_size;
q_data->q_size = q->properties.queue_size;
qp->queue_size = q_data->q_size;
static int post_recv_io(struct rtrs_clt_con *con, size_t q_size)
for (i = 0; i < q_size; i++) {
size_t q_size = 0;
q_size = SERVICE_CON_QUEUE_DEPTH;
q_size = clt_path->queue_depth;
q_size *= 2;
err = post_recv_io(to_clt_con(clt_path->s.con[cid]), q_size);
size_t q_size;
q_size = SERVICE_CON_QUEUE_DEPTH;
q_size = srv->queue_depth;
err = post_recv_io(to_srv_con(srv_path->s.con[cid]), q_size);
static int post_recv_io(struct rtrs_srv_con *con, size_t q_size)
for (i = 0; i < q_size; i++) {
return q_size(&mq->dirty) == 0u;
iface_q->qhdr->q_size = IFACEQ_QUEUE_SIZE / sizeof(u32);
u32 q_size;
qsize = qhdr->q_size;
qsize = qhdr->q_size;
u32 q_size;
qhdr->q_size = IFACEQ_QUEUE_SIZE / 4;
unsigned int q_size = q_data->format.plane_fmt[i].sizeimage;
if (sizes[i] < q_size)
return (qinfo->q_size - msgq_occupied(msgq, qinfo) - 1);
qinfo->q_size = msgq_size;
qinfo->q_low = qinfo->q_size >> 1;
qinfo->q_mask = qinfo->q_size - 1;
avail, qinfo->q_size);
if (wr_idx >= qinfo->q_size) {
wr_idx, qinfo->q_size);
max_msg_to_process = BCM_VK_MSG_PROC_MAX_LOOP * qinfo->q_size;
if ((rd_idx >= qinfo->q_size) ||
(src_size > (qinfo->q_size - 1))) {
rd_idx, src_size, qinfo->q_size);
u32 q_size;
dma_free_coherent(dev, qcq->q_size,
qcq->q_size = PDS_PAGE_SIZE +
qcq->q_size + qcq->cq_size,
qcq->q_size = PDS_PAGE_SIZE + (num_descs * desc_size);
qcq->q_base = dma_alloc_coherent(dev, qcq->q_size,
dma_free_coherent(dev, qcq->q_size, qcq->q_base, qcq->q_base_pa);
u32 q_size;
debugfs_create_x32("q_size", 0400, qcq_dentry, &qcq->q_size);
u32 q_size;
q_size = txq_depth * BFI_TXQ_WI_SIZE;
q_size = ALIGN(q_size, PAGE_SIZE);
page_count = q_size >> PAGE_SHIFT;
lio_dma_free(oct, q_size, iq->base_addr, iq->base_addr_dma);
u64 desc_size = 0, q_size;
q_size = iq->max_count * desc_size;
lio_dma_free(oct, (u32)q_size, iq->base_addr,
u32 q_size;
q_size = (u32)conf->instr_type * num_descs;
iq->base_addr = lio_dma_alloc(oct, q_size, &iq->base_addr_dma);
lio_dma_free(oct, q_size, iq->base_addr, iq->base_addr_dma);
debugfs_create_x32("q_size", 0400, qcq_dentry, &qcq->q_size);
swap(a->q_size, b->q_size);
dma_free_coherent(dev, qcq->q_size, qcq->q_base, qcq->q_base_pa);
int q_size;
q_size = ALIGN(num_descs * desc_size, PAGE_SIZE);
new->q_size = PAGE_SIZE + q_size +
new->q_base = dma_alloc_coherent(dev, new->q_size,
new->cq.base = PTR_ALIGN(new->q.base + q_size, PAGE_SIZE);
new->cq.base_pa = ALIGN(new->q_base_pa + q_size, PAGE_SIZE);
new->q_size = PAGE_SIZE + (num_descs * desc_size);
new->q_base = dma_alloc_coherent(dev, new->q_size, &new->q_base_pa,
dma_free_coherent(dev, new->q_size, new->q_base, new->q_base_pa);
memset(qcq->q_base, 0, qcq->q_size);
u32 q_size;
u16 q_size;
u16 q_size;
q_sz = mei->shared_mem.q_size[SAP_DIRECTION_ME_TO_HOST][SAP_QUEUE_IDX_NOTIF];
q_sz = mei->shared_mem.q_size[SAP_DIRECTION_ME_TO_HOST][SAP_QUEUE_IDX_DATA];
size_t q_size[SAP_DIRECTION_MAX][SAP_QUEUE_IDX_MAX];
mem->q_size[dir][queue] =
q_sz = mei->shared_mem.q_size[SAP_DIRECTION_HOST_TO_ME][SAP_QUEUE_IDX_NOTIF];
q_sz = mei->shared_mem.q_size[SAP_DIRECTION_HOST_TO_ME][SAP_QUEUE_IDX_DATA];
int wait, i, q_index, q_size;
q_size = vhost->scsi_scrqs.active_queues;
q_size = 1;
for (q_index = 0; q_index < q_size; q_index++) {
for (q_index = 0; q_index < q_size; q_index++) {
const size_t q_size = __member_size(q);
if (p_size == SIZE_MAX && q_size == SIZE_MAX)
const size_t q_size = __member_size(q);
if (p_size == SIZE_MAX && q_size == SIZE_MAX)
const size_t q_size = __member_size(q);
if (p_size == SIZE_MAX && q_size == SIZE_MAX)
const size_t q_size,
if (__compiletime_lessthan(q_size_field, q_size) &&
__compiletime_lessthan(q_size, size))
else if (q_size != SIZE_MAX && q_size < size)
fortify_panic(func, FORTIFY_READ, q_size, size, true);
#define __fortify_memcpy_chk(p, q, size, p_size, q_size, \
const size_t __q_size = (q_size); \
const size_t q_size = __struct_size(q);
if (__compiletime_lessthan(q_size, size))
else if (q_size < size)
fortify_panic(FORTIFY_FUNC_memcmp, FORTIFY_READ, q_size, size, INT_MIN);
const size_t q_size = __member_size(q);
__builtin_constant_p(q_size) &&
p_size == SIZE_MAX && q_size == SIZE_MAX)
int q_size = 0;
q_size += size;
atomic_add(q_size, &sk->sk_rmem_alloc);