BLKIF_MAX_SEGMENTS_PER_REQUEST
int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST;
int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST;
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST;
int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST;
gnttab_map_grant_ref_t mapops[BLKIF_MAX_SEGMENTS_PER_REQUEST];
if (segs > BLKIF_MAX_SEGMENTS_PER_REQUEST)
BLKIF_MAX_SEGMENTS_PER_REQUEST;
gnttab_unmap_grant_ref_t unmapops[BLKIF_MAX_SEGMENTS_PER_REQUEST];
ASSERT(segs <= BLKIF_MAX_SEGMENTS_PER_REQUEST);
grant_handle_t xr_page_hdls[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page xr_plist[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page *xr_pplist[BLKIF_MAX_SEGMENTS_PER_REQUEST];
#define XDB_MAX_IO_PAGES(v) ((v)->xs_nentry * BLKIF_MAX_SEGMENTS_PER_REQUEST)
* BLKIF_MAX_SEGMENTS_PER_REQUEST \
xdb_seg_t xr_segs[BLKIF_MAX_SEGMENTS_PER_REQUEST];
ASSERT(seg <= BLKIF_MAX_SEGMENTS_PER_REQUEST);
BLKIF_MAX_SEGMENTS_PER_REQUEST, /* maximum number of segments */
BLKIF_MAX_SEGMENTS_PER_REQUEST, &gh) == -1) {
BLKIF_MAX_SEGMENTS_PER_REQUEST);
ASSERT(gs->gs_ngrefs < BLKIF_MAX_SEGMENTS_PER_REQUEST);
grant_ref_t gs_ge[BLKIF_MAX_SEGMENTS_PER_REQUEST];
#define XB_MAX_XFER (XB_MAX_SEGLEN * BLKIF_MAX_SEGMENTS_PER_REQUEST)
grant_ref_t gref[BLKIF_MAX_SEGMENTS_PER_REQUEST];
(BLKIF_RING_SIZE * BLKIF_MAX_SEGMENTS_PER_REQUEST * PAGESIZE)
((uintptr_t)base + (id * BLKIF_MAX_SEGMENTS_PER_REQUEST * PAGESIZE))
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];