BLKIF_MAX_SEGMENTS_PER_REQUEST
n = min_t(unsigned int, BLKIF_MAX_SEGMENTS_PER_REQUEST,
n = min_t(unsigned int, BLKIF_MAX_SEGMENTS_PER_REQUEST,
(nseg > BLKIF_MAX_SEGMENTS_PER_REQUEST)) ||
struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
if (++segs_to_unmap == BLKIF_MAX_SEGMENTS_PER_REQUEST ||
struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
if (++segs_to_unmap == BLKIF_MAX_SEGMENTS_PER_REQUEST) {
struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page *unmap_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
unsigned int batch = min(num, BLKIF_MAX_SEGMENTS_PER_REQUEST);
struct gnttab_map_grant_ref map[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct page *pages_to_gnt[BLKIF_MAX_SEGMENTS_PER_REQUEST];
if (segs_to_map == BLKIF_MAX_SEGMENTS_PER_REQUEST)
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
BLKIF_MAX_SEGMENTS_PER_REQUEST));
grants = BLKIF_MAX_SEGMENTS_PER_REQUEST;
if (indirect_segments <= BLKIF_MAX_SEGMENTS_PER_REQUEST)
if (xen_blkif_max_segments < BLKIF_MAX_SEGMENTS_PER_REQUEST)
xen_blkif_max_segments = BLKIF_MAX_SEGMENTS_PER_REQUEST;
grant_idx >= BLKIF_MAX_SEGMENTS_PER_REQUEST)) {
grant_idx -= BLKIF_MAX_SEGMENTS_PER_REQUEST;
first->u.rw.nr_segments = BLKIF_MAX_SEGMENTS_PER_REQUEST;
second->u.rw.nr_segments = nr_segments - BLKIF_MAX_SEGMENTS_PER_REQUEST;
(BLKIF_MAX_SEGMENTS_PER_REQUEST * XEN_PAGE_SIZE) / 512;
if (max_grefs > BLKIF_MAX_SEGMENTS_PER_REQUEST)
num_grant > BLKIF_MAX_SEGMENTS_PER_REQUEST;
if (num_grant > BLKIF_MAX_SEGMENTS_PER_REQUEST &&
#define HAS_EXTRA_REQ (BLKIF_MAX_SEGMENTS_PER_REQUEST < XEN_PFN_PER_PAGE)
BLKIF_MAX_SEGMENTS_PER_REQUEST;
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];