MAXPHYS
c = ulmin(iov->iov_len, MAXPHYS);
#define BYTES_PER_DUMP MAXPHYS /* must be a multiple of pagesize */
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
#define BYTES_PER_DUMP MAXPHYS /* must be a multiple of pagesize */
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
#define OCTMMC_MAX_DMASEG MIN(MAXPHYS, (1u << 18))
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
c = ulmin(iov->iov_len, MAXPHYS);
KASSERT(ncookies < MAXPHYS / PAGE_SIZE);
int vsd_map_idx[MAXPHYS / PAGE_SIZE];
ai.max_xfer_sz = MAXPHYS / DEV_BSIZE;
struct ldc_cookie cookie[MAXPHYS / PAGE_SIZE];
ai.max_xfer_sz = MAXPHYS / sc->sc_vdisk_block_size;
c = ulmin(iov->iov_len, MAXPHYS);
error = bus_dmamap_create(sc->sc_dmat, MAXPHYS, AMLMMC_NDESC,
error = bus_dmamap_create(sc->sc_dmat, MAXPHYS, DWMMC_NDESC,
if (bus_dmamap_create(sc->aac_dmat, MAXPHYS, AAC_MAXSGENTRIES,
MAXPHYS, 0, BUS_DMA_NOWAIT, &cm->cm_datamap)) {
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS, AHCI_MAX_PRDT,
#define AHD_NSEG (roundup(atop(MAXPHYS) + 1, 16))
AHC_MAXTRANSFER_SIZE, AHC_NSEG, MAXPHYS, 0,
#define AHC_NSEG (roundup(atop(MAXPHYS) + 1, 16))
#define AHC_MAXTRANSFER_SIZE MIN(MAXPHYS,0x00ffffff)
sc->cfg.maxsg = MAXPHYS / PAGE_SIZE;
if (sc->maxsg > MAXPHYS / PAGE_SIZE)
sc->maxsg = MAXPHYS / PAGE_SIZE;
MAXPHYS, sc->sc_max_sgl, MAXPHYS, 0,
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS,
sc->sc_max_sgl_len, MAXPHYS, 0,
u_int64_t prpl[MAXPHYS / PAGE_SIZE];
sc->sc_mdts = MAXPHYS;
#define OOSIOP_NSG (MIN(atop(MAXPHYS) + 1, 32)) /* Max S/G operation */
#if MAXPHYS > (PAGE_SIZE * (OSIOP_NSG - 1))
#define OSIOP_MAX_XFER MAXPHYS
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS,
QLA_MAX_SEGS, MAXPHYS, 0,
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS,
QLW_MAX_SEGS, MAXPHYS, 0,
#define RTSX_DMA_DATA_BUFSIZE MAXPHYS
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS, SILI_DMA_SEGS,
MAXPHYS, 0, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW,
error = bus_dmamap_create(sc->sc_c.sc_dmat, MAXPHYS, SIOP_NSG,
MAXPHYS, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
if (atareq->datalen > MAXPHYS)
MAXPHYS, BUS_DMA_NOWAIT|BUS_DMA_ALLOCNOW)) {
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS, ARC_SGL_MAXLEN,
MAXPHYS, 0, 0, &ccb->ccb_dmamap) != 0) {
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS, IPS_MAXSGS,
MAXPHYS, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
MAXPHYS, sc->sc_max_sgl, MAXPHYS, 0,
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS, sc->sc_max_sgl,
MAXPHYS, 0,
#define NIDEDMA_TABLES (MAXPHYS/PAGE_SIZE + 1)
#define MDL_SIZE (MAXPHYS / MDL_SEG_SIZE + 1) /* no hardware limit? */
if (bus_dmamap_create(esc->sc_dmat, MAXPHYS, MDL_SIZE, MDL_SEG_SIZE,
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS,
QLE_MAX_SEGS-1, MAXPHYS, 0,
#define VMWPVS_MAXSGL (MAXPHYS / PAGE_SIZE)
if (bus_dmamap_create(sc->sc_dmat, MAXPHYS,
VMWPVS_MAXSGL, MAXPHYS, 0,
#define HVS_MAX_SGE (MAXPHYS / PAGE_SIZE + 1)
error = bus_dmamap_create(sc->sc_dmat, MAXPHYS, HVS_MAX_SGE,
#define SEG_MAX (MAXPHYS/PAGE_SIZE + 1)
r = bus_dmamap_create(sc->sc_virtio->sc_dmat, MAXPHYS,
SEG_MAX, MAXPHYS, 0,
#define SEG_MAX (MAXPHYS/PAGE_SIZE + 1)
r = bus_dmamap_create(vsc->sc_dmat, MAXPHYS, SEG_MAX, MAXPHYS,
error = bus_dmamap_create(sc->sc_dmat, MAXPHYS, 2 *
if (size > MAXPHYS)
sc->sc_max_seg = saa->max_seg ? saa->max_seg : MAXPHYS;
error = bus_dmamap_create(sc->sc_dmat, MAXPHYS, SDMMC_MAXNSEGS,
#define SDMMC_MAXNSEGS ((MAXPHYS / PAGE_SIZE) + 1)
dma_bufsize = (size > MAXPHYS) ? MAXPHYS : size;
bufsize = (size > MAXPHYS) ? MAXPHYS : size;
length = MIN(MIN(leftover, chunkend - lbaoffs), MAXPHYS);
dma_free(crwu->cr_dmabuf, MAXPHYS);
crwu->cr_dmabuf = dma_alloc(MAXPHYS, PR_WAITOK);
crwu->cr_crp = crypto_getreq(MAXPHYS >> DEV_BSHIFT);
(MAXPHYS / sd->sd_meta->ssdi.ssd_strip_size + 1) *
sd->sd_meta->ssdi.ssd_strip_size = MAXPHYS;
sd->sd_meta->ssdi.ssd_strip_size = MAXPHYS;
sd->sd_meta->ssdi.ssd_strip_size = MAXPHYS;
if (bp->b_bcount > MAXPHYS)
bp->b_bcount = MAXPHYS;
chunk = MIN(csize, MAXPHYS);
#define VMMAP_MAXLEN MAXPHYS
i += MAXPHYS, blkctr += btodb(MAXPHYS)) {
if (hibernate_block_io(hib, blkctr, MAXPHYS,
tempva = (vaddr_t)km_alloc(MAXPHYS + PAGE_SIZE, &kv_any, &kp_none,
if (compressed_size - processed >= MAXPHYS)
read_size = MAXPHYS;
KASSERT(num_io_pages <= MAXPHYS/PAGE_SIZE + 1);
for (i = 0; i < hibernate_chunk_table_size; i += MAXPHYS) {
MAXPHYS, IO_TYPE_CHK))) {
bufkvm &= ~(MAXPHYS - 1);
howmany = MAXPHYS / size;
bcstats.kvaslots = bcstats.kvaslots_avail = size / MAXPHYS;
buf_kva_start += MAXPHYS;
if (oldp != NULL && *oldlenp > MAXPHYS)
if (oldp != NULL && *oldlenp > MAXPHYS)
if (screq->datalen > MAXPHYS)
if (atareq->datalen > MAXPHYS)
#define RESERVE_PAGES (RESERVE_SLOTS * MAXPHYS / PAGE_SIZE)
struct vm_page *pgs[MAXPHYS >> PAGE_SHIFT];
KASSERT(npages <= MAXPHYS >> PAGE_SHIFT);
if (sz > MAXPHYS)