check_add_overflow
if (check_add_overflow(offset, sizeof(struct ffa_composite_mem_region), &checked_offset)) {
if (check_add_overflow(call_end, INT_MIN, &low_limit))
if (check_add_overflow(start, len, &end) || end > bdev_nr_bytes(bdev))
if (check_add_overflow(start, len, &end) ||
if (check_add_overflow(start, length, &end))
if (check_add_overflow(start_sect, nr_sects, &end_sect)) {
if (check_add_overflow(len_1, len_2, &len))
if (check_add_overflow(sizeof(struct asymmetric_key_id), len, &kid_sz))
if (check_add_overflow(exp_info.size, va_ent[i].len, &exp_info.size)) {
if (!range || check_add_overflow(addr, size, &addr_end))
if (!range || check_add_overflow(start, size, &end)) {
if (check_add_overflow(in_trans->addr, resources->xferred_dma_size, &xfer_start_addr))
if (check_add_overflow(xfer_start_addr, remaining, &end))
if (check_add_overflow(slice_ent[i].offset, slice_ent[i].size, &total) ||
if (check_add_overflow(dump_length,
if (check_add_overflow(dump_length, (u64)sizeof(*dump_meta), &dump_length)) {
if (check_add_overflow(dump_length, mul_bytes, &dump_length)) {
check_add_overflow(sizeof(*swnodes) +
check_add_overflow(sizeof(*props) + sizeof(*package), alloc_size,
if (unlikely(check_add_overflow(sizeof(struct devres),
if (check_add_overflow(init_hdr_offset, init_hdr_size, &end) || end > vram_size) {
if (check_add_overflow(saddr, size, &tmp)
|| check_add_overflow(offset, size, &tmp)
return check_add_overflow(addr, range, &end);
if (unlikely(check_add_overflow(sizeof(*dr), size, &tot_size)))
if (check_add_overflow(mul_u32_u32(height, fb->pitches[color_plane]),
if (check_add_overflow(fb->base.offsets[i], 64, &end)) {
check_add_overflow(args->offset, args->size, &offset_plus_size) ||
if (check_add_overflow(start, size - 1, &end))
if (check_add_overflow(offset, size, &offset_plus_size))
if (check_add_overflow(start, size - 1, &end))
if (check_add_overflow(init->kernel_managed_addr,
if (check_add_overflow(sizeof(*args), size, &args_size))
if (check_add_overflow(sizeof(*args), size, &args_size)) {
if (check_add_overflow(*offset, cmd->words, &next_offset)) {
if (check_add_overflow(vma->vm_pgoff << PAGE_SHIFT, vm_size, &vm_end))
if (check_add_overflow(user_pfn, (page_count - 1), &check))
if (check_add_overflow(offset, conn_param->private_data_len, &req.private_data_len))
if (check_add_overflow(offset, conn_param->private_data_len, &req.private_data_len))
if (check_add_overflow(xa_first, npages, &xa_last))
check_add_overflow(wrs_offset, wrs_size, &wrs_size))
if (check_add_overflow(curr_base, curr_len, &end) ||
if (check_add_overflow(offset, (unsigned long)size, &end))
if (check_add_overflow(umem_odp->umem.address,
if (check_add_overflow(size, pbundle->internal_used, &new_used))
if (check_add_overflow(offset, (u64)odp->umem.address,
if (check_add_overflow(*agg_avg, avg_bw, agg_avg))
if (check_add_overflow(iova, len - 1, &last))
if (WARN_ON(!phys || check_add_overflow(phys, size, &end)))
if (WARN_ON(check_add_overflow(iova, size, &end)))
if (!check_add_overflow(offset, pgsize_next, &offset_end) &&
WARN_ON(check_add_overflow(iova, length - 1, &last_iova)))
if (check_add_overflow(iova, length - 1, &last_iova))
if (check_add_overflow(iova, length - 1, &last_iova))
if (check_add_overflow(iova, length - 1, &last))
if (check_add_overflow(iova, length - 1, &area->node.last))
if (check_add_overflow(start_byte, length - 1, &area->pages_node.last))
if (check_add_overflow(start, length - 1, &last))
if (check_add_overflow(bitmap->iova, bitmap->length - 1, &last_iova))
if (check_add_overflow(iova, length - 1, &last_iova))
if (check_add_overflow(*start, iova_alignment - 1, &aligned_start))
if (check_add_overflow(iova, length - 1, &iova_last))
if (check_add_overflow((unsigned long)uptr, length, &end))
rc = check_add_overflow(pages->npinned, npages, &pages->npinned);
if (check_add_overflow(start, length - 1, &last))
check_add_overflow((uintptr_t)uptr, (uintptr_t)length, &end)) {
check_add_overflow((uintptr_t)uptr, (uintptr_t)length, &end))
if (check_add_overflow(offset, cmd->length, &length))
if (check_add_overflow(length, PAGE_SIZE - 1, &length))
if (check_add_overflow(cmd->nesting_parent_iova, cmd->length - 1,
if (check_add_overflow(bfr->iova, bfr->size - 1, &bfr_end))
if (check_add_overflow((u32)sizeof(*meta_hdr),
if (check_add_overflow(sem->u.sem.count, count, &sum) ||
if (check_add_overflow(req.start, req.len, &end) || end > mtd->size)
if (check_add_overflow(req.start, req.len, &end) || end > mtd->size) {
if (check_add_overflow(req.start, req.length, &end_peb))
if (check_add_overflow(pfa_ptr, pfa_len - 1, &max_tlv)) {
if (check_add_overflow(next_tlv, 2, &next_tlv) ||
check_add_overflow(next_tlv, tlv_len, &next_tlv)) {
if (check_add_overflow(start_qid, num_queues, &end_qid)) {
if (check_add_overflow(max_io_eqs, MLX5_ESW_MAX_CTRL_EQS, &max_eqs)) {
if (check_add_overflow((int)fixed, inlen, &inlen)) {
if (check_add_overflow(header_offset, EFX_SNICIMAGE_HEADER_MINLEN,
if (check_add_overflow(header_offset, header_len, &header_end))
if (check_add_overflow(header_end, payload_size, &payload_end) ||
if (check_add_overflow(header_offset, EFX_SNICBUNDLE_HEADER_LEN,
if (check_add_overflow(header_offset, EFX_REFLASH_HEADER_LENGTH_OFST +
if (check_add_overflow(header_offset, header_len, &trailer_offset) ||
check_add_overflow(trailer_offset, payload_size, &trailer_offset) ||
check_add_overflow(trailer_offset, EFX_REFLASH_TRAILER_LEN,
else if (size && check_add_overflow(r->start, size - 1, &r->end))
if (check_add_overflow(ktime_get_real_seconds(),
if (check_add_overflow(priv->sleep_sec, div64_u64(apival, priv->rtc_hz),
check_add_overflow(data_len, 8, &data_len) ||
check_add_overflow((loff_t)len, *pos, &requested_length))
if (check_add_overflow(state_size, state_size / 10, &inc_state_size))
check_add_overflow((loff_t)len, *pos, &requested_length))
if ((check_add_overflow(nvdev->usemem.memphys,
check_add_overflow(PFN_PHYS(pgoff), req_len, &end))
check_add_overflow((loff_t)len, *pos, &requested_length))
check_add_overflow(len, *offs, &end))
if (check_add_overflow(start, dma_ranges[i].offset,
check_add_overflow(phys_vec[i].paddr,
if (check_add_overflow(length, len, &length))
check_add_overflow(user_iova, iova_size - 1, &iova_end)))
if (check_add_overflow(iova, size - 1, &iova_end)) {
if (check_add_overflow(iova, size - 1, &iova_end) ||
check_add_overflow(vaddr, size - 1, &vaddr_end))
if (check_add_overflow(iova, size - 1, &iova_end))
check_add_overflow(user_iova, iova_size - 1, &iova_end))
if (check_add_overflow(range.iova, range.length, &iova_end) ||
if (check_add_overflow(report.iova, report.length, &iova_end) ||
if (check_add_overflow(FONT_EXTRA_WORDS * sizeof(int), size, &alloc_size))
check_add_overflow(range->start, range->len, &range_end))
if (unlikely(check_add_overflow(start, len, &offset) || offset > eb->len))
if (unlikely(check_add_overflow(btrfs_file_extent_num_bytes(leaf, fi),
if (unlikely(check_add_overflow(logical, length, &chunk_end))) {
!check_add_overflow(ofs, inode->i_size, &end) &&
!check_add_overflow(ofs, sizeof(*ih), &ofs) &&
if (unlikely(check_add_overflow(map->m_pa, map->m_plen, &pend) ||
if (check_add_overflow(max(bprm->argc, 1), bprm->envc, &ptr_size) ||
if (unlikely(check_add_overflow(sizeof(*sm), seq->count, &kbufsize)))
} else if (check_add_overflow(prev_lcn, dlcn, &lcn)) {
if (check_add_overflow(vcn64, len, &next_vcn))
if (check_add_overflow(vcn64, len, &vcn64))
if (check_add_overflow(offset, len, &sum))
if (check_add_overflow(pos, len, &tmp))
if (check_add_overflow(p->found_pages, n_pages, &total_pages) ||
if (unlikely(check_add_overflow(pos, len, &tmp)))
check_add_overflow((unsigned long)ea, next, &addr))
if (check_add_overflow(off, count, &len) || len > iov->iov_len)
if (unlikely(check_add_overflow(*noff, sizeof(*t_hdr), noff)))
if (check_add_overflow(off, rem, &len) || len > rsp_iov->iov_len)
if (check_add_overflow(doff, dlen, &len) || len > rem)
if (unlikely(check_add_overflow(input_len, sizeof(*req), &len) ||
check_add_overflow(dacloffset, sizeof(struct smb_acl), &dacl_struct_end) ||
check_add_overflow(dacl_offset, sizeof(struct smb_acl), &dacl_struct_end) ||
check_add_overflow(len, ppntsd_size - dacl_offset, &len))
check_add_overflow(len, tmp, &len))
if (check_add_overflow(len, 5 * sizeof(struct smb_ace), &len))
check_add_overflow(len, tmp, &len))
if (check_add_overflow(bloc->logicalBlockNum, offset, &blk) ||
check_add_overflow(blk, count, &blk) ||
if (check_add_overflow(sizeof(struct allocExtDesc),
if (check_add_overflow(map->s_partition_root, map->s_partition_len,
if (check_add_overflow(map->s_partition_len,
if (check_add_overflow(req->ip1_bcount, bmbt_blocks, &req->ip1_bcount))
if (check_add_overflow(req->ip2_bcount, bmbt_blocks, &req->ip2_bcount))
if (check_add_overflow(resblks, bmbt_blocks, &resblks))
if (check_add_overflow(resblks, bmbt_blocks, &resblks))
if (check_add_overflow(resblks, rmapbt_blocks, &resblks))
if (check_add_overflow(resblks, rmapbt_blocks, &resblks))
if (check_add_overflow(ifp->if_nextents, (uint64_t)delta,
if (glen == 0 || check_add_overflow(gbno, glen - 1, &end))
check_add_overflow(logres, step_size, &logres))
if (check_add_overflow(fxr->file1_offset, fxr->length, &tmp) ||
check_add_overflow(fxr->file2_offset, fxr->length, &tmp))
if (check_add_overflow(len, extra, &len))
check_add_overflow((x), v, &v); \
if (check_add_overflow(addend1, addend2, &bytes))
if (check_add_overflow(offset, size, \
if (check_add_overflow(offset, groupname##_##name##__sz,\
if (check_add_overflow(cl->filter_cnt, 1, &res))
if (check_add_overflow(range.off, range.len, &end))
if (unlikely(check_add_overflow(offset, size, &end) ||
if (check_add_overflow((unsigned long)p->addr, size, &tmp_check))
if (check_add_overflow(reg->user_addr, reg->size, &end))
if (check_add_overflow(uaddr, len, &end))
if (check_add_overflow(end, PAGE_SIZE - 1, &end))
if (check_add_overflow(sizeof(struct io_uring_recvmsg_out),
if (check_add_overflow(hdr, controllen, &hdr))
if (unlikely(check_add_overflow(buf_addr, (u64)len, &buf_end)))
if (check_add_overflow(arg->nr, arg->src_off, &off) || off > nbufs)
if (check_add_overflow(arg->nr, arg->dst_off, &nbufs))
if (unlikely(check_add_overflow(total_len, iov_len, &total_len)))
if (unlikely(check_add_overflow(total_len, iov[i].iov_len,
if (check_add_overflow(up->offset, nr_args, &tmp))
if (check_add_overflow(base, acct_len, &tmp))
if (check_add_overflow(start, t->size - 1, &end)) {
if (check_add_overflow(smin_ptr, smin_val, &dst_reg->smin_value) ||
check_add_overflow(smax_ptr, smax_val, &dst_reg->smax_value)) {
if (check_add_overflow(umin_ptr, umin_val, &dst_reg->umin_value) ||
check_add_overflow(umax_ptr, umax_val, &dst_reg->umax_value)) {
if (check_add_overflow(*dst_smin, src_reg->s32_min_value, dst_smin) ||
check_add_overflow(*dst_smax, src_reg->s32_max_value, dst_smax)) {
min_overflow = check_add_overflow(*dst_umin, umin_val, dst_umin);
max_overflow = check_add_overflow(*dst_umax, umax_val, dst_umax);
if (check_add_overflow(*dst_smin, src_reg->smin_value, dst_smin) ||
check_add_overflow(*dst_smax, src_reg->smax_value, dst_smax)) {
min_overflow = check_add_overflow(*dst_umin, umin_val, dst_umin);
max_overflow = check_add_overflow(*dst_umax, umax_val, dst_umax);
if (check_add_overflow(reg->umin_value, reg->off, &min_index) ||
if (check_add_overflow(reg->umax_value, reg->off, &max_index) ||
if (check_add_overflow(insn->imm, delta, &imm))
if (check_add_overflow(insn->off, delta, &off))
if (check_add_overflow(note_off, note_size, ¬e_end))
if (check_add_overflow(new_off, ALIGN(name_sz, 4), &new_off) ||
check_add_overflow(new_off, ALIGN(desc_sz, 4), &new_off) ||
if (check_add_overflow(ctx->bytecount_lo, len, &ctx->bytecount_lo))
if (check_add_overflow(start, len, &end))
check_add_overflow(start, length - 1,
if (check_add_overflow(addr, len, &addr_end))
if (check_add_overflow(size, off, &end_index) ||
if (check_add_overflow(old_nelems, trans->nelems, &nelems))
if (check_add_overflow(set->size, set->ndeact, &maxsize))
if (check_add_overflow(maxsize, delta, &maxsize))
if (check_add_overflow(priv->rate, priv->burst, &rate_with_burst))
check_add_overflow(rules, extra, &rules_alloc))
if (check_add_overflow(ret, NFT_PIPAPO_ALIGN_HEADROOM, &ret))
if (check_add_overflow(new_max_intervals, new_max_intervals,
if (check_add_overflow(new_max_intervals, delta,
if (check_add_overflow(head->rc_pageoff, len, &total))
if (check_add_overflow(addr, len, &end) || end > pool->addrs_cnt)
check_add_overflow((loff_t)len, *pos, &requested_length))
if (check_add_overflow(phys, (phys_addr_t)size - 1, &end_phys))
if (check_add_overflow(s->size, n, &new_capacity))
check_add_overflow(sizeof(*lctx), lctx->ctx_len, &required_len) ||
if (check_add_overflow(bytes, c, &bytes))
if (check_add_overflow(iova, size - 1, &last) ||
if (check_add_overflow(iova, size - 1, &last) ||