kmalloc_flex
b = kmalloc_flex(*b, data, size);
cmem = kmalloc_flex(*cmem, ranges, nr_ranges);
cmem = kmalloc_flex(*cmem, ranges, nr_ranges);
ctx->switch_log = kmalloc_flex(*ctx->switch_log, log,
cmem = kmalloc_flex(*cmem, ranges, nr_ranges);
ib = kmalloc_flex(*ib, data, nr_ptrs, GFP_DMA | GFP_KERNEL);
bia = kmalloc_flex(*bia, bvecs, nr_vecs, gfp_mask);
bmd = kmalloc_flex(*bmd, iov, data->nr_segs, gfp_mask);
et = kmalloc_flex(*et, tags, nr_tags, gfp);
conn = kmalloc_flex(*conn, remote_name, csi2_res_src_length + 1);
buf = kmalloc_flex(*buf, sg, pages);
fw_counters = kmalloc_flex(*fw_counters, ae_counters, ae_count);
r = kmalloc_flex(*r, data, a->length);
mock = kmalloc_flex(*mock, pages, npages);
msg = kmalloc_flex(*msg, msg, num_dwords, GFP_ATOMIC);
p = kmalloc_flex(*p, seqno, KSYNCMAP);
objs = kmalloc_flex(*objs, objs, nents);
ctx = kmalloc_flex(*ctx, msg.payload, payload_size, GFP_ATOMIC);
pkey_cache = kmalloc_flex(*pkey_cache, table,
work = kmalloc_flex(*work, path, paths);
dev = kmalloc_flex(*dev, port, device->phys_port_cnt);
table = kmalloc_flex(*table, icm, num_icm);
db_tab = kmalloc_flex(*db_tab, page, npages);
chunk = kmalloc_flex(*chunk, page_list,
sc = kmalloc_flex(*sc, stripe, stripes);
r10b = kmalloc_flex(*r10b, devs, conf->copies, GFP_NOIO);
q = kmalloc_flex(*q, entries, total_count);
dg_info = kmalloc_flex(*dg_info, msg_payload,
array = kmalloc_flex(*array, entries, capacity, GFP_ATOMIC);
newcfi = kmalloc_flex(*newcfi, chips, numvirtchips);
cfi->cfiq = kmalloc_flex(*cfi->cfiq, EraseRegionInfo, num_erase_regions);
retcfi = kmalloc_flex(*retcfi, chips, cfi.numchips);
cfi->cfiq = kmalloc_flex(*cfi->cfiq, EraseRegionInfo, num_erase_regions);
mc_list = kmalloc_flex(*mc_list, mc,
msg_ind = kmalloc_flex(*msg_ind, msg, len, GFP_ATOMIC);
res = kmalloc_flex(*res, cmd.data, len, GFP_ATOMIC);
new_data = kmalloc_flex(*new_data, data, len, GFP_ATOMIC);
event = kmalloc_flex(*event, data, length,
desc = kmalloc_flex(*desc, nsids, NVMET_MAX_NAMESPACES);
stb_data_arr = kmalloc_flex(*stb_data_arr, data, fsize);
stb_data_arr = kmalloc_flex(*stb_data_arr, data, fsize);
rpl_16byte_wwid_list = kmalloc_flex(*rpl_16byte_wwid_list, lun_entries,
ctx->peripherals = kmalloc_flex(*ctx->peripherals, array, num_slaves);
ctx->peripherals = kmalloc_flex(*ctx->peripherals, array, num_slaves);
raw_data = kmalloc_flex(*raw_data, data, len);
tmr = kmalloc_flex(*tmr, tmr_cmd_ids, cmd_cnt, GFP_NOIO);
buf = kmalloc_flex(*buf, buf, maxframe, GFP_ATOMIC);
buf = kmalloc_flex(*buf, buf, maxframe);
p = kmalloc_flex(*p, data, 2 * size, GFP_ATOMIC | __GFP_NOWARN);
urb = kmalloc_flex(*urb, iso_frame_desc, iso_packets, mem_flags);
buf = kmalloc_flex(*buf, storage, data_len);
pkt = kmalloc_flex(*pkt, data, data_length, GFP_ATOMIC);
rb = kmalloc_flex(*rb, msg, len);
preflist = kmalloc_flex(*preflist, prefs, max_prefs);
acl = kmalloc_flex(*acl, data, size);
acl = kmalloc_flex(*acl, data, size);
acl = kmalloc_flex(*acl, data, size);
acl = kmalloc_flex(*acl, data, size);
item = kmalloc_flex(*item, data, data_len, GFP_NOFS);
perm = kmalloc_flex(*perm, pool_ns, pool_ns_len + 1, GFP_NOFS);
(*ea_inode_array) = kmalloc_flex(**ea_inode_array, inodes,
new_array = kmalloc_flex(**ea_inode_array, inodes,
handle = kmalloc_flex(*handle, f_handle, f_handle.handle_bytes);
header = kmalloc_flex(*header, a_entries, acl->a_count);
p = kmalloc_flex(*p, buckets, n, gfp_flags);
tb = kmalloc_flex(*tb, buf, len);
p = kmalloc_flex(*p, name, len + 1);
acl = kmalloc_flex(*acl, a_entries, count, flags);
walk = walk->next = kmalloc_flex(*walk, entries, len);
de = kmalloc_flex(*de, name, nlen);
sig_data = kmalloc_flex(*sig_data, ctx, lsmctx.len);
tree = kmalloc_flex(*tree, pathname, sz);
kip = kmalloc_flex(*kip, slot_used, slots_per_page(c));
struct tp_probes *p = kmalloc_flex(*p, probes, count);
new_s = kmalloc_flex(*new_s, index_key, keylen);
stats = kmalloc_flex(*stats, stacks, limit, GFP_NOWAIT);
mlru = kmalloc_flex(*mlru, node, nr_node_ids, gfp);
anon_name = kmalloc_flex(*anon_name, name, count);
new = kmalloc_flex(*new, entries, size, GFP_KERNEL_ACCOUNT);
new_wi_state = kmalloc_flex(*new_wi_state, iw_table, nr_node_ids);
rp = kmalloc_flex(*rp, addr, i);
rp = kmalloc_flex(*rp, entry, count, GFP_ATOMIC);
monmap = kmalloc_flex(*monmap, mon_inst, num_mon, GFP_NOIO);
req = kmalloc_flex(*req, r_ops, num_ops, gfp_flags);
r = kmalloc_flex(*r, steps, yes, GFP_NOFS);
return kmalloc_flex(struct flush_backlogs, w, nr_cpu_ids);
md_dst = kmalloc_flex(*md_dst, u.tun_info.options, optslen, flags);
saved_syn = kmalloc_flex(*saved_syn, data, len, GFP_ATOMIC);
new = kmalloc_flex(*new, f, old->field_count, GFP_KERNEL_ACCOUNT);
m = kmalloc_flex(*m, f, field_count);
e = kmalloc_flex(*e, stamps, nstamps_max, GFP_ATOMIC);
stab = kmalloc_flex(*stab, data, tsize);
chunk = kmalloc_flex(*chunk, ch_segments, segcount);
dfitem = kmalloc_flex(*dfitem, data, data_len, GFP_ATOMIC);
str = kmalloc_flex(*str, name, size, gfp);
cache = kmalloc_flex(*cache, str, str_len, GFP_ATOMIC);
ctx = kmalloc_flex(*ctx, ctx_str, str_len, GFP_ATOMIC);
ctx = kmalloc_flex(*ctx, ctx_str, str_len + 1, gfp);
p = kmalloc_flex(*p, conns, len);
new_bus = kmalloc_flex(*bus, range, bus->dev_count + 1,
new_bus = kmalloc_flex(*bus, range, bus->dev_count - 1,