GFP_ATOMIC
return (linux_alloc_pages(GFP_ATOMIC, order));
ctxm->pg_info = kcalloc(n, sizeof(*ctxm->pg_info), GFP_ATOMIC);
resolve_dmac_work = kzalloc(sizeof(*resolve_dmac_work), GFP_ATOMIC);
dcb_work = kzalloc(sizeof(*dcb_work), GFP_ATOMIC);
dbq_work = kzalloc(sizeof(*dbq_work), GFP_ATOMIC);
dbq_work = kzalloc(sizeof(*dbq_work), GFP_ATOMIC);
dbq_work = kzalloc(sizeof(*dbq_work), GFP_ATOMIC);
nq_work = kzalloc(sizeof(*nq_work), GFP_ATOMIC);
nq_work = kzalloc(sizeof(*nq_work), GFP_ATOMIC);
reg_elem = ECORE_ZALLOC(sizeof(*reg_elem), GFP_ATOMIC, sc);
return ECORE_ZALLOC(sizeof(struct ecore_exeq_elem), GFP_ATOMIC,
new_cmd = ECORE_ZALLOC(total_sz, GFP_ATOMIC, sc);
elem = ECORE_CALLOC(len, sizeof(*elem), GFP_ATOMIC, sc);
if (!idr_pre_get(idr, lock ? GFP_KERNEL : GFP_ATOMIC))
post_terminate(qhp, NULL, internal ? GFP_ATOMIC : GFP_KERNEL);
__c4iw_ep_disconnect(ep, abort, internal ? GFP_ATOMIC :
new_send = kzalloc(sizeof(*new_send), GFP_ATOMIC);
child_listen_node = kzalloc(sizeof(*child_listen_node), GFP_ATOMIC);
cm_node = kzalloc(sizeof(*cm_node), GFP_ATOMIC);
work = kzalloc(sizeof(*work), GFP_ATOMIC);
event = kzalloc(sizeof(*event), GFP_ATOMIC);
entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
work = kzalloc(sizeof(*work), GFP_ATOMIC);
ah = kzalloc(sizeof(*ah), GFP_ATOMIC);
cmpl = kzalloc(sizeof(*cmpl), GFP_ATOMIC);
cmpl = kzalloc(sizeof(*cmpl), GFP_ATOMIC);
cqp_request = kzalloc(sizeof(*cqp_request), GFP_ATOMIC);
INIT_RADIX_TREE(&cq_table->tree, GFP_ATOMIC);
INIT_RADIX_TREE(&dev->qp_table_tree, GFP_ATOMIC);
INIT_RADIX_TREE(&srq_table->tree, GFP_ATOMIC);
filter = kzalloc(sizeof(struct mlx4_en_filter), GFP_ATOMIC);
tmp = kzalloc(sizeof(struct mlx4_en_addr_list), GFP_ATOMIC);
tmp = kzalloc(sizeof(struct mlx4_en_addr_list), GFP_ATOMIC);
IB_MGMT_MAD_DATA, GFP_ATOMIC,
dm = kcalloc(ports, sizeof(*dm), GFP_ATOMIC);
dm[i] = kmalloc(sizeof (struct mlx4_ib_demux_work), GFP_ATOMIC);
port_gid_table->gids[free].ctx = kmalloc(sizeof(*port_gid_table->gids[free].ctx), GFP_ATOMIC);
ew = kmalloc(sizeof(*ew), GFP_ATOMIC);
ew = kmalloc(sizeof(*ew), GFP_ATOMIC);
ew = kmalloc(sizeof *ew, GFP_ATOMIC);
gids = kmalloc(sizeof(*gids) * MLX4_MAX_PORT_GIDS, GFP_ATOMIC);
gids = kmalloc(sizeof(*gids) * MLX4_MAX_PORT_GIDS, GFP_ATOMIC);
work = kmalloc(sizeof(*work), GFP_ATOMIC);
gfp_t alloc_flags = cbk ? GFP_ATOMIC : GFP_KERNEL;
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
curr_match = kmalloc(sizeof(*curr_match), GFP_ATOMIC);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
req = kzalloc(sizeof(*req), GFP_ATOMIC);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
context = kzalloc(sizeof(*context), GFP_ATOMIC);
xfer_trans = kzalloc(sizeof(*xfer_trans), GFP_ATOMIC);
soft_wc = kmalloc(sizeof(*soft_wc), GFP_ATOMIC);
GFP_ATOMIC);
pas = (__be64 *)get_zeroed_page(GFP_ATOMIC);
array->page_list[p].page = (void **) get_zeroed_page(GFP_ATOMIC);
ah->av = kmalloc(sizeof *ah->av, GFP_ATOMIC);
av = kmalloc(sizeof *av, GFP_ATOMIC);
GFP_ATOMIC, &ah->avdma);
IB_MGMT_MAD_DATA, GFP_ATOMIC,
cq->resize_buf = kmalloc(sizeof *cq->resize_buf, GFP_ATOMIC);
ifp->needed_tailroom, GFP_ATOMIC);
reset_data = kzalloc(sizeof(*reset_data), GFP_ATOMIC);
wq_data = kzalloc(sizeof(*wq_data), GFP_ATOMIC);
hb_timer_data = kzalloc(sizeof(*hb_timer_data), GFP_ATOMIC);
stop_data = kzalloc(sizeof(*stop_data), GFP_ATOMIC);
p_hwfn->unzip_buf = OSAL_ZALLOC(p_hwfn->p_dev, GFP_ATOMIC,
p_cmd_elem = OSAL_ZALLOC(p_hwfn->p_dev, GFP_ATOMIC,
p_ent = OSAL_ZALLOC(p_hwfn->p_dev, GFP_ATOMIC, sizeof(*p_ent));
packet = kzalloc(sizeof(*packet), GFP_ATOMIC);
ep = kzalloc(sizeof(*ep), GFP_ATOMIC);
work = kzalloc(sizeof(*work), GFP_ATOMIC);
work = kmalloc(sizeof *work, GFP_ATOMIC);
GFP_ATOMIC,
work = kmalloc(sizeof *work, GFP_ATOMIC);
GFP_ATOMIC,
private_data = kzalloc(req.private_data_len, GFP_ATOMIC);
private_data = kzalloc(req.private_data_len, GFP_ATOMIC);
p = kmemdup(event->private_data, event->private_data_len, GFP_ATOMIC);
*method = kzalloc(sizeof **method, GFP_ATOMIC);
*class = kzalloc(sizeof **class, GFP_ATOMIC);
vendor = kzalloc(sizeof *vendor, GFP_ATOMIC);
vendor_class = kzalloc(sizeof *vendor_class, GFP_ATOMIC);
GFP_ATOMIC);
GFP_ATOMIC);
local = kmalloc(sizeof *local, GFP_ATOMIC);
mad_priv = alloc_mad_private(mad_size, GFP_ATOMIC);
entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
work = kmalloc(sizeof(*work), GFP_ATOMIC);
work = kmalloc(sizeof(*work), GFP_ATOMIC);
entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
(flags & RDMA_CREATE_AH_SLEEPABLE) ? GFP_KERNEL : GFP_ATOMIC);
tx = kzalloc(sizeof *tx, GFP_ATOMIC);
path = kzalloc(sizeof *path, GFP_ATOMIC);
1000, GFP_ATOMIC,
GFP_ATOMIC,
mcast = kzalloc(sizeof *mcast, can_sleep ? GFP_KERNEL : GFP_ATOMIC);