alloc_wr
ud = alloc_wr(next_size, user_wr->num_sge);
rdma = alloc_wr(next_size, user_wr->num_sge);
atomic = alloc_wr(next_size, user_wr->num_sge);
next = alloc_wr(next_size, user_wr->num_sge);
struct sk_buff *skb = alloc_wr(sizeof(struct ulp_mem_io) +
struct sk_buff *skb = alloc_wr(sizeof(struct ulp_mem_io) +
struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0,
struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0,
skb = alloc_wr(sizeof(*req), 0, GFP_ATOMIC);
skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_ATOMIC);
csk->cpl_close = alloc_wr(sizeof(struct cpl_close_con_req), 0,
csk->cpl_abort_req = alloc_wr(sizeof(struct cpl_abort_req), 0,
csk->cpl_abort_rpl = alloc_wr(sizeof(struct cpl_abort_rpl), 0,
skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_KERNEL);
skb = alloc_wr(size, 0, GFP_ATOMIC);
skb = alloc_wr(size6, 0, GFP_ATOMIC);
csk->cpl_close = alloc_wr(sizeof(struct cpl_close_con_req),
csk->cpl_abort_req = alloc_wr(sizeof(struct cpl_abort_req),
csk->cpl_abort_rpl = alloc_wr(sizeof(struct cpl_abort_rpl),
skb = alloc_wr(size, 0, GFP_NOIO);
skb = alloc_wr(size6, 0, GFP_NOIO);
struct sk_buff *skb = alloc_wr(wr_len, 0, GFP_ATOMIC);
skb = alloc_wr(sizeof(*req), 0, GFP_KERNEL);
skb = alloc_wr(sizeof(*req), 0, GFP_KERNEL);
skb = alloc_wr(len, 0, GFP_ATOMIC);
skb = alloc_wr(sizeof(*req), 0, GFP_ATOMIC);
skb = alloc_wr(flowclen, 0, GFP_ATOMIC);