"csk 0x%p,%u,0x%lx,%u, credit %u, dack %u.\n",
csk, csk->state, csk->flags, csk->tid, credits, dack);
- skb = alloc_cpl(sizeof(*req), 0, GFP_ATOMIC);
+ skb = alloc_wr(sizeof(*req), 0, GFP_ATOMIC);
if (!skb) {
pr_info("csk 0x%p, credit %u, OOM.\n", csk, credits);
return 0;
cxgbi_sock_get(csk);
spin_lock_bh(&csk->lock);
- skb = alloc_cpl(sizeof(struct cpl_act_open_req), 0, GFP_ATOMIC);
+ skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_ATOMIC);
if (!skb)
cxgbi_sock_fail_act_open(csk, -ENOMEM);
else {
*/
static int alloc_cpls(struct cxgbi_sock *csk)
{
- csk->cpl_close = alloc_cpl(sizeof(struct cpl_close_con_req), 0,
+ csk->cpl_close = alloc_wr(sizeof(struct cpl_close_con_req), 0,
GFP_KERNEL);
if (!csk->cpl_close)
return -ENOMEM;
- csk->cpl_abort_req = alloc_cpl(sizeof(struct cpl_abort_req), 0,
+ csk->cpl_abort_req = alloc_wr(sizeof(struct cpl_abort_req), 0,
GFP_KERNEL);
if (!csk->cpl_abort_req)
goto free_cpl_skbs;
- csk->cpl_abort_rpl = alloc_cpl(sizeof(struct cpl_abort_rpl), 0,
+ csk->cpl_abort_rpl = alloc_wr(sizeof(struct cpl_abort_rpl), 0,
GFP_KERNEL);
if (!csk->cpl_abort_rpl)
goto free_cpl_skbs;
cxgbi_sock_set_flag(csk, CTPF_HAS_ATID);
cxgbi_sock_get(csk);
- skb = alloc_cpl(sizeof(struct cpl_act_open_req), 0, GFP_KERNEL);
+ skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_KERNEL);
if (!skb)
goto rel_resource;
skb->sk = (struct sock *)csk;
"ddp 0x%p, idx %d, cnt %d.\n", ddp, idx, cnt);
for (i = 0; i < cnt; i++) {
- struct sk_buff *skb = alloc_cpl(sizeof(struct ulp_mem_io) +
+ struct sk_buff *skb = alloc_wr(sizeof(struct ulp_mem_io) +
PPOD_SIZE, 0, gfp);
if (skb) {
ddp->gl_skb[idx + i] = skb;
static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk,
unsigned int tid, int pg_idx, bool reply)
{
- struct sk_buff *skb = alloc_cpl(sizeof(struct cpl_set_tcb_field), 0,
+ struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0,
GFP_KERNEL);
struct cpl_set_tcb_field *req;
u64 val = pg_idx < DDP_PGIDX_MAX ? pg_idx : 0;
static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid,
int hcrc, int dcrc, int reply)
{
- struct sk_buff *skb = alloc_cpl(sizeof(struct cpl_set_tcb_field), 0,
+ struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0,
GFP_KERNEL);
struct cpl_set_tcb_field *req;
u64 val = (hcrc ? 1 : 0) | (dcrc ? 2 : 0);
"csk 0x%p,%u,0x%lx,%u, credit %u.\n",
csk, csk->state, csk->flags, csk->tid, credits);
- skb = alloc_cpl(sizeof(*req), 0, GFP_ATOMIC);
+ skb = alloc_wr(sizeof(*req), 0, GFP_ATOMIC);
if (!skb) {
pr_info("csk 0x%p, credit %u, OOM.\n", csk, credits);
return 0;
int flowclen, i;
flowclen = 80;
- skb = alloc_cpl(flowclen, 0, GFP_ATOMIC);
+ skb = alloc_wr(flowclen, 0, GFP_ATOMIC);
flowc = (struct fw_flowc_wr *)skb->head;
flowc->op_to_nparams =
htonl(FW_WR_OP(FW_FLOWC_WR) | FW_FLOWC_WR_NPARAMS(8));
cxgbi_sock_get(csk);
spin_lock_bh(&csk->lock);
- skb = alloc_cpl(sizeof(struct cpl_act_open_req), 0, GFP_ATOMIC);
+ skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_ATOMIC);
if (!skb)
cxgbi_sock_fail_act_open(csk, -ENOMEM);
else {
static int alloc_cpls(struct cxgbi_sock *csk)
{
- csk->cpl_close = alloc_cpl(sizeof(struct cpl_close_con_req),
- 0, GFP_NOIO);
+ csk->cpl_close = alloc_wr(sizeof(struct cpl_close_con_req),
+ 0, GFP_KERNEL);
if (!csk->cpl_close)
return -ENOMEM;
- csk->cpl_abort_req = alloc_cpl(sizeof(struct cpl_abort_req),
- 0, GFP_NOIO);
+ csk->cpl_abort_req = alloc_wr(sizeof(struct cpl_abort_req),
+ 0, GFP_KERNEL);
if (!csk->cpl_abort_req)
goto free_cpls;
- csk->cpl_abort_rpl = alloc_cpl(sizeof(struct cpl_abort_rpl),
- 0, GFP_NOIO);
+ csk->cpl_abort_rpl = alloc_wr(sizeof(struct cpl_abort_rpl),
+ 0, GFP_KERNEL);
if (!csk->cpl_abort_rpl)
goto free_cpls;
return 0;
}
cxgbi_sock_get(csk);
- skb = alloc_cpl(sizeof(struct cpl_act_open_req), 0, GFP_NOIO);
+ skb = alloc_wr(sizeof(struct cpl_act_open_req), 0, GFP_KERNEL);
if (!skb)
goto rel_resource;
skb->sk = (struct sock *)csk;
dlen = PPOD_SIZE * npods;
pm_addr = idx * PPOD_SIZE + ddp->llimit;
- skb = alloc_cpl(sizeof(*req) + sizeof(*sgl), dlen, GFP_ATOMIC);
+ skb = alloc_wr(sizeof(*req) + sizeof(*sgl), dlen, GFP_ATOMIC);
if (!skb) {
pr_err("cdev 0x%p, idx %u, npods %u, OOM.\n",
cdev, idx, npods);
if (!pg_idx)
return 0;
- skb = alloc_cpl(sizeof(*req), 0, GFP_KERNEL);
+ skb = alloc_wr(sizeof(*req), 0, GFP_KERNEL);
if (!skb)
return -ENOMEM;
val = TCB_ULP_RAW(val);
val |= TCB_ULP_TYPE(ULP2_MODE_ISCSI);
- skb = alloc_cpl(sizeof(*req), 0, GFP_KERNEL);
+ skb = alloc_wr(sizeof(*req), 0, GFP_KERNEL);
if (!skb)
return -ENOMEM;
if (pgl == NULL) {
unsigned int len = 64 - sizeof(struct rsp_ctrl) - 8;
- skb = alloc_cpl(len, 0, GFP_ATOMIC);
+ skb = alloc_wr(len, 0, GFP_ATOMIC);
if (!skb)
goto nomem;
skb_copy_to_linear_data(skb, &rsp[1], len);
return wscale;
}
-static inline struct sk_buff *alloc_cpl(int cpl_len, int dlen, gfp_t gfp)
+static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp)
{
- int wrlen = roundup(cpl_len, 16);
struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp);
if (skb) {
__skb_put(skb, wrlen);
memset(skb->head, 0, wrlen + dlen);
} else
- pr_info("alloc cpl skb %u+%u, OOM.\n", cpl_len, dlen);
+ pr_info("alloc cpl wr skb %u+%u, OOM.\n", wrlen, dlen);
return skb;
}