@@ -92,7 +92,7 @@ void rxe_mw_cleanup(struct rxe_pool_elem *elem);
/* rxe_net.c */
struct sk_buff *rxe_init_packet(struct rxe_dev *rxe, struct rxe_av *av,
- int paylen, struct rxe_pkt_info *pkt);
+ struct rxe_pkt_info *pkt);
int rxe_prepare(struct rxe_av *av, struct rxe_pkt_info *pkt,
struct sk_buff *skb);
int rxe_xmit_packet(struct rxe_qp *qp, struct rxe_pkt_info *pkt,
@@ -443,7 +443,7 @@ int rxe_xmit_packet(struct rxe_qp *qp, struct rxe_pkt_info *pkt,
}
struct sk_buff *rxe_init_packet(struct rxe_dev *rxe, struct rxe_av *av,
- int paylen, struct rxe_pkt_info *pkt)
+ struct rxe_pkt_info *pkt)
{
unsigned int hdr_len;
struct sk_buff *skb = NULL;
@@ -468,7 +468,7 @@ struct sk_buff *rxe_init_packet(struct rxe_dev *rxe, struct rxe_av *av,
rcu_read_unlock();
goto out;
}
- skb = alloc_skb(paylen + hdr_len + LL_RESERVED_SPACE(ndev),
+ skb = alloc_skb(pkt->paylen + hdr_len + LL_RESERVED_SPACE(ndev),
GFP_ATOMIC);
if (unlikely(!skb)) {
@@ -489,7 +489,7 @@ struct sk_buff *rxe_init_packet(struct rxe_dev *rxe, struct rxe_av *av,
pkt->rxe = rxe;
pkt->port_num = port_num;
- pkt->hdr = skb_put(skb, paylen);
+ pkt->hdr = skb_put(skb, pkt->paylen);
pkt->mask |= RXE_GRH_MASK;
out:
@@ -456,51 +456,76 @@ static int rxe_init_payload(struct rxe_qp *qp, struct rxe_send_wqe *wqe,
return err;
}
-static struct sk_buff *init_req_packet(struct rxe_qp *qp,
- struct rxe_av *av,
- struct rxe_send_wqe *wqe,
- int opcode, u32 payload,
- struct rxe_pkt_info *pkt)
+static struct sk_buff *rxe_init_req_packet(struct rxe_qp *qp,
+ struct rxe_send_wqe *wqe,
+ int opcode, u32 payload,
+ struct rxe_pkt_info *pkt)
{
struct rxe_dev *rxe = to_rdev(qp->ibqp.device);
struct sk_buff *skb;
- int pad = (-payload) & 0x3;
- int paylen;
+ struct rxe_av *av;
+ struct rxe_ah *ah;
+ void *padp;
+ int pad;
+ int err = -EINVAL;
+
+ pkt->rxe = rxe;
+ pkt->opcode = opcode;
+ pkt->qp = qp;
+ pkt->psn = qp->req.psn;
+ pkt->mask = rxe_opcode[opcode].mask;
+ pkt->wqe = wqe;
+ pkt->port_num = 1;
+
+ /* get address vector and address handle for UD qps only */
+ av = rxe_get_av(pkt, &ah);
+ if (unlikely(!av))
+ goto err_out;
/* length from start of bth to end of icrc */
- paylen = rxe_opcode[opcode].length + payload + pad + RXE_ICRC_SIZE;
- pkt->paylen = paylen;
+ pad = (-payload) & 0x3;
+ pkt->paylen = rxe_opcode[opcode].length + payload +
+ pad + RXE_ICRC_SIZE;
/* init skb */
- skb = rxe_init_packet(rxe, av, paylen, pkt);
+ skb = rxe_init_packet(rxe, av, pkt);
if (unlikely(!skb))
- return NULL;
+ goto err_out;
rxe_init_roce_hdrs(qp, wqe, pkt, pad);
- return skb;
-}
+ if (pkt->mask & RXE_WRITE_OR_SEND_MASK) {
+ err = rxe_init_payload(qp, wqe, pkt, payload);
+ if (err)
+ goto err_out;
+ }
-static int finish_packet(struct rxe_qp *qp, struct rxe_av *av,
- struct rxe_send_wqe *wqe, struct rxe_pkt_info *pkt,
- struct sk_buff *skb, u32 payload)
-{
- int err;
+ if (pad) {
+ padp = payload_addr(pkt) + payload;
+ memset(padp, 0, pad);
+ }
+ /* IP and UDP network headers */
err = rxe_prepare(av, pkt, skb);
if (err)
- return err;
+ goto err_out;
- if (pkt->mask & RXE_WRITE_OR_SEND_MASK) {
- err = rxe_init_payload(qp, wqe, pkt, payload);
- if (bth_pad(pkt)) {
- u8 *pad = payload_addr(pkt) + payload;
+ if (ah)
+ rxe_put(ah);
- memset(pad, 0, bth_pad(pkt));
- }
- }
+ return skb;
- return 0;
+err_out:
+ if (err == -EFAULT)
+ wqe->status = IB_WC_LOC_PROT_ERR;
+ else
+ wqe->status = IB_WC_LOC_QP_OP_ERR;
+ if (skb)
+ kfree_skb(skb);
+ if (ah)
+ rxe_put(ah);
+
+ return NULL;
}
static void update_wqe_state(struct rxe_qp *qp,
@@ -630,7 +655,6 @@ static int rxe_do_local_ops(struct rxe_qp *qp, struct rxe_send_wqe *wqe)
int rxe_requester(void *arg)
{
struct rxe_qp *qp = (struct rxe_qp *)arg;
- struct rxe_dev *rxe = to_rdev(qp->ibqp.device);
struct rxe_pkt_info pkt;
struct sk_buff *skb;
struct rxe_send_wqe *wqe;
@@ -643,8 +667,6 @@ int rxe_requester(void *arg)
struct rxe_send_wqe rollback_wqe;
u32 rollback_psn;
struct rxe_queue *q = qp->sq.queue;
- struct rxe_ah *ah;
- struct rxe_av *av;
if (!rxe_get(qp))
return -EAGAIN;
@@ -753,44 +775,9 @@ int rxe_requester(void *arg)
payload = mtu;
}
- pkt.rxe = rxe;
- pkt.opcode = opcode;
- pkt.qp = qp;
- pkt.psn = qp->req.psn;
- pkt.mask = rxe_opcode[opcode].mask;
- pkt.wqe = wqe;
-
- av = rxe_get_av(&pkt, &ah);
- if (unlikely(!av)) {
- pr_err("qp#%d Failed no address vector\n", qp_num(qp));
- wqe->status = IB_WC_LOC_QP_OP_ERR;
- goto err;
- }
-
- skb = init_req_packet(qp, av, wqe, opcode, payload, &pkt);
- if (unlikely(!skb)) {
- pr_err("qp#%d Failed allocating skb\n", qp_num(qp));
- wqe->status = IB_WC_LOC_QP_OP_ERR;
- if (ah)
- rxe_put(ah);
- goto err;
- }
-
- err = finish_packet(qp, av, wqe, &pkt, skb, payload);
- if (unlikely(err)) {
- pr_debug("qp#%d Error during finish packet\n", qp_num(qp));
- if (err == -EFAULT)
- wqe->status = IB_WC_LOC_PROT_ERR;
- else
- wqe->status = IB_WC_LOC_QP_OP_ERR;
- kfree_skb(skb);
- if (ah)
- rxe_put(ah);
+ skb = rxe_init_req_packet(qp, wqe, opcode, payload, &pkt);
+ if (unlikely(!skb))
goto err;
- }
-
- if (ah)
- rxe_put(ah);
/*
* To prevent a race on wqe access between requester and completer,
@@ -665,22 +665,19 @@ static struct sk_buff *prepare_ack_packet(struct rxe_qp *qp,
int pad;
int err;
- /*
- * allocate packet
- */
pad = (-payload) & 0x3;
paylen = rxe_opcode[opcode].length + payload + pad + RXE_ICRC_SIZE;
- skb = rxe_init_packet(rxe, &qp->pri_av, paylen, ack);
- if (!skb)
- return NULL;
-
ack->qp = qp;
ack->opcode = opcode;
ack->mask = rxe_opcode[opcode].mask;
ack->paylen = paylen;
ack->psn = psn;
+ skb = rxe_init_packet(rxe, &qp->pri_av, ack);
+ if (!skb)
+ return NULL;
+
bth_init(ack, opcode, 0, 0, pad, IB_DEFAULT_PKEY_FULL,
qp->attr.dest_qp_num, 0, psn);
Isolate all the code to build a request packet into a single subroutine called rxe_init_req_packet(). Signed-off-by: Bob Pearson <rpearsonhpe@gmail.com> --- drivers/infiniband/sw/rxe/rxe_loc.h | 2 +- drivers/infiniband/sw/rxe/rxe_net.c | 6 +- drivers/infiniband/sw/rxe/rxe_req.c | 121 ++++++++++++--------------- drivers/infiniband/sw/rxe/rxe_resp.c | 11 +-- 4 files changed, 62 insertions(+), 78 deletions(-)