|
@@ -90,8 +90,7 @@ cxgb_mk_tid_release(struct sk_buff *skb, u32 len, u32 tid, u16 chan)
|
|
|
{
|
|
|
struct cpl_tid_release *req;
|
|
|
|
|
|
- req = __skb_put(skb, len);
|
|
|
- memset(req, 0, len);
|
|
|
+ req = __skb_put_zero(skb, len);
|
|
|
|
|
|
INIT_TP_WR(req, tid);
|
|
|
OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_TID_RELEASE, tid));
|
|
@@ -104,8 +103,7 @@ cxgb_mk_close_con_req(struct sk_buff *skb, u32 len, u32 tid, u16 chan,
|
|
|
{
|
|
|
struct cpl_close_con_req *req;
|
|
|
|
|
|
- req = __skb_put(skb, len);
|
|
|
- memset(req, 0, len);
|
|
|
+ req = __skb_put_zero(skb, len);
|
|
|
|
|
|
INIT_TP_WR(req, tid);
|
|
|
OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_CLOSE_CON_REQ, tid));
|
|
@@ -119,8 +117,7 @@ cxgb_mk_abort_req(struct sk_buff *skb, u32 len, u32 tid, u16 chan,
|
|
|
{
|
|
|
struct cpl_abort_req *req;
|
|
|
|
|
|
- req = __skb_put(skb, len);
|
|
|
- memset(req, 0, len);
|
|
|
+ req = __skb_put_zero(skb, len);
|
|
|
|
|
|
INIT_TP_WR(req, tid);
|
|
|
OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_ABORT_REQ, tid));
|
|
@@ -134,8 +131,7 @@ cxgb_mk_abort_rpl(struct sk_buff *skb, u32 len, u32 tid, u16 chan)
|
|
|
{
|
|
|
struct cpl_abort_rpl *rpl;
|
|
|
|
|
|
- rpl = __skb_put(skb, len);
|
|
|
- memset(rpl, 0, len);
|
|
|
+ rpl = __skb_put_zero(skb, len);
|
|
|
|
|
|
INIT_TP_WR(rpl, tid);
|
|
|
OPCODE_TID(rpl) = cpu_to_be32(MK_OPCODE_TID(CPL_ABORT_RPL, tid));
|
|
@@ -149,8 +145,7 @@ cxgb_mk_rx_data_ack(struct sk_buff *skb, u32 len, u32 tid, u16 chan,
|
|
|
{
|
|
|
struct cpl_rx_data_ack *req;
|
|
|
|
|
|
- req = __skb_put(skb, len);
|
|
|
- memset(req, 0, len);
|
|
|
+ req = __skb_put_zero(skb, len);
|
|
|
|
|
|
INIT_TP_WR(req, tid);
|
|
|
OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_RX_DATA_ACK, tid));
|