|
@@ -33,27 +33,165 @@
|
|
|
*/
|
|
|
|
|
|
#include "cxgb4.h"
|
|
|
+#include "t4_regs.h"
|
|
|
#include "l2t.h"
|
|
|
#include "t4fw_api.h"
|
|
|
#include "cxgb4_filter.h"
|
|
|
|
|
|
+static inline bool is_field_set(u32 val, u32 mask)
|
|
|
+{
|
|
|
+ return val || mask;
|
|
|
+}
|
|
|
+
|
|
|
+static inline bool unsupported(u32 conf, u32 conf_mask, u32 val, u32 mask)
|
|
|
+{
|
|
|
+ return !(conf & conf_mask) && is_field_set(val, mask);
|
|
|
+}
|
|
|
+
|
|
|
+/* Validate filter spec against configuration done on the card. */
|
|
|
+static int validate_filter(struct net_device *dev,
|
|
|
+ struct ch_filter_specification *fs)
|
|
|
+{
|
|
|
+ struct adapter *adapter = netdev2adap(dev);
|
|
|
+ u32 fconf, iconf;
|
|
|
+
|
|
|
+ /* Check for unconfigured fields being used. */
|
|
|
+ fconf = adapter->params.tp.vlan_pri_map;
|
|
|
+ iconf = adapter->params.tp.ingress_config;
|
|
|
+
|
|
|
+ if (unsupported(fconf, FCOE_F, fs->val.fcoe, fs->mask.fcoe) ||
|
|
|
+ unsupported(fconf, PORT_F, fs->val.iport, fs->mask.iport) ||
|
|
|
+ unsupported(fconf, TOS_F, fs->val.tos, fs->mask.tos) ||
|
|
|
+ unsupported(fconf, ETHERTYPE_F, fs->val.ethtype,
|
|
|
+ fs->mask.ethtype) ||
|
|
|
+ unsupported(fconf, MACMATCH_F, fs->val.macidx, fs->mask.macidx) ||
|
|
|
+ unsupported(fconf, MPSHITTYPE_F, fs->val.matchtype,
|
|
|
+ fs->mask.matchtype) ||
|
|
|
+ unsupported(fconf, FRAGMENTATION_F, fs->val.frag, fs->mask.frag) ||
|
|
|
+ unsupported(fconf, PROTOCOL_F, fs->val.proto, fs->mask.proto) ||
|
|
|
+ unsupported(fconf, VNIC_ID_F, fs->val.pfvf_vld,
|
|
|
+ fs->mask.pfvf_vld) ||
|
|
|
+ unsupported(fconf, VNIC_ID_F, fs->val.ovlan_vld,
|
|
|
+ fs->mask.ovlan_vld) ||
|
|
|
+ unsupported(fconf, VLAN_F, fs->val.ivlan_vld, fs->mask.ivlan_vld))
|
|
|
+ return -EOPNOTSUPP;
|
|
|
+
|
|
|
+ /* T4 inconveniently uses the same FT_VNIC_ID_W bits for both the Outer
|
|
|
+ * VLAN Tag and PF/VF/VFvld fields based on VNIC_F being set
|
|
|
+ * in TP_INGRESS_CONFIG. Hense the somewhat crazy checks
|
|
|
+ * below. Additionally, since the T4 firmware interface also
|
|
|
+ * carries that overlap, we need to translate any PF/VF
|
|
|
+ * specification into that internal format below.
|
|
|
+ */
|
|
|
+ if (is_field_set(fs->val.pfvf_vld, fs->mask.pfvf_vld) &&
|
|
|
+ is_field_set(fs->val.ovlan_vld, fs->mask.ovlan_vld))
|
|
|
+ return -EOPNOTSUPP;
|
|
|
+ if (unsupported(iconf, VNIC_F, fs->val.pfvf_vld, fs->mask.pfvf_vld) ||
|
|
|
+ (is_field_set(fs->val.ovlan_vld, fs->mask.ovlan_vld) &&
|
|
|
+ (iconf & VNIC_F)))
|
|
|
+ return -EOPNOTSUPP;
|
|
|
+ if (fs->val.pf > 0x7 || fs->val.vf > 0x7f)
|
|
|
+ return -ERANGE;
|
|
|
+ fs->mask.pf &= 0x7;
|
|
|
+ fs->mask.vf &= 0x7f;
|
|
|
+
|
|
|
+ /* If the user is requesting that the filter action loop
|
|
|
+ * matching packets back out one of our ports, make sure that
|
|
|
+ * the egress port is in range.
|
|
|
+ */
|
|
|
+ if (fs->action == FILTER_SWITCH &&
|
|
|
+ fs->eport >= adapter->params.nports)
|
|
|
+ return -ERANGE;
|
|
|
+
|
|
|
+ /* Don't allow various trivially obvious bogus out-of-range values... */
|
|
|
+ if (fs->val.iport >= adapter->params.nports)
|
|
|
+ return -ERANGE;
|
|
|
+
|
|
|
+ /* T4 doesn't support removing VLAN Tags for loop back filters. */
|
|
|
+ if (is_t4(adapter->params.chip) &&
|
|
|
+ fs->action == FILTER_SWITCH &&
|
|
|
+ (fs->newvlan == VLAN_REMOVE ||
|
|
|
+ fs->newvlan == VLAN_REWRITE))
|
|
|
+ return -EOPNOTSUPP;
|
|
|
+
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static unsigned int get_filter_steerq(struct net_device *dev,
|
|
|
+ struct ch_filter_specification *fs)
|
|
|
+{
|
|
|
+ struct adapter *adapter = netdev2adap(dev);
|
|
|
+ unsigned int iq;
|
|
|
+
|
|
|
+ /* If the user has requested steering matching Ingress Packets
|
|
|
+ * to a specific Queue Set, we need to make sure it's in range
|
|
|
+ * for the port and map that into the Absolute Queue ID of the
|
|
|
+ * Queue Set's Response Queue.
|
|
|
+ */
|
|
|
+ if (!fs->dirsteer) {
|
|
|
+ if (fs->iq)
|
|
|
+ return -EINVAL;
|
|
|
+ iq = 0;
|
|
|
+ } else {
|
|
|
+ struct port_info *pi = netdev_priv(dev);
|
|
|
+
|
|
|
+ /* If the iq id is greater than the number of qsets,
|
|
|
+ * then assume it is an absolute qid.
|
|
|
+ */
|
|
|
+ if (fs->iq < pi->nqsets)
|
|
|
+ iq = adapter->sge.ethrxq[pi->first_qset +
|
|
|
+ fs->iq].rspq.abs_id;
|
|
|
+ else
|
|
|
+ iq = fs->iq;
|
|
|
+ }
|
|
|
+
|
|
|
+ return iq;
|
|
|
+}
|
|
|
+
|
|
|
+static int cxgb4_set_ftid(struct tid_info *t, int fidx, int family)
|
|
|
+{
|
|
|
+ spin_lock_bh(&t->ftid_lock);
|
|
|
+
|
|
|
+ if (test_bit(fidx, t->ftid_bmap)) {
|
|
|
+ spin_unlock_bh(&t->ftid_lock);
|
|
|
+ return -EBUSY;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (family == PF_INET)
|
|
|
+ __set_bit(fidx, t->ftid_bmap);
|
|
|
+ else
|
|
|
+ bitmap_allocate_region(t->ftid_bmap, fidx, 2);
|
|
|
+
|
|
|
+ spin_unlock_bh(&t->ftid_lock);
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static void cxgb4_clear_ftid(struct tid_info *t, int fidx, int family)
|
|
|
+{
|
|
|
+ spin_lock_bh(&t->ftid_lock);
|
|
|
+ if (family == PF_INET)
|
|
|
+ __clear_bit(fidx, t->ftid_bmap);
|
|
|
+ else
|
|
|
+ bitmap_release_region(t->ftid_bmap, fidx, 2);
|
|
|
+ spin_unlock_bh(&t->ftid_lock);
|
|
|
+}
|
|
|
+
|
|
|
/* Delete the filter at a specified index. */
|
|
|
static int del_filter_wr(struct adapter *adapter, int fidx)
|
|
|
{
|
|
|
struct filter_entry *f = &adapter->tids.ftid_tab[fidx];
|
|
|
struct fw_filter_wr *fwr;
|
|
|
- unsigned int len, ftid;
|
|
|
struct sk_buff *skb;
|
|
|
+ unsigned int len;
|
|
|
|
|
|
len = sizeof(*fwr);
|
|
|
- ftid = adapter->tids.ftid_base + fidx;
|
|
|
|
|
|
skb = alloc_skb(len, GFP_KERNEL);
|
|
|
if (!skb)
|
|
|
return -ENOMEM;
|
|
|
|
|
|
fwr = (struct fw_filter_wr *)__skb_put(skb, len);
|
|
|
- t4_mk_filtdelwr(ftid, fwr, adapter->sge.fw_evtq.abs_id);
|
|
|
+ t4_mk_filtdelwr(f->tid, fwr, adapter->sge.fw_evtq.abs_id);
|
|
|
|
|
|
/* Mark the filter as "pending" and ship off the Filter Work Request.
|
|
|
* When we get the Work Request Reply we'll clear the pending status.
|
|
@@ -74,7 +212,6 @@ int set_filter_wr(struct adapter *adapter, int fidx)
|
|
|
struct filter_entry *f = &adapter->tids.ftid_tab[fidx];
|
|
|
struct fw_filter_wr *fwr;
|
|
|
struct sk_buff *skb;
|
|
|
- unsigned int ftid;
|
|
|
|
|
|
skb = alloc_skb(sizeof(*fwr), GFP_KERNEL);
|
|
|
if (!skb)
|
|
@@ -94,8 +231,6 @@ int set_filter_wr(struct adapter *adapter, int fidx)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- ftid = adapter->tids.ftid_base + fidx;
|
|
|
-
|
|
|
fwr = (struct fw_filter_wr *)__skb_put(skb, sizeof(*fwr));
|
|
|
memset(fwr, 0, sizeof(*fwr));
|
|
|
|
|
@@ -110,7 +245,7 @@ int set_filter_wr(struct adapter *adapter, int fidx)
|
|
|
fwr->op_pkd = htonl(FW_WR_OP_V(FW_FILTER_WR));
|
|
|
fwr->len16_pkd = htonl(FW_WR_LEN16_V(sizeof(*fwr) / 16));
|
|
|
fwr->tid_to_iq =
|
|
|
- htonl(FW_FILTER_WR_TID_V(ftid) |
|
|
|
+ htonl(FW_FILTER_WR_TID_V(f->tid) |
|
|
|
FW_FILTER_WR_RQTYPE_V(f->fs.type) |
|
|
|
FW_FILTER_WR_NOREPLY_V(0) |
|
|
|
FW_FILTER_WR_IQ_V(f->fs.iq));
|
|
@@ -235,33 +370,341 @@ void clear_filter(struct adapter *adap, struct filter_entry *f)
|
|
|
memset(f, 0, sizeof(*f));
|
|
|
}
|
|
|
|
|
|
+void clear_all_filters(struct adapter *adapter)
|
|
|
+{
|
|
|
+ unsigned int i;
|
|
|
+
|
|
|
+ if (adapter->tids.ftid_tab) {
|
|
|
+ struct filter_entry *f = &adapter->tids.ftid_tab[0];
|
|
|
+ unsigned int max_ftid = adapter->tids.nftids +
|
|
|
+ adapter->tids.nsftids;
|
|
|
+
|
|
|
+ for (i = 0; i < max_ftid; i++, f++)
|
|
|
+ if (f->valid || f->pending)
|
|
|
+ clear_filter(adapter, f);
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+/* Fill up default masks for set match fields. */
|
|
|
+static void fill_default_mask(struct ch_filter_specification *fs)
|
|
|
+{
|
|
|
+ unsigned int lip = 0, lip_mask = 0;
|
|
|
+ unsigned int fip = 0, fip_mask = 0;
|
|
|
+ unsigned int i;
|
|
|
+
|
|
|
+ if (fs->val.iport && !fs->mask.iport)
|
|
|
+ fs->mask.iport |= ~0;
|
|
|
+ if (fs->val.fcoe && !fs->mask.fcoe)
|
|
|
+ fs->mask.fcoe |= ~0;
|
|
|
+ if (fs->val.matchtype && !fs->mask.matchtype)
|
|
|
+ fs->mask.matchtype |= ~0;
|
|
|
+ if (fs->val.macidx && !fs->mask.macidx)
|
|
|
+ fs->mask.macidx |= ~0;
|
|
|
+ if (fs->val.ethtype && !fs->mask.ethtype)
|
|
|
+ fs->mask.ethtype |= ~0;
|
|
|
+ if (fs->val.ivlan && !fs->mask.ivlan)
|
|
|
+ fs->mask.ivlan |= ~0;
|
|
|
+ if (fs->val.ovlan && !fs->mask.ovlan)
|
|
|
+ fs->mask.ovlan |= ~0;
|
|
|
+ if (fs->val.frag && !fs->mask.frag)
|
|
|
+ fs->mask.frag |= ~0;
|
|
|
+ if (fs->val.tos && !fs->mask.tos)
|
|
|
+ fs->mask.tos |= ~0;
|
|
|
+ if (fs->val.proto && !fs->mask.proto)
|
|
|
+ fs->mask.proto |= ~0;
|
|
|
+
|
|
|
+ for (i = 0; i < ARRAY_SIZE(fs->val.lip); i++) {
|
|
|
+ lip |= fs->val.lip[i];
|
|
|
+ lip_mask |= fs->mask.lip[i];
|
|
|
+ fip |= fs->val.fip[i];
|
|
|
+ fip_mask |= fs->mask.fip[i];
|
|
|
+ }
|
|
|
+
|
|
|
+ if (lip && !lip_mask)
|
|
|
+ memset(fs->mask.lip, ~0, sizeof(fs->mask.lip));
|
|
|
+
|
|
|
+ if (fip && !fip_mask)
|
|
|
+ memset(fs->mask.fip, ~0, sizeof(fs->mask.lip));
|
|
|
+
|
|
|
+ if (fs->val.lport && !fs->mask.lport)
|
|
|
+ fs->mask.lport = ~0;
|
|
|
+ if (fs->val.fport && !fs->mask.fport)
|
|
|
+ fs->mask.fport = ~0;
|
|
|
+}
|
|
|
+
|
|
|
+/* Check a Chelsio Filter Request for validity, convert it into our internal
|
|
|
+ * format and send it to the hardware. Return 0 on success, an error number
|
|
|
+ * otherwise. We attach any provided filter operation context to the internal
|
|
|
+ * filter specification in order to facilitate signaling completion of the
|
|
|
+ * operation.
|
|
|
+ */
|
|
|
+int __cxgb4_set_filter(struct net_device *dev, int filter_id,
|
|
|
+ struct ch_filter_specification *fs,
|
|
|
+ struct filter_ctx *ctx)
|
|
|
+{
|
|
|
+ struct adapter *adapter = netdev2adap(dev);
|
|
|
+ unsigned int max_fidx, fidx, iq;
|
|
|
+ struct filter_entry *f;
|
|
|
+ u32 iconf;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ max_fidx = adapter->tids.nftids;
|
|
|
+ if (filter_id != (max_fidx + adapter->tids.nsftids - 1) &&
|
|
|
+ filter_id >= max_fidx)
|
|
|
+ return -E2BIG;
|
|
|
+
|
|
|
+ fill_default_mask(fs);
|
|
|
+
|
|
|
+ ret = validate_filter(dev, fs);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+
|
|
|
+ iq = get_filter_steerq(dev, fs);
|
|
|
+ if (iq < 0)
|
|
|
+ return iq;
|
|
|
+
|
|
|
+ /* IPv6 filters occupy four slots and must be aligned on
|
|
|
+ * four-slot boundaries. IPv4 filters only occupy a single
|
|
|
+ * slot and have no alignment requirements but writing a new
|
|
|
+ * IPv4 filter into the middle of an existing IPv6 filter
|
|
|
+ * requires clearing the old IPv6 filter and hence we prevent
|
|
|
+ * insertion.
|
|
|
+ */
|
|
|
+ if (fs->type == 0) { /* IPv4 */
|
|
|
+ /* If our IPv4 filter isn't being written to a
|
|
|
+ * multiple of four filter index and there's an IPv6
|
|
|
+ * filter at the multiple of 4 base slot, then we
|
|
|
+ * prevent insertion.
|
|
|
+ */
|
|
|
+ fidx = filter_id & ~0x3;
|
|
|
+ if (fidx != filter_id &&
|
|
|
+ adapter->tids.ftid_tab[fidx].fs.type) {
|
|
|
+ f = &adapter->tids.ftid_tab[fidx];
|
|
|
+ if (f->valid) {
|
|
|
+ dev_err(adapter->pdev_dev,
|
|
|
+ "Invalid location. IPv6 requires 4 slots and is occupying slots %u to %u\n",
|
|
|
+ fidx, fidx + 3);
|
|
|
+ return -EINVAL;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ } else { /* IPv6 */
|
|
|
+ /* Ensure that the IPv6 filter is aligned on a
|
|
|
+ * multiple of 4 boundary.
|
|
|
+ */
|
|
|
+ if (filter_id & 0x3) {
|
|
|
+ dev_err(adapter->pdev_dev,
|
|
|
+ "Invalid location. IPv6 must be aligned on a 4-slot boundary\n");
|
|
|
+ return -EINVAL;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Check all except the base overlapping IPv4 filter slots. */
|
|
|
+ for (fidx = filter_id + 1; fidx < filter_id + 4; fidx++) {
|
|
|
+ f = &adapter->tids.ftid_tab[fidx];
|
|
|
+ if (f->valid) {
|
|
|
+ dev_err(adapter->pdev_dev,
|
|
|
+ "Invalid location. IPv6 requires 4 slots and an IPv4 filter exists at %u\n",
|
|
|
+ fidx);
|
|
|
+ return -EINVAL;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Check to make sure that provided filter index is not
|
|
|
+ * already in use by someone else
|
|
|
+ */
|
|
|
+ f = &adapter->tids.ftid_tab[filter_id];
|
|
|
+ if (f->valid)
|
|
|
+ return -EBUSY;
|
|
|
+
|
|
|
+ fidx = filter_id + adapter->tids.ftid_base;
|
|
|
+ ret = cxgb4_set_ftid(&adapter->tids, filter_id,
|
|
|
+ fs->type ? PF_INET6 : PF_INET);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+
|
|
|
+ /* Check to make sure the filter requested is writable ... */
|
|
|
+ ret = writable_filter(f);
|
|
|
+ if (ret) {
|
|
|
+ /* Clear the bits we have set above */
|
|
|
+ cxgb4_clear_ftid(&adapter->tids, filter_id,
|
|
|
+ fs->type ? PF_INET6 : PF_INET);
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Clear out any old resources being used by the filter before
|
|
|
+ * we start constructing the new filter.
|
|
|
+ */
|
|
|
+ if (f->valid)
|
|
|
+ clear_filter(adapter, f);
|
|
|
+
|
|
|
+ /* Convert the filter specification into our internal format.
|
|
|
+ * We copy the PF/VF specification into the Outer VLAN field
|
|
|
+ * here so the rest of the code -- including the interface to
|
|
|
+ * the firmware -- doesn't have to constantly do these checks.
|
|
|
+ */
|
|
|
+ f->fs = *fs;
|
|
|
+ f->fs.iq = iq;
|
|
|
+ f->dev = dev;
|
|
|
+
|
|
|
+ iconf = adapter->params.tp.ingress_config;
|
|
|
+ if (iconf & VNIC_F) {
|
|
|
+ f->fs.val.ovlan = (fs->val.pf << 13) | fs->val.vf;
|
|
|
+ f->fs.mask.ovlan = (fs->mask.pf << 13) | fs->mask.vf;
|
|
|
+ f->fs.val.ovlan_vld = fs->val.pfvf_vld;
|
|
|
+ f->fs.mask.ovlan_vld = fs->mask.pfvf_vld;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Attempt to set the filter. If we don't succeed, we clear
|
|
|
+ * it and return the failure.
|
|
|
+ */
|
|
|
+ f->ctx = ctx;
|
|
|
+ f->tid = fidx; /* Save the actual tid */
|
|
|
+ ret = set_filter_wr(adapter, filter_id);
|
|
|
+ if (ret) {
|
|
|
+ cxgb4_clear_ftid(&adapter->tids, filter_id,
|
|
|
+ fs->type ? PF_INET6 : PF_INET);
|
|
|
+ clear_filter(adapter, f);
|
|
|
+ }
|
|
|
+
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+/* Check a delete filter request for validity and send it to the hardware.
|
|
|
+ * Return 0 on success, an error number otherwise. We attach any provided
|
|
|
+ * filter operation context to the internal filter specification in order to
|
|
|
+ * facilitate signaling completion of the operation.
|
|
|
+ */
|
|
|
+int __cxgb4_del_filter(struct net_device *dev, int filter_id,
|
|
|
+ struct filter_ctx *ctx)
|
|
|
+{
|
|
|
+ struct adapter *adapter = netdev2adap(dev);
|
|
|
+ struct filter_entry *f;
|
|
|
+ unsigned int max_fidx;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ max_fidx = adapter->tids.nftids;
|
|
|
+ if (filter_id != (max_fidx + adapter->tids.nsftids - 1) &&
|
|
|
+ filter_id >= max_fidx)
|
|
|
+ return -E2BIG;
|
|
|
+
|
|
|
+ f = &adapter->tids.ftid_tab[filter_id];
|
|
|
+ ret = writable_filter(f);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+
|
|
|
+ if (f->valid) {
|
|
|
+ f->ctx = ctx;
|
|
|
+ cxgb4_clear_ftid(&adapter->tids, filter_id,
|
|
|
+ f->fs.type ? PF_INET6 : PF_INET);
|
|
|
+ return del_filter_wr(adapter, filter_id);
|
|
|
+ }
|
|
|
+
|
|
|
+ /* If the caller has passed in a Completion Context then we need to
|
|
|
+ * mark it as a successful completion so they don't stall waiting
|
|
|
+ * for it.
|
|
|
+ */
|
|
|
+ if (ctx) {
|
|
|
+ ctx->result = 0;
|
|
|
+ complete(&ctx->completion);
|
|
|
+ }
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+int cxgb4_set_filter(struct net_device *dev, int filter_id,
|
|
|
+ struct ch_filter_specification *fs)
|
|
|
+{
|
|
|
+ struct filter_ctx ctx;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ init_completion(&ctx.completion);
|
|
|
+
|
|
|
+ ret = __cxgb4_set_filter(dev, filter_id, fs, &ctx);
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ /* Wait for reply */
|
|
|
+ ret = wait_for_completion_timeout(&ctx.completion, 10 * HZ);
|
|
|
+ if (!ret)
|
|
|
+ return -ETIMEDOUT;
|
|
|
+
|
|
|
+ ret = ctx.result;
|
|
|
+out:
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+int cxgb4_del_filter(struct net_device *dev, int filter_id)
|
|
|
+{
|
|
|
+ struct filter_ctx ctx;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ init_completion(&ctx.completion);
|
|
|
+
|
|
|
+ ret = __cxgb4_del_filter(dev, filter_id, &ctx);
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ /* Wait for reply */
|
|
|
+ ret = wait_for_completion_timeout(&ctx.completion, 10 * HZ);
|
|
|
+ if (!ret)
|
|
|
+ return -ETIMEDOUT;
|
|
|
+
|
|
|
+ ret = ctx.result;
|
|
|
+out:
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
/* Handle a filter write/deletion reply. */
|
|
|
void filter_rpl(struct adapter *adap, const struct cpl_set_tcb_rpl *rpl)
|
|
|
{
|
|
|
- unsigned int idx = GET_TID(rpl);
|
|
|
- unsigned int nidx = idx - adap->tids.ftid_base;
|
|
|
- struct filter_entry *f;
|
|
|
- unsigned int ret;
|
|
|
+ unsigned int tid = GET_TID(rpl);
|
|
|
+ struct filter_entry *f = NULL;
|
|
|
+ unsigned int max_fidx;
|
|
|
+ int idx;
|
|
|
|
|
|
- if (idx >= adap->tids.ftid_base && nidx <
|
|
|
- (adap->tids.nftids + adap->tids.nsftids)) {
|
|
|
- idx = nidx;
|
|
|
- ret = TCB_COOKIE_G(rpl->cookie);
|
|
|
+ max_fidx = adap->tids.nftids + adap->tids.nsftids;
|
|
|
+ /* Get the corresponding filter entry for this tid */
|
|
|
+ if (adap->tids.ftid_tab) {
|
|
|
+ /* Check this in normal filter region */
|
|
|
+ idx = tid - adap->tids.ftid_base;
|
|
|
+ if (idx >= max_fidx)
|
|
|
+ return;
|
|
|
f = &adap->tids.ftid_tab[idx];
|
|
|
+ if (f->tid != tid)
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* We found the filter entry for this tid */
|
|
|
+ if (f) {
|
|
|
+ unsigned int ret = TCB_COOKIE_G(rpl->cookie);
|
|
|
+ struct filter_ctx *ctx;
|
|
|
+
|
|
|
+ /* Pull off any filter operation context attached to the
|
|
|
+ * filter.
|
|
|
+ */
|
|
|
+ ctx = f->ctx;
|
|
|
+ f->ctx = NULL;
|
|
|
|
|
|
if (ret == FW_FILTER_WR_FLT_DELETED) {
|
|
|
/* Clear the filter when we get confirmation from the
|
|
|
* hardware that the filter has been deleted.
|
|
|
*/
|
|
|
clear_filter(adap, f);
|
|
|
+ if (ctx)
|
|
|
+ ctx->result = 0;
|
|
|
} else if (ret == FW_FILTER_WR_SMT_TBL_FULL) {
|
|
|
dev_err(adap->pdev_dev, "filter %u setup failed due to full SMT\n",
|
|
|
idx);
|
|
|
clear_filter(adap, f);
|
|
|
+ if (ctx)
|
|
|
+ ctx->result = -ENOMEM;
|
|
|
} else if (ret == FW_FILTER_WR_FLT_ADDED) {
|
|
|
f->smtidx = (be64_to_cpu(rpl->oldval) >> 24) & 0xff;
|
|
|
f->pending = 0; /* asynchronous setup completed */
|
|
|
f->valid = 1;
|
|
|
+ if (ctx) {
|
|
|
+ ctx->result = 0;
|
|
|
+ ctx->tid = idx;
|
|
|
+ }
|
|
|
} else {
|
|
|
/* Something went wrong. Issue a warning about the
|
|
|
* problem and clear everything out.
|
|
@@ -269,6 +712,10 @@ void filter_rpl(struct adapter *adap, const struct cpl_set_tcb_rpl *rpl)
|
|
|
dev_err(adap->pdev_dev, "filter %u setup failed with error %u\n",
|
|
|
idx, ret);
|
|
|
clear_filter(adap, f);
|
|
|
+ if (ctx)
|
|
|
+ ctx->result = -EINVAL;
|
|
|
}
|
|
|
+ if (ctx)
|
|
|
+ complete(&ctx->completion);
|
|
|
}
|
|
|
}
|