|
@@ -3,6 +3,7 @@
|
|
* Copyright (c) 2016 Mellanox Technologies. All rights reserved.
|
|
* Copyright (c) 2016 Mellanox Technologies. All rights reserved.
|
|
* Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
|
|
* Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
|
|
* Copyright (c) 2016 Ido Schimmel <idosch@mellanox.com>
|
|
* Copyright (c) 2016 Ido Schimmel <idosch@mellanox.com>
|
|
|
|
+ * Copyright (c) 2016 Yotam Gigi <yotamg@mellanox.com>
|
|
*
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
* modification, are permitted provided that the following conditions are met:
|
|
@@ -38,6 +39,10 @@
|
|
#include <linux/rhashtable.h>
|
|
#include <linux/rhashtable.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/in6.h>
|
|
#include <linux/in6.h>
|
|
|
|
+#include <linux/notifier.h>
|
|
|
|
+#include <net/netevent.h>
|
|
|
|
+#include <net/neighbour.h>
|
|
|
|
+#include <net/arp.h>
|
|
|
|
|
|
#include "spectrum.h"
|
|
#include "spectrum.h"
|
|
#include "core.h"
|
|
#include "core.h"
|
|
@@ -112,6 +117,8 @@ enum mlxsw_sp_fib_entry_type {
|
|
MLXSW_SP_FIB_ENTRY_TYPE_TRAP,
|
|
MLXSW_SP_FIB_ENTRY_TYPE_TRAP,
|
|
};
|
|
};
|
|
|
|
|
|
|
|
+struct mlxsw_sp_nexthop_group;
|
|
|
|
+
|
|
struct mlxsw_sp_fib_entry {
|
|
struct mlxsw_sp_fib_entry {
|
|
struct rhash_head ht_node;
|
|
struct rhash_head ht_node;
|
|
struct mlxsw_sp_fib_key key;
|
|
struct mlxsw_sp_fib_key key;
|
|
@@ -119,6 +126,8 @@ struct mlxsw_sp_fib_entry {
|
|
u8 added:1;
|
|
u8 added:1;
|
|
u16 rif; /* used for action local */
|
|
u16 rif; /* used for action local */
|
|
struct mlxsw_sp_vr *vr;
|
|
struct mlxsw_sp_vr *vr;
|
|
|
|
+ struct list_head nexthop_group_node;
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_group;
|
|
};
|
|
};
|
|
|
|
|
|
struct mlxsw_sp_fib {
|
|
struct mlxsw_sp_fib {
|
|
@@ -544,6 +553,949 @@ static void mlxsw_sp_vrs_init(struct mlxsw_sp *mlxsw_sp)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+struct mlxsw_sp_neigh_key {
|
|
|
|
+ unsigned char addr[sizeof(struct in6_addr)];
|
|
|
|
+ struct net_device *dev;
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+struct mlxsw_sp_neigh_entry {
|
|
|
|
+ struct rhash_head ht_node;
|
|
|
|
+ struct mlxsw_sp_neigh_key key;
|
|
|
|
+ u16 rif;
|
|
|
|
+ struct neighbour *n;
|
|
|
|
+ bool offloaded;
|
|
|
|
+ struct delayed_work dw;
|
|
|
|
+ struct mlxsw_sp_port *mlxsw_sp_port;
|
|
|
|
+ unsigned char ha[ETH_ALEN];
|
|
|
|
+ struct list_head nexthop_list; /* list of nexthops using
|
|
|
|
+ * this neigh entry
|
|
|
|
+ */
|
|
|
|
+ struct list_head nexthop_neighs_list_node;
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+static const struct rhashtable_params mlxsw_sp_neigh_ht_params = {
|
|
|
|
+ .key_offset = offsetof(struct mlxsw_sp_neigh_entry, key),
|
|
|
|
+ .head_offset = offsetof(struct mlxsw_sp_neigh_entry, ht_node),
|
|
|
|
+ .key_len = sizeof(struct mlxsw_sp_neigh_key),
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+static int
|
|
|
|
+mlxsw_sp_neigh_entry_insert(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry)
|
|
|
|
+{
|
|
|
|
+ return rhashtable_insert_fast(&mlxsw_sp->router.neigh_ht,
|
|
|
|
+ &neigh_entry->ht_node,
|
|
|
|
+ mlxsw_sp_neigh_ht_params);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_neigh_entry_remove(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry)
|
|
|
|
+{
|
|
|
|
+ rhashtable_remove_fast(&mlxsw_sp->router.neigh_ht,
|
|
|
|
+ &neigh_entry->ht_node,
|
|
|
|
+ mlxsw_sp_neigh_ht_params);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neigh_update_hw(struct work_struct *work);
|
|
|
|
+
|
|
|
|
+static struct mlxsw_sp_neigh_entry *
|
|
|
|
+mlxsw_sp_neigh_entry_create(const void *addr, size_t addr_len,
|
|
|
|
+ struct net_device *dev, u16 rif,
|
|
|
|
+ struct neighbour *n)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+
|
|
|
|
+ neigh_entry = kzalloc(sizeof(*neigh_entry), GFP_ATOMIC);
|
|
|
|
+ if (!neigh_entry)
|
|
|
|
+ return NULL;
|
|
|
|
+ memcpy(neigh_entry->key.addr, addr, addr_len);
|
|
|
|
+ neigh_entry->key.dev = dev;
|
|
|
|
+ neigh_entry->rif = rif;
|
|
|
|
+ neigh_entry->n = n;
|
|
|
|
+ INIT_DELAYED_WORK(&neigh_entry->dw, mlxsw_sp_router_neigh_update_hw);
|
|
|
|
+ INIT_LIST_HEAD(&neigh_entry->nexthop_list);
|
|
|
|
+ return neigh_entry;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_neigh_entry_destroy(struct mlxsw_sp_neigh_entry *neigh_entry)
|
|
|
|
+{
|
|
|
|
+ kfree(neigh_entry);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct mlxsw_sp_neigh_entry *
|
|
|
|
+mlxsw_sp_neigh_entry_lookup(struct mlxsw_sp *mlxsw_sp, const void *addr,
|
|
|
|
+ size_t addr_len, struct net_device *dev)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_key key = {{ 0 } };
|
|
|
|
+
|
|
|
|
+ memcpy(key.addr, addr, addr_len);
|
|
|
|
+ key.dev = dev;
|
|
|
|
+ return rhashtable_lookup_fast(&mlxsw_sp->router.neigh_ht,
|
|
|
|
+ &key, mlxsw_sp_neigh_ht_params);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+int mlxsw_sp_router_neigh_construct(struct net_device *dev,
|
|
|
|
+ struct neighbour *n)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_port *mlxsw_sp_port = netdev_priv(dev);
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+ struct mlxsw_sp_rif *r;
|
|
|
|
+ u32 dip;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ if (n->tbl != &arp_tbl)
|
|
|
|
+ return 0;
|
|
|
|
+
|
|
|
|
+ dip = ntohl(*((__be32 *) n->primary_key));
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, &dip, sizeof(dip),
|
|
|
|
+ n->dev);
|
|
|
|
+ if (neigh_entry) {
|
|
|
|
+ WARN_ON(neigh_entry->n != n);
|
|
|
|
+ return 0;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ r = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
|
|
|
|
+ if (WARN_ON(!r))
|
|
|
|
+ return -EINVAL;
|
|
|
|
+
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_create(&dip, sizeof(dip), n->dev,
|
|
|
|
+ r->rif, n);
|
|
|
|
+ if (!neigh_entry)
|
|
|
|
+ return -ENOMEM;
|
|
|
|
+ err = mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
|
|
|
|
+ if (err)
|
|
|
|
+ goto err_neigh_entry_insert;
|
|
|
|
+ return 0;
|
|
|
|
+
|
|
|
|
+err_neigh_entry_insert:
|
|
|
|
+ mlxsw_sp_neigh_entry_destroy(neigh_entry);
|
|
|
|
+ return err;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+void mlxsw_sp_router_neigh_destroy(struct net_device *dev,
|
|
|
|
+ struct neighbour *n)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_port *mlxsw_sp_port = netdev_priv(dev);
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+ u32 dip;
|
|
|
|
+
|
|
|
|
+ if (n->tbl != &arp_tbl)
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ dip = ntohl(*((__be32 *) n->primary_key));
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, &dip, sizeof(dip),
|
|
|
|
+ n->dev);
|
|
|
|
+ if (!neigh_entry)
|
|
|
|
+ return;
|
|
|
|
+ mlxsw_sp_neigh_entry_remove(mlxsw_sp, neigh_entry);
|
|
|
|
+ mlxsw_sp_neigh_entry_destroy(neigh_entry);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_router_neighs_update_interval_init(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ unsigned long interval = NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME);
|
|
|
|
+
|
|
|
|
+ mlxsw_sp->router.neighs_update.interval = jiffies_to_msecs(interval);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neigh_ent_ipv4_process(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ char *rauhtd_pl,
|
|
|
|
+ int ent_index)
|
|
|
|
+{
|
|
|
|
+ struct net_device *dev;
|
|
|
|
+ struct neighbour *n;
|
|
|
|
+ __be32 dipn;
|
|
|
|
+ u32 dip;
|
|
|
|
+ u16 rif;
|
|
|
|
+
|
|
|
|
+ mlxsw_reg_rauhtd_ent_ipv4_unpack(rauhtd_pl, ent_index, &rif, &dip);
|
|
|
|
+
|
|
|
|
+ if (!mlxsw_sp->rifs[rif]) {
|
|
|
|
+ dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
|
|
|
|
+ return;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ dipn = htonl(dip);
|
|
|
|
+ dev = mlxsw_sp->rifs[rif]->dev;
|
|
|
|
+ n = neigh_lookup(&arp_tbl, &dipn, dev);
|
|
|
|
+ if (!n) {
|
|
|
|
+ netdev_err(dev, "Failed to find matching neighbour for IP=%pI4h\n",
|
|
|
|
+ &dip);
|
|
|
|
+ return;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ netdev_dbg(dev, "Updating neighbour with IP=%pI4h\n", &dip);
|
|
|
|
+ neigh_event_send(n, NULL);
|
|
|
|
+ neigh_release(n);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neigh_rec_ipv4_process(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ char *rauhtd_pl,
|
|
|
|
+ int rec_index)
|
|
|
|
+{
|
|
|
|
+ u8 num_entries;
|
|
|
|
+ int i;
|
|
|
|
+
|
|
|
|
+ num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
|
|
|
|
+ rec_index);
|
|
|
|
+ /* Hardware starts counting at 0, so add 1. */
|
|
|
|
+ num_entries++;
|
|
|
|
+
|
|
|
|
+ /* Each record consists of several neighbour entries. */
|
|
|
|
+ for (i = 0; i < num_entries; i++) {
|
|
|
|
+ int ent_index;
|
|
|
|
+
|
|
|
|
+ ent_index = rec_index * MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC + i;
|
|
|
|
+ mlxsw_sp_router_neigh_ent_ipv4_process(mlxsw_sp, rauhtd_pl,
|
|
|
|
+ ent_index);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neigh_rec_process(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ char *rauhtd_pl, int rec_index)
|
|
|
|
+{
|
|
|
|
+ switch (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, rec_index)) {
|
|
|
|
+ case MLXSW_REG_RAUHTD_TYPE_IPV4:
|
|
|
|
+ mlxsw_sp_router_neigh_rec_ipv4_process(mlxsw_sp, rauhtd_pl,
|
|
|
|
+ rec_index);
|
|
|
|
+ break;
|
|
|
|
+ case MLXSW_REG_RAUHTD_TYPE_IPV6:
|
|
|
|
+ WARN_ON_ONCE(1);
|
|
|
|
+ break;
|
|
|
|
+ }
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ char *rauhtd_pl;
|
|
|
|
+ u8 num_rec;
|
|
|
|
+ int i, err;
|
|
|
|
+
|
|
|
|
+ rauhtd_pl = kmalloc(MLXSW_REG_RAUHTD_LEN, GFP_KERNEL);
|
|
|
|
+ if (!rauhtd_pl)
|
|
|
|
+ return -ENOMEM;
|
|
|
|
+
|
|
|
|
+ /* Make sure the neighbour's netdev isn't removed in the
|
|
|
|
+ * process.
|
|
|
|
+ */
|
|
|
|
+ rtnl_lock();
|
|
|
|
+ do {
|
|
|
|
+ mlxsw_reg_rauhtd_pack(rauhtd_pl, MLXSW_REG_RAUHTD_TYPE_IPV4);
|
|
|
|
+ err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(rauhtd),
|
|
|
|
+ rauhtd_pl);
|
|
|
|
+ if (err) {
|
|
|
|
+ dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to dump neighbour talbe\n");
|
|
|
|
+ break;
|
|
|
|
+ }
|
|
|
|
+ num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
|
|
|
|
+ for (i = 0; i < num_rec; i++)
|
|
|
|
+ mlxsw_sp_router_neigh_rec_process(mlxsw_sp, rauhtd_pl,
|
|
|
|
+ i);
|
|
|
|
+ } while (num_rec);
|
|
|
|
+ rtnl_unlock();
|
|
|
|
+
|
|
|
|
+ kfree(rauhtd_pl);
|
|
|
|
+ return err;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neighs_update_nh(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+
|
|
|
|
+ /* Take RTNL mutex here to prevent lists from changes */
|
|
|
|
+ rtnl_lock();
|
|
|
|
+ list_for_each_entry(neigh_entry, &mlxsw_sp->router.nexthop_neighs_list,
|
|
|
|
+ nexthop_neighs_list_node) {
|
|
|
|
+ /* If this neigh have nexthops, make the kernel think this neigh
|
|
|
|
+ * is active regardless of the traffic.
|
|
|
|
+ */
|
|
|
|
+ if (!list_empty(&neigh_entry->nexthop_list))
|
|
|
|
+ neigh_event_send(neigh_entry->n, NULL);
|
|
|
|
+ }
|
|
|
|
+ rtnl_unlock();
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_router_neighs_update_work_schedule(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ unsigned long interval = mlxsw_sp->router.neighs_update.interval;
|
|
|
|
+
|
|
|
|
+ mlxsw_core_schedule_dw(&mlxsw_sp->router.neighs_update.dw,
|
|
|
|
+ msecs_to_jiffies(interval));
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neighs_update_work(struct work_struct *work)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp = container_of(work, struct mlxsw_sp,
|
|
|
|
+ router.neighs_update.dw.work);
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ err = mlxsw_sp_router_neighs_update_rauhtd(mlxsw_sp);
|
|
|
|
+ if (err)
|
|
|
|
+ dev_err(mlxsw_sp->bus_info->dev, "Could not update kernel for neigh activity");
|
|
|
|
+
|
|
|
|
+ mlxsw_sp_router_neighs_update_nh(mlxsw_sp);
|
|
|
|
+
|
|
|
|
+ mlxsw_sp_router_neighs_update_work_schedule(mlxsw_sp);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_probe_unresolved_nexthops(struct work_struct *work)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp = container_of(work, struct mlxsw_sp,
|
|
|
|
+ router.nexthop_probe_dw.work);
|
|
|
|
+
|
|
|
|
+ /* Iterate over nexthop neighbours, find those who are unresolved and
|
|
|
|
+ * send arp on them. This solves the chicken-egg problem when
|
|
|
|
+ * the nexthop wouldn't get offloaded until the neighbor is resolved
|
|
|
|
+ * but it wouldn't get resolved ever in case traffic is flowing in HW
|
|
|
|
+ * using different nexthop.
|
|
|
|
+ *
|
|
|
|
+ * Take RTNL mutex here to prevent lists from changes.
|
|
|
|
+ */
|
|
|
|
+ rtnl_lock();
|
|
|
|
+ list_for_each_entry(neigh_entry, &mlxsw_sp->router.nexthop_neighs_list,
|
|
|
|
+ nexthop_neighs_list_node) {
|
|
|
|
+ if (!(neigh_entry->n->nud_state & NUD_VALID) &&
|
|
|
|
+ !list_empty(&neigh_entry->nexthop_list))
|
|
|
|
+ neigh_event_send(neigh_entry->n, NULL);
|
|
|
|
+ }
|
|
|
|
+ rtnl_unlock();
|
|
|
|
+
|
|
|
|
+ mlxsw_core_schedule_dw(&mlxsw_sp->router.nexthop_probe_dw,
|
|
|
|
+ MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry,
|
|
|
|
+ bool removing);
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_router_neigh_update_hw(struct work_struct *work)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry =
|
|
|
|
+ container_of(work, struct mlxsw_sp_neigh_entry, dw.work);
|
|
|
|
+ struct neighbour *n = neigh_entry->n;
|
|
|
|
+ struct mlxsw_sp_port *mlxsw_sp_port = neigh_entry->mlxsw_sp_port;
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
|
|
|
|
+ char rauht_pl[MLXSW_REG_RAUHT_LEN];
|
|
|
|
+ struct net_device *dev;
|
|
|
|
+ bool entry_connected;
|
|
|
|
+ u8 nud_state;
|
|
|
|
+ bool updating;
|
|
|
|
+ bool removing;
|
|
|
|
+ bool adding;
|
|
|
|
+ u32 dip;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ read_lock_bh(&n->lock);
|
|
|
|
+ dip = ntohl(*((__be32 *) n->primary_key));
|
|
|
|
+ memcpy(neigh_entry->ha, n->ha, sizeof(neigh_entry->ha));
|
|
|
|
+ nud_state = n->nud_state;
|
|
|
|
+ dev = n->dev;
|
|
|
|
+ read_unlock_bh(&n->lock);
|
|
|
|
+
|
|
|
|
+ entry_connected = nud_state & NUD_VALID;
|
|
|
|
+ adding = (!neigh_entry->offloaded) && entry_connected;
|
|
|
|
+ updating = neigh_entry->offloaded && entry_connected;
|
|
|
|
+ removing = neigh_entry->offloaded && !entry_connected;
|
|
|
|
+
|
|
|
|
+ if (adding || updating) {
|
|
|
|
+ mlxsw_reg_rauht_pack4(rauht_pl, MLXSW_REG_RAUHT_OP_WRITE_ADD,
|
|
|
|
+ neigh_entry->rif,
|
|
|
|
+ neigh_entry->ha, dip);
|
|
|
|
+ err = mlxsw_reg_write(mlxsw_sp->core,
|
|
|
|
+ MLXSW_REG(rauht), rauht_pl);
|
|
|
|
+ if (err) {
|
|
|
|
+ netdev_err(dev, "Could not add neigh %pI4h\n", &dip);
|
|
|
|
+ neigh_entry->offloaded = false;
|
|
|
|
+ } else {
|
|
|
|
+ neigh_entry->offloaded = true;
|
|
|
|
+ }
|
|
|
|
+ mlxsw_sp_nexthop_neigh_update(mlxsw_sp, neigh_entry, false);
|
|
|
|
+ } else if (removing) {
|
|
|
|
+ mlxsw_reg_rauht_pack4(rauht_pl, MLXSW_REG_RAUHT_OP_WRITE_DELETE,
|
|
|
|
+ neigh_entry->rif,
|
|
|
|
+ neigh_entry->ha, dip);
|
|
|
|
+ err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht),
|
|
|
|
+ rauht_pl);
|
|
|
|
+ if (err) {
|
|
|
|
+ netdev_err(dev, "Could not delete neigh %pI4h\n", &dip);
|
|
|
|
+ neigh_entry->offloaded = true;
|
|
|
|
+ } else {
|
|
|
|
+ neigh_entry->offloaded = false;
|
|
|
|
+ }
|
|
|
|
+ mlxsw_sp_nexthop_neigh_update(mlxsw_sp, neigh_entry, true);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ neigh_release(n);
|
|
|
|
+ mlxsw_sp_port_dev_put(mlxsw_sp_port);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_router_netevent_event(struct notifier_block *unused,
|
|
|
|
+ unsigned long event, void *ptr)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+ struct mlxsw_sp_port *mlxsw_sp_port;
|
|
|
|
+ struct mlxsw_sp *mlxsw_sp;
|
|
|
|
+ unsigned long interval;
|
|
|
|
+ struct net_device *dev;
|
|
|
|
+ struct neigh_parms *p;
|
|
|
|
+ struct neighbour *n;
|
|
|
|
+ u32 dip;
|
|
|
|
+
|
|
|
|
+ switch (event) {
|
|
|
|
+ case NETEVENT_DELAY_PROBE_TIME_UPDATE:
|
|
|
|
+ p = ptr;
|
|
|
|
+
|
|
|
|
+ /* We don't care about changes in the default table. */
|
|
|
|
+ if (!p->dev || p->tbl != &arp_tbl)
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+
|
|
|
|
+ /* We are in atomic context and can't take RTNL mutex,
|
|
|
|
+ * so use RCU variant to walk the device chain.
|
|
|
|
+ */
|
|
|
|
+ mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(p->dev);
|
|
|
|
+ if (!mlxsw_sp_port)
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+
|
|
|
|
+ mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
|
|
|
|
+ interval = jiffies_to_msecs(NEIGH_VAR(p, DELAY_PROBE_TIME));
|
|
|
|
+ mlxsw_sp->router.neighs_update.interval = interval;
|
|
|
|
+
|
|
|
|
+ mlxsw_sp_port_dev_put(mlxsw_sp_port);
|
|
|
|
+ break;
|
|
|
|
+ case NETEVENT_NEIGH_UPDATE:
|
|
|
|
+ n = ptr;
|
|
|
|
+ dev = n->dev;
|
|
|
|
+
|
|
|
|
+ if (n->tbl != &arp_tbl)
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+
|
|
|
|
+ mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(dev);
|
|
|
|
+ if (!mlxsw_sp_port)
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+
|
|
|
|
+ mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
|
|
|
|
+ dip = ntohl(*((__be32 *) n->primary_key));
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp,
|
|
|
|
+ &dip,
|
|
|
|
+ sizeof(__be32),
|
|
|
|
+ dev);
|
|
|
|
+ if (WARN_ON(!neigh_entry) || WARN_ON(neigh_entry->n != n)) {
|
|
|
|
+ mlxsw_sp_port_dev_put(mlxsw_sp_port);
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+ }
|
|
|
|
+ neigh_entry->mlxsw_sp_port = mlxsw_sp_port;
|
|
|
|
+
|
|
|
|
+ /* Take a reference to ensure the neighbour won't be
|
|
|
|
+ * destructed until we drop the reference in delayed
|
|
|
|
+ * work.
|
|
|
|
+ */
|
|
|
|
+ neigh_clone(n);
|
|
|
|
+ if (!mlxsw_core_schedule_dw(&neigh_entry->dw, 0)) {
|
|
|
|
+ neigh_release(n);
|
|
|
|
+ mlxsw_sp_port_dev_put(mlxsw_sp_port);
|
|
|
|
+ }
|
|
|
|
+ break;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ return NOTIFY_DONE;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct notifier_block mlxsw_sp_router_netevent_nb __read_mostly = {
|
|
|
|
+ .notifier_call = mlxsw_sp_router_netevent_event,
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_neigh_init(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ err = rhashtable_init(&mlxsw_sp->router.neigh_ht,
|
|
|
|
+ &mlxsw_sp_neigh_ht_params);
|
|
|
|
+ if (err)
|
|
|
|
+ return err;
|
|
|
|
+
|
|
|
|
+ /* Initialize the polling interval according to the default
|
|
|
|
+ * table.
|
|
|
|
+ */
|
|
|
|
+ mlxsw_sp_router_neighs_update_interval_init(mlxsw_sp);
|
|
|
|
+
|
|
|
|
+ err = register_netevent_notifier(&mlxsw_sp_router_netevent_nb);
|
|
|
|
+ if (err)
|
|
|
|
+ goto err_register_netevent_notifier;
|
|
|
|
+
|
|
|
|
+ /* Create the delayed works for the activity_update */
|
|
|
|
+ INIT_DELAYED_WORK(&mlxsw_sp->router.neighs_update.dw,
|
|
|
|
+ mlxsw_sp_router_neighs_update_work);
|
|
|
|
+ INIT_DELAYED_WORK(&mlxsw_sp->router.nexthop_probe_dw,
|
|
|
|
+ mlxsw_sp_router_probe_unresolved_nexthops);
|
|
|
|
+ mlxsw_core_schedule_dw(&mlxsw_sp->router.neighs_update.dw, 0);
|
|
|
|
+ mlxsw_core_schedule_dw(&mlxsw_sp->router.nexthop_probe_dw, 0);
|
|
|
|
+ return 0;
|
|
|
|
+
|
|
|
|
+err_register_netevent_notifier:
|
|
|
|
+ rhashtable_destroy(&mlxsw_sp->router.neigh_ht);
|
|
|
|
+ return err;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_neigh_fini(struct mlxsw_sp *mlxsw_sp)
|
|
|
|
+{
|
|
|
|
+ cancel_delayed_work_sync(&mlxsw_sp->router.neighs_update.dw);
|
|
|
|
+ cancel_delayed_work_sync(&mlxsw_sp->router.nexthop_probe_dw);
|
|
|
|
+ unregister_netevent_notifier(&mlxsw_sp_router_netevent_nb);
|
|
|
|
+ rhashtable_destroy(&mlxsw_sp->router.neigh_ht);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+struct mlxsw_sp_nexthop {
|
|
|
|
+ struct list_head neigh_list_node; /* member of neigh entry list */
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp; /* pointer back to the group
|
|
|
|
+ * this belongs to
|
|
|
|
+ */
|
|
|
|
+ u8 should_offload:1, /* set indicates this neigh is connected and
|
|
|
|
+ * should be put to KVD linear area of this group.
|
|
|
|
+ */
|
|
|
|
+ offloaded:1, /* set in case the neigh is actually put into
|
|
|
|
+ * KVD linear area of this group.
|
|
|
|
+ */
|
|
|
|
+ update:1; /* set indicates that MAC of this neigh should be
|
|
|
|
+ * updated in HW
|
|
|
|
+ */
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+struct mlxsw_sp_nexthop_group {
|
|
|
|
+ struct list_head list; /* node in mlxsw->router.nexthop_group_list */
|
|
|
|
+ struct list_head fib_list; /* list of fib entries that use this group */
|
|
|
|
+ u8 adj_index_valid:1;
|
|
|
|
+ u32 adj_index;
|
|
|
|
+ u16 ecmp_size;
|
|
|
|
+ u16 count;
|
|
|
|
+ struct mlxsw_sp_nexthop nexthops[0];
|
|
|
|
+};
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_adj_index_mass_update_vr(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_vr *vr,
|
|
|
|
+ u32 adj_index, u16 ecmp_size,
|
|
|
|
+ u32 new_adj_index,
|
|
|
|
+ u16 new_ecmp_size)
|
|
|
|
+{
|
|
|
|
+ char raleu_pl[MLXSW_REG_RALEU_LEN];
|
|
|
|
+
|
|
|
|
+ mlxsw_reg_raleu_pack(raleu_pl, vr->proto, vr->id,
|
|
|
|
+ adj_index, ecmp_size,
|
|
|
|
+ new_adj_index, new_ecmp_size);
|
|
|
|
+ return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raleu), raleu_pl);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_adj_index_mass_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp,
|
|
|
|
+ u32 old_adj_index, u16 old_ecmp_size)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry;
|
|
|
|
+ struct mlxsw_sp_vr *vr = NULL;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
|
|
|
|
+ if (vr == fib_entry->vr)
|
|
|
|
+ continue;
|
|
|
|
+ vr = fib_entry->vr;
|
|
|
|
+ err = mlxsw_sp_adj_index_mass_update_vr(mlxsw_sp, vr,
|
|
|
|
+ old_adj_index,
|
|
|
|
+ old_ecmp_size,
|
|
|
|
+ nh_grp->adj_index,
|
|
|
|
+ nh_grp->ecmp_size);
|
|
|
|
+ if (err)
|
|
|
|
+ return err;
|
|
|
|
+ }
|
|
|
|
+ return 0;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_nexthop_mac_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
|
|
|
|
+ struct mlxsw_sp_nexthop *nh)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
|
|
|
|
+ char ratr_pl[MLXSW_REG_RATR_LEN];
|
|
|
|
+
|
|
|
|
+ mlxsw_reg_ratr_pack(ratr_pl, MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY,
|
|
|
|
+ true, adj_index, neigh_entry->rif);
|
|
|
|
+ mlxsw_reg_ratr_eth_entry_pack(ratr_pl, neigh_entry->ha);
|
|
|
|
+ return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int
|
|
|
|
+mlxsw_sp_nexthop_group_mac_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp)
|
|
|
|
+{
|
|
|
|
+ u32 adj_index = nh_grp->adj_index; /* base */
|
|
|
|
+ struct mlxsw_sp_nexthop *nh;
|
|
|
|
+ int i;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ nh = &nh_grp->nexthops[i];
|
|
|
|
+
|
|
|
|
+ if (!nh->should_offload) {
|
|
|
|
+ nh->offloaded = 0;
|
|
|
|
+ continue;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ if (nh->update) {
|
|
|
|
+ err = mlxsw_sp_nexthop_mac_update(mlxsw_sp,
|
|
|
|
+ adj_index, nh);
|
|
|
|
+ if (err)
|
|
|
|
+ return err;
|
|
|
|
+ nh->update = 0;
|
|
|
|
+ nh->offloaded = 1;
|
|
|
|
+ }
|
|
|
|
+ adj_index++;
|
|
|
|
+ }
|
|
|
|
+ return 0;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry);
|
|
|
|
+
|
|
|
|
+static int
|
|
|
|
+mlxsw_sp_nexthop_fib_entries_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
|
|
|
|
+ err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
|
|
|
|
+ if (err)
|
|
|
|
+ return err;
|
|
|
|
+ }
|
|
|
|
+ return 0;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_nexthop_group_refresh(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop *nh;
|
|
|
|
+ bool offload_change = false;
|
|
|
|
+ u32 adj_index;
|
|
|
|
+ u16 ecmp_size = 0;
|
|
|
|
+ bool old_adj_index_valid;
|
|
|
|
+ u32 old_adj_index;
|
|
|
|
+ u16 old_ecmp_size;
|
|
|
|
+ int ret;
|
|
|
|
+ int i;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ nh = &nh_grp->nexthops[i];
|
|
|
|
+
|
|
|
|
+ if (nh->should_offload ^ nh->offloaded) {
|
|
|
|
+ offload_change = true;
|
|
|
|
+ if (nh->should_offload)
|
|
|
|
+ nh->update = 1;
|
|
|
|
+ }
|
|
|
|
+ if (nh->should_offload)
|
|
|
|
+ ecmp_size++;
|
|
|
|
+ }
|
|
|
|
+ if (!offload_change) {
|
|
|
|
+ /* Nothing was added or removed, so no need to reallocate. Just
|
|
|
|
+ * update MAC on existing adjacency indexes.
|
|
|
|
+ */
|
|
|
|
+ err = mlxsw_sp_nexthop_group_mac_update(mlxsw_sp, nh_grp);
|
|
|
|
+ if (err) {
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
|
|
|
|
+ goto set_trap;
|
|
|
|
+ }
|
|
|
|
+ return;
|
|
|
|
+ }
|
|
|
|
+ if (!ecmp_size)
|
|
|
|
+ /* No neigh of this group is connected so we just set
|
|
|
|
+ * the trap and let everthing flow through kernel.
|
|
|
|
+ */
|
|
|
|
+ goto set_trap;
|
|
|
|
+
|
|
|
|
+ ret = mlxsw_sp_kvdl_alloc(mlxsw_sp, ecmp_size);
|
|
|
|
+ if (ret < 0) {
|
|
|
|
+ /* We ran out of KVD linear space, just set the
|
|
|
|
+ * trap and let everything flow through kernel.
|
|
|
|
+ */
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to allocate KVD linear area for nexthop group.\n");
|
|
|
|
+ goto set_trap;
|
|
|
|
+ }
|
|
|
|
+ adj_index = ret;
|
|
|
|
+ old_adj_index_valid = nh_grp->adj_index_valid;
|
|
|
|
+ old_adj_index = nh_grp->adj_index;
|
|
|
|
+ old_ecmp_size = nh_grp->ecmp_size;
|
|
|
|
+ nh_grp->adj_index_valid = 1;
|
|
|
|
+ nh_grp->adj_index = adj_index;
|
|
|
|
+ nh_grp->ecmp_size = ecmp_size;
|
|
|
|
+ err = mlxsw_sp_nexthop_group_mac_update(mlxsw_sp, nh_grp);
|
|
|
|
+ if (err) {
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
|
|
|
|
+ goto set_trap;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ if (!old_adj_index_valid) {
|
|
|
|
+ /* The trap was set for fib entries, so we have to call
|
|
|
|
+ * fib entry update to unset it and use adjacency index.
|
|
|
|
+ */
|
|
|
|
+ err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
|
|
|
|
+ if (err) {
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to add adjacency index to fib entries.\n");
|
|
|
|
+ goto set_trap;
|
|
|
|
+ }
|
|
|
|
+ return;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ err = mlxsw_sp_adj_index_mass_update(mlxsw_sp, nh_grp,
|
|
|
|
+ old_adj_index, old_ecmp_size);
|
|
|
|
+ mlxsw_sp_kvdl_free(mlxsw_sp, old_adj_index);
|
|
|
|
+ if (err) {
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to mass-update adjacency index for nexthop group.\n");
|
|
|
|
+ goto set_trap;
|
|
|
|
+ }
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+set_trap:
|
|
|
|
+ old_adj_index_valid = nh_grp->adj_index_valid;
|
|
|
|
+ nh_grp->adj_index_valid = 0;
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ nh = &nh_grp->nexthops[i];
|
|
|
|
+ nh->offloaded = 0;
|
|
|
|
+ }
|
|
|
|
+ err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
|
|
|
|
+ if (err)
|
|
|
|
+ dev_warn(mlxsw_sp->bus_info->dev, "Failed to set traps for fib entries.\n");
|
|
|
|
+ if (old_adj_index_valid)
|
|
|
|
+ mlxsw_sp_kvdl_free(mlxsw_sp, nh_grp->adj_index);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void __mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp_nexthop *nh,
|
|
|
|
+ bool removing)
|
|
|
|
+{
|
|
|
|
+ if (!removing && !nh->should_offload)
|
|
|
|
+ nh->should_offload = 1;
|
|
|
|
+ else if (removing && nh->offloaded)
|
|
|
|
+ nh->should_offload = 0;
|
|
|
|
+ nh->update = 1;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry,
|
|
|
|
+ bool removing)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop *nh;
|
|
|
|
+
|
|
|
|
+ /* Take RTNL mutex here to prevent lists from changes */
|
|
|
|
+ rtnl_lock();
|
|
|
|
+ list_for_each_entry(nh, &neigh_entry->nexthop_list,
|
|
|
|
+ neigh_list_node) {
|
|
|
|
+ __mlxsw_sp_nexthop_neigh_update(nh, removing);
|
|
|
|
+ mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
|
|
|
|
+ }
|
|
|
|
+ rtnl_unlock();
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_nexthop_init(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp,
|
|
|
|
+ struct mlxsw_sp_nexthop *nh,
|
|
|
|
+ struct fib_nh *fib_nh)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry;
|
|
|
|
+ u32 gwip = ntohl(fib_nh->nh_gw);
|
|
|
|
+ struct net_device *dev = fib_nh->nh_dev;
|
|
|
|
+ struct neighbour *n;
|
|
|
|
+ u8 nud_state;
|
|
|
|
+
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, &gwip,
|
|
|
|
+ sizeof(gwip), dev);
|
|
|
|
+ if (!neigh_entry) {
|
|
|
|
+ __be32 gwipn = htonl(gwip);
|
|
|
|
+
|
|
|
|
+ n = neigh_create(&arp_tbl, &gwipn, dev);
|
|
|
|
+ if (IS_ERR(n))
|
|
|
|
+ return PTR_ERR(n);
|
|
|
|
+ neigh_event_send(n, NULL);
|
|
|
|
+ neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, &gwip,
|
|
|
|
+ sizeof(gwip), dev);
|
|
|
|
+ if (!neigh_entry) {
|
|
|
|
+ neigh_release(n);
|
|
|
|
+ return -EINVAL;
|
|
|
|
+ }
|
|
|
|
+ } else {
|
|
|
|
+ /* Take a reference of neigh here ensuring that neigh would
|
|
|
|
+ * not be detructed before the nexthop entry is finished.
|
|
|
|
+ * The second branch takes the reference in neith_create()
|
|
|
|
+ */
|
|
|
|
+ n = neigh_entry->n;
|
|
|
|
+ neigh_clone(n);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ /* If that is the first nexthop connected to that neigh, add to
|
|
|
|
+ * nexthop_neighs_list
|
|
|
|
+ */
|
|
|
|
+ if (list_empty(&neigh_entry->nexthop_list))
|
|
|
|
+ list_add_tail(&neigh_entry->nexthop_neighs_list_node,
|
|
|
|
+ &mlxsw_sp->router.nexthop_neighs_list);
|
|
|
|
+
|
|
|
|
+ nh->nh_grp = nh_grp;
|
|
|
|
+ nh->neigh_entry = neigh_entry;
|
|
|
|
+ list_add_tail(&nh->neigh_list_node, &neigh_entry->nexthop_list);
|
|
|
|
+ read_lock_bh(&n->lock);
|
|
|
|
+ nud_state = n->nud_state;
|
|
|
|
+ read_unlock_bh(&n->lock);
|
|
|
|
+ __mlxsw_sp_nexthop_neigh_update(nh, !(nud_state & NUD_VALID));
|
|
|
|
+
|
|
|
|
+ return 0;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_nexthop_fini(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop *nh)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
|
|
|
|
+
|
|
|
|
+ list_del(&nh->neigh_list_node);
|
|
|
|
+
|
|
|
|
+ /* If that is the last nexthop connected to that neigh, remove from
|
|
|
|
+ * nexthop_neighs_list
|
|
|
|
+ */
|
|
|
|
+ if (list_empty(&nh->neigh_entry->nexthop_list))
|
|
|
|
+ list_del(&nh->neigh_entry->nexthop_neighs_list_node);
|
|
|
|
+
|
|
|
|
+ neigh_release(neigh_entry->n);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct mlxsw_sp_nexthop_group *
|
|
|
|
+mlxsw_sp_nexthop_group_create(struct mlxsw_sp *mlxsw_sp, struct fib_info *fi)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp;
|
|
|
|
+ struct mlxsw_sp_nexthop *nh;
|
|
|
|
+ struct fib_nh *fib_nh;
|
|
|
|
+ size_t alloc_size;
|
|
|
|
+ int i;
|
|
|
|
+ int err;
|
|
|
|
+
|
|
|
|
+ alloc_size = sizeof(*nh_grp) +
|
|
|
|
+ fi->fib_nhs * sizeof(struct mlxsw_sp_nexthop);
|
|
|
|
+ nh_grp = kzalloc(alloc_size, GFP_KERNEL);
|
|
|
|
+ if (!nh_grp)
|
|
|
|
+ return ERR_PTR(-ENOMEM);
|
|
|
|
+ INIT_LIST_HEAD(&nh_grp->fib_list);
|
|
|
|
+ nh_grp->count = fi->fib_nhs;
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ nh = &nh_grp->nexthops[i];
|
|
|
|
+ fib_nh = &fi->fib_nh[i];
|
|
|
|
+ err = mlxsw_sp_nexthop_init(mlxsw_sp, nh_grp, nh, fib_nh);
|
|
|
|
+ if (err)
|
|
|
|
+ goto err_nexthop_init;
|
|
|
|
+ }
|
|
|
|
+ list_add_tail(&nh_grp->list, &mlxsw_sp->router.nexthop_group_list);
|
|
|
|
+ mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
|
|
|
|
+ return nh_grp;
|
|
|
|
+
|
|
|
|
+err_nexthop_init:
|
|
|
|
+ for (i--; i >= 0; i--)
|
|
|
|
+ mlxsw_sp_nexthop_fini(mlxsw_sp, nh);
|
|
|
|
+ kfree(nh_grp);
|
|
|
|
+ return ERR_PTR(err);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_nexthop_group_destroy(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop *nh;
|
|
|
|
+ int i;
|
|
|
|
+
|
|
|
|
+ list_del(&nh_grp->list);
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ nh = &nh_grp->nexthops[i];
|
|
|
|
+ mlxsw_sp_nexthop_fini(mlxsw_sp, nh);
|
|
|
|
+ }
|
|
|
|
+ kfree(nh_grp);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static bool mlxsw_sp_nexthop_match(struct mlxsw_sp_nexthop *nh,
|
|
|
|
+ struct fib_info *fi)
|
|
|
|
+{
|
|
|
|
+ int i;
|
|
|
|
+
|
|
|
|
+ for (i = 0; i < fi->fib_nhs; i++) {
|
|
|
|
+ struct fib_nh *fib_nh = &fi->fib_nh[i];
|
|
|
|
+ u32 gwip = ntohl(fib_nh->nh_gw);
|
|
|
|
+
|
|
|
|
+ if (memcmp(nh->neigh_entry->key.addr,
|
|
|
|
+ &gwip, sizeof(u32)) == 0 &&
|
|
|
|
+ nh->neigh_entry->key.dev == fib_nh->nh_dev)
|
|
|
|
+ return true;
|
|
|
|
+ }
|
|
|
|
+ return false;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static bool mlxsw_sp_nexthop_group_match(struct mlxsw_sp_nexthop_group *nh_grp,
|
|
|
|
+ struct fib_info *fi)
|
|
|
|
+{
|
|
|
|
+ int i;
|
|
|
|
+
|
|
|
|
+ if (nh_grp->count != fi->fib_nhs)
|
|
|
|
+ return false;
|
|
|
|
+ for (i = 0; i < nh_grp->count; i++) {
|
|
|
|
+ struct mlxsw_sp_nexthop *nh = &nh_grp->nexthops[i];
|
|
|
|
+
|
|
|
|
+ if (!mlxsw_sp_nexthop_match(nh, fi))
|
|
|
|
+ return false;
|
|
|
|
+ }
|
|
|
|
+ return true;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct mlxsw_sp_nexthop_group *
|
|
|
|
+mlxsw_sp_nexthop_group_find(struct mlxsw_sp *mlxsw_sp, struct fib_info *fi)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp;
|
|
|
|
+
|
|
|
|
+ list_for_each_entry(nh_grp, &mlxsw_sp->router.nexthop_group_list,
|
|
|
|
+ list) {
|
|
|
|
+ if (mlxsw_sp_nexthop_group_match(nh_grp, fi))
|
|
|
|
+ return nh_grp;
|
|
|
|
+ }
|
|
|
|
+ return NULL;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int mlxsw_sp_nexthop_group_get(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry,
|
|
|
|
+ struct fib_info *fi)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp;
|
|
|
|
+
|
|
|
|
+ nh_grp = mlxsw_sp_nexthop_group_find(mlxsw_sp, fi);
|
|
|
|
+ if (!nh_grp) {
|
|
|
|
+ nh_grp = mlxsw_sp_nexthop_group_create(mlxsw_sp, fi);
|
|
|
|
+ if (IS_ERR(nh_grp))
|
|
|
|
+ return PTR_ERR(nh_grp);
|
|
|
|
+ }
|
|
|
|
+ list_add_tail(&fib_entry->nexthop_group_node, &nh_grp->fib_list);
|
|
|
|
+ fib_entry->nh_group = nh_grp;
|
|
|
|
+ return 0;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void mlxsw_sp_nexthop_group_put(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry)
|
|
|
|
+{
|
|
|
|
+ struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
|
|
|
|
+
|
|
|
|
+ list_del(&fib_entry->nexthop_group_node);
|
|
|
|
+ if (!list_empty(&nh_grp->fib_list))
|
|
|
|
+ return;
|
|
|
|
+ mlxsw_sp_nexthop_group_destroy(mlxsw_sp, nh_grp);
|
|
|
|
+}
|
|
|
|
+
|
|
static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
|
|
static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
|
|
{
|
|
{
|
|
char rgcr_pl[MLXSW_REG_RGCR_LEN];
|
|
char rgcr_pl[MLXSW_REG_RGCR_LEN];
|
|
@@ -565,19 +1517,54 @@ int mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
|
|
{
|
|
{
|
|
int err;
|
|
int err;
|
|
|
|
|
|
|
|
+ INIT_LIST_HEAD(&mlxsw_sp->router.nexthop_neighs_list);
|
|
|
|
+ INIT_LIST_HEAD(&mlxsw_sp->router.nexthop_group_list);
|
|
err = __mlxsw_sp_router_init(mlxsw_sp);
|
|
err = __mlxsw_sp_router_init(mlxsw_sp);
|
|
if (err)
|
|
if (err)
|
|
return err;
|
|
return err;
|
|
mlxsw_sp_lpm_init(mlxsw_sp);
|
|
mlxsw_sp_lpm_init(mlxsw_sp);
|
|
mlxsw_sp_vrs_init(mlxsw_sp);
|
|
mlxsw_sp_vrs_init(mlxsw_sp);
|
|
- return 0;
|
|
|
|
|
|
+ return mlxsw_sp_neigh_init(mlxsw_sp);
|
|
}
|
|
}
|
|
|
|
|
|
void mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
|
|
void mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
|
|
{
|
|
{
|
|
|
|
+ mlxsw_sp_neigh_fini(mlxsw_sp);
|
|
__mlxsw_sp_router_fini(mlxsw_sp);
|
|
__mlxsw_sp_router_fini(mlxsw_sp);
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+static int mlxsw_sp_fib_entry_op4_remote(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry,
|
|
|
|
+ enum mlxsw_reg_ralue_op op)
|
|
|
|
+{
|
|
|
|
+ char ralue_pl[MLXSW_REG_RALUE_LEN];
|
|
|
|
+ u32 *p_dip = (u32 *) fib_entry->key.addr;
|
|
|
|
+ struct mlxsw_sp_vr *vr = fib_entry->vr;
|
|
|
|
+ enum mlxsw_reg_ralue_trap_action trap_action;
|
|
|
|
+ u16 trap_id = 0;
|
|
|
|
+ u32 adjacency_index = 0;
|
|
|
|
+ u16 ecmp_size = 0;
|
|
|
|
+
|
|
|
|
+ /* In case the nexthop group adjacency index is valid, use it
|
|
|
|
+ * with provided ECMP size. Otherwise, setup trap and pass
|
|
|
|
+ * traffic to kernel.
|
|
|
|
+ */
|
|
|
|
+ if (fib_entry->nh_group->adj_index_valid) {
|
|
|
|
+ trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
|
|
|
|
+ adjacency_index = fib_entry->nh_group->adj_index;
|
|
|
|
+ ecmp_size = fib_entry->nh_group->ecmp_size;
|
|
|
|
+ } else {
|
|
|
|
+ trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
|
|
|
|
+ trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ mlxsw_reg_ralue_pack4(ralue_pl, vr->proto, op, vr->id,
|
|
|
|
+ fib_entry->key.prefix_len, *p_dip);
|
|
|
|
+ mlxsw_reg_ralue_act_remote_pack(ralue_pl, trap_action, trap_id,
|
|
|
|
+ adjacency_index, ecmp_size);
|
|
|
|
+ return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
|
|
|
|
+}
|
|
|
|
+
|
|
static int mlxsw_sp_fib_entry_op4_local(struct mlxsw_sp *mlxsw_sp,
|
|
static int mlxsw_sp_fib_entry_op4_local(struct mlxsw_sp *mlxsw_sp,
|
|
struct mlxsw_sp_fib_entry *fib_entry,
|
|
struct mlxsw_sp_fib_entry *fib_entry,
|
|
enum mlxsw_reg_ralue_op op)
|
|
enum mlxsw_reg_ralue_op op)
|
|
@@ -614,7 +1601,7 @@ static int mlxsw_sp_fib_entry_op4(struct mlxsw_sp *mlxsw_sp,
|
|
{
|
|
{
|
|
switch (fib_entry->type) {
|
|
switch (fib_entry->type) {
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
|
|
- return -EINVAL;
|
|
|
|
|
|
+ return mlxsw_sp_fib_entry_op4_remote(mlxsw_sp, fib_entry, op);
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
|
|
return mlxsw_sp_fib_entry_op4_local(mlxsw_sp, fib_entry, op);
|
|
return mlxsw_sp_fib_entry_op4_local(mlxsw_sp, fib_entry, op);
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_TRAP:
|
|
case MLXSW_SP_FIB_ENTRY_TYPE_TRAP:
|
|
@@ -694,7 +1681,17 @@ mlxsw_sp_router_fib4_entry_init(struct mlxsw_sp *mlxsw_sp,
|
|
fib_entry->rif = r->rif;
|
|
fib_entry->rif = r->rif;
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
- return -EINVAL;
|
|
|
|
|
|
+ fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
|
|
|
|
+ return mlxsw_sp_nexthop_group_get(mlxsw_sp, fib_entry, fi);
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void
|
|
|
|
+mlxsw_sp_router_fib4_entry_fini(struct mlxsw_sp *mlxsw_sp,
|
|
|
|
+ struct mlxsw_sp_fib_entry *fib_entry)
|
|
|
|
+{
|
|
|
|
+ if (fib_entry->type != MLXSW_SP_FIB_ENTRY_TYPE_REMOTE)
|
|
|
|
+ return;
|
|
|
|
+ mlxsw_sp_nexthop_group_put(mlxsw_sp, fib_entry);
|
|
}
|
|
}
|
|
|
|
|
|
static int
|
|
static int
|
|
@@ -738,6 +1735,7 @@ mlxsw_sp_router_fib4_add_prepare(struct mlxsw_sp_port *mlxsw_sp_port,
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
err_alloc_info:
|
|
err_alloc_info:
|
|
|
|
+ mlxsw_sp_router_fib4_entry_fini(mlxsw_sp, fib_entry);
|
|
err_fib4_entry_init:
|
|
err_fib4_entry_init:
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
err_fib_entry_create:
|
|
err_fib_entry_create:
|
|
@@ -772,6 +1770,7 @@ mlxsw_sp_router_fib4_add_commit(struct mlxsw_sp_port *mlxsw_sp_port,
|
|
err_fib_entry_add:
|
|
err_fib_entry_add:
|
|
mlxsw_sp_fib_entry_remove(vr->fib, fib_entry);
|
|
mlxsw_sp_fib_entry_remove(vr->fib, fib_entry);
|
|
err_fib_entry_insert:
|
|
err_fib_entry_insert:
|
|
|
|
+ mlxsw_sp_router_fib4_entry_fini(mlxsw_sp, fib_entry);
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
mlxsw_sp_vr_put(mlxsw_sp, vr);
|
|
mlxsw_sp_vr_put(mlxsw_sp, vr);
|
|
return err;
|
|
return err;
|
|
@@ -808,6 +1807,7 @@ int mlxsw_sp_router_fib4_del(struct mlxsw_sp_port *mlxsw_sp_port,
|
|
}
|
|
}
|
|
mlxsw_sp_fib_entry_del(mlxsw_sp_port->mlxsw_sp, fib_entry);
|
|
mlxsw_sp_fib_entry_del(mlxsw_sp_port->mlxsw_sp, fib_entry);
|
|
mlxsw_sp_fib_entry_remove(vr->fib, fib_entry);
|
|
mlxsw_sp_fib_entry_remove(vr->fib, fib_entry);
|
|
|
|
+ mlxsw_sp_router_fib4_entry_fini(mlxsw_sp, fib_entry);
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
mlxsw_sp_fib_entry_destroy(fib_entry);
|
|
mlxsw_sp_vr_put(mlxsw_sp, vr);
|
|
mlxsw_sp_vr_put(mlxsw_sp, vr);
|
|
return 0;
|
|
return 0;
|