#include "datapath.h"
#include "table.h"
#include "chain.h"
+#include "dp_dev.h"
#include "forward.h"
#include "flow.h"
-#include "datapath_t.h"
#include "compat.h"
static struct genl_family dp_genl_family;
static struct genl_multicast_group mc_group;
-int dp_dev_setup(struct net_device *dev);
-
/* It's hard to imagine wanting more than one datapath, but... */
#define DP_MAX 32
static int dp_maint_func(void *data);
static int send_port_status(struct net_bridge_port *p, uint8_t status);
static int dp_genl_openflow_done(struct netlink_callback *);
+static struct net_bridge_port *new_nbp(struct datapath *,
+ struct net_device *, int port_no);
+static int del_switch_port(struct net_bridge_port *);
/* nla_shrink - reduce amount of space reserved by nla_reserve
* @skb: socket buffer from which to recover room
static int
send_openflow_skb(struct sk_buff *skb, const struct sender *sender)
{
- int err = (sender
- ? genlmsg_unicast(skb, sender->pid)
- : genlmsg_multicast(skb, 0, mc_group.id, GFP_ATOMIC));
- if (err && net_ratelimit())
- printk(KERN_WARNING "send_openflow_skb: send failed: %d\n",
- err);
- return err;
+ return (sender
+ ? genlmsg_unicast(skb, sender->pid)
+ : genlmsg_multicast(skb, 0, mc_group.id, GFP_ATOMIC));
}
/* Generates a unique datapath id. It incorporates the datapath index
if (dp == NULL)
goto err_unlock;
+ /* Setup our "of" device */
+ err = dp_dev_setup(dp);
+ if (err)
+ goto err_free_dp;
+
dp->dp_idx = dp_idx;
dp->id = gen_datapath_id(dp_idx);
dp->chain = chain_create(dp);
if (dp->chain == NULL)
- goto err_free_dp;
+ goto err_destroy_dp_dev;
INIT_LIST_HEAD(&dp->port_list);
-#if 0
- /* Setup our "of" device */
- dp->dev.priv = dp;
- rtnl_lock();
- err = dp_dev_setup(&dp->dev);
- rtnl_unlock();
- if (err != 0)
- printk("datapath: problem setting up 'of' device\n");
-#endif
+ dp->local_port = new_nbp(dp, dp->netdev, OFPP_LOCAL);
+ if (IS_ERR(dp->local_port)) {
+ err = PTR_ERR(dp->local_port);
+ goto err_destroy_local_port;
+ }
dp->flags = 0;
dp->miss_send_len = OFP_DEFAULT_MISS_SEND_LEN;
dp->dp_task = kthread_run(dp_maint_func, dp, "dp%d", dp_idx);
if (IS_ERR(dp->dp_task))
- goto err_free_dp;
+ goto err_destroy_chain;
rcu_assign_pointer(dps[dp_idx], dp);
mutex_unlock(&dp_mutex);
return 0;
+err_destroy_local_port:
+ del_switch_port(dp->local_port);
+err_destroy_chain:
+ chain_destroy(dp->chain);
+err_destroy_dp_dev:
+ dp_dev_destroy(dp);
err_free_dp:
kfree(dp);
err_unlock:
}
static struct net_bridge_port *new_nbp(struct datapath *dp,
- struct net_device *dev)
+ struct net_device *dev, int port_no)
{
struct net_bridge_port *p;
- int port_no;
- port_no = find_portno(dp);
- if (port_no < 0)
- return ERR_PTR(port_no);
+ if (dev->br_port != NULL)
+ return ERR_PTR(-EBUSY);
p = kzalloc(sizeof(*p), GFP_KERNEL);
if (p == NULL)
return ERR_PTR(-ENOMEM);
- p->dp = dp;
+ rtnl_lock();
+ dev_set_promiscuity(dev, 1);
+ rtnl_unlock();
dev_hold(dev);
+ p->dp = dp;
p->dev = dev;
p->port_no = port_no;
+ if (port_no != OFPP_LOCAL)
+ rcu_assign_pointer(dev->br_port, p);
+ if (port_no < OFPP_MAX)
+ rcu_assign_pointer(dp->ports[port_no], p);
+ list_add_rcu(&p->node, &dp->port_list);
return p;
}
int add_switch_port(struct datapath *dp, struct net_device *dev)
{
struct net_bridge_port *p;
+ int port_no;
- if (dev->flags & IFF_LOOPBACK || dev->type != ARPHRD_ETHER)
+ if (dev->flags & IFF_LOOPBACK || dev->type != ARPHRD_ETHER
+ || is_dp_dev(dev))
return -EINVAL;
- if (dev->br_port != NULL)
- return -EBUSY;
+ port_no = find_portno(dp);
+ if (port_no < 0)
+ return port_no;
- p = new_nbp(dp, dev);
+ p = new_nbp(dp, dev, port_no);
if (IS_ERR(p))
return PTR_ERR(p);
- dev_hold(dev);
- rcu_assign_pointer(dev->br_port, p);
- rtnl_lock();
- dev_set_promiscuity(dev, 1);
- rtnl_unlock();
-
- rcu_assign_pointer(dp->ports[p->port_no], p);
- list_add_rcu(&p->node, &dp->port_list);
-
/* Notify the ctlpath that this port has been added */
send_port_status(p, OFPPR_ADD);
dev_set_promiscuity(p->dev, -1);
rtnl_unlock();
list_del_rcu(&p->node);
- rcu_assign_pointer(p->dp->ports[p->port_no], NULL);
+ if (p->port_no != OFPP_LOCAL)
+ rcu_assign_pointer(p->dp->ports[p->port_no], NULL);
rcu_assign_pointer(p->dev->br_port, NULL);
/* Then wait until no one is still using it, and destroy it. */
{
struct net_bridge_port *p, *n;
-#if 0
- /* Unregister the "of" device of this dp */
- rtnl_lock();
- unregister_netdevice(&dp->dev);
- rtnl_unlock();
-#endif
-
kthread_stop(dp->dp_task);
/* Drop references to DP. */
del_switch_port(p);
rcu_assign_pointer(dps[dp->dp_idx], NULL);
+ /* Kill off local_port dev references from buffered packets that have
+ * associated dst entries. */
+ synchronize_rcu();
+ fwd_discard_all();
+
+ /* Destroy dp->netdev. (Must follow deleting switch ports since
+ * dp->local_port has a reference to it.) */
+ dp_dev_destroy(dp);
+
/* Wait until no longer in use, then destroy it. */
synchronize_rcu();
chain_destroy(dp->chain);
struct datapath *dp = (struct datapath *) data;
while (!kthread_should_stop()) {
-#if 1
chain_timeout(dp->chain);
-#else
- int count = chain_timeout(dp->chain);
- chain_print_stats(dp->chain);
- if (count)
- printk("%d flows timed out\n", count);
-#endif
msleep_interruptible(MAINT_SLEEP_MSECS);
}
do_port_input(struct net_bridge_port *p, struct sk_buff *skb)
{
/* Push the Ethernet header back on. */
- if (skb->protocol == htons(ETH_P_8021Q))
- skb_push(skb, VLAN_ETH_HLEN);
- else
- skb_push(skb, ETH_HLEN);
+ skb_push(skb, ETH_HLEN);
fwd_port_input(p->dp->chain, skb, p->port_no);
}
/* Forwarding output path.
* Based on net/bridge/br_forward.c. */
-/* Don't forward packets to originating port. If we're flooding,
- * then don't send out ports with flooding disabled.
- */
-static inline int should_deliver(const struct net_bridge_port *p,
- const struct sk_buff *skb, int flood)
-{
- if (skb->dev == p->dev)
- return 0;
-
- if (flood && (p->flags & BRIDGE_PORT_NO_FLOOD))
- return 0;
-
- return 1;
-}
-
static inline unsigned packet_length(const struct sk_buff *skb)
{
int length = skb->len - ETH_HLEN;
static int
output_all(struct datapath *dp, struct sk_buff *skb, int flood)
{
+ u32 disable = flood ? BRIDGE_PORT_NO_FLOOD : 0;
struct net_bridge_port *p;
- int prev_port;
+ int prev_port = -1;
- prev_port = -1;
list_for_each_entry_rcu (p, &dp->port_list, node) {
- if (!should_deliver(p, skb, flood))
+ if (skb->dev == p->dev || p->flags & disable)
continue;
if (prev_port != -1) {
struct sk_buff *clone = skb_clone(skb, GFP_ATOMIC);
int dp_set_origin(struct datapath *dp, uint16_t in_port,
struct sk_buff *skb)
{
- if (in_port < OFPP_MAX && dp->ports[in_port]) {
- skb->dev = dp->ports[in_port]->dev;
+ struct net_bridge_port *p = (in_port < OFPP_MAX ? dp->ports[in_port]
+ : in_port == OFPP_LOCAL ? dp->local_port
+ : NULL);
+ if (p) {
+ skb->dev = p->dev;
return 0;
}
return -ENOENT;
*/
int dp_output_port(struct datapath *dp, struct sk_buff *skb, int out_port)
{
- struct net_bridge_port *p;
- int len = skb->len;
-
BUG_ON(!skb);
if (out_port == OFPP_FLOOD)
return output_all(dp, skb, 1);
return dp_output_control(dp, skb, fwd_save_skb(skb), 0,
OFPR_ACTION);
else if (out_port == OFPP_TABLE) {
+ struct net_bridge_port *p = skb->dev->br_port;
struct sw_flow_key key;
struct sw_flow *flow;
- flow_extract(skb, skb->dev->br_port->port_no, &key);
+ flow_extract(skb, p ? p->port_no : OFPP_LOCAL, &key);
flow = chain_lookup(dp->chain, &key);
if (likely(flow != NULL)) {
flow_used(flow, skb);
return 0;
}
return -ESRCH;
- } else if (out_port >= OFPP_MAX)
- goto bad_port;
+ } else if (out_port == OFPP_LOCAL) {
+ struct net_device *dev = dp->netdev;
+ return dev ? dp_dev_recv(dev, skb) : -ESRCH;
+ } else if (out_port >= 0 && out_port < OFPP_MAX) {
+ struct net_bridge_port *p = dp->ports[out_port];
+ int len = skb->len;
+ if (p == NULL)
+ goto bad_port;
+ skb->dev = p->dev;
+ if (packet_length(skb) > skb->dev->mtu) {
+ printk("dropped over-mtu packet: %d > %d\n",
+ packet_length(skb), skb->dev->mtu);
+ kfree_skb(skb);
+ return -E2BIG;
+ }
- p = dp->ports[out_port];
- if (p == NULL)
- goto bad_port;
+ dev_queue_xmit(skb);
- skb->dev = p->dev;
- if (packet_length(skb) > skb->dev->mtu) {
- printk("dropped over-mtu packet: %d > %d\n",
- packet_length(skb), skb->dev->mtu);
- kfree_skb(skb);
- return -E2BIG;
+ return len;
}
- dev_queue_xmit(skb);
-
- return len;
-
bad_port:
kfree_skb(skb);
if (net_ratelimit())
* forward the whole packet? */
struct sk_buff *f_skb;
struct ofp_packet_in *opi;
+ struct net_bridge_port *p;
size_t fwd_len, opi_len;
int err;
}
opi->buffer_id = htonl(buffer_id);
opi->total_len = htons(skb->len);
- opi->in_port = htons(skb->dev->br_port->port_no);
+ p = skb->dev->br_port;
+ opi->in_port = htons(p ? p->port_no : OFPP_LOCAL);
opi->reason = reason;
opi->pad = 0;
memcpy(opi->data, skb_mac_header(skb), fwd_len);
int
dp_update_port_flags(struct datapath *dp, const struct ofp_phy_port *opp)
{
- struct net_bridge_port *p;
-
- p = dp->ports[htons(opp->port_no)];
-
+ int port_no = ntohs(opp->port_no);
+ struct net_bridge_port *p = (port_no < OFPP_MAX ? dp->ports[port_no]
+ : port_no == OFPP_LOCAL ? dp->local_port
+ : NULL);
/* Make sure the port id hasn't changed since this was sent */
- if (!p || memcmp(opp->hw_addr, p->dev->dev_addr, ETH_ALEN) != 0)
+ if (!p || memcmp(opp->hw_addr, p->dev->dev_addr, ETH_ALEN))
return -1;
-
p->flags = htonl(opp->flags);
-
return 0;
}
return send_openflow_skb(skb, sender);
}
+int
+dp_send_echo_reply(struct datapath *dp, const struct sender *sender,
+ const struct ofp_header *rq)
+{
+ struct sk_buff *skb;
+ struct ofp_header *reply;
+
+ reply = alloc_openflow_skb(dp, ntohs(rq->length), OFPT_ECHO_REPLY,
+ sender, &skb);
+ if (!reply)
+ return -ENOMEM;
+
+ memcpy(reply + 1, rq + 1, ntohs(rq->length) - sizeof *rq);
+ return send_openflow_skb(skb, sender);
+}
+
/* Generic Netlink interface.
*
* See netlink(7) for an introduction to netlink. See
.dumpit = dp_genl_openflow_dumpit,
};
-static struct nla_policy dp_genl_benchmark_policy[DP_GENL_A_MAX + 1] = {
- [DP_GENL_A_DP_IDX] = { .type = NLA_U32 },
- [DP_GENL_A_NPACKETS] = { .type = NLA_U32 },
- [DP_GENL_A_PSIZE] = { .type = NLA_U32 },
-};
-
-static struct genl_ops dp_genl_ops_benchmark_nl = {
- .cmd = DP_GENL_C_BENCHMARK_NL,
- .flags = GENL_ADMIN_PERM, /* Requires CAP_NET_ADMIN privilege. */
- .policy = dp_genl_benchmark_policy,
- .doit = dp_genl_benchmark_nl,
- .dumpit = NULL,
-};
-
static struct genl_ops *dp_genl_all_ops[] = {
/* Keep this operation first. Generic Netlink dispatching
* looks up operations with linear search, so we want it at the
&dp_genl_ops_query_dp,
&dp_genl_ops_add_port,
&dp_genl_ops_del_port,
- &dp_genl_ops_benchmark_nl,
};
static int dp_init_netlink(void)