X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=datapath%2Factions.c;h=7fe2f5454b44203108e7b97c6d1a408f8b2bd972;hb=e2f3178f0582eda302bdc5629189b6a56d9fbcdd;hp=78712c6a9603e524423c8137514f263287ca196a;hpb=b85d8d61a6c3b9c2fa578fa62c795a93544764af;p=sliver-openvswitch.git diff --git a/datapath/actions.c b/datapath/actions.c index 78712c6a9..7fe2f5454 100644 --- a/datapath/actions.c +++ b/datapath/actions.c @@ -1,35 +1,45 @@ /* - * Distributed under the terms of the GNU GPL version 2. - * Copyright (c) 2007, 2008, 2009, 2010, 2011 Nicira Networks. + * Copyright (c) 2007-2014 Nicira, Inc. * - * Significant portions of this file may be copied from parts of the Linux - * kernel, by Linus Torvalds and others. + * This program is free software; you can redistribute it and/or + * modify it under the terms of version 2 of the GNU General Public + * License as published by the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, but + * WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA + * 02110-1301, USA */ -/* Functions for executing flow actions. */ +#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include #include #include +#include +#include #include #include #include #include #include -#include #include +#include #include +#include +#include -#include "actions.h" -#include "checksum.h" #include "datapath.h" -#include "loop_counter.h" -#include "openvswitch/datapath-protocol.h" #include "vlan.h" #include "vport.h" -static int do_execute_actions(struct datapath *, struct sk_buff *, - struct sw_flow_actions *acts); +static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, + const struct nlattr *attr, int len, bool keep_skb); static int make_writable(struct sk_buff *skb, int write_len) { @@ -39,216 +49,511 @@ static int make_writable(struct sk_buff *skb, int write_len) return pskb_expand_head(skb, 0, 0, GFP_ATOMIC); } -static int strip_vlan(struct sk_buff *skb) +/* remove VLAN header from packet and update csum accordingly. */ +static int __pop_vlan_tci(struct sk_buff *skb, __be16 *current_tci) { - struct ethhdr *eh; + struct vlan_hdr *vhdr; int err; - if (vlan_tx_tag_present(skb)) { - vlan_set_tci(skb, 0); - return 0; - } - - if (unlikely(skb->protocol != htons(ETH_P_8021Q) || - skb->len < VLAN_ETH_HLEN)) - return 0; - err = make_writable(skb, VLAN_ETH_HLEN); if (unlikely(err)) return err; - if (get_ip_summed(skb) == OVS_CSUM_COMPLETE) + if (skb->ip_summed == CHECKSUM_COMPLETE) skb->csum = csum_sub(skb->csum, csum_partial(skb->data - + ETH_HLEN, VLAN_HLEN, 0)); + + (2 * ETH_ALEN), VLAN_HLEN, 0)); - memmove(skb->data + VLAN_HLEN, skb->data, 2 * ETH_ALEN); + vhdr = (struct vlan_hdr *)(skb->data + ETH_HLEN); + *current_tci = vhdr->h_vlan_TCI; - eh = (struct ethhdr *)skb_pull(skb, VLAN_HLEN); + memmove(skb->data + VLAN_HLEN, skb->data, 2 * ETH_ALEN); + __skb_pull(skb, VLAN_HLEN); - skb->protocol = eh->h_proto; + vlan_set_encap_proto(skb, vhdr); skb->mac_header += VLAN_HLEN; + skb_reset_mac_len(skb); return 0; } -static int modify_vlan_tci(struct sk_buff *skb, __be16 tci) +static int pop_vlan(struct sk_buff *skb) { - if (!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_8021Q)) { - int err; + __be16 tci; + int err; - if (unlikely(skb->len < VLAN_ETH_HLEN)) + if (likely(vlan_tx_tag_present(skb))) { + vlan_set_tci(skb, 0); + } else { + if (unlikely(skb->protocol != htons(ETH_P_8021Q) || + skb->len < VLAN_ETH_HLEN)) return 0; - err = strip_vlan(skb); - if (unlikely(err)) + err = __pop_vlan_tci(skb, &tci); + if (err) return err; } + /* move next vlan tag to hw accel tag */ + if (likely(skb->protocol != htons(ETH_P_8021Q) || + skb->len < VLAN_ETH_HLEN)) + return 0; + + err = __pop_vlan_tci(skb, &tci); + if (unlikely(err)) + return err; + + __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), ntohs(tci)); + return 0; +} + +static int push_vlan(struct sk_buff *skb, const struct ovs_action_push_vlan *vlan) +{ + if (unlikely(vlan_tx_tag_present(skb))) { + u16 current_tag; + + /* push down current VLAN tag */ + current_tag = vlan_tx_tag_get(skb); + + if (!__vlan_put_tag(skb, skb->vlan_proto, current_tag)) + return -ENOMEM; + + if (skb->ip_summed == CHECKSUM_COMPLETE) + skb->csum = csum_add(skb->csum, csum_partial(skb->data + + (2 * ETH_ALEN), VLAN_HLEN, 0)); + + } + __vlan_hwaccel_put_tag(skb, vlan->vlan_tpid, ntohs(vlan->vlan_tci) & ~VLAN_TAG_PRESENT); + return 0; +} + +static int set_eth_addr(struct sk_buff *skb, + const struct ovs_key_ethernet *eth_key) +{ + int err; + err = make_writable(skb, ETH_HLEN); + if (unlikely(err)) + return err; + + skb_postpull_rcsum(skb, eth_hdr(skb), ETH_ALEN * 2); + + ether_addr_copy(eth_hdr(skb)->h_source, eth_key->eth_src); + ether_addr_copy(eth_hdr(skb)->h_dest, eth_key->eth_dst); - __vlan_hwaccel_put_tag(skb, ntohs(tci)); + ovs_skb_postpush_rcsum(skb, eth_hdr(skb), ETH_ALEN * 2); return 0; } -static bool is_ip(struct sk_buff *skb) +static void set_ip_addr(struct sk_buff *skb, struct iphdr *nh, + __be32 *addr, __be32 new_addr) { - return (OVS_CB(skb)->flow->key.eth.type == htons(ETH_P_IP) && - skb->transport_header > skb->network_header); + int transport_len = skb->len - skb_transport_offset(skb); + + if (nh->protocol == IPPROTO_TCP) { + if (likely(transport_len >= sizeof(struct tcphdr))) + inet_proto_csum_replace4(&tcp_hdr(skb)->check, skb, + *addr, new_addr, 1); + } else if (nh->protocol == IPPROTO_UDP) { + if (likely(transport_len >= sizeof(struct udphdr))) { + struct udphdr *uh = udp_hdr(skb); + + if (uh->check || skb->ip_summed == CHECKSUM_PARTIAL) { + inet_proto_csum_replace4(&uh->check, skb, + *addr, new_addr, 1); + if (!uh->check) + uh->check = CSUM_MANGLED_0; + } + } + } + + csum_replace4(&nh->check, *addr, new_addr); + skb_clear_hash(skb); + *addr = new_addr; } -static __sum16 *get_l4_checksum(struct sk_buff *skb) +static void update_ipv6_checksum(struct sk_buff *skb, u8 l4_proto, + __be32 addr[4], const __be32 new_addr[4]) { - u8 nw_proto = OVS_CB(skb)->flow->key.ip.proto; int transport_len = skb->len - skb_transport_offset(skb); - if (nw_proto == IPPROTO_TCP) { + + if (l4_proto == IPPROTO_TCP) { if (likely(transport_len >= sizeof(struct tcphdr))) - return &tcp_hdr(skb)->check; - } else if (nw_proto == IPPROTO_UDP) { - if (likely(transport_len >= sizeof(struct udphdr))) - return &udp_hdr(skb)->check; + inet_proto_csum_replace16(&tcp_hdr(skb)->check, skb, + addr, new_addr, 1); + } else if (l4_proto == IPPROTO_UDP) { + if (likely(transport_len >= sizeof(struct udphdr))) { + struct udphdr *uh = udp_hdr(skb); + + if (uh->check || skb->ip_summed == CHECKSUM_PARTIAL) { + inet_proto_csum_replace16(&uh->check, skb, + addr, new_addr, 1); + if (!uh->check) + uh->check = CSUM_MANGLED_0; + } + } } - return NULL; } -static int set_nw_addr(struct sk_buff *skb, const struct nlattr *a) +static void set_ipv6_addr(struct sk_buff *skb, u8 l4_proto, + __be32 addr[4], const __be32 new_addr[4], + bool recalculate_csum) +{ + if (recalculate_csum) + update_ipv6_checksum(skb, l4_proto, addr, new_addr); + + skb_clear_hash(skb); + memcpy(addr, new_addr, sizeof(__be32[4])); +} + +static void set_ipv6_tc(struct ipv6hdr *nh, u8 tc) +{ + nh->priority = tc >> 4; + nh->flow_lbl[0] = (nh->flow_lbl[0] & 0x0F) | ((tc & 0x0F) << 4); +} + +static void set_ipv6_fl(struct ipv6hdr *nh, u32 fl) +{ + nh->flow_lbl[0] = (nh->flow_lbl[0] & 0xF0) | (fl & 0x000F0000) >> 16; + nh->flow_lbl[1] = (fl & 0x0000FF00) >> 8; + nh->flow_lbl[2] = fl & 0x000000FF; +} + +static void set_ip_ttl(struct sk_buff *skb, struct iphdr *nh, u8 new_ttl) +{ + csum_replace2(&nh->check, htons(nh->ttl << 8), htons(new_ttl << 8)); + nh->ttl = new_ttl; +} + +static int set_ipv4(struct sk_buff *skb, const struct ovs_key_ipv4 *ipv4_key) { - __be32 new_nwaddr = nla_get_be32(a); struct iphdr *nh; - __sum16 *check; - __be32 *nwaddr; int err; - if (unlikely(!is_ip(skb))) - return 0; - err = make_writable(skb, skb_network_offset(skb) + sizeof(struct iphdr)); if (unlikely(err)) return err; nh = ip_hdr(skb); - nwaddr = nla_type(a) == ODP_ACTION_ATTR_SET_NW_SRC ? &nh->saddr : &nh->daddr; - check = get_l4_checksum(skb); - if (likely(check)) - inet_proto_csum_replace4(check, skb, *nwaddr, new_nwaddr, 1); - csum_replace4(&nh->check, *nwaddr, new_nwaddr); + if (ipv4_key->ipv4_src != nh->saddr) + set_ip_addr(skb, nh, &nh->saddr, ipv4_key->ipv4_src); + + if (ipv4_key->ipv4_dst != nh->daddr) + set_ip_addr(skb, nh, &nh->daddr, ipv4_key->ipv4_dst); - skb_clear_rxhash(skb); + if (ipv4_key->ipv4_tos != nh->tos) + ipv4_change_dsfield(nh, 0, ipv4_key->ipv4_tos); - *nwaddr = new_nwaddr; + if (ipv4_key->ipv4_ttl != nh->ttl) + set_ip_ttl(skb, nh, ipv4_key->ipv4_ttl); return 0; } -static int set_nw_tos(struct sk_buff *skb, u8 nw_tos) +static int set_ipv6(struct sk_buff *skb, const struct ovs_key_ipv6 *ipv6_key) { - struct iphdr *nh = ip_hdr(skb); - u8 old, new; + struct ipv6hdr *nh; int err; - - if (unlikely(!is_ip(skb))) - return 0; + __be32 *saddr; + __be32 *daddr; err = make_writable(skb, skb_network_offset(skb) + - sizeof(struct iphdr)); + sizeof(struct ipv6hdr)); if (unlikely(err)) return err; - /* Set the DSCP bits and preserve the ECN bits. */ - old = nh->tos; - new = nw_tos | (nh->tos & INET_ECN_MASK); - csum_replace4(&nh->check, (__force __be32)old, - (__force __be32)new); - nh->tos = new; + nh = ipv6_hdr(skb); + saddr = (__be32 *)&nh->saddr; + daddr = (__be32 *)&nh->daddr; + + if (memcmp(ipv6_key->ipv6_src, saddr, sizeof(ipv6_key->ipv6_src))) + set_ipv6_addr(skb, ipv6_key->ipv6_proto, saddr, + ipv6_key->ipv6_src, true); + + if (memcmp(ipv6_key->ipv6_dst, daddr, sizeof(ipv6_key->ipv6_dst))) { + unsigned int offset = 0; + int flags = OVS_IP6T_FH_F_SKIP_RH; + bool recalc_csum = true; + + if (ipv6_ext_hdr(nh->nexthdr)) + recalc_csum = ipv6_find_hdr(skb, &offset, + NEXTHDR_ROUTING, NULL, + &flags) != NEXTHDR_ROUTING; + + set_ipv6_addr(skb, ipv6_key->ipv6_proto, daddr, + ipv6_key->ipv6_dst, recalc_csum); + } + + set_ipv6_tc(nh, ipv6_key->ipv6_tclass); + set_ipv6_fl(nh, ntohl(ipv6_key->ipv6_label)); + nh->hop_limit = ipv6_key->ipv6_hlimit; return 0; } -static int set_tp_port(struct sk_buff *skb, const struct nlattr *a) +/* Must follow make_writable() since that can move the skb data. */ +static void set_tp_port(struct sk_buff *skb, __be16 *port, + __be16 new_port, __sum16 *check) +{ + inet_proto_csum_replace2(check, skb, *port, new_port, 0); + *port = new_port; + skb_clear_hash(skb); +} + +static void set_udp_port(struct sk_buff *skb, __be16 *port, __be16 new_port) +{ + struct udphdr *uh = udp_hdr(skb); + + if (uh->check && skb->ip_summed != CHECKSUM_PARTIAL) { + set_tp_port(skb, port, new_port, &uh->check); + + if (!uh->check) + uh->check = CSUM_MANGLED_0; + } else { + *port = new_port; + skb_clear_hash(skb); + } +} + +static int set_udp(struct sk_buff *skb, const struct ovs_key_udp *udp_port_key) { - struct udphdr *th; - __sum16 *check; - __be16 *port; + struct udphdr *uh; int err; - if (unlikely(!is_ip(skb))) - return 0; + err = make_writable(skb, skb_transport_offset(skb) + + sizeof(struct udphdr)); + if (unlikely(err)) + return err; + + uh = udp_hdr(skb); + if (udp_port_key->udp_src != uh->source) + set_udp_port(skb, &uh->source, udp_port_key->udp_src); + + if (udp_port_key->udp_dst != uh->dest) + set_udp_port(skb, &uh->dest, udp_port_key->udp_dst); + + return 0; +} + +static int set_tcp(struct sk_buff *skb, const struct ovs_key_tcp *tcp_port_key) +{ + struct tcphdr *th; + int err; err = make_writable(skb, skb_transport_offset(skb) + sizeof(struct tcphdr)); if (unlikely(err)) return err; - /* Must follow make_writable() since that can move the skb data. */ - check = get_l4_checksum(skb); - if (unlikely(!check)) - return 0; + th = tcp_hdr(skb); + if (tcp_port_key->tcp_src != th->source) + set_tp_port(skb, &th->source, tcp_port_key->tcp_src, &th->check); - /* - * Update port and checksum. - * - * This is OK because source and destination port numbers are at the - * same offsets in both UDP and TCP headers, and get_l4_checksum() only - * supports those protocols. - */ - th = udp_hdr(skb); - port = nla_type(a) == ODP_ACTION_ATTR_SET_TP_SRC ? &th->source : &th->dest; - inet_proto_csum_replace2(check, skb, *port, nla_get_be16(a), 0); - *port = nla_get_be16(a); - skb_clear_rxhash(skb); + if (tcp_port_key->tcp_dst != th->dest) + set_tp_port(skb, &th->dest, tcp_port_key->tcp_dst, &th->check); return 0; } -static void do_output(struct datapath *dp, struct sk_buff *skb, int out_port) +static int set_sctp(struct sk_buff *skb, + const struct ovs_key_sctp *sctp_port_key) { - struct vport *p; + struct sctphdr *sh; + int err; + unsigned int sctphoff = skb_transport_offset(skb); - if (!skb) - goto error; + err = make_writable(skb, sctphoff + sizeof(struct sctphdr)); + if (unlikely(err)) + return err; - p = rcu_dereference(dp->ports[out_port]); - if (!p) - goto error; + sh = sctp_hdr(skb); + if (sctp_port_key->sctp_src != sh->source || + sctp_port_key->sctp_dst != sh->dest) { + __le32 old_correct_csum, new_csum, old_csum; - vport_send(p, skb); - return; + old_csum = sh->checksum; + old_correct_csum = sctp_compute_cksum(skb, sctphoff); + + sh->source = sctp_port_key->sctp_src; + sh->dest = sctp_port_key->sctp_dst; + + new_csum = sctp_compute_cksum(skb, sctphoff); + + /* Carry any checksum errors through. */ + sh->checksum = old_csum ^ old_correct_csum ^ new_csum; + + skb_clear_hash(skb); + } -error: - kfree_skb(skb); + return 0; } -static int output_userspace(struct datapath *dp, struct sk_buff *skb, u64 arg) +static int do_output(struct datapath *dp, struct sk_buff *skb, int out_port) { - struct dp_upcall_info upcall; + struct vport *vport; - skb = skb_clone(skb, GFP_ATOMIC); - if (!skb) + if (unlikely(!skb)) return -ENOMEM; - upcall.cmd = ODP_PACKET_CMD_ACTION; - upcall.key = &OVS_CB(skb)->flow->key; - upcall.userdata = arg; - upcall.sample_pool = 0; - upcall.actions = NULL; - upcall.actions_len = 0; - return dp_upcall(dp, skb, &upcall); + vport = ovs_vport_rcu(dp, out_port); + if (unlikely(!vport)) { + kfree_skb(skb); + return -ENODEV; + } + + ovs_vport_send(vport, skb); + return 0; +} + +static int output_userspace(struct datapath *dp, struct sk_buff *skb, + const struct nlattr *attr) +{ + struct dp_upcall_info upcall; + const struct nlattr *a; + int rem; + + BUG_ON(!OVS_CB(skb)->pkt_key); + + upcall.cmd = OVS_PACKET_CMD_ACTION; + upcall.key = OVS_CB(skb)->pkt_key; + upcall.userdata = NULL; + upcall.portid = 0; + + for (a = nla_data(attr), rem = nla_len(attr); rem > 0; + a = nla_next(a, &rem)) { + switch (nla_type(a)) { + case OVS_USERSPACE_ATTR_USERDATA: + upcall.userdata = a; + break; + + case OVS_USERSPACE_ATTR_PID: + upcall.portid = nla_get_u32(a); + break; + } + } + + return ovs_dp_upcall(dp, skb, &upcall); +} + +static int sample(struct datapath *dp, struct sk_buff *skb, + const struct nlattr *attr) +{ + const struct nlattr *acts_list = NULL; + const struct nlattr *a; + int rem; + + for (a = nla_data(attr), rem = nla_len(attr); rem > 0; + a = nla_next(a, &rem)) { + switch (nla_type(a)) { + case OVS_SAMPLE_ATTR_PROBABILITY: + if (prandom_u32() >= nla_get_u32(a)) + return 0; + break; + + case OVS_SAMPLE_ATTR_ACTIONS: + acts_list = a; + break; + } + } + + return do_execute_actions(dp, skb, nla_data(acts_list), + nla_len(acts_list), true); +} + +static void execute_hash(struct sk_buff *skb, const struct nlattr *attr) +{ + struct sw_flow_key *key = OVS_CB(skb)->pkt_key; + struct ovs_action_hash *hash_act = nla_data(attr); + u32 hash = 0; + + /* OVS_HASH_ALG_L4 is the only possible hash algorithm. */ + hash = skb_get_hash(skb); + hash = jhash_1word(hash, hash_act->hash_basis); + if (!hash) + hash = 0x1; + + key->ovs_flow_hash = hash; +} + +static int execute_set_action(struct sk_buff *skb, + const struct nlattr *nested_attr) +{ + int err = 0; + + switch (nla_type(nested_attr)) { + case OVS_KEY_ATTR_PRIORITY: + skb->priority = nla_get_u32(nested_attr); + break; + + case OVS_KEY_ATTR_SKB_MARK: + skb->mark = nla_get_u32(nested_attr); + break; + + case OVS_KEY_ATTR_IPV4_TUNNEL: + OVS_CB(skb)->tun_key = nla_data(nested_attr); + break; + + case OVS_KEY_ATTR_ETHERNET: + err = set_eth_addr(skb, nla_data(nested_attr)); + break; + + case OVS_KEY_ATTR_IPV4: + err = set_ipv4(skb, nla_data(nested_attr)); + break; + + case OVS_KEY_ATTR_IPV6: + err = set_ipv6(skb, nla_data(nested_attr)); + break; + + case OVS_KEY_ATTR_TCP: + err = set_tcp(skb, nla_data(nested_attr)); + break; + + case OVS_KEY_ATTR_UDP: + err = set_udp(skb, nla_data(nested_attr)); + break; + + case OVS_KEY_ATTR_SCTP: + err = set_sctp(skb, nla_data(nested_attr)); + break; + } + + return err; +} + +static int execute_recirc(struct datapath *dp, struct sk_buff *skb, + const struct nlattr *a) +{ + struct sw_flow_key recirc_key; + const struct vport *p = OVS_CB(skb)->input_vport; + uint32_t hash = OVS_CB(skb)->pkt_key->ovs_flow_hash; + int err; + + err = ovs_flow_extract(skb, p->port_no, &recirc_key); + if (err) + return err; + + recirc_key.ovs_flow_hash = hash; + recirc_key.recirc_id = nla_get_u32(a); + + ovs_dp_process_packet_with_key(skb, &recirc_key, true); + + return 0; } /* Execute a list of actions against 'skb'. */ static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, - struct sw_flow_actions *acts) + const struct nlattr *attr, int len, bool keep_skb) { /* Every output action needs a separate clone of 'skb', but the common * case is just a single output action, so that doing a clone and * then freeing the original skbuff is wasteful. So the following code * is slightly obscure just to avoid that. */ int prev_port = -1; - u32 priority = skb->priority; const struct nlattr *a; int rem; - for (a = acts->actions, rem = acts->actions_len; rem > 0; + for (a = attr, rem = len; rem > 0; a = nla_next(a, &rem)) { int err = 0; @@ -258,58 +563,53 @@ static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, } switch (nla_type(a)) { - case ODP_ACTION_ATTR_OUTPUT: + case OVS_ACTION_ATTR_OUTPUT: prev_port = nla_get_u32(a); break; - case ODP_ACTION_ATTR_USERSPACE: - err = output_userspace(dp, skb, nla_get_u64(a)); + case OVS_ACTION_ATTR_USERSPACE: + output_userspace(dp, skb, a); break; - case ODP_ACTION_ATTR_SET_TUNNEL: - OVS_CB(skb)->tun_id = nla_get_be64(a); + case OVS_ACTION_ATTR_HASH: + execute_hash(skb, a); break; - case ODP_ACTION_ATTR_SET_DL_TCI: - err = modify_vlan_tci(skb, nla_get_be16(a)); + case OVS_ACTION_ATTR_PUSH_VLAN: + err = push_vlan(skb, nla_data(a)); + if (unlikely(err)) /* skb already freed. */ + return err; break; - case ODP_ACTION_ATTR_STRIP_VLAN: - err = strip_vlan(skb); + case OVS_ACTION_ATTR_POP_VLAN: + err = pop_vlan(skb); break; - case ODP_ACTION_ATTR_SET_DL_SRC: - err = make_writable(skb, ETH_HLEN); - if (likely(!err)) - memcpy(eth_hdr(skb)->h_source, nla_data(a), ETH_ALEN); - break; + case OVS_ACTION_ATTR_RECIRC: { + struct sk_buff *recirc_skb; + const bool last_action = (a->nla_len == rem); - case ODP_ACTION_ATTR_SET_DL_DST: - err = make_writable(skb, ETH_HLEN); - if (likely(!err)) - memcpy(eth_hdr(skb)->h_dest, nla_data(a), ETH_ALEN); - break; + if (!last_action || keep_skb) + recirc_skb = skb_clone(skb, GFP_ATOMIC); + else + recirc_skb = skb; - case ODP_ACTION_ATTR_SET_NW_SRC: - case ODP_ACTION_ATTR_SET_NW_DST: - err = set_nw_addr(skb, a); - break; + err = execute_recirc(dp, recirc_skb, a); - case ODP_ACTION_ATTR_SET_NW_TOS: - err = set_nw_tos(skb, nla_get_u8(a)); - break; + if (last_action || err) + return err; - case ODP_ACTION_ATTR_SET_TP_SRC: - case ODP_ACTION_ATTR_SET_TP_DST: - err = set_tp_port(skb, a); break; + } - case ODP_ACTION_ATTR_SET_PRIORITY: - skb->priority = nla_get_u32(a); + case OVS_ACTION_ATTR_SET: + err = execute_set_action(skb, nla_data(a)); break; - case ODP_ACTION_ATTR_POP_PRIORITY: - skb->priority = priority; + case OVS_ACTION_ATTR_SAMPLE: + err = sample(dp, skb, a); + if (unlikely(err)) /* skb already freed. */ + return err; break; } @@ -319,51 +619,56 @@ static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, } } - if (prev_port != -1) + if (prev_port != -1) { + if (keep_skb) + skb = skb_clone(skb, GFP_ATOMIC); + do_output(dp, skb, prev_port); - else + } else if (!keep_skb) consume_skb(skb); return 0; } -static void sflow_sample(struct datapath *dp, struct sk_buff *skb, - struct sw_flow_actions *acts) +/* We limit the number of times that we pass into execute_actions() + * to avoid blowing out the stack in the event that we have a loop. + * + * Each loop adds some (estimated) cost to the kernel stack. + * The loop terminates when the max cost is exceeded. + * */ +#define RECIRC_STACK_COST 1 +#define DEFAULT_STACK_COST 4 +/* Allow up to 4 regular services, and up to 3 recirculations */ +#define MAX_STACK_COST (DEFAULT_STACK_COST * 4 + RECIRC_STACK_COST * 3) + +struct loop_counter { + u8 stack_cost; /* loop stack cost. */ + bool looping; /* Loop detected? */ +}; + +static DEFINE_PER_CPU(struct loop_counter, loop_counters); + +static int loop_suppress(struct datapath *dp, struct sw_flow_actions *actions) { - struct sk_buff *nskb; - struct vport *p = OVS_CB(skb)->vport; - struct dp_upcall_info upcall; - - if (unlikely(!p)) - return; - - atomic_inc(&p->sflow_pool); - if (net_random() >= dp->sflow_probability) - return; - - nskb = skb_clone(skb, GFP_ATOMIC); - if (unlikely(!nskb)) - return; - - upcall.cmd = ODP_PACKET_CMD_SAMPLE; - upcall.key = &OVS_CB(skb)->flow->key; - upcall.userdata = 0; - upcall.sample_pool = atomic_read(&p->sflow_pool); - upcall.actions = acts->actions; - upcall.actions_len = acts->actions_len; - dp_upcall(dp, nskb, &upcall); + if (net_ratelimit()) + pr_warn("%s: flow loop detected, dropping\n", + ovs_dp_name(dp)); + actions->actions_len = 0; + return -ELOOP; } /* Execute a list of actions against 'skb'. */ -int execute_actions(struct datapath *dp, struct sk_buff *skb) +int ovs_execute_actions(struct datapath *dp, struct sk_buff *skb, bool recirc) { struct sw_flow_actions *acts = rcu_dereference(OVS_CB(skb)->flow->sf_acts); + const u8 stack_cost = recirc ? RECIRC_STACK_COST : DEFAULT_STACK_COST; struct loop_counter *loop; int error; /* Check whether we've looped too much. */ - loop = loop_get_counter(); - if (unlikely(++loop->count > MAX_LOOPS)) + loop = &__get_cpu_var(loop_counters); + loop->stack_cost += stack_cost; + if (unlikely(loop->stack_cost > MAX_STACK_COST)) loop->looping = true; if (unlikely(loop->looping)) { error = loop_suppress(dp, acts); @@ -371,21 +676,19 @@ int execute_actions(struct datapath *dp, struct sk_buff *skb) goto out_loop; } - /* Really execute actions. */ - if (dp->sflow_probability) - sflow_sample(dp, skb, acts); - OVS_CB(skb)->tun_id = 0; - error = do_execute_actions(dp, skb, acts); + OVS_CB(skb)->tun_key = NULL; + error = do_execute_actions(dp, skb, acts->actions, + acts->actions_len, false); /* Check whether sub-actions looped too much. */ if (unlikely(loop->looping)) error = loop_suppress(dp, acts); out_loop: - /* Decrement loop counter. */ - if (!--loop->count) + /* Decrement loop stack cost. */ + loop->stack_cost -= stack_cost; + if (!loop->stack_cost) loop->looping = false; - loop_put_counter(); return error; }