/*
- * Copyright (c) 2010 Nicira Networks.
- * Distributed under the terms of the GNU GPL version 2.
+ * Copyright (c) 2007-2012 Nicira, Inc.
*
- * Significant portions of this file may be copied from parts of the Linux
- * kernel, by Linus Torvalds and others.
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of version 2 of the GNU General Public
+ * License as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful, but
+ * WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
+ * 02110-1301, USA
*/
#ifndef TUNNEL_H
#define TUNNEL_H 1
#include <linux/version.h>
+#include <net/net_namespace.h>
+#include <net/netns/generic.h>
#include "flow.h"
-#include "openvswitch/tunnel.h"
-#include "table.h"
#include "vport.h"
-/*
- * The absolute minimum fragment size. Note that there are many other
- * definitions of the minimum MTU.
- */
-#define IP_MIN_MTU 68
-
-/*
- * One of these goes in struct tnl_ops and in tnl_find_port().
- * These values are in the same namespace as other TNL_T_* values, so
- * only the least significant 10 bits are available to define protocol
- * identifiers.
- */
-#define TNL_T_PROTO_GRE 0
-#define TNL_T_PROTO_CAPWAP 1
-
-/* These flags are only needed when calling tnl_find_port(). */
-#define TNL_T_KEY_EXACT (1 << 10)
-#define TNL_T_KEY_MATCH (1 << 11)
-#define TNL_T_KEY_EITHER (TNL_T_KEY_EXACT | TNL_T_KEY_MATCH)
-
-struct tnl_mutable_config {
- struct rcu_head rcu;
-
- unsigned seq; /* Sequence number to identify this config. */
-
- u32 tunnel_type; /* Set of TNL_T_* flags that define lookup. */
- unsigned tunnel_hlen; /* Tunnel header length. */
-
- unsigned char eth_addr[ETH_ALEN];
- unsigned mtu;
-
- struct tnl_port_config port_config;
-};
-
struct tnl_ops {
- u32 tunnel_type; /* Put the TNL_T_PROTO_* type in here. */
u8 ipproto; /* The IP protocol for the tunnel. */
/*
* Returns the length of the tunnel header that will be added in
- * build_header() (i.e. excludes the IP header). Returns a negative
- * error code if the configuration is invalid.
+ * build_header() (i.e. excludes the IP header).
*/
- int (*hdr_len)(const struct tnl_port_config *);
-
+ int (*hdr_len)(const struct ovs_key_ipv4_tunnel *);
/*
- * Builds the static portion of the tunnel header, which is stored in
- * the header cache. In general the performance of this function is
- * not too important as we try to only call it when building the cache
- * so it is preferable to shift as much work as possible here. However,
- * in some circumstances caching is disabled and this function will be
- * called for every packet, so try not to make it too slow.
- */
- void (*build_header)(const struct vport *,
- const struct tnl_mutable_config *, void *header);
-
- /*
- * Updates the cached header of a packet to match the actual packet
- * data. Typical things that might need to be updated are length,
- * checksum, etc. The IP header will have already been updated and this
- * is the final step before transmission. Returns a linked list of
- * completed SKBs (multiple packets may be generated in the event
- * of fragmentation).
- */
- struct sk_buff *(*update_header)(const struct vport *,
- const struct tnl_mutable_config *,
- struct dst_entry *, struct sk_buff *);
-};
-
-#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
-/*
- * On these kernels we have a fast mechanism to tell if the ARP cache for a
- * particular destination has changed.
- */
-#define HAVE_HH_SEQ
-#endif
-#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,27)
-/*
- * On these kernels we have a fast mechanism to tell if the routing table
- * has changed.
- */
-#define HAVE_RT_GENID
-#endif
-#if !defined(HAVE_HH_SEQ) || !defined(HAVE_RT_GENID)
-/* If we can't detect all system changes directly we need to use a timeout. */
-#define NEED_CACHE_TIMEOUT
-#endif
-struct tnl_cache {
- struct rcu_head rcu;
-
- int len; /* Length of data to be memcpy'd from cache. */
-
- /* Sequence number of mutable->seq from which this cache was generated. */
- unsigned mutable_seq;
-
-#ifdef HAVE_HH_SEQ
- /*
- * The sequence number from the seqlock protecting the hardware header
- * cache (in the ARP cache). Since every write increments the counter
- * this gives us an easy way to tell if it has changed.
- */
- unsigned hh_seq;
-#endif
-
-#ifdef NEED_CACHE_TIMEOUT
- /*
- * If we don't have direct mechanisms to detect all important changes in
- * the system fall back to an expiration time. This expiration time
- * can be relatively short since at high rates there will be millions of
- * packets per second, so we'll still get plenty of benefit from the
- * cache. Note that if something changes we may blackhole packets
- * until the expiration time (depending on what changed and the kernel
- * version we may be able to detect the change sooner). Expiration is
- * expressed as a time in jiffies.
- */
- unsigned long expiration;
-#endif
-
- /*
- * The routing table entry that is the result of looking up the tunnel
- * endpoints. It also contains a sequence number (called a generation
- * ID) that can be compared to a global sequence to tell if the routing
- * table has changed (and therefore there is a potential that this
- * cached route has been invalidated).
- */
- struct rtable *rt;
-
- /*
- * If the output device for tunnel traffic is an OVS internal device,
- * the flow of that datapath. Since all tunnel traffic will have the
- * same headers this allows us to cache the flow lookup. NULL if the
- * output device is not OVS or if there is no flow installed.
- */
- struct sw_flow *flow;
-
- /* The cached header follows after padding for alignment. */
+ * Builds header for given SKB. Space will have already been
+ * allocated at the start of the packet equal
+ * to sizeof(struct iphdr) + value returned by hdr_len().
+ */
+ void (*build_header)(const struct vport *, struct sk_buff *,
+ int tunnel_hlen);
};
struct tnl_vport {
struct rcu_head rcu;
- struct tbl_node tbl_node;
+ __be16 dst_port;
char name[IFNAMSIZ];
const struct tnl_ops *tnl_ops;
-
- struct tnl_mutable_config *mutable; /* Protected by RCU. */
-
- /*
- * ID of last fragment sent (for tunnel protocols with direct support
- * fragmentation). If the protocol relies on IP fragmentation then
- * this is not needed.
- */
- atomic_t frag_id;
-
- spinlock_t cache_lock;
- struct tnl_cache *cache; /* Protected by RCU/cache_lock. */
-
-#ifdef NEED_CACHE_TIMEOUT
- /*
- * If we must rely on expiration time to invalidate the cache, this is
- * the interval. It is randomized within a range (defined by
- * MAX_CACHE_EXP in tunnel.c) to avoid synchronized expirations caused
- * by creation of a large number of tunnels at a one time.
- */
- unsigned long cache_exp_interval;
-#endif
};
-struct vport *tnl_create(const struct vport_parms *, const struct vport_ops *,
- const struct tnl_ops *);
-int tnl_modify(struct vport *, const void __user *config);
-int tnl_destroy(struct vport *);
-int tnl_set_mtu(struct vport *vport, int mtu);
-int tnl_set_addr(struct vport *vport, const unsigned char *addr);
-const char *tnl_get_name(const struct vport *vport);
-const unsigned char *tnl_get_addr(const struct vport *vport);
-int tnl_get_mtu(const struct vport *vport);
-int tnl_send(struct vport *vport, struct sk_buff *skb);
-void tnl_rcv(struct vport *vport, struct sk_buff *skb);
+struct vport *ovs_tnl_create(const struct vport_parms *, const struct vport_ops *,
+ const struct tnl_ops *);
+void ovs_tnl_destroy(struct vport *);
-struct vport *tnl_find_port(__be32 saddr, __be32 daddr, __be32 key,
- int tunnel_type,
- const struct tnl_mutable_config **mutable);
-bool tnl_frag_needed(struct vport *vport,
- const struct tnl_mutable_config *mutable,
- struct sk_buff *skb, unsigned int mtu, __be32 flow_key);
-void tnl_free_linked_skbs(struct sk_buff *skb);
+const char *ovs_tnl_get_name(const struct vport *vport);
+int ovs_tnl_send(struct vport *vport, struct sk_buff *skb);
+void ovs_tnl_rcv(struct vport *vport, struct sk_buff *skb);
+u16 ovs_tnl_get_src_port(struct sk_buff *skb);
static inline struct tnl_vport *tnl_vport_priv(const struct vport *vport)
{
return vport_priv(vport);
}
+static inline void tnl_tun_key_init(struct ovs_key_ipv4_tunnel *tun_key,
+ const struct iphdr *iph, __be64 tun_id, u32 tun_flags)
+{
+ tun_key->tun_id = tun_id;
+ tun_key->ipv4_src = iph->saddr;
+ tun_key->ipv4_dst = iph->daddr;
+ tun_key->ipv4_tos = iph->tos;
+ tun_key->ipv4_ttl = iph->ttl;
+ tun_key->tun_flags = tun_flags;
+
+ /* clear struct padding. */
+ memset((unsigned char*) tun_key + OVS_TUNNEL_KEY_SIZE, 0,
+ sizeof(*tun_key) - OVS_TUNNEL_KEY_SIZE);
+}
#endif /* tunnel.h */