2010-08-10 20:11:48 -04:00
|
|
|
/*
|
|
|
|
|
* Copyright (c) 2010 Nicira Networks.
|
|
|
|
|
* Distributed under the terms of the GNU GPL version 2.
|
|
|
|
|
*
|
|
|
|
|
* Significant portions of this file may be copied from parts of the Linux
|
|
|
|
|
* kernel, by Linus Torvalds and others.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#ifndef TUNNEL_H
|
|
|
|
|
#define TUNNEL_H 1
|
|
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
#include <linux/version.h>
|
|
|
|
|
|
|
|
|
|
#include "flow.h"
|
2010-08-10 20:11:48 -04:00
|
|
|
#include "openvswitch/tunnel.h"
|
|
|
|
|
#include "table.h"
|
|
|
|
|
#include "vport.h"
|
|
|
|
|
|
2010-08-16 10:32:41 -04:00
|
|
|
/*
|
|
|
|
|
* The absolute minimum fragment size. Note that there are many other
|
|
|
|
|
* definitions of the minimum MTU.
|
|
|
|
|
*/
|
2010-08-10 20:11:48 -04:00
|
|
|
#define IP_MIN_MTU 68
|
|
|
|
|
|
|
|
|
|
/*
|
2010-08-27 13:55:02 -07:00
|
|
|
* One of these goes in struct tnl_ops and in tnl_find_port().
|
2010-08-10 20:11:48 -04:00
|
|
|
* These values are in the same namespace as other TNL_T_* values, so
|
2010-08-27 13:55:02 -07:00
|
|
|
* only the least significant 10 bits are available to define protocol
|
|
|
|
|
* identifiers.
|
2010-08-10 20:11:48 -04:00
|
|
|
*/
|
|
|
|
|
#define TNL_T_PROTO_GRE 0
|
2010-08-11 20:55:58 -04:00
|
|
|
#define TNL_T_PROTO_CAPWAP 1
|
2010-08-10 20:11:48 -04:00
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
/* These flags are only needed when calling tnl_find_port(). */
|
2010-08-10 20:11:48 -04:00
|
|
|
#define TNL_T_KEY_EXACT (1 << 10)
|
|
|
|
|
#define TNL_T_KEY_MATCH (1 << 11)
|
|
|
|
|
#define TNL_T_KEY_EITHER (TNL_T_KEY_EXACT | TNL_T_KEY_MATCH)
|
|
|
|
|
|
|
|
|
|
struct tnl_mutable_config {
|
|
|
|
|
struct rcu_head rcu;
|
|
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
unsigned seq; /* Sequence number to identify this config. */
|
2010-08-10 20:11:48 -04:00
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
u32 tunnel_type; /* Set of TNL_T_* flags that define lookup. */
|
|
|
|
|
unsigned tunnel_hlen; /* Tunnel header length. */
|
|
|
|
|
|
|
|
|
|
unsigned char eth_addr[ETH_ALEN];
|
|
|
|
|
unsigned mtu;
|
2010-08-10 20:11:48 -04:00
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
struct tnl_port_config port_config;
|
2010-08-10 20:11:48 -04:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
struct tnl_ops {
|
2010-08-27 13:55:02 -07:00
|
|
|
u32 tunnel_type; /* Put the TNL_T_PROTO_* type in here. */
|
|
|
|
|
u8 ipproto; /* The IP protocol for the tunnel. */
|
2010-08-10 20:11:48 -04:00
|
|
|
|
2010-08-16 10:32:41 -04:00
|
|
|
/*
|
2010-08-27 13:55:02 -07:00
|
|
|
* Returns the length of the tunnel header that will be added in
|
2010-08-16 10:32:41 -04:00
|
|
|
* build_header() (i.e. excludes the IP header). Returns a negative
|
|
|
|
|
* error code if the configuration is invalid.
|
|
|
|
|
*/
|
2010-08-10 20:11:48 -04:00
|
|
|
int (*hdr_len)(const struct tnl_port_config *);
|
2010-08-16 10:32:41 -04:00
|
|
|
|
|
|
|
|
/*
|
2010-08-27 13:55:02 -07:00
|
|
|
* Builds the static portion of the tunnel header, which is stored in
|
|
|
|
|
* the header cache. In general the performance of this function is
|
|
|
|
|
* not too important as we try to only call it when building the cache
|
|
|
|
|
* so it is preferable to shift as much work as possible here. However,
|
|
|
|
|
* in some circumstances caching is disabled and this function will be
|
|
|
|
|
* called for every packet, so try not to make it too slow.
|
|
|
|
|
*/
|
|
|
|
|
void (*build_header)(const struct vport *,
|
|
|
|
|
const struct tnl_mutable_config *, void *header);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Updates the cached header of a packet to match the actual packet
|
|
|
|
|
* data. Typical things that might need to be updated are length,
|
|
|
|
|
* checksum, etc. The IP header will have already been updated and this
|
|
|
|
|
* is the final step before transmission. Returns a linked list of
|
|
|
|
|
* completed SKBs (multiple packets may be generated in the event
|
|
|
|
|
* of fragmentation).
|
|
|
|
|
*/
|
|
|
|
|
struct sk_buff *(*update_header)(const struct vport *,
|
|
|
|
|
const struct tnl_mutable_config *,
|
|
|
|
|
struct dst_entry *, struct sk_buff *);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
|
|
|
|
|
/*
|
|
|
|
|
* On these kernels we have a fast mechanism to tell if the ARP cache for a
|
|
|
|
|
* particular destination has changed.
|
|
|
|
|
*/
|
|
|
|
|
#define HAVE_HH_SEQ
|
|
|
|
|
#endif
|
|
|
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,27)
|
|
|
|
|
/*
|
|
|
|
|
* On these kernels we have a fast mechanism to tell if the routing table
|
|
|
|
|
* has changed.
|
|
|
|
|
*/
|
|
|
|
|
#define HAVE_RT_GENID
|
|
|
|
|
#endif
|
|
|
|
|
#if !defined(HAVE_HH_SEQ) || !defined(HAVE_RT_GENID)
|
|
|
|
|
/* If we can't detect all system changes directly we need to use a timeout. */
|
|
|
|
|
#define NEED_CACHE_TIMEOUT
|
|
|
|
|
#endif
|
|
|
|
|
struct tnl_cache {
|
|
|
|
|
struct rcu_head rcu;
|
|
|
|
|
|
|
|
|
|
int len; /* Length of data to be memcpy'd from cache. */
|
|
|
|
|
|
|
|
|
|
/* Sequence number of mutable->seq from which this cache was generated. */
|
|
|
|
|
unsigned mutable_seq;
|
|
|
|
|
|
|
|
|
|
#ifdef HAVE_HH_SEQ
|
|
|
|
|
/*
|
|
|
|
|
* The sequence number from the seqlock protecting the hardware header
|
|
|
|
|
* cache (in the ARP cache). Since every write increments the counter
|
|
|
|
|
* this gives us an easy way to tell if it has changed.
|
|
|
|
|
*/
|
|
|
|
|
unsigned hh_seq;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#ifdef NEED_CACHE_TIMEOUT
|
|
|
|
|
/*
|
|
|
|
|
* If we don't have direct mechanisms to detect all important changes in
|
|
|
|
|
* the system fall back to an expiration time. This expiration time
|
|
|
|
|
* can be relatively short since at high rates there will be millions of
|
|
|
|
|
* packets per second, so we'll still get plenty of benefit from the
|
|
|
|
|
* cache. Note that if something changes we may blackhole packets
|
|
|
|
|
* until the expiration time (depending on what changed and the kernel
|
|
|
|
|
* version we may be able to detect the change sooner). Expiration is
|
|
|
|
|
* expressed as a time in jiffies.
|
2010-08-16 10:32:41 -04:00
|
|
|
*/
|
2010-08-27 13:55:02 -07:00
|
|
|
unsigned long expiration;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* The routing table entry that is the result of looking up the tunnel
|
|
|
|
|
* endpoints. It also contains a sequence number (called a generation
|
|
|
|
|
* ID) that can be compared to a global sequence to tell if the routing
|
|
|
|
|
* table has changed (and therefore there is a potential that this
|
|
|
|
|
* cached route has been invalidated).
|
|
|
|
|
*/
|
|
|
|
|
struct rtable *rt;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* If the output device for tunnel traffic is an OVS internal device,
|
|
|
|
|
* the flow of that datapath. Since all tunnel traffic will have the
|
|
|
|
|
* same headers this allows us to cache the flow lookup. NULL if the
|
|
|
|
|
* output device is not OVS or if there is no flow installed.
|
|
|
|
|
*/
|
|
|
|
|
struct sw_flow *flow;
|
|
|
|
|
|
|
|
|
|
/* The cached header follows after padding for alignment. */
|
2010-08-10 20:11:48 -04:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
struct tnl_vport {
|
|
|
|
|
struct rcu_head rcu;
|
|
|
|
|
struct tbl_node tbl_node;
|
|
|
|
|
|
|
|
|
|
char name[IFNAMSIZ];
|
|
|
|
|
const struct tnl_ops *tnl_ops;
|
|
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
struct tnl_mutable_config *mutable; /* Protected by RCU. */
|
2010-08-16 10:32:41 -04:00
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
/*
|
|
|
|
|
* ID of last fragment sent (for tunnel protocols with direct support
|
|
|
|
|
* fragmentation). If the protocol relies on IP fragmentation then
|
|
|
|
|
* this is not needed.
|
|
|
|
|
*/
|
2010-08-16 10:32:41 -04:00
|
|
|
atomic_t frag_id;
|
2010-08-27 13:55:02 -07:00
|
|
|
|
|
|
|
|
spinlock_t cache_lock;
|
|
|
|
|
struct tnl_cache *cache; /* Protected by RCU/cache_lock. */
|
|
|
|
|
|
|
|
|
|
#ifdef NEED_CACHE_TIMEOUT
|
|
|
|
|
/*
|
|
|
|
|
* If we must rely on expiration time to invalidate the cache, this is
|
|
|
|
|
* the interval. It is randomized within a range (defined by
|
|
|
|
|
* MAX_CACHE_EXP in tunnel.c) to avoid synchronized expirations caused
|
|
|
|
|
* by creation of a large number of tunnels at a one time.
|
|
|
|
|
*/
|
|
|
|
|
unsigned long cache_exp_interval;
|
|
|
|
|
#endif
|
2010-08-10 20:11:48 -04:00
|
|
|
};
|
|
|
|
|
|
2010-11-04 16:32:57 -07:00
|
|
|
struct vport *tnl_create(const struct vport_parms *, const struct vport_ops *,
|
2010-08-10 20:11:48 -04:00
|
|
|
const struct tnl_ops *);
|
2010-12-03 14:41:38 -08:00
|
|
|
int tnl_modify(struct vport *, struct odp_port *);
|
2010-08-10 20:11:48 -04:00
|
|
|
int tnl_destroy(struct vport *);
|
|
|
|
|
int tnl_set_mtu(struct vport *vport, int mtu);
|
|
|
|
|
int tnl_set_addr(struct vport *vport, const unsigned char *addr);
|
|
|
|
|
const char *tnl_get_name(const struct vport *vport);
|
|
|
|
|
const unsigned char *tnl_get_addr(const struct vport *vport);
|
|
|
|
|
int tnl_get_mtu(const struct vport *vport);
|
|
|
|
|
int tnl_send(struct vport *vport, struct sk_buff *skb);
|
|
|
|
|
void tnl_rcv(struct vport *vport, struct sk_buff *skb);
|
|
|
|
|
|
|
|
|
|
struct vport *tnl_find_port(__be32 saddr, __be32 daddr, __be32 key,
|
|
|
|
|
int tunnel_type,
|
|
|
|
|
const struct tnl_mutable_config **mutable);
|
|
|
|
|
bool tnl_frag_needed(struct vport *vport,
|
|
|
|
|
const struct tnl_mutable_config *mutable,
|
|
|
|
|
struct sk_buff *skb, unsigned int mtu, __be32 flow_key);
|
2010-08-27 13:55:02 -07:00
|
|
|
void tnl_free_linked_skbs(struct sk_buff *skb);
|
2010-08-10 20:11:48 -04:00
|
|
|
|
|
|
|
|
static inline struct tnl_vport *tnl_vport_priv(const struct vport *vport)
|
|
|
|
|
{
|
|
|
|
|
return vport_priv(vport);
|
|
|
|
|
}
|
|
|
|
|
|
2010-08-27 13:55:02 -07:00
|
|
|
|
2010-08-10 20:11:48 -04:00
|
|
|
#endif /* tunnel.h */
|