mirror of
https://github.com/openvswitch/ovs
synced 2025-10-25 15:07:05 +00:00
Most kernels provide some form of ip fragmentation. However, until recently many of them would always send ICMP responses for over_MTU packets, even when operating in bridge mode. Backport the check to ensure this doesn't occur. Signed-off-by: Joe Stringer <joestringer@nicira.com> Acked-by: Pravin B Shelar <pshelar@nicira.com>
111 lines
2.8 KiB
C
111 lines
2.8 KiB
C
#ifndef __NET_IP_WRAPPER_H
|
|
#define __NET_IP_WRAPPER_H 1
|
|
|
|
#include_next <net/ip.h>
|
|
|
|
#include <net/route.h>
|
|
#include <linux/version.h>
|
|
|
|
#ifndef HAVE_IP_IS_FRAGMENT
|
|
static inline bool ip_is_fragment(const struct iphdr *iph)
|
|
{
|
|
return (iph->frag_off & htons(IP_MF | IP_OFFSET)) != 0;
|
|
}
|
|
#endif
|
|
|
|
#ifndef HAVE_INET_GET_LOCAL_PORT_RANGE_USING_NET
|
|
static inline void rpl_inet_get_local_port_range(struct net *net, int *low,
|
|
int *high)
|
|
{
|
|
inet_get_local_port_range(low, high);
|
|
}
|
|
#define inet_get_local_port_range rpl_inet_get_local_port_range
|
|
|
|
#endif
|
|
|
|
/* IPv4 datagram length is stored into 16bit field (tot_len) */
|
|
#ifndef IP_MAX_MTU
|
|
#define IP_MAX_MTU 0xFFFFU
|
|
#endif
|
|
|
|
#ifndef HAVE_IP_SKB_DST_MTU
|
|
static inline bool rpl_ip_sk_use_pmtu(const struct sock *sk)
|
|
{
|
|
return inet_sk(sk)->pmtudisc < IP_PMTUDISC_PROBE;
|
|
}
|
|
#define ip_sk_use_pmtu rpl_ip_sk_use_pmtu
|
|
|
|
static inline unsigned int ip_dst_mtu_maybe_forward(const struct dst_entry *dst,
|
|
bool forwarding)
|
|
{
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,14,0)
|
|
struct net *net = dev_net(dst->dev);
|
|
|
|
if (net->ipv4.sysctl_ip_fwd_use_pmtu ||
|
|
dst_metric_locked(dst, RTAX_MTU) ||
|
|
!forwarding)
|
|
return dst_mtu(dst);
|
|
#endif
|
|
|
|
return min(dst->dev->mtu, IP_MAX_MTU);
|
|
}
|
|
|
|
static inline unsigned int rpl_ip_skb_dst_mtu(const struct sk_buff *skb)
|
|
{
|
|
if (!skb->sk || ip_sk_use_pmtu(skb->sk)) {
|
|
bool forwarding = IPCB(skb)->flags & IPSKB_FORWARDED;
|
|
return ip_dst_mtu_maybe_forward(skb_dst(skb), forwarding);
|
|
} else {
|
|
return min(skb_dst(skb)->dev->mtu, IP_MAX_MTU);
|
|
}
|
|
}
|
|
#define ip_skb_dst_mtu rpl_ip_skb_dst_mtu
|
|
#endif /* HAVE_IP_SKB_DST_MTU */
|
|
|
|
#ifdef HAVE_IP_FRAGMENT_TAKES_SOCK
|
|
#define OVS_VPORT_OUTPUT_PARAMS struct sock *sock, struct sk_buff *skb
|
|
#else
|
|
#define OVS_VPORT_OUTPUT_PARAMS struct sk_buff *skb
|
|
#endif
|
|
|
|
#ifdef OVS_FRAGMENT_BACKPORT
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(4,2,0)
|
|
static inline bool ip_defrag_user_in_between(u32 user,
|
|
enum ip_defrag_users lower_bond,
|
|
enum ip_defrag_users upper_bond)
|
|
{
|
|
return user >= lower_bond && user <= upper_bond;
|
|
}
|
|
#endif
|
|
|
|
#ifndef HAVE_IP_DO_FRAGMENT
|
|
static inline int rpl_ip_do_fragment(struct sock *sk, struct sk_buff *skb,
|
|
int (*output)(OVS_VPORT_OUTPUT_PARAMS))
|
|
{
|
|
unsigned int mtu = ip_skb_dst_mtu(skb);
|
|
struct iphdr *iph = ip_hdr(skb);
|
|
struct rtable *rt = skb_rtable(skb);
|
|
struct net_device *dev = rt->dst.dev;
|
|
|
|
if (unlikely(((iph->frag_off & htons(IP_DF)) && !skb->ignore_df) ||
|
|
(IPCB(skb)->frag_max_size &&
|
|
IPCB(skb)->frag_max_size > mtu))) {
|
|
|
|
pr_warn("Dropping packet in ip_do_fragment()\n");
|
|
IP_INC_STATS(dev_net(dev), IPSTATS_MIB_FRAGFAILS);
|
|
kfree_skb(skb);
|
|
return -EMSGSIZE;
|
|
}
|
|
|
|
#ifndef HAVE_IP_FRAGMENT_TAKES_SOCK
|
|
return ip_fragment(skb, output);
|
|
#else
|
|
return ip_fragment(sk, skb, output);
|
|
#endif
|
|
}
|
|
#define ip_do_fragment rpl_ip_do_fragment
|
|
#endif /* IP_DO_FRAGMENT */
|
|
#endif /* OVS_FRAGMENT_BACKPORT */
|
|
#endif
|