mirror of
https://github.com/openvswitch/ovs
synced 2025-08-22 09:58:01 +00:00
Rather than mark with a offload flags + mark with a segmentation size, simply rely on the netdev implementation which sets a segmentation size when appropriate. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
1337 lines
39 KiB
C
1337 lines
39 KiB
C
/*
|
||
* Copyright (c) 2016 Nicira, Inc.
|
||
* Copyright (c) 2016 Red Hat, Inc.
|
||
*
|
||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||
* you may not use this file except in compliance with the License.
|
||
* You may obtain a copy of the License at:
|
||
*
|
||
* http://www.apache.org/licenses/LICENSE-2.0
|
||
*
|
||
* Unless required by applicable law or agreed to in writing, software
|
||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
* See the License for the specific language governing permissions and
|
||
* limitations under the License.
|
||
*/
|
||
|
||
#include <config.h>
|
||
|
||
#include "netdev-native-tnl.h"
|
||
|
||
#include <errno.h>
|
||
#include <fcntl.h>
|
||
#include <sys/socket.h>
|
||
#include <net/if.h>
|
||
#include <sys/types.h>
|
||
#include <netinet/in.h>
|
||
#include <netinet/ip.h>
|
||
#include <netinet/ip6.h>
|
||
#include <sys/ioctl.h>
|
||
|
||
#include <stdlib.h>
|
||
#include <sys/time.h>
|
||
|
||
#include "byte-order.h"
|
||
#include "coverage.h"
|
||
#include "csum.h"
|
||
#include "dp-packet.h"
|
||
#include "netdev.h"
|
||
#include "netdev-vport.h"
|
||
#include "netdev-vport-private.h"
|
||
#include "odp-netlink.h"
|
||
#include "packets.h"
|
||
#include "seq.h"
|
||
#include "unaligned.h"
|
||
#include "unixctl.h"
|
||
#include "util.h"
|
||
#include "openvswitch/vlog.h"
|
||
|
||
VLOG_DEFINE_THIS_MODULE(native_tnl);
|
||
static struct vlog_rate_limit err_rl = VLOG_RATE_LIMIT_INIT(60, 5);
|
||
|
||
COVERAGE_DEFINE(native_tnl_l3csum_checked);
|
||
COVERAGE_DEFINE(native_tnl_l3csum_err);
|
||
COVERAGE_DEFINE(native_tnl_l4csum_checked);
|
||
COVERAGE_DEFINE(native_tnl_l4csum_err);
|
||
|
||
#define VXLAN_HLEN (sizeof(struct udp_header) + \
|
||
sizeof(struct vxlanhdr))
|
||
|
||
#define GENEVE_BASE_HLEN (sizeof(struct udp_header) + \
|
||
sizeof(struct genevehdr))
|
||
|
||
#define GTPU_HLEN (sizeof(struct udp_header) + \
|
||
sizeof(struct gtpuhdr))
|
||
|
||
uint16_t tnl_udp_port_min = 32768;
|
||
uint16_t tnl_udp_port_max = 61000;
|
||
|
||
ovs_be16
|
||
netdev_tnl_get_src_port(struct dp_packet *packet)
|
||
{
|
||
uint32_t hash;
|
||
|
||
if (OVS_LIKELY(dp_packet_rss_valid(packet))) {
|
||
hash = dp_packet_get_rss_hash(packet);
|
||
} else {
|
||
struct flow flow;
|
||
|
||
flow_extract(packet, &flow);
|
||
hash = flow_hash_5tuple(&flow, 0);
|
||
|
||
dp_packet_set_rss_hash(packet, hash);
|
||
}
|
||
|
||
hash = ((uint64_t) hash * (tnl_udp_port_max - tnl_udp_port_min)) >> 32;
|
||
|
||
return htons(hash + tnl_udp_port_min);
|
||
}
|
||
|
||
static void *
|
||
ip_extract_tnl_md(struct dp_packet *packet, struct flow_tnl *tnl,
|
||
unsigned int *hlen)
|
||
{
|
||
void *nh;
|
||
struct ip_header *ip;
|
||
struct ovs_16aligned_ip6_hdr *ip6;
|
||
void *l4;
|
||
int l3_size;
|
||
|
||
nh = dp_packet_l3(packet);
|
||
ip = nh;
|
||
ip6 = nh;
|
||
l4 = dp_packet_l4(packet);
|
||
|
||
if (!nh || !l4) {
|
||
return NULL;
|
||
}
|
||
|
||
*hlen = sizeof(struct eth_header);
|
||
|
||
l3_size = dp_packet_size(packet) -
|
||
((char *)nh - (char *)dp_packet_data(packet));
|
||
|
||
if (IP_VER(ip->ip_ihl_ver) == 4) {
|
||
bool bad_csum = dp_packet_ip_checksum_bad(packet);
|
||
ovs_be32 ip_src, ip_dst;
|
||
|
||
/* A packet coming from a network device might have the
|
||
* csum already checked. In this case, skip the check. */
|
||
if (OVS_UNLIKELY(!bad_csum && dp_packet_ip_checksum_unknown(packet))) {
|
||
COVERAGE_INC(native_tnl_l3csum_checked);
|
||
if (csum(ip, IP_IHL(ip->ip_ihl_ver) * 4)) {
|
||
dp_packet_ip_checksum_set_bad(packet);
|
||
bad_csum = true;
|
||
} else {
|
||
dp_packet_ip_checksum_set_good(packet);
|
||
}
|
||
}
|
||
if (OVS_UNLIKELY(bad_csum)) {
|
||
COVERAGE_INC(native_tnl_l3csum_err);
|
||
VLOG_WARN_RL(&err_rl, "ip packet has invalid checksum");
|
||
return NULL;
|
||
}
|
||
|
||
if (ntohs(ip->ip_tot_len) > l3_size) {
|
||
VLOG_WARN_RL(&err_rl, "ip packet is truncated (IP length %d, actual %d)",
|
||
ntohs(ip->ip_tot_len), l3_size);
|
||
return NULL;
|
||
}
|
||
if (IP_IHL(ip->ip_ihl_ver) * 4 > sizeof(struct ip_header)) {
|
||
VLOG_WARN_RL(&err_rl, "ip options not supported on tunnel packets "
|
||
"(%d bytes)", IP_IHL(ip->ip_ihl_ver) * 4);
|
||
return NULL;
|
||
}
|
||
|
||
ip_src = get_16aligned_be32(&ip->ip_src);
|
||
ip_dst = get_16aligned_be32(&ip->ip_dst);
|
||
|
||
tnl->ip_src = ip_src;
|
||
tnl->ip_dst = ip_dst;
|
||
tnl->ip_tos = ip->ip_tos;
|
||
tnl->ip_ttl = ip->ip_ttl;
|
||
|
||
*hlen += IP_HEADER_LEN;
|
||
|
||
} else if (IP_VER(ip->ip_ihl_ver) == 6) {
|
||
ovs_be32 tc_flow = get_16aligned_be32(&ip6->ip6_flow);
|
||
|
||
memcpy(tnl->ipv6_src.s6_addr, ip6->ip6_src.be16, sizeof ip6->ip6_src);
|
||
memcpy(tnl->ipv6_dst.s6_addr, ip6->ip6_dst.be16, sizeof ip6->ip6_dst);
|
||
|
||
tnl->ip_tos = ntohl(tc_flow) >> 20;
|
||
tnl->ip_ttl = ip6->ip6_hlim;
|
||
|
||
*hlen += packet->l4_ofs - packet->l3_ofs;
|
||
|
||
} else {
|
||
VLOG_WARN_RL(&err_rl, "ipv4 packet has invalid version (%d)",
|
||
IP_VER(ip->ip_ihl_ver));
|
||
return NULL;
|
||
}
|
||
|
||
return l4;
|
||
}
|
||
|
||
/* Pushes the 'size' bytes of 'header' into the headroom of 'packet',
|
||
* reallocating the packet if necessary. 'header' should contain an Ethernet
|
||
* header, followed by an IPv4 header (without options), and an L4 header.
|
||
*
|
||
* This function sets the IP header's ip_tot_len field (which should be zeroed
|
||
* as part of 'header') and puts its value into '*ip_tot_size' as well. Also
|
||
* updates IP header checksum if not offloaded, as well as the l3 and l4
|
||
* offsets in the 'packet'.
|
||
*
|
||
* Return pointer to the L4 header added to 'packet'. */
|
||
void *
|
||
netdev_tnl_push_ip_header(struct dp_packet *packet, const void *header,
|
||
int size, int *ip_tot_size, ovs_be32 ipv6_label)
|
||
{
|
||
struct eth_header *eth;
|
||
struct ip_header *ip;
|
||
struct ovs_16aligned_ip6_hdr *ip6;
|
||
|
||
eth = dp_packet_push_uninit(packet, size);
|
||
*ip_tot_size = dp_packet_size(packet) - sizeof (struct eth_header);
|
||
|
||
memcpy(eth, header, size);
|
||
/* The encapsulated packet has type Ethernet. Adjust dp_packet. */
|
||
packet->packet_type = htonl(PT_ETH);
|
||
dp_packet_reset_offsets(packet);
|
||
packet->l3_ofs = sizeof (struct eth_header);
|
||
|
||
if (netdev_tnl_is_header_ipv6(header)) {
|
||
ip6 = netdev_tnl_ipv6_hdr(eth);
|
||
*ip_tot_size -= IPV6_HEADER_LEN;
|
||
ip6->ip6_plen = htons(*ip_tot_size);
|
||
packet_set_ipv6_flow_label(&ip6->ip6_flow, ipv6_label);
|
||
dp_packet_ip_checksum_set_unknown(packet);
|
||
|
||
packet->l4_ofs = dp_packet_size(packet) - *ip_tot_size;
|
||
|
||
return ip6 + 1;
|
||
} else {
|
||
ip = netdev_tnl_ip_hdr(eth);
|
||
ip->ip_tot_len = htons(*ip_tot_size);
|
||
*ip_tot_size -= IP_HEADER_LEN;
|
||
/* Postpone checksum to when the packet is pushed to the port. */
|
||
dp_packet_ip_checksum_set_partial(packet);
|
||
|
||
packet->l4_ofs = dp_packet_size(packet) - *ip_tot_size;
|
||
|
||
return ip + 1;
|
||
}
|
||
}
|
||
|
||
static void *
|
||
udp_extract_tnl_md(struct dp_packet *packet, struct flow_tnl *tnl,
|
||
unsigned int *hlen)
|
||
{
|
||
struct udp_header *udp;
|
||
|
||
udp = ip_extract_tnl_md(packet, tnl, hlen);
|
||
if (!udp) {
|
||
return NULL;
|
||
}
|
||
|
||
if (udp->udp_csum) {
|
||
bool bad_csum = dp_packet_l4_checksum_bad(packet);
|
||
|
||
if (OVS_UNLIKELY(!bad_csum && dp_packet_l4_checksum_unknown(packet))) {
|
||
uint32_t csum;
|
||
COVERAGE_INC(native_tnl_l4csum_checked);
|
||
if (netdev_tnl_is_header_ipv6(dp_packet_data(packet))) {
|
||
csum = packet_csum_pseudoheader6(dp_packet_l3(packet));
|
||
} else {
|
||
csum = packet_csum_pseudoheader(dp_packet_l3(packet));
|
||
}
|
||
|
||
csum = csum_continue(csum, udp, dp_packet_size(packet) -
|
||
((const unsigned char *)udp -
|
||
(const unsigned char *)dp_packet_eth(packet)
|
||
));
|
||
if (csum_finish(csum)) {
|
||
dp_packet_l4_checksum_set_bad(packet);
|
||
bad_csum = true;
|
||
} else {
|
||
dp_packet_l4_checksum_set_good(packet);
|
||
}
|
||
}
|
||
if (OVS_UNLIKELY(bad_csum)) {
|
||
COVERAGE_INC(native_tnl_l4csum_err);
|
||
return NULL;
|
||
}
|
||
tnl->flags |= FLOW_TNL_F_CSUM;
|
||
}
|
||
|
||
tnl->tp_src = udp->udp_src;
|
||
tnl->tp_dst = udp->udp_dst;
|
||
|
||
return udp + 1;
|
||
}
|
||
|
||
static void
|
||
tnl_ol_push(struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
packet->offloads <<= DP_PACKET_OL_SHIFT_COUNT;
|
||
|
||
if (data->tnl_type == OVS_VPORT_TYPE_GENEVE) {
|
||
dp_packet_tunnel_set_geneve(packet);
|
||
} else if (data->tnl_type == OVS_VPORT_TYPE_VXLAN) {
|
||
dp_packet_tunnel_set_vxlan(packet);
|
||
} else if (data->tnl_type == OVS_VPORT_TYPE_GRE ||
|
||
data->tnl_type == OVS_VPORT_TYPE_IP6GRE) {
|
||
dp_packet_tunnel_set_gre(packet);
|
||
}
|
||
}
|
||
|
||
static void
|
||
tnl_ol_pop(struct dp_packet *packet, int off)
|
||
{
|
||
packet->offloads >>= DP_PACKET_OL_SHIFT_COUNT;
|
||
|
||
dp_packet_reset_packet(packet, off);
|
||
}
|
||
|
||
void
|
||
netdev_tnl_push_udp_header(const struct netdev *netdev OVS_UNUSED,
|
||
struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
uint16_t l3_ofs = packet->l3_ofs;
|
||
uint16_t l4_ofs = packet->l4_ofs;
|
||
struct udp_header *udp;
|
||
ovs_be16 udp_src;
|
||
int ip_tot_size;
|
||
|
||
/* We may need to re-calculate the hash and this has to be done before
|
||
* modifying the packet. */
|
||
udp_src = netdev_tnl_get_src_port(packet);
|
||
|
||
tnl_ol_push(packet, data);
|
||
udp = netdev_tnl_push_ip_header(packet, data->header, data->header_len,
|
||
&ip_tot_size, 0);
|
||
|
||
udp->udp_src = udp_src;
|
||
udp->udp_len = htons(ip_tot_size);
|
||
|
||
dp_packet_l4_proto_set_udp(packet);
|
||
if (udp->udp_csum) {
|
||
dp_packet_l4_checksum_set_partial(packet);
|
||
} else {
|
||
dp_packet_l4_checksum_set_good(packet);
|
||
}
|
||
|
||
if (l3_ofs != UINT16_MAX) {
|
||
packet->inner_l3_ofs = l3_ofs + data->header_len;
|
||
}
|
||
if (l4_ofs != UINT16_MAX) {
|
||
packet->inner_l4_ofs = l4_ofs + data->header_len;
|
||
}
|
||
}
|
||
|
||
static void *
|
||
eth_build_header(struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
uint16_t eth_proto = params->is_ipv6 ? ETH_TYPE_IPV6 : ETH_TYPE_IP;
|
||
struct eth_header *eth;
|
||
|
||
memset(data->header, 0, sizeof data->header);
|
||
|
||
eth = (struct eth_header *)data->header;
|
||
eth->eth_dst = params->dmac;
|
||
eth->eth_src = params->smac;
|
||
eth->eth_type = htons(eth_proto);
|
||
data->header_len = sizeof(struct eth_header);
|
||
return eth + 1;
|
||
}
|
||
|
||
void *
|
||
netdev_tnl_ip_build_header(struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params,
|
||
uint8_t next_proto, ovs_be32 ipv6_label)
|
||
{
|
||
void *l3;
|
||
|
||
l3 = eth_build_header(data, params);
|
||
if (!params->is_ipv6) {
|
||
ovs_be32 ip_src = in6_addr_get_mapped_ipv4(params->s_ip);
|
||
struct ip_header *ip;
|
||
|
||
ip = (struct ip_header *) l3;
|
||
|
||
ip->ip_ihl_ver = IP_IHL_VER(5, 4);
|
||
ip->ip_tos = params->flow->tunnel.ip_tos;
|
||
ip->ip_ttl = params->flow->tunnel.ip_ttl;
|
||
ip->ip_proto = next_proto;
|
||
put_16aligned_be32(&ip->ip_src, ip_src);
|
||
put_16aligned_be32(&ip->ip_dst, params->flow->tunnel.ip_dst);
|
||
|
||
ip->ip_frag_off = (params->flow->tunnel.flags & FLOW_TNL_F_DONT_FRAGMENT) ?
|
||
htons(IP_DF) : 0;
|
||
|
||
/* The checksum will be calculated when the headers are pushed
|
||
* to the packet if offloading is not enabled. */
|
||
|
||
data->header_len += IP_HEADER_LEN;
|
||
return ip + 1;
|
||
} else {
|
||
struct ovs_16aligned_ip6_hdr *ip6;
|
||
|
||
ip6 = (struct ovs_16aligned_ip6_hdr *) l3;
|
||
|
||
put_16aligned_be32(&ip6->ip6_flow, htonl(6 << 28) |
|
||
htonl(params->flow->tunnel.ip_tos << 20) |
|
||
(ipv6_label & htonl(IPV6_LABEL_MASK)));
|
||
ip6->ip6_hlim = params->flow->tunnel.ip_ttl;
|
||
ip6->ip6_nxt = next_proto;
|
||
memcpy(&ip6->ip6_src, params->s_ip, sizeof(ovs_be32[4]));
|
||
memcpy(&ip6->ip6_dst, ¶ms->flow->tunnel.ipv6_dst, sizeof(ovs_be32[4]));
|
||
|
||
data->header_len += IPV6_HEADER_LEN;
|
||
return ip6 + 1;
|
||
}
|
||
}
|
||
|
||
static void *
|
||
udp_build_header(const struct netdev_tunnel_config *tnl_cfg,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
struct udp_header *udp;
|
||
|
||
udp = netdev_tnl_ip_build_header(data, params, IPPROTO_UDP, 0);
|
||
udp->udp_dst = tnl_cfg->dst_port;
|
||
|
||
if (params->flow->tunnel.flags & FLOW_TNL_F_CSUM) {
|
||
/* Write a value in now to mark that we should compute the checksum
|
||
* later. 0xffff is handy because it is transparent to the
|
||
* calculation. */
|
||
udp->udp_csum = htons(0xffff);
|
||
}
|
||
data->header_len += sizeof *udp;
|
||
return udp + 1;
|
||
}
|
||
|
||
static int
|
||
gre_header_len(ovs_be16 flags)
|
||
{
|
||
int hlen = 4;
|
||
|
||
if (flags & htons(GRE_CSUM)) {
|
||
hlen += 4;
|
||
}
|
||
if (flags & htons(GRE_KEY)) {
|
||
hlen += 4;
|
||
}
|
||
if (flags & htons(GRE_SEQ)) {
|
||
hlen += 4;
|
||
}
|
||
return hlen;
|
||
}
|
||
|
||
static int
|
||
parse_gre_header(struct dp_packet *packet,
|
||
struct flow_tnl *tnl)
|
||
{
|
||
const struct gre_base_hdr *greh;
|
||
ovs_16aligned_be32 *options;
|
||
int hlen;
|
||
unsigned int ulen;
|
||
uint16_t greh_protocol;
|
||
|
||
greh = ip_extract_tnl_md(packet, tnl, &ulen);
|
||
if (!greh) {
|
||
return -EINVAL;
|
||
}
|
||
|
||
if (greh->flags & ~(htons(GRE_CSUM | GRE_KEY | GRE_SEQ))) {
|
||
return -EINVAL;
|
||
}
|
||
|
||
hlen = ulen + gre_header_len(greh->flags);
|
||
if (hlen > dp_packet_size(packet)) {
|
||
return -EINVAL;
|
||
}
|
||
|
||
options = (ovs_16aligned_be32 *)(greh + 1);
|
||
if (greh->flags & htons(GRE_CSUM)) {
|
||
ovs_be16 pkt_csum;
|
||
|
||
pkt_csum = csum(greh, dp_packet_size(packet) -
|
||
((const unsigned char *)greh -
|
||
(const unsigned char *)dp_packet_eth(packet)));
|
||
if (pkt_csum) {
|
||
return -EINVAL;
|
||
}
|
||
tnl->flags = FLOW_TNL_F_CSUM;
|
||
options++;
|
||
}
|
||
|
||
if (greh->flags & htons(GRE_KEY)) {
|
||
tnl->tun_id = be32_to_be64(get_16aligned_be32(options));
|
||
tnl->flags |= FLOW_TNL_F_KEY;
|
||
options++;
|
||
}
|
||
|
||
if (greh->flags & htons(GRE_SEQ)) {
|
||
options++;
|
||
}
|
||
|
||
/* Set the new packet type depending on the GRE protocol field. */
|
||
greh_protocol = ntohs(greh->protocol);
|
||
if (greh_protocol == ETH_TYPE_TEB) {
|
||
packet->packet_type = htonl(PT_ETH);
|
||
} else if (greh_protocol >= ETH_TYPE_MIN) {
|
||
/* Allow all GRE protocol values above 0x5ff as Ethertypes. */
|
||
packet->packet_type = PACKET_TYPE_BE(OFPHTN_ETHERTYPE, greh_protocol);
|
||
} else {
|
||
return -EINVAL;
|
||
}
|
||
|
||
return hlen;
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_gre_pop_header(struct dp_packet *packet)
|
||
{
|
||
const void *data_dp = dp_packet_data(packet);
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
int hlen = sizeof(struct eth_header) + 4;
|
||
|
||
ovs_assert(data_dp);
|
||
|
||
hlen += netdev_tnl_is_header_ipv6(data_dp) ?
|
||
IPV6_HEADER_LEN : IP_HEADER_LEN;
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (hlen > dp_packet_size(packet)) {
|
||
goto err;
|
||
}
|
||
|
||
hlen = parse_gre_header(packet, tnl);
|
||
if (hlen < 0) {
|
||
goto err;
|
||
}
|
||
|
||
tnl_ol_pop(packet, hlen);
|
||
|
||
return packet;
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
void
|
||
netdev_gre_push_header(const struct netdev *netdev,
|
||
struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
struct netdev_vport *dev = netdev_vport_cast(netdev);
|
||
uint16_t l3_ofs = packet->l3_ofs;
|
||
uint16_t l4_ofs = packet->l4_ofs;
|
||
struct gre_base_hdr *greh;
|
||
int ip_tot_size;
|
||
|
||
tnl_ol_push(packet, data);
|
||
|
||
greh = netdev_tnl_push_ip_header(packet, data->header, data->header_len,
|
||
&ip_tot_size, 0);
|
||
|
||
if (greh->flags & htons(GRE_CSUM)) {
|
||
ovs_be16 *csum_opt = (ovs_be16 *) (greh + 1);
|
||
*csum_opt = csum(greh, ip_tot_size);
|
||
}
|
||
|
||
if (greh->flags & htons(GRE_SEQ)) {
|
||
if (!dp_packet_get_tso_segsz(packet)) {
|
||
/* Last 4 bytes are GRE seqno. */
|
||
int seq_ofs = gre_header_len(greh->flags) - 4;
|
||
ovs_16aligned_be32 *seq_opt =
|
||
ALIGNED_CAST(ovs_16aligned_be32 *, (char *) greh + seq_ofs);
|
||
|
||
put_16aligned_be32(seq_opt,
|
||
htonl(atomic_count_inc(&dev->gre_seqno)));
|
||
} else {
|
||
VLOG_WARN_RL(&err_rl, "Cannot use GRE Sequence numbers with TSO.");
|
||
}
|
||
}
|
||
|
||
if (l3_ofs != UINT16_MAX) {
|
||
packet->inner_l3_ofs = l3_ofs + data->header_len;
|
||
}
|
||
if (l4_ofs != UINT16_MAX) {
|
||
packet->inner_l4_ofs = l4_ofs + data->header_len;
|
||
}
|
||
}
|
||
|
||
int
|
||
netdev_gre_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
const struct netdev_tunnel_config *tnl_cfg;
|
||
struct gre_base_hdr *greh;
|
||
ovs_16aligned_be32 *options;
|
||
unsigned int hlen;
|
||
|
||
greh = netdev_tnl_ip_build_header(data, params, IPPROTO_GRE, 0);
|
||
|
||
if (params->flow->packet_type == htonl(PT_ETH)) {
|
||
greh->protocol = htons(ETH_TYPE_TEB);
|
||
} else if (pt_ns(params->flow->packet_type) == OFPHTN_ETHERTYPE) {
|
||
greh->protocol = pt_ns_type_be(params->flow->packet_type);
|
||
} else {
|
||
return EINVAL;
|
||
}
|
||
greh->flags = 0;
|
||
|
||
options = (ovs_16aligned_be32 *) (greh + 1);
|
||
if (params->flow->tunnel.flags & FLOW_TNL_F_CSUM) {
|
||
greh->flags |= htons(GRE_CSUM);
|
||
put_16aligned_be32(options, 0);
|
||
options++;
|
||
}
|
||
|
||
tnl_cfg = netdev_get_tunnel_config(netdev);
|
||
|
||
if (tnl_cfg->out_key_present) {
|
||
greh->flags |= htons(GRE_KEY);
|
||
put_16aligned_be32(options, be64_to_be32(params->flow->tunnel.tun_id));
|
||
options++;
|
||
}
|
||
|
||
if (tnl_cfg->set_seq) {
|
||
greh->flags |= htons(GRE_SEQ);
|
||
/* seqno is updated at push header */
|
||
options++;
|
||
}
|
||
|
||
hlen = (uint8_t *) options - (uint8_t *) greh;
|
||
|
||
data->header_len += hlen;
|
||
if (!params->is_ipv6) {
|
||
data->tnl_type = OVS_VPORT_TYPE_GRE;
|
||
} else {
|
||
data->tnl_type = OVS_VPORT_TYPE_IP6GRE;
|
||
}
|
||
return 0;
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_erspan_pop_header(struct dp_packet *packet)
|
||
{
|
||
const struct gre_base_hdr *greh;
|
||
const struct erspan_base_hdr *ersh;
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
int hlen = sizeof(struct eth_header);
|
||
unsigned int ulen;
|
||
uint16_t greh_protocol;
|
||
|
||
hlen += netdev_tnl_is_header_ipv6(dp_packet_data(packet)) ?
|
||
IPV6_HEADER_LEN : IP_HEADER_LEN;
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (hlen > dp_packet_size(packet)) {
|
||
goto err;
|
||
}
|
||
|
||
greh = ip_extract_tnl_md(packet, tnl, &ulen);
|
||
if (!greh) {
|
||
goto err;
|
||
}
|
||
|
||
greh_protocol = ntohs(greh->protocol);
|
||
if (greh_protocol != ETH_TYPE_ERSPAN1 &&
|
||
greh_protocol != ETH_TYPE_ERSPAN2) {
|
||
goto err;
|
||
}
|
||
|
||
if (greh->flags & ~htons(GRE_SEQ)) {
|
||
goto err;
|
||
}
|
||
|
||
ersh = ERSPAN_HDR(greh);
|
||
tnl->tun_id = be16_to_be64(htons(get_sid(ersh)));
|
||
tnl->erspan_ver = ersh->ver;
|
||
|
||
if (ersh->ver == 1) {
|
||
ovs_16aligned_be32 *index = ALIGNED_CAST(ovs_16aligned_be32 *,
|
||
ersh + 1);
|
||
tnl->erspan_idx = ntohl(get_16aligned_be32(index));
|
||
tnl->flags |= FLOW_TNL_F_KEY;
|
||
hlen = ulen + ERSPAN_GREHDR_LEN + sizeof *ersh + ERSPAN_V1_MDSIZE;
|
||
} else if (ersh->ver == 2) {
|
||
struct erspan_md2 *md2 = ALIGNED_CAST(struct erspan_md2 *, ersh + 1);
|
||
tnl->erspan_dir = md2->dir;
|
||
tnl->erspan_hwid = get_hwid(md2);
|
||
tnl->flags |= FLOW_TNL_F_KEY;
|
||
hlen = ulen + ERSPAN_GREHDR_LEN + sizeof *ersh + ERSPAN_V2_MDSIZE;
|
||
} else {
|
||
VLOG_WARN_RL(&err_rl, "ERSPAN version error %d", ersh->ver);
|
||
goto err;
|
||
}
|
||
|
||
if (hlen > dp_packet_size(packet)) {
|
||
goto err;
|
||
}
|
||
|
||
tnl_ol_pop(packet, hlen);
|
||
|
||
return packet;
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
void
|
||
netdev_erspan_push_header(const struct netdev *netdev,
|
||
struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
struct netdev_vport *dev = netdev_vport_cast(netdev);
|
||
struct erspan_base_hdr *ersh;
|
||
struct gre_base_hdr *greh;
|
||
struct erspan_md2 *md2;
|
||
int ip_tot_size;
|
||
|
||
greh = netdev_tnl_push_ip_header(packet, data->header, data->header_len,
|
||
&ip_tot_size, 0);
|
||
|
||
/* update GRE seqno */
|
||
ovs_16aligned_be32 *seqno = (ovs_16aligned_be32 *) (greh + 1);
|
||
put_16aligned_be32(seqno, htonl(atomic_count_inc(&dev->gre_seqno)));
|
||
|
||
/* update v2 timestamp */
|
||
if (greh->protocol == htons(ETH_TYPE_ERSPAN2)) {
|
||
ersh = ERSPAN_HDR(greh);
|
||
md2 = ALIGNED_CAST(struct erspan_md2 *, ersh + 1);
|
||
put_16aligned_be32(&md2->timestamp, get_erspan_ts(ERSPAN_100US));
|
||
}
|
||
}
|
||
|
||
int
|
||
netdev_erspan_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
const struct netdev_tunnel_config *tnl_cfg;
|
||
struct gre_base_hdr *greh;
|
||
struct erspan_base_hdr *ersh;
|
||
unsigned int hlen;
|
||
uint32_t tun_id;
|
||
int erspan_ver;
|
||
uint16_t sid;
|
||
|
||
greh = netdev_tnl_ip_build_header(data, params, IPPROTO_GRE, 0);
|
||
ersh = ERSPAN_HDR(greh);
|
||
|
||
tun_id = ntohl(be64_to_be32(params->flow->tunnel.tun_id));
|
||
/* ERSPAN only has 10-bit session ID */
|
||
if (tun_id & ~ERSPAN_SID_MASK) {
|
||
return EINVAL;
|
||
} else {
|
||
sid = (uint16_t) tun_id;
|
||
}
|
||
|
||
tnl_cfg = netdev_get_tunnel_config(netdev);
|
||
|
||
if (tnl_cfg->erspan_ver_flow) {
|
||
erspan_ver = params->flow->tunnel.erspan_ver;
|
||
} else {
|
||
erspan_ver = tnl_cfg->erspan_ver;
|
||
}
|
||
|
||
if (erspan_ver == 1) {
|
||
greh->protocol = htons(ETH_TYPE_ERSPAN1);
|
||
greh->flags = htons(GRE_SEQ);
|
||
ersh->ver = 1;
|
||
set_sid(ersh, sid);
|
||
|
||
uint32_t erspan_idx = (tnl_cfg->erspan_idx_flow
|
||
? params->flow->tunnel.erspan_idx
|
||
: tnl_cfg->erspan_idx);
|
||
put_16aligned_be32(ALIGNED_CAST(ovs_16aligned_be32 *, ersh + 1),
|
||
htonl(erspan_idx));
|
||
|
||
hlen = ERSPAN_GREHDR_LEN + sizeof *ersh + ERSPAN_V1_MDSIZE;
|
||
} else if (erspan_ver == 2) {
|
||
struct erspan_md2 *md2 = ALIGNED_CAST(struct erspan_md2 *, ersh + 1);
|
||
|
||
greh->protocol = htons(ETH_TYPE_ERSPAN2);
|
||
greh->flags = htons(GRE_SEQ);
|
||
ersh->ver = 2;
|
||
set_sid(ersh, sid);
|
||
|
||
md2->sgt = 0; /* security group tag */
|
||
md2->gra = 0;
|
||
put_16aligned_be32(&md2->timestamp, 0);
|
||
|
||
if (tnl_cfg->erspan_hwid_flow) {
|
||
set_hwid(md2, params->flow->tunnel.erspan_hwid);
|
||
} else {
|
||
set_hwid(md2, tnl_cfg->erspan_hwid);
|
||
}
|
||
|
||
if (tnl_cfg->erspan_dir_flow) {
|
||
md2->dir = params->flow->tunnel.erspan_dir;
|
||
} else {
|
||
md2->dir = tnl_cfg->erspan_dir;
|
||
}
|
||
|
||
hlen = ERSPAN_GREHDR_LEN + sizeof *ersh + ERSPAN_V2_MDSIZE;
|
||
} else {
|
||
VLOG_WARN_RL(&err_rl, "ERSPAN version error %d", tnl_cfg->erspan_ver);
|
||
return EINVAL;
|
||
}
|
||
|
||
data->header_len += hlen;
|
||
|
||
if (params->is_ipv6) {
|
||
data->tnl_type = OVS_VPORT_TYPE_IP6ERSPAN;
|
||
} else {
|
||
data->tnl_type = OVS_VPORT_TYPE_ERSPAN;
|
||
}
|
||
return 0;
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_gtpu_pop_header(struct dp_packet *packet)
|
||
{
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
struct gtpuhdr *gtph;
|
||
unsigned int gtpu_hlen;
|
||
unsigned int hlen;
|
||
|
||
ovs_assert(packet->l3_ofs > 0);
|
||
ovs_assert(packet->l4_ofs > 0);
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (GTPU_HLEN > dp_packet_l4_size(packet)) {
|
||
goto err;
|
||
}
|
||
|
||
gtph = udp_extract_tnl_md(packet, tnl, &hlen);
|
||
if (!gtph) {
|
||
goto err;
|
||
}
|
||
|
||
tnl->gtpu_flags = gtph->md.flags;
|
||
tnl->gtpu_msgtype = gtph->md.msgtype;
|
||
tnl->tun_id = be32_to_be64(get_16aligned_be32(>ph->teid));
|
||
|
||
if (tnl->gtpu_msgtype == GTPU_MSGTYPE_GPDU) {
|
||
struct ip_header *ip;
|
||
|
||
if (gtph->md.flags & GTPU_S_MASK) {
|
||
gtpu_hlen = GTPU_HLEN + sizeof(struct gtpuhdr_opt);
|
||
} else {
|
||
gtpu_hlen = GTPU_HLEN;
|
||
}
|
||
ip = ALIGNED_CAST(struct ip_header *, (char *)gtph + gtpu_hlen);
|
||
|
||
if (IP_VER(ip->ip_ihl_ver) == 4) {
|
||
packet->packet_type = htonl(PT_IPV4);
|
||
} else if (IP_VER(ip->ip_ihl_ver) == 6) {
|
||
packet->packet_type = htonl(PT_IPV6);
|
||
} else {
|
||
VLOG_WARN_RL(&err_rl, "GTP-U: Receive non-IP packet.");
|
||
}
|
||
tnl_ol_pop(packet, hlen + gtpu_hlen);
|
||
} else {
|
||
/* non-GPDU GTP-U messages, ex: echo request, end marker.
|
||
* Users should redirect these packets to controller, or.
|
||
* any application that handles GTP-U messages, so keep
|
||
* the original packet.
|
||
*/
|
||
packet->packet_type = htonl(PT_ETH);
|
||
VLOG_WARN_ONCE("Receive non-GPDU msgtype: %"PRIu8,
|
||
gtph->md.msgtype);
|
||
}
|
||
|
||
return packet;
|
||
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
void
|
||
netdev_gtpu_push_header(const struct netdev *netdev,
|
||
struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
struct netdev_vport *dev = netdev_vport_cast(netdev);
|
||
struct udp_header *udp;
|
||
struct gtpuhdr *gtpuh;
|
||
ovs_be16 udp_src;
|
||
int ip_tot_size;
|
||
unsigned int payload_len;
|
||
|
||
/* We may need to re-calculate the hash and this has to be done before
|
||
* modifying the packet. */
|
||
udp_src = netdev_tnl_get_src_port(packet);
|
||
|
||
payload_len = dp_packet_size(packet);
|
||
udp = netdev_tnl_push_ip_header(packet, data->header, data->header_len,
|
||
&ip_tot_size, 0);
|
||
udp->udp_src = udp_src;
|
||
udp->udp_len = htons(ip_tot_size);
|
||
/* Postpone checksum to the egress netdev. */
|
||
dp_packet_l4_proto_set_udp(packet);
|
||
dp_packet_l4_checksum_set_partial(packet);
|
||
|
||
gtpuh = ALIGNED_CAST(struct gtpuhdr *, udp + 1);
|
||
|
||
if (gtpuh->md.flags & GTPU_S_MASK) {
|
||
ovs_be16 *seqno = ALIGNED_CAST(ovs_be16 *, gtpuh + 1);
|
||
*seqno = htons(atomic_count_inc(&dev->gre_seqno));
|
||
payload_len += sizeof(struct gtpuhdr_opt);
|
||
}
|
||
gtpuh->len = htons(payload_len);
|
||
}
|
||
|
||
int
|
||
netdev_gtpu_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
const struct netdev_tunnel_config *tnl_cfg;
|
||
struct gtpuhdr *gtph;
|
||
unsigned int gtpu_hlen;
|
||
|
||
tnl_cfg = netdev_get_tunnel_config(netdev);
|
||
|
||
gtph = udp_build_header(tnl_cfg, data, params);
|
||
|
||
/* Set to default if not set in flow. */
|
||
gtph->md.flags = params->flow->tunnel.gtpu_flags ?
|
||
params->flow->tunnel.gtpu_flags : GTPU_FLAGS_DEFAULT;
|
||
gtph->md.msgtype = params->flow->tunnel.gtpu_msgtype ?
|
||
params->flow->tunnel.gtpu_msgtype : GTPU_MSGTYPE_GPDU;
|
||
put_16aligned_be32(>ph->teid,
|
||
be64_to_be32(params->flow->tunnel.tun_id));
|
||
|
||
gtpu_hlen = sizeof *gtph;
|
||
if (tnl_cfg->set_seq) {
|
||
gtph->md.flags |= GTPU_S_MASK;
|
||
gtpu_hlen += sizeof(struct gtpuhdr_opt);
|
||
}
|
||
|
||
data->header_len += gtpu_hlen;
|
||
data->tnl_type = OVS_VPORT_TYPE_GTPU;
|
||
|
||
return 0;
|
||
}
|
||
|
||
int
|
||
netdev_srv6_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
const struct netdev_tunnel_config *tnl_cfg;
|
||
union ovs_16aligned_in6_addr *s;
|
||
const struct in6_addr *segs;
|
||
struct srv6_base_hdr *srh;
|
||
ovs_be16 dl_type;
|
||
int nr_segs;
|
||
int i;
|
||
|
||
tnl_cfg = netdev_get_tunnel_config(netdev);
|
||
if (tnl_cfg->srv6_num_segs) {
|
||
nr_segs = tnl_cfg->srv6_num_segs;
|
||
segs = tnl_cfg->srv6_segs;
|
||
} else {
|
||
/*
|
||
* If explicit segment list setting is omitted, tunnel destination
|
||
* is considered to be the first segment list.
|
||
*/
|
||
nr_segs = 1;
|
||
segs = ¶ms->flow->tunnel.ipv6_dst;
|
||
}
|
||
|
||
if (!ipv6_addr_equals(&segs[0], ¶ms->flow->tunnel.ipv6_dst)) {
|
||
return EINVAL;
|
||
}
|
||
|
||
/* Writes the netdev_srv6_flowlabel enum value to the ipv6
|
||
* flowlabel field. It must later be replaced by a valid value
|
||
* in the header push. */
|
||
srh = netdev_tnl_ip_build_header(data, params, IPPROTO_ROUTING,
|
||
htonl(tnl_cfg->srv6_flowlabel));
|
||
|
||
srh->rt_hdr.segments_left = nr_segs - 1;
|
||
srh->rt_hdr.type = IPV6_SRCRT_TYPE_4;
|
||
srh->rt_hdr.hdrlen = 2 * nr_segs;
|
||
srh->last_entry = nr_segs - 1;
|
||
srh->flags = 0;
|
||
srh->tag = 0;
|
||
|
||
dl_type = params->flow->dl_type;
|
||
if (dl_type == htons(ETH_TYPE_IP)) {
|
||
srh->rt_hdr.nexthdr = IPPROTO_IPIP;
|
||
} else if (dl_type == htons(ETH_TYPE_IPV6)) {
|
||
srh->rt_hdr.nexthdr = IPPROTO_IPV6;
|
||
} else {
|
||
return EOPNOTSUPP;
|
||
}
|
||
|
||
s = (union ovs_16aligned_in6_addr *) (srh + 1);
|
||
for (i = 0; i < nr_segs; i++) {
|
||
/* Segment list is written to the header in reverse order. */
|
||
memcpy(s, &segs[nr_segs - i - 1], sizeof *s);
|
||
s++;
|
||
}
|
||
|
||
data->header_len += sizeof *srh + 8 * srh->rt_hdr.hdrlen;
|
||
data->tnl_type = OVS_VPORT_TYPE_SRV6;
|
||
|
||
return 0;
|
||
}
|
||
|
||
void
|
||
netdev_srv6_push_header(const struct netdev *netdev OVS_UNUSED,
|
||
struct dp_packet *packet,
|
||
const struct ovs_action_push_tnl *data)
|
||
{
|
||
struct ovs_16aligned_ip6_hdr *inner_ip6, *outer_ip6;
|
||
enum netdev_srv6_flowlabel srv6_flowlabel;
|
||
ovs_be32 ipv6_label = 0;
|
||
int ip_tot_size;
|
||
uint32_t flow;
|
||
|
||
inner_ip6 = dp_packet_l3(packet);
|
||
outer_ip6 = netdev_tnl_ipv6_hdr((void *) data->header);
|
||
srv6_flowlabel = ntohl(get_16aligned_be32(&outer_ip6->ip6_flow)) &
|
||
IPV6_LABEL_MASK;
|
||
|
||
switch (srv6_flowlabel) {
|
||
case SRV6_FLOWLABEL_COPY:
|
||
flow = ntohl(get_16aligned_be32(&inner_ip6->ip6_flow));
|
||
ipv6_label = (flow >> 28) == 6 ? htonl(flow & IPV6_LABEL_MASK) : 0;
|
||
break;
|
||
|
||
case SRV6_FLOWLABEL_ZERO:
|
||
ipv6_label = 0;
|
||
break;
|
||
|
||
case SRV6_FLOWLABEL_COMPUTE:
|
||
ipv6_label = htonl(dp_packet_get_rss_hash(packet) & IPV6_LABEL_MASK);
|
||
break;
|
||
}
|
||
|
||
netdev_tnl_push_ip_header(packet, data->header,
|
||
data->header_len, &ip_tot_size, ipv6_label);
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_srv6_pop_header(struct dp_packet *packet)
|
||
{
|
||
const struct ovs_16aligned_ip6_hdr *nh = dp_packet_l3(packet);
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
const struct ip6_rt_hdr *rt_hdr;
|
||
uint8_t nw_proto = nh->ip6_nxt;
|
||
const void *data = nh + 1;
|
||
uint8_t nw_frag = 0;
|
||
unsigned int hlen;
|
||
size_t size;
|
||
|
||
/*
|
||
* Verifies that the routing header is present in the IPv6
|
||
* extension headers and that its type is SRv6.
|
||
*/
|
||
size = dp_packet_l3_size(packet);
|
||
if (size < IPV6_HEADER_LEN) {
|
||
goto err;
|
||
}
|
||
size -= IPV6_HEADER_LEN;
|
||
|
||
if (!parse_ipv6_ext_hdrs(&data, &size, &nw_proto, &nw_frag,
|
||
NULL, &rt_hdr)) {
|
||
goto err;
|
||
}
|
||
|
||
if (!rt_hdr || rt_hdr->type != IPV6_SRCRT_TYPE_4) {
|
||
goto err;
|
||
}
|
||
|
||
if (rt_hdr->segments_left > 0) {
|
||
VLOG_WARN_RL(&err_rl, "invalid srv6 segments_left=%d\n",
|
||
rt_hdr->segments_left);
|
||
goto err;
|
||
}
|
||
|
||
if (rt_hdr->nexthdr == IPPROTO_IPIP) {
|
||
packet->packet_type = htonl(PT_IPV4);
|
||
} else if (rt_hdr->nexthdr == IPPROTO_IPV6) {
|
||
packet->packet_type = htonl(PT_IPV6);
|
||
} else {
|
||
goto err;
|
||
}
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (!ip_extract_tnl_md(packet, tnl, &hlen)) {
|
||
goto err;
|
||
}
|
||
|
||
tnl_ol_pop(packet, hlen);
|
||
|
||
return packet;
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_vxlan_pop_header(struct dp_packet *packet)
|
||
{
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
struct vxlanhdr *vxh;
|
||
unsigned int hlen;
|
||
ovs_be32 vx_flags;
|
||
enum packet_type next_pt = PT_ETH;
|
||
|
||
ovs_assert(packet->l3_ofs > 0);
|
||
ovs_assert(packet->l4_ofs > 0);
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (VXLAN_HLEN > dp_packet_l4_size(packet)) {
|
||
goto err;
|
||
}
|
||
|
||
vxh = udp_extract_tnl_md(packet, tnl, &hlen);
|
||
if (!vxh) {
|
||
goto err;
|
||
}
|
||
|
||
vx_flags = get_16aligned_be32(&vxh->vx_flags);
|
||
if (vx_flags & htonl(VXLAN_HF_GPE)) {
|
||
vx_flags &= htonl(~VXLAN_GPE_USED_BITS);
|
||
/* Drop the OAM packets */
|
||
if (vxh->vx_gpe.flags & VXLAN_GPE_FLAGS_O) {
|
||
goto err;
|
||
}
|
||
switch (vxh->vx_gpe.next_protocol) {
|
||
case VXLAN_GPE_NP_IPV4:
|
||
next_pt = PT_IPV4;
|
||
break;
|
||
case VXLAN_GPE_NP_IPV6:
|
||
next_pt = PT_IPV6;
|
||
break;
|
||
case VXLAN_GPE_NP_NSH:
|
||
next_pt = PT_NSH;
|
||
break;
|
||
case VXLAN_GPE_NP_ETHERNET:
|
||
next_pt = PT_ETH;
|
||
break;
|
||
default:
|
||
goto err;
|
||
}
|
||
}
|
||
|
||
if (vx_flags != htonl(VXLAN_FLAGS) ||
|
||
(get_16aligned_be32(&vxh->vx_vni) & htonl(0xff))) {
|
||
VLOG_WARN_RL(&err_rl, "invalid vxlan flags=%#x vni=%#x\n",
|
||
ntohl(vx_flags),
|
||
ntohl(get_16aligned_be32(&vxh->vx_vni)));
|
||
goto err;
|
||
}
|
||
tnl->tun_id = htonll(ntohl(get_16aligned_be32(&vxh->vx_vni)) >> 8);
|
||
tnl->flags |= FLOW_TNL_F_KEY;
|
||
|
||
packet->packet_type = htonl(next_pt);
|
||
tnl_ol_pop(packet, hlen + VXLAN_HLEN);
|
||
if (next_pt != PT_ETH) {
|
||
packet->l3_ofs = 0;
|
||
}
|
||
|
||
return packet;
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
int
|
||
netdev_vxlan_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
const struct netdev_tunnel_config *tnl_cfg;
|
||
struct vxlanhdr *vxh;
|
||
|
||
tnl_cfg = netdev_get_tunnel_config(netdev);
|
||
|
||
vxh = udp_build_header(tnl_cfg, data, params);
|
||
|
||
if (tnl_cfg->exts & (1 << OVS_VXLAN_EXT_GPE)) {
|
||
put_16aligned_be32(&vxh->vx_flags, htonl(VXLAN_FLAGS | VXLAN_HF_GPE));
|
||
put_16aligned_be32(&vxh->vx_vni,
|
||
htonl(ntohll(params->flow->tunnel.tun_id) << 8));
|
||
if (params->flow->packet_type == htonl(PT_ETH)) {
|
||
vxh->vx_gpe.next_protocol = VXLAN_GPE_NP_ETHERNET;
|
||
} else if (pt_ns(params->flow->packet_type) == OFPHTN_ETHERTYPE) {
|
||
switch (pt_ns_type(params->flow->packet_type)) {
|
||
case ETH_TYPE_IP:
|
||
vxh->vx_gpe.next_protocol = VXLAN_GPE_NP_IPV4;
|
||
break;
|
||
case ETH_TYPE_IPV6:
|
||
vxh->vx_gpe.next_protocol = VXLAN_GPE_NP_IPV6;
|
||
break;
|
||
case ETH_TYPE_NSH:
|
||
vxh->vx_gpe.next_protocol = VXLAN_GPE_NP_NSH;
|
||
break;
|
||
case ETH_TYPE_TEB:
|
||
vxh->vx_gpe.next_protocol = VXLAN_GPE_NP_ETHERNET;
|
||
break;
|
||
default:
|
||
return EINVAL;
|
||
}
|
||
} else {
|
||
return EINVAL;
|
||
}
|
||
} else {
|
||
put_16aligned_be32(&vxh->vx_flags, htonl(VXLAN_FLAGS));
|
||
put_16aligned_be32(&vxh->vx_vni,
|
||
htonl(ntohll(params->flow->tunnel.tun_id) << 8));
|
||
}
|
||
|
||
data->header_len += sizeof *vxh;
|
||
data->tnl_type = OVS_VPORT_TYPE_VXLAN;
|
||
return 0;
|
||
}
|
||
|
||
struct dp_packet *
|
||
netdev_geneve_pop_header(struct dp_packet *packet)
|
||
{
|
||
struct pkt_metadata *md = &packet->md;
|
||
struct flow_tnl *tnl = &md->tunnel;
|
||
struct genevehdr *gnh;
|
||
unsigned int hlen, opts_len, ulen;
|
||
|
||
pkt_metadata_init_tnl(md);
|
||
if (GENEVE_BASE_HLEN > dp_packet_l4_size(packet)) {
|
||
VLOG_WARN_RL(&err_rl, "geneve packet too small: min header=%u packet size=%"PRIuSIZE"\n",
|
||
(unsigned int)GENEVE_BASE_HLEN, dp_packet_l4_size(packet));
|
||
goto err;
|
||
}
|
||
|
||
gnh = udp_extract_tnl_md(packet, tnl, &ulen);
|
||
if (!gnh) {
|
||
goto err;
|
||
}
|
||
|
||
opts_len = gnh->opt_len * 4;
|
||
hlen = ulen + GENEVE_BASE_HLEN + opts_len;
|
||
if (hlen > dp_packet_size(packet)) {
|
||
VLOG_WARN_RL(&err_rl, "geneve packet too small: header len=%u packet size=%u\n",
|
||
hlen, dp_packet_size(packet));
|
||
goto err;
|
||
}
|
||
|
||
if (gnh->ver != 0) {
|
||
VLOG_WARN_RL(&err_rl, "unknown geneve version: %"PRIu8"\n", gnh->ver);
|
||
goto err;
|
||
}
|
||
|
||
if (gnh->proto_type != htons(ETH_TYPE_TEB)) {
|
||
VLOG_WARN_RL(&err_rl, "unknown geneve encapsulated protocol: %#x\n",
|
||
ntohs(gnh->proto_type));
|
||
goto err;
|
||
}
|
||
|
||
tnl->flags |= gnh->oam ? FLOW_TNL_F_OAM : 0;
|
||
tnl->tun_id = htonll(ntohl(get_16aligned_be32(&gnh->vni)) >> 8);
|
||
tnl->flags |= FLOW_TNL_F_KEY;
|
||
|
||
memcpy(tnl->metadata.opts.gnv, gnh->options, opts_len);
|
||
tnl->metadata.present.len = opts_len;
|
||
tnl->flags |= FLOW_TNL_F_UDPIF;
|
||
|
||
packet->packet_type = htonl(PT_ETH);
|
||
tnl_ol_pop(packet, hlen);
|
||
|
||
return packet;
|
||
err:
|
||
dp_packet_delete(packet);
|
||
return NULL;
|
||
}
|
||
|
||
int
|
||
netdev_geneve_build_header(const struct netdev *netdev,
|
||
struct ovs_action_push_tnl *data,
|
||
const struct netdev_tnl_build_header_params *params)
|
||
{
|
||
struct genevehdr *gnh;
|
||
int opt_len;
|
||
bool crit_opt;
|
||
|
||
gnh = udp_build_header(netdev_get_tunnel_config(netdev), data, params);
|
||
|
||
put_16aligned_be32(&gnh->vni, htonl(ntohll(params->flow->tunnel.tun_id) << 8));
|
||
|
||
opt_len = tun_metadata_to_geneve_header(¶ms->flow->tunnel,
|
||
gnh->options, &crit_opt);
|
||
|
||
gnh->opt_len = opt_len / 4;
|
||
gnh->oam = !!(params->flow->tunnel.flags & FLOW_TNL_F_OAM);
|
||
gnh->critical = crit_opt ? 1 : 0;
|
||
gnh->proto_type = htons(ETH_TYPE_TEB);
|
||
|
||
data->header_len += sizeof *gnh + opt_len;
|
||
data->tnl_type = OVS_VPORT_TYPE_GENEVE;
|
||
return 0;
|
||
}
|
||
|
||
|
||
void
|
||
netdev_tnl_egress_port_range(struct unixctl_conn *conn, int argc,
|
||
const char *argv[], void *aux OVS_UNUSED)
|
||
{
|
||
int val1, val2;
|
||
|
||
if (argc < 3) {
|
||
struct ds ds = DS_EMPTY_INITIALIZER;
|
||
|
||
ds_put_format(&ds, "Tunnel UDP source port range: %"PRIu16"-%"PRIu16"\n",
|
||
tnl_udp_port_min, tnl_udp_port_max);
|
||
|
||
unixctl_command_reply(conn, ds_cstr(&ds));
|
||
ds_destroy(&ds);
|
||
return;
|
||
}
|
||
|
||
if (argc != 3) {
|
||
return;
|
||
}
|
||
|
||
val1 = atoi(argv[1]);
|
||
if (val1 <= 0 || val1 > UINT16_MAX) {
|
||
unixctl_command_reply(conn, "Invalid min.");
|
||
return;
|
||
}
|
||
val2 = atoi(argv[2]);
|
||
if (val2 <= 0 || val2 > UINT16_MAX) {
|
||
unixctl_command_reply(conn, "Invalid max.");
|
||
return;
|
||
}
|
||
|
||
if (val1 > val2) {
|
||
tnl_udp_port_min = val2;
|
||
tnl_udp_port_max = val1;
|
||
} else {
|
||
tnl_udp_port_min = val1;
|
||
tnl_udp_port_max = val2;
|
||
}
|
||
seq_change(tnl_conf_seq);
|
||
|
||
unixctl_command_reply(conn, "OK");
|
||
}
|