mirror of
https://github.com/openvswitch/ovs
synced 2025-08-22 09:58:01 +00:00
Upon tunnel output failure, due to routing failure for example, add an explicit drop action, so it will appear in the dp-flows for better visibility for that case. For those, add additional drop reasons. Reviewed-by: Roi Dayan <roid@nvidia.com> Signed-off-by: Eli Britstein <elibr@nvidia.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
1285 lines
42 KiB
C
1285 lines
42 KiB
C
/*
|
|
* Copyright (c) 2009, 2010, 2011, 2012, 2013, 2014, 2015 Nicira, Inc.
|
|
* Copyright (c) 2013 Simon Horman
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <config.h>
|
|
#include "odp-execute.h"
|
|
#include "odp-execute-private.h"
|
|
#include <sys/types.h>
|
|
#include <netinet/in.h>
|
|
#include <arpa/inet.h>
|
|
#include <netinet/icmp6.h>
|
|
#include <netinet/ip6.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#include "coverage.h"
|
|
#include "dp-packet.h"
|
|
#include "dpif.h"
|
|
#include "netlink.h"
|
|
#include "odp-netlink.h"
|
|
#include "odp-util.h"
|
|
#include "packets.h"
|
|
#include "flow.h"
|
|
#include "unaligned.h"
|
|
#include "util.h"
|
|
#include "csum.h"
|
|
#include "conntrack.h"
|
|
#include "openvswitch/vlog.h"
|
|
#include "unixctl.h"
|
|
|
|
VLOG_DEFINE_THIS_MODULE(odp_execute);
|
|
COVERAGE_DEFINE(datapath_drop_sample_error);
|
|
COVERAGE_DEFINE(datapath_drop_nsh_decap_error);
|
|
COVERAGE_DEFINE(drop_action_of_pipeline);
|
|
COVERAGE_DEFINE(drop_action_bridge_not_found);
|
|
COVERAGE_DEFINE(drop_action_recursion_too_deep);
|
|
COVERAGE_DEFINE(drop_action_too_many_resubmit);
|
|
COVERAGE_DEFINE(drop_action_stack_too_deep);
|
|
COVERAGE_DEFINE(drop_action_no_recirculation_context);
|
|
COVERAGE_DEFINE(drop_action_recirculation_conflict);
|
|
COVERAGE_DEFINE(drop_action_too_many_mpls_labels);
|
|
COVERAGE_DEFINE(drop_action_invalid_tunnel_metadata);
|
|
COVERAGE_DEFINE(drop_action_unsupported_packet_type);
|
|
COVERAGE_DEFINE(drop_action_congestion);
|
|
COVERAGE_DEFINE(drop_action_forwarding_disabled);
|
|
COVERAGE_DEFINE(drop_action_tunnel_routing_failed);
|
|
COVERAGE_DEFINE(drop_action_tunnel_output_no_ethernet);
|
|
COVERAGE_DEFINE(drop_action_tunnel_neigh_cache_miss);
|
|
COVERAGE_DEFINE(drop_action_tunnel_header_build_failed);
|
|
|
|
static void
|
|
dp_update_drop_action_counter(enum xlate_error drop_reason,
|
|
int delta)
|
|
{
|
|
static struct vlog_rate_limit rl = VLOG_RATE_LIMIT_INIT(1, 5);
|
|
|
|
switch (drop_reason) {
|
|
case XLATE_OK:
|
|
COVERAGE_ADD(drop_action_of_pipeline, delta);
|
|
break;
|
|
case XLATE_BRIDGE_NOT_FOUND:
|
|
COVERAGE_ADD(drop_action_bridge_not_found, delta);
|
|
break;
|
|
case XLATE_RECURSION_TOO_DEEP:
|
|
COVERAGE_ADD(drop_action_recursion_too_deep, delta);
|
|
break;
|
|
case XLATE_TOO_MANY_RESUBMITS:
|
|
COVERAGE_ADD(drop_action_too_many_resubmit, delta);
|
|
break;
|
|
case XLATE_STACK_TOO_DEEP:
|
|
COVERAGE_ADD(drop_action_stack_too_deep, delta);
|
|
break;
|
|
case XLATE_NO_RECIRCULATION_CONTEXT:
|
|
COVERAGE_ADD(drop_action_no_recirculation_context, delta);
|
|
break;
|
|
case XLATE_RECIRCULATION_CONFLICT:
|
|
COVERAGE_ADD(drop_action_recirculation_conflict, delta);
|
|
break;
|
|
case XLATE_TOO_MANY_MPLS_LABELS:
|
|
COVERAGE_ADD(drop_action_too_many_mpls_labels, delta);
|
|
break;
|
|
case XLATE_INVALID_TUNNEL_METADATA:
|
|
COVERAGE_ADD(drop_action_invalid_tunnel_metadata, delta);
|
|
break;
|
|
case XLATE_UNSUPPORTED_PACKET_TYPE:
|
|
COVERAGE_ADD(drop_action_unsupported_packet_type, delta);
|
|
break;
|
|
case XLATE_CONGESTION_DROP:
|
|
COVERAGE_ADD(drop_action_congestion, delta);
|
|
break;
|
|
case XLATE_FORWARDING_DISABLED:
|
|
COVERAGE_ADD(drop_action_forwarding_disabled, delta);
|
|
break;
|
|
case XLATE_TUNNEL_ROUTING_FAILED:
|
|
COVERAGE_ADD(drop_action_tunnel_routing_failed, delta);
|
|
break;
|
|
case XLATE_TUNNEL_OUTPUT_NO_ETHERNET:
|
|
COVERAGE_ADD(drop_action_tunnel_output_no_ethernet, delta);
|
|
break;
|
|
case XLATE_TUNNEL_NEIGH_CACHE_MISS:
|
|
COVERAGE_ADD(drop_action_tunnel_neigh_cache_miss, delta);
|
|
break;
|
|
case XLATE_TUNNEL_HEADER_BUILD_FAILED:
|
|
COVERAGE_ADD(drop_action_tunnel_header_build_failed, delta);
|
|
break;
|
|
case XLATE_MAX:
|
|
default:
|
|
VLOG_ERR_RL(&rl, "Invalid Drop reason type: %d", drop_reason);
|
|
}
|
|
}
|
|
|
|
/* Masked copy of an ethernet address. 'src' is already properly masked. */
|
|
static void
|
|
ether_addr_copy_masked(struct eth_addr *dst, const struct eth_addr src,
|
|
const struct eth_addr mask)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(dst->be16); i++) {
|
|
dst->be16[i] = src.be16[i] | (dst->be16[i] & ~mask.be16[i]);
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_eth_set_addrs(struct dp_packet *packet, const struct ovs_key_ethernet *key,
|
|
const struct ovs_key_ethernet *mask)
|
|
{
|
|
struct eth_header *eh = dp_packet_eth(packet);
|
|
|
|
if (eh) {
|
|
if (!mask) {
|
|
eh->eth_src = key->eth_src;
|
|
eh->eth_dst = key->eth_dst;
|
|
} else {
|
|
ether_addr_copy_masked(&eh->eth_src, key->eth_src, mask->eth_src);
|
|
ether_addr_copy_masked(&eh->eth_dst, key->eth_dst, mask->eth_dst);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_set_ipv4(struct dp_packet *packet, const struct ovs_key_ipv4 *key,
|
|
const struct ovs_key_ipv4 *mask)
|
|
{
|
|
struct ip_header *nh = dp_packet_l3(packet);
|
|
ovs_be32 ip_src_nh;
|
|
ovs_be32 ip_dst_nh;
|
|
ovs_be32 new_ip_src;
|
|
ovs_be32 new_ip_dst;
|
|
uint8_t new_tos;
|
|
uint8_t new_ttl;
|
|
|
|
ovs_assert(nh);
|
|
|
|
if (mask->ipv4_src) {
|
|
ip_src_nh = get_16aligned_be32(&nh->ip_src);
|
|
new_ip_src = key->ipv4_src | (ip_src_nh & ~mask->ipv4_src);
|
|
|
|
if (ip_src_nh != new_ip_src) {
|
|
packet_set_ipv4_addr(packet, &nh->ip_src, new_ip_src);
|
|
}
|
|
}
|
|
|
|
if (mask->ipv4_dst) {
|
|
ip_dst_nh = get_16aligned_be32(&nh->ip_dst);
|
|
new_ip_dst = key->ipv4_dst | (ip_dst_nh & ~mask->ipv4_dst);
|
|
|
|
if (ip_dst_nh != new_ip_dst) {
|
|
packet_set_ipv4_addr(packet, &nh->ip_dst, new_ip_dst);
|
|
}
|
|
}
|
|
|
|
if (mask->ipv4_tos) {
|
|
new_tos = key->ipv4_tos | (nh->ip_tos & ~mask->ipv4_tos);
|
|
|
|
if (nh->ip_tos != new_tos) {
|
|
if (dp_packet_hwol_tx_ip_csum(packet)) {
|
|
dp_packet_ol_reset_ip_csum_good(packet);
|
|
} else {
|
|
nh->ip_csum = recalc_csum16(nh->ip_csum,
|
|
htons((uint16_t) nh->ip_tos),
|
|
htons((uint16_t) new_tos));
|
|
}
|
|
|
|
nh->ip_tos = new_tos;
|
|
}
|
|
}
|
|
|
|
if (OVS_LIKELY(mask->ipv4_ttl)) {
|
|
new_ttl = key->ipv4_ttl | (nh->ip_ttl & ~mask->ipv4_ttl);
|
|
|
|
if (OVS_LIKELY(nh->ip_ttl != new_ttl)) {
|
|
if (dp_packet_hwol_tx_ip_csum(packet)) {
|
|
dp_packet_ol_reset_ip_csum_good(packet);
|
|
} else {
|
|
nh->ip_csum = recalc_csum16(nh->ip_csum,
|
|
htons(nh->ip_ttl << 8),
|
|
htons(new_ttl << 8));
|
|
}
|
|
|
|
nh->ip_ttl = new_ttl;
|
|
}
|
|
}
|
|
}
|
|
|
|
static struct in6_addr *
|
|
mask_ipv6_addr(const ovs_16aligned_be32 *old, const struct in6_addr *addr,
|
|
const struct in6_addr *mask, struct in6_addr *masked)
|
|
{
|
|
#ifdef s6_addr32
|
|
for (int i = 0; i < 4; i++) {
|
|
masked->s6_addr32[i] = addr->s6_addr32[i]
|
|
| (get_16aligned_be32(&old[i]) & ~mask->s6_addr32[i]);
|
|
}
|
|
#else
|
|
const uint8_t *old8 = (const uint8_t *)old;
|
|
for (int i = 0; i < 16; i++) {
|
|
masked->s6_addr[i] = addr->s6_addr[i] | (old8[i] & ~mask->s6_addr[i]);
|
|
}
|
|
#endif
|
|
return masked;
|
|
}
|
|
|
|
static void
|
|
odp_set_ipv6(struct dp_packet *packet, const struct ovs_key_ipv6 *key,
|
|
const struct ovs_key_ipv6 *mask)
|
|
{
|
|
struct ovs_16aligned_ip6_hdr *nh = dp_packet_l3(packet);
|
|
struct in6_addr sbuf, dbuf;
|
|
uint8_t old_tc = ntohl(get_16aligned_be32(&nh->ip6_flow)) >> 20;
|
|
ovs_be32 old_fl = get_16aligned_be32(&nh->ip6_flow) & htonl(0xfffff);
|
|
|
|
packet_set_ipv6(
|
|
packet,
|
|
mask_ipv6_addr(nh->ip6_src.be32, &key->ipv6_src, &mask->ipv6_src,
|
|
&sbuf),
|
|
mask_ipv6_addr(nh->ip6_dst.be32, &key->ipv6_dst, &mask->ipv6_dst,
|
|
&dbuf),
|
|
key->ipv6_tclass | (old_tc & ~mask->ipv6_tclass),
|
|
key->ipv6_label | (old_fl & ~mask->ipv6_label),
|
|
key->ipv6_hlimit | (nh->ip6_hlim & ~mask->ipv6_hlimit));
|
|
}
|
|
|
|
static void
|
|
odp_set_tcp(struct dp_packet *packet, const struct ovs_key_tcp *key,
|
|
const struct ovs_key_tcp *mask)
|
|
{
|
|
struct tcp_header *th = dp_packet_l4(packet);
|
|
|
|
if (OVS_LIKELY(th && dp_packet_get_tcp_payload(packet))) {
|
|
packet_set_tcp_port(packet,
|
|
key->tcp_src | (th->tcp_src & ~mask->tcp_src),
|
|
key->tcp_dst | (th->tcp_dst & ~mask->tcp_dst));
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_set_udp(struct dp_packet *packet, const struct ovs_key_udp *key,
|
|
const struct ovs_key_udp *mask)
|
|
{
|
|
struct udp_header *uh = dp_packet_l4(packet);
|
|
|
|
if (OVS_LIKELY(uh && dp_packet_get_udp_payload(packet))) {
|
|
packet_set_udp_port(packet,
|
|
key->udp_src | (uh->udp_src & ~mask->udp_src),
|
|
key->udp_dst | (uh->udp_dst & ~mask->udp_dst));
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_set_sctp(struct dp_packet *packet, const struct ovs_key_sctp *key,
|
|
const struct ovs_key_sctp *mask)
|
|
{
|
|
struct sctp_header *sh = dp_packet_l4(packet);
|
|
|
|
if (OVS_LIKELY(sh && dp_packet_get_sctp_payload(packet))) {
|
|
packet_set_sctp_port(packet,
|
|
key->sctp_src | (sh->sctp_src & ~mask->sctp_src),
|
|
key->sctp_dst | (sh->sctp_dst & ~mask->sctp_dst));
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_set_tunnel_action(const struct nlattr *a, struct flow_tnl *tun_key)
|
|
{
|
|
ovs_assert(odp_tun_key_from_attr(a, tun_key, NULL) != ODP_FIT_ERROR);
|
|
}
|
|
|
|
static void
|
|
set_arp(struct dp_packet *packet, const struct ovs_key_arp *key,
|
|
const struct ovs_key_arp *mask)
|
|
{
|
|
struct arp_eth_header *arp = dp_packet_l3(packet);
|
|
|
|
ovs_assert(arp);
|
|
|
|
if (!mask) {
|
|
arp->ar_op = key->arp_op;
|
|
arp->ar_sha = key->arp_sha;
|
|
put_16aligned_be32(&arp->ar_spa, key->arp_sip);
|
|
arp->ar_tha = key->arp_tha;
|
|
put_16aligned_be32(&arp->ar_tpa, key->arp_tip);
|
|
} else {
|
|
ovs_be32 ar_spa = get_16aligned_be32(&arp->ar_spa);
|
|
ovs_be32 ar_tpa = get_16aligned_be32(&arp->ar_tpa);
|
|
|
|
arp->ar_op = key->arp_op | (arp->ar_op & ~mask->arp_op);
|
|
ether_addr_copy_masked(&arp->ar_sha, key->arp_sha, mask->arp_sha);
|
|
put_16aligned_be32(&arp->ar_spa,
|
|
key->arp_sip | (ar_spa & ~mask->arp_sip));
|
|
ether_addr_copy_masked(&arp->ar_tha, key->arp_tha, mask->arp_tha);
|
|
put_16aligned_be32(&arp->ar_tpa,
|
|
key->arp_tip | (ar_tpa & ~mask->arp_tip));
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_set_nd_ext(struct dp_packet *packet, const struct ovs_key_nd_extensions
|
|
*key, const struct ovs_key_nd_extensions *mask)
|
|
{
|
|
const struct ovs_nd_msg *ns = dp_packet_l4(packet);
|
|
ovs_16aligned_be32 reserved = ns->rso_flags;
|
|
uint8_t opt_type = ns->options[0].type;
|
|
|
|
if (mask->nd_reserved) {
|
|
put_16aligned_be32(&reserved, key->nd_reserved);
|
|
}
|
|
if (mask->nd_options_type) {
|
|
opt_type = key->nd_options_type;
|
|
}
|
|
packet_set_nd_ext(packet, reserved, opt_type);
|
|
}
|
|
|
|
static void
|
|
odp_set_nd(struct dp_packet *packet, const struct ovs_key_nd *key,
|
|
const struct ovs_key_nd *mask)
|
|
{
|
|
const struct ovs_nd_msg *ns = dp_packet_l4(packet);
|
|
const struct ovs_nd_lla_opt *lla_opt = dp_packet_get_nd_payload(packet);
|
|
|
|
if (OVS_LIKELY(ns && lla_opt)) {
|
|
int bytes_remain = dp_packet_l4_size(packet) - sizeof(*ns);
|
|
struct in6_addr tgt_buf;
|
|
struct eth_addr sll_buf = eth_addr_zero;
|
|
struct eth_addr tll_buf = eth_addr_zero;
|
|
|
|
while (bytes_remain >= ND_LLA_OPT_LEN && lla_opt->len != 0) {
|
|
if (lla_opt->type == ND_OPT_SOURCE_LINKADDR
|
|
&& lla_opt->len == 1) {
|
|
sll_buf = lla_opt->mac;
|
|
ether_addr_copy_masked(&sll_buf, key->nd_sll, mask->nd_sll);
|
|
|
|
/* A packet can only contain one SLL or TLL option */
|
|
break;
|
|
} else if (lla_opt->type == ND_OPT_TARGET_LINKADDR
|
|
&& lla_opt->len == 1) {
|
|
tll_buf = lla_opt->mac;
|
|
ether_addr_copy_masked(&tll_buf, key->nd_tll, mask->nd_tll);
|
|
|
|
/* A packet can only contain one SLL or TLL option */
|
|
break;
|
|
}
|
|
|
|
lla_opt += lla_opt->len;
|
|
bytes_remain -= lla_opt->len * ND_LLA_OPT_LEN;
|
|
}
|
|
|
|
packet_set_nd(packet,
|
|
mask_ipv6_addr(ns->target.be32, &key->nd_target,
|
|
&mask->nd_target, &tgt_buf),
|
|
sll_buf,
|
|
tll_buf);
|
|
}
|
|
}
|
|
|
|
/* Set the NSH header. Assumes the NSH header is present and matches the
|
|
* MD format of the key. The slow path must take case of that. */
|
|
static void
|
|
odp_set_nsh(struct dp_packet *packet, const struct nlattr *a, bool has_mask)
|
|
{
|
|
struct ovs_key_nsh key, mask;
|
|
struct nsh_hdr *nsh = dp_packet_l3(packet);
|
|
uint8_t mdtype = nsh_md_type(nsh);
|
|
ovs_be32 path_hdr;
|
|
|
|
if (has_mask) {
|
|
odp_nsh_key_from_attr(a, &key, &mask, NULL);
|
|
} else {
|
|
odp_nsh_key_from_attr(a, &key, NULL, NULL);
|
|
}
|
|
|
|
if (!has_mask) {
|
|
nsh_set_flags_and_ttl(nsh, key.flags, key.ttl);
|
|
put_16aligned_be32(&nsh->path_hdr, key.path_hdr);
|
|
switch (mdtype) {
|
|
case NSH_M_TYPE1:
|
|
for (int i = 0; i < 4; i++) {
|
|
put_16aligned_be32(&nsh->md1.context[i], key.context[i]);
|
|
}
|
|
break;
|
|
case NSH_M_TYPE2:
|
|
default:
|
|
/* No support for setting any other metadata format yet. */
|
|
break;
|
|
}
|
|
} else {
|
|
uint8_t flags = nsh_get_flags(nsh);
|
|
uint8_t ttl = nsh_get_ttl(nsh);
|
|
|
|
flags = key.flags | (flags & ~mask.flags);
|
|
ttl = key.ttl | (ttl & ~mask.ttl);
|
|
nsh_set_flags_and_ttl(nsh, flags, ttl);
|
|
|
|
uint32_t spi = ntohl(nsh_get_spi(nsh));
|
|
uint8_t si = nsh_get_si(nsh);
|
|
uint32_t spi_mask = nsh_path_hdr_to_spi_uint32(mask.path_hdr);
|
|
uint8_t si_mask = nsh_path_hdr_to_si(mask.path_hdr);
|
|
if (spi_mask == 0x00ffffff) {
|
|
spi_mask = UINT32_MAX;
|
|
}
|
|
spi = nsh_path_hdr_to_spi_uint32(key.path_hdr) | (spi & ~spi_mask);
|
|
si = nsh_path_hdr_to_si(key.path_hdr) | (si & ~si_mask);
|
|
path_hdr = nsh_get_path_hdr(nsh);
|
|
nsh_path_hdr_set_spi(&path_hdr, htonl(spi));
|
|
nsh_path_hdr_set_si(&path_hdr, si);
|
|
put_16aligned_be32(&nsh->path_hdr, path_hdr);
|
|
switch (mdtype) {
|
|
case NSH_M_TYPE1:
|
|
for (int i = 0; i < 4; i++) {
|
|
ovs_be32 p = get_16aligned_be32(&nsh->md1.context[i]);
|
|
ovs_be32 k = key.context[i];
|
|
ovs_be32 m = mask.context[i];
|
|
put_16aligned_be32(&nsh->md1.context[i], k | (p & ~m));
|
|
}
|
|
break;
|
|
case NSH_M_TYPE2:
|
|
default:
|
|
/* No support for setting any other metadata format yet. */
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_execute_set_action(struct dp_packet *packet, const struct nlattr *a)
|
|
{
|
|
enum ovs_key_attr type = nl_attr_type(a);
|
|
const struct ovs_key_ipv4 *ipv4_key;
|
|
const struct ovs_key_ipv6 *ipv6_key;
|
|
struct pkt_metadata *md = &packet->md;
|
|
|
|
switch (type) {
|
|
case OVS_KEY_ATTR_PRIORITY:
|
|
md->skb_priority = nl_attr_get_u32(a);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_TUNNEL:
|
|
odp_set_tunnel_action(a, &md->tunnel);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_SKB_MARK:
|
|
md->pkt_mark = nl_attr_get_u32(a);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ETHERNET:
|
|
odp_eth_set_addrs(packet, nl_attr_get(a), NULL);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_NSH: {
|
|
odp_set_nsh(packet, a, false);
|
|
break;
|
|
}
|
|
|
|
case OVS_KEY_ATTR_IPV4:
|
|
ipv4_key = nl_attr_get_unspec(a, sizeof(struct ovs_key_ipv4));
|
|
packet_set_ipv4(packet, ipv4_key->ipv4_src,
|
|
ipv4_key->ipv4_dst, ipv4_key->ipv4_tos,
|
|
ipv4_key->ipv4_ttl);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_IPV6:
|
|
ipv6_key = nl_attr_get_unspec(a, sizeof(struct ovs_key_ipv6));
|
|
packet_set_ipv6(packet, &ipv6_key->ipv6_src, &ipv6_key->ipv6_dst,
|
|
ipv6_key->ipv6_tclass, ipv6_key->ipv6_label,
|
|
ipv6_key->ipv6_hlimit);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_TCP:
|
|
if (OVS_LIKELY(dp_packet_get_tcp_payload(packet))) {
|
|
const struct ovs_key_tcp *tcp_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_tcp));
|
|
|
|
packet_set_tcp_port(packet, tcp_key->tcp_src,
|
|
tcp_key->tcp_dst);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_UDP:
|
|
if (OVS_LIKELY(dp_packet_get_udp_payload(packet))) {
|
|
const struct ovs_key_udp *udp_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_udp));
|
|
|
|
packet_set_udp_port(packet, udp_key->udp_src,
|
|
udp_key->udp_dst);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_SCTP:
|
|
if (OVS_LIKELY(dp_packet_get_sctp_payload(packet))) {
|
|
const struct ovs_key_sctp *sctp_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_sctp));
|
|
|
|
packet_set_sctp_port(packet, sctp_key->sctp_src,
|
|
sctp_key->sctp_dst);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_MPLS:
|
|
set_mpls_lse(packet, nl_attr_get_be32(a));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ARP:
|
|
set_arp(packet, nl_attr_get(a), NULL);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ICMP:
|
|
case OVS_KEY_ATTR_ICMPV6:
|
|
if (OVS_LIKELY(dp_packet_get_icmp_payload(packet))) {
|
|
const struct ovs_key_icmp *icmp_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_icmp));
|
|
|
|
packet_set_icmp(packet, icmp_key->icmp_type, icmp_key->icmp_code);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ND:
|
|
if (OVS_LIKELY(dp_packet_get_nd_payload(packet))) {
|
|
const struct ovs_key_nd *nd_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_nd));
|
|
packet_set_nd(packet, &nd_key->nd_target, nd_key->nd_sll,
|
|
nd_key->nd_tll);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ND_EXTENSIONS:
|
|
if (OVS_LIKELY(dp_packet_get_nd_payload(packet))) {
|
|
const struct ovs_key_nd_extensions *nd_ext_key
|
|
= nl_attr_get_unspec(a, sizeof(struct ovs_key_nd_extensions));
|
|
ovs_16aligned_be32 rso_flags;
|
|
put_16aligned_be32(&rso_flags, nd_ext_key->nd_reserved);
|
|
packet_set_nd_ext(packet, rso_flags, nd_ext_key->nd_options_type);
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_DP_HASH:
|
|
md->dp_hash = nl_attr_get_u32(a);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_RECIRC_ID:
|
|
md->recirc_id = nl_attr_get_u32(a);
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_UNSPEC:
|
|
case OVS_KEY_ATTR_PACKET_TYPE:
|
|
case OVS_KEY_ATTR_ENCAP:
|
|
case OVS_KEY_ATTR_ETHERTYPE:
|
|
case OVS_KEY_ATTR_IN_PORT:
|
|
case OVS_KEY_ATTR_VLAN:
|
|
case OVS_KEY_ATTR_TCP_FLAGS:
|
|
case OVS_KEY_ATTR_CT_STATE:
|
|
case OVS_KEY_ATTR_CT_ORIG_TUPLE_IPV4:
|
|
case OVS_KEY_ATTR_CT_ORIG_TUPLE_IPV6:
|
|
case OVS_KEY_ATTR_CT_ZONE:
|
|
case OVS_KEY_ATTR_CT_MARK:
|
|
case OVS_KEY_ATTR_CT_LABELS:
|
|
case OVS_KEY_ATTR_TUNNEL_INFO:
|
|
case __OVS_KEY_ATTR_MAX:
|
|
default:
|
|
OVS_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_execute_masked_set_action(struct dp_packet *packet,
|
|
const struct nlattr *a)
|
|
{
|
|
struct pkt_metadata *md = &packet->md;
|
|
enum ovs_key_attr type = nl_attr_type(a);
|
|
struct mpls_hdr *mh;
|
|
|
|
switch (type) {
|
|
case OVS_KEY_ATTR_PRIORITY:
|
|
md->skb_priority = nl_attr_get_u32(a)
|
|
| (md->skb_priority & ~*odp_get_key_mask(a, uint32_t));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_SKB_MARK:
|
|
md->pkt_mark = nl_attr_get_u32(a)
|
|
| (md->pkt_mark & ~*odp_get_key_mask(a, uint32_t));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ETHERNET:
|
|
odp_eth_set_addrs(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_ethernet));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_NSH: {
|
|
odp_set_nsh(packet, a, true);
|
|
break;
|
|
}
|
|
|
|
case OVS_KEY_ATTR_IPV4:
|
|
odp_set_ipv4(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_ipv4));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_IPV6:
|
|
odp_set_ipv6(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_ipv6));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_TCP:
|
|
odp_set_tcp(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_tcp));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_UDP:
|
|
odp_set_udp(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_udp));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_SCTP:
|
|
odp_set_sctp(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_sctp));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_MPLS:
|
|
mh = dp_packet_l2_5(packet);
|
|
if (mh) {
|
|
put_16aligned_be32(&mh->mpls_lse, nl_attr_get_be32(a)
|
|
| (get_16aligned_be32(&mh->mpls_lse)
|
|
& ~*odp_get_key_mask(a, ovs_be32)));
|
|
}
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ARP:
|
|
set_arp(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_arp));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ND:
|
|
odp_set_nd(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_nd));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_ND_EXTENSIONS:
|
|
odp_set_nd_ext(packet, nl_attr_get(a),
|
|
odp_get_key_mask(a, struct ovs_key_nd_extensions));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_DP_HASH:
|
|
md->dp_hash = nl_attr_get_u32(a)
|
|
| (md->dp_hash & ~*odp_get_key_mask(a, uint32_t));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_RECIRC_ID:
|
|
md->recirc_id = nl_attr_get_u32(a)
|
|
| (md->recirc_id & ~*odp_get_key_mask(a, uint32_t));
|
|
break;
|
|
|
|
case OVS_KEY_ATTR_TUNNEL: /* Masked data not supported for tunnel. */
|
|
case OVS_KEY_ATTR_PACKET_TYPE:
|
|
case OVS_KEY_ATTR_UNSPEC:
|
|
case OVS_KEY_ATTR_CT_STATE:
|
|
case OVS_KEY_ATTR_CT_ZONE:
|
|
case OVS_KEY_ATTR_CT_MARK:
|
|
case OVS_KEY_ATTR_CT_LABELS:
|
|
case OVS_KEY_ATTR_CT_ORIG_TUPLE_IPV4:
|
|
case OVS_KEY_ATTR_CT_ORIG_TUPLE_IPV6:
|
|
case OVS_KEY_ATTR_ENCAP:
|
|
case OVS_KEY_ATTR_ETHERTYPE:
|
|
case OVS_KEY_ATTR_IN_PORT:
|
|
case OVS_KEY_ATTR_VLAN:
|
|
case OVS_KEY_ATTR_ICMP:
|
|
case OVS_KEY_ATTR_ICMPV6:
|
|
case OVS_KEY_ATTR_TCP_FLAGS:
|
|
case OVS_KEY_ATTR_TUNNEL_INFO:
|
|
case __OVS_KEY_ATTR_MAX:
|
|
default:
|
|
OVS_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_execute_sample(void *dp, struct dp_packet *packet, bool steal,
|
|
const struct nlattr *action,
|
|
odp_execute_cb dp_execute_action)
|
|
{
|
|
const struct nlattr *subactions = NULL;
|
|
const struct nlattr *a;
|
|
struct dp_packet_batch pb;
|
|
size_t left;
|
|
|
|
NL_NESTED_FOR_EACH_UNSAFE (a, left, action) {
|
|
int type = nl_attr_type(a);
|
|
|
|
switch ((enum ovs_sample_attr) type) {
|
|
case OVS_SAMPLE_ATTR_PROBABILITY:
|
|
if (random_uint32() >= nl_attr_get_u32(a)) {
|
|
if (steal) {
|
|
COVERAGE_INC(datapath_drop_sample_error);
|
|
dp_packet_delete(packet);
|
|
}
|
|
return;
|
|
}
|
|
break;
|
|
|
|
case OVS_SAMPLE_ATTR_ACTIONS:
|
|
subactions = a;
|
|
break;
|
|
|
|
case OVS_SAMPLE_ATTR_UNSPEC:
|
|
case __OVS_SAMPLE_ATTR_MAX:
|
|
default:
|
|
OVS_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
if (!steal) {
|
|
/* The 'subactions' may modify the packet, but the modification
|
|
* should not propagate beyond this sample action. Make a copy
|
|
* the packet in case we don't own the packet, so that the
|
|
* 'subactions' are only applid to the clone. 'odp_execute_actions'
|
|
* will free the clone. */
|
|
packet = dp_packet_clone(packet);
|
|
}
|
|
dp_packet_batch_init_packet(&pb, packet);
|
|
odp_execute_actions(dp, &pb, true, nl_attr_get(subactions),
|
|
nl_attr_get_size(subactions), dp_execute_action);
|
|
}
|
|
|
|
static void
|
|
odp_execute_clone(void *dp, struct dp_packet_batch *batch, bool steal,
|
|
const struct nlattr *actions,
|
|
odp_execute_cb dp_execute_action)
|
|
{
|
|
if (!steal) {
|
|
/* The 'actions' may modify the packet, but the modification
|
|
* should not propagate beyond this clone action. Make a copy
|
|
* the packet in case we don't own the packet, so that the
|
|
* 'actions' are only applied to the clone. 'odp_execute_actions'
|
|
* will free the clone. */
|
|
struct dp_packet_batch clone_pkt_batch;
|
|
dp_packet_batch_clone(&clone_pkt_batch, batch);
|
|
dp_packet_batch_reset_cutlen(batch);
|
|
odp_execute_actions(dp, &clone_pkt_batch, true, nl_attr_get(actions),
|
|
nl_attr_get_size(actions), dp_execute_action);
|
|
}
|
|
else {
|
|
odp_execute_actions(dp, batch, true, nl_attr_get(actions),
|
|
nl_attr_get_size(actions), dp_execute_action);
|
|
}
|
|
}
|
|
|
|
static void
|
|
odp_execute_check_pkt_len(void *dp, struct dp_packet *packet, bool steal,
|
|
const struct nlattr *action,
|
|
odp_execute_cb dp_execute_action)
|
|
{
|
|
static const struct nl_policy ovs_cpl_policy[] = {
|
|
[OVS_CHECK_PKT_LEN_ATTR_PKT_LEN] = { .type = NL_A_U16 },
|
|
[OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_GREATER] = { .type = NL_A_NESTED },
|
|
[OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_LESS_EQUAL]
|
|
= { .type = NL_A_NESTED },
|
|
};
|
|
struct nlattr *attrs[ARRAY_SIZE(ovs_cpl_policy)];
|
|
|
|
if (!nl_parse_nested(action, ovs_cpl_policy, attrs, ARRAY_SIZE(attrs))) {
|
|
OVS_NOT_REACHED();
|
|
}
|
|
|
|
const struct nlattr *a;
|
|
struct dp_packet_batch pb;
|
|
uint32_t size = dp_packet_get_send_len(packet)
|
|
- dp_packet_l2_pad_size(packet);
|
|
|
|
a = attrs[OVS_CHECK_PKT_LEN_ATTR_PKT_LEN];
|
|
if (size > nl_attr_get_u16(a)) {
|
|
a = attrs[OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_GREATER];
|
|
} else {
|
|
a = attrs[OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_LESS_EQUAL];
|
|
}
|
|
|
|
if (!steal) {
|
|
/* The 'subactions' may modify the packet, but the modification
|
|
* should not propagate beyond this action. Make a copy
|
|
* the packet in case we don't own the packet, so that the
|
|
* 'subactions' are only applid to check_pkt_len. 'odp_execute_actions'
|
|
* will free the clone. */
|
|
packet = dp_packet_clone(packet);
|
|
}
|
|
/* If nl_attr_get(a) is NULL, the packet will be freed by
|
|
* odp_execute_actions. */
|
|
dp_packet_batch_init_packet(&pb, packet);
|
|
odp_execute_actions(dp, &pb, true, nl_attr_get(a), nl_attr_get_size(a),
|
|
dp_execute_action);
|
|
}
|
|
|
|
static bool
|
|
requires_datapath_assistance(const struct nlattr *a)
|
|
{
|
|
enum ovs_action_attr type = nl_attr_type(a);
|
|
|
|
switch (type) {
|
|
/* These only make sense in the context of a datapath. */
|
|
case OVS_ACTION_ATTR_OUTPUT:
|
|
case OVS_ACTION_ATTR_LB_OUTPUT:
|
|
case OVS_ACTION_ATTR_TUNNEL_PUSH:
|
|
case OVS_ACTION_ATTR_TUNNEL_POP:
|
|
case OVS_ACTION_ATTR_USERSPACE:
|
|
case OVS_ACTION_ATTR_RECIRC:
|
|
case OVS_ACTION_ATTR_CT:
|
|
case OVS_ACTION_ATTR_METER:
|
|
case OVS_ACTION_ATTR_PSAMPLE:
|
|
return true;
|
|
|
|
case OVS_ACTION_ATTR_SET:
|
|
case OVS_ACTION_ATTR_SET_MASKED:
|
|
case OVS_ACTION_ATTR_PUSH_VLAN:
|
|
case OVS_ACTION_ATTR_POP_VLAN:
|
|
case OVS_ACTION_ATTR_HASH:
|
|
case OVS_ACTION_ATTR_PUSH_MPLS:
|
|
case OVS_ACTION_ATTR_POP_MPLS:
|
|
case OVS_ACTION_ATTR_TRUNC:
|
|
case OVS_ACTION_ATTR_PUSH_ETH:
|
|
case OVS_ACTION_ATTR_POP_ETH:
|
|
case OVS_ACTION_ATTR_CLONE:
|
|
case OVS_ACTION_ATTR_PUSH_NSH:
|
|
case OVS_ACTION_ATTR_POP_NSH:
|
|
case OVS_ACTION_ATTR_CT_CLEAR:
|
|
case OVS_ACTION_ATTR_CHECK_PKT_LEN:
|
|
case OVS_ACTION_ATTR_ADD_MPLS:
|
|
case OVS_ACTION_ATTR_DEC_TTL:
|
|
case OVS_ACTION_ATTR_DROP:
|
|
return false;
|
|
|
|
case OVS_ACTION_ATTR_SAMPLE: {
|
|
/* Nested "psample" actions rely on the datapath executing the
|
|
* parent "sample", storing the probability and making it available
|
|
* when the nested "psample" is run. */
|
|
const struct nlattr *attr;
|
|
unsigned int left;
|
|
|
|
NL_NESTED_FOR_EACH (attr, left, a) {
|
|
if (nl_attr_type(attr) == OVS_SAMPLE_ATTR_ACTIONS) {
|
|
const struct nlattr *act;
|
|
unsigned int act_left;
|
|
|
|
NL_NESTED_FOR_EACH (act, act_left, attr) {
|
|
if (nl_attr_type(act) == OVS_ACTION_ATTR_PSAMPLE) {
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_UNSPEC:
|
|
case __OVS_ACTION_ATTR_MAX:
|
|
OVS_NOT_REACHED();
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static void
|
|
action_pop_vlan(struct dp_packet_batch *batch,
|
|
const struct nlattr *a OVS_UNUSED)
|
|
{
|
|
struct dp_packet *packet;
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
eth_pop_vlan(packet);
|
|
}
|
|
}
|
|
|
|
static void
|
|
action_push_vlan(struct dp_packet_batch *batch, const struct nlattr *a)
|
|
{
|
|
struct dp_packet *packet;
|
|
const struct ovs_action_push_vlan *vlan = nl_attr_get(a);
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
eth_push_vlan(packet, vlan->vlan_tpid, vlan->vlan_tci);
|
|
}
|
|
}
|
|
|
|
static void
|
|
action_set_masked(struct dp_packet_batch *batch, const struct nlattr *a)
|
|
{
|
|
const struct nlattr *key = nl_attr_get(a);
|
|
struct dp_packet *packet;
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
odp_execute_masked_set_action(packet, key);
|
|
}
|
|
}
|
|
|
|
/* Implementation of the scalar actions impl init function. Build up the
|
|
* array of func ptrs here. */
|
|
int
|
|
odp_action_scalar_init(struct odp_execute_action_impl *self)
|
|
{
|
|
/* Set function pointers for actions that can be applied directly, these
|
|
* are identified by OVS_ACTION_ATTR_*. */
|
|
self->funcs[OVS_ACTION_ATTR_POP_VLAN] = action_pop_vlan;
|
|
self->funcs[OVS_ACTION_ATTR_PUSH_VLAN] = action_push_vlan;
|
|
self->funcs[OVS_ACTION_ATTR_SET_MASKED] = action_set_masked;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* The active function pointers on the datapath. ISA optimized implementations
|
|
* are enabled by plugging them into this static arary, which is consulted when
|
|
* applying actions on the datapath. */
|
|
static ATOMIC(struct odp_execute_action_impl *) actions_active_impl;
|
|
|
|
static int
|
|
odp_actions_impl_set(const char *name)
|
|
{
|
|
struct odp_execute_action_impl *active;
|
|
active = odp_execute_action_set(name);
|
|
if (!active) {
|
|
VLOG_ERR("Failed setting action implementation to %s", name);
|
|
return 1;
|
|
}
|
|
|
|
atomic_store_relaxed(&actions_active_impl, active);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
action_impl_set(struct unixctl_conn *conn, int argc OVS_UNUSED,
|
|
const char *argv[], void *aux OVS_UNUSED)
|
|
{
|
|
struct ds reply = DS_EMPTY_INITIALIZER;
|
|
|
|
int err = odp_actions_impl_set(argv[1]);
|
|
if (err) {
|
|
ds_put_format(&reply,
|
|
"Error: unknown action implementation, %s, specified!",
|
|
argv[1]);
|
|
unixctl_command_reply_error(conn, ds_cstr(&reply));
|
|
} else {
|
|
ds_put_format(&reply, "Action implementation set to %s.", argv[1]);
|
|
unixctl_command_reply(conn, ds_cstr(&reply));
|
|
}
|
|
|
|
ds_destroy(&reply);
|
|
}
|
|
|
|
static void
|
|
action_impl_show(struct unixctl_conn *conn, int argc OVS_UNUSED,
|
|
const char *argv[] OVS_UNUSED, void *aux OVS_UNUSED)
|
|
{
|
|
struct ds reply = DS_EMPTY_INITIALIZER;
|
|
|
|
odp_execute_action_get_info(&reply);
|
|
unixctl_command_reply(conn, ds_cstr(&reply));
|
|
ds_destroy(&reply);
|
|
}
|
|
|
|
static void
|
|
odp_execute_unixctl_init(void)
|
|
{
|
|
unixctl_command_register("odp-execute/action-impl-set", "name",
|
|
1, 1, action_impl_set,
|
|
NULL);
|
|
unixctl_command_register("odp-execute/action-impl-show", "",
|
|
0, 0, action_impl_show,
|
|
NULL);
|
|
}
|
|
|
|
void
|
|
odp_execute_init(void)
|
|
{
|
|
static struct ovsthread_once once = OVSTHREAD_ONCE_INITIALIZER;
|
|
if (ovsthread_once_start(&once)) {
|
|
odp_execute_action_init();
|
|
#ifdef ACTIONS_AUTOVALIDATOR_DEFAULT
|
|
odp_actions_impl_set("autovalidator");
|
|
#else
|
|
odp_actions_impl_set("scalar");
|
|
#endif
|
|
odp_execute_unixctl_init();
|
|
ovsthread_once_done(&once);
|
|
}
|
|
}
|
|
|
|
/* Executes all of the 'actions_len' bytes of datapath actions in 'actions' on
|
|
* the packets in 'batch'. If 'steal' is true, possibly modifies and
|
|
* definitely free the packets in 'batch', otherwise leaves 'batch' unchanged.
|
|
*
|
|
* Some actions (e.g. output actions) can only be executed by a datapath. This
|
|
* function implements those actions by passing the action and the packets to
|
|
* 'dp_execute_action' (along with 'dp'). If 'dp_execute_action' is passed a
|
|
* true 'steal' parameter then it must definitely free the packets passed into
|
|
* it. The packet can be modified whether 'steal' is false or true. If a
|
|
* packet is removed from the batch, then the fate of the packet is determined
|
|
* by the code that does this removal, irrespective of the value of 'steal'.
|
|
* Otherwise, if the packet is not removed from the batch and 'steal' is false
|
|
* then the packet could either be cloned or not. */
|
|
void
|
|
odp_execute_actions(void *dp, struct dp_packet_batch *batch, bool steal,
|
|
const struct nlattr *actions, size_t actions_len,
|
|
odp_execute_cb dp_execute_action)
|
|
{
|
|
struct dp_packet *packet;
|
|
const struct nlattr *a;
|
|
unsigned int left;
|
|
|
|
NL_ATTR_FOR_EACH_UNSAFE (a, left, actions, actions_len) {
|
|
int type = nl_attr_type(a);
|
|
enum ovs_action_attr attr_type = (enum ovs_action_attr) type;
|
|
bool last_action = (left <= NLA_ALIGN(a->nla_len));
|
|
|
|
if (requires_datapath_assistance(a)) {
|
|
if (dp_execute_action) {
|
|
/* Allow 'dp_execute_action' to steal the packet data if we do
|
|
* not need it any more. */
|
|
bool should_steal = steal && last_action;
|
|
|
|
dp_execute_action(dp, batch, a, should_steal);
|
|
|
|
if (last_action || dp_packet_batch_is_empty(batch)) {
|
|
/* We do not need to free the packets.
|
|
* Either dp_execute_actions() has stolen them
|
|
* or the batch is freed due to errors. In either
|
|
* case we do not need to execute further actions.
|
|
*/
|
|
return;
|
|
}
|
|
}
|
|
continue;
|
|
}
|
|
|
|
/* If type is set in the active actions implementation, call the
|
|
* function-pointer and continue to the next action. */
|
|
if (attr_type <= OVS_ACTION_ATTR_MAX) {
|
|
/* Read the action implementation pointer atomically to avoid
|
|
* non-atomic read causing corruption if being written by another
|
|
* thread simultaneously. */
|
|
struct odp_execute_action_impl *actions_impl;
|
|
atomic_read_relaxed(&actions_active_impl, &actions_impl);
|
|
|
|
if (actions_impl && actions_impl->funcs[attr_type]) {
|
|
actions_impl->funcs[attr_type](batch, a);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/* If the action was not handled by the active function pointers above,
|
|
* process them by switching on the type below. */
|
|
|
|
switch (attr_type) {
|
|
case OVS_ACTION_ATTR_HASH: {
|
|
const struct ovs_action_hash *hash_act = nl_attr_get(a);
|
|
|
|
/* Calculate a hash value directly. This might not match the
|
|
* value computed by the datapath, but it is much less expensive,
|
|
* and the current use case (bonding) does not require a strict
|
|
* match to work properly. */
|
|
switch (hash_act->hash_alg) {
|
|
case OVS_HASH_ALG_L4: {
|
|
struct flow flow;
|
|
uint32_t hash;
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
/* RSS hash can be used here instead of 5tuple for
|
|
* performance reasons. */
|
|
if (dp_packet_rss_valid(packet)) {
|
|
hash = dp_packet_get_rss_hash(packet);
|
|
hash = hash_int(hash, hash_act->hash_basis);
|
|
} else {
|
|
flow_extract(packet, &flow);
|
|
hash = flow_hash_5tuple(&flow, hash_act->hash_basis);
|
|
}
|
|
packet->md.dp_hash = hash;
|
|
}
|
|
break;
|
|
}
|
|
case OVS_HASH_ALG_SYM_L4: {
|
|
struct flow flow;
|
|
uint32_t hash;
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
flow_extract(packet, &flow);
|
|
hash = flow_hash_symmetric_l3l4(&flow,
|
|
hash_act->hash_basis,
|
|
false);
|
|
packet->md.dp_hash = hash;
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
/* Assert on unknown hash algorithm. */
|
|
OVS_NOT_REACHED();
|
|
}
|
|
break;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_PUSH_MPLS: {
|
|
const struct ovs_action_push_mpls *mpls = nl_attr_get(a);
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
push_mpls(packet, mpls->mpls_ethertype, mpls->mpls_lse);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_POP_MPLS:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
pop_mpls(packet, nl_attr_get_be16(a));
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_SET:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
odp_execute_set_action(packet, nl_attr_get(a));
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_SAMPLE:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
odp_execute_sample(dp, packet, steal && last_action, a,
|
|
dp_execute_action);
|
|
}
|
|
|
|
if (last_action) {
|
|
/* We do not need to free the packets. odp_execute_sample() has
|
|
* stolen them*/
|
|
return;
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_TRUNC: {
|
|
const struct ovs_action_trunc *trunc =
|
|
nl_attr_get_unspec(a, sizeof *trunc);
|
|
|
|
batch->trunc = true;
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
dp_packet_set_cutlen(packet, trunc->max_len);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_CLONE:
|
|
odp_execute_clone(dp, batch, steal && last_action, a,
|
|
dp_execute_action);
|
|
if (last_action) {
|
|
/* We do not need to free the packets. odp_execute_clone() has
|
|
* stolen them. */
|
|
return;
|
|
}
|
|
break;
|
|
case OVS_ACTION_ATTR_METER:
|
|
/* Not implemented yet. */
|
|
break;
|
|
case OVS_ACTION_ATTR_PUSH_ETH: {
|
|
const struct ovs_action_push_eth *eth = nl_attr_get(a);
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
push_eth(packet, ð->addresses.eth_dst,
|
|
ð->addresses.eth_src);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_POP_ETH:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
pop_eth(packet);
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_PUSH_NSH: {
|
|
uint32_t buffer[NSH_HDR_MAX_LEN / 4];
|
|
struct nsh_hdr *nsh_hdr = ALIGNED_CAST(struct nsh_hdr *, buffer);
|
|
nsh_reset_ver_flags_ttl_len(nsh_hdr);
|
|
odp_nsh_hdr_from_attr(nl_attr_get(a), nsh_hdr, NSH_HDR_MAX_LEN);
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
push_nsh(packet, nsh_hdr);
|
|
}
|
|
break;
|
|
}
|
|
case OVS_ACTION_ATTR_POP_NSH: {
|
|
size_t i;
|
|
const size_t num = dp_packet_batch_size(batch);
|
|
|
|
DP_PACKET_BATCH_REFILL_FOR_EACH (i, num, packet, batch) {
|
|
if (pop_nsh(packet)) {
|
|
dp_packet_batch_refill(batch, packet, i);
|
|
} else {
|
|
COVERAGE_INC(datapath_drop_nsh_decap_error);
|
|
dp_packet_delete(packet);
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
case OVS_ACTION_ATTR_CT_CLEAR:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
conntrack_clear(packet);
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_CHECK_PKT_LEN:
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
odp_execute_check_pkt_len(dp, packet, steal && last_action, a,
|
|
dp_execute_action);
|
|
}
|
|
|
|
if (last_action) {
|
|
/* We do not need to free the packets.
|
|
* odp_execute_check_pkt_len() has stolen them. */
|
|
return;
|
|
}
|
|
break;
|
|
|
|
case OVS_ACTION_ATTR_ADD_MPLS: {
|
|
const struct ovs_action_add_mpls *mpls = nl_attr_get(a);
|
|
bool l3_flag = mpls->tun_flags & OVS_MPLS_L3_TUNNEL_FLAG_MASK;
|
|
|
|
DP_PACKET_BATCH_FOR_EACH (i, packet, batch) {
|
|
add_mpls(packet, mpls->mpls_ethertype, mpls->mpls_lse,
|
|
l3_flag);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case OVS_ACTION_ATTR_DROP:{
|
|
const enum xlate_error *drop_reason = nl_attr_get(a);
|
|
|
|
dp_update_drop_action_counter(*drop_reason,
|
|
dp_packet_batch_size(batch));
|
|
dp_packet_delete_batch(batch, steal);
|
|
return;
|
|
}
|
|
case OVS_ACTION_ATTR_OUTPUT:
|
|
case OVS_ACTION_ATTR_LB_OUTPUT:
|
|
case OVS_ACTION_ATTR_TUNNEL_PUSH:
|
|
case OVS_ACTION_ATTR_TUNNEL_POP:
|
|
case OVS_ACTION_ATTR_USERSPACE:
|
|
case OVS_ACTION_ATTR_RECIRC:
|
|
case OVS_ACTION_ATTR_CT:
|
|
case OVS_ACTION_ATTR_UNSPEC:
|
|
case OVS_ACTION_ATTR_DEC_TTL:
|
|
case OVS_ACTION_ATTR_PSAMPLE:
|
|
case __OVS_ACTION_ATTR_MAX:
|
|
/* The following actions are handled by the scalar implementation. */
|
|
case OVS_ACTION_ATTR_POP_VLAN:
|
|
case OVS_ACTION_ATTR_PUSH_VLAN:
|
|
case OVS_ACTION_ATTR_SET_MASKED:
|
|
OVS_NOT_REACHED();
|
|
}
|
|
|
|
/* Do not add any generic processing here, as it won't be executed when
|
|
* an ISA-specific action implementation exists. */
|
|
}
|
|
|
|
dp_packet_delete_batch(batch, steal);
|
|
}
|