2
0
mirror of https://github.com/openvswitch/ovs synced 2025-08-22 09:58:01 +00:00
ovs/lib/dp-packet.c

640 lines
21 KiB
C
Raw Normal View History

/*
* Copyright (c) 2008, 2009, 2010, 2011, 2012, 2013, 2016 Nicira, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <config.h>
#include <stdlib.h>
#include <string.h>
#include "dp-packet.h"
#include "netdev-afxdp.h"
#include "netdev-dpdk.h"
#include "netdev-provider.h"
#include "openvswitch/dynamic-string.h"
#include "util.h"
static void
dp_packet_init__(struct dp_packet *b, size_t allocated, enum dp_packet_source source)
{
dp_packet_set_allocated(b, allocated);
b->source = source;
dp_packet_reset_offsets(b);
pkt_metadata_init(&b->md, 0);
dp_packet_reset_cutlen(b);
dp_packet_reset_offload(b);
dp_packet_set_tso_segsz(b, 0);
/* Initialize implementation-specific fields of dp_packet. */
dp_packet_init_specific(b);
userspace: Add packet_type in dp_packet and flow This commit adds a packet_type attribute to the structs dp_packet and flow to explicitly carry the type of the packet as prepration for the introduction of the so-called packet type-aware pipeline (PTAP) in OVS. The packet_type is a big-endian 32 bit integer with the encoding as specified in OpenFlow verion 1.5. The upper 16 bits contain the packet type name space. Pre-defined values are defined in openflow-common.h: enum ofp_header_type_namespaces { OFPHTN_ONF = 0, /* ONF namespace. */ OFPHTN_ETHERTYPE = 1, /* ns_type is an Ethertype. */ OFPHTN_IP_PROTO = 2, /* ns_type is a IP protocol number. */ OFPHTN_UDP_TCP_PORT = 3, /* ns_type is a TCP or UDP port. */ OFPHTN_IPV4_OPTION = 4, /* ns_type is an IPv4 option number. */ }; The lower 16 bits specify the actual type in the context of the name space. Only name spaces 0 and 1 will be supported for now. For name space OFPHTN_ONF the relevant packet type is 0 (Ethernet). This is the default packet_type in OVS and the only one supported so far. Packets of type (OFPHTN_ONF, 0) are called Ethernet packets. In name space OFPHTN_ETHERTYPE the type is the Ethertype of the packet. A packet of type (OFPHTN_ETHERTYPE, <Ethertype>) is a standard L2 packet whith the Ethernet header (and any VLAN tags) removed to expose the L3 (or L2.5) payload of the packet. These will simply be called L3 packets. The Ethernet address fields dl_src and dl_dst in struct flow are not applicable for an L3 packet and must be zero. However, to maintain compatibility with the large code base, we have chosen to copy the Ethertype of an L3 packet into the the dl_type field of struct flow. This does not mean that it will be possible to match on dl_type for L3 packets with PTAP later on. Matching must be done on packet_type instead. New dp_packets are initialized with packet_type Ethernet. Ports that receive L3 packets will have to explicitly adjust the packet_type. Signed-off-by: Jean Tourrilhes <jt@labs.hpe.com> Signed-off-by: Jan Scheurich <jan.scheurich@ericsson.com> Co-authored-by: Zoltan Balogh <zoltan.balogh@ericsson.com> Signed-off-by: Ben Pfaff <blp@ovn.org>
2017-04-25 16:29:59 +00:00
/* By default assume the packet type to be Ethernet. */
b->packet_type = htonl(PT_ETH);
}
static void
dp_packet_use__(struct dp_packet *b, void *base, size_t allocated,
enum dp_packet_source source)
{
dp_packet_set_base(b, base);
dp_packet_set_data(b, base);
dp_packet_set_size(b, 0);
dp_packet_init__(b, allocated, source);
}
/* Initializes 'b' as an empty dp_packet that contains the 'allocated' bytes of
* memory starting at 'base'. 'base' should be the first byte of a region
* obtained from malloc(). It will be freed (with free()) if 'b' is resized or
* freed. */
void
dp_packet_use(struct dp_packet *b, void *base, size_t allocated)
{
dp_packet_use__(b, base, allocated, DPBUF_MALLOC);
}
#if HAVE_AF_XDP
/* Initialize 'b' as an empty dp_packet that contains
* memory starting at AF_XDP umem base.
*/
void
dp_packet_use_afxdp(struct dp_packet *b, void *data, size_t allocated,
size_t headroom)
{
dp_packet_set_base(b, (char *)data - headroom);
dp_packet_set_data(b, data);
dp_packet_set_size(b, 0);
dp_packet_init__(b, allocated, DPBUF_AFXDP);
}
#endif
/* Initializes 'b' as an empty dp_packet that contains the 'allocated' bytes of
* memory starting at 'base'. 'base' should point to a buffer on the stack.
* (Nothing actually relies on 'base' being allocated on the stack. It could
* be static or malloc()'d memory. But stack space is the most common use
* case.)
*
* 'base' should be appropriately aligned. Using an array of uint32_t or
* uint64_t for the buffer is a reasonable way to ensure appropriate alignment
* for 32- or 64-bit data.
*
* An dp_packet operation that requires reallocating data will copy the provided
* buffer into a malloc()'d buffer. Thus, it is wise to call dp_packet_uninit()
* on an dp_packet initialized by this function, so that if it expanded into the
* heap, that memory is freed. */
void
dp_packet_use_stub(struct dp_packet *b, void *base, size_t allocated)
{
dp_packet_use__(b, base, allocated, DPBUF_STUB);
}
/* Initializes 'b' as an dp_packet whose data starts at 'data' and continues for
* 'size' bytes. This is appropriate for an dp_packet that will be used to
* inspect existing data, without moving it around or reallocating it, and
* generally without modifying it at all.
*
* An dp_packet operation that requires reallocating data will assert-fail if this
* function was used to initialize it. */
void
dp_packet_use_const(struct dp_packet *b, const void *data, size_t size)
{
dp_packet_use__(b, CONST_CAST(void *, data), size, DPBUF_STACK);
dp_packet_set_size(b, size);
}
/* Initializes 'b' as a DPDK dp-packet, which must have been allocated from a
* DPDK memory pool. */
void
dp_packet_init_dpdk(struct dp_packet *b)
{
b->source = DPBUF_DPDK;
}
/* Initializes 'b' as an empty dp_packet with an initial capacity of 'size'
* bytes. */
void
dp_packet_init(struct dp_packet *b, size_t size)
{
dp_packet_use(b, size ? xmalloc(size) : NULL, size);
}
/* Frees memory that 'b' points to. */
void
dp_packet_uninit(struct dp_packet *b)
{
if (b) {
if (b->source == DPBUF_MALLOC) {
free(dp_packet_base(b));
} else if (b->source == DPBUF_DPDK) {
free_dpdk_buf(b);
} else if (b->source == DPBUF_AFXDP) {
free_afxdp_buf(b);
}
}
}
/* Creates and returns a new dp_packet with an initial capacity of 'size'
* bytes. */
struct dp_packet *
dp_packet_new(size_t size)
{
#ifdef DPDK_NETDEV
struct dp_packet *b = xmalloc_cacheline(sizeof *b);
#else
struct dp_packet *b = xmalloc(sizeof *b);
#endif
dp_packet_init(b, size);
return b;
}
/* Creates and returns a new dp_packet with an initial capacity of 'size +
* headroom' bytes, reserving the first 'headroom' bytes as headroom. */
struct dp_packet *
dp_packet_new_with_headroom(size_t size, size_t headroom)
{
struct dp_packet *b = dp_packet_new(size + headroom);
dp_packet_reserve(b, headroom);
return b;
}
/* Creates and returns a new dp_packet that initially contains a copy of the
* 'dp_packet_size(buffer)' bytes of data starting at 'buffer->data' with no headroom or
* tailroom. */
struct dp_packet *
dp_packet_clone(const struct dp_packet *buffer)
{
ovs_assert(buffer);
return dp_packet_clone_with_headroom(buffer, 0);
}
/* Creates and returns a new dp_packet whose data are copied from 'buffer'.
* The returned dp_packet will additionally have 'headroom' bytes of
* headroom. */
struct dp_packet *
dp_packet_clone_with_headroom(const struct dp_packet *buffer, size_t headroom)
{
const void *data_dp = dp_packet_data(buffer);
struct dp_packet *new_buffer;
uint32_t mark;
ovs_assert(data_dp);
new_buffer = dp_packet_clone_data_with_headroom(data_dp,
dp_packet_size(buffer),
headroom);
/* Copy the following fields into the returned buffer: l2_pad_size,
* l2_5_ofs, l3_ofs, l4_ofs, cutlen, packet_type, offloads and md. */
memcpy(&new_buffer->l2_pad_size, &buffer->l2_pad_size,
sizeof(struct dp_packet) -
offsetof(struct dp_packet, l2_pad_size));
dp_packet_set_tso_segsz(new_buffer, dp_packet_get_tso_segsz(buffer));
if (dp_packet_rss_valid(buffer)) {
dp_packet_set_rss_hash(new_buffer, dp_packet_get_rss_hash(buffer));
}
if (dp_packet_has_flow_mark(buffer, &mark)) {
dp_packet_set_flow_mark(new_buffer, mark);
}
return new_buffer;
}
/* Creates and returns a new dp_packet that initially contains a copy of the
* 'size' bytes of data starting at 'data' with no headroom or tailroom. */
struct dp_packet *
dp_packet_clone_data(const void *data, size_t size)
{
return dp_packet_clone_data_with_headroom(data, size, 0);
}
/* Creates and returns a new dp_packet that initially contains 'headroom' bytes of
* headroom followed by a copy of the 'size' bytes of data starting at
* 'data'. */
struct dp_packet *
dp_packet_clone_data_with_headroom(const void *data, size_t size, size_t headroom)
{
struct dp_packet *b = dp_packet_new_with_headroom(size, headroom);
dp_packet_put(b, data, size);
return b;
}
static void
dp_packet_copy__(struct dp_packet *b, uint8_t *new_base,
size_t new_headroom, size_t new_tailroom)
{
const uint8_t *old_base = dp_packet_base(b);
size_t old_headroom = dp_packet_headroom(b);
size_t old_tailroom = dp_packet_tailroom(b);
size_t copy_headroom = MIN(old_headroom, new_headroom);
size_t copy_tailroom = MIN(old_tailroom, new_tailroom);
memcpy(&new_base[new_headroom - copy_headroom],
&old_base[old_headroom - copy_headroom],
copy_headroom + dp_packet_size(b) + copy_tailroom);
}
/* Reallocates 'b' so that it has exactly 'new_headroom' and 'new_tailroom'
* bytes of headroom and tailroom, respectively. */
void
dp_packet_resize(struct dp_packet *b, size_t new_headroom, size_t new_tailroom)
{
void *new_base, *new_data;
size_t new_allocated;
new_allocated = new_headroom + dp_packet_size(b) + new_tailroom;
switch (b->source) {
case DPBUF_DPDK:
OVS_NOT_REACHED();
case DPBUF_MALLOC:
if (new_headroom == dp_packet_headroom(b)) {
new_base = xrealloc(dp_packet_base(b), new_allocated);
} else {
new_base = xmalloc(new_allocated);
dp_packet_copy__(b, new_base, new_headroom, new_tailroom);
free(dp_packet_base(b));
}
break;
case DPBUF_STACK:
OVS_NOT_REACHED();
case DPBUF_AFXDP:
OVS_NOT_REACHED();
case DPBUF_STUB:
b->source = DPBUF_MALLOC;
new_base = xmalloc(new_allocated);
dp_packet_copy__(b, new_base, new_headroom, new_tailroom);
break;
default:
OVS_NOT_REACHED();
}
dp_packet_set_allocated(b, new_allocated);
dp_packet_set_base(b, new_base);
new_data = (char *) new_base + new_headroom;
if (dp_packet_data(b) != new_data) {
dp_packet_set_data(b, new_data);
}
}
/* Ensures that 'b' has room for at least 'size' bytes at its tail end,
* reallocating and copying its data if necessary. Its headroom, if any, is
* preserved. */
void
dp_packet_prealloc_tailroom(struct dp_packet *b, size_t size)
{
if ((size && !dp_packet_base(b)) || (size > dp_packet_tailroom(b))) {
dp_packet_resize(b, dp_packet_headroom(b), MAX(size, 64));
}
}
/* Ensures that 'b' has room for at least 'size' bytes at its head,
* reallocating and copying its data if necessary. Its tailroom, if any, is
* preserved. */
void
dp_packet_prealloc_headroom(struct dp_packet *b, size_t size)
{
if (size > dp_packet_headroom(b)) {
dp_packet_resize(b, MAX(size, 64), dp_packet_tailroom(b));
}
}
/* Shifts all of the data within the allocated space in 'b' by 'delta' bytes.
* For example, a 'delta' of 1 would cause each byte of data to move one byte
* forward (from address 'p' to 'p+1'), and a 'delta' of -1 would cause each
* byte to move one byte backward (from 'p' to 'p-1'). */
void
dp_packet_shift(struct dp_packet *b, int delta)
{
ovs_assert(delta > 0 ? delta <= dp_packet_tailroom(b)
: delta < 0 ? -delta <= dp_packet_headroom(b)
: true);
if (delta != 0) {
const void *data_dp = dp_packet_data(b);
char *dst = (char *) data_dp + delta;
ovs_assert(data_dp);
memmove(dst, data_dp, dp_packet_size(b));
dp_packet_set_data(b, dst);
}
}
/* Appends 'size' bytes of data to the tail end of 'b', reallocating and
* copying its data if necessary. Returns a pointer to the first byte of the
* new data, which is left uninitialized. */
void *
dp_packet_put_uninit(struct dp_packet *b, size_t size)
{
void *p;
dp_packet_prealloc_tailroom(b, size);
p = dp_packet_tail(b);
dp_packet_set_size(b, dp_packet_size(b) + size);
return p;
}
/* Appends 'size' zeroed bytes to the tail end of 'b'. Data in 'b' is
* reallocated and copied if necessary. Returns a pointer to the first byte of
* the data's location in the dp_packet. */
void *
dp_packet_put_zeros(struct dp_packet *b, size_t size)
{
void *dst = dp_packet_put_uninit(b, size);
nullable_memset(dst, 0, size);
return dst;
}
/* Appends the 'size' bytes of data in 'p' to the tail end of 'b'. Data in 'b'
* is reallocated and copied if necessary. Returns a pointer to the first
* byte of the data's location in the dp_packet. */
void *
dp_packet_put(struct dp_packet *b, const void *p, size_t size)
{
void *dst = dp_packet_put_uninit(b, size);
nullable_memcpy(dst, p, size);
return dst;
}
/* Parses as many pairs of hex digits as possible (possibly separated by
* spaces) from the beginning of 's', appending bytes for their values to 'b'.
* Returns the first character of 's' that is not the first of a pair of hex
* digits. If 'n' is nonnull, stores the number of bytes added to 'b' in
* '*n'. */
char *
dp_packet_put_hex(struct dp_packet *b, const char *s, size_t *n)
{
size_t initial_size = dp_packet_size(b);
for (;;) {
uint8_t byte;
bool ok;
s += strspn(s, " \t\r\n");
byte = hexits_value(s, 2, &ok);
if (!ok) {
if (n) {
*n = dp_packet_size(b) - initial_size;
}
return CONST_CAST(char *, s);
}
dp_packet_put(b, &byte, 1);
s += 2;
}
}
/* Reserves 'size' bytes of headroom so that they can be later allocated with
* dp_packet_push_uninit() without reallocating the dp_packet. */
void
dp_packet_reserve(struct dp_packet *b, size_t size)
{
ovs_assert(!dp_packet_size(b));
dp_packet_prealloc_tailroom(b, size);
dp_packet_set_data(b, (char*)dp_packet_data(b) + size);
}
/* Reserves 'headroom' bytes at the head and 'tailroom' at the end so that
* they can be later allocated with dp_packet_push_uninit() or
* dp_packet_put_uninit() without reallocating the dp_packet. */
void
dp_packet_reserve_with_tailroom(struct dp_packet *b, size_t headroom,
size_t tailroom)
{
ovs_assert(!dp_packet_size(b));
dp_packet_prealloc_tailroom(b, headroom + tailroom);
dp_packet_set_data(b, (char*)dp_packet_data(b) + headroom);
}
/* Prefixes 'size' bytes to the head end of 'b', reallocating and copying its
* data if necessary. Returns a pointer to the first byte of the data's
* location in the dp_packet. The new data is left uninitialized. */
void *
dp_packet_push_uninit(struct dp_packet *b, size_t size)
{
dp_packet_prealloc_headroom(b, size);
dp_packet_set_data(b, (char*)dp_packet_data(b) - size);
dp_packet_set_size(b, dp_packet_size(b) + size);
return dp_packet_data(b);
}
/* Prefixes 'size' zeroed bytes to the head end of 'b', reallocating and
* copying its data if necessary. Returns a pointer to the first byte of the
* data's location in the dp_packet. */
void *
dp_packet_push_zeros(struct dp_packet *b, size_t size)
{
void *dst = dp_packet_push_uninit(b, size);
nullable_memset(dst, 0, size);
return dst;
}
/* Copies the 'size' bytes starting at 'p' to the head end of 'b', reallocating
* and copying its data if necessary. Returns a pointer to the first byte of
* the data's location in the dp_packet. */
void *
dp_packet_push(struct dp_packet *b, const void *p, size_t size)
{
void *dst = dp_packet_push_uninit(b, size);
nullable_memcpy(dst, p, size);
return dst;
}
/* Returns the data in 'b' as a block of malloc()'d memory and frees the buffer
* within 'b'. (If 'b' itself was dynamically allocated, e.g. with
* dp_packet_new(), then it should still be freed with, e.g., dp_packet_delete().) */
void *
dp_packet_steal_data(struct dp_packet *b)
{
void *p;
ovs_assert(b->source != DPBUF_DPDK);
ovs_assert(b->source != DPBUF_AFXDP);
if (b->source == DPBUF_MALLOC && dp_packet_data(b) == dp_packet_base(b)) {
p = dp_packet_data(b);
} else {
p = xmemdup(dp_packet_data(b), dp_packet_size(b));
if (b->source == DPBUF_MALLOC) {
free(dp_packet_base(b));
}
}
dp_packet_set_base(b, NULL);
dp_packet_set_data(b, NULL);
return p;
}
static inline void
dp_packet_adjust_layer_offset(uint16_t *offset, int increment)
{
if (*offset != UINT16_MAX) {
*offset += increment;
}
}
/* Adjust the size of the l2_5 portion of the dp_packet, updating the l2
* pointer and the layer offsets. The caller is responsible for
* modifying the contents. */
void *
dp_packet_resize_l2_5(struct dp_packet *b, int increment)
{
if (increment >= 0) {
dp_packet_push_uninit(b, increment);
} else {
dp_packet_pull(b, -increment);
}
/* Adjust layer offsets after l2_5. */
dp_packet_adjust_layer_offset(&b->l3_ofs, increment);
dp_packet_adjust_layer_offset(&b->l4_ofs, increment);
dp_packet_adjust_layer_offset(&b->inner_l3_ofs, increment);
dp_packet_adjust_layer_offset(&b->inner_l4_ofs, increment);
return dp_packet_data(b);
}
/* Adjust the size of the l2 portion of the dp_packet, updating the l2
* pointer and the layer offsets. The caller is responsible for
* modifying the contents. */
void *
dp_packet_resize_l2(struct dp_packet *b, int increment)
{
dp_packet_resize_l2_5(b, increment);
dp_packet_adjust_layer_offset(&b->l2_5_ofs, increment);
return dp_packet_data(b);
}
bool
dp_packet_compare_offsets(struct dp_packet *b1, struct dp_packet *b2,
struct ds *err_str)
{
if ((b1->l2_pad_size != b2->l2_pad_size) ||
(b1->l2_5_ofs != b2->l2_5_ofs) ||
(b1->l3_ofs != b2->l3_ofs) ||
(b1->l4_ofs != b2->l4_ofs) ||
(b1->inner_l3_ofs != b2->inner_l3_ofs) ||
(b1->inner_l4_ofs != b2->inner_l4_ofs)) {
if (err_str) {
ds_put_format(err_str, "Packet offset comparison failed\n");
ds_put_format(err_str, "Buffer 1 offsets: l2_pad_size %u,"
" l2_5_ofs : %u l3_ofs %u, l4_ofs %u,"
" inner_l3_ofs %u, inner_l4_ofs %u\n",
b1->l2_pad_size, b1->l2_5_ofs,
b1->l3_ofs, b1->l4_ofs,
b1->inner_l3_ofs, b1->inner_l4_ofs);
ds_put_format(err_str, "Buffer 2 offsets: l2_pad_size %u,"
" l2_5_ofs : %u l3_ofs %u, l4_ofs %u,"
" inner_l3_ofs %u, inner_l4_ofs %u\n",
b2->l2_pad_size, b2->l2_5_ofs,
b2->l3_ofs, b2->l4_ofs,
b2->inner_l3_ofs, b2->inner_l4_ofs);
}
return false;
}
return true;
}
/* Checks if the packet 'p' is compatible with netdev_ol_flags 'flags'
* and if not, updates the packet with the software fall back. */
void
dp_packet_ol_send_prepare(struct dp_packet *p, uint64_t flags)
{
dp-packet: Rework IP checksum offloads. As the packet traverses through OVS, offloading Tx flags must be carefully evaluated and updated which results in a bit of complexity because of a separate "outer" Tx offloading flag coming from DPDK API, and a "normal"/"inner" Tx offloading flag. On the other hand, the DPDK mbuf API specifies 4 status when it comes to IP checksums: - RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN: no information about the RX IP checksum - RTE_MBUF_F_RX_IP_CKSUM_BAD: the IP checksum in the packet is wrong - RTE_MBUF_F_RX_IP_CKSUM_GOOD: the IP checksum in the packet is valid - RTE_MBUF_F_RX_IP_CKSUM_NONE: the IP checksum is not correct in the packet data, but the integrity of the IP header is verified. This patch changes OVS API so that OVS code only tracks the status of the checksum of the "current" L3 header and let the Tx flags aspect to the netdev-* implementations. With this API, the flow extraction can be cleaned up. During packet processing, OVS can simply look for the IP checksum validity (either good, or partial) before changing some IP header, and then mark the checksum as partial. In the conntrack case, when natting packets, the checksum status of the inner part (ICMP error case) must be forced temporarily as unknown to force checksum resolution. When tunneling comes into play, IP checksums status is bit-shifted for future considerations in the processing if, for example, the tunnel header gets decapsulated again, or in the netdev-* implementations that support tunnel offloading. Finally, netdev-* implementations only need to care about packets in partial status: a good checksum does not need touching, a bad checksum has been updated by kept as bad by OVS, an unknown checksum is either an IPv6 or if it was an IPv4, OVS updated it too (keeping it good or bad accordingly). Rename current API for consistency with dp_packet_(inner_)?ip_checksum_. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
2025-06-17 09:20:57 +02:00
if (!dp_packet_ip_checksum_partial(p)
&& !dp_packet_l4_checksum_partial(p)
dp-packet: Rework IP checksum offloads. As the packet traverses through OVS, offloading Tx flags must be carefully evaluated and updated which results in a bit of complexity because of a separate "outer" Tx offloading flag coming from DPDK API, and a "normal"/"inner" Tx offloading flag. On the other hand, the DPDK mbuf API specifies 4 status when it comes to IP checksums: - RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN: no information about the RX IP checksum - RTE_MBUF_F_RX_IP_CKSUM_BAD: the IP checksum in the packet is wrong - RTE_MBUF_F_RX_IP_CKSUM_GOOD: the IP checksum in the packet is valid - RTE_MBUF_F_RX_IP_CKSUM_NONE: the IP checksum is not correct in the packet data, but the integrity of the IP header is verified. This patch changes OVS API so that OVS code only tracks the status of the checksum of the "current" L3 header and let the Tx flags aspect to the netdev-* implementations. With this API, the flow extraction can be cleaned up. During packet processing, OVS can simply look for the IP checksum validity (either good, or partial) before changing some IP header, and then mark the checksum as partial. In the conntrack case, when natting packets, the checksum status of the inner part (ICMP error case) must be forced temporarily as unknown to force checksum resolution. When tunneling comes into play, IP checksums status is bit-shifted for future considerations in the processing if, for example, the tunnel header gets decapsulated again, or in the netdev-* implementations that support tunnel offloading. Finally, netdev-* implementations only need to care about packets in partial status: a good checksum does not need touching, a bad checksum has been updated by kept as bad by OVS, an unknown checksum is either an IPv6 or if it was an IPv4, OVS updated it too (keeping it good or bad accordingly). Rename current API for consistency with dp_packet_(inner_)?ip_checksum_. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
2025-06-17 09:20:57 +02:00
&& !dp_packet_inner_ip_checksum_partial(p)
&& !dp_packet_inner_l4_checksum_partial(p)) {
/* Only checksumming needs actions. */
return;
}
if (!dp_packet_tunnel(p)) {
dp-packet: Rework IP checksum offloads. As the packet traverses through OVS, offloading Tx flags must be carefully evaluated and updated which results in a bit of complexity because of a separate "outer" Tx offloading flag coming from DPDK API, and a "normal"/"inner" Tx offloading flag. On the other hand, the DPDK mbuf API specifies 4 status when it comes to IP checksums: - RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN: no information about the RX IP checksum - RTE_MBUF_F_RX_IP_CKSUM_BAD: the IP checksum in the packet is wrong - RTE_MBUF_F_RX_IP_CKSUM_GOOD: the IP checksum in the packet is valid - RTE_MBUF_F_RX_IP_CKSUM_NONE: the IP checksum is not correct in the packet data, but the integrity of the IP header is verified. This patch changes OVS API so that OVS code only tracks the status of the checksum of the "current" L3 header and let the Tx flags aspect to the netdev-* implementations. With this API, the flow extraction can be cleaned up. During packet processing, OVS can simply look for the IP checksum validity (either good, or partial) before changing some IP header, and then mark the checksum as partial. In the conntrack case, when natting packets, the checksum status of the inner part (ICMP error case) must be forced temporarily as unknown to force checksum resolution. When tunneling comes into play, IP checksums status is bit-shifted for future considerations in the processing if, for example, the tunnel header gets decapsulated again, or in the netdev-* implementations that support tunnel offloading. Finally, netdev-* implementations only need to care about packets in partial status: a good checksum does not need touching, a bad checksum has been updated by kept as bad by OVS, an unknown checksum is either an IPv6 or if it was an IPv4, OVS updated it too (keeping it good or bad accordingly). Rename current API for consistency with dp_packet_(inner_)?ip_checksum_. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
2025-06-17 09:20:57 +02:00
if (dp_packet_ip_checksum_partial(p)
&& !(flags & NETDEV_TX_OFFLOAD_IPV4_CKSUM)) {
dp_packet_ip_set_header_csum(p, false);
}
if (dp_packet_l4_checksum_partial(p)) {
if (dp_packet_l4_proto_tcp(p)) {
if (!(flags & NETDEV_TX_OFFLOAD_TCP_CKSUM)) {
packet_tcp_complete_csum(p, false);
}
} else if (dp_packet_l4_proto_udp(p)) {
if (!(flags & NETDEV_TX_OFFLOAD_UDP_CKSUM)) {
packet_udp_complete_csum(p, false);
}
} else {
ovs_assert(dp_packet_l4_proto_sctp(p));
if (!(flags & NETDEV_TX_OFFLOAD_SCTP_CKSUM)) {
packet_sctp_complete_csum(p, false);
}
}
}
return;
}
if (dp_packet_tunnel_geneve(p)
|| dp_packet_tunnel_vxlan(p)) {
/* If the TX interface doesn't support UDP tunnel offload but does
* support inner checksum offload and an outer UDP checksum is
* required, then we can't offload inner checksum either. As that would
* invalidate the outer checksum. */
if (!(flags & NETDEV_TX_OFFLOAD_OUTER_UDP_CKSUM)
&& dp_packet_l4_checksum_partial(p)) {
flags &= ~(NETDEV_TX_OFFLOAD_TCP_CKSUM |
NETDEV_TX_OFFLOAD_UDP_CKSUM |
NETDEV_TX_OFFLOAD_SCTP_CKSUM |
NETDEV_TX_OFFLOAD_IPV4_CKSUM);
}
}
dp-packet: Rework IP checksum offloads. As the packet traverses through OVS, offloading Tx flags must be carefully evaluated and updated which results in a bit of complexity because of a separate "outer" Tx offloading flag coming from DPDK API, and a "normal"/"inner" Tx offloading flag. On the other hand, the DPDK mbuf API specifies 4 status when it comes to IP checksums: - RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN: no information about the RX IP checksum - RTE_MBUF_F_RX_IP_CKSUM_BAD: the IP checksum in the packet is wrong - RTE_MBUF_F_RX_IP_CKSUM_GOOD: the IP checksum in the packet is valid - RTE_MBUF_F_RX_IP_CKSUM_NONE: the IP checksum is not correct in the packet data, but the integrity of the IP header is verified. This patch changes OVS API so that OVS code only tracks the status of the checksum of the "current" L3 header and let the Tx flags aspect to the netdev-* implementations. With this API, the flow extraction can be cleaned up. During packet processing, OVS can simply look for the IP checksum validity (either good, or partial) before changing some IP header, and then mark the checksum as partial. In the conntrack case, when natting packets, the checksum status of the inner part (ICMP error case) must be forced temporarily as unknown to force checksum resolution. When tunneling comes into play, IP checksums status is bit-shifted for future considerations in the processing if, for example, the tunnel header gets decapsulated again, or in the netdev-* implementations that support tunnel offloading. Finally, netdev-* implementations only need to care about packets in partial status: a good checksum does not need touching, a bad checksum has been updated by kept as bad by OVS, an unknown checksum is either an IPv6 or if it was an IPv4, OVS updated it too (keeping it good or bad accordingly). Rename current API for consistency with dp_packet_(inner_)?ip_checksum_. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
2025-06-17 09:20:57 +02:00
if (dp_packet_inner_ip_checksum_partial(p)
&& !(flags & NETDEV_TX_OFFLOAD_IPV4_CKSUM)) {
dp_packet_ip_set_header_csum(p, true);
}
if (dp_packet_inner_l4_checksum_partial(p)) {
if (dp_packet_inner_l4_proto_tcp(p)) {
if (!(flags & NETDEV_TX_OFFLOAD_TCP_CKSUM)) {
packet_tcp_complete_csum(p, true);
}
} else if (dp_packet_inner_l4_proto_udp(p)) {
if (!(flags & NETDEV_TX_OFFLOAD_UDP_CKSUM)) {
packet_udp_complete_csum(p, true);
}
} else {
ovs_assert(dp_packet_inner_l4_proto_sctp(p));
if (!(flags & NETDEV_TX_OFFLOAD_SCTP_CKSUM)) {
packet_sctp_complete_csum(p, true);
}
}
}
dp-packet: Rework IP checksum offloads. As the packet traverses through OVS, offloading Tx flags must be carefully evaluated and updated which results in a bit of complexity because of a separate "outer" Tx offloading flag coming from DPDK API, and a "normal"/"inner" Tx offloading flag. On the other hand, the DPDK mbuf API specifies 4 status when it comes to IP checksums: - RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN: no information about the RX IP checksum - RTE_MBUF_F_RX_IP_CKSUM_BAD: the IP checksum in the packet is wrong - RTE_MBUF_F_RX_IP_CKSUM_GOOD: the IP checksum in the packet is valid - RTE_MBUF_F_RX_IP_CKSUM_NONE: the IP checksum is not correct in the packet data, but the integrity of the IP header is verified. This patch changes OVS API so that OVS code only tracks the status of the checksum of the "current" L3 header and let the Tx flags aspect to the netdev-* implementations. With this API, the flow extraction can be cleaned up. During packet processing, OVS can simply look for the IP checksum validity (either good, or partial) before changing some IP header, and then mark the checksum as partial. In the conntrack case, when natting packets, the checksum status of the inner part (ICMP error case) must be forced temporarily as unknown to force checksum resolution. When tunneling comes into play, IP checksums status is bit-shifted for future considerations in the processing if, for example, the tunnel header gets decapsulated again, or in the netdev-* implementations that support tunnel offloading. Finally, netdev-* implementations only need to care about packets in partial status: a good checksum does not need touching, a bad checksum has been updated by kept as bad by OVS, an unknown checksum is either an IPv6 or if it was an IPv4, OVS updated it too (keeping it good or bad accordingly). Rename current API for consistency with dp_packet_(inner_)?ip_checksum_. Signed-off-by: David Marchand <david.marchand@redhat.com> Signed-off-by: Ilya Maximets <i.maximets@ovn.org>
2025-06-17 09:20:57 +02:00
if (dp_packet_ip_checksum_partial(p)
&& !(flags & NETDEV_TX_OFFLOAD_OUTER_IP_CKSUM)) {
dp_packet_ip_set_header_csum(p, false);
}
if (dp_packet_l4_checksum_partial(p)) {
ovs_assert(dp_packet_l4_proto_udp(p));
if (!(flags & NETDEV_TX_OFFLOAD_OUTER_UDP_CKSUM)) {
packet_udp_complete_csum(p, false);
}
}
}