2
0
mirror of https://github.com/openvswitch/ovs synced 2025-08-31 06:15:47 +00:00

netdev-linux: Prepend the std packet in the TSO packet

Usually TSO packets are close to 50k, 60k bytes long, so to
to copy less bytes when receiving a packet from the kernel
change the approach. Instead of extending the MTU sized
packet received and append with remaining TSO data from
the TSO buffer, allocate a TSO packet with enough headroom
to prepend the std packet data.

Fixes: 29cf9c1b3b ("userspace: Add TCP Segmentation Offload support")
Suggested-by: Ben Pfaff <blp@ovn.org>
Signed-off-by: Flavio Leitner <fbl@sysclose.org>
Signed-off-by: Ben Pfaff <blp@ovn.org>
This commit is contained in:
Flavio Leitner
2020-02-03 18:45:50 -03:00
committed by Ben Pfaff
parent 2297cbe6cc
commit 73858f9dbe
4 changed files with 78 additions and 52 deletions

View File

@@ -243,8 +243,8 @@ dp_packet_copy__(struct dp_packet *b, uint8_t *new_base,
/* Reallocates 'b' so that it has exactly 'new_headroom' and 'new_tailroom'
* bytes of headroom and tailroom, respectively. */
static void
dp_packet_resize__(struct dp_packet *b, size_t new_headroom, size_t new_tailroom)
void
dp_packet_resize(struct dp_packet *b, size_t new_headroom, size_t new_tailroom)
{
void *new_base, *new_data;
size_t new_allocated;
@@ -297,7 +297,7 @@ void
dp_packet_prealloc_tailroom(struct dp_packet *b, size_t size)
{
if (size > dp_packet_tailroom(b)) {
dp_packet_resize__(b, dp_packet_headroom(b), MAX(size, 64));
dp_packet_resize(b, dp_packet_headroom(b), MAX(size, 64));
}
}
@@ -308,7 +308,7 @@ void
dp_packet_prealloc_headroom(struct dp_packet *b, size_t size)
{
if (size > dp_packet_headroom(b)) {
dp_packet_resize__(b, MAX(size, 64), dp_packet_tailroom(b));
dp_packet_resize(b, MAX(size, 64), dp_packet_tailroom(b));
}
}

View File

@@ -152,6 +152,8 @@ struct dp_packet *dp_packet_clone_with_headroom(const struct dp_packet *,
struct dp_packet *dp_packet_clone_data(const void *, size_t);
struct dp_packet *dp_packet_clone_data_with_headroom(const void *, size_t,
size_t headroom);
void dp_packet_resize(struct dp_packet *b, size_t new_headroom,
size_t new_tailroom);
static inline void dp_packet_delete(struct dp_packet *);
static inline void *dp_packet_at(const struct dp_packet *, size_t offset,

View File

@@ -45,7 +45,8 @@ struct netdev_rxq_linux {
struct netdev_rxq up;
bool is_tap;
int fd;
char *aux_bufs[NETDEV_MAX_BURST]; /* Batch of preallocated TSO buffers. */
struct dp_packet *aux_bufs[NETDEV_MAX_BURST]; /* Preallocated TSO
packets. */
};
int netdev_linux_construct(struct netdev *);

View File

@@ -1052,15 +1052,6 @@ static struct netdev_rxq *
netdev_linux_rxq_alloc(void)
{
struct netdev_rxq_linux *rx = xzalloc(sizeof *rx);
if (userspace_tso_enabled()) {
int i;
/* Allocate auxiliay buffers to receive TSO packets. */
for (i = 0; i < NETDEV_MAX_BURST; i++) {
rx->aux_bufs[i] = xmalloc(LINUX_RXQ_TSO_MAX_LEN);
}
}
return &rx->up;
}
@@ -1172,7 +1163,7 @@ netdev_linux_rxq_destruct(struct netdev_rxq *rxq_)
}
for (i = 0; i < NETDEV_MAX_BURST; i++) {
free(rx->aux_bufs[i]);
dp_packet_delete(rx->aux_bufs[i]);
}
}
@@ -1238,13 +1229,18 @@ netdev_linux_batch_rxq_recv_sock(struct netdev_rxq_linux *rx, int mtu,
virtio_net_hdr_size = 0;
}
std_len = VLAN_ETH_HEADER_LEN + mtu + virtio_net_hdr_size;
/* The length here needs to be accounted in the same way when the
* aux_buf is allocated so that it can be prepended to TSO buffer. */
std_len = virtio_net_hdr_size + VLAN_ETH_HEADER_LEN + mtu;
for (i = 0; i < NETDEV_MAX_BURST; i++) {
buffers[i] = dp_packet_new_with_headroom(std_len, DP_NETDEV_HEADROOM);
iovs[i][IOV_PACKET].iov_base = dp_packet_data(buffers[i]);
iovs[i][IOV_PACKET].iov_len = std_len;
iovs[i][IOV_AUXBUF].iov_base = rx->aux_bufs[i];
iovs[i][IOV_AUXBUF].iov_len = LINUX_RXQ_TSO_MAX_LEN;
if (iovlen == IOV_TSO_SIZE) {
iovs[i][IOV_AUXBUF].iov_base = dp_packet_data(rx->aux_bufs[i]);
iovs[i][IOV_AUXBUF].iov_len = dp_packet_tailroom(rx->aux_bufs[i]);
}
mmsgs[i].msg_hdr.msg_name = NULL;
mmsgs[i].msg_hdr.msg_namelen = 0;
mmsgs[i].msg_hdr.msg_iov = iovs[i];
@@ -1268,6 +1264,8 @@ netdev_linux_batch_rxq_recv_sock(struct netdev_rxq_linux *rx, int mtu,
}
for (i = 0; i < retval; i++) {
struct dp_packet *pkt;
if (mmsgs[i].msg_len < ETH_HEADER_LEN) {
struct netdev *netdev_ = netdev_rxq_get_netdev(&rx->up);
struct netdev_linux *netdev = netdev_linux_cast(netdev_);
@@ -1280,29 +1278,29 @@ netdev_linux_batch_rxq_recv_sock(struct netdev_rxq_linux *rx, int mtu,
}
if (mmsgs[i].msg_len > std_len) {
/* Build a single linear TSO packet by expanding the current packet
* to append the data received in the aux_buf. */
size_t extra_len = mmsgs[i].msg_len - std_len;
/* Build a single linear TSO packet by prepending the data from
* std_len buffer to the aux_buf. */
pkt = rx->aux_bufs[i];
dp_packet_set_size(pkt, mmsgs[i].msg_len - std_len);
dp_packet_push(pkt, dp_packet_data(buffers[i]), std_len);
/* The headroom should be the same in buffers[i], pkt and
* DP_NETDEV_HEADROOM. */
dp_packet_resize(pkt, DP_NETDEV_HEADROOM, 0);
dp_packet_delete(buffers[i]);
rx->aux_bufs[i] = NULL;
} else {
dp_packet_set_size(buffers[i], mmsgs[i].msg_len);
pkt = buffers[i];
}
dp_packet_set_size(buffers[i], dp_packet_size(buffers[i])
+ std_len);
dp_packet_prealloc_tailroom(buffers[i], extra_len);
memcpy(dp_packet_tail(buffers[i]), rx->aux_bufs[i], extra_len);
dp_packet_set_size(buffers[i], dp_packet_size(buffers[i])
+ extra_len);
} else {
dp_packet_set_size(buffers[i], dp_packet_size(buffers[i])
+ mmsgs[i].msg_len);
}
if (virtio_net_hdr_size && netdev_linux_parse_vnet_hdr(buffers[i])) {
if (virtio_net_hdr_size && netdev_linux_parse_vnet_hdr(pkt)) {
struct netdev *netdev_ = netdev_rxq_get_netdev(&rx->up);
struct netdev_linux *netdev = netdev_linux_cast(netdev_);
/* Unexpected error situation: the virtio header is not present
* or corrupted. Drop the packet but continue in case next ones
* are correct. */
dp_packet_delete(buffers[i]);
dp_packet_delete(pkt);
netdev->rx_dropped += 1;
VLOG_WARN_RL(&rl, "%s: Dropped packet: Invalid virtio net header",
netdev_get_name(netdev_));
@@ -1325,16 +1323,16 @@ netdev_linux_batch_rxq_recv_sock(struct netdev_rxq_linux *rx, int mtu,
struct eth_header *eth;
bool double_tagged;
eth = dp_packet_data(buffers[i]);
eth = dp_packet_data(pkt);
double_tagged = eth->eth_type == htons(ETH_TYPE_VLAN_8021Q);
eth_push_vlan(buffers[i],
eth_push_vlan(pkt,
auxdata_to_vlan_tpid(aux, double_tagged),
htons(aux->tp_vlan_tci));
break;
}
}
dp_packet_batch_add(batch, buffers[i]);
dp_packet_batch_add(batch, pkt);
}
/* Delete unused buffers. */
@@ -1354,7 +1352,6 @@ static int
netdev_linux_batch_rxq_recv_tap(struct netdev_rxq_linux *rx, int mtu,
struct dp_packet_batch *batch)
{
struct dp_packet *buffer;
int virtio_net_hdr_size;
ssize_t retval;
size_t std_len;
@@ -1372,16 +1369,22 @@ netdev_linux_batch_rxq_recv_tap(struct netdev_rxq_linux *rx, int mtu,
virtio_net_hdr_size = 0;
}
std_len = VLAN_ETH_HEADER_LEN + mtu + virtio_net_hdr_size;
/* The length here needs to be accounted in the same way when the
* aux_buf is allocated so that it can be prepended to TSO buffer. */
std_len = virtio_net_hdr_size + VLAN_ETH_HEADER_LEN + mtu;
for (i = 0; i < NETDEV_MAX_BURST; i++) {
struct dp_packet *buffer;
struct dp_packet *pkt;
struct iovec iov[IOV_TSO_SIZE];
/* Assume Ethernet port. No need to set packet_type. */
buffer = dp_packet_new_with_headroom(std_len, DP_NETDEV_HEADROOM);
iov[IOV_PACKET].iov_base = dp_packet_data(buffer);
iov[IOV_PACKET].iov_len = std_len;
iov[IOV_AUXBUF].iov_base = rx->aux_bufs[i];
iov[IOV_AUXBUF].iov_len = LINUX_RXQ_TSO_MAX_LEN;
if (iovlen == IOV_TSO_SIZE) {
iov[IOV_AUXBUF].iov_base = dp_packet_data(rx->aux_bufs[i]);
iov[IOV_AUXBUF].iov_len = dp_packet_tailroom(rx->aux_bufs[i]);
}
do {
retval = readv(rx->fd, iov, iovlen);
@@ -1393,33 +1396,36 @@ netdev_linux_batch_rxq_recv_tap(struct netdev_rxq_linux *rx, int mtu,
}
if (retval > std_len) {
/* Build a single linear TSO packet by expanding the current packet
* to append the data received in the aux_buf. */
size_t extra_len = retval - std_len;
dp_packet_set_size(buffer, dp_packet_size(buffer) + std_len);
dp_packet_prealloc_tailroom(buffer, extra_len);
memcpy(dp_packet_tail(buffer), rx->aux_bufs[i], extra_len);
dp_packet_set_size(buffer, dp_packet_size(buffer) + extra_len);
/* Build a single linear TSO packet by prepending the data from
* std_len buffer to the aux_buf. */
pkt = rx->aux_bufs[i];
dp_packet_set_size(pkt, retval - std_len);
dp_packet_push(pkt, dp_packet_data(buffer), std_len);
/* The headroom should be the same in buffers[i], pkt and
* DP_NETDEV_HEADROOM. */
dp_packet_resize(pkt, DP_NETDEV_HEADROOM, 0);
dp_packet_delete(buffer);
rx->aux_bufs[i] = NULL;
} else {
dp_packet_set_size(buffer, dp_packet_size(buffer) + retval);
pkt = buffer;
}
if (virtio_net_hdr_size && netdev_linux_parse_vnet_hdr(buffer)) {
if (virtio_net_hdr_size && netdev_linux_parse_vnet_hdr(pkt)) {
struct netdev *netdev_ = netdev_rxq_get_netdev(&rx->up);
struct netdev_linux *netdev = netdev_linux_cast(netdev_);
/* Unexpected error situation: the virtio header is not present
* or corrupted. Drop the packet but continue in case next ones
* are correct. */
dp_packet_delete(buffer);
dp_packet_delete(pkt);
netdev->rx_dropped += 1;
VLOG_WARN_RL(&rl, "%s: Dropped packet: Invalid virtio net header",
netdev_get_name(netdev_));
continue;
}
dp_packet_batch_add(batch, buffer);
dp_packet_batch_add(batch, pkt);
}
if ((i == 0) && (retval < 0)) {
@@ -1442,6 +1448,23 @@ netdev_linux_rxq_recv(struct netdev_rxq *rxq_, struct dp_packet_batch *batch,
mtu = ETH_PAYLOAD_MAX;
}
if (userspace_tso_enabled()) {
/* Allocate TSO packets. The packet has enough headroom to store
* a full non-TSO packet. When a TSO packet is received, the data
* from non-TSO buffer (std_len) is prepended to the TSO packet
* (aux_buf). */
size_t std_len = sizeof(struct virtio_net_hdr) + VLAN_ETH_HEADER_LEN
+ DP_NETDEV_HEADROOM + mtu;
size_t data_len = LINUX_RXQ_TSO_MAX_LEN - std_len;
for (int i = 0; i < NETDEV_MAX_BURST; i++) {
if (rx->aux_bufs[i]) {
continue;
}
rx->aux_bufs[i] = dp_packet_new_with_headroom(data_len, std_len);
}
}
dp_packet_batch_init(batch);
retval = (rx->is_tap
? netdev_linux_batch_rxq_recv_tap(rx, mtu, batch)