From patchwork Tue Oct 13 13:45:39 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Andrew Rybchenko X-Patchwork-Id: 80540 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 21C76A04B7; Tue, 13 Oct 2020 15:57:34 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 5EC3C1BAA0; Tue, 13 Oct 2020 15:46:54 +0200 (CEST) Received: from dispatch1-us1.ppe-hosted.com (dispatch1-us1.ppe-hosted.com [67.231.154.164]) by dpdk.org (Postfix) with ESMTP id ABC711DB8A for ; Tue, 13 Oct 2020 15:46:05 +0200 (CEST) Received: from mx1-us1.ppe-hosted.com (unknown [10.110.50.150]) by dispatch1-us1.ppe-hosted.com (PPE Hosted ESMTP Server) with ESMTP id 73C882007E for ; Tue, 13 Oct 2020 13:46:03 +0000 (UTC) Received: from us4-mdac16-47.at1.mdlocal (unknown [10.110.50.130]) by mx1-us1.ppe-hosted.com (PPE Hosted ESMTP Server) with ESMTP id 7233A800A7 for ; Tue, 13 Oct 2020 13:46:03 +0000 (UTC) X-Virus-Scanned: Proofpoint Essentials engine Received: from mx1-us1.ppe-hosted.com (unknown [10.110.49.106]) by mx1-us1.ppe-hosted.com (PPE Hosted ESMTP Server) with ESMTPS id 11D0710004F for ; Tue, 13 Oct 2020 13:46:03 +0000 (UTC) Received: from webmail.solarflare.com (uk.solarflare.com [193.34.186.16]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (No client certificate requested) by mx1-us1.ppe-hosted.com (PPE Hosted ESMTP Server) with ESMTPS id CF29BB40068 for ; Tue, 13 Oct 2020 13:46:02 +0000 (UTC) Received: from ukex01.SolarFlarecom.com (10.17.10.4) by ukex01.SolarFlarecom.com (10.17.10.4) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Tue, 13 Oct 2020 14:45:56 +0100 Received: from opal.uk.solarflarecom.com (10.17.10.1) by ukex01.SolarFlarecom.com (10.17.10.4) with Microsoft SMTP Server id 15.0.1497.2 via Frontend Transport; Tue, 13 Oct 2020 14:45:56 +0100 Received: from ukv-loginhost.uk.solarflarecom.com (ukv-loginhost.uk.solarflarecom.com [10.17.10.39]) by opal.uk.solarflarecom.com (8.13.8/8.13.8) with ESMTP id 09DDjusn006083 for ; Tue, 13 Oct 2020 14:45:56 +0100 Received: from ukv-loginhost.uk.solarflarecom.com (localhost [127.0.0.1]) by ukv-loginhost.uk.solarflarecom.com (Postfix) with ESMTP id 685E11613AB for ; Tue, 13 Oct 2020 14:45:56 +0100 (BST) From: Andrew Rybchenko To: Date: Tue, 13 Oct 2020 14:45:39 +0100 Message-ID: <1602596753-32282-23-git-send-email-arybchenko@solarflare.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1602596753-32282-1-git-send-email-arybchenko@solarflare.com> References: <1602596753-32282-1-git-send-email-arybchenko@solarflare.com> MIME-Version: 1.0 X-TM-AS-Product-Ver: SMEX-12.5.0.1300-8.6.1012-25722.003 X-TM-AS-Result: No-14.401400-8.000000-10 X-TMASE-MatchedRID: fVkdlGqLoLY6WGg4O5YW8EdAWPMBu8kQcUeqyCu6iFkWhfQgHJ/xHsiT Wug2C4DNNNN0dk/lb45w5T4Iaj538mJZXQNDzktSZobkeVFowYUvKK/+8XMSRLIFKaUww8967gW uEPosBb9GOVSl8PHlYcYGS5HBZwU4uGH8gro7lJEyIyttzvQ993607foZgOWyqPGqHIPGZiOhwz E4cmZ285t2TBH1ptM49W4OM0HGdzu/grWuik/IvYdlc1JaOB1TXNY8BwboMUCMUViaYYbK3Iu6u 5Cfj8Oaw01OnNToscV89C1b9Tu9HYfGISHW1BqpGjzBgnFZvQ6khIH8pCh4lq//6X0io9HxvnwP kToqUsI7Wg6fPaF64IAy6p60ZV62fJ5/bZ6npdg7AFczfjr/7J5G7WHar402IV7TlCx3v+iP+r1 n3TUiEaQyBeqYzcCnFt9n6eJi/xQ= X-TM-AS-User-Approved-Sender: Yes X-TM-AS-User-Blocked-Sender: No X-TMASE-Result: 10--14.401400-8.000000 X-TMASE-Version: SMEX-12.5.0.1300-8.6.1012-25722.003 X-MDID: 1602596763-vzIdqLc6S1Jk X-PPE-DISP: 1602596763;vzIdqLc6S1Jk Subject: [dpdk-dev] [PATCH 22/36] net/sfc: support tunnels for EF100 native Tx datapath X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Add support for outer IPv4/UDP and inner IPv4/UDP/TCP checksum offloads. Use partial checksum offload for inner TCP/UDP offload. Signed-off-by: Andrew Rybchenko --- doc/guides/nics/sfc_efx.rst | 2 +- drivers/net/sfc/sfc_ef100_tx.c | 93 ++++++++++++++++++++++++++++++++-- 2 files changed, 90 insertions(+), 5 deletions(-) diff --git a/doc/guides/nics/sfc_efx.rst b/doc/guides/nics/sfc_efx.rst index 0e32d0c6d9..f3135fdd70 100644 --- a/doc/guides/nics/sfc_efx.rst +++ b/doc/guides/nics/sfc_efx.rst @@ -329,7 +329,7 @@ boolean parameters value. is even more faster then **ef10** but does not support multi-segment mbufs, disallows multiple mempools and neglects mbuf reference counters. **ef100** chooses EF100 native datapath which supports multi-segment - mbufs, IPv4 and TCP/UDP checksum offloads. + mbufs, inner/outer IPv4 and TCP/UDP checksum offloads. - ``perf_profile`` [auto|throughput|low-latency] (default **throughput**) diff --git a/drivers/net/sfc/sfc_ef100_tx.c b/drivers/net/sfc/sfc_ef100_tx.c index 0dba5c8eee..20d4d1cf9c 100644 --- a/drivers/net/sfc/sfc_ef100_tx.c +++ b/drivers/net/sfc/sfc_ef100_tx.c @@ -11,6 +11,7 @@ #include #include +#include #include "efx.h" #include "efx_types.h" @@ -96,8 +97,21 @@ sfc_ef100_tx_prepare_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, for (i = 0; i < nb_pkts; i++) { struct rte_mbuf *m = tx_pkts[i]; unsigned int max_nb_header_segs = 0; + bool calc_phdr_cksum = false; int ret; + /* + * Partial checksum offload is used in the case of + * inner TCP/UDP checksum offload. It requires + * pseudo-header checksum which is calculated below, + * but requires contiguous packet headers. + */ + if ((m->ol_flags & PKT_TX_TUNNEL_MASK) && + (m->ol_flags & PKT_TX_L4_MASK)) { + calc_phdr_cksum = true; + max_nb_header_segs = 1; + } + ret = sfc_dp_tx_prepare_pkt(m, max_nb_header_segs, 0, 0, txq->max_fill_level, 0, 0); if (unlikely(ret != 0)) { @@ -109,6 +123,19 @@ sfc_ef100_tx_prepare_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, rte_errno = EINVAL; break; } + + if (calc_phdr_cksum) { + /* + * Full checksum offload does IPv4 header checksum + * and does not require any assistance. + */ + ret = rte_net_intel_cksum_flags_prepare(m, + m->ol_flags & ~PKT_TX_IP_CKSUM); + if (unlikely(ret != 0)) { + rte_errno = -ret; + break; + } + } } return i; @@ -215,19 +242,75 @@ sfc_ef100_tx_reap(struct sfc_ef100_txq *txq) sfc_ef100_tx_reap_num_descs(txq, sfc_ef100_tx_process_events(txq)); } +static uint8_t +sfc_ef100_tx_qdesc_cso_inner_l3(uint64_t tx_tunnel) +{ + uint8_t inner_l3; + + switch (tx_tunnel) { + case PKT_TX_TUNNEL_VXLAN: + inner_l3 = ESE_GZ_TX_DESC_CS_INNER_L3_VXLAN; + break; + case PKT_TX_TUNNEL_GENEVE: + inner_l3 = ESE_GZ_TX_DESC_CS_INNER_L3_GENEVE; + break; + default: + inner_l3 = ESE_GZ_TX_DESC_CS_INNER_L3_OFF; + break; + } + return inner_l3; +} + static void sfc_ef100_tx_qdesc_send_create(const struct rte_mbuf *m, efx_oword_t *tx_desc) { bool outer_l3; bool outer_l4; + uint8_t inner_l3; + uint8_t partial_en; + uint16_t part_cksum_w; + uint16_t l4_offset_w; + + if ((m->ol_flags & PKT_TX_TUNNEL_MASK) == 0) { + outer_l3 = (m->ol_flags & PKT_TX_IP_CKSUM); + outer_l4 = (m->ol_flags & PKT_TX_L4_MASK); + inner_l3 = ESE_GZ_TX_DESC_CS_INNER_L3_OFF; + partial_en = ESE_GZ_TX_DESC_CSO_PARTIAL_EN_OFF; + part_cksum_w = 0; + l4_offset_w = 0; + } else { + outer_l3 = (m->ol_flags & PKT_TX_OUTER_IP_CKSUM); + outer_l4 = (m->ol_flags & PKT_TX_OUTER_UDP_CKSUM); + inner_l3 = sfc_ef100_tx_qdesc_cso_inner_l3(m->ol_flags & + PKT_TX_TUNNEL_MASK); + + switch (m->ol_flags & PKT_TX_L4_MASK) { + case PKT_TX_TCP_CKSUM: + partial_en = ESE_GZ_TX_DESC_CSO_PARTIAL_EN_TCP; + part_cksum_w = offsetof(struct rte_tcp_hdr, cksum) >> 1; + break; + case PKT_TX_UDP_CKSUM: + partial_en = ESE_GZ_TX_DESC_CSO_PARTIAL_EN_UDP; + part_cksum_w = offsetof(struct rte_udp_hdr, + dgram_cksum) >> 1; + break; + default: + partial_en = ESE_GZ_TX_DESC_CSO_PARTIAL_EN_OFF; + part_cksum_w = 0; + break; + } + l4_offset_w = (m->outer_l2_len + m->outer_l3_len + + m->l2_len + m->l3_len) >> 1; + } - outer_l3 = (m->ol_flags & PKT_TX_IP_CKSUM); - outer_l4 = (m->ol_flags & PKT_TX_L4_MASK); - - EFX_POPULATE_OWORD_6(*tx_desc, + EFX_POPULATE_OWORD_10(*tx_desc, ESF_GZ_TX_SEND_ADDR, rte_mbuf_data_iova(m), ESF_GZ_TX_SEND_LEN, rte_pktmbuf_data_len(m), ESF_GZ_TX_SEND_NUM_SEGS, m->nb_segs, + ESF_GZ_TX_SEND_CSO_PARTIAL_START_W, l4_offset_w, + ESF_GZ_TX_SEND_CSO_PARTIAL_CSUM_W, part_cksum_w, + ESF_GZ_TX_SEND_CSO_PARTIAL_EN, partial_en, + ESF_GZ_TX_SEND_CSO_INNER_L3, inner_l3, ESF_GZ_TX_SEND_CSO_OUTER_L3, outer_l3, ESF_GZ_TX_SEND_CSO_OUTER_L4, outer_l4, ESF_GZ_TX_DESC_TYPE, ESE_GZ_TX_DESC_TYPE_SEND); @@ -603,6 +686,8 @@ struct sfc_dp_tx sfc_ef100_tx = { .features = SFC_DP_TX_FEAT_MULTI_PROCESS, .dev_offload_capa = 0, .queue_offload_capa = DEV_TX_OFFLOAD_IPV4_CKSUM | + DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM | + DEV_TX_OFFLOAD_OUTER_UDP_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | DEV_TX_OFFLOAD_MULTI_SEGS,