From patchwork Thu Feb 28 19:20:59 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Ananyev, Konstantin" X-Patchwork-Id: 50683 X-Patchwork-Delegate: gakhil@marvell.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id CD1324CBD; Thu, 28 Feb 2019 20:21:21 +0100 (CET) Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by dpdk.org (Postfix) with ESMTP id B34384C8F for ; Thu, 28 Feb 2019 20:21:18 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga006.fm.intel.com ([10.253.24.20]) by orsmga102.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 28 Feb 2019 11:21:17 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.58,424,1544515200"; d="scan'208";a="322983973" Received: from sivswdev08.ir.intel.com (HELO localhost.localdomain) ([10.237.217.47]) by fmsmga006.fm.intel.com with ESMTP; 28 Feb 2019 11:21:16 -0800 From: Konstantin Ananyev To: dev@dpdk.org Cc: akhil.goyal@nxp.com, olivier.matz@6wind.com, Konstantin Ananyev Date: Thu, 28 Feb 2019 19:20:59 +0000 Message-Id: <1551381661-21078-5-git-send-email-konstantin.ananyev@intel.com> X-Mailer: git-send-email 1.7.0.7 In-Reply-To: <1551381661-21078-1-git-send-email-konstantin.ananyev@intel.com> References: <1551381661-21078-1-git-send-email-konstantin.ananyev@intel.com> Subject: [dpdk-dev] [PATCH 4/6] ipsec: change the way unprocessed mbufs are accounted X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" As was pointed in one of previous reviews - we can avoid updating contents of mbuf array for successfully processed packets. Instead store indexes of failed packets, to move them beyond the good ones later. Signed-off-by: Konstantin Ananyev --- lib/librte_ipsec/sa.c | 166 +++++++++++++++++++++++------------------- 1 file changed, 93 insertions(+), 73 deletions(-) diff --git a/lib/librte_ipsec/sa.c b/lib/librte_ipsec/sa.c index 74ed3b41e..e4e3d7fc3 100644 --- a/lib/librte_ipsec/sa.c +++ b/lib/librte_ipsec/sa.c @@ -411,14 +411,31 @@ rte_ipsec_sa_init(struct rte_ipsec_sa *sa, const struct rte_ipsec_sa_prm *prm, return sz; } -static inline void -mbuf_bulk_copy(struct rte_mbuf *dst[], struct rte_mbuf * const src[], - uint32_t num) +/* + * Move bad (unprocessed) mbufs beyond the good (processed) ones. + * dr[] contains the indexes of bad mbufs insinde the mb[]. + */ +static void +mbuf_bad_move(struct rte_mbuf *mb[], const uint32_t dr[], uint32_t num, + uint32_t drn) { - uint32_t i; + uint32_t i, j, k; + struct rte_mbuf *drb[drn]; + + j = 0; + k = 0; - for (i = 0; i != num; i++) - dst[i] = src[i]; + /* copy bad ones into a temp place */ + for (i = 0; i != num; i++) { + if (j != drn && i == dr[j]) + drb[j++] = mb[i]; + else + mb[k++] = mb[i]; + } + + /* copy bad ones after the good ones */ + for (i = 0; i != drn; i++) + mb[k + i] = drb[i]; } /* @@ -603,7 +620,7 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -627,17 +644,17 @@ outb_tun_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, 0, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + mbuf_bad_move(mb, dr, n, n - k); return k; } @@ -738,7 +755,7 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -766,17 +783,17 @@ outb_trs_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], outb_pkt_xprepare(sa, sqc, &icv); lksd_none_cop_prepare(cop[k], cs, mb[i]); esp_outb_cop_prepare(cop[k], sa, iv, &icv, l2 + l3, rc); - mb[k++] = mb[i]; + k++; /* failure, put packet into the death-row */ } else { - dr[i - k] = mb[i]; + dr[i - k] = i; rte_errno = -rc; } } /* copy not prepared mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + mbuf_bad_move(mb, dr, n, n - k); return k; } @@ -924,7 +941,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], struct rte_cryptodev_sym_session *cs; struct replay_sqn *rsn; union sym_op_data icv; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; cs = ss->crypto.ses; @@ -941,10 +958,9 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], hl, rc); } - if (rc == 0) - mb[k++] = mb[i]; - else { - dr[i - k] = mb[i]; + k += (rc == 0); + if (rc != 0) { + dr[i - k] = i; rte_errno = -rc; } } @@ -953,7 +969,7 @@ inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], /* copy not prepared mbufs beyond good ones */ if (k != num && k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + mbuf_bad_move(mb, dr, num, num - k); return k; } @@ -1105,7 +1121,7 @@ esp_inb_trs_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb, */ static inline uint16_t esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], - struct rte_mbuf *mb[], struct rte_mbuf *dr[], uint16_t num) + uint32_t dr[], uint16_t num) { uint32_t i, k; struct replay_sqn *rsn; @@ -1115,9 +1131,9 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[], k = 0; for (i = 0; i != num; i++) { if (esn_inb_update_sqn(rsn, sa, sqn[i]) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } rsn_update_finish(sa, rsn); @@ -1131,10 +1147,10 @@ static uint16_t inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; struct rte_ipsec_sa *sa; uint32_t sqn[num]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1144,23 +1160,27 @@ inb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_tun_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + mbuf_bad_move(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + mbuf_bad_move(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1170,10 +1190,10 @@ static uint16_t inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { - uint32_t i, k; + uint32_t i, k, n; uint32_t sqn[num]; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1183,23 +1203,27 @@ inb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], for (i = 0; i != num; i++) { /* good packet */ if (esp_inb_trs_single_pkt_process(sa, mb[i], sqn + k) == 0) - mb[k++] = mb[i]; + k++; /* bad packet, will drop from furhter processing */ else - dr[i - k] = mb[i]; + dr[i - k] = i; } - /* update seq # and replay winow */ - k = esp_inb_rsn_update(sa, sqn, mb, dr + i - k, k); - /* handle unprocessed mbufs */ - if (k != num) { + if (k != num && k != 0) + mbuf_bad_move(mb, dr, num, num - k); + + /* update SQN and replay winow */ + n = esp_inb_rsn_update(sa, sqn, dr, k); + + /* handle mbufs with wrong SQN */ + if (n != k && n != 0) + mbuf_bad_move(mb, dr, k, k - n); + + if (n != num) rte_errno = EBADMSG; - if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); - } - return k; + return n; } /* @@ -1215,7 +1239,7 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint32_t i, k, icv_len, *icv; struct rte_mbuf *ml; struct rte_ipsec_sa *sa; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1228,16 +1252,16 @@ outb_sqh_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], icv = rte_pktmbuf_mtod_offset(ml, void *, ml->data_len - icv_len); remove_sqh(icv, icv_len); - mb[k++] = mb[i]; + k++; } else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + mbuf_bad_move(mb, dr, num, num - k); } return k; @@ -1257,23 +1281,23 @@ pkt_flag_process(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[], uint16_t num) { uint32_t i, k; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; RTE_SET_USED(ss); k = 0; for (i = 0; i != num; i++) { if ((mb[i]->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED) == 0) - mb[k++] = mb[i]; + k++; else - dr[i - k] = mb[i]; + dr[i - k] = i; } /* handle unprocessed mbufs */ if (k != num) { rte_errno = EBADMSG; if (k != 0) - mbuf_bulk_copy(mb + k, dr, num - k); + mbuf_bad_move(mb, dr, num, num - k); } return k; @@ -1314,7 +1338,7 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1332,22 +1356,20 @@ inline_outb_tun_pkt_process(const struct rte_ipsec_session *ss, /* try to update the packet itself */ rc = esp_outb_tun_pkt_prepare(sa, sqc, iv, mb[i], &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + mbuf_bad_move(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; } @@ -1366,7 +1388,7 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, struct rte_ipsec_sa *sa; union sym_op_data icv; uint64_t iv[IPSEC_MAX_IV_QWORD]; - struct rte_mbuf *dr[num]; + uint32_t dr[num]; sa = ss->sa; @@ -1388,22 +1410,20 @@ inline_outb_trs_pkt_process(const struct rte_ipsec_session *ss, rc = esp_outb_trs_pkt_prepare(sa, sqc, iv, mb[i], l2, l3, &icv); - /* success, update mbuf fields */ - if (rc >= 0) - mb[k++] = mb[i]; + k += (rc >= 0); + /* failure, put packet into the death-row */ - else { - dr[i - k] = mb[i]; + if (rc < 0) { + dr[i - k] = i; rte_errno = -rc; } } - inline_outb_mbuf_prepare(ss, mb, k); - /* copy not processed mbufs beyond good ones */ if (k != n && k != 0) - mbuf_bulk_copy(mb + k, dr, n - k); + mbuf_bad_move(mb, dr, n, n - k); + inline_outb_mbuf_prepare(ss, mb, k); return k; }