get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/51983/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 51983,
    "url": "http://patches.dpdk.org/api/patches/51983/?format=api",
    "web_url": "http://patches.dpdk.org/project/dpdk/patch/20190401125656.7636-8-konstantin.ananyev@intel.com/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<20190401125656.7636-8-konstantin.ananyev@intel.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/20190401125656.7636-8-konstantin.ananyev@intel.com",
    "date": "2019-04-01T12:56:54",
    "name": "[v5,7/9] ipsec: reorder packet process for esp inbound",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "5c8abdf7b1083fb85a9b1f4abb81d01ef44b094e",
    "submitter": {
        "id": 33,
        "url": "http://patches.dpdk.org/api/people/33/?format=api",
        "name": "Ananyev, Konstantin",
        "email": "konstantin.ananyev@intel.com"
    },
    "delegate": {
        "id": 6690,
        "url": "http://patches.dpdk.org/api/users/6690/?format=api",
        "username": "akhil",
        "first_name": "akhil",
        "last_name": "goyal",
        "email": "gakhil@marvell.com"
    },
    "mbox": "http://patches.dpdk.org/project/dpdk/patch/20190401125656.7636-8-konstantin.ananyev@intel.com/mbox/",
    "series": [
        {
            "id": 4024,
            "url": "http://patches.dpdk.org/api/series/4024/?format=api",
            "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=4024",
            "date": "2019-04-01T12:56:47",
            "name": "Few small improvements for ipsec library",
            "version": 5,
            "mbox": "http://patches.dpdk.org/series/4024/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/51983/comments/",
    "check": "success",
    "checks": "http://patches.dpdk.org/api/patches/51983/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@dpdk.org",
        "Delivered-To": "patchwork@dpdk.org",
        "Received": [
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 07FA95681;\n\tMon,  1 Apr 2019 14:57:34 +0200 (CEST)",
            "from mga01.intel.com (mga01.intel.com [192.55.52.88])\n\tby dpdk.org (Postfix) with ESMTP id 7CBED4D3A\n\tfor <dev@dpdk.org>; Mon,  1 Apr 2019 14:57:29 +0200 (CEST)",
            "from fmsmga008.fm.intel.com ([10.253.24.58])\n\tby fmsmga101.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384;\n\t01 Apr 2019 05:57:29 -0700",
            "from sivswdev08.ir.intel.com ([10.237.217.47])\n\tby fmsmga008.fm.intel.com with ESMTP; 01 Apr 2019 05:57:27 -0700"
        ],
        "X-Amp-Result": "SKIPPED(no attachment in message)",
        "X-Amp-File-Uploaded": "False",
        "X-ExtLoop1": "1",
        "X-IronPort-AV": "E=Sophos;i=\"5.60,296,1549958400\"; d=\"scan'208\";a=\"136564345\"",
        "From": "Konstantin Ananyev <konstantin.ananyev@intel.com>",
        "To": "dev@dpdk.org",
        "Cc": "akhil.goyal@nxp.com, olivier.matz@6wind.com,\n\tKonstantin Ananyev <konstantin.ananyev@intel.com>",
        "Date": "Mon,  1 Apr 2019 13:56:54 +0100",
        "Message-Id": "<20190401125656.7636-8-konstantin.ananyev@intel.com>",
        "X-Mailer": "git-send-email 2.18.0",
        "In-Reply-To": "<20190401125656.7636-1-konstantin.ananyev@intel.com>",
        "References": "<20190329102726.27716-1-konstantin.ananyev@intel.com>\n\t<20190401125656.7636-1-konstantin.ananyev@intel.com>",
        "Subject": "[dpdk-dev] [PATCH v5 7/9] ipsec: reorder packet process for esp\n\tinbound",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n\t<mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n\t<mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "Change the order of operations for esp inbound post-process:\n- read mbuf metadata and esp tail first for all packets in the burst\n  first to minimize stalls due to load latency.\n- move code that is common for both transport and tunnel modes into\n  separate functions to reduce code duplication.\n- add extra check for packet consitency\n\nSigned-off-by: Konstantin Ananyev <konstantin.ananyev@intel.com>\n---\n lib/librte_ipsec/esp_inb.c   | 351 ++++++++++++++++++++++-------------\n lib/librte_ipsec/ipsec_sqn.h |   4 -\n 2 files changed, 227 insertions(+), 128 deletions(-)",
    "diff": "diff --git a/lib/librte_ipsec/esp_inb.c b/lib/librte_ipsec/esp_inb.c\nindex 8d1171556..138ed0450 100644\n--- a/lib/librte_ipsec/esp_inb.c\n+++ b/lib/librte_ipsec/esp_inb.c\n@@ -15,8 +15,11 @@\n #include \"misc.h\"\n #include \"pad.h\"\n \n+typedef uint16_t (*esp_inb_process_t)(const struct rte_ipsec_sa *sa,\n+\tstruct rte_mbuf *mb[], uint32_t sqn[], uint32_t dr[], uint16_t num);\n+\n /*\n- * setup crypto op and crypto sym op for ESP inbound tunnel packet.\n+ * setup crypto op and crypto sym op for ESP inbound packet.\n  */\n static inline void\n inb_cop_prepare(struct rte_crypto_op *cop,\n@@ -216,111 +219,239 @@ esp_inb_pkt_prepare(const struct rte_ipsec_session *ss, struct rte_mbuf *mb[],\n }\n \n /*\n- * process ESP inbound tunnel packet.\n+ * Start with processing inbound packet.\n+ * This is common part for both tunnel and transport mode.\n+ * Extract information that will be needed later from mbuf metadata and\n+ * actual packet data:\n+ * - mbuf for packet's last segment\n+ * - length of the L2/L3 headers\n+ * - esp tail structure\n  */\n-static inline int\n-inb_tun_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb,\n-\tuint32_t *sqn)\n+static inline void\n+process_step1(struct rte_mbuf *mb, uint32_t tlen, struct rte_mbuf **ml,\n+\tstruct esp_tail *espt, uint32_t *hlen)\n {\n-\tuint32_t hlen, icv_len, tlen;\n-\tstruct esp_hdr *esph;\n-\tstruct esp_tail *espt;\n-\tstruct rte_mbuf *ml;\n-\tchar *pd;\n+\tconst struct esp_tail *pt;\n \n-\tif (mb->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED)\n-\t\treturn -EBADMSG;\n+\tml[0] = rte_pktmbuf_lastseg(mb);\n+\thlen[0] = mb->l2_len + mb->l3_len;\n+\tpt = rte_pktmbuf_mtod_offset(ml[0], const struct esp_tail *,\n+\t\tml[0]->data_len - tlen);\n+\tespt[0] = pt[0];\n+}\n \n-\ticv_len = sa->icv_len;\n+/*\n+ * packet checks for transport mode:\n+ * - no reported IPsec related failures in ol_flags\n+ * - tail length is valid\n+ * - padding bytes are valid\n+ */\n+static inline int32_t\n+trs_process_check(const struct rte_mbuf *mb, const struct rte_mbuf *ml,\n+\tstruct esp_tail espt, uint32_t hlen, uint32_t tlen)\n+{\n+\tconst uint8_t *pd;\n+\tint32_t ofs;\n \n-\tml = rte_pktmbuf_lastseg(mb);\n-\tespt = rte_pktmbuf_mtod_offset(ml, struct esp_tail *,\n-\t\tml->data_len - icv_len - sizeof(*espt));\n+\tofs = ml->data_len - tlen;\n+\tpd = rte_pktmbuf_mtod_offset(ml, const uint8_t *, ofs);\n \n-\t/*\n-\t * check padding and next proto.\n-\t * return an error if something is wrong.\n-\t */\n-\tpd = (char *)espt - espt->pad_len;\n-\tif (espt->next_proto != sa->proto ||\n-\t\t\tmemcmp(pd, esp_pad_bytes, espt->pad_len))\n-\t\treturn -EINVAL;\n+\treturn ((mb->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED) != 0 ||\n+\t\tofs < 0 || tlen + hlen > mb->pkt_len ||\n+\t\t(espt.pad_len != 0 && memcmp(pd, esp_pad_bytes, espt.pad_len)));\n+}\n+\n+/*\n+ * packet checks for tunnel mode:\n+ * - same as for trasnport mode\n+ * - esp tail next proto contains expected for that SA value\n+ */\n+static inline int32_t\n+tun_process_check(const struct rte_mbuf *mb, struct rte_mbuf *ml,\n+\tstruct esp_tail espt, uint32_t hlen, const uint32_t tlen, uint8_t proto)\n+{\n+\treturn (trs_process_check(mb, ml, espt, hlen, tlen) ||\n+\t\tespt.next_proto != proto);\n+}\n+\n+/*\n+ * step two for tunnel mode:\n+ * - read SQN value (for future use)\n+ * - cut of ICV, ESP tail and padding bytes\n+ * - cut of ESP header and IV, also if needed - L2/L3 headers\n+ *   (controlled by *adj* value)\n+ */\n+static inline void *\n+tun_process_step2(struct rte_mbuf *mb, struct rte_mbuf *ml, uint32_t hlen,\n+\tuint32_t adj, uint32_t tlen, uint32_t *sqn)\n+{\n+\tconst struct esp_hdr *ph;\n+\n+\t/* read SQN value */\n+\tph = rte_pktmbuf_mtod_offset(mb, const struct esp_hdr *, hlen);\n+\tsqn[0] = ph->seq;\n \n \t/* cut of ICV, ESP tail and padding bytes */\n-\ttlen = icv_len + sizeof(*espt) + espt->pad_len;\n \tml->data_len -= tlen;\n \tmb->pkt_len -= tlen;\n \n \t/* cut of L2/L3 headers, ESP header and IV */\n-\thlen = mb->l2_len + mb->l3_len;\n-\tesph = rte_pktmbuf_mtod_offset(mb, struct esp_hdr *, hlen);\n-\trte_pktmbuf_adj(mb, hlen + sa->ctp.cipher.offset);\n+\treturn rte_pktmbuf_adj(mb, adj);\n+}\n+\n+/*\n+ * step two for transport mode:\n+ * - read SQN value (for future use)\n+ * - cut of ICV, ESP tail and padding bytes\n+ * - cut of ESP header and IV\n+ * - move L2/L3 header to fill the gap after ESP header removal\n+ */\n+static inline void *\n+trs_process_step2(struct rte_mbuf *mb, struct rte_mbuf *ml, uint32_t hlen,\n+\tuint32_t adj, uint32_t tlen, uint32_t *sqn)\n+{\n+\tchar *np, *op;\n+\n+\t/* get start of the packet before modifications */\n+\top = rte_pktmbuf_mtod(mb, char *);\n+\n+\t/* cut off ESP header and IV */\n+\tnp = tun_process_step2(mb, ml, hlen, adj, tlen, sqn);\n+\n+\t/* move header bytes to fill the gap after ESP header removal */\n+\tremove_esph(np, op, hlen);\n+\treturn np;\n+}\n \n-\t/* retrieve SQN for later check */\n-\t*sqn = rte_be_to_cpu_32(esph->seq);\n+/*\n+ * step three for transport mode:\n+ * update mbuf metadata:\n+ * - packet_type\n+ * - ol_flags\n+ */\n+static inline void\n+trs_process_step3(struct rte_mbuf *mb)\n+{\n+\t/* reset mbuf packet type */\n+\tmb->packet_type &= (RTE_PTYPE_L2_MASK | RTE_PTYPE_L3_MASK);\n \n+\t/* clear the PKT_RX_SEC_OFFLOAD flag if set */\n+\tmb->ol_flags &= ~PKT_RX_SEC_OFFLOAD;\n+}\n+\n+/*\n+ * step three for tunnel mode:\n+ * update mbuf metadata:\n+ * - packet_type\n+ * - ol_flags\n+ * - tx_offload\n+ */\n+static inline void\n+tun_process_step3(struct rte_mbuf *mb, uint64_t txof_msk, uint64_t txof_val)\n+{\n \t/* reset mbuf metatdata: L2/L3 len, packet type */\n \tmb->packet_type = RTE_PTYPE_UNKNOWN;\n-\tmb->tx_offload = (mb->tx_offload & sa->tx_offload.msk) |\n-\t\tsa->tx_offload.val;\n+\tmb->tx_offload = (mb->tx_offload & txof_msk) | txof_val;\n \n \t/* clear the PKT_RX_SEC_OFFLOAD flag if set */\n-\tmb->ol_flags &= ~(mb->ol_flags & PKT_RX_SEC_OFFLOAD);\n-\treturn 0;\n+\tmb->ol_flags &= ~PKT_RX_SEC_OFFLOAD;\n }\n \n+\n /*\n- * process ESP inbound transport packet.\n+ * *process* function for tunnel packets\n  */\n-static inline int\n-inb_trs_single_pkt_process(struct rte_ipsec_sa *sa, struct rte_mbuf *mb,\n-\tuint32_t *sqn)\n+static inline uint16_t\n+tun_process(const struct rte_ipsec_sa *sa, struct rte_mbuf *mb[],\n+\tuint32_t sqn[], uint32_t dr[], uint16_t num)\n {\n-\tuint32_t hlen, icv_len, l2len, l3len, tlen;\n-\tstruct esp_hdr *esph;\n-\tstruct esp_tail *espt;\n-\tstruct rte_mbuf *ml;\n-\tchar *np, *op, *pd;\n+\tuint32_t adj, i, k, tl;\n+\tuint32_t hl[num];\n+\tstruct esp_tail espt[num];\n+\tstruct rte_mbuf *ml[num];\n \n-\tif (mb->ol_flags & PKT_RX_SEC_OFFLOAD_FAILED)\n-\t\treturn -EBADMSG;\n+\tconst uint32_t tlen = sa->icv_len + sizeof(espt[0]);\n+\tconst uint32_t cofs = sa->ctp.cipher.offset;\n \n-\ticv_len = sa->icv_len;\n+\t/*\n+\t * to minimize stalls due to load latency,\n+\t * read mbufs metadata and esp tail first.\n+\t */\n+\tfor (i = 0; i != num; i++)\n+\t\tprocess_step1(mb[i], tlen, &ml[i], &espt[i], &hl[i]);\n \n-\tml = rte_pktmbuf_lastseg(mb);\n-\tespt = rte_pktmbuf_mtod_offset(ml, struct esp_tail *,\n-\t\tml->data_len - icv_len - sizeof(*espt));\n+\tk = 0;\n+\tfor (i = 0; i != num; i++) {\n \n-\t/* check padding, return an error if something is wrong. */\n-\tpd = (char *)espt - espt->pad_len;\n-\tif (memcmp(pd, esp_pad_bytes, espt->pad_len))\n-\t\treturn -EINVAL;\n+\t\tadj = hl[i] + cofs;\n+\t\ttl = tlen + espt[i].pad_len;\n \n-\t/* cut of ICV, ESP tail and padding bytes */\n-\ttlen = icv_len + sizeof(*espt) + espt->pad_len;\n-\tml->data_len -= tlen;\n-\tmb->pkt_len -= tlen;\n+\t\t/* check that packet is valid */\n+\t\tif (tun_process_check(mb[i], ml[i], espt[i], adj, tl,\n+\t\t\t\t\tsa->proto) == 0) {\n \n-\t/* retrieve SQN for later check */\n-\tl2len = mb->l2_len;\n-\tl3len = mb->l3_len;\n-\thlen = l2len + l3len;\n-\top = rte_pktmbuf_mtod(mb, char *);\n-\tesph = (struct esp_hdr *)(op + hlen);\n-\t*sqn = rte_be_to_cpu_32(esph->seq);\n+\t\t\t/* modify packet's layout */\n+\t\t\ttun_process_step2(mb[i], ml[i], hl[i], adj,\n+\t\t\t\ttl, sqn + k);\n+\t\t\t/* update mbuf's metadata */\n+\t\t\ttun_process_step3(mb[i], sa->tx_offload.msk,\n+\t\t\t\tsa->tx_offload.val);\n+\t\t\tk++;\n+\t\t} else\n+\t\t\tdr[i - k] = i;\n+\t}\n \n-\t/* cut off ESP header and IV, update L3 header */\n-\tnp = rte_pktmbuf_adj(mb, sa->ctp.cipher.offset);\n-\tremove_esph(np, op, hlen);\n-\tupdate_trs_l3hdr(sa, np + l2len, mb->pkt_len, l2len, l3len,\n-\t\t\tespt->next_proto);\n+\treturn k;\n+}\n \n-\t/* reset mbuf packet type */\n-\tmb->packet_type &= (RTE_PTYPE_L2_MASK | RTE_PTYPE_L3_MASK);\n \n-\t/* clear the PKT_RX_SEC_OFFLOAD flag if set */\n-\tmb->ol_flags &= ~(mb->ol_flags & PKT_RX_SEC_OFFLOAD);\n-\treturn 0;\n+/*\n+ * *process* function for tunnel packets\n+ */\n+static inline uint16_t\n+trs_process(const struct rte_ipsec_sa *sa, struct rte_mbuf *mb[],\n+\tuint32_t sqn[], uint32_t dr[], uint16_t num)\n+{\n+\tchar *np;\n+\tuint32_t i, k, l2, tl;\n+\tuint32_t hl[num];\n+\tstruct esp_tail espt[num];\n+\tstruct rte_mbuf *ml[num];\n+\n+\tconst uint32_t tlen = sa->icv_len + sizeof(espt[0]);\n+\tconst uint32_t cofs = sa->ctp.cipher.offset;\n+\n+\t/*\n+\t * to minimize stalls due to load latency,\n+\t * read mbufs metadata and esp tail first.\n+\t */\n+\tfor (i = 0; i != num; i++)\n+\t\tprocess_step1(mb[i], tlen, &ml[i], &espt[i], &hl[i]);\n+\n+\tk = 0;\n+\tfor (i = 0; i != num; i++) {\n+\n+\t\ttl = tlen + espt[i].pad_len;\n+\t\tl2 = mb[i]->l2_len;\n+\n+\t\t/* check that packet is valid */\n+\t\tif (trs_process_check(mb[i], ml[i], espt[i], hl[i] + cofs,\n+\t\t\t\ttl) == 0) {\n+\n+\t\t\t/* modify packet's layout */\n+\t\t\tnp = trs_process_step2(mb[i], ml[i], hl[i], cofs, tl,\n+\t\t\t\tsqn + k);\n+\t\t\tupdate_trs_l3hdr(sa, np + l2, mb[i]->pkt_len,\n+\t\t\t\tl2, hl[i] - l2, espt[i].next_proto);\n+\n+\t\t\t/* update mbuf's metadata */\n+\t\t\ttrs_process_step3(mb[i]);\n+\t\t\tk++;\n+\t\t} else\n+\t\t\tdr[i - k] = i;\n+\t}\n+\n+\treturn k;\n }\n \n /*\n@@ -333,11 +464,15 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[],\n \tuint32_t i, k;\n \tstruct replay_sqn *rsn;\n \n+\t/* replay not enabled */\n+\tif (sa->replay.win_sz == 0)\n+\t\treturn num;\n+\n \trsn = rsn_update_start(sa);\n \n \tk = 0;\n \tfor (i = 0; i != num; i++) {\n-\t\tif (esn_inb_update_sqn(rsn, sa, sqn[i]) == 0)\n+\t\tif (esn_inb_update_sqn(rsn, sa, rte_be_to_cpu_32(sqn[i])) == 0)\n \t\t\tk++;\n \t\telse\n \t\t\tdr[i - k] = i;\n@@ -348,13 +483,13 @@ esp_inb_rsn_update(struct rte_ipsec_sa *sa, const uint32_t sqn[],\n }\n \n /*\n- * process group of ESP inbound tunnel packets.\n+ * process group of ESP inbound packets.\n  */\n-uint16_t\n-esp_inb_tun_pkt_process(const struct rte_ipsec_session *ss,\n-\tstruct rte_mbuf *mb[], uint16_t num)\n+static inline uint16_t\n+esp_inb_pkt_process(const struct rte_ipsec_session *ss,\n+\tstruct rte_mbuf *mb[], uint16_t num, esp_inb_process_t process)\n {\n-\tuint32_t i, k, n;\n+\tuint32_t k, n;\n \tstruct rte_ipsec_sa *sa;\n \tuint32_t sqn[num];\n \tuint32_t dr[num];\n@@ -362,16 +497,7 @@ esp_inb_tun_pkt_process(const struct rte_ipsec_session *ss,\n \tsa = ss->sa;\n \n \t/* process packets, extract seq numbers */\n-\n-\tk = 0;\n-\tfor (i = 0; i != num; i++) {\n-\t\t/* good packet */\n-\t\tif (inb_tun_single_pkt_process(sa, mb[i], sqn + k) == 0)\n-\t\t\tk++;\n-\t\t/* bad packet, will drop from furhter processing */\n-\t\telse\n-\t\t\tdr[i - k] = i;\n-\t}\n+\tk = process(sa, mb, sqn, dr, num);\n \n \t/* handle unprocessed mbufs */\n \tif (k != num && k != 0)\n@@ -390,6 +516,16 @@ esp_inb_tun_pkt_process(const struct rte_ipsec_session *ss,\n \treturn n;\n }\n \n+/*\n+ * process group of ESP inbound tunnel packets.\n+ */\n+uint16_t\n+esp_inb_tun_pkt_process(const struct rte_ipsec_session *ss,\n+\tstruct rte_mbuf *mb[], uint16_t num)\n+{\n+\treturn esp_inb_pkt_process(ss, mb, num, tun_process);\n+}\n+\n /*\n  * process group of ESP inbound transport packets.\n  */\n@@ -397,38 +533,5 @@ uint16_t\n esp_inb_trs_pkt_process(const struct rte_ipsec_session *ss,\n \tstruct rte_mbuf *mb[], uint16_t num)\n {\n-\tuint32_t i, k, n;\n-\tuint32_t sqn[num];\n-\tstruct rte_ipsec_sa *sa;\n-\tuint32_t dr[num];\n-\n-\tsa = ss->sa;\n-\n-\t/* process packets, extract seq numbers */\n-\n-\tk = 0;\n-\tfor (i = 0; i != num; i++) {\n-\t\t/* good packet */\n-\t\tif (inb_trs_single_pkt_process(sa, mb[i], sqn + k) == 0)\n-\t\t\tk++;\n-\t\t/* bad packet, will drop from furhter processing */\n-\t\telse\n-\t\t\tdr[i - k] = i;\n-\t}\n-\n-\t/* handle unprocessed mbufs */\n-\tif (k != num && k != 0)\n-\t\tmove_bad_mbufs(mb, dr, num, num - k);\n-\n-\t/* update SQN and replay winow */\n-\tn = esp_inb_rsn_update(sa, sqn, dr, k);\n-\n-\t/* handle mbufs with wrong SQN */\n-\tif (n != k && n != 0)\n-\t\tmove_bad_mbufs(mb, dr, k, k - n);\n-\n-\tif (n != num)\n-\t\trte_errno = EBADMSG;\n-\n-\treturn n;\n+\treturn esp_inb_pkt_process(ss, mb, num, trs_process);\n }\ndiff --git a/lib/librte_ipsec/ipsec_sqn.h b/lib/librte_ipsec/ipsec_sqn.h\nindex 4ba079d75..0c2f76a7a 100644\n--- a/lib/librte_ipsec/ipsec_sqn.h\n+++ b/lib/librte_ipsec/ipsec_sqn.h\n@@ -152,10 +152,6 @@ esn_inb_update_sqn(struct replay_sqn *rsn, const struct rte_ipsec_sa *sa,\n {\n \tuint32_t bit, bucket, last_bucket, new_bucket, diff, i;\n \n-\t/* replay not enabled */\n-\tif (sa->replay.win_sz == 0)\n-\t\treturn 0;\n-\n \t/* handle ESN */\n \tif (IS_ESN(sa))\n \t\tsqn = reconstruct_esn(rsn->sqn, sqn, sa->replay.win_sz);\n",
    "prefixes": [
        "v5",
        "7/9"
    ]
}