Show a patch.

GET /api/patches/276/
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 276,
    "url": "http://patches.dpdk.org/api/patches/276/",
    "web_url": "http://patches.dpdk.org/patch/276/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk"
    },
    "msgid": "<1409567080-27083-4-git-send-email-david.marchand@6wind.com>",
    "date": "2014-09-01T10:24:26",
    "name": "[dpdk-dev,v2,03/17] ixgbe: clean log messages",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "2438ac5cac907ff4b4675e906bc5869165996b8d",
    "submitter": {
        "id": 3,
        "url": "http://patches.dpdk.org/api/people/3/",
        "name": "David Marchand",
        "email": "david.marchand@6wind.com"
    },
    "delegate": null,
    "mbox": "http://patches.dpdk.org/patch/276/mbox/",
    "series": [],
    "comments": "http://patches.dpdk.org/api/patches/276/comments/",
    "check": "pending",
    "checks": "http://patches.dpdk.org/api/patches/276/checks/",
    "tags": {},
    "headers": {
        "Return-Path": "<david.marchand@6wind.com>",
        "References": "<1409567080-27083-1-git-send-email-david.marchand@6wind.com>",
        "X-Mailman-Version": "2.1.15",
        "From": "David Marchand <david.marchand@6wind.com>",
        "X-List-Received-Date": "Mon, 01 Sep 2014 10:20:25 -0000",
        "X-BeenThere": "dev@dpdk.org",
        "Message-Id": "<1409567080-27083-4-git-send-email-david.marchand@6wind.com>",
        "X-Google-DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n\td=1e100.net; s=20130820;\n\th=x-gm-message-state:from:to:subject:date:message-id:in-reply-to\n\t:references;\n\tbh=7m7cfPnEVKn/S84YsOYVuT7147FspNFMm5FIz+bwYcI=;\n\tb=QxixjlVdqTcoeOl8g/xUB1LQQJu4cOYgoYqflaXUJhd0VFJ3/NnvITpF7bLCNNSCtv\n\thNS8OjEylK315d9UdYjekmYFQ3vSRlRy9u5NUTYHHR5GLahx8FYuMsIsKVI70DXAoAZp\n\t2X2vX9XaCct02H60JsC7f5ZjMjFLVMpoOS0w6M8j3eJWIJj8r+QAtJttkmfGY8F46lTB\n\tAuFu8uRqySK+5kE4fiPgADMZ+4HkR4bXwxXSQu7hDnFb7Aou4UtH51ru3CesCUhrnODi\n\tLjqFotR9jo0cRF9dnztbhAh0c1MxQEpPHhVzZ8+CDdaSHZOMtajMgOEOek7cmhSrJW3n\n\tQBNA==",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "X-Received": "by 10.180.91.40 with SMTP id cb8mr20813644wib.45.1409567091324; \n\tMon, 01 Sep 2014 03:24:51 -0700 (PDT)",
        "X-Mailer": "git-send-email 1.7.10.4",
        "Precedence": "list",
        "X-Gm-Message-State": "ALoCoQnUr/9vGDD1ls9MZjIdZo0zQ7/f75AhF5qIvTd0Eai5PBL9GGEiCiu2LkeOoTW0nYh0Q6aV",
        "Date": "Mon,  1 Sep 2014 12:24:26 +0200",
        "Subject": "[dpdk-dev] [PATCH v2 03/17] ixgbe: clean log messages",
        "List-Archive": "<http://dpdk.org/ml/archives/dev/>",
        "Received": [
            "from mail-we0-f177.google.com (mail-we0-f177.google.com\n\t[74.125.82.177]) by dpdk.org (Postfix) with ESMTP id 70C4BB3A7\n\tfor <dev@dpdk.org>; Mon,  1 Sep 2014 12:20:24 +0200 (CEST)",
            "by mail-we0-f177.google.com with SMTP id u56so5204416wes.36\n\tfor <dev@dpdk.org>; Mon, 01 Sep 2014 03:24:51 -0700 (PDT)",
            "from alcyon.dev.6wind.com (guy78-3-82-239-227-177.fbx.proxad.net.\n\t[82.239.227.177]) by mx.google.com with ESMTPSA id\n\tlm18sm24287018wic.22.2014.09.01.03.24.49 for <dev@dpdk.org>\n\t(version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128);\n\tMon, 01 Sep 2014 03:24:50 -0700 (PDT)"
        ],
        "List-Subscribe": "<http://dpdk.org/ml/listinfo/dev>,\n\t<mailto:dev-request@dpdk.org?subject=subscribe>",
        "List-Id": "patches and discussions about DPDK <dev.dpdk.org>",
        "In-Reply-To": "<1409567080-27083-1-git-send-email-david.marchand@6wind.com>",
        "List-Unsubscribe": "<http://dpdk.org/ml/options/dev>,\n\t<mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "To": "dev@dpdk.org"
    },
    "content": "Clean log messages:\n- remove leading \\n in some messages,\n- remove trailing \\n in some messages,\n- split multi lines messages,\n- replace some PMD_INIT_LOG(DEBUG, \"some_func\") with PMD_INIT_FUNC_TRACE().\n\nSigned-off-by: David Marchand <david.marchand@6wind.com>\n---\n lib/librte_pmd_ixgbe/ixgbe_ethdev.c |  150 +++++++++++++++++------------------\n lib/librte_pmd_ixgbe/ixgbe_fdir.c   |    6 +-\n lib/librte_pmd_ixgbe/ixgbe_rxtx.c   |   93 +++++++++++-----------\n 3 files changed, 124 insertions(+), 125 deletions(-)",
    "diff": "diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c\nindex a8a7ed6..1419494 100644\n--- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c\n+++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c\n@@ -547,12 +547,12 @@ ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev *eth_dev,\n \tif ((hw->mac.type != ixgbe_mac_82599EB) && (hw->mac.type != ixgbe_mac_X540))\n \t\treturn -ENOSYS;\n \n-\tPMD_INIT_LOG(INFO, \"Setting port %d, %s queue_id %d to stat index %d\\n\",\n+\tPMD_INIT_LOG(INFO, \"Setting port %d, %s queue_id %d to stat index %d\",\n \t\t     (int)(eth_dev->data->port_id), is_rx ? \"RX\" : \"TX\", queue_id, stat_idx);\n \n \tn = (uint8_t)(queue_id / NB_QMAP_FIELDS_PER_QSM_REG);\n \tif (n >= IXGBE_NB_STAT_MAPPING_REGS) {\n-\t\tPMD_INIT_LOG(ERR, \"Nb of stat mapping registers exceeded\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"Nb of stat mapping registers exceeded\");\n \t\treturn -EIO;\n \t}\n \toffset = (uint8_t)(queue_id % NB_QMAP_FIELDS_PER_QSM_REG);\n@@ -572,19 +572,20 @@ ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev *eth_dev,\n \telse\n \t\tstat_mappings->rqsmr[n] |= qsmr_mask;\n \n-\tPMD_INIT_LOG(INFO, \"Set port %d, %s queue_id %d to stat index %d\\n\"\n-\t\t     \"%s[%d] = 0x%08x\\n\",\n-\t\t     (int)(eth_dev->data->port_id), is_rx ? \"RX\" : \"TX\", queue_id, stat_idx,\n-\t\t     is_rx ? \"RQSMR\" : \"TQSM\",n, is_rx ? stat_mappings->rqsmr[n] : stat_mappings->tqsm[n]);\n+\tPMD_INIT_LOG(INFO, \"Set port %d, %s queue_id %d to stat index %d\",\n+\t\t     (int)(eth_dev->data->port_id), is_rx ? \"RX\" : \"TX\",\n+\t\t     queue_id, stat_idx);\n+\tPMD_INIT_LOG(INFO, \"%s[%d] = 0x%08x\", is_rx ? \"RQSMR\" : \"TQSM\", n,\n+\t\t     is_rx ? stat_mappings->rqsmr[n] : stat_mappings->tqsm[n]);\n \n \t/* Now write the mapping in the appropriate register */\n \tif (is_rx) {\n-\t\tPMD_INIT_LOG(INFO, \"Write 0x%x to RX IXGBE stat mapping reg:%d\\n\",\n+\t\tPMD_INIT_LOG(INFO, \"Write 0x%x to RX IXGBE stat mapping reg:%d\",\n \t\t\t     stat_mappings->rqsmr[n], n);\n \t\tIXGBE_WRITE_REG(hw, IXGBE_RQSMR(n), stat_mappings->rqsmr[n]);\n \t}\n \telse {\n-\t\tPMD_INIT_LOG(INFO, \"Write 0x%x to TX IXGBE stat mapping reg:%d\\n\",\n+\t\tPMD_INIT_LOG(INFO, \"Write 0x%x to TX IXGBE stat mapping reg:%d\",\n \t\t\t     stat_mappings->tqsm[n], n);\n \t\tIXGBE_WRITE_REG(hw, IXGBE_TQSM(n), stat_mappings->tqsm[n]);\n \t}\n@@ -790,12 +791,13 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \n \tif (diag == IXGBE_ERR_EEPROM_VERSION) {\n \t\tPMD_INIT_LOG(ERR, \"This device is a pre-production adapter/\"\n-\t\t    \"LOM.  Please be aware there may be issues associated \"\n-\t\t    \"with your hardware.\\n If you are experiencing problems \"\n+\t\t    \"LOM. Please be aware there may be issues associated \"\n+\t\t    \"with your hardware.\");\n+\t\tPMD_INIT_LOG(ERR, \"If you are experiencing problems \"\n \t\t    \"please contact your Intel or hardware representative \"\n-\t\t    \"who provided you with this hardware.\\n\");\n+\t\t    \"who provided you with this hardware.\");\n \t} else if (diag == IXGBE_ERR_SFP_NOT_SUPPORTED)\n-\t\tPMD_INIT_LOG(ERR, \"Unsupported SFP+ Module\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"Unsupported SFP+ Module\");\n \tif (diag) {\n \t\tPMD_INIT_LOG(ERR, \"Hardware Initialization Failure: %d\", diag);\n \t\treturn -EIO;\n@@ -811,10 +813,9 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \teth_dev->data->mac_addrs = rte_zmalloc(\"ixgbe\", ETHER_ADDR_LEN *\n \t\t\thw->mac.num_rar_entries, 0);\n \tif (eth_dev->data->mac_addrs == NULL) {\n-\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\"Failed to allocate %u bytes needed to store \"\n-\t\t\t\"MAC addresses\",\n-\t\t\tETHER_ADDR_LEN * hw->mac.num_rar_entries);\n+\t\tPMD_INIT_LOG(ERR, \"Failed to allocate %u bytes needed to store \"\n+\t\t\t     \"MAC addresses\",\n+\t\t\t     ETHER_ADDR_LEN * hw->mac.num_rar_entries);\n \t\treturn -ENOMEM;\n \t}\n \t/* Copy the permanent MAC address */\n@@ -825,9 +826,9 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \teth_dev->data->hash_mac_addrs = rte_zmalloc(\"ixgbe\", ETHER_ADDR_LEN *\n \t\t\tIXGBE_VMDQ_NUM_UC_MAC, 0);\n \tif (eth_dev->data->hash_mac_addrs == NULL) {\n-\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\"Failed to allocate %d bytes needed to store MAC addresses\",\n-\t\t\tETHER_ADDR_LEN * IXGBE_VMDQ_NUM_UC_MAC);\n+\t\tPMD_INIT_LOG(ERR, \"Failed to allocate %d bytes needed to store \"\n+\t\t\t     \"MAC addresses\",\n+\t\t\t     ETHER_ADDR_LEN * IXGBE_VMDQ_NUM_UC_MAC);\n \t\treturn -ENOMEM;\n \t}\n \n@@ -849,12 +850,11 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \tIXGBE_WRITE_FLUSH(hw);\n \n \tif (ixgbe_is_sfp(hw) && hw->phy.sfp_type != ixgbe_sfp_type_not_present)\n-\t\tPMD_INIT_LOG(DEBUG,\n-\t\t\t     \"MAC: %d, PHY: %d, SFP+: %d<n\",\n+\t\tPMD_INIT_LOG(DEBUG, \"MAC: %d, PHY: %d, SFP+: %d\",\n \t\t\t     (int) hw->mac.type, (int) hw->phy.type,\n \t\t\t     (int) hw->phy.sfp_type);\n \telse\n-\t\tPMD_INIT_LOG(DEBUG, \"MAC: %d, PHY: %d\\n\",\n+\t\tPMD_INIT_LOG(DEBUG, \"MAC: %d, PHY: %d\",\n \t\t\t     (int) hw->mac.type, (int) hw->phy.type);\n \n \tPMD_INIT_LOG(DEBUG, \"port %d vendorID=0x%x deviceID=0x%x\",\n@@ -933,7 +933,7 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \t\tIXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(eth_dev->data->dev_private);\n \tstruct ether_addr *perm_addr = (struct ether_addr *) hw->mac.perm_addr;\n \n-\tPMD_INIT_LOG(DEBUG, \"eth_ixgbevf_dev_init\");\n+\tPMD_INIT_FUNC_TRACE();\n \n \teth_dev->dev_ops = &ixgbevf_eth_dev_ops;\n \teth_dev->rx_pkt_burst = &ixgbe_recv_pkts;\n@@ -963,7 +963,8 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \t/* Initialize the shared code (base driver) */\n \tdiag = ixgbe_init_shared_code(hw);\n \tif (diag != IXGBE_SUCCESS) {\n-\t\tPMD_INIT_LOG(ERR, \"Shared code init failed for ixgbevf: %d\", diag);\n+\t\tPMD_INIT_LOG(ERR, \"Shared code init failed for ixgbevf: %d\",\n+\t\t\t     diag);\n \t\treturn -EIO;\n \t}\n \n@@ -996,10 +997,9 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \teth_dev->data->mac_addrs = rte_zmalloc(\"ixgbevf\", ETHER_ADDR_LEN *\n \t\t\thw->mac.num_rar_entries, 0);\n \tif (eth_dev->data->mac_addrs == NULL) {\n-\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\"Failed to allocate %u bytes needed to store \"\n-\t\t\t\"MAC addresses\",\n-\t\t\tETHER_ADDR_LEN * hw->mac.num_rar_entries);\n+\t\tPMD_INIT_LOG(ERR, \"Failed to allocate %u bytes needed to store \"\n+\t\t\t     \"MAC addresses\",\n+\t\t\t     ETHER_ADDR_LEN * hw->mac.num_rar_entries);\n \t\treturn -ENOMEM;\n \t}\n \n@@ -1033,13 +1033,14 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,\n \t\t\tbreak;\n \n \t\tdefault:\n-\t\t\tPMD_INIT_LOG(ERR, \"VF Initialization Failure: %d\", diag);\n+\t\t\tPMD_INIT_LOG(ERR, \"VF Initialization Failure: %d\",\n+\t\t\t\t     diag);\n \t\t\treturn (-EIO);\n \t}\n \n-\tPMD_INIT_LOG(DEBUG, \"\\nport %d vendorID=0x%x deviceID=0x%x mac.type=%s\\n\",\n-\t\t\t eth_dev->data->port_id, pci_dev->id.vendor_id, pci_dev->id.device_id,\n-\t\t\t \"ixgbe_mac_82599_vf\");\n+\tPMD_INIT_LOG(DEBUG, \"port %d vendorID=0x%x deviceID=0x%x mac.type=%s\",\n+\t\t     eth_dev->data->port_id, pci_dev->id.vendor_id,\n+\t\t     pci_dev->id.device_id, \"ixgbe_mac_82599_vf\");\n \n \treturn 0;\n }\n@@ -1416,8 +1417,8 @@ ixgbe_dev_start(struct rte_eth_dev *dev)\n \n \t/* IXGBE devices don't support half duplex */\n \tif ((dev->data->dev_conf.link_duplex != ETH_LINK_AUTONEG_DUPLEX) &&\n-\t\t\t(dev->data->dev_conf.link_duplex != ETH_LINK_FULL_DUPLEX)) {\n-\t\tPMD_INIT_LOG(ERR, \"Invalid link_duplex (%hu) for port %hhu\\n\",\n+\t    (dev->data->dev_conf.link_duplex != ETH_LINK_FULL_DUPLEX)) {\n+\t\tPMD_INIT_LOG(ERR, \"Invalid link_duplex (%hu) for port %hhu\",\n \t\t\t\tdev->data->dev_conf.link_duplex,\n \t\t\t\tdev->data->port_id);\n \t\treturn -EINVAL;\n@@ -1443,7 +1444,7 @@ ixgbe_dev_start(struct rte_eth_dev *dev)\n \t/* This can fail when allocating mbufs for descriptor rings */\n \terr = ixgbe_dev_rx_init(dev);\n \tif (err) {\n-\t\tPMD_INIT_LOG(ERR, \"Unable to initialize RX hardware\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"Unable to initialize RX hardware\");\n \t\tgoto error;\n \t}\n \n@@ -1490,9 +1491,9 @@ ixgbe_dev_start(struct rte_eth_dev *dev)\n \t\tspeed = IXGBE_LINK_SPEED_10GB_FULL;\n \t\tbreak;\n \tdefault:\n-\t\tPMD_INIT_LOG(ERR, \"Invalid link_speed (%hu) for port %hhu\\n\",\n-\t\t\t\tdev->data->dev_conf.link_speed,\n-\t\t\t\tdev->data->port_id);\n+\t\tPMD_INIT_LOG(ERR, \"Invalid link_speed (%hu) for port %hhu\",\n+\t\t\t     dev->data->dev_conf.link_speed,\n+\t\t\t     dev->data->port_id);\n \t\tgoto error;\n \t}\n \n@@ -1598,10 +1599,8 @@ ixgbe_dev_set_link_up(struct rte_eth_dev *dev)\n #ifdef RTE_NIC_BYPASS\n \t\tif (hw->device_id == IXGBE_DEV_ID_82599_BYPASS) {\n \t\t\t/* Not suported in bypass mode */\n-\t\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\t\"\\nSet link up is not supported \"\n-\t\t\t\t\"by device id 0x%x\\n\",\n-\t\t\t\thw->device_id);\n+\t\t\tPMD_INIT_LOG(ERR, \"Set link up is not supported \"\n+\t\t\t\t     \"by device id 0x%x\", hw->device_id);\n \t\t\treturn -ENOTSUP;\n \t\t}\n #endif\n@@ -1610,8 +1609,8 @@ ixgbe_dev_set_link_up(struct rte_eth_dev *dev)\n \t\treturn 0;\n \t}\n \n-\tPMD_INIT_LOG(ERR, \"\\nSet link up is not supported by device id 0x%x\\n\",\n-\t\thw->device_id);\n+\tPMD_INIT_LOG(ERR, \"Set link up is not supported by device id 0x%x\",\n+\t\t     hw->device_id);\n \treturn -ENOTSUP;\n }\n \n@@ -1627,10 +1626,8 @@ ixgbe_dev_set_link_down(struct rte_eth_dev *dev)\n #ifdef RTE_NIC_BYPASS\n \t\tif (hw->device_id == IXGBE_DEV_ID_82599_BYPASS) {\n \t\t\t/* Not suported in bypass mode */\n-\t\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\t\"\\nSet link down is not supported \"\n-\t\t\t\t\"by device id 0x%x\\n\",\n-\t\t\t\t hw->device_id);\n+\t\t\tPMD_INIT_LOG(ERR, \"Set link down is not supported \"\n+\t\t\t\t     \"by device id 0x%x\", hw->device_id);\n \t\t\treturn -ENOTSUP;\n \t\t}\n #endif\n@@ -1639,9 +1636,8 @@ ixgbe_dev_set_link_down(struct rte_eth_dev *dev)\n \t\treturn 0;\n \t}\n \n-\tPMD_INIT_LOG(ERR,\n-\t\t\"\\nSet link down is not supported by device id 0x%x\\n\",\n-\t\t hw->device_id);\n+\tPMD_INIT_LOG(ERR, \"Set link down is not supported by device id 0x%x\",\n+\t\t     hw->device_id);\n \treturn -ENOTSUP;\n }\n \n@@ -2179,7 +2175,7 @@ ixgbe_dev_interrupt_action(struct rte_eth_dev *dev)\n \tstruct rte_eth_link link;\n \tint intr_enable_delay = false;\n \n-\tPMD_DRV_LOG(DEBUG, \"intr action type %d\\n\", intr->flags);\n+\tPMD_DRV_LOG(DEBUG, \"intr action type %d\", intr->flags);\n \n \tif (intr->flags & IXGBE_FLAG_MAILBOX) {\n \t\tixgbe_pf_mbx_process(dev);\n@@ -2209,7 +2205,8 @@ ixgbe_dev_interrupt_action(struct rte_eth_dev *dev)\n \n \tif (intr_enable_delay) {\n \t\tif (rte_eal_alarm_set(timeout * 1000,\n-\t\t\t\t      ixgbe_dev_interrupt_delayed_handler, (void*)dev) < 0)\n+\t\t\t\t      ixgbe_dev_interrupt_delayed_handler,\n+\t\t\t\t      (void *)dev) < 0)\n \t\t\tPMD_DRV_LOG(ERR, \"Error setting alarm\");\n \t} else {\n \t\tPMD_DRV_LOG(DEBUG, \"enable intr immediately\");\n@@ -2256,7 +2253,7 @@ ixgbe_dev_interrupt_delayed_handler(void *param)\n \t\t_rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC);\n \t}\n \n-\tPMD_DRV_LOG(DEBUG, \"enable intr in delayed handler S[%08x]\\n\", eicr);\n+\tPMD_DRV_LOG(DEBUG, \"enable intr in delayed handler S[%08x]\", eicr);\n \tixgbe_enable_intr(dev);\n \trte_intr_enable(&(dev->pci_dev->intr_handle));\n }\n@@ -2370,7 +2367,7 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)\n \tif (fc_conf->autoneg != !hw->fc.disable_fc_autoneg)\n \t\treturn -ENOTSUP;\n \trx_buf_size = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(0));\n-\tPMD_INIT_LOG(DEBUG, \"Rx packet buffer size = 0x%x \\n\", rx_buf_size);\n+\tPMD_INIT_LOG(DEBUG, \"Rx packet buffer size = 0x%x\", rx_buf_size);\n \n \t/*\n \t * At least reserve one Ethernet frame for watermark\n@@ -2379,8 +2376,8 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)\n \tmax_high_water = (rx_buf_size - ETHER_MAX_LEN) >> IXGBE_RXPBSIZE_SHIFT;\n \tif ((fc_conf->high_water > max_high_water) ||\n \t\t(fc_conf->high_water < fc_conf->low_water)) {\n-\t\tPMD_INIT_LOG(ERR, \"Invalid high/low water setup value in KB\\n\");\n-\t\tPMD_INIT_LOG(ERR, \"High_water must <=  0x%x\\n\", max_high_water);\n+\t\tPMD_INIT_LOG(ERR, \"Invalid high/low water setup value in KB\");\n+\t\tPMD_INIT_LOG(ERR, \"High_water must <= 0x%x\", max_high_water);\n \t\treturn (-EINVAL);\n \t}\n \n@@ -2412,7 +2409,7 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)\n \t\treturn 0;\n \t}\n \n-\tPMD_INIT_LOG(ERR, \"ixgbe_fc_enable = 0x%x \\n\", err);\n+\tPMD_INIT_LOG(ERR, \"ixgbe_fc_enable = 0x%x\", err);\n \treturn -EIO;\n }\n \n@@ -2442,13 +2439,13 @@ ixgbe_dcb_pfc_enable_generic(struct ixgbe_hw *hw,uint8_t tc_num)\n \tif (hw->fc.current_mode & ixgbe_fc_tx_pause) {\n \t\t /* High/Low water can not be 0 */\n \t\tif( (!hw->fc.high_water[tc_num])|| (!hw->fc.low_water[tc_num])) {\n-\t\t\tPMD_INIT_LOG(ERR,\"Invalid water mark configuration\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \"Invalid water mark configuration\");\n \t\t\tret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;\n \t\t\tgoto out;\n \t\t}\n \n \t\tif(hw->fc.low_water[tc_num] >= hw->fc.high_water[tc_num]) {\n-\t\t\tPMD_INIT_LOG(ERR,\"Invalid water mark configuration\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \"Invalid water mark configuration\");\n \t\t\tret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;\n \t\t\tgoto out;\n \t\t}\n@@ -2592,7 +2589,7 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_pfc_conf *p\n \tixgbe_dcb_unpack_map_cee(dcb_config, IXGBE_DCB_RX_CONFIG, map);\n \ttc_num = map[pfc_conf->priority];\n \trx_buf_size = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(tc_num));\n-\tPMD_INIT_LOG(DEBUG, \"Rx packet buffer size = 0x%x \\n\", rx_buf_size);\n+\tPMD_INIT_LOG(DEBUG, \"Rx packet buffer size = 0x%x\", rx_buf_size);\n \t/*\n \t * At least reserve one Ethernet frame for watermark\n \t * high_water/low_water in kilo bytes for ixgbe\n@@ -2600,8 +2597,8 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_pfc_conf *p\n \tmax_high_water = (rx_buf_size - ETHER_MAX_LEN) >> IXGBE_RXPBSIZE_SHIFT;\n \tif ((pfc_conf->fc.high_water > max_high_water) ||\n \t\t(pfc_conf->fc.high_water <= pfc_conf->fc.low_water)) {\n-\t\tPMD_INIT_LOG(ERR, \"Invalid high/low water setup value in KB\\n\");\n-\t\tPMD_INIT_LOG(ERR, \"High_water must <=  0x%x\\n\", max_high_water);\n+\t\tPMD_INIT_LOG(ERR, \"Invalid high/low water setup value in KB\");\n+\t\tPMD_INIT_LOG(ERR, \"High_water must <= 0x%x\", max_high_water);\n \t\treturn (-EINVAL);\n \t}\n \n@@ -2617,7 +2614,7 @@ ixgbe_priority_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_pfc_conf *p\n \tif ((err == IXGBE_SUCCESS) || (err == IXGBE_ERR_FC_NOT_NEGOTIATED))\n \t\treturn 0;\n \n-\tPMD_INIT_LOG(ERR, \"ixgbe_dcb_pfc_enable = 0x%x \\n\", err);\n+\tPMD_INIT_LOG(ERR, \"ixgbe_dcb_pfc_enable = 0x%x\", err);\n \treturn -EIO;\n }\n \n@@ -2764,7 +2761,7 @@ ixgbe_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu)\n static void\n ixgbevf_intr_disable(struct ixgbe_hw *hw)\n {\n-\tPMD_INIT_LOG(DEBUG, \"ixgbevf_intr_disable\");\n+\tPMD_INIT_FUNC_TRACE();\n \n \t/* Clear interrupt mask to stop from interrupts being generated */\n \tIXGBE_WRITE_REG(hw, IXGBE_VTEIMC, IXGBE_VF_IRQ_CLEAR_MASK);\n@@ -2777,8 +2774,8 @@ ixgbevf_dev_configure(struct rte_eth_dev *dev)\n {\n \tstruct rte_eth_conf* conf = &dev->data->dev_conf;\n \n-\tPMD_INIT_LOG(DEBUG, \"\\nConfigured Virtual Function port id: %d\\n\",\n-\t\tdev->data->port_id);\n+\tPMD_INIT_LOG(DEBUG, \"Configured Virtual Function port id: %d\",\n+\t\t     dev->data->port_id);\n \n \t/*\n \t * VF has no ability to enable/disable HW CRC\n@@ -2786,12 +2783,12 @@ ixgbevf_dev_configure(struct rte_eth_dev *dev)\n \t */\n #ifndef RTE_LIBRTE_IXGBE_PF_DISABLE_STRIP_CRC\n \tif (!conf->rxmode.hw_strip_crc) {\n-\t\tPMD_INIT_LOG(INFO, \"VF can't disable HW CRC Strip\\n\");\n+\t\tPMD_INIT_LOG(INFO, \"VF can't disable HW CRC Strip\");\n \t\tconf->rxmode.hw_strip_crc = 1;\n \t}\n #else\n \tif (conf->rxmode.hw_strip_crc) {\n-\t\tPMD_INIT_LOG(INFO, \"VF can't enable HW CRC Strip\\n\");\n+\t\tPMD_INIT_LOG(INFO, \"VF can't enable HW CRC Strip\");\n \t\tconf->rxmode.hw_strip_crc = 0;\n \t}\n #endif\n@@ -2806,7 +2803,7 @@ ixgbevf_dev_start(struct rte_eth_dev *dev)\n \t\tIXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);\n \tint err, mask = 0;\n \n-\tPMD_INIT_LOG(DEBUG, \"ixgbevf_dev_start\");\n+\tPMD_INIT_FUNC_TRACE();\n \n \thw->mac.ops.reset_hw(hw);\n \n@@ -2818,7 +2815,7 @@ ixgbevf_dev_start(struct rte_eth_dev *dev)\n \t/* This can fail when allocating mbufs for descriptor rings */\n \terr = ixgbevf_dev_rx_init(dev);\n \tif (err) {\n-\t\tPMD_INIT_LOG(ERR, \"Unable to initialize RX hardware (%d)\\n\", err);\n+\t\tPMD_INIT_LOG(ERR, \"Unable to initialize RX hardware (%d)\", err);\n \t\tixgbe_dev_clear_queues(dev);\n \t\treturn err;\n \t}\n@@ -2841,7 +2838,7 @@ ixgbevf_dev_stop(struct rte_eth_dev *dev)\n {\n \tstruct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);\n \n-\tPMD_INIT_LOG(DEBUG, \"ixgbevf_dev_stop\");\n+\tPMD_INIT_FUNC_TRACE();\n \n \thw->adapter_stopped = TRUE;\n \tixgbe_stop_adapter(hw);\n@@ -2860,7 +2857,7 @@ ixgbevf_dev_close(struct rte_eth_dev *dev)\n {\n \tstruct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);\n \n-\tPMD_INIT_LOG(DEBUG, \"ixgbevf_dev_close\");\n+\tPMD_INIT_FUNC_TRACE();\n \n \tixgbe_reset_hw(hw);\n \n@@ -2969,7 +2966,7 @@ ixgbe_vmdq_mode_check(struct ixgbe_hw *hw)\n \t/* we only need to do this if VMDq is enabled */\n \treg_val = IXGBE_READ_REG(hw, IXGBE_VT_CTL);\n \tif (!(reg_val & IXGBE_VT_CTL_VT_ENABLE)) {\n-\t\tPMD_INIT_LOG(ERR, \"VMDq must be enabled for this setting\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"VMDq must be enabled for this setting\");\n \t\treturn (-1);\n \t}\n \n@@ -3098,7 +3095,7 @@ ixgbe_set_pool_rx_mode(struct rte_eth_dev *dev, uint16_t pool,\n \n \tif (hw->mac.type == ixgbe_mac_82598EB) {\n \t\tPMD_INIT_LOG(ERR, \"setting VF receive mode set should be done\"\n-\t\t\t\" on 82599 hardware and newer\\n\");\n+\t\t\t     \" on 82599 hardware and newer\");\n \t\treturn (-ENOTSUP);\n \t}\n \tif (ixgbe_vmdq_mode_check(hw) < 0)\n@@ -3513,8 +3510,7 @@ ixgbevf_remove_mac_addr(struct rte_eth_dev *dev, uint32_t index)\n \t\t\tcontinue;\n \t\tdiag = ixgbevf_set_uc_addr_vf(hw, 2, mac_addr->addr_bytes);\n \t\tif (diag != 0)\n-\t\t\tPMD_DRV_LOG(ERR,\n-\t\t\t\t    \"Adding again MAC address \"\n+\t\t\tPMD_DRV_LOG(ERR, \"Adding again MAC address \"\n \t\t\t\t    \"%02x:%02x:%02x:%02x:%02x:%02x failed \"\n \t\t\t\t    \"diag=%d\",\n \t\t\t\t    mac_addr->addr_bytes[0],\ndiff --git a/lib/librte_pmd_ixgbe/ixgbe_fdir.c b/lib/librte_pmd_ixgbe/ixgbe_fdir.c\nindex 6c0a530..8819aac 100644\n--- a/lib/librte_pmd_ixgbe/ixgbe_fdir.c\n+++ b/lib/librte_pmd_ixgbe/ixgbe_fdir.c\n@@ -112,7 +112,7 @@ static void fdir_enable_82599(struct ixgbe_hw *hw, u32 fdirctrl)\n \t}\n \n \tif (i >= IXGBE_FDIR_INIT_DONE_POLL)\n-\t\tPMD_INIT_LOG(WARNING, \"Flow Director poll time exceeded!\\n\");\n+\t\tPMD_INIT_LOG(WARNING, \"Flow Director poll time exceeded!\");\n }\n \n /*\n@@ -381,7 +381,7 @@ fdir_add_signature_filter_82599(struct ixgbe_hw *hw,\n \tfdirhashcmd |= fdirhash;\n \tIXGBE_WRITE_REG64(hw, IXGBE_FDIRHASH, fdirhashcmd);\n \n-\tPMD_INIT_LOG(DEBUG, \"Tx Queue=%x hash=%x\\n\", queue, (u32)fdirhashcmd);\n+\tPMD_INIT_LOG(DEBUG, \"Tx Queue=%x hash=%x\", queue, (u32)fdirhashcmd);\n }\n \n /*\n@@ -614,7 +614,7 @@ fdir_set_input_mask_82599(struct ixgbe_hw *hw,\n \t\t/* use the L4 protocol mask for raw IPv4/IPv6 traffic */\n \t\tfdirm |= IXGBE_FDIRM_L4P;\n \t\tif (input_mask->dst_port_mask || input_mask->src_port_mask) {\n-\t\t\tPMD_INIT_LOG(ERR, \" Error on src/dst port mask\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \" Error on src/dst port mask\");\n \t\t\treturn -EINVAL;\n \t\t}\n \t}\ndiff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c\nindex 46962bc..981df60 100644\n--- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c\n+++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c\n@@ -490,8 +490,7 @@ ixgbe_xmit_cleanup(struct igb_tx_queue *txq)\n \tdesc_to_clean_to = sw_ring[desc_to_clean_to].last_id;\n \tif (! (txr[desc_to_clean_to].wb.status & IXGBE_TXD_STAT_DD))\n \t{\n-\t\tPMD_TX_FREE_LOG(DEBUG,\n-\t\t\t\t\"TX descriptor %4u is not done\"\n+\t\tPMD_TX_FREE_LOG(DEBUG, \"TX descriptor %4u is not done\"\n \t\t\t\t\"(port=%d queue=%d)\",\n \t\t\t\tdesc_to_clean_to,\n \t\t\t\ttxq->port_id, txq->queue_id);\n@@ -507,8 +506,7 @@ ixgbe_xmit_cleanup(struct igb_tx_queue *txq)\n \t\tnb_tx_to_clean = (uint16_t)(desc_to_clean_to -\n \t\t\t\t\t\tlast_desc_cleaned);\n \n-\tPMD_TX_FREE_LOG(DEBUG,\n-\t\t\t\"Cleaning %4u TX descriptors: %4u to %4u \"\n+\tPMD_TX_FREE_LOG(DEBUG, \"Cleaning %4u TX descriptors: %4u to %4u \"\n \t\t\t\"(port=%d queue=%d)\",\n \t\t\tnb_tx_to_clean, last_desc_cleaned, desc_to_clean_to,\n \t\t\ttxq->port_id, txq->queue_id);\n@@ -614,7 +612,7 @@ ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,\n \t\t\ttx_last = (uint16_t) (tx_last - txq->nb_tx_desc);\n \n \t\tPMD_TX_LOG(DEBUG, \"port_id=%u queue_id=%u pktlen=%u\"\n-\t\t\t   \" tx_first=%u tx_last=%u\\n\",\n+\t\t\t   \" tx_first=%u tx_last=%u\",\n \t\t\t   (unsigned) txq->port_id,\n \t\t\t   (unsigned) txq->queue_id,\n \t\t\t   (unsigned) pkt_len,\n@@ -627,8 +625,7 @@ ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,\n \t\t * nb_used better be less than or equal to txq->tx_rs_thresh\n \t\t */\n \t\tif (nb_used > txq->nb_tx_free) {\n-\t\t\tPMD_TX_FREE_LOG(DEBUG,\n-\t\t\t\t\t\"Not enough free TX descriptors \"\n+\t\t\tPMD_TX_FREE_LOG(DEBUG, \"Not enough free TX descriptors \"\n \t\t\t\t\t\"nb_used=%4u nb_free=%4u \"\n \t\t\t\t\t\"(port=%d queue=%d)\",\n \t\t\t\t\tnb_used, txq->nb_tx_free,\n@@ -1066,7 +1063,7 @@ rx_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \t\tif (ixgbe_rx_alloc_bufs(rxq) != 0) {\n \t\t\tint i, j;\n \t\t\tPMD_RX_LOG(DEBUG, \"RX mbuf alloc failed port_id=%u \"\n-\t\t\t\t   \"queue_id=%u\\n\", (unsigned) rxq->port_id,\n+\t\t\t\t   \"queue_id=%u\", (unsigned) rxq->port_id,\n \t\t\t\t   (unsigned) rxq->queue_id);\n \n \t\t\trte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed +=\n@@ -1193,7 +1190,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \t\t * frames to its peer(s).\n \t\t */\n \t\tPMD_RX_LOG(DEBUG, \"port_id=%u queue_id=%u rx_id=%u \"\n-\t\t\t   \"ext_err_stat=0x%08x pkt_len=%u\\n\",\n+\t\t\t   \"ext_err_stat=0x%08x pkt_len=%u\",\n \t\t\t   (unsigned) rxq->port_id, (unsigned) rxq->queue_id,\n \t\t\t   (unsigned) rx_id, (unsigned) staterr,\n \t\t\t   (unsigned) rte_le_to_cpu_16(rxd.wb.upper.length));\n@@ -1201,7 +1198,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \t\tnmb = rte_rxmbuf_alloc(rxq->mb_pool);\n \t\tif (nmb == NULL) {\n \t\t\tPMD_RX_LOG(DEBUG, \"RX mbuf alloc failed port_id=%u \"\n-\t\t\t\t   \"queue_id=%u\\n\", (unsigned) rxq->port_id,\n+\t\t\t\t   \"queue_id=%u\", (unsigned) rxq->port_id,\n \t\t\t\t   (unsigned) rxq->queue_id);\n \t\t\trte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed++;\n \t\t\tbreak;\n@@ -1296,7 +1293,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \tnb_hold = (uint16_t) (nb_hold + rxq->nb_rx_hold);\n \tif (nb_hold > rxq->rx_free_thresh) {\n \t\tPMD_RX_LOG(DEBUG, \"port_id=%u queue_id=%u rx_tail=%u \"\n-\t\t\t   \"nb_hold=%u nb_rx=%u\\n\",\n+\t\t\t   \"nb_hold=%u nb_rx=%u\",\n \t\t\t   (unsigned) rxq->port_id, (unsigned) rxq->queue_id,\n \t\t\t   (unsigned) rx_id, (unsigned) nb_hold,\n \t\t\t   (unsigned) nb_rx);\n@@ -1383,8 +1380,8 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \t\t * to happen by sending specific \"back-pressure\" flow control\n \t\t * frames to its peer(s).\n \t\t */\n-\t\tPMD_RX_LOG(DEBUG, \"\\nport_id=%u queue_id=%u rx_id=%u \"\n-\t\t\t   \"staterr=0x%x data_len=%u\\n\",\n+\t\tPMD_RX_LOG(DEBUG, \"port_id=%u queue_id=%u rx_id=%u \"\n+\t\t\t   \"staterr=0x%x data_len=%u\",\n \t\t\t   (unsigned) rxq->port_id, (unsigned) rxq->queue_id,\n \t\t\t   (unsigned) rx_id, (unsigned) staterr,\n \t\t\t   (unsigned) rte_le_to_cpu_16(rxd.wb.upper.length));\n@@ -1392,7 +1389,7 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \t\tnmb = rte_rxmbuf_alloc(rxq->mb_pool);\n \t\tif (nmb == NULL) {\n \t\t\tPMD_RX_LOG(DEBUG, \"RX mbuf alloc failed port_id=%u \"\n-\t\t\t\t   \"queue_id=%u\\n\", (unsigned) rxq->port_id,\n+\t\t\t\t   \"queue_id=%u\", (unsigned) rxq->port_id,\n \t\t\t\t   (unsigned) rxq->queue_id);\n \t\t\trte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed++;\n \t\t\tbreak;\n@@ -1561,7 +1558,7 @@ ixgbe_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,\n \tnb_hold = (uint16_t) (nb_hold + rxq->nb_rx_hold);\n \tif (nb_hold > rxq->rx_free_thresh) {\n \t\tPMD_RX_LOG(DEBUG, \"port_id=%u queue_id=%u rx_tail=%u \"\n-\t\t\t   \"nb_hold=%u nb_rx=%u\\n\",\n+\t\t\t   \"nb_hold=%u nb_rx=%u\",\n \t\t\t   (unsigned) rxq->port_id, (unsigned) rxq->queue_id,\n \t\t\t   (unsigned) rx_id, (unsigned) nb_hold,\n \t\t\t   (unsigned) nb_rx);\n@@ -1767,7 +1764,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \tif (tx_rs_thresh >= (nb_desc - 2)) {\n \t\tPMD_INIT_LOG(ERR, \"tx_rs_thresh must be less than the number \"\n \t\t\t     \"of TX descriptors minus 2. (tx_rs_thresh=%u \"\n-\t\t\t     \"port=%d queue=%d)\\n\", (unsigned int)tx_rs_thresh,\n+\t\t\t     \"port=%d queue=%d)\", (unsigned int)tx_rs_thresh,\n \t\t\t     (int)dev->data->port_id, (int)queue_idx);\n \t\treturn -(EINVAL);\n \t}\n@@ -1775,7 +1772,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \t\tPMD_INIT_LOG(ERR, \"tx_rs_thresh must be less than the \"\n \t\t\t     \"tx_free_thresh must be less than the number of \"\n \t\t\t     \"TX descriptors minus 3. (tx_free_thresh=%u \"\n-\t\t\t     \"port=%d queue=%d)\\n\",\n+\t\t\t     \"port=%d queue=%d)\",\n \t\t\t     (unsigned int)tx_free_thresh,\n \t\t\t     (int)dev->data->port_id, (int)queue_idx);\n \t\treturn -(EINVAL);\n@@ -1783,7 +1780,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \tif (tx_rs_thresh > tx_free_thresh) {\n \t\tPMD_INIT_LOG(ERR, \"tx_rs_thresh must be less than or equal to \"\n \t\t\t     \"tx_free_thresh. (tx_free_thresh=%u \"\n-\t\t\t     \"tx_rs_thresh=%u port=%d queue=%d)\\n\",\n+\t\t\t     \"tx_rs_thresh=%u port=%d queue=%d)\",\n \t\t\t     (unsigned int)tx_free_thresh,\n \t\t\t     (unsigned int)tx_rs_thresh,\n \t\t\t     (int)dev->data->port_id,\n@@ -1793,7 +1790,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \tif ((nb_desc % tx_rs_thresh) != 0) {\n \t\tPMD_INIT_LOG(ERR, \"tx_rs_thresh must be a divisor of the \"\n \t\t\t     \"number of TX descriptors. (tx_rs_thresh=%u \"\n-\t\t\t     \"port=%d queue=%d)\\n\", (unsigned int)tx_rs_thresh,\n+\t\t\t     \"port=%d queue=%d)\", (unsigned int)tx_rs_thresh,\n \t\t\t     (int)dev->data->port_id, (int)queue_idx);\n \t\treturn -(EINVAL);\n \t}\n@@ -1807,7 +1804,7 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \tif ((tx_rs_thresh > 1) && (tx_conf->tx_thresh.wthresh != 0)) {\n \t\tPMD_INIT_LOG(ERR, \"TX WTHRESH must be set to 0 if \"\n \t\t\t     \"tx_rs_thresh is greater than 1. (tx_rs_thresh=%u \"\n-\t\t\t     \"port=%d queue=%d)\\n\", (unsigned int)tx_rs_thresh,\n+\t\t\t     \"port=%d queue=%d)\", (unsigned int)tx_rs_thresh,\n \t\t\t     (int)dev->data->port_id, (int)queue_idx);\n \t\treturn -(EINVAL);\n \t}\n@@ -1873,26 +1870,32 @@ ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev,\n \t\tixgbe_tx_queue_release(txq);\n \t\treturn (-ENOMEM);\n \t}\n-\tPMD_INIT_LOG(DEBUG, \"sw_ring=%p hw_ring=%p dma_addr=0x%\"PRIx64\"\\n\",\n+\tPMD_INIT_LOG(DEBUG, \"sw_ring=%p hw_ring=%p dma_addr=0x%\"PRIx64,\n \t\t     txq->sw_ring, txq->tx_ring, txq->tx_ring_phys_addr);\n \n \t/* Use a simple Tx queue (no offloads, no multi segs) if possible */\n \tif (((txq->txq_flags & IXGBE_SIMPLE_FLAGS) == IXGBE_SIMPLE_FLAGS) &&\n \t    (txq->tx_rs_thresh >= RTE_PMD_IXGBE_TX_MAX_BURST)) {\n-\t\tPMD_INIT_LOG(INFO, \"Using simple tx code path\\n\");\n+\t\tPMD_INIT_LOG(INFO, \"Using simple tx code path\");\n #ifdef RTE_IXGBE_INC_VECTOR\n \t\tif (txq->tx_rs_thresh <= RTE_IXGBE_TX_MAX_FREE_BUF_SZ &&\n \t\t    ixgbe_txq_vec_setup(txq, socket_id) == 0) {\n-\t\t\tPMD_INIT_LOG(INFO, \"Vector tx enabled.\\n\");\n+\t\t\tPMD_INIT_LOG(INFO, \"Vector tx enabled.\");\n \t\t\tdev->tx_pkt_burst = ixgbe_xmit_pkts_vec;\n \t\t}\n \t\telse\n #endif\n \t\t\tdev->tx_pkt_burst = ixgbe_xmit_pkts_simple;\n \t} else {\n-\t\tPMD_INIT_LOG(INFO, \"Using full-featured tx code path\\n\");\n-\t\tPMD_INIT_LOG(INFO, \" - txq_flags = %lx [IXGBE_SIMPLE_FLAGS=%lx]\\n\", (long unsigned)txq->txq_flags, (long unsigned)IXGBE_SIMPLE_FLAGS);\n-\t\tPMD_INIT_LOG(INFO, \" - tx_rs_thresh = %lu [RTE_PMD_IXGBE_TX_MAX_BURST=%lu]\\n\", (long unsigned)txq->tx_rs_thresh, (long unsigned)RTE_PMD_IXGBE_TX_MAX_BURST);\n+\t\tPMD_INIT_LOG(INFO, \"Using full-featured tx code path\");\n+\t\tPMD_INIT_LOG(INFO, \" - txq_flags = %lx \"\n+\t\t\t     \"[IXGBE_SIMPLE_FLAGS=%lx]\",\n+\t\t\t     (long unsigned)txq->txq_flags,\n+\t\t\t     (long unsigned)IXGBE_SIMPLE_FLAGS);\n+\t\tPMD_INIT_LOG(INFO, \" - tx_rs_thresh = %lu \"\n+\t\t\t     \"[RTE_PMD_IXGBE_TX_MAX_BURST=%lu]\",\n+\t\t\t     (long unsigned)txq->tx_rs_thresh,\n+\t\t\t     (long unsigned)RTE_PMD_IXGBE_TX_MAX_BURST);\n \t\tdev->tx_pkt_burst = ixgbe_xmit_pkts;\n \t}\n \n@@ -2152,7 +2155,7 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev,\n \t\tixgbe_rx_queue_release(rxq);\n \t\treturn (-ENOMEM);\n \t}\n-\tPMD_INIT_LOG(DEBUG, \"sw_ring=%p hw_ring=%p dma_addr=0x%\"PRIx64\"\\n\",\n+\tPMD_INIT_LOG(DEBUG, \"sw_ring=%p hw_ring=%p dma_addr=0x%\"PRIx64,\n \t\t     rxq->sw_ring, rxq->rx_ring, rxq->rx_ring_phys_addr);\n \n \t/*\n@@ -2166,13 +2169,13 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev,\n #ifdef RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC\n \t\tPMD_INIT_LOG(DEBUG, \"Rx Burst Bulk Alloc Preconditions are \"\n \t\t\t     \"satisfied. Rx Burst Bulk Alloc function will be \"\n-\t\t\t     \"used on port=%d, queue=%d.\\n\",\n+\t\t\t     \"used on port=%d, queue=%d.\",\n \t\t\t     rxq->port_id, rxq->queue_id);\n \t\tdev->rx_pkt_burst = ixgbe_recv_pkts_bulk_alloc;\n #ifdef RTE_IXGBE_INC_VECTOR\n \t\tif (!ixgbe_rx_vec_condition_check(dev)) {\n \t\t\tPMD_INIT_LOG(INFO, \"Vector rx enabled, please make \"\n-\t\t\t\t     \"sure RX burst size no less than 32.\\n\");\n+\t\t\t\t     \"sure RX burst size no less than 32.\");\n \t\t\tixgbe_rxq_vec_setup(rxq, socket_id);\n \t\t\tdev->rx_pkt_burst = ixgbe_recv_pkts_vec;\n \t\t}\n@@ -2182,7 +2185,7 @@ ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev,\n \t\tPMD_INIT_LOG(DEBUG, \"Rx Burst Bulk Alloc Preconditions \"\n \t\t\t     \"are not satisfied, Scattered Rx is requested, \"\n \t\t\t     \"or RTE_LIBRTE_IXGBE_RX_ALLOW_BULK_ALLOC is not \"\n-\t\t\t     \"enabled (port=%d, queue=%d).\\n\",\n+\t\t\t     \"enabled (port=%d, queue=%d).\",\n \t\t\t     rxq->port_id, rxq->queue_id);\n \t}\n \tdev->data->rx_queues[queue_idx] = rxq;\n@@ -2201,7 +2204,7 @@ ixgbe_dev_rx_queue_count(struct rte_eth_dev *dev, uint16_t rx_queue_id)\n \tuint32_t desc = 0;\n \n \tif (rx_queue_id >= dev->data->nb_rx_queues) {\n-\t\tPMD_RX_LOG(ERR, \"Invalid RX queue id=%d\\n\", rx_queue_id);\n+\t\tPMD_RX_LOG(ERR, \"Invalid RX queue id=%d\", rx_queue_id);\n \t\treturn 0;\n \t}\n \n@@ -2917,7 +2920,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,\n \t\tixgbe_dcb_rx_hw_config(hw, dcb_config);\n \t\tbreak;\n \tdefault:\n-\t\tPMD_INIT_LOG(ERR, \"Incorrect DCB RX mode configuration\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"Incorrect DCB RX mode configuration\");\n \t\tbreak;\n \t}\n \tswitch (dev->data->dev_conf.txmode.mq_mode) {\n@@ -2939,7 +2942,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,\n \t\tixgbe_dcb_tx_hw_config(hw, dcb_config);\n \t\tbreak;\n \tdefault:\n-\t\tPMD_INIT_LOG(ERR, \"Incorrect DCB TX mode configuration\\n\");\n+\t\tPMD_INIT_LOG(ERR, \"Incorrect DCB TX mode configuration\");\n \t\tbreak;\n \t}\n \n@@ -3210,7 +3213,7 @@ ixgbe_alloc_rx_queue_mbufs(struct igb_rx_queue *rxq)\n \t\tvolatile union ixgbe_adv_rx_desc *rxd;\n \t\tstruct rte_mbuf *mbuf = rte_rxmbuf_alloc(rxq->mb_pool);\n \t\tif (mbuf == NULL) {\n-\t\t\tPMD_INIT_LOG(ERR, \"RX mbuf alloc failed queue_id=%u\\n\",\n+\t\t\tPMD_INIT_LOG(ERR, \"RX mbuf alloc failed queue_id=%u\",\n \t\t\t\t     (unsigned) rxq->queue_id);\n \t\t\treturn (-ENOMEM);\n \t\t}\n@@ -3282,7 +3285,7 @@ ixgbe_dev_mq_rx_configure(struct rte_eth_dev *dev)\n \t\t\tIXGBE_WRITE_REG(hw, IXGBE_MRQC, IXGBE_MRQC_VMDQRT8TCEN);\n \t\t\tbreak;\n \t\tdefault:\n-\t\t\tPMD_INIT_LOG(ERR, \"invalid pool number in IOV mode\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \"invalid pool number in IOV mode\");\n \t\t}\n \t}\n \n@@ -3335,7 +3338,7 @@ ixgbe_dev_mq_tx_configure(struct rte_eth_dev *dev)\n \t\t\tbreak;\n \t\tdefault:\n \t\t\tmtqc = IXGBE_MTQC_64Q_1PB;\n-\t\t\tPMD_INIT_LOG(ERR, \"invalid pool number in IOV mode\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \"invalid pool number in IOV mode\");\n \t\t}\n \t\tIXGBE_WRITE_REG(hw, IXGBE_MTQC, mtqc);\n \t}\n@@ -3603,7 +3606,7 @@ ixgbe_setup_loopback_link_82599(struct ixgbe_hw *hw)\n \tif (ixgbe_verify_lesm_fw_enabled_82599(hw)) {\n \t\tif (hw->mac.ops.acquire_swfw_sync(hw, IXGBE_GSSR_MAC_CSR_SM) !=\n \t\t\t\tIXGBE_SUCCESS) {\n-\t\t\tPMD_INIT_LOG(ERR, \"Could not enable loopback mode\\n\");\n+\t\t\tPMD_INIT_LOG(ERR, \"Could not enable loopback mode\");\n \t\t\t/* ignore error */\n \t\t\treturn;\n \t\t}\n@@ -3699,7 +3702,7 @@ ixgbe_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id)\n \t\t/* Allocate buffers for descriptor rings */\n \t\tif (ixgbe_alloc_rx_queue_mbufs(rxq) != 0) {\n \t\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\t\"Could not alloc mbuf for queue:%d\\n\",\n+\t\t\t\t\"Could not alloc mbuf for queue:%d\",\n \t\t\t\trx_queue_id);\n \t\t\treturn -1;\n \t\t}\n@@ -3715,7 +3718,7 @@ ixgbe_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id)\n \t\t} while (--poll_ms && !(rxdctl & IXGBE_RXDCTL_ENABLE));\n \t\tif (!poll_ms)\n \t\t\tPMD_INIT_LOG(ERR, \"Could not enable \"\n-\t\t\t\t     \"Rx Queue %d\\n\", rx_queue_id);\n+\t\t\t\t     \"Rx Queue %d\", rx_queue_id);\n \t\trte_wmb();\n \t\tIXGBE_WRITE_REG(hw, IXGBE_RDH(rxq->reg_idx), 0);\n \t\tIXGBE_WRITE_REG(hw, IXGBE_RDT(rxq->reg_idx), rxq->nb_rx_desc - 1);\n@@ -3754,7 +3757,7 @@ ixgbe_dev_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id)\n \t\t} while (--poll_ms && (rxdctl | IXGBE_RXDCTL_ENABLE));\n \t\tif (!poll_ms)\n \t\t\tPMD_INIT_LOG(ERR, \"Could not disable \"\n-\t\t\t\t     \"Rx Queue %d\\n\", rx_queue_id);\n+\t\t\t\t     \"Rx Queue %d\", rx_queue_id);\n \n \t\trte_delay_us(RTE_IXGBE_WAIT_100_US);\n \n@@ -3797,7 +3800,7 @@ ixgbe_dev_tx_queue_start(struct rte_eth_dev *dev, uint16_t tx_queue_id)\n \t\t\t} while (--poll_ms && !(txdctl & IXGBE_TXDCTL_ENABLE));\n \t\t\tif (!poll_ms)\n \t\t\t\tPMD_INIT_LOG(ERR, \"Could not enable \"\n-\t\t\t\t\t     \"Tx Queue %d\\n\", tx_queue_id);\n+\t\t\t\t\t     \"Tx Queue %d\", tx_queue_id);\n \t\t}\n \t\trte_wmb();\n \t\tIXGBE_WRITE_REG(hw, IXGBE_TDH(txq->reg_idx), 0);\n@@ -3838,7 +3841,7 @@ ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)\n \t\t\t} while (--poll_ms && (txtdh != txtdt));\n \t\t\tif (!poll_ms)\n \t\t\t\tPMD_INIT_LOG(ERR,\n-\t\t\t\t\"Tx Queue %d is not empty when stopping.\\n\",\n+\t\t\t\t\"Tx Queue %d is not empty when stopping.\",\n \t\t\t\ttx_queue_id);\n \t\t}\n \n@@ -3856,7 +3859,7 @@ ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)\n \t\t\t} while (--poll_ms && (txdctl | IXGBE_TXDCTL_ENABLE));\n \t\t\tif (!poll_ms)\n \t\t\t\tPMD_INIT_LOG(ERR, \"Could not disable \"\n-\t\t\t\t\t     \"Tx Queue %d\\n\", tx_queue_id);\n+\t\t\t\t\t     \"Tx Queue %d\", tx_queue_id);\n \t\t}\n \n \t\tif (txq->ops != NULL) {\n@@ -4073,7 +4076,7 @@ ixgbevf_dev_rxtx_start(struct rte_eth_dev *dev)\n \t\t} while (--poll_ms && !(txdctl & IXGBE_TXDCTL_ENABLE));\n \t\tif (!poll_ms)\n \t\t\tPMD_INIT_LOG(ERR, \"Could not enable \"\n-\t\t\t\t\t \"Tx Queue %d\\n\", i);\n+\t\t\t\t\t \"Tx Queue %d\", i);\n \t}\n \tfor (i = 0; i < dev->data->nb_rx_queues; i++) {\n \n@@ -4091,7 +4094,7 @@ ixgbevf_dev_rxtx_start(struct rte_eth_dev *dev)\n \t\t} while (--poll_ms && !(rxdctl & IXGBE_RXDCTL_ENABLE));\n \t\tif (!poll_ms)\n \t\t\tPMD_INIT_LOG(ERR, \"Could not enable \"\n-\t\t\t\t\t \"Rx Queue %d\\n\", i);\n+\t\t\t\t\t \"Rx Queue %d\", i);\n \t\trte_wmb();\n \t\tIXGBE_WRITE_REG(hw, IXGBE_VFRDT(i), rxq->nb_rx_desc - 1);\n \n",
    "prefixes": [
        "dpdk-dev",
        "v2",
        "03/17"
    ]
}