Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/132139/?format=api
http://patches.dpdk.org/api/patches/132139/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/patch/20230928121830.62608-2-juraj.linkes@pantheon.tech/", "project": { "id": 1, "url": "http://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<20230928121830.62608-2-juraj.linkes@pantheon.tech>", "list_archive_url": "https://inbox.dpdk.org/dev/20230928121830.62608-2-juraj.linkes@pantheon.tech", "date": "2023-09-28T12:18:30", "name": "[v3,2/2] dts: reformat to 100 line length", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": true, "hash": "f01d12f864faab4e69a54b54b17895bda9f1ed52", "submitter": { "id": 1626, "url": "http://patches.dpdk.org/api/people/1626/?format=api", "name": "Juraj Linkeš", "email": "juraj.linkes@pantheon.tech" }, "delegate": { "id": 24651, "url": "http://patches.dpdk.org/api/users/24651/?format=api", "username": "dmarchand", "first_name": "David", "last_name": "Marchand", "email": "david.marchand@redhat.com" }, "mbox": "http://patches.dpdk.org/project/dpdk/patch/20230928121830.62608-2-juraj.linkes@pantheon.tech/mbox/", "series": [ { "id": 29681, "url": "http://patches.dpdk.org/api/series/29681/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=29681", "date": "2023-09-28T12:18:29", "name": "[v3,1/2] doc: increase python max line length to 100", "version": 3, "mbox": "http://patches.dpdk.org/series/29681/mbox/" } ], "comments": "http://patches.dpdk.org/api/patches/132139/comments/", "check": "success", "checks": "http://patches.dpdk.org/api/patches/132139/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 0E3BE42661;\n\tThu, 28 Sep 2023 14:18:42 +0200 (CEST)", "from mails.dpdk.org (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 3C9FE406B6;\n\tThu, 28 Sep 2023 14:18:39 +0200 (CEST)", "from mail-wr1-f52.google.com (mail-wr1-f52.google.com\n [209.85.221.52]) by mails.dpdk.org (Postfix) with ESMTP id 17DC3402CA\n for <dev@dpdk.org>; Thu, 28 Sep 2023 14:18:36 +0200 (CEST)", "by mail-wr1-f52.google.com with SMTP id\n ffacd0b85a97d-3231d67aff2so9235188f8f.0\n for <dev@dpdk.org>; Thu, 28 Sep 2023 05:18:36 -0700 (PDT)", "from jlinkes-PT-Latitude-5530.. (ip-46.34.242.20.o2inet.sk.\n [46.34.242.20]) by smtp.gmail.com with ESMTPSA id\n m26-20020a056000025a00b003233a31a467sm5716869wrz.34.2023.09.28.05.18.34\n (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256);\n Thu, 28 Sep 2023 05:18:35 -0700 (PDT)" ], "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=pantheon.tech; s=google; t=1695903516; x=1696508316; darn=dpdk.org;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:from:to:cc:subject:date\n :message-id:reply-to;\n bh=0bSPwL99MsUgX27L646/Q8kDfaQgIS2mJ/UClo9UCbk=;\n b=tZtbXtBGG+RwcdT7IE8G/y5J/YVEB1HGYODctlZptLWjI7LGL3S51MUjkhLUeX4TFn\n TX2Gq6M9qIUhE9WGEwrXn/WRuA4OC5LOBVXX9kvUPlxRrHSJ26Qy1NFAVaPyseJlxzLm\n grlu8PHTvPbwXkF+8+D7rqLQxojWxr6g0t6CXh0Jcjuc2gzwGmg5kYJprTyBkVxePsG9\n TJ5zji8U5YL5K+cboygn7DCqRoZ+HqMvTZEzztdntFHa0vXLGHQw+37etwXKmA5ChKOp\n dLjkRwNdW0iZFRimsZYZ92yX5w2NS6R5edVetEETuALdvz2s5fIZIGmGRyz8DH0N9NDI\n 5ORA==", "X-Google-DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=1e100.net; s=20230601; t=1695903516; x=1696508316;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc\n :subject:date:message-id:reply-to;\n bh=0bSPwL99MsUgX27L646/Q8kDfaQgIS2mJ/UClo9UCbk=;\n b=gKpEAxddRe0clsaw7foNjHVefep1ATg6Aqk4swT8QJ7Sp9YAJmK+Iepu1RFYmhhfPW\n StG3wDlXt/VwP99NvZ05aSTS4fbG8XGeBE+pNUfxHubl2BM3ehfhK8POi3IrJfXCZ7LB\n k5shpawkgHUZir6ws3utUeSkPcpvmdVHAfz0Khw+W8qEjz/m6boLyQwNlKTGqUh6S7Uv\n htK3Jmfoy1IMZ8wTrTOMytxxcdzA1DF3/p9fuw+MtXfrCExZieDpI9pJOA+gHEmn3yWn\n XM6tjWxoDKQxIIslW+d+XrdmKpUqKx3qVbqTfGv0TbwaiFcE3HqN8vzTfCvTeGvyEptE\n oyUg==", "X-Gm-Message-State": "AOJu0YyTAjZAc80tIDFNXL7LvtvzIkQxDIss7kcwcpBvwNGPEiQuE+C6\n HWopmnkdV6rstGPf+KTCfx9QeA==", "X-Google-Smtp-Source": "\n AGHT+IG4qXYe6LbQkKznUbCJ2Dg7DRZYu2ndUI9uckTBDnhgQ//O49xr1onSIgCrzv0AzfOTsiJctA==", "X-Received": "by 2002:a5d:45c9:0:b0:31f:f1f4:ca8b with SMTP id\n b9-20020a5d45c9000000b0031ff1f4ca8bmr1064999wrs.40.1695903515407;\n Thu, 28 Sep 2023 05:18:35 -0700 (PDT)", "From": "=?utf-8?q?Juraj_Linke=C5=A1?= <juraj.linkes@pantheon.tech>", "To": "thomas@monjalon.net, Honnappa.Nagarahalli@arm.com,\n bruce.richardson@intel.com, jspewock@iol.unh.edu, probb@iol.unh.edu,\n stephen@networkplumber.org", "Cc": "dev@dpdk.org, =?utf-8?q?Juraj_Linke=C5=A1?= <juraj.linkes@pantheon.tech>", "Subject": "[PATCH v3 2/2] dts: reformat to 100 line length", "Date": "Thu, 28 Sep 2023 14:18:30 +0200", "Message-Id": "<20230928121830.62608-2-juraj.linkes@pantheon.tech>", "X-Mailer": "git-send-email 2.34.1", "In-Reply-To": "<20230928121830.62608-1-juraj.linkes@pantheon.tech>", "References": "<20230926121013.23351-1-juraj.linkes@pantheon.tech>\n <20230928121830.62608-1-juraj.linkes@pantheon.tech>", "MIME-Version": "1.0", "Content-Type": "text/plain; charset=UTF-8", "Content-Transfer-Encoding": "8bit", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org" }, "content": "Reformat to 100 from the previous 88 to unify with C recommendations.\n\nSigned-off-by: Juraj Linkeš <juraj.linkes@pantheon.tech>\n---\n dts/framework/config/__init__.py | 20 ++-----\n dts/framework/dts.py | 15 ++---\n dts/framework/exception.py | 5 +-\n dts/framework/remote_session/__init__.py | 4 +-\n dts/framework/remote_session/linux_session.py | 39 ++++---------\n dts/framework/remote_session/posix_session.py | 30 +++-------\n .../remote/interactive_remote_session.py | 7 +--\n .../remote/interactive_shell.py | 4 +-\n .../remote_session/remote/remote_session.py | 8 +--\n .../remote_session/remote/ssh_session.py | 16 ++----\n .../remote_session/remote/testpmd_shell.py | 8 +--\n dts/framework/settings.py | 15 ++---\n dts/framework/test_result.py | 16 ++----\n dts/framework/test_suite.py | 57 +++++--------------\n .../capturing_traffic_generator.py | 7 +--\n dts/framework/testbed_model/hw/cpu.py | 20 ++-----\n dts/framework/testbed_model/node.py | 8 +--\n dts/framework/testbed_model/scapy.py | 19 ++-----\n dts/framework/testbed_model/sut_node.py | 40 ++++---------\n dts/framework/testbed_model/tg_node.py | 7 +--\n dts/framework/utils.py | 20 ++-----\n dts/tests/TestSuite_hello_world.py | 4 +-\n dts/tests/TestSuite_smoke_tests.py | 11 +---\n 23 files changed, 99 insertions(+), 281 deletions(-)", "diff": "diff --git a/dts/framework/config/__init__.py b/dts/framework/config/__init__.py\nindex cb7e00ba34..9b32cf0532 100644\n--- a/dts/framework/config/__init__.py\n+++ b/dts/framework/config/__init__.py\n@@ -140,9 +140,7 @@ def from_dict(d: dict) -> Union[\"SutNodeConfiguration\", \"TGNodeConfiguration\"]:\n \n if \"traffic_generator\" in d:\n return TGNodeConfiguration(\n- traffic_generator=TrafficGeneratorConfig.from_dict(\n- d[\"traffic_generator\"]\n- ),\n+ traffic_generator=TrafficGeneratorConfig.from_dict(d[\"traffic_generator\"]),\n **common_config,\n )\n else:\n@@ -249,9 +247,7 @@ def from_dict(\n build_targets: list[BuildTargetConfiguration] = list(\n map(BuildTargetConfiguration.from_dict, d[\"build_targets\"])\n )\n- test_suites: list[TestSuiteConfig] = list(\n- map(TestSuiteConfig.from_dict, d[\"test_suites\"])\n- )\n+ test_suites: list[TestSuiteConfig] = list(map(TestSuiteConfig.from_dict, d[\"test_suites\"]))\n sut_name = d[\"system_under_test_node\"][\"node_name\"]\n skip_smoke_tests = d.get(\"skip_smoke_tests\", False)\n assert sut_name in node_map, f\"Unknown SUT {sut_name} in execution {d}\"\n@@ -268,9 +264,7 @@ def from_dict(\n ), f\"Invalid TG configuration {traffic_generator_node}\"\n \n vdevs = (\n- d[\"system_under_test_node\"][\"vdevs\"]\n- if \"vdevs\" in d[\"system_under_test_node\"]\n- else []\n+ d[\"system_under_test_node\"][\"vdevs\"] if \"vdevs\" in d[\"system_under_test_node\"] else []\n )\n return ExecutionConfiguration(\n build_targets=build_targets,\n@@ -299,9 +293,7 @@ def from_dict(d: dict) -> \"Configuration\":\n assert len(nodes) == len(node_map), \"Duplicate node names are not allowed\"\n \n executions: list[ExecutionConfiguration] = list(\n- map(\n- ExecutionConfiguration.from_dict, d[\"executions\"], [node_map for _ in d]\n- )\n+ map(ExecutionConfiguration.from_dict, d[\"executions\"], [node_map for _ in d])\n )\n \n return Configuration(executions=executions)\n@@ -315,9 +307,7 @@ def load_config() -> Configuration:\n with open(SETTINGS.config_file_path, \"r\") as f:\n config_data = yaml.safe_load(f)\n \n- schema_path = os.path.join(\n- pathlib.Path(__file__).parent.resolve(), \"conf_yaml_schema.json\"\n- )\n+ schema_path = os.path.join(pathlib.Path(__file__).parent.resolve(), \"conf_yaml_schema.json\")\n \n with open(schema_path, \"r\") as f:\n schema = json.load(f)\ndiff --git a/dts/framework/dts.py b/dts/framework/dts.py\nindex f773f0c38d..25d6942d81 100644\n--- a/dts/framework/dts.py\n+++ b/dts/framework/dts.py\n@@ -92,9 +92,7 @@ def _run_execution(\n Run the given execution. This involves running the execution setup as well as\n running all build targets in the given execution.\n \"\"\"\n- dts_logger.info(\n- f\"Running execution with SUT '{execution.system_under_test_node.name}'.\"\n- )\n+ dts_logger.info(f\"Running execution with SUT '{execution.system_under_test_node.name}'.\")\n execution_result = result.add_execution(sut_node.config)\n execution_result.add_sut_info(sut_node.node_info)\n \n@@ -107,9 +105,7 @@ def _run_execution(\n \n else:\n for build_target in execution.build_targets:\n- _run_build_target(\n- sut_node, tg_node, build_target, execution, execution_result\n- )\n+ _run_build_target(sut_node, tg_node, build_target, execution, execution_result)\n \n finally:\n try:\n@@ -170,13 +166,10 @@ def _run_all_suites(\n execution.test_suites[:0] = [TestSuiteConfig.from_dict(\"smoke_tests\")]\n for test_suite_config in execution.test_suites:\n try:\n- _run_single_suite(\n- sut_node, tg_node, execution, build_target_result, test_suite_config\n- )\n+ _run_single_suite(sut_node, tg_node, execution, build_target_result, test_suite_config)\n except BlockingTestSuiteError as e:\n dts_logger.exception(\n- f\"An error occurred within {test_suite_config.test_suite}. \"\n- \"Skipping build target...\"\n+ f\"An error occurred within {test_suite_config.test_suite}. Skipping build target.\"\n )\n result.add_error(e)\n end_build_target = True\ndiff --git a/dts/framework/exception.py b/dts/framework/exception.py\nindex 001a5a5496..b362e42924 100644\n--- a/dts/framework/exception.py\n+++ b/dts/framework/exception.py\n@@ -116,10 +116,7 @@ def __init__(self, command: str, command_return_code: int):\n self.command_return_code = command_return_code\n \n def __str__(self) -> str:\n- return (\n- f\"Command {self.command} returned a non-zero exit code: \"\n- f\"{self.command_return_code}\"\n- )\n+ return f\"Command {self.command} returned a non-zero exit code: {self.command_return_code}\"\n \n \n class RemoteDirectoryExistsError(DTSError):\ndiff --git a/dts/framework/remote_session/__init__.py b/dts/framework/remote_session/__init__.py\nindex 00b6d1f03a..6124417bd7 100644\n--- a/dts/framework/remote_session/__init__.py\n+++ b/dts/framework/remote_session/__init__.py\n@@ -30,9 +30,7 @@\n )\n \n \n-def create_session(\n- node_config: NodeConfiguration, name: str, logger: DTSLOG\n-) -> OSSession:\n+def create_session(node_config: NodeConfiguration, name: str, logger: DTSLOG) -> OSSession:\n match node_config.os:\n case OS.linux:\n return LinuxSession(node_config, name, logger)\ndiff --git a/dts/framework/remote_session/linux_session.py b/dts/framework/remote_session/linux_session.py\nindex a3f1a6bf3b..fd877fbfae 100644\n--- a/dts/framework/remote_session/linux_session.py\n+++ b/dts/framework/remote_session/linux_session.py\n@@ -82,9 +82,7 @@ def setup_hugepages(self, hugepage_amount: int, force_first_numa: bool) -> None:\n self._mount_huge_pages()\n \n def _get_hugepage_size(self) -> int:\n- hugepage_size = self.send_command(\n- \"awk '/Hugepagesize/ {print $2}' /proc/meminfo\"\n- ).stdout\n+ hugepage_size = self.send_command(\"awk '/Hugepagesize/ {print $2}' /proc/meminfo\").stdout\n return int(hugepage_size)\n \n def _get_hugepages_total(self) -> int:\n@@ -120,13 +118,9 @@ def _supports_numa(self) -> bool:\n # there's no reason to do any numa specific configuration)\n return len(self._numa_nodes) > 1\n \n- def _configure_huge_pages(\n- self, amount: int, size: int, force_first_numa: bool\n- ) -> None:\n+ def _configure_huge_pages(self, amount: int, size: int, force_first_numa: bool) -> None:\n self._logger.info(\"Configuring Hugepages.\")\n- hugepage_config_path = (\n- f\"/sys/kernel/mm/hugepages/hugepages-{size}kB/nr_hugepages\"\n- )\n+ hugepage_config_path = f\"/sys/kernel/mm/hugepages/hugepages-{size}kB/nr_hugepages\"\n if force_first_numa and self._supports_numa():\n # clear non-numa hugepages\n self.send_command(f\"echo 0 | tee {hugepage_config_path}\", privileged=True)\n@@ -135,24 +129,18 @@ def _configure_huge_pages(\n f\"/hugepages-{size}kB/nr_hugepages\"\n )\n \n- self.send_command(\n- f\"echo {amount} | tee {hugepage_config_path}\", privileged=True\n- )\n+ self.send_command(f\"echo {amount} | tee {hugepage_config_path}\", privileged=True)\n \n def update_ports(self, ports: list[Port]) -> None:\n self._logger.debug(\"Gathering port info.\")\n for port in ports:\n- assert (\n- port.node == self.name\n- ), \"Attempted to gather port info on the wrong node\"\n+ assert port.node == self.name, \"Attempted to gather port info on the wrong node\"\n \n port_info_list = self._get_lshw_info()\n for port in ports:\n for port_info in port_info_list:\n if f\"pci@{port.pci}\" == port_info.get(\"businfo\"):\n- self._update_port_attr(\n- port, port_info.get(\"logicalname\"), \"logical_name\"\n- )\n+ self._update_port_attr(port, port_info.get(\"logicalname\"), \"logical_name\")\n self._update_port_attr(port, port_info.get(\"serial\"), \"mac_address\")\n port_info_list.remove(port_info)\n break\n@@ -163,25 +151,18 @@ def _get_lshw_info(self) -> list[LshwOutput]:\n output = self.send_command(\"lshw -quiet -json -C network\", verify=True)\n return json.loads(output.stdout)\n \n- def _update_port_attr(\n- self, port: Port, attr_value: str | None, attr_name: str\n- ) -> None:\n+ def _update_port_attr(self, port: Port, attr_value: str | None, attr_name: str) -> None:\n if attr_value:\n setattr(port, attr_name, attr_value)\n- self._logger.debug(\n- f\"Found '{attr_name}' of port {port.pci}: '{attr_value}'.\"\n- )\n+ self._logger.debug(f\"Found '{attr_name}' of port {port.pci}: '{attr_value}'.\")\n else:\n self._logger.warning(\n- f\"Attempted to get '{attr_name}' of port {port.pci}, \"\n- f\"but it doesn't exist.\"\n+ f\"Attempted to get '{attr_name}' of port {port.pci}, but it doesn't exist.\"\n )\n \n def configure_port_state(self, port: Port, enable: bool) -> None:\n state = \"up\" if enable else \"down\"\n- self.send_command(\n- f\"ip link set dev {port.logical_name} {state}\", privileged=True\n- )\n+ self.send_command(f\"ip link set dev {port.logical_name} {state}\", privileged=True)\n \n def configure_port_ip_address(\n self,\ndiff --git a/dts/framework/remote_session/posix_session.py b/dts/framework/remote_session/posix_session.py\nindex 5da0516e05..a29e2e8280 100644\n--- a/dts/framework/remote_session/posix_session.py\n+++ b/dts/framework/remote_session/posix_session.py\n@@ -94,8 +94,7 @@ def extract_remote_tarball(\n expected_dir: str | PurePath | None = None,\n ) -> None:\n self.send_command(\n- f\"tar xfm {remote_tarball_path} \"\n- f\"-C {PurePosixPath(remote_tarball_path).parent}\",\n+ f\"tar xfm {remote_tarball_path} -C {PurePosixPath(remote_tarball_path).parent}\",\n 60,\n )\n if expected_dir:\n@@ -125,8 +124,7 @@ def build_dpdk(\n self._logger.info(\"Configuring DPDK build from scratch.\")\n self.remove_remote_dir(remote_dpdk_build_dir)\n self.send_command(\n- f\"meson setup \"\n- f\"{meson_args} {remote_dpdk_dir} {remote_dpdk_build_dir}\",\n+ f\"meson setup {meson_args} {remote_dpdk_dir} {remote_dpdk_build_dir}\",\n timeout,\n verify=True,\n env=env_vars,\n@@ -140,9 +138,7 @@ def build_dpdk(\n raise DPDKBuildError(f\"DPDK build failed when doing '{e.command}'.\")\n \n def get_dpdk_version(self, build_dir: str | PurePath) -> str:\n- out = self.send_command(\n- f\"cat {self.join_remote_path(build_dir, 'VERSION')}\", verify=True\n- )\n+ out = self.send_command(f\"cat {self.join_remote_path(build_dir, 'VERSION')}\", verify=True)\n return out.stdout\n \n def kill_cleanup_dpdk_apps(self, dpdk_prefix_list: Iterable[str]) -> None:\n@@ -156,9 +152,7 @@ def kill_cleanup_dpdk_apps(self, dpdk_prefix_list: Iterable[str]) -> None:\n self._check_dpdk_hugepages(dpdk_runtime_dirs)\n self._remove_dpdk_runtime_dirs(dpdk_runtime_dirs)\n \n- def _get_dpdk_runtime_dirs(\n- self, dpdk_prefix_list: Iterable[str]\n- ) -> list[PurePosixPath]:\n+ def _get_dpdk_runtime_dirs(self, dpdk_prefix_list: Iterable[str]) -> list[PurePosixPath]:\n prefix = PurePosixPath(\"/var\", \"run\", \"dpdk\")\n if not dpdk_prefix_list:\n remote_prefixes = self._list_remote_dirs(prefix)\n@@ -174,9 +168,7 @@ def _list_remote_dirs(self, remote_path: str | PurePath) -> list[str] | None:\n Return a list of directories of the remote_dir.\n If remote_path doesn't exist, return None.\n \"\"\"\n- out = self.send_command(\n- f\"ls -l {remote_path} | awk '/^d/ {{print $NF}}'\"\n- ).stdout\n+ out = self.send_command(f\"ls -l {remote_path} | awk '/^d/ {{print $NF}}'\").stdout\n if \"No such file or directory\" in out:\n return None\n else:\n@@ -200,9 +192,7 @@ def _remote_files_exists(self, remote_path: PurePath) -> bool:\n result = self.send_command(f\"test -e {remote_path}\")\n return not result.return_code\n \n- def _check_dpdk_hugepages(\n- self, dpdk_runtime_dirs: Iterable[str | PurePath]\n- ) -> None:\n+ def _check_dpdk_hugepages(self, dpdk_runtime_dirs: Iterable[str | PurePath]) -> None:\n for dpdk_runtime_dir in dpdk_runtime_dirs:\n hugepage_info = PurePosixPath(dpdk_runtime_dir, \"hugepage_info\")\n if self._remote_files_exists(hugepage_info):\n@@ -213,9 +203,7 @@ def _check_dpdk_hugepages(\n self._logger.warning(out)\n self._logger.warning(\"*******************************************\")\n \n- def _remove_dpdk_runtime_dirs(\n- self, dpdk_runtime_dirs: Iterable[str | PurePath]\n- ) -> None:\n+ def _remove_dpdk_runtime_dirs(self, dpdk_runtime_dirs: Iterable[str | PurePath]) -> None:\n for dpdk_runtime_dir in dpdk_runtime_dirs:\n self.remove_remote_dir(dpdk_runtime_dir)\n \n@@ -245,6 +233,4 @@ def get_node_info(self) -> NodeInfo:\n SETTINGS.timeout,\n ).stdout.split(\"\\n\")\n kernel_version = self.send_command(\"uname -r\", SETTINGS.timeout).stdout\n- return NodeInfo(\n- os_release_info[0].strip(), os_release_info[1].strip(), kernel_version\n- )\n+ return NodeInfo(os_release_info[0].strip(), os_release_info[1].strip(), kernel_version)\ndiff --git a/dts/framework/remote_session/remote/interactive_remote_session.py b/dts/framework/remote_session/remote/interactive_remote_session.py\nindex 9085a668e8..098ded1bb0 100644\n--- a/dts/framework/remote_session/remote/interactive_remote_session.py\n+++ b/dts/framework/remote_session/remote/interactive_remote_session.py\n@@ -73,9 +73,7 @@ def __init__(self, node_config: NodeConfiguration, _logger: DTSLOG) -> None:\n f\"Initializing interactive connection for {self.username}@{self.hostname}\"\n )\n self._connect()\n- self._logger.info(\n- f\"Interactive connection successful for {self.username}@{self.hostname}\"\n- )\n+ self._logger.info(f\"Interactive connection successful for {self.username}@{self.hostname}\")\n \n def _connect(self) -> None:\n \"\"\"Establish a connection to the node.\n@@ -108,8 +106,7 @@ def _connect(self) -> None:\n self._logger.debug(traceback.format_exc())\n self._logger.warning(e)\n self._logger.info(\n- \"Retrying interactive session connection: \"\n- f\"retry number {retry_attempt +1}\"\n+ f\"Retrying interactive session connection: retry number {retry_attempt +1}\"\n )\n else:\n break\ndiff --git a/dts/framework/remote_session/remote/interactive_shell.py b/dts/framework/remote_session/remote/interactive_shell.py\nindex c24376b2a8..4db19fb9b3 100644\n--- a/dts/framework/remote_session/remote/interactive_shell.py\n+++ b/dts/framework/remote_session/remote/interactive_shell.py\n@@ -85,9 +85,7 @@ def __init__(\n self._app_args = app_args\n self._start_application(get_privileged_command)\n \n- def _start_application(\n- self, get_privileged_command: Callable[[str], str] | None\n- ) -> None:\n+ def _start_application(self, get_privileged_command: Callable[[str], str] | None) -> None:\n \"\"\"Starts a new interactive application based on the path to the app.\n \n This method is often overridden by subclasses as their process for\ndiff --git a/dts/framework/remote_session/remote/remote_session.py b/dts/framework/remote_session/remote/remote_session.py\nindex 0647d93de4..719f7d1ef7 100644\n--- a/dts/framework/remote_session/remote/remote_session.py\n+++ b/dts/framework/remote_session/remote/remote_session.py\n@@ -96,9 +96,7 @@ def send_command(\n If verify is True, check the return code of the executed command\n and raise a RemoteCommandExecutionError if the command failed.\n \"\"\"\n- self._logger.info(\n- f\"Sending: '{command}'\" + (f\" with env vars: '{env}'\" if env else \"\")\n- )\n+ self._logger.info(f\"Sending: '{command}'\" + (f\" with env vars: '{env}'\" if env else \"\"))\n result = self._send_command(command, timeout, env)\n if verify and result.return_code:\n self._logger.debug(\n@@ -112,9 +110,7 @@ def send_command(\n return result\n \n @abstractmethod\n- def _send_command(\n- self, command: str, timeout: float, env: dict | None\n- ) -> CommandResult:\n+ def _send_command(self, command: str, timeout: float, env: dict | None) -> CommandResult:\n \"\"\"\n Use the underlying protocol to execute the command using optional env vars\n and return CommandResult.\ndiff --git a/dts/framework/remote_session/remote/ssh_session.py b/dts/framework/remote_session/remote/ssh_session.py\nindex 8d127f1601..1a7ee649ab 100644\n--- a/dts/framework/remote_session/remote/ssh_session.py\n+++ b/dts/framework/remote_session/remote/ssh_session.py\n@@ -80,9 +80,7 @@ def _connect(self) -> None:\n if error not in errors:\n errors.append(error)\n \n- self._logger.info(\n- f\"Retrying connection: retry number {retry_attempt + 1}.\"\n- )\n+ self._logger.info(f\"Retrying connection: retry number {retry_attempt + 1}.\")\n \n else:\n break\n@@ -92,9 +90,7 @@ def _connect(self) -> None:\n def is_alive(self) -> bool:\n return self.session.is_connected\n \n- def _send_command(\n- self, command: str, timeout: float, env: dict | None\n- ) -> CommandResult:\n+ def _send_command(self, command: str, timeout: float, env: dict | None) -> CommandResult:\n \"\"\"Send a command and return the result of the execution.\n \n Args:\n@@ -107,9 +103,7 @@ def _send_command(\n SSHTimeoutError: The command execution timed out.\n \"\"\"\n try:\n- output = self.session.run(\n- command, env=env, warn=True, hide=True, timeout=timeout\n- )\n+ output = self.session.run(command, env=env, warn=True, hide=True, timeout=timeout)\n \n except (UnexpectedExit, ThreadException) as e:\n self._logger.exception(e)\n@@ -119,9 +113,7 @@ def _send_command(\n self._logger.exception(e)\n raise SSHTimeoutError(command, e.result.stderr) from e\n \n- return CommandResult(\n- self.name, command, output.stdout, output.stderr, output.return_code\n- )\n+ return CommandResult(self.name, command, output.stdout, output.stderr, output.return_code)\n \n def copy_from(\n self,\ndiff --git a/dts/framework/remote_session/remote/testpmd_shell.py b/dts/framework/remote_session/remote/testpmd_shell.py\nindex 1455b5a199..08ac311016 100644\n--- a/dts/framework/remote_session/remote/testpmd_shell.py\n+++ b/dts/framework/remote_session/remote/testpmd_shell.py\n@@ -21,13 +21,9 @@ class TestPmdShell(InteractiveShell):\n path: PurePath = PurePath(\"app\", \"dpdk-testpmd\")\n dpdk_app: bool = True\n _default_prompt: str = \"testpmd>\"\n- _command_extra_chars: str = (\n- \"\\n\" # We want to append an extra newline to every command\n- )\n+ _command_extra_chars: str = \"\\n\" # We want to append an extra newline to every command\n \n- def _start_application(\n- self, get_privileged_command: Callable[[str], str] | None\n- ) -> None:\n+ def _start_application(self, get_privileged_command: Callable[[str], str] | None) -> None:\n \"\"\"See \"_start_application\" in InteractiveShell.\"\"\"\n self._app_args += \" -- -i\"\n super()._start_application(get_privileged_command)\ndiff --git a/dts/framework/settings.py b/dts/framework/settings.py\nindex cfa39d011b..974793a11a 100644\n--- a/dts/framework/settings.py\n+++ b/dts/framework/settings.py\n@@ -72,9 +72,8 @@ class _Settings:\n \n def _get_parser() -> argparse.ArgumentParser:\n parser = argparse.ArgumentParser(\n- description=\"Run DPDK test suites. All options may be specified with \"\n- \"the environment variables provided in brackets. \"\n- \"Command line arguments have higher priority.\",\n+ description=\"Run DPDK test suites. All options may be specified with the environment \"\n+ \"variables provided in brackets. Command line arguments have higher priority.\",\n formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n )\n \n@@ -82,8 +81,7 @@ def _get_parser() -> argparse.ArgumentParser:\n \"--config-file\",\n action=_env_arg(\"DTS_CFG_FILE\"),\n default=\"conf.yaml\",\n- help=\"[DTS_CFG_FILE] configuration file that describes the test cases, SUTs \"\n- \"and targets.\",\n+ help=\"[DTS_CFG_FILE] configuration file that describes the test cases, SUTs and targets.\",\n )\n \n parser.add_argument(\n@@ -100,8 +98,7 @@ def _get_parser() -> argparse.ArgumentParser:\n action=_env_arg(\"DTS_TIMEOUT\"),\n default=15,\n type=float,\n- help=\"[DTS_TIMEOUT] The default timeout for all DTS operations except for \"\n- \"compiling DPDK.\",\n+ help=\"[DTS_TIMEOUT] The default timeout for all DTS operations except for compiling DPDK.\",\n )\n \n parser.add_argument(\n@@ -170,9 +167,7 @@ def _get_settings() -> _Settings:\n timeout=parsed_args.timeout,\n verbose=(parsed_args.verbose == \"Y\"),\n skip_setup=(parsed_args.skip_setup == \"Y\"),\n- dpdk_tarball_path=Path(\n- DPDKGitTarball(parsed_args.tarball, parsed_args.output_dir)\n- )\n+ dpdk_tarball_path=Path(DPDKGitTarball(parsed_args.tarball, parsed_args.output_dir))\n if not os.path.exists(parsed_args.tarball)\n else Path(parsed_args.tarball),\n compile_timeout=parsed_args.compile_timeout,\ndiff --git a/dts/framework/test_result.py b/dts/framework/test_result.py\nindex f0fbe80f6f..4c2e7e2418 100644\n--- a/dts/framework/test_result.py\n+++ b/dts/framework/test_result.py\n@@ -83,9 +83,7 @@ def __iadd__(self, other: Result) -> \"Statistics\":\n \"\"\"\n self[other.name] += 1\n self[\"PASS RATE\"] = (\n- float(self[Result.PASS.name])\n- * 100\n- / sum(self[result.name] for result in Result)\n+ float(self[Result.PASS.name]) * 100 / sum(self[result.name] for result in Result)\n )\n return self\n \n@@ -135,9 +133,7 @@ def _get_setup_teardown_errors(self) -> list[Exception]:\n \n def _get_inner_errors(self) -> list[Exception]:\n return [\n- error\n- for inner_result in self._inner_results\n- for error in inner_result.get_errors()\n+ error for inner_result in self._inner_results for error in inner_result.get_errors()\n ]\n \n def get_errors(self) -> list[Exception]:\n@@ -174,9 +170,7 @@ def add_stats(self, statistics: Statistics) -> None:\n statistics += self.result\n \n def __bool__(self) -> bool:\n- return (\n- bool(self.setup_result) and bool(self.teardown_result) and bool(self.result)\n- )\n+ return bool(self.setup_result) and bool(self.teardown_result) and bool(self.result)\n \n \n class TestSuiteResult(BaseResult):\n@@ -247,9 +241,7 @@ def __init__(self, sut_node: NodeConfiguration):\n super(ExecutionResult, self).__init__()\n self.sut_node = sut_node\n \n- def add_build_target(\n- self, build_target: BuildTargetConfiguration\n- ) -> BuildTargetResult:\n+ def add_build_target(self, build_target: BuildTargetConfiguration) -> BuildTargetResult:\n build_target_result = BuildTargetResult(build_target)\n self._inner_results.append(build_target_result)\n return build_target_result\ndiff --git a/dts/framework/test_suite.py b/dts/framework/test_suite.py\nindex 3b890c0451..4a7907ec33 100644\n--- a/dts/framework/test_suite.py\n+++ b/dts/framework/test_suite.py\n@@ -102,9 +102,7 @@ def _process_links(self) -> None:\n tg_port.peer,\n tg_port.identifier,\n ):\n- self._port_links.append(\n- PortLink(sut_port=sut_port, tg_port=tg_port)\n- )\n+ self._port_links.append(PortLink(sut_port=sut_port, tg_port=tg_port))\n \n def set_up_suite(self) -> None:\n \"\"\"\n@@ -151,9 +149,7 @@ def configure_testbed_ipv4(self, restore: bool = False) -> None:\n def _configure_ipv4_forwarding(self, enable: bool) -> None:\n self.sut_node.configure_ipv4_forwarding(enable)\n \n- def send_packet_and_capture(\n- self, packet: Packet, duration: float = 1\n- ) -> list[Packet]:\n+ def send_packet_and_capture(self, packet: Packet, duration: float = 1) -> list[Packet]:\n \"\"\"\n Send a packet through the appropriate interface and\n receive on the appropriate interface.\n@@ -202,21 +198,15 @@ def verify(self, condition: bool, failure_description: str) -> None:\n self._fail_test_case_verify(failure_description)\n \n def _fail_test_case_verify(self, failure_description: str) -> None:\n- self._logger.debug(\n- \"A test case failed, showing the last 10 commands executed on SUT:\"\n- )\n+ self._logger.debug(\"A test case failed, showing the last 10 commands executed on SUT:\")\n for command_res in self.sut_node.main_session.remote_session.history[-10:]:\n self._logger.debug(command_res.command)\n- self._logger.debug(\n- \"A test case failed, showing the last 10 commands executed on TG:\"\n- )\n+ self._logger.debug(\"A test case failed, showing the last 10 commands executed on TG:\")\n for command_res in self.tg_node.main_session.remote_session.history[-10:]:\n self._logger.debug(command_res.command)\n raise TestCaseVerifyError(failure_description)\n \n- def verify_packets(\n- self, expected_packet: Packet, received_packets: list[Packet]\n- ) -> None:\n+ def verify_packets(self, expected_packet: Packet, received_packets: list[Packet]) -> None:\n for received_packet in received_packets:\n if self._compare_packets(expected_packet, received_packet):\n break\n@@ -225,17 +215,11 @@ def verify_packets(\n f\"The expected packet {get_packet_summaries(expected_packet)} \"\n f\"not found among received {get_packet_summaries(received_packets)}\"\n )\n- self._fail_test_case_verify(\n- \"An expected packet not found among received packets.\"\n- )\n+ self._fail_test_case_verify(\"An expected packet not found among received packets.\")\n \n- def _compare_packets(\n- self, expected_packet: Packet, received_packet: Packet\n- ) -> bool:\n+ def _compare_packets(self, expected_packet: Packet, received_packet: Packet) -> bool:\n self._logger.debug(\n- \"Comparing packets: \\n\"\n- f\"{expected_packet.summary()}\\n\"\n- f\"{received_packet.summary()}\"\n+ f\"Comparing packets: \\n{expected_packet.summary()}\\n{received_packet.summary()}\"\n )\n \n l3 = IP in expected_packet.layers()\n@@ -262,14 +246,10 @@ def _compare_packets(\n expected_payload = expected_payload.payload\n \n if expected_payload:\n- self._logger.debug(\n- f\"The expected packet did not contain {expected_payload}.\"\n- )\n+ self._logger.debug(f\"The expected packet did not contain {expected_payload}.\")\n return False\n if received_payload and received_payload.__class__ != Padding:\n- self._logger.debug(\n- \"The received payload had extra layers which were not padding.\"\n- )\n+ self._logger.debug(\"The received payload had extra layers which were not padding.\")\n return False\n return True\n \n@@ -296,10 +276,7 @@ def _verify_l2_frame(self, received_packet: Ether, l3: bool) -> bool:\n \n def _verify_l3_packet(self, received_packet: IP, expected_packet: IP) -> bool:\n self._logger.debug(\"Looking at the IP layer.\")\n- if (\n- received_packet.src != expected_packet.src\n- or received_packet.dst != expected_packet.dst\n- ):\n+ if received_packet.src != expected_packet.src or received_packet.dst != expected_packet.dst:\n return False\n return True\n \n@@ -373,9 +350,7 @@ def _get_test_cases(self, test_case_regex: str) -> list[MethodType]:\n if self._should_be_executed(test_case_name, test_case_regex):\n filtered_test_cases.append(test_case)\n cases_str = \", \".join((x.__name__ for x in filtered_test_cases))\n- self._logger.debug(\n- f\"Found test cases '{cases_str}' in {self.__class__.__name__}.\"\n- )\n+ self._logger.debug(f\"Found test cases '{cases_str}' in {self.__class__.__name__}.\")\n return filtered_test_cases\n \n def _should_be_executed(self, test_case_name: str, test_case_regex: str) -> bool:\n@@ -445,9 +420,7 @@ def _execute_test_case(\n self._logger.exception(f\"Test case execution ERROR: {test_case_name}\")\n test_case_result.update(Result.ERROR, e)\n except KeyboardInterrupt:\n- self._logger.error(\n- f\"Test case execution INTERRUPTED by user: {test_case_name}\"\n- )\n+ self._logger.error(f\"Test case execution INTERRUPTED by user: {test_case_name}\")\n test_case_result.update(Result.SKIP)\n raise KeyboardInterrupt(\"Stop DTS\")\n \n@@ -464,9 +437,7 @@ def is_test_suite(object) -> bool:\n try:\n testcase_module = importlib.import_module(testsuite_module_path)\n except ModuleNotFoundError as e:\n- raise ConfigurationError(\n- f\"Test suite '{testsuite_module_path}' not found.\"\n- ) from e\n+ raise ConfigurationError(f\"Test suite '{testsuite_module_path}' not found.\") from e\n return [\n test_suite_class\n for _, test_suite_class in inspect.getmembers(testcase_module, is_test_suite)\ndiff --git a/dts/framework/testbed_model/capturing_traffic_generator.py b/dts/framework/testbed_model/capturing_traffic_generator.py\nindex ab98987f8e..e6512061d7 100644\n--- a/dts/framework/testbed_model/capturing_traffic_generator.py\n+++ b/dts/framework/testbed_model/capturing_traffic_generator.py\n@@ -100,8 +100,7 @@ def send_packets_and_capture(\n \"\"\"\n self._logger.debug(get_packet_summaries(packets))\n self._logger.debug(\n- f\"Sending packet on {send_port.logical_name}, \"\n- f\"receiving on {receive_port.logical_name}.\"\n+ f\"Sending packet on {send_port.logical_name}, receiving on {receive_port.logical_name}.\"\n )\n received_packets = self._send_packets_and_capture(\n packets,\n@@ -110,9 +109,7 @@ def send_packets_and_capture(\n duration,\n )\n \n- self._logger.debug(\n- f\"Received packets: {get_packet_summaries(received_packets)}\"\n- )\n+ self._logger.debug(f\"Received packets: {get_packet_summaries(received_packets)}\")\n self._write_capture_from_packets(capture_name, received_packets)\n return received_packets\n \ndiff --git a/dts/framework/testbed_model/hw/cpu.py b/dts/framework/testbed_model/hw/cpu.py\nindex d1918a12dc..cbc5fe7fff 100644\n--- a/dts/framework/testbed_model/hw/cpu.py\n+++ b/dts/framework/testbed_model/hw/cpu.py\n@@ -54,9 +54,7 @@ def __init__(self, lcore_list: list[int] | list[str] | list[LogicalCore] | str):\n \n # the input lcores may not be sorted\n self._lcore_list.sort()\n- self._lcore_str = (\n- f'{\",\".join(self._get_consecutive_lcores_range(self._lcore_list))}'\n- )\n+ self._lcore_str = f'{\",\".join(self._get_consecutive_lcores_range(self._lcore_list))}'\n \n @property\n def lcore_list(self) -> list[int]:\n@@ -70,15 +68,11 @@ def _get_consecutive_lcores_range(self, lcore_ids_list: list[int]) -> list[str]:\n segment.append(lcore_id)\n else:\n formatted_core_list.append(\n- f\"{segment[0]}-{segment[-1]}\"\n- if len(segment) > 1\n- else f\"{segment[0]}\"\n+ f\"{segment[0]}-{segment[-1]}\" if len(segment) > 1 else f\"{segment[0]}\"\n )\n current_core_index = lcore_ids_list.index(lcore_id)\n formatted_core_list.extend(\n- self._get_consecutive_lcores_range(\n- lcore_ids_list[current_core_index:]\n- )\n+ self._get_consecutive_lcores_range(lcore_ids_list[current_core_index:])\n )\n segment.clear()\n break\n@@ -125,9 +119,7 @@ def __init__(\n self._filter_specifier = filter_specifier\n \n # sorting by core is needed in case hyperthreading is enabled\n- self._lcores_to_filter = sorted(\n- lcore_list, key=lambda x: x.core, reverse=not ascending\n- )\n+ self._lcores_to_filter = sorted(lcore_list, key=lambda x: x.core, reverse=not ascending)\n self.filter()\n \n @abstractmethod\n@@ -220,9 +212,7 @@ def _filter_cores_from_socket(\n else:\n # we have enough lcores per this core\n continue\n- elif self._filter_specifier.cores_per_socket > len(\n- lcore_count_per_core_map\n- ):\n+ elif self._filter_specifier.cores_per_socket > len(lcore_count_per_core_map):\n # only add cores if we need more\n lcore_count_per_core_map[lcore.core] = 1\n filtered_lcores.append(lcore)\ndiff --git a/dts/framework/testbed_model/node.py b/dts/framework/testbed_model/node.py\nindex fc01e0bf8e..ef700d8114 100644\n--- a/dts/framework/testbed_model/node.py\n+++ b/dts/framework/testbed_model/node.py\n@@ -103,18 +103,14 @@ def _tear_down_execution(self) -> None:\n is not decorated so that the derived class doesn't have to use the decorator.\n \"\"\"\n \n- def set_up_build_target(\n- self, build_target_config: BuildTargetConfiguration\n- ) -> None:\n+ def set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None:\n \"\"\"\n Perform the build target setup that will be done for each build target\n tested on this node.\n \"\"\"\n self._set_up_build_target(build_target_config)\n \n- def _set_up_build_target(\n- self, build_target_config: BuildTargetConfiguration\n- ) -> None:\n+ def _set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None:\n \"\"\"\n This method exists to be optionally overwritten by derived classes and\n is not decorated so that the derived class doesn't have to use the decorator.\ndiff --git a/dts/framework/testbed_model/scapy.py b/dts/framework/testbed_model/scapy.py\nindex af0d4dbb25..9083e92b3d 100644\n--- a/dts/framework/testbed_model/scapy.py\n+++ b/dts/framework/testbed_model/scapy.py\n@@ -96,9 +96,7 @@ def scapy_send_packets_and_capture(\n return [scapy_packet.build() for scapy_packet in sniffer.stop(join=True)]\n \n \n-def scapy_send_packets(\n- xmlrpc_packets: list[xmlrpc.client.Binary], send_iface: str\n-) -> None:\n+def scapy_send_packets(xmlrpc_packets: list[xmlrpc.client.Binary], send_iface: str) -> None:\n \"\"\"RPC function to send packets.\n \n The function is meant to be executed on the remote TG node.\n@@ -197,9 +195,7 @@ class ScapyTrafficGenerator(CapturingTrafficGenerator):\n def __init__(self, tg_node: TGNode, config: ScapyTrafficGeneratorConfig):\n self._config = config\n self._tg_node = tg_node\n- self._logger = getLogger(\n- f\"{self._tg_node.name} {self._config.traffic_generator_type}\"\n- )\n+ self._logger = getLogger(f\"{self._tg_node.name} {self._config.traffic_generator_type}\")\n \n assert (\n self._tg_node.config.os == OS.linux\n@@ -218,9 +214,7 @@ def __init__(self, tg_node: TGNode, config: ScapyTrafficGeneratorConfig):\n self._start_xmlrpc_server_in_remote_python(xmlrpc_server_listen_port)\n \n # connect to the server\n- server_url = (\n- f\"http://{self._tg_node.config.hostname}:{xmlrpc_server_listen_port}\"\n- )\n+ server_url = f\"http://{self._tg_node.config.hostname}:{xmlrpc_server_listen_port}\"\n self.rpc_server_proxy = xmlrpc.client.ServerProxy(\n server_url, allow_none=True, verbose=SETTINGS.verbose\n )\n@@ -240,17 +234,14 @@ def _start_xmlrpc_server_in_remote_python(self, listen_port: int):\n src = inspect.getsource(QuittableXMLRPCServer)\n # Lines with only whitespace break the repl if in the middle of a function\n # or class, so strip all lines containing only whitespace\n- src = \"\\n\".join(\n- [line for line in src.splitlines() if not line.isspace() and line != \"\"]\n- )\n+ src = \"\\n\".join([line for line in src.splitlines() if not line.isspace() and line != \"\"])\n \n spacing = \"\\n\" * 4\n \n # execute it in the python terminal\n self.session.send_command(spacing + src + spacing)\n self.session.send_command(\n- f\"server = QuittableXMLRPCServer(('0.0.0.0', {listen_port}));\"\n- f\"server.serve_forever()\",\n+ f\"server = QuittableXMLRPCServer(('0.0.0.0', {listen_port}));server.serve_forever()\",\n \"XMLRPC OK\",\n )\n \ndiff --git a/dts/framework/testbed_model/sut_node.py b/dts/framework/testbed_model/sut_node.py\nindex 202aebfd06..d5d8905be7 100644\n--- a/dts/framework/testbed_model/sut_node.py\n+++ b/dts/framework/testbed_model/sut_node.py\n@@ -129,9 +129,7 @@ def remote_dpdk_build_dir(self) -> PurePath:\n @property\n def dpdk_version(self) -> str:\n if self._dpdk_version is None:\n- self._dpdk_version = self.main_session.get_dpdk_version(\n- self._remote_dpdk_dir\n- )\n+ self._dpdk_version = self.main_session.get_dpdk_version(self._remote_dpdk_dir)\n return self._dpdk_version\n \n @property\n@@ -149,8 +147,7 @@ def compiler_version(self) -> str:\n )\n else:\n self._logger.warning(\n- \"Failed to get compiler version because\"\n- \"_build_target_config is None.\"\n+ \"Failed to get compiler version because _build_target_config is None.\"\n )\n return \"\"\n return self._compiler_version\n@@ -163,9 +160,7 @@ def get_build_target_info(self) -> BuildTargetInfo:\n def _guess_dpdk_remote_dir(self) -> PurePath:\n return self.main_session.guess_dpdk_remote_dir(self._remote_tmp_dir)\n \n- def _set_up_build_target(\n- self, build_target_config: BuildTargetConfiguration\n- ) -> None:\n+ def _set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None:\n \"\"\"\n Setup DPDK on the SUT node.\n \"\"\"\n@@ -177,23 +172,18 @@ def _set_up_build_target(\n self._copy_dpdk_tarball()\n self._build_dpdk()\n \n- def _configure_build_target(\n- self, build_target_config: BuildTargetConfiguration\n- ) -> None:\n+ def _configure_build_target(self, build_target_config: BuildTargetConfiguration) -> None:\n \"\"\"\n Populate common environment variables and set build target config.\n \"\"\"\n self._env_vars = {}\n self._build_target_config = build_target_config\n- self._env_vars.update(\n- self.main_session.get_dpdk_build_env_vars(build_target_config.arch)\n- )\n+ self._env_vars.update(self.main_session.get_dpdk_build_env_vars(build_target_config.arch))\n self._env_vars[\"CC\"] = build_target_config.compiler.name\n if build_target_config.compiler_wrapper:\n self._env_vars[\"CC\"] = (\n- f\"'{build_target_config.compiler_wrapper} \"\n- f\"{build_target_config.compiler.name}'\"\n- )\n+ f\"'{build_target_config.compiler_wrapper} {build_target_config.compiler.name}'\"\n+ ) # fmt: skip\n \n @Node.skip_setup\n def _copy_dpdk_tarball(self) -> None:\n@@ -224,9 +214,7 @@ def _copy_dpdk_tarball(self) -> None:\n self.main_session.remove_remote_dir(self._remote_dpdk_dir)\n \n # then extract to remote path\n- self.main_session.extract_remote_tarball(\n- remote_tarball_path, self._remote_dpdk_dir\n- )\n+ self.main_session.extract_remote_tarball(remote_tarball_path, self._remote_dpdk_dir)\n \n @Node.skip_setup\n def _build_dpdk(self) -> None:\n@@ -263,9 +251,7 @@ def build_dpdk_app(self, app_name: str, **meson_dpdk_args: str | bool) -> PurePa\n )\n \n if app_name == \"all\":\n- return self.main_session.join_remote_path(\n- self.remote_dpdk_build_dir, \"examples\"\n- )\n+ return self.main_session.join_remote_path(self.remote_dpdk_build_dir, \"examples\")\n return self.main_session.join_remote_path(\n self.remote_dpdk_build_dir, \"examples\", f\"dpdk-{app_name}\"\n )\n@@ -319,9 +305,7 @@ def create_eal_parameters(\n '-c 0xf -a 0000:88:00.0 --file-prefix=dpdk_1112_20190809143420';\n \"\"\"\n \n- lcore_list = LogicalCoreList(\n- self.filter_lcores(lcore_filter_specifier, ascending_cores)\n- )\n+ lcore_list = LogicalCoreList(self.filter_lcores(lcore_filter_specifier, ascending_cores))\n \n if append_prefix_timestamp:\n prefix = f\"{prefix}_{self._dpdk_timestamp}\"\n@@ -386,6 +370,4 @@ def create_interactive_shell(\n self.remote_dpdk_build_dir, shell_cls.path\n )\n \n- return super().create_interactive_shell(\n- shell_cls, timeout, privileged, str(eal_parameters)\n- )\n+ return super().create_interactive_shell(shell_cls, timeout, privileged, str(eal_parameters))\ndiff --git a/dts/framework/testbed_model/tg_node.py b/dts/framework/testbed_model/tg_node.py\nindex 27025cfa31..79a55663b5 100644\n--- a/dts/framework/testbed_model/tg_node.py\n+++ b/dts/framework/testbed_model/tg_node.py\n@@ -45,9 +45,7 @@ class TGNode(Node):\n \n def __init__(self, node_config: TGNodeConfiguration):\n super(TGNode, self).__init__(node_config)\n- self.traffic_generator = create_traffic_generator(\n- self, node_config.traffic_generator\n- )\n+ self.traffic_generator = create_traffic_generator(self, node_config.traffic_generator)\n self._logger.info(f\"Created node: {self.name}\")\n \n def send_packet_and_capture(\n@@ -94,6 +92,5 @@ def create_traffic_generator(\n return ScapyTrafficGenerator(tg_node, traffic_generator_config)\n case _:\n raise ConfigurationError(\n- \"Unknown traffic generator: \"\n- f\"{traffic_generator_config.traffic_generator_type}\"\n+ f\"Unknown traffic generator: {traffic_generator_config.traffic_generator_type}\"\n )\ndiff --git a/dts/framework/utils.py b/dts/framework/utils.py\nindex d27c2c5b5f..d098d364ff 100644\n--- a/dts/framework/utils.py\n+++ b/dts/framework/utils.py\n@@ -19,9 +19,7 @@\n \n class StrEnum(Enum):\n @staticmethod\n- def _generate_next_value_(\n- name: str, start: int, count: int, last_values: object\n- ) -> str:\n+ def _generate_next_value_(name: str, start: int, count: int, last_values: object) -> str:\n return name\n \n def __str__(self) -> str:\n@@ -32,9 +30,7 @@ def __str__(self) -> str:\n \n \n def check_dts_python_version() -> None:\n- if sys.version_info.major < 3 or (\n- sys.version_info.major == 3 and sys.version_info.minor < 10\n- ):\n+ if sys.version_info.major < 3 or (sys.version_info.major == 3 and sys.version_info.minor < 10):\n print(\n RED(\n (\n@@ -60,9 +56,7 @@ def expand_range(range_str: str) -> list[int]:\n range_boundaries = range_str.split(\"-\")\n # will throw an exception when items in range_boundaries can't be converted,\n # serving as type check\n- expanded_range.extend(\n- range(int(range_boundaries[0]), int(range_boundaries[-1]) + 1)\n- )\n+ expanded_range.extend(range(int(range_boundaries[0]), int(range_boundaries[-1]) + 1))\n \n return expanded_range\n \n@@ -71,9 +65,7 @@ def get_packet_summaries(packets: list[Packet]):\n if len(packets) == 1:\n packet_summaries = packets[0].summary()\n else:\n- packet_summaries = json.dumps(\n- list(map(lambda pkt: pkt.summary(), packets)), indent=4\n- )\n+ packet_summaries = json.dumps(list(map(lambda pkt: pkt.summary(), packets)), indent=4)\n return f\"Packet contents: \\n{packet_summaries}\"\n \n \n@@ -94,9 +86,7 @@ class MesonArgs(object):\n _default_library: str\n \n def __init__(self, default_library: str | None = None, **dpdk_args: str | bool):\n- self._default_library = (\n- f\"--default-library={default_library}\" if default_library else \"\"\n- )\n+ self._default_library = f\"--default-library={default_library}\" if default_library else \"\"\n self._dpdk_args = \" \".join(\n (\n f\"-D{dpdk_arg_name}={dpdk_arg_value}\"\ndiff --git a/dts/tests/TestSuite_hello_world.py b/dts/tests/TestSuite_hello_world.py\nindex 7e3d95c0cf..768ba1cfa8 100644\n--- a/dts/tests/TestSuite_hello_world.py\n+++ b/dts/tests/TestSuite_hello_world.py\n@@ -34,9 +34,7 @@ def test_hello_world_single_core(self) -> None:\n # get the first usable core\n lcore_amount = LogicalCoreCount(1, 1, 1)\n lcores = LogicalCoreCountFilter(self.sut_node.lcores, lcore_amount).filter()\n- eal_para = self.sut_node.create_eal_parameters(\n- lcore_filter_specifier=lcore_amount\n- )\n+ eal_para = self.sut_node.create_eal_parameters(lcore_filter_specifier=lcore_amount)\n result = self.sut_node.run_dpdk_app(self.app_helloworld_path, eal_para)\n self.verify(\n f\"hello from core {int(lcores[0])}\" in result.stdout,\ndiff --git a/dts/tests/TestSuite_smoke_tests.py b/dts/tests/TestSuite_smoke_tests.py\nindex 4a269df75b..670053c6a5 100644\n--- a/dts/tests/TestSuite_smoke_tests.py\n+++ b/dts/tests/TestSuite_smoke_tests.py\n@@ -45,13 +45,10 @@ def test_driver_tests(self) -> None:\n for dev in self.sut_node.virtual_devices:\n vdev_args += f\"--vdev {dev} \"\n vdev_args = vdev_args[:-1]\n- driver_tests_command = (\n- f\"meson test -C {self.dpdk_build_dir_path} --suite driver-tests\"\n- )\n+ driver_tests_command = f\"meson test -C {self.dpdk_build_dir_path} --suite driver-tests\"\n if vdev_args:\n self._logger.info(\n- \"Running driver tests with the following virtual \"\n- f\"devices: {vdev_args}\"\n+ f\"Running driver tests with the following virtual devices: {vdev_args}\"\n )\n driver_tests_command += f' --test-args \"{vdev_args}\"'\n \n@@ -67,9 +64,7 @@ def test_devices_listed_in_testpmd(self) -> None:\n Test:\n Uses testpmd driver to verify that devices have been found by testpmd.\n \"\"\"\n- testpmd_driver = self.sut_node.create_interactive_shell(\n- TestPmdShell, privileged=True\n- )\n+ testpmd_driver = self.sut_node.create_interactive_shell(TestPmdShell, privileged=True)\n dev_list = [str(x) for x in testpmd_driver.get_devices()]\n for nic in self.nics_in_node:\n self.verify(\n", "prefixes": [ "v3", "2/2" ] }{ "id": 132139, "url": "