Just a few more comments.

On Mon, Jul 10, 2023 at 6:23 PM <jspew...@iol.unh.edu> wrote:
>
> From: Jeremy Spewock <jspew...@iol.unh.edu>
>
> Adds a new test suite for running smoke tests that verify general
> configuration aspects of the system under test. If any of these tests
> fail, the DTS execution terminates as part of a "fail-fast" model.
>
> Signed-off-by: Jeremy Spewock <jspew...@iol.unh.edu>
> ---
>  dts/conf.yaml                                 |  17 +-
>  dts/framework/config/__init__.py              | 116 +++++++++--
>  dts/framework/config/conf_yaml_schema.json    | 142 +++++++++++++-
>  dts/framework/dts.py                          |  88 ++++++---
>  dts/framework/exception.py                    |  12 ++
>  dts/framework/remote_session/__init__.py      |  10 +-
>  dts/framework/remote_session/os_session.py    |  24 ++-
>  dts/framework/remote_session/posix_session.py |  29 ++-
>  .../remote_session/remote/__init__.py         |  10 +
>  .../remote/interactive_remote_session.py      | 118 ++++++++++++
>  .../remote/interactive_shell.py               |  99 ++++++++++
>  .../remote_session/remote/testpmd_shell.py    |  67 +++++++
>  dts/framework/test_result.py                  |  37 +++-
>  dts/framework/test_suite.py                   |  21 +-
>  dts/framework/testbed_model/node.py           |   2 +
>  dts/framework/testbed_model/sut_node.py       | 180 +++++++++++++-----
>  dts/tests/TestSuite_smoke_tests.py            | 118 ++++++++++++
>  17 files changed, 994 insertions(+), 96 deletions(-)
>  create mode 100644 
> dts/framework/remote_session/remote/interactive_remote_session.py
>  create mode 100644 dts/framework/remote_session/remote/interactive_shell.py
>  create mode 100644 dts/framework/remote_session/remote/testpmd_shell.py
>  create mode 100644 dts/tests/TestSuite_smoke_tests.py
>
> diff --git a/dts/conf.yaml b/dts/conf.yaml
> index a9bd8a3e..2717de13 100644
> --- a/dts/conf.yaml
> +++ b/dts/conf.yaml
> @@ -10,9 +10,13 @@ executions:
>          compiler_wrapper: ccache
>      perf: false
>      func: true
> +    skip_smoke_tests: false #optional flag that allow you to ski smoke tests

Typo: ski
Also put a space after #

>      test_suites:
>        - hello_world
> -    system_under_test: "SUT 1"
> +    system_under_test:
> +      node_name: "SUT 1"
> +      vdevs: #optional: if removed vdevs won't be used in the execution

Missing space after #
The sentence after hugepages has a comma in it, let's unify those.

> +        - "crypto_openssl"
>  nodes:
>    - name: "SUT 1"
>      hostname: sut1.change.me.localhost
> @@ -20,6 +24,17 @@ nodes:
>      arch: x86_64
>      os: linux
>      lcores: ""
> +    ports:

I'm comparing my version with this patch and I've just noticed this -
let's put the ports at the end (after hugepages). This way we'll have
the configuration sorted into sections of sorts:
Cores/cpu config
Memory config
Port/devices config

> +      - pci: "0000:00:08.0"
> +        os_driver_for_dpdk: vfio-pci #OS driver that DPDK will use

Missing space after #

> +        os_driver: i40e
> +        peer_node: "TG 1"
> +        peer_pci: "0000:00:08.0"
> +      - pci: "0000:00:08.1"
> +        os_driver_for_dpdk: vfio-pci
> +        os_driver: i40e
> +        peer_node: "TG 1"
> +        peer_pci: "0000:00:08.1"
>      use_first_core: false
>      memory_channels: 4
>      hugepages:  # optional; if removed, will use system hugepage 
> configuration
> diff --git a/dts/framework/config/__init__.py 
> b/dts/framework/config/__init__.py
> index ebb0823f..75ac1cbe 100644
> --- a/dts/framework/config/__init__.py
> +++ b/dts/framework/config/__init__.py
> @@ -12,6 +12,7 @@
>  import pathlib
>  from dataclasses import dataclass
>  from enum import Enum, auto, unique
> +from pathlib import PurePath
>  from typing import Any, TypedDict
>
>  import warlock  # type: ignore
> @@ -72,6 +73,20 @@ class HugepageConfiguration:
>      force_first_numa: bool
>
>
> +@dataclass(slots=True, frozen=True)
> +class PortConfig:
> +    node: str
> +    pci: str
> +    os_driver_for_dpdk: str
> +    os_driver: str
> +    peer_node: str
> +    peer_pci: str
> +
> +    @staticmethod
> +    def from_dict(node: str, d: dict) -> "PortConfig":
> +        return PortConfig(node=node, **d)
> +
> +
>  @dataclass(slots=True, frozen=True)
>  class NodeConfiguration:
>      name: str
> @@ -84,6 +99,7 @@ class NodeConfiguration:
>      use_first_core: bool
>      memory_channels: int
>      hugepages: HugepageConfiguration | None
> +    ports: list[PortConfig]
>
>      @staticmethod
>      def from_dict(d: dict) -> "NodeConfiguration":
> @@ -92,18 +108,43 @@ def from_dict(d: dict) -> "NodeConfiguration":
>              if "force_first_numa" not in hugepage_config:
>                  hugepage_config["force_first_numa"] = False
>              hugepage_config = HugepageConfiguration(**hugepage_config)
> +        common_config = {
> +            "name": d["name"],
> +            "hostname": d["hostname"],
> +            "user": d["user"],
> +            "password": d.get("password"),
> +            "arch": Architecture(d["arch"]),
> +            "os": OS(d["os"]),
> +            "lcores": d.get("lcores", "1"),
> +            "use_first_core": d.get("use_first_core", False),
> +            "memory_channels": d.get("memory_channels", 1),
> +            "hugepages": hugepage_config,
> +            "ports": [PortConfig.from_dict(d["name"], port) for port in 
> d["ports"]],
> +        }
> +
> +        return NodeConfiguration(**common_config)
> +
> +
> +@dataclass(slots=True)

Looks like this could be frozen as well. This should work even in the
future, as I imagine we'll get all the info we need just once.

> +class NodeInfo:
> +    """Class to hold important versions within the node.
> +
> +    This class, unlike the NodeConfiguration class, cannot be generated at 
> the start.
> +    This is because we need to initialize a connection with the node before 
> we can
> +    collect the information needed in this class. Therefore, it cannot be a 
> part of
> +    the configuration class above.
> +    """
>
> -        return NodeConfiguration(
> -            name=d["name"],
> -            hostname=d["hostname"],
> -            user=d["user"],
> -            password=d.get("password"),
> -            arch=Architecture(d["arch"]),
> -            os=OS(d["os"]),
> -            lcores=d.get("lcores", "1"),
> -            use_first_core=d.get("use_first_core", False),
> -            memory_channels=d.get("memory_channels", 1),
> -            hugepages=hugepage_config,
> +    os_name: str
> +    os_version: str
> +    kernel_version: str
> +
> +    @staticmethod
> +    def from_dict(d: dict):
> +        return NodeInfo(
> +            os_name=d["os_name"],
> +            os_version=d["os_version"],
> +            kernel_version=d["kernel_version"],
>          )

We don't need the from_dict method, as we can instantiate this class
right away (node_info = NodeInfo(os_name=d["os_name"],
os_version=d["os_version"], kernel_version=d["kernel_version"])).

The other classes need this method because we're doing some processing
before instantiating the classes - this one doesn't need it.

>
>
> @@ -128,6 +169,24 @@ def from_dict(d: dict) -> "BuildTargetConfiguration":
>          )
>
>
> +@dataclass(slots=True)

Can also be frozen.

> +class BuildTargetInfo:
> +    """Class to hold important versions within the build target.
> +
> +    This is very similar to the NodeVersionInfo class, it just instead holds 
> information

References renamed class.

> +    for the build target.
> +    """
> +
> +    dpdk_version: str
> +    compiler_version: str
> +
> +    @staticmethod
> +    def from_dict(d: dict):
> +        return BuildTargetInfo(
> +            dpdk_version=d["dpdk_version"], 
> compiler_version=d["compiler_version"]
> +        )

Same as above.

> +
> +
>  class TestSuiteConfigDict(TypedDict):
>      suite: str
>      cases: list[str]
> @@ -157,6 +216,8 @@ class ExecutionConfiguration:
>      func: bool
>      test_suites: list[TestSuiteConfig]
>      system_under_test: NodeConfiguration
> +    vdevs: list[str]
> +    skip_smoke_tests: bool
>
>      @staticmethod
>      def from_dict(d: dict, node_map: dict) -> "ExecutionConfiguration":
> @@ -166,15 +227,20 @@ def from_dict(d: dict, node_map: dict) -> 
> "ExecutionConfiguration":
>          test_suites: list[TestSuiteConfig] = list(
>              map(TestSuiteConfig.from_dict, d["test_suites"])
>          )
> -        sut_name = d["system_under_test"]
> +        sut_name = d["system_under_test"]["node_name"]
> +        skip_smoke_tests = d.get("skip_smoke_tests", False)
>          assert sut_name in node_map, f"Unknown SUT {sut_name} in execution 
> {d}"
> -
> +        vdevs = (
> +            d["system_under_test"]["vdevs"] if "vdevs" in 
> d["system_under_test"] else []
> +        )
>          return ExecutionConfiguration(
>              build_targets=build_targets,
>              perf=d["perf"],
>              func=d["func"],
> +            skip_smoke_tests=skip_smoke_tests,
>              test_suites=test_suites,
>              system_under_test=node_map[sut_name],
> +            vdevs=vdevs,
>          )
>
>
> @@ -221,3 +287,27 @@ def load_config() -> Configuration:
>
>
>  CONFIGURATION = load_config()
> +
> +
> +@unique
> +class InteractiveApp(Enum):
> +    """An enum that represents different supported interactive applications
> +
> +    The values in this enum must all be set to objects that have a key called
> +    "default_path" where "default_path" represents a PurPath object for the 
> path
> +    to the application. This default path will be passed into the handler 
> class
> +    for the application so that it can start the application. For every key 
> other
> +    than the default shell option, the path will be appended to the path to 
> the DPDK
> +    build directory for the current SUT node.
> +    """
> +
> +    shell = {"default_path": PurePath()}
> +    testpmd = {"default_path": PurePath("app", "dpdk-testpmd")}
> +
> +    def get_path(self) -> PurePath:
> +        """A method for getting the default paths of an application
> +
> +        Returns:
> +            String array that represents an OS agnostic path to the 
> application.
> +        """
> +        return self.value["default_path"]
> diff --git a/dts/framework/config/conf_yaml_schema.json 
> b/dts/framework/config/conf_yaml_schema.json
> index ca2d4a1e..61f52b43 100644
> --- a/dts/framework/config/conf_yaml_schema.json
> +++ b/dts/framework/config/conf_yaml_schema.json
> @@ -6,6 +6,76 @@
>        "type": "string",
>        "description": "A unique identifier for a node"
>      },
> +    "NIC": {
> +      "type": "string",
> +      "enum": [
> +        "ALL",
> +        "ConnectX3_MT4103",
> +        "ConnectX4_LX_MT4117",
> +        "ConnectX4_MT4115",
> +        "ConnectX5_MT4119",
> +        "ConnectX5_MT4121",
> +        "I40E_10G-10G_BASE_T_BC",
> +        "I40E_10G-10G_BASE_T_X722",
> +        "I40E_10G-SFP_X722",
> +        "I40E_10G-SFP_XL710",
> +        "I40E_10G-X722_A0",
> +        "I40E_1G-1G_BASE_T_X722",
> +        "I40E_25G-25G_SFP28",
> +        "I40E_40G-QSFP_A",
> +        "I40E_40G-QSFP_B",
> +        "IAVF-ADAPTIVE_VF",
> +        "IAVF-VF",
> +        "IAVF_10G-X722_VF",
> +        "ICE_100G-E810C_QSFP",
> +        "ICE_25G-E810C_SFP",
> +        "ICE_25G-E810_XXV_SFP",
> +        "IGB-I350_VF",
> +        "IGB_1G-82540EM",
> +        "IGB_1G-82545EM_COPPER",
> +        "IGB_1G-82571EB_COPPER",
> +        "IGB_1G-82574L",
> +        "IGB_1G-82576",
> +        "IGB_1G-82576_QUAD_COPPER",
> +        "IGB_1G-82576_QUAD_COPPER_ET2",
> +        "IGB_1G-82580_COPPER",
> +        "IGB_1G-I210_COPPER",
> +        "IGB_1G-I350_COPPER",
> +        "IGB_1G-I354_SGMII",
> +        "IGB_1G-PCH_LPTLP_I218_LM",
> +        "IGB_1G-PCH_LPTLP_I218_V",
> +        "IGB_1G-PCH_LPT_I217_LM",
> +        "IGB_1G-PCH_LPT_I217_V",
> +        "IGB_2.5G-I354_BACKPLANE_2_5GBPS",
> +        "IGC-I225_LM",
> +        "IGC-I226_LM",
> +        "IXGBE_10G-82599_SFP",
> +        "IXGBE_10G-82599_SFP_SF_QP",
> +        "IXGBE_10G-82599_T3_LOM",
> +        "IXGBE_10G-82599_VF",
> +        "IXGBE_10G-X540T",
> +        "IXGBE_10G-X540_VF",
> +        "IXGBE_10G-X550EM_A_SFP",
> +        "IXGBE_10G-X550EM_X_10G_T",
> +        "IXGBE_10G-X550EM_X_SFP",
> +        "IXGBE_10G-X550EM_X_VF",
> +        "IXGBE_10G-X550T",
> +        "IXGBE_10G-X550_VF",
> +        "brcm_57414",
> +        "brcm_P2100G",
> +        "cavium_0011",
> +        "cavium_a034",
> +        "cavium_a063",
> +        "cavium_a064",
> +        "fastlinq_ql41000",
> +        "fastlinq_ql41000_vf",
> +        "fastlinq_ql45000",
> +        "fastlinq_ql45000_vf",
> +        "hi1822",
> +        "virtio"
> +      ]
> +    },
> +
>      "ARCH": {
>        "type": "string",
>        "enum": [
> @@ -94,6 +164,19 @@
>          "amount"
>        ]
>      },
> +    "pci_address": {
> +      "type": "string",
> +      "pattern": "^[\\da-fA-F]{4}:[\\da-fA-F]{2}:[\\da-fA-F]{2}.\\d:?\\w*$"
> +    },
> +    "port_peer_address": {
> +      "description": "Peer is a TRex port, and IXIA port or a PCI address",
> +      "oneOf": [
> +        {
> +          "description": "PCI peer port",
> +          "$ref": "#/definitions/pci_address"
> +        }
> +      ]
> +    },
>      "test_suite": {
>        "type": "string",
>        "enum": [
> @@ -165,6 +248,44 @@
>            },
>            "hugepages": {
>              "$ref": "#/definitions/hugepages"
> +          },
> +          "ports": {
> +            "type": "array",
> +            "items": {
> +              "type": "object",
> +              "description": "Each port should be described on both sides of 
> the connection. This makes configuration slightly more verbose but greatly 
> simplifies implementation. If there are an inconsistencies, then DTS will not 
> run until that issue is fixed. An example inconsistency would be port 1, node 
> 1 says it is connected to port 1, node 2, but port 1, node 2 says it is 
> connected to port 2, node 1.",
> +              "properties": {
> +                "pci": {
> +                  "$ref": "#/definitions/pci_address",
> +                  "description": "The local PCI address of the port"
> +                },
> +                "os_driver_for_dpdk": {
> +                  "type": "string",
> +                  "description": "The driver that the kernel should bind 
> this device to for DPDK to use it. (ex: vfio-pci)"
> +                },
> +                "os_driver": {
> +                  "type": "string",
> +                  "description": "The driver normally used by this port (ex: 
> i40e)"
> +                },
> +                "peer_node": {
> +                  "type": "string",
> +                  "description": "The name of the node the peer port is on"
> +                },
> +                "peer_pci": {
> +                  "$ref": "#/definitions/pci_address",
> +                  "description": "The PCI address of the peer port"
> +                }
> +              },
> +              "additionalProperties": false,
> +              "required": [
> +                "pci",
> +                "os_driver_for_dpdk",
> +                "os_driver",
> +                "peer_node",
> +                "peer_pci"
> +              ]
> +            },
> +            "minimum": 1
>            }
>          },
>          "additionalProperties": false,
> @@ -211,8 +332,27 @@
>                ]
>              }
>            },
> +          "skip_smoke_tests": {
> +            "description": "Optional field that allows you to skip smoke 
> testing",
> +            "type": "boolean"
> +          },
>            "system_under_test": {
> -            "$ref": "#/definitions/node_name"
> +            "type":"object",
> +            "properties": {
> +              "node_name": {
> +                "$ref": "#/definitions/node_name"
> +              },
> +              "vdevs": {
> +                "description": "Opentional list of names of vdevs to be used 
> in execution",
> +                "type": "array",
> +                "items": {
> +                  "type": "string"
> +                }
> +              }
> +            },
> +            "required": [
> +              "node_name"
> +            ]
>            }
>          },
>          "additionalProperties": false,
> diff --git a/dts/framework/dts.py b/dts/framework/dts.py
> index 05022845..1b67938f 100644
> --- a/dts/framework/dts.py
> +++ b/dts/framework/dts.py
> @@ -5,7 +5,13 @@
>
>  import sys
>
> -from .config import CONFIGURATION, BuildTargetConfiguration, 
> ExecutionConfiguration
> +from .config import (
> +    CONFIGURATION,
> +    BuildTargetConfiguration,
> +    ExecutionConfiguration,
> +    TestSuiteConfig,
> +)
> +from .exception import BlockingTestSuiteError
>  from .logger import DTSLOG, getLogger
>  from .test_result import BuildTargetResult, DTSResult, ExecutionResult, 
> Result
>  from .test_suite import get_test_suites
> @@ -82,7 +88,7 @@ def _run_execution(
>      running all build targets in the given execution.
>      """
>      dts_logger.info(f"Running execution with SUT 
> '{execution.system_under_test.name}'.")
> -    execution_result = result.add_execution(sut_node.config)
> +    execution_result = result.add_execution(sut_node.config, 
> sut_node.node_info)
>
>      try:
>          sut_node.set_up_execution(execution)
> @@ -118,14 +124,15 @@ def _run_build_target(
>
>      try:
>          sut_node.set_up_build_target(build_target)
> -        result.dpdk_version = sut_node.dpdk_version
> +        # result.dpdk_version = sut_node.dpdk_version
> +        
> build_target_result.add_build_target_versions(sut_node.get_build_target_info())
>          build_target_result.update_setup(Result.PASS)
>      except Exception as e:
>          dts_logger.exception("Build target setup failed.")
>          build_target_result.update_setup(Result.FAIL, e)
>
>      else:
> -        _run_suites(sut_node, execution, build_target_result)
> +        _run_all_suites(sut_node, execution, build_target_result)
>
>      finally:
>          try:
> @@ -136,7 +143,7 @@ def _run_build_target(
>              build_target_result.update_teardown(Result.FAIL, e)
>
>
> -def _run_suites(
> +def _run_all_suites(
>      sut_node: SutNode,
>      execution: ExecutionConfiguration,
>      build_target_result: BuildTargetResult,
> @@ -146,27 +153,62 @@ def _run_suites(
>      with possibly only a subset of test cases.
>      If no subset is specified, run all test cases.
>      """
> +    end_build_target = False
> +    if not execution.skip_smoke_tests:
> +        execution.test_suites[:0] = 
> [TestSuiteConfig.from_dict("smoke_tests")]
>      for test_suite_config in execution.test_suites:
>          try:
> -            full_suite_path = 
> f"tests.TestSuite_{test_suite_config.test_suite}"
> -            test_suite_classes = get_test_suites(full_suite_path)
> -            suites_str = ", ".join((x.__name__ for x in test_suite_classes))
> -            dts_logger.debug(
> -                f"Found test suites '{suites_str}' in '{full_suite_path}'."
> +            _run_single_suite(
> +                sut_node, execution, build_target_result, test_suite_config
>              )
> -        except Exception as e:
> -            dts_logger.exception("An error occurred when searching for test 
> suites.")
> -            result.update_setup(Result.ERROR, e)
> -
> -        else:
> -            for test_suite_class in test_suite_classes:
> -                test_suite = test_suite_class(
> -                    sut_node,
> -                    test_suite_config.test_cases,
> -                    execution.func,
> -                    build_target_result,
> -                )
> -                test_suite.run()
> +        except BlockingTestSuiteError as e:
> +            dts_logger.exception(
> +                f"An error occurred within {test_suite_config.test_suite}. "
> +                "Skipping build target..."
> +            )
> +            result.add_error(e)
> +            end_build_target = True
> +        # if a blocking test failed and we need to bail out of suite 
> executions
> +        if end_build_target:
> +            break
> +
> +
> +def _run_single_suite(
> +    sut_node: SutNode,
> +    execution: ExecutionConfiguration,
> +    build_target_result: BuildTargetResult,
> +    test_suite_config: TestSuiteConfig,
> +) -> None:
> +    """Runs a single test suite.
> +
> +    Args:
> +        sut_node: Node to run tests on.
> +        execution: Execution the test case belongs to.
> +        build_target_result: Build target configuration test case is run on
> +        test_suite_config: Test suite configuration
> +
> +    Raises:
> +        BlockingTestSuiteError: If a test suite that was marked as blocking 
> fails.
> +    """
> +    try:
> +        full_suite_path = f"tests.TestSuite_{test_suite_config.test_suite}"
> +        test_suite_classes = get_test_suites(full_suite_path)
> +        suites_str = ", ".join((x.__name__ for x in test_suite_classes))
> +        dts_logger.debug(f"Found test suites '{suites_str}' in 
> '{full_suite_path}'.")
> +    except Exception as e:
> +        dts_logger.exception("An error occurred when searching for test 
> suites.")
> +        result.update_setup(Result.ERROR, e)
> +
> +    else:
> +        for test_suite_class in test_suite_classes:
> +            test_suite = test_suite_class(
> +                sut_node,
> +                test_suite_config.test_cases,
> +                execution.func,
> +                build_target_result,
> +                result,
> +            )
> +            test_suite.run()
>
>
>  def _exit_dts() -> None:
> diff --git a/dts/framework/exception.py b/dts/framework/exception.py
> index ca353d98..dfb12df4 100644
> --- a/dts/framework/exception.py
> +++ b/dts/framework/exception.py
> @@ -25,6 +25,7 @@ class ErrorSeverity(IntEnum):
>      SSH_ERR = 4
>      DPDK_BUILD_ERR = 10
>      TESTCASE_VERIFY_ERR = 20
> +    BLOCKING_TESTSUITE_ERR = 25
>
>
>  class DTSError(Exception):
> @@ -144,3 +145,14 @@ def __init__(self, value: str):
>
>      def __str__(self) -> str:
>          return repr(self.value)
> +
> +
> +class BlockingTestSuiteError(DTSError):
> +    suite_name: str
> +    severity: ClassVar[ErrorSeverity] = ErrorSeverity.BLOCKING_TESTSUITE_ERR
> +
> +    def __init__(self, suite_name: str) -> None:
> +        self.suite_name = suite_name
> +
> +    def __str__(self) -> str:
> +        return f"Blocking suite {self.suite_name} failed."
> diff --git a/dts/framework/remote_session/__init__.py 
> b/dts/framework/remote_session/__init__.py
> index ee221503..4fe32d35 100644
> --- a/dts/framework/remote_session/__init__.py
> +++ b/dts/framework/remote_session/__init__.py
> @@ -1,5 +1,6 @@
>  # SPDX-License-Identifier: BSD-3-Clause
>  # Copyright(c) 2023 PANTHEON.tech s.r.o.
> +# Copyright(c) 2023 University of New Hampshire
>
>  """
>  The package provides modules for managing remote connections to a remote 
> host (node),
> @@ -17,7 +18,14 @@
>
>  from .linux_session import LinuxSession
>  from .os_session import OSSession
> -from .remote import CommandResult, RemoteSession, SSHSession
> +from .remote import (
> +    CommandResult,
> +    InteractiveRemoteSession,
> +    InteractiveShell,
> +    RemoteSession,
> +    SSHSession,
> +    TestPmdShell,
> +)
>
>
>  def create_session(
> diff --git a/dts/framework/remote_session/os_session.py 
> b/dts/framework/remote_session/os_session.py
> index 4c48ae25..4346ecc4 100644
> --- a/dts/framework/remote_session/os_session.py
> +++ b/dts/framework/remote_session/os_session.py
> @@ -6,13 +6,19 @@
>  from collections.abc import Iterable
>  from pathlib import PurePath
>
> -from framework.config import Architecture, NodeConfiguration
> +from framework.config import Architecture, NodeConfiguration, NodeInfo
>  from framework.logger import DTSLOG
>  from framework.settings import SETTINGS
>  from framework.testbed_model import LogicalCore
>  from framework.utils import EnvVarsDict, MesonArgs
>
> -from .remote import CommandResult, RemoteSession, create_remote_session
> +from .remote import (
> +    CommandResult,
> +    InteractiveRemoteSession,
> +    RemoteSession,
> +    create_interactive_session,
> +    create_remote_session,
> +)
>
>
>  class OSSession(ABC):
> @@ -26,6 +32,7 @@ class OSSession(ABC):
>      name: str
>      _logger: DTSLOG
>      remote_session: RemoteSession
> +    interactive_session: InteractiveRemoteSession
>
>      def __init__(
>          self,
> @@ -37,6 +44,7 @@ def __init__(
>          self.name = name
>          self._logger = logger
>          self.remote_session = create_remote_session(node_config, name, 
> logger)
> +        self.interactive_session = create_interactive_session(node_config, 
> name, logger)
>
>      def close(self, force: bool = False) -> None:
>          """
> @@ -173,3 +181,15 @@ def setup_hugepages(self, hugepage_amount: int, 
> force_first_numa: bool) -> None:
>          if needed and mount the hugepages if needed.
>          If force_first_numa is True, configure hugepages just on the first 
> socket.
>          """
> +
> +    @abstractmethod
> +    def get_compiler_version(self, compiler_name: str) -> str:
> +        """
> +        Get installed version of compiler used for DPDK
> +        """
> +
> +    @abstractmethod
> +    def get_node_info(self) -> NodeInfo:
> +        """
> +        Collect information about the node
> +        """
> diff --git a/dts/framework/remote_session/posix_session.py 
> b/dts/framework/remote_session/posix_session.py
> index d38062e8..f8ec159f 100644
> --- a/dts/framework/remote_session/posix_session.py
> +++ b/dts/framework/remote_session/posix_session.py
> @@ -6,7 +6,7 @@
>  from collections.abc import Iterable
>  from pathlib import PurePath, PurePosixPath
>
> -from framework.config import Architecture
> +from framework.config import Architecture, NodeInfo
>  from framework.exception import DPDKBuildError, RemoteCommandExecutionError
>  from framework.settings import SETTINGS
>  from framework.utils import EnvVarsDict, MesonArgs
> @@ -219,3 +219,30 @@ def _remove_dpdk_runtime_dirs(
>
>      def get_dpdk_file_prefix(self, dpdk_prefix) -> str:
>          return ""
> +
> +    def get_compiler_version(self, compiler_name: str) -> str:
> +        match compiler_name:
> +            case "gcc":
> +                return self.send_command(f"{compiler_name} --version", 
> 60).stdout.split(
> +                    "\n"
> +                )[0]

The timeouts are still there.

> +            case "clang":
> +                return self.send_command(f"{compiler_name} --version", 
> 60).stdout.split(
> +                    "\n"
> +                )[0]
> +            case "msvc":
> +                return self.send_command("cl", 60).stdout
> +            case "icc":
> +                return self.send_command(f"{compiler_name} -V", 60).stdout
> +            case _:
> +                raise ValueError(f"Unknown compiler {compiler_name}")
> +
> +    def get_node_info(self) -> NodeInfo:
> +        os_release_info = self.send_command(
> +            "awk -F= '$1 ~ /^NAME$|^VERSION$/ {print $2}' /etc/os-release",
> +            SETTINGS.timeout,
> +        ).stdout.split("\n")
> +        kernel_version = self.send_command("uname -r", 
> SETTINGS.timeout).stdout
> +        return NodeInfo(
> +            os_release_info[0].strip(), os_release_info[1].strip(), 
> kernel_version
> +        )
> diff --git a/dts/framework/remote_session/remote/__init__.py 
> b/dts/framework/remote_session/remote/__init__.py
> index 8a151221..224598a8 100644
> --- a/dts/framework/remote_session/remote/__init__.py
> +++ b/dts/framework/remote_session/remote/__init__.py
> @@ -1,16 +1,26 @@
>  # SPDX-License-Identifier: BSD-3-Clause
>  # Copyright(c) 2023 PANTHEON.tech s.r.o.
> +# Copyright(c) 2022-2023 University of New Hampshire

There are other instances of the copyright statement. It's not
necessary to change them all to just 2023, but I'd say it's
preferable.

>
>  # pylama:ignore=W0611
>
>  from framework.config import NodeConfiguration
>  from framework.logger import DTSLOG
>
> +from .interactive_remote_session import InteractiveRemoteSession
> +from .interactive_shell import InteractiveShell
>  from .remote_session import CommandResult, RemoteSession
>  from .ssh_session import SSHSession
> +from .testpmd_shell import TestPmdShell
>
>
>  def create_remote_session(
>      node_config: NodeConfiguration, name: str, logger: DTSLOG
>  ) -> RemoteSession:
>      return SSHSession(node_config, name, logger)
> +
> +
> +def create_interactive_session(
> +    node_config: NodeConfiguration, name: str, logger: DTSLOG
> +) -> InteractiveRemoteSession:
> +    return InteractiveRemoteSession(node_config, logger)
> diff --git 
> a/dts/framework/remote_session/remote/interactive_remote_session.py 
> b/dts/framework/remote_session/remote/interactive_remote_session.py
> new file mode 100644
> index 00000000..e145d35d
> --- /dev/null
> +++ b/dts/framework/remote_session/remote/interactive_remote_session.py
> @@ -0,0 +1,118 @@
> +# SPDX-License-Identifier: BSD-3-Clause
> +# Copyright(c) 2023 University of New Hampshire
> +
> +import socket
> +import traceback
> +from pathlib import PurePath
> +from typing import Union
> +
> +from paramiko import AutoAddPolicy, SSHClient, Transport  # type: ignore
> +from paramiko.ssh_exception import (  # type: ignore
> +    AuthenticationException,
> +    BadHostKeyException,
> +    NoValidConnectionsError,
> +    SSHException,
> +)
> +
> +from framework.config import InteractiveApp, NodeConfiguration
> +from framework.exception import SSHConnectionError
> +from framework.logger import DTSLOG
> +
> +from .interactive_shell import InteractiveShell
> +from .testpmd_shell import TestPmdShell
> +
> +
> +class InteractiveRemoteSession:
> +    hostname: str
> +    ip: str
> +    port: int
> +    username: str
> +    password: str
> +    _logger: DTSLOG
> +    _node_config: NodeConfiguration
> +    session: SSHClient
> +    _transport: Transport | None
> +
> +    def __init__(self, node_config: NodeConfiguration, _logger: DTSLOG) -> 
> None:
> +        self._node_config = node_config
> +        self._logger = _logger
> +        self.hostname = node_config.hostname
> +        self.username = node_config.user
> +        self.password = node_config.password if node_config.password else ""
> +        port = "22"
> +        self.ip = node_config.hostname
> +        if ":" in node_config.hostname:
> +            self.ip, port = node_config.hostname.split(":")
> +        self.port = int(port)
> +        self._logger.info(
> +            f"Initializing interactive connection for 
> {self.username}@{self.hostname}"
> +        )
> +        self._connect()
> +        self._logger.info(
> +            f"Interactive connection successful for 
> {self.username}@{self.hostname}"
> +        )
> +
> +    def _connect(self) -> None:
> +        client = SSHClient()
> +        client.set_missing_host_key_policy(AutoAddPolicy)
> +        self.session = client
> +        retry_attempts = 10
> +        for retry_attempt in range(retry_attempts):
> +            try:
> +                client.connect(
> +                    self.ip,
> +                    username=self.username,
> +                    port=self.port,
> +                    password=self.password,
> +                    timeout=20 if self.port else 10,
> +                )
> +            except (TypeError, BadHostKeyException, AuthenticationException) 
> as e:
> +                self._logger.exception(e)
> +                raise SSHConnectionError(self.hostname) from e
> +            except (NoValidConnectionsError, socket.error, SSHException) as 
> e:
> +                self._logger.debug(traceback.format_exc())
> +                self._logger.warning(e)
> +                self._logger.info(
> +                    "Retrying interactive session connection: "
> +                    f"retry number {retry_attempt +1}"
> +                )
> +            else:
> +                break
> +        else:
> +            raise SSHConnectionError(self.hostname)
> +        # Interactive sessions are used on an "as needed" basis so we have
> +        # to set a keepalive
> +        self._transport = self.session.get_transport()
> +        if self._transport is not None:
> +            self._transport.set_keepalive(30)
> +
> +    def create_interactive_shell(
> +        self,
> +        shell_type: InteractiveApp,
> +        path_to_app: PurePath,
> +        eal_parameters: str,
> +        timeout: float,
> +    ) -> Union[InteractiveShell, TestPmdShell]:
> +        """
> +        See "create_interactive_shell" in SutNode
> +        """
> +        match (shell_type):
> +            case InteractiveApp.shell:
> +                return InteractiveShell(
> +                    self.session, self._logger, path_to_app, timeout
> +                )
> +            case InteractiveApp.testpmd:
> +                return TestPmdShell(
> +                    self.session,
> +                    self._logger,
> +                    path_to_app,
> +                    timeout=timeout,
> +                    eal_flags=eal_parameters,
> +                )
> +            case _:
> +                self._logger.info(
> +                    f"Unhandled app type {shell_type.name}, defaulting to 
> shell."
> +                )
> +                return InteractiveShell(
> +                    self.session, self._logger, path_to_app, timeout
> +                )
> diff --git a/dts/framework/remote_session/remote/interactive_shell.py 
> b/dts/framework/remote_session/remote/interactive_shell.py
> new file mode 100644
> index 00000000..4b0735c8
> --- /dev/null
> +++ b/dts/framework/remote_session/remote/interactive_shell.py
> @@ -0,0 +1,99 @@
> +from pathlib import PurePath
> +
> +from paramiko import Channel, SSHClient, channel  # type: ignore
> +
> +from framework.logger import DTSLOG
> +from framework.settings import SETTINGS
> +
> +
> +class InteractiveShell:
> +
> +    _interactive_session: SSHClient
> +    _stdin: channel.ChannelStdinFile
> +    _stdout: channel.ChannelFile
> +    _ssh_channel: Channel
> +    _logger: DTSLOG
> +    _timeout: float
> +    _path_to_app: PurePath
> +
> +    def __init__(
> +        self,
> +        interactive_session: SSHClient,
> +        logger: DTSLOG,
> +        path_to_app: PurePath,
> +        timeout: float = SETTINGS.timeout,
> +    ) -> None:
> +        self._interactive_session = interactive_session
> +        self._ssh_channel = self._interactive_session.invoke_shell()
> +        self._stdin = self._ssh_channel.makefile_stdin("w")
> +        self._stdout = self._ssh_channel.makefile("r")
> +        self._ssh_channel.settimeout(timeout)
> +        self._ssh_channel.set_combine_stderr(True)  # combines stdout and 
> stderr streams
> +        self._logger = logger
> +        self._timeout = timeout
> +        self._path_to_app = path_to_app
> +
> +
> +    def send_command_no_output(self, command: str) -> None:
> +        """Send command to channel without recording output.
> +
> +        This method will not verify any input or output, it will simply 
> assume the
> +        command succeeded. This method will also consume all output in the 
> buffer
> +        after executing the command.
> +        """
> +        self._logger.info(
> +            f"Sending command {command.strip()} and not collecting output"
> +        )
> +        self._stdin.write(f"{command}\n")
> +        self._stdin.flush()
> +        self.empty_stdout_buffer()
> +
> +    def empty_stdout_buffer(self) -> None:
> +        """Removes all data from the stdout buffer.
> +
> +        Because of the way paramiko handles read buffers, there is no way to 
> effectively
> +        remove data from, or "flush", read buffers. This method essentially 
> moves our
> +        offset on the buffer to the end and thus "removes" the data from the 
> buffer.
> +        Timeouts are thrown on read operations of paramiko pipes based on 
> whether data
> +        had been received before timeout so we assume that if we reach the 
> timeout then
> +        we are at the end of the buffer.
> +        """
> +        self._ssh_channel.settimeout(0.5)
> +        try:
> +            for line in self._stdout:
> +                pass
> +        except TimeoutError:
> +            pass
> +        self._ssh_channel.settimeout(self._timeout)  # reset timeout
> +
> +    def send_command_get_output(self, command: str, prompt: str) -> str:
> +        """Send a command and get all output before the expected ending 
> string.
> +
> +        Lines that expect input are not included in the stdout buffer so 
> they cannot be
> +        used for expect. For example, if you were prompted to log into 
> something
> +        with a username and password, you cannot expect "username:" because 
> it won't
> +        yet be in the stdout buffer. A work around for this could be 
> consuming an
> +        extra newline character to force the current prompt into the stdout 
> buffer.
> +
> +        Returns:
> +            All output in the buffer before expected string
> +        """
> +        self._logger.info(f"Sending command {command.strip()}...")
> +        self._stdin.write(f"{command}\n")
> +        self._stdin.flush()
> +        out: str = ""
> +        for line in self._stdout:
> +            out += line
> +            if prompt in line and not line.rstrip().endswith(
> +                command.rstrip()
> +            ):  # ignore line that sent command
> +                break
> +        self._logger.debug(f"Got output: {out}")
> +        return out
> +
> +    def close(self) -> None:
> +        self._stdin.close()
> +        self._ssh_channel.close()
> +
> +    def __del__(self) -> None:
> +        self.close()
> diff --git a/dts/framework/remote_session/remote/testpmd_shell.py 
> b/dts/framework/remote_session/remote/testpmd_shell.py
> new file mode 100644
> index 00000000..bde3468b
> --- /dev/null
> +++ b/dts/framework/remote_session/remote/testpmd_shell.py
> @@ -0,0 +1,67 @@
> +# SPDX-License-Identifier: BSD-3-Clause
> +# Copyright(c) 2023 University of New Hampshire
> +
> +
> +from pathlib import PurePath
> +
> +from paramiko import SSHClient  # type: ignore
> +
> +from framework.logger import DTSLOG
> +from framework.settings import SETTINGS
> +
> +from .interactive_shell import InteractiveShell
> +
> +
> +class TestPmdShell(InteractiveShell):
> +    expected_prompt: str = "testpmd>"
> +    _eal_flags: str
> +
> +    def __init__(
> +        self,
> +        interactive_session: SSHClient,
> +        logger: DTSLOG,
> +        path_to_testpmd: PurePath,
> +        eal_flags: str,
> +        timeout: float = SETTINGS.timeout,
> +    ) -> None:
> +        """Initializes an interactive testpmd session using specified 
> parameters."""
> +        self._eal_flags = eal_flags
> +
> +        super(TestPmdShell, self).__init__(
> +            interactive_session,
> +            logger=logger,
> +            path_to_app=path_to_testpmd,
> +            timeout=timeout,
> +        )
> +        self._start_application()
> +
> +    def _start_application(self) -> None:
> +        """Starts a new interactive testpmd shell using _path_to_app.
> +        """
> +        self.send_command(
> +            f"{self._path_to_app} {self._eal_flags} -- -i",
> +        )
> +
> +    def send_command(self, command: str, prompt: str = expected_prompt) -> 
> str:
> +        """Specific way of handling the command for testpmd
> +
> +        An extra newline character is consumed in order to force the current 
> line into
> +        the stdout buffer.
> +        """
> +        return self.send_command_get_output(f"{command}\n", prompt)
> +
> +    def get_devices(self) -> list[str]:
> +        """Get a list of device names that are known to testpmd
> +
> +        Uses the device info listed in testpmd and then parses the output to
> +        return only the names of the devices.
> +
> +        Returns:
> +            A list of strings representing device names (e.g. 0000:14:00.1)
> +        """
> +        dev_info: str = self.send_command("show device info all")
> +        dev_list: list[str] = []
> +        for line in dev_info.split("\n"):
> +            if "device name:" in line.lower():
> +                dev_list.append(line.strip().split(": ")[1].strip())
> +        return dev_list
> diff --git a/dts/framework/test_result.py b/dts/framework/test_result.py
> index 74391982..2436eb7f 100644
> --- a/dts/framework/test_result.py
> +++ b/dts/framework/test_result.py
> @@ -1,5 +1,6 @@
>  # SPDX-License-Identifier: BSD-3-Clause
>  # Copyright(c) 2023 PANTHEON.tech s.r.o.
> +# Copyright(c) 2022-2023 University of New Hampshire
>
>  """
>  Generic result container and reporters
> @@ -13,9 +14,11 @@
>      OS,
>      Architecture,
>      BuildTargetConfiguration,
> +    BuildTargetInfo,
>      Compiler,
>      CPUType,
>      NodeConfiguration,
> +    NodeInfo,
>  )
>  from .exception import DTSError, ErrorSeverity
>  from .logger import DTSLOG
> @@ -67,12 +70,14 @@ class Statistics(dict):
>      Using a dict provides a convenient way to format the data.
>      """
>
> -    def __init__(self, dpdk_version):
> +    def __init__(self, output_info: dict[str, str] | None):
>          super(Statistics, self).__init__()
>          for result in Result:
>              self[result.name] = 0
>          self["PASS RATE"] = 0.0
> -        self["DPDK VERSION"] = dpdk_version
> +        if output_info:
> +            for info_key, info_val in output_info.items():
> +                self[info_key] = info_val
>
>      def __iadd__(self, other: Result) -> "Statistics":
>          """
> @@ -206,6 +211,8 @@ class BuildTargetResult(BaseResult):
>      os: OS
>      cpu: CPUType
>      compiler: Compiler
> +    compiler_version: str | None
> +    dpdk_version: str | None
>
>      def __init__(self, build_target: BuildTargetConfiguration):
>          super(BuildTargetResult, self).__init__()
> @@ -213,6 +220,12 @@ def __init__(self, build_target: 
> BuildTargetConfiguration):
>          self.os = build_target.os
>          self.cpu = build_target.cpu
>          self.compiler = build_target.compiler
> +        self.compiler_version = None
> +        self.dpdk_version = None
> +
> +    def add_build_target_versions(self, versions: BuildTargetInfo) -> None:
> +        self.compiler_version = versions.compiler_version
> +        self.dpdk_version = versions.dpdk_version
>
>      def add_test_suite(self, test_suite_name: str) -> TestSuiteResult:
>          test_suite_result = TestSuiteResult(test_suite_name)
> @@ -228,10 +241,17 @@ class ExecutionResult(BaseResult):
>      """
>
>      sut_node: NodeConfiguration
> +    sut_os_name: str
> +    sut_os_version: str
> +    sut_kernel_version: str
>
> -    def __init__(self, sut_node: NodeConfiguration):
> +    def __init__(self, sut_node: NodeConfiguration, sut_version_info: 
> NodeInfo):
>          super(ExecutionResult, self).__init__()
>          self.sut_node = sut_node
> +        self.sut_version_info = sut_version_info
> +        self.sut_os_name = sut_version_info.os_name
> +        self.sut_os_version = sut_version_info.os_version
> +        self.sut_kernel_version = sut_version_info.kernel_version
>
>      def add_build_target(
>          self, build_target: BuildTargetConfiguration
> @@ -258,6 +278,7 @@ class DTSResult(BaseResult):
>      """
>
>      dpdk_version: str | None
> +    output: dict | None
>      _logger: DTSLOG
>      _errors: list[Exception]
>      _return_code: ErrorSeverity
> @@ -267,14 +288,17 @@ class DTSResult(BaseResult):
>      def __init__(self, logger: DTSLOG):
>          super(DTSResult, self).__init__()
>          self.dpdk_version = None
> +        self.output = None
>          self._logger = logger
>          self._errors = []
>          self._return_code = ErrorSeverity.NO_ERR
>          self._stats_result = None
>          self._stats_filename = os.path.join(SETTINGS.output_dir, 
> "statistics.txt")
>
> -    def add_execution(self, sut_node: NodeConfiguration) -> ExecutionResult:
> -        execution_result = ExecutionResult(sut_node)
> +    def add_execution(
> +        self, sut_node: NodeConfiguration, sut_version_info: NodeInfo
> +    ) -> ExecutionResult:
> +        execution_result = ExecutionResult(sut_node, sut_version_info)
>          self._inner_results.append(execution_result)
>          return execution_result
>
> @@ -296,7 +320,8 @@ def process(self) -> None:
>              for error in self._errors:
>                  self._logger.debug(repr(error))
>
> -        self._stats_result = Statistics(self.dpdk_version)
> +        self._stats_result = Statistics(self.output)
> +        # add information gathered from the smoke tests to the statistics
>          self.add_stats(self._stats_result)
>          with open(self._stats_filename, "w+") as stats_file:
>              stats_file.write(str(self._stats_result))
> diff --git a/dts/framework/test_suite.py b/dts/framework/test_suite.py
> index 0705f38f..5df5d2a6 100644
> --- a/dts/framework/test_suite.py
> +++ b/dts/framework/test_suite.py
> @@ -11,10 +11,21 @@
>  import re
>  from types import MethodType
>
> -from .exception import ConfigurationError, SSHTimeoutError, 
> TestCaseVerifyError
> +from .exception import (
> +    BlockingTestSuiteError,
> +    ConfigurationError,
> +    SSHTimeoutError,
> +    TestCaseVerifyError,
> +)
>  from .logger import DTSLOG, getLogger
>  from .settings import SETTINGS
> -from .test_result import BuildTargetResult, Result, TestCaseResult, 
> TestSuiteResult
> +from .test_result import (
> +    BuildTargetResult,
> +    DTSResult,
> +    Result,
> +    TestCaseResult,
> +    TestSuiteResult,
> +)
>  from .testbed_model import SutNode
>
>
> @@ -37,10 +48,12 @@ class TestSuite(object):
>      """
>
>      sut_node: SutNode
> +    is_blocking = False
>      _logger: DTSLOG
>      _test_cases_to_run: list[str]
>      _func: bool
>      _result: TestSuiteResult
> +    _dts_result: DTSResult
>
>      def __init__(
>          self,
> @@ -48,6 +61,7 @@ def __init__(
>          test_cases: list[str],
>          func: bool,
>          build_target_result: BuildTargetResult,
> +        dts_result: DTSResult,
>      ):
>          self.sut_node = sut_node
>          self._logger = getLogger(self.__class__.__name__)
> @@ -55,6 +69,7 @@ def __init__(
>          self._test_cases_to_run.extend(SETTINGS.test_cases)
>          self._func = func
>          self._result = 
> build_target_result.add_test_suite(self.__class__.__name__)
> +        self._dts_result = dts_result
>
>      def set_up_suite(self) -> None:
>          """
> @@ -118,6 +133,8 @@ def run(self) -> None:
>                      f"the next test suite may be affected."
>                  )
>                  self._result.update_setup(Result.ERROR, e)
> +            if len(self._result.get_errors()) > 0 and self.is_blocking:
> +                raise BlockingTestSuiteError(test_suite_name)
>
>      def _execute_test_suite(self) -> None:
>          """
> diff --git a/dts/framework/testbed_model/node.py 
> b/dts/framework/testbed_model/node.py
> index d48fafe6..c5147e0e 100644
> --- a/dts/framework/testbed_model/node.py
> +++ b/dts/framework/testbed_model/node.py
> @@ -40,6 +40,7 @@ class Node(object):
>      lcores: list[LogicalCore]
>      _logger: DTSLOG
>      _other_sessions: list[OSSession]
> +    _execution_config: ExecutionConfiguration
>
>      def __init__(self, node_config: NodeConfiguration):
>          self.config = node_config
> @@ -64,6 +65,7 @@ def set_up_execution(self, execution_config: 
> ExecutionConfiguration) -> None:
>          """
>          self._setup_hugepages()
>          self._set_up_execution(execution_config)
> +        self._execution_config = execution_config
>
>      def _set_up_execution(self, execution_config: ExecutionConfiguration) -> 
> None:
>          """
> diff --git a/dts/framework/testbed_model/sut_node.py 
> b/dts/framework/testbed_model/sut_node.py
> index 2b2b50d9..9b17ac3d 100644
> --- a/dts/framework/testbed_model/sut_node.py
> +++ b/dts/framework/testbed_model/sut_node.py
> @@ -1,14 +1,27 @@
>  # SPDX-License-Identifier: BSD-3-Clause
>  # Copyright(c) 2010-2014 Intel Corporation
>  # Copyright(c) 2023 PANTHEON.tech s.r.o.
> +# Copyright(c) 2022-2023 University of New Hampshire
>
>  import os
>  import tarfile
>  import time
>  from pathlib import PurePath
> -
> -from framework.config import BuildTargetConfiguration, NodeConfiguration
> -from framework.remote_session import CommandResult, OSSession
> +from typing import Union
> +
> +from framework.config import (
> +    BuildTargetConfiguration,
> +    BuildTargetInfo,
> +    InteractiveApp,
> +    NodeConfiguration,
> +    NodeInfo,
> +)
> +from framework.remote_session import (
> +    CommandResult,
> +    InteractiveShell,
> +    OSSession,
> +    TestPmdShell,
> +)
>  from framework.settings import SETTINGS
>  from framework.utils import EnvVarsDict, MesonArgs
>
> @@ -16,6 +29,52 @@
>  from .node import Node
>
>
> +class EalParameters(object):
> +    def __init__(
> +        self,
> +        lcore_list: LogicalCoreList,
> +        memory_channels: int,
> +        prefix: str,
> +        no_pci: bool,
> +        vdevs: list[VirtualDevice],
> +        other_eal_param: str,
> +    ):
> +        """
> +        Generate eal parameters character string;
> +        :param lcore_list: the list of logical cores to use.
> +        :param memory_channels: the number of memory channels to use.
> +        :param prefix: set file prefix string, eg:
> +                        prefix='vf'
> +        :param no_pci: switch of disable PCI bus eg:
> +                        no_pci=True
> +        :param vdevs: virtual device list, eg:
> +                        vdevs=[
> +                            VirtualDevice('net_ring0'),
> +                            VirtualDevice('net_ring1')
> +                        ]
> +        :param other_eal_param: user defined DPDK eal parameters, eg:
> +                        other_eal_param='--single-file-segments'
> +        """
> +        self._lcore_list = f"-l {lcore_list}"
> +        self._memory_channels = f"-n {memory_channels}"
> +        self._prefix = prefix
> +        if prefix:
> +            self._prefix = f"--file-prefix={prefix}"
> +        self._no_pci = "--no-pci" if no_pci else ""
> +        self._vdevs = " ".join(f"--vdev {vdev}" for vdev in vdevs)
> +        self._other_eal_param = other_eal_param
> +
> +    def __str__(self) -> str:
> +        return (
> +            f"{self._lcore_list} "
> +            f"{self._memory_channels} "
> +            f"{self._prefix} "
> +            f"{self._no_pci} "
> +            f"{self._vdevs} "
> +            f"{self._other_eal_param}"
> +        )
> +
> +
>  class SutNode(Node):
>      """
>      A class for managing connections to the System under Test, providing
> @@ -30,9 +89,11 @@ class SutNode(Node):
>      _env_vars: EnvVarsDict
>      _remote_tmp_dir: PurePath
>      __remote_dpdk_dir: PurePath | None
> -    _dpdk_version: str | None
>      _app_compile_timeout: float
>      _dpdk_kill_session: OSSession | None
> +    _dpdk_version: str | None
> +    _node_info: NodeInfo | None
> +    _compiler_version: str | None
>
>      def __init__(self, node_config: NodeConfiguration):
>          super(SutNode, self).__init__(node_config)
> @@ -41,12 +102,14 @@ def __init__(self, node_config: NodeConfiguration):
>          self._env_vars = EnvVarsDict()
>          self._remote_tmp_dir = self.main_session.get_remote_tmp_dir()
>          self.__remote_dpdk_dir = None
> -        self._dpdk_version = None
>          self._app_compile_timeout = 90
>          self._dpdk_kill_session = None
>          self._dpdk_timestamp = (
>              f"{str(os.getpid())}_{time.strftime('%Y%m%d%H%M%S', 
> time.localtime())}"
>          )
> +        self._dpdk_version = None
> +        self._node_info = None
> +        self._compiler_version = None
>
>      @property
>      def _remote_dpdk_dir(self) -> PurePath:
> @@ -75,6 +138,32 @@ def dpdk_version(self) -> str:
>              )
>          return self._dpdk_version
>
> +    @property
> +    def node_info(self) -> NodeInfo:
> +        if self._node_info is None:
> +            self._node_info = self.main_session.get_node_info()
> +        return self._node_info
> +
> +    @property
> +    def compiler_version(self) -> str:
> +        if self._compiler_version is None:
> +            if self._build_target_config is not None:
> +                self._compiler_version = 
> self.main_session.get_compiler_version(
> +                    self._build_target_config.compiler.name
> +                )
> +            else:
> +                self._logger.warning(
> +                    "Failed to get compiler version because"
> +                    "_build_target_config is None."
> +                )
> +                return ""
> +        return self._compiler_version
> +
> +    def get_build_target_info(self) -> BuildTargetInfo:
> +        return BuildTargetInfo(
> +            dpdk_version=self.dpdk_version, 
> compiler_version=self.compiler_version
> +        )
> +
>      def _guess_dpdk_remote_dir(self) -> PurePath:
>          return self.main_session.guess_dpdk_remote_dir(self._remote_tmp_dir)
>
> @@ -84,6 +173,10 @@ def _set_up_build_target(
>          """
>          Setup DPDK on the SUT node.
>          """
> +        # we want to ensure that dpdk_version and compiler_version is reset 
> for new
> +        # build targets
> +        self._dpdk_version = None
> +        self._compiler_version = None
>          self._configure_build_target(build_target_config)
>          self._copy_dpdk_tarball()
>          self._build_dpdk()
> @@ -262,48 +355,43 @@ def run_dpdk_app(
>              f"{app_path} {eal_args}", timeout, verify=True
>          )
>
> -
> -class EalParameters(object):
> -    def __init__(
> +    def create_interactive_shell(
>          self,
> -        lcore_list: LogicalCoreList,
> -        memory_channels: int,
> -        prefix: str,
> -        no_pci: bool,
> -        vdevs: list[VirtualDevice],
> -        other_eal_param: str,
> -    ):
> +        shell_type: InteractiveApp,
> +        path_to_app: PurePath | None = None,
> +        timeout: float = SETTINGS.timeout,
> +        eal_parameters: EalParameters | None = None,
> +    ) -> Union[InteractiveShell, TestPmdShell]:
> +        """Create a handler for an interactive session.
> +
> +        This method is a factory that calls a method in OSSession to create 
> shells for
> +        different DPDK applications.
> +
> +        Args:
> +            shell_type: Enum value representing the desired application.
> +            path_to_app: Represents a path to the application you are 
> attempting to
> +                launch. This path will be executed at the start of the app
> +                initialization. If one isn't provided, the default specified 
> in the
> +                enumeration will be used.
> +            timeout: Timeout for reading output from the SSH channel. If you 
> are
> +                reading from the buffer and don't receive any data within 
> the timeout
> +                it will throw an error.
> +            eal_parameters: List of EAL parameters to use to launch the app. 
> This is
> +                ignored for base "shell" types.
> +        Returns:
> +            Instance of the desired interactive application.
>          """
> -        Generate eal parameters character string;
> -        :param lcore_list: the list of logical cores to use.
> -        :param memory_channels: the number of memory channels to use.
> -        :param prefix: set file prefix string, eg:
> -                        prefix='vf'
> -        :param no_pci: switch of disable PCI bus eg:
> -                        no_pci=True
> -        :param vdevs: virtual device list, eg:
> -                        vdevs=[
> -                            VirtualDevice('net_ring0'),
> -                            VirtualDevice('net_ring1')
> -                        ]
> -        :param other_eal_param: user defined DPDK eal parameters, eg:
> -                        other_eal_param='--single-file-segments'
> -        """
> -        self._lcore_list = f"-l {lcore_list}"
> -        self._memory_channels = f"-n {memory_channels}"
> -        self._prefix = prefix
> -        if prefix:
> -            self._prefix = f"--file-prefix={prefix}"
> -        self._no_pci = "--no-pci" if no_pci else ""
> -        self._vdevs = " ".join(f"--vdev {vdev}" for vdev in vdevs)
> -        self._other_eal_param = other_eal_param
> +        # if we just want a default shell, there is no need to append the 
> DPDK build
> +        # directory to the path
> +        default_path = shell_type.get_path()
>
> -    def __str__(self) -> str:
> -        return (
> -            f"{self._lcore_list} "
> -            f"{self._memory_channels} "
> -            f"{self._prefix} "
> -            f"{self._no_pci} "
> -            f"{self._vdevs} "
> -            f"{self._other_eal_param}"
> +        if shell_type != InteractiveApp.shell:
> +            default_path = self.main_session.join_remote_path(
> +                self.remote_dpdk_build_dir, shell_type.get_path()
> +            )
> +        return 
> self.main_session.interactive_session.create_interactive_shell(
> +            shell_type,
> +            path_to_app if path_to_app else default_path,
> +            str(eal_parameters) if eal_parameters else "",
> +            timeout,
>          )

I forgot to mention that I'd like to change to structure a bit here -
calling self.main_session.create_interactive_shell() would makes more
sense to me, as the interactive_session is basically an implementation
detail.

> diff --git a/dts/tests/TestSuite_smoke_tests.py 
> b/dts/tests/TestSuite_smoke_tests.py
> new file mode 100644
> index 00000000..b7e70ee1
> --- /dev/null
> +++ b/dts/tests/TestSuite_smoke_tests.py
> @@ -0,0 +1,118 @@
> +# SPDX-License-Identifier: BSD-3-Clause
> +# Copyright(c) 2023 University of New Hampshire
> +
> +import re
> +
> +from framework.config import InteractiveApp
> +from framework.remote_session import TestPmdShell
> +from framework.settings import SETTINGS
> +from framework.test_suite import TestSuite
> +
> +
> +class SmokeTests(TestSuite):
> +    is_blocking = True
> +    # dicts in this list are expected to have two keys:
> +    # "pci_address" and "current_driver"
> +    nics_in_node: list[dict[str, str]] = []
> +
> +    def set_up_suite(self) -> None:
> +        """
> +        Setup:
> +            Set the build directory path and generate a list of NICs in the 
> SUT node.
> +        """
> +        self.dpdk_build_dir_path = self.sut_node.remote_dpdk_build_dir
> +        for nic in self.sut_node.config.ports:
> +            new_dict = {
> +                "pci_address": nic.pci,
> +                "current_driver": nic.os_driver.strip(),
> +            }
> +            self.nics_in_node.append(new_dict)
> +
> +    def test_unit_tests(self) -> None:
> +        """
> +        Test:
> +            Run the fast-test unit-test suite through meson.
> +        """
> +        self.sut_node.main_session.send_command(
> +            f"meson test -C {self.dpdk_build_dir_path} --suite fast-tests",
> +            300,
> +            verify=True,
> +        )
> +
> +    def test_driver_tests(self) -> None:
> +        """
> +        Test:
> +            Run the driver-test unit-test suite through meson.
> +        """
> +        list_of_vdevs = ""
> +        for dev in self.sut_node._execution_config.vdevs:
> +            list_of_vdevs += f"--vdev {dev} "
> +        list_of_vdevs = list_of_vdevs[:-1]
> +        if list_of_vdevs:
> +            self._logger.info(
> +                "Running driver tests with the following virtual "
> +                f"devices: {list_of_vdevs}"
> +            )
> +            self.sut_node.main_session.send_command(
> +                f"meson test -C {self.dpdk_build_dir_path} --suite 
> driver-tests "
> +                f'--test-args "{list_of_vdevs}"',
> +                300,
> +                verify=True,
> +            )
> +        else:
> +            self.sut_node.main_session.send_command(
> +                f"meson test -C {self.dpdk_build_dir_path} --suite 
> driver-tests",
> +                300,
> +                verify=True,
> +            )
> +
> +    def test_devices_listed_in_testpmd(self) -> None:
> +        """
> +        Test:
> +            Uses testpmd driver to verify that devices have been found by 
> testpmd.
> +        """
> +        testpmd_driver = 
> self.sut_node.create_interactive_shell(InteractiveApp.testpmd)
> +        # We know it should always be a TestPmdShell but mypy doesn't
> +        assert isinstance(testpmd_driver, TestPmdShell)
> +        dev_list: list[str] = testpmd_driver.get_devices()
> +        for nic in self.nics_in_node:
> +            self.verify(
> +                nic["pci_address"] in dev_list,
> +                f"Device {nic['pci_address']} was not listed in testpmd's 
> available devices, "
> +                "please check your configuration",
> +            )
> +
> +    def test_device_bound_to_driver(self) -> None:
> +        """
> +        Test:
> +            Ensure that all drivers listed in the config are bound to the 
> correct driver.
> +        """
> +        path_to_devbind = self.sut_node.main_session.join_remote_path(
> +            self.sut_node._remote_dpdk_dir, "usertools", "dpdk-devbind.py"
> +        )
> +
> +        regex_for_pci_address = "/[0-9]{4}:[0-9]{2}:[0-9]{2}.[0-9]{1}/"

This shouldn't be tucked away in a test case. It should be in some
high-level module - let's put it to utils.py for the time being.

> +        all_nics_in_dpdk_devbind = self.sut_node.main_session.send_command(
> +            f"{path_to_devbind} --status | awk '{regex_for_pci_address}'",
> +            SETTINGS.timeout,
> +        ).stdout
> +
> +        for nic in self.nics_in_node:
> +            # This regular expression finds the line in the above string 
> that starts
> +            # with the address for the nic we are on in the loop and then 
> captures the
> +            # name of the driver in a group
> +            devbind_info_for_nic = re.search(
> +                f"{nic['pci_address']}[^\\n]*drv=([\\d\\w]*) [^\\n]*",
> +                all_nics_in_dpdk_devbind,
> +            )
> +            self.verify(
> +                devbind_info_for_nic is not None,
> +                f"Failed to find configured device ({nic['pci_address']}) 
> using dpdk-devbind.py",
> +            )
> +            # We know this isn't None, but mypy doesn't
> +            assert devbind_info_for_nic is not None
> +            self.verify(
> +                devbind_info_for_nic.group(1) == nic["current_driver"],
> +                f"Driver for device {nic['pci_address']} does not match 
> driver listed in "
> +                f"configuration (bound to {devbind_info_for_nic.group(1)})",
> +            )
> --
> 2.41.0
>

Reply via email to