From: Jeremy Spewock <jspew...@iol.unh.edu>

Adds a new test suite for running smoke tests that verify general
configuration aspects of the system under test. If any of these tests
fail, the DTS execution terminates as part of a "fail-fast" model.

Signed-off-by: Jeremy Spewock <jspew...@iol.unh.edu>
---
 dts/conf.yaml                                 |  17 +-
 dts/framework/config/__init__.py              | 116 +++++++++--
 dts/framework/config/conf_yaml_schema.json    | 142 +++++++++++++-
 dts/framework/dts.py                          |  88 ++++++---
 dts/framework/exception.py                    |  12 ++
 dts/framework/remote_session/__init__.py      |  10 +-
 dts/framework/remote_session/os_session.py    |  24 ++-
 dts/framework/remote_session/posix_session.py |  29 ++-
 .../remote_session/remote/__init__.py         |  10 +
 .../remote/interactive_remote_session.py      | 118 ++++++++++++
 .../remote/interactive_shell.py               |  99 ++++++++++
 .../remote_session/remote/testpmd_shell.py    |  67 +++++++
 dts/framework/test_result.py                  |  37 +++-
 dts/framework/test_suite.py                   |  21 +-
 dts/framework/testbed_model/node.py           |   2 +
 dts/framework/testbed_model/sut_node.py       | 180 +++++++++++++-----
 dts/tests/TestSuite_smoke_tests.py            | 118 ++++++++++++
 17 files changed, 994 insertions(+), 96 deletions(-)
 create mode 100644 
dts/framework/remote_session/remote/interactive_remote_session.py
 create mode 100644 dts/framework/remote_session/remote/interactive_shell.py
 create mode 100644 dts/framework/remote_session/remote/testpmd_shell.py
 create mode 100644 dts/tests/TestSuite_smoke_tests.py

diff --git a/dts/conf.yaml b/dts/conf.yaml
index a9bd8a3e..2717de13 100644
--- a/dts/conf.yaml
+++ b/dts/conf.yaml
@@ -10,9 +10,13 @@ executions:
         compiler_wrapper: ccache
     perf: false
     func: true
+    skip_smoke_tests: false #optional flag that allow you to ski smoke tests
     test_suites:
       - hello_world
-    system_under_test: "SUT 1"
+    system_under_test:
+      node_name: "SUT 1"
+      vdevs: #optional: if removed vdevs won't be used in the execution
+        - "crypto_openssl"
 nodes:
   - name: "SUT 1"
     hostname: sut1.change.me.localhost
@@ -20,6 +24,17 @@ nodes:
     arch: x86_64
     os: linux
     lcores: ""
+    ports:
+      - pci: "0000:00:08.0"
+        os_driver_for_dpdk: vfio-pci #OS driver that DPDK will use
+        os_driver: i40e
+        peer_node: "TG 1"
+        peer_pci: "0000:00:08.0"
+      - pci: "0000:00:08.1"
+        os_driver_for_dpdk: vfio-pci
+        os_driver: i40e
+        peer_node: "TG 1"
+        peer_pci: "0000:00:08.1"
     use_first_core: false
     memory_channels: 4
     hugepages:  # optional; if removed, will use system hugepage configuration
diff --git a/dts/framework/config/__init__.py b/dts/framework/config/__init__.py
index ebb0823f..75ac1cbe 100644
--- a/dts/framework/config/__init__.py
+++ b/dts/framework/config/__init__.py
@@ -12,6 +12,7 @@
 import pathlib
 from dataclasses import dataclass
 from enum import Enum, auto, unique
+from pathlib import PurePath
 from typing import Any, TypedDict
 
 import warlock  # type: ignore
@@ -72,6 +73,20 @@ class HugepageConfiguration:
     force_first_numa: bool
 
 
+@dataclass(slots=True, frozen=True)
+class PortConfig:
+    node: str
+    pci: str
+    os_driver_for_dpdk: str
+    os_driver: str
+    peer_node: str
+    peer_pci: str
+
+    @staticmethod
+    def from_dict(node: str, d: dict) -> "PortConfig":
+        return PortConfig(node=node, **d)
+
+
 @dataclass(slots=True, frozen=True)
 class NodeConfiguration:
     name: str
@@ -84,6 +99,7 @@ class NodeConfiguration:
     use_first_core: bool
     memory_channels: int
     hugepages: HugepageConfiguration | None
+    ports: list[PortConfig]
 
     @staticmethod
     def from_dict(d: dict) -> "NodeConfiguration":
@@ -92,18 +108,43 @@ def from_dict(d: dict) -> "NodeConfiguration":
             if "force_first_numa" not in hugepage_config:
                 hugepage_config["force_first_numa"] = False
             hugepage_config = HugepageConfiguration(**hugepage_config)
+        common_config = {
+            "name": d["name"],
+            "hostname": d["hostname"],
+            "user": d["user"],
+            "password": d.get("password"),
+            "arch": Architecture(d["arch"]),
+            "os": OS(d["os"]),
+            "lcores": d.get("lcores", "1"),
+            "use_first_core": d.get("use_first_core", False),
+            "memory_channels": d.get("memory_channels", 1),
+            "hugepages": hugepage_config,
+            "ports": [PortConfig.from_dict(d["name"], port) for port in 
d["ports"]],
+        }
+
+        return NodeConfiguration(**common_config)
+
+
+@dataclass(slots=True)
+class NodeInfo:
+    """Class to hold important versions within the node.
+
+    This class, unlike the NodeConfiguration class, cannot be generated at the 
start.
+    This is because we need to initialize a connection with the node before we 
can
+    collect the information needed in this class. Therefore, it cannot be a 
part of
+    the configuration class above.
+    """
 
-        return NodeConfiguration(
-            name=d["name"],
-            hostname=d["hostname"],
-            user=d["user"],
-            password=d.get("password"),
-            arch=Architecture(d["arch"]),
-            os=OS(d["os"]),
-            lcores=d.get("lcores", "1"),
-            use_first_core=d.get("use_first_core", False),
-            memory_channels=d.get("memory_channels", 1),
-            hugepages=hugepage_config,
+    os_name: str
+    os_version: str
+    kernel_version: str
+
+    @staticmethod
+    def from_dict(d: dict):
+        return NodeInfo(
+            os_name=d["os_name"],
+            os_version=d["os_version"],
+            kernel_version=d["kernel_version"],
         )
 
 
@@ -128,6 +169,24 @@ def from_dict(d: dict) -> "BuildTargetConfiguration":
         )
 
 
+@dataclass(slots=True)
+class BuildTargetInfo:
+    """Class to hold important versions within the build target.
+
+    This is very similar to the NodeVersionInfo class, it just instead holds 
information
+    for the build target.
+    """
+
+    dpdk_version: str
+    compiler_version: str
+
+    @staticmethod
+    def from_dict(d: dict):
+        return BuildTargetInfo(
+            dpdk_version=d["dpdk_version"], 
compiler_version=d["compiler_version"]
+        )
+
+
 class TestSuiteConfigDict(TypedDict):
     suite: str
     cases: list[str]
@@ -157,6 +216,8 @@ class ExecutionConfiguration:
     func: bool
     test_suites: list[TestSuiteConfig]
     system_under_test: NodeConfiguration
+    vdevs: list[str]
+    skip_smoke_tests: bool
 
     @staticmethod
     def from_dict(d: dict, node_map: dict) -> "ExecutionConfiguration":
@@ -166,15 +227,20 @@ def from_dict(d: dict, node_map: dict) -> 
"ExecutionConfiguration":
         test_suites: list[TestSuiteConfig] = list(
             map(TestSuiteConfig.from_dict, d["test_suites"])
         )
-        sut_name = d["system_under_test"]
+        sut_name = d["system_under_test"]["node_name"]
+        skip_smoke_tests = d.get("skip_smoke_tests", False)
         assert sut_name in node_map, f"Unknown SUT {sut_name} in execution {d}"
-
+        vdevs = (
+            d["system_under_test"]["vdevs"] if "vdevs" in 
d["system_under_test"] else []
+        )
         return ExecutionConfiguration(
             build_targets=build_targets,
             perf=d["perf"],
             func=d["func"],
+            skip_smoke_tests=skip_smoke_tests,
             test_suites=test_suites,
             system_under_test=node_map[sut_name],
+            vdevs=vdevs,
         )
 
 
@@ -221,3 +287,27 @@ def load_config() -> Configuration:
 
 
 CONFIGURATION = load_config()
+
+
+@unique
+class InteractiveApp(Enum):
+    """An enum that represents different supported interactive applications
+
+    The values in this enum must all be set to objects that have a key called
+    "default_path" where "default_path" represents a PurPath object for the 
path
+    to the application. This default path will be passed into the handler class
+    for the application so that it can start the application. For every key 
other
+    than the default shell option, the path will be appended to the path to 
the DPDK
+    build directory for the current SUT node.
+    """
+
+    shell = {"default_path": PurePath()}
+    testpmd = {"default_path": PurePath("app", "dpdk-testpmd")}
+
+    def get_path(self) -> PurePath:
+        """A method for getting the default paths of an application
+
+        Returns:
+            String array that represents an OS agnostic path to the 
application.
+        """
+        return self.value["default_path"]
diff --git a/dts/framework/config/conf_yaml_schema.json 
b/dts/framework/config/conf_yaml_schema.json
index ca2d4a1e..61f52b43 100644
--- a/dts/framework/config/conf_yaml_schema.json
+++ b/dts/framework/config/conf_yaml_schema.json
@@ -6,6 +6,76 @@
       "type": "string",
       "description": "A unique identifier for a node"
     },
+    "NIC": {
+      "type": "string",
+      "enum": [
+        "ALL",
+        "ConnectX3_MT4103",
+        "ConnectX4_LX_MT4117",
+        "ConnectX4_MT4115",
+        "ConnectX5_MT4119",
+        "ConnectX5_MT4121",
+        "I40E_10G-10G_BASE_T_BC",
+        "I40E_10G-10G_BASE_T_X722",
+        "I40E_10G-SFP_X722",
+        "I40E_10G-SFP_XL710",
+        "I40E_10G-X722_A0",
+        "I40E_1G-1G_BASE_T_X722",
+        "I40E_25G-25G_SFP28",
+        "I40E_40G-QSFP_A",
+        "I40E_40G-QSFP_B",
+        "IAVF-ADAPTIVE_VF",
+        "IAVF-VF",
+        "IAVF_10G-X722_VF",
+        "ICE_100G-E810C_QSFP",
+        "ICE_25G-E810C_SFP",
+        "ICE_25G-E810_XXV_SFP",
+        "IGB-I350_VF",
+        "IGB_1G-82540EM",
+        "IGB_1G-82545EM_COPPER",
+        "IGB_1G-82571EB_COPPER",
+        "IGB_1G-82574L",
+        "IGB_1G-82576",
+        "IGB_1G-82576_QUAD_COPPER",
+        "IGB_1G-82576_QUAD_COPPER_ET2",
+        "IGB_1G-82580_COPPER",
+        "IGB_1G-I210_COPPER",
+        "IGB_1G-I350_COPPER",
+        "IGB_1G-I354_SGMII",
+        "IGB_1G-PCH_LPTLP_I218_LM",
+        "IGB_1G-PCH_LPTLP_I218_V",
+        "IGB_1G-PCH_LPT_I217_LM",
+        "IGB_1G-PCH_LPT_I217_V",
+        "IGB_2.5G-I354_BACKPLANE_2_5GBPS",
+        "IGC-I225_LM",
+        "IGC-I226_LM",
+        "IXGBE_10G-82599_SFP",
+        "IXGBE_10G-82599_SFP_SF_QP",
+        "IXGBE_10G-82599_T3_LOM",
+        "IXGBE_10G-82599_VF",
+        "IXGBE_10G-X540T",
+        "IXGBE_10G-X540_VF",
+        "IXGBE_10G-X550EM_A_SFP",
+        "IXGBE_10G-X550EM_X_10G_T",
+        "IXGBE_10G-X550EM_X_SFP",
+        "IXGBE_10G-X550EM_X_VF",
+        "IXGBE_10G-X550T",
+        "IXGBE_10G-X550_VF",
+        "brcm_57414",
+        "brcm_P2100G",
+        "cavium_0011",
+        "cavium_a034",
+        "cavium_a063",
+        "cavium_a064",
+        "fastlinq_ql41000",
+        "fastlinq_ql41000_vf",
+        "fastlinq_ql45000",
+        "fastlinq_ql45000_vf",
+        "hi1822",
+        "virtio"
+      ]
+    },
+
     "ARCH": {
       "type": "string",
       "enum": [
@@ -94,6 +164,19 @@
         "amount"
       ]
     },
+    "pci_address": {
+      "type": "string",
+      "pattern": "^[\\da-fA-F]{4}:[\\da-fA-F]{2}:[\\da-fA-F]{2}.\\d:?\\w*$"
+    },
+    "port_peer_address": {
+      "description": "Peer is a TRex port, and IXIA port or a PCI address",
+      "oneOf": [
+        {
+          "description": "PCI peer port",
+          "$ref": "#/definitions/pci_address"
+        }
+      ]
+    },
     "test_suite": {
       "type": "string",
       "enum": [
@@ -165,6 +248,44 @@
           },
           "hugepages": {
             "$ref": "#/definitions/hugepages"
+          },
+          "ports": {
+            "type": "array",
+            "items": {
+              "type": "object",
+              "description": "Each port should be described on both sides of 
the connection. This makes configuration slightly more verbose but greatly 
simplifies implementation. If there are an inconsistencies, then DTS will not 
run until that issue is fixed. An example inconsistency would be port 1, node 1 
says it is connected to port 1, node 2, but port 1, node 2 says it is connected 
to port 2, node 1.",
+              "properties": {
+                "pci": {
+                  "$ref": "#/definitions/pci_address",
+                  "description": "The local PCI address of the port"
+                },
+                "os_driver_for_dpdk": {
+                  "type": "string",
+                  "description": "The driver that the kernel should bind this 
device to for DPDK to use it. (ex: vfio-pci)"
+                },
+                "os_driver": {
+                  "type": "string",
+                  "description": "The driver normally used by this port (ex: 
i40e)"
+                },
+                "peer_node": {
+                  "type": "string",
+                  "description": "The name of the node the peer port is on"
+                },
+                "peer_pci": {
+                  "$ref": "#/definitions/pci_address",
+                  "description": "The PCI address of the peer port"
+                }
+              },
+              "additionalProperties": false,
+              "required": [
+                "pci",
+                "os_driver_for_dpdk",
+                "os_driver",
+                "peer_node",
+                "peer_pci"
+              ]
+            },
+            "minimum": 1
           }
         },
         "additionalProperties": false,
@@ -211,8 +332,27 @@
               ]
             }
           },
+          "skip_smoke_tests": {
+            "description": "Optional field that allows you to skip smoke 
testing",
+            "type": "boolean"
+          },
           "system_under_test": {
-            "$ref": "#/definitions/node_name"
+            "type":"object",
+            "properties": {
+              "node_name": {
+                "$ref": "#/definitions/node_name"
+              },
+              "vdevs": {
+                "description": "Opentional list of names of vdevs to be used 
in execution",
+                "type": "array",
+                "items": {
+                  "type": "string"
+                }
+              }
+            },
+            "required": [
+              "node_name"
+            ]
           }
         },
         "additionalProperties": false,
diff --git a/dts/framework/dts.py b/dts/framework/dts.py
index 05022845..1b67938f 100644
--- a/dts/framework/dts.py
+++ b/dts/framework/dts.py
@@ -5,7 +5,13 @@
 
 import sys
 
-from .config import CONFIGURATION, BuildTargetConfiguration, 
ExecutionConfiguration
+from .config import (
+    CONFIGURATION,
+    BuildTargetConfiguration,
+    ExecutionConfiguration,
+    TestSuiteConfig,
+)
+from .exception import BlockingTestSuiteError
 from .logger import DTSLOG, getLogger
 from .test_result import BuildTargetResult, DTSResult, ExecutionResult, Result
 from .test_suite import get_test_suites
@@ -82,7 +88,7 @@ def _run_execution(
     running all build targets in the given execution.
     """
     dts_logger.info(f"Running execution with SUT 
'{execution.system_under_test.name}'.")
-    execution_result = result.add_execution(sut_node.config)
+    execution_result = result.add_execution(sut_node.config, 
sut_node.node_info)
 
     try:
         sut_node.set_up_execution(execution)
@@ -118,14 +124,15 @@ def _run_build_target(
 
     try:
         sut_node.set_up_build_target(build_target)
-        result.dpdk_version = sut_node.dpdk_version
+        # result.dpdk_version = sut_node.dpdk_version
+        
build_target_result.add_build_target_versions(sut_node.get_build_target_info())
         build_target_result.update_setup(Result.PASS)
     except Exception as e:
         dts_logger.exception("Build target setup failed.")
         build_target_result.update_setup(Result.FAIL, e)
 
     else:
-        _run_suites(sut_node, execution, build_target_result)
+        _run_all_suites(sut_node, execution, build_target_result)
 
     finally:
         try:
@@ -136,7 +143,7 @@ def _run_build_target(
             build_target_result.update_teardown(Result.FAIL, e)
 
 
-def _run_suites(
+def _run_all_suites(
     sut_node: SutNode,
     execution: ExecutionConfiguration,
     build_target_result: BuildTargetResult,
@@ -146,27 +153,62 @@ def _run_suites(
     with possibly only a subset of test cases.
     If no subset is specified, run all test cases.
     """
+    end_build_target = False
+    if not execution.skip_smoke_tests:
+        execution.test_suites[:0] = [TestSuiteConfig.from_dict("smoke_tests")]
     for test_suite_config in execution.test_suites:
         try:
-            full_suite_path = f"tests.TestSuite_{test_suite_config.test_suite}"
-            test_suite_classes = get_test_suites(full_suite_path)
-            suites_str = ", ".join((x.__name__ for x in test_suite_classes))
-            dts_logger.debug(
-                f"Found test suites '{suites_str}' in '{full_suite_path}'."
+            _run_single_suite(
+                sut_node, execution, build_target_result, test_suite_config
             )
-        except Exception as e:
-            dts_logger.exception("An error occurred when searching for test 
suites.")
-            result.update_setup(Result.ERROR, e)
-
-        else:
-            for test_suite_class in test_suite_classes:
-                test_suite = test_suite_class(
-                    sut_node,
-                    test_suite_config.test_cases,
-                    execution.func,
-                    build_target_result,
-                )
-                test_suite.run()
+        except BlockingTestSuiteError as e:
+            dts_logger.exception(
+                f"An error occurred within {test_suite_config.test_suite}. "
+                "Skipping build target..."
+            )
+            result.add_error(e)
+            end_build_target = True
+        # if a blocking test failed and we need to bail out of suite executions
+        if end_build_target:
+            break
+
+
+def _run_single_suite(
+    sut_node: SutNode,
+    execution: ExecutionConfiguration,
+    build_target_result: BuildTargetResult,
+    test_suite_config: TestSuiteConfig,
+) -> None:
+    """Runs a single test suite.
+
+    Args:
+        sut_node: Node to run tests on.
+        execution: Execution the test case belongs to.
+        build_target_result: Build target configuration test case is run on
+        test_suite_config: Test suite configuration
+
+    Raises:
+        BlockingTestSuiteError: If a test suite that was marked as blocking 
fails.
+    """
+    try:
+        full_suite_path = f"tests.TestSuite_{test_suite_config.test_suite}"
+        test_suite_classes = get_test_suites(full_suite_path)
+        suites_str = ", ".join((x.__name__ for x in test_suite_classes))
+        dts_logger.debug(f"Found test suites '{suites_str}' in 
'{full_suite_path}'.")
+    except Exception as e:
+        dts_logger.exception("An error occurred when searching for test 
suites.")
+        result.update_setup(Result.ERROR, e)
+
+    else:
+        for test_suite_class in test_suite_classes:
+            test_suite = test_suite_class(
+                sut_node,
+                test_suite_config.test_cases,
+                execution.func,
+                build_target_result,
+                result,
+            )
+            test_suite.run()
 
 
 def _exit_dts() -> None:
diff --git a/dts/framework/exception.py b/dts/framework/exception.py
index ca353d98..dfb12df4 100644
--- a/dts/framework/exception.py
+++ b/dts/framework/exception.py
@@ -25,6 +25,7 @@ class ErrorSeverity(IntEnum):
     SSH_ERR = 4
     DPDK_BUILD_ERR = 10
     TESTCASE_VERIFY_ERR = 20
+    BLOCKING_TESTSUITE_ERR = 25
 
 
 class DTSError(Exception):
@@ -144,3 +145,14 @@ def __init__(self, value: str):
 
     def __str__(self) -> str:
         return repr(self.value)
+
+
+class BlockingTestSuiteError(DTSError):
+    suite_name: str
+    severity: ClassVar[ErrorSeverity] = ErrorSeverity.BLOCKING_TESTSUITE_ERR
+
+    def __init__(self, suite_name: str) -> None:
+        self.suite_name = suite_name
+
+    def __str__(self) -> str:
+        return f"Blocking suite {self.suite_name} failed."
diff --git a/dts/framework/remote_session/__init__.py 
b/dts/framework/remote_session/__init__.py
index ee221503..4fe32d35 100644
--- a/dts/framework/remote_session/__init__.py
+++ b/dts/framework/remote_session/__init__.py
@@ -1,5 +1,6 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2023 PANTHEON.tech s.r.o.
+# Copyright(c) 2023 University of New Hampshire
 
 """
 The package provides modules for managing remote connections to a remote host 
(node),
@@ -17,7 +18,14 @@
 
 from .linux_session import LinuxSession
 from .os_session import OSSession
-from .remote import CommandResult, RemoteSession, SSHSession
+from .remote import (
+    CommandResult,
+    InteractiveRemoteSession,
+    InteractiveShell,
+    RemoteSession,
+    SSHSession,
+    TestPmdShell,
+)
 
 
 def create_session(
diff --git a/dts/framework/remote_session/os_session.py 
b/dts/framework/remote_session/os_session.py
index 4c48ae25..4346ecc4 100644
--- a/dts/framework/remote_session/os_session.py
+++ b/dts/framework/remote_session/os_session.py
@@ -6,13 +6,19 @@
 from collections.abc import Iterable
 from pathlib import PurePath
 
-from framework.config import Architecture, NodeConfiguration
+from framework.config import Architecture, NodeConfiguration, NodeInfo
 from framework.logger import DTSLOG
 from framework.settings import SETTINGS
 from framework.testbed_model import LogicalCore
 from framework.utils import EnvVarsDict, MesonArgs
 
-from .remote import CommandResult, RemoteSession, create_remote_session
+from .remote import (
+    CommandResult,
+    InteractiveRemoteSession,
+    RemoteSession,
+    create_interactive_session,
+    create_remote_session,
+)
 
 
 class OSSession(ABC):
@@ -26,6 +32,7 @@ class OSSession(ABC):
     name: str
     _logger: DTSLOG
     remote_session: RemoteSession
+    interactive_session: InteractiveRemoteSession
 
     def __init__(
         self,
@@ -37,6 +44,7 @@ def __init__(
         self.name = name
         self._logger = logger
         self.remote_session = create_remote_session(node_config, name, logger)
+        self.interactive_session = create_interactive_session(node_config, 
name, logger)
 
     def close(self, force: bool = False) -> None:
         """
@@ -173,3 +181,15 @@ def setup_hugepages(self, hugepage_amount: int, 
force_first_numa: bool) -> None:
         if needed and mount the hugepages if needed.
         If force_first_numa is True, configure hugepages just on the first 
socket.
         """
+
+    @abstractmethod
+    def get_compiler_version(self, compiler_name: str) -> str:
+        """
+        Get installed version of compiler used for DPDK
+        """
+
+    @abstractmethod
+    def get_node_info(self) -> NodeInfo:
+        """
+        Collect information about the node
+        """
diff --git a/dts/framework/remote_session/posix_session.py 
b/dts/framework/remote_session/posix_session.py
index d38062e8..f8ec159f 100644
--- a/dts/framework/remote_session/posix_session.py
+++ b/dts/framework/remote_session/posix_session.py
@@ -6,7 +6,7 @@
 from collections.abc import Iterable
 from pathlib import PurePath, PurePosixPath
 
-from framework.config import Architecture
+from framework.config import Architecture, NodeInfo
 from framework.exception import DPDKBuildError, RemoteCommandExecutionError
 from framework.settings import SETTINGS
 from framework.utils import EnvVarsDict, MesonArgs
@@ -219,3 +219,30 @@ def _remove_dpdk_runtime_dirs(
 
     def get_dpdk_file_prefix(self, dpdk_prefix) -> str:
         return ""
+
+    def get_compiler_version(self, compiler_name: str) -> str:
+        match compiler_name:
+            case "gcc":
+                return self.send_command(f"{compiler_name} --version", 
60).stdout.split(
+                    "\n"
+                )[0]
+            case "clang":
+                return self.send_command(f"{compiler_name} --version", 
60).stdout.split(
+                    "\n"
+                )[0]
+            case "msvc":
+                return self.send_command("cl", 60).stdout
+            case "icc":
+                return self.send_command(f"{compiler_name} -V", 60).stdout
+            case _:
+                raise ValueError(f"Unknown compiler {compiler_name}")
+
+    def get_node_info(self) -> NodeInfo:
+        os_release_info = self.send_command(
+            "awk -F= '$1 ~ /^NAME$|^VERSION$/ {print $2}' /etc/os-release",
+            SETTINGS.timeout,
+        ).stdout.split("\n")
+        kernel_version = self.send_command("uname -r", SETTINGS.timeout).stdout
+        return NodeInfo(
+            os_release_info[0].strip(), os_release_info[1].strip(), 
kernel_version
+        )
diff --git a/dts/framework/remote_session/remote/__init__.py 
b/dts/framework/remote_session/remote/__init__.py
index 8a151221..224598a8 100644
--- a/dts/framework/remote_session/remote/__init__.py
+++ b/dts/framework/remote_session/remote/__init__.py
@@ -1,16 +1,26 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2023 PANTHEON.tech s.r.o.
+# Copyright(c) 2022-2023 University of New Hampshire
 
 # pylama:ignore=W0611
 
 from framework.config import NodeConfiguration
 from framework.logger import DTSLOG
 
+from .interactive_remote_session import InteractiveRemoteSession
+from .interactive_shell import InteractiveShell
 from .remote_session import CommandResult, RemoteSession
 from .ssh_session import SSHSession
+from .testpmd_shell import TestPmdShell
 
 
 def create_remote_session(
     node_config: NodeConfiguration, name: str, logger: DTSLOG
 ) -> RemoteSession:
     return SSHSession(node_config, name, logger)
+
+
+def create_interactive_session(
+    node_config: NodeConfiguration, name: str, logger: DTSLOG
+) -> InteractiveRemoteSession:
+    return InteractiveRemoteSession(node_config, logger)
diff --git a/dts/framework/remote_session/remote/interactive_remote_session.py 
b/dts/framework/remote_session/remote/interactive_remote_session.py
new file mode 100644
index 00000000..e145d35d
--- /dev/null
+++ b/dts/framework/remote_session/remote/interactive_remote_session.py
@@ -0,0 +1,118 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2023 University of New Hampshire
+
+import socket
+import traceback
+from pathlib import PurePath
+from typing import Union
+
+from paramiko import AutoAddPolicy, SSHClient, Transport  # type: ignore
+from paramiko.ssh_exception import (  # type: ignore
+    AuthenticationException,
+    BadHostKeyException,
+    NoValidConnectionsError,
+    SSHException,
+)
+
+from framework.config import InteractiveApp, NodeConfiguration
+from framework.exception import SSHConnectionError
+from framework.logger import DTSLOG
+
+from .interactive_shell import InteractiveShell
+from .testpmd_shell import TestPmdShell
+
+
+class InteractiveRemoteSession:
+    hostname: str
+    ip: str
+    port: int
+    username: str
+    password: str
+    _logger: DTSLOG
+    _node_config: NodeConfiguration
+    session: SSHClient
+    _transport: Transport | None
+
+    def __init__(self, node_config: NodeConfiguration, _logger: DTSLOG) -> 
None:
+        self._node_config = node_config
+        self._logger = _logger
+        self.hostname = node_config.hostname
+        self.username = node_config.user
+        self.password = node_config.password if node_config.password else ""
+        port = "22"
+        self.ip = node_config.hostname
+        if ":" in node_config.hostname:
+            self.ip, port = node_config.hostname.split(":")
+        self.port = int(port)
+        self._logger.info(
+            f"Initializing interactive connection for 
{self.username}@{self.hostname}"
+        )
+        self._connect()
+        self._logger.info(
+            f"Interactive connection successful for 
{self.username}@{self.hostname}"
+        )
+
+    def _connect(self) -> None:
+        client = SSHClient()
+        client.set_missing_host_key_policy(AutoAddPolicy)
+        self.session = client
+        retry_attempts = 10
+        for retry_attempt in range(retry_attempts):
+            try:
+                client.connect(
+                    self.ip,
+                    username=self.username,
+                    port=self.port,
+                    password=self.password,
+                    timeout=20 if self.port else 10,
+                )
+            except (TypeError, BadHostKeyException, AuthenticationException) 
as e:
+                self._logger.exception(e)
+                raise SSHConnectionError(self.hostname) from e
+            except (NoValidConnectionsError, socket.error, SSHException) as e:
+                self._logger.debug(traceback.format_exc())
+                self._logger.warning(e)
+                self._logger.info(
+                    "Retrying interactive session connection: "
+                    f"retry number {retry_attempt +1}"
+                )
+            else:
+                break
+        else:
+            raise SSHConnectionError(self.hostname)
+        # Interactive sessions are used on an "as needed" basis so we have
+        # to set a keepalive
+        self._transport = self.session.get_transport()
+        if self._transport is not None:
+            self._transport.set_keepalive(30)
+
+    def create_interactive_shell(
+        self,
+        shell_type: InteractiveApp,
+        path_to_app: PurePath,
+        eal_parameters: str,
+        timeout: float,
+    ) -> Union[InteractiveShell, TestPmdShell]:
+        """
+        See "create_interactive_shell" in SutNode
+        """
+        match (shell_type):
+            case InteractiveApp.shell:
+                return InteractiveShell(
+                    self.session, self._logger, path_to_app, timeout
+                )
+            case InteractiveApp.testpmd:
+                return TestPmdShell(
+                    self.session,
+                    self._logger,
+                    path_to_app,
+                    timeout=timeout,
+                    eal_flags=eal_parameters,
+                )
+            case _:
+                self._logger.info(
+                    f"Unhandled app type {shell_type.name}, defaulting to 
shell."
+                )
+                return InteractiveShell(
+                    self.session, self._logger, path_to_app, timeout
+                )
diff --git a/dts/framework/remote_session/remote/interactive_shell.py 
b/dts/framework/remote_session/remote/interactive_shell.py
new file mode 100644
index 00000000..4b0735c8
--- /dev/null
+++ b/dts/framework/remote_session/remote/interactive_shell.py
@@ -0,0 +1,99 @@
+from pathlib import PurePath
+
+from paramiko import Channel, SSHClient, channel  # type: ignore
+
+from framework.logger import DTSLOG
+from framework.settings import SETTINGS
+
+
+class InteractiveShell:
+
+    _interactive_session: SSHClient
+    _stdin: channel.ChannelStdinFile
+    _stdout: channel.ChannelFile
+    _ssh_channel: Channel
+    _logger: DTSLOG
+    _timeout: float
+    _path_to_app: PurePath
+
+    def __init__(
+        self,
+        interactive_session: SSHClient,
+        logger: DTSLOG,
+        path_to_app: PurePath,
+        timeout: float = SETTINGS.timeout,
+    ) -> None:
+        self._interactive_session = interactive_session
+        self._ssh_channel = self._interactive_session.invoke_shell()
+        self._stdin = self._ssh_channel.makefile_stdin("w")
+        self._stdout = self._ssh_channel.makefile("r")
+        self._ssh_channel.settimeout(timeout)
+        self._ssh_channel.set_combine_stderr(True)  # combines stdout and 
stderr streams
+        self._logger = logger
+        self._timeout = timeout
+        self._path_to_app = path_to_app
+
+
+    def send_command_no_output(self, command: str) -> None:
+        """Send command to channel without recording output.
+
+        This method will not verify any input or output, it will simply assume 
the
+        command succeeded. This method will also consume all output in the 
buffer
+        after executing the command.
+        """
+        self._logger.info(
+            f"Sending command {command.strip()} and not collecting output"
+        )
+        self._stdin.write(f"{command}\n")
+        self._stdin.flush()
+        self.empty_stdout_buffer()
+
+    def empty_stdout_buffer(self) -> None:
+        """Removes all data from the stdout buffer.
+
+        Because of the way paramiko handles read buffers, there is no way to 
effectively
+        remove data from, or "flush", read buffers. This method essentially 
moves our
+        offset on the buffer to the end and thus "removes" the data from the 
buffer.
+        Timeouts are thrown on read operations of paramiko pipes based on 
whether data
+        had been received before timeout so we assume that if we reach the 
timeout then
+        we are at the end of the buffer.
+        """
+        self._ssh_channel.settimeout(0.5)
+        try:
+            for line in self._stdout:
+                pass
+        except TimeoutError:
+            pass
+        self._ssh_channel.settimeout(self._timeout)  # reset timeout
+
+    def send_command_get_output(self, command: str, prompt: str) -> str:
+        """Send a command and get all output before the expected ending string.
+
+        Lines that expect input are not included in the stdout buffer so they 
cannot be
+        used for expect. For example, if you were prompted to log into 
something
+        with a username and password, you cannot expect "username:" because it 
won't
+        yet be in the stdout buffer. A work around for this could be consuming 
an
+        extra newline character to force the current prompt into the stdout 
buffer.
+
+        Returns:
+            All output in the buffer before expected string
+        """
+        self._logger.info(f"Sending command {command.strip()}...")
+        self._stdin.write(f"{command}\n")
+        self._stdin.flush()
+        out: str = ""
+        for line in self._stdout:
+            out += line
+            if prompt in line and not line.rstrip().endswith(
+                command.rstrip()
+            ):  # ignore line that sent command
+                break
+        self._logger.debug(f"Got output: {out}")
+        return out
+
+    def close(self) -> None:
+        self._stdin.close()
+        self._ssh_channel.close()
+
+    def __del__(self) -> None:
+        self.close()
diff --git a/dts/framework/remote_session/remote/testpmd_shell.py 
b/dts/framework/remote_session/remote/testpmd_shell.py
new file mode 100644
index 00000000..bde3468b
--- /dev/null
+++ b/dts/framework/remote_session/remote/testpmd_shell.py
@@ -0,0 +1,67 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2023 University of New Hampshire
+
+
+from pathlib import PurePath
+
+from paramiko import SSHClient  # type: ignore
+
+from framework.logger import DTSLOG
+from framework.settings import SETTINGS
+
+from .interactive_shell import InteractiveShell
+
+
+class TestPmdShell(InteractiveShell):
+    expected_prompt: str = "testpmd>"
+    _eal_flags: str
+
+    def __init__(
+        self,
+        interactive_session: SSHClient,
+        logger: DTSLOG,
+        path_to_testpmd: PurePath,
+        eal_flags: str,
+        timeout: float = SETTINGS.timeout,
+    ) -> None:
+        """Initializes an interactive testpmd session using specified 
parameters."""
+        self._eal_flags = eal_flags
+
+        super(TestPmdShell, self).__init__(
+            interactive_session,
+            logger=logger,
+            path_to_app=path_to_testpmd,
+            timeout=timeout,
+        )
+        self._start_application()
+
+    def _start_application(self) -> None:
+        """Starts a new interactive testpmd shell using _path_to_app.
+        """
+        self.send_command(
+            f"{self._path_to_app} {self._eal_flags} -- -i",
+        )
+
+    def send_command(self, command: str, prompt: str = expected_prompt) -> str:
+        """Specific way of handling the command for testpmd
+
+        An extra newline character is consumed in order to force the current 
line into
+        the stdout buffer.
+        """
+        return self.send_command_get_output(f"{command}\n", prompt)
+
+    def get_devices(self) -> list[str]:
+        """Get a list of device names that are known to testpmd
+
+        Uses the device info listed in testpmd and then parses the output to
+        return only the names of the devices.
+
+        Returns:
+            A list of strings representing device names (e.g. 0000:14:00.1)
+        """
+        dev_info: str = self.send_command("show device info all")
+        dev_list: list[str] = []
+        for line in dev_info.split("\n"):
+            if "device name:" in line.lower():
+                dev_list.append(line.strip().split(": ")[1].strip())
+        return dev_list
diff --git a/dts/framework/test_result.py b/dts/framework/test_result.py
index 74391982..2436eb7f 100644
--- a/dts/framework/test_result.py
+++ b/dts/framework/test_result.py
@@ -1,5 +1,6 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2023 PANTHEON.tech s.r.o.
+# Copyright(c) 2022-2023 University of New Hampshire
 
 """
 Generic result container and reporters
@@ -13,9 +14,11 @@
     OS,
     Architecture,
     BuildTargetConfiguration,
+    BuildTargetInfo,
     Compiler,
     CPUType,
     NodeConfiguration,
+    NodeInfo,
 )
 from .exception import DTSError, ErrorSeverity
 from .logger import DTSLOG
@@ -67,12 +70,14 @@ class Statistics(dict):
     Using a dict provides a convenient way to format the data.
     """
 
-    def __init__(self, dpdk_version):
+    def __init__(self, output_info: dict[str, str] | None):
         super(Statistics, self).__init__()
         for result in Result:
             self[result.name] = 0
         self["PASS RATE"] = 0.0
-        self["DPDK VERSION"] = dpdk_version
+        if output_info:
+            for info_key, info_val in output_info.items():
+                self[info_key] = info_val
 
     def __iadd__(self, other: Result) -> "Statistics":
         """
@@ -206,6 +211,8 @@ class BuildTargetResult(BaseResult):
     os: OS
     cpu: CPUType
     compiler: Compiler
+    compiler_version: str | None
+    dpdk_version: str | None
 
     def __init__(self, build_target: BuildTargetConfiguration):
         super(BuildTargetResult, self).__init__()
@@ -213,6 +220,12 @@ def __init__(self, build_target: BuildTargetConfiguration):
         self.os = build_target.os
         self.cpu = build_target.cpu
         self.compiler = build_target.compiler
+        self.compiler_version = None
+        self.dpdk_version = None
+
+    def add_build_target_versions(self, versions: BuildTargetInfo) -> None:
+        self.compiler_version = versions.compiler_version
+        self.dpdk_version = versions.dpdk_version
 
     def add_test_suite(self, test_suite_name: str) -> TestSuiteResult:
         test_suite_result = TestSuiteResult(test_suite_name)
@@ -228,10 +241,17 @@ class ExecutionResult(BaseResult):
     """
 
     sut_node: NodeConfiguration
+    sut_os_name: str
+    sut_os_version: str
+    sut_kernel_version: str
 
-    def __init__(self, sut_node: NodeConfiguration):
+    def __init__(self, sut_node: NodeConfiguration, sut_version_info: 
NodeInfo):
         super(ExecutionResult, self).__init__()
         self.sut_node = sut_node
+        self.sut_version_info = sut_version_info
+        self.sut_os_name = sut_version_info.os_name
+        self.sut_os_version = sut_version_info.os_version
+        self.sut_kernel_version = sut_version_info.kernel_version
 
     def add_build_target(
         self, build_target: BuildTargetConfiguration
@@ -258,6 +278,7 @@ class DTSResult(BaseResult):
     """
 
     dpdk_version: str | None
+    output: dict | None
     _logger: DTSLOG
     _errors: list[Exception]
     _return_code: ErrorSeverity
@@ -267,14 +288,17 @@ class DTSResult(BaseResult):
     def __init__(self, logger: DTSLOG):
         super(DTSResult, self).__init__()
         self.dpdk_version = None
+        self.output = None
         self._logger = logger
         self._errors = []
         self._return_code = ErrorSeverity.NO_ERR
         self._stats_result = None
         self._stats_filename = os.path.join(SETTINGS.output_dir, 
"statistics.txt")
 
-    def add_execution(self, sut_node: NodeConfiguration) -> ExecutionResult:
-        execution_result = ExecutionResult(sut_node)
+    def add_execution(
+        self, sut_node: NodeConfiguration, sut_version_info: NodeInfo
+    ) -> ExecutionResult:
+        execution_result = ExecutionResult(sut_node, sut_version_info)
         self._inner_results.append(execution_result)
         return execution_result
 
@@ -296,7 +320,8 @@ def process(self) -> None:
             for error in self._errors:
                 self._logger.debug(repr(error))
 
-        self._stats_result = Statistics(self.dpdk_version)
+        self._stats_result = Statistics(self.output)
+        # add information gathered from the smoke tests to the statistics
         self.add_stats(self._stats_result)
         with open(self._stats_filename, "w+") as stats_file:
             stats_file.write(str(self._stats_result))
diff --git a/dts/framework/test_suite.py b/dts/framework/test_suite.py
index 0705f38f..5df5d2a6 100644
--- a/dts/framework/test_suite.py
+++ b/dts/framework/test_suite.py
@@ -11,10 +11,21 @@
 import re
 from types import MethodType
 
-from .exception import ConfigurationError, SSHTimeoutError, TestCaseVerifyError
+from .exception import (
+    BlockingTestSuiteError,
+    ConfigurationError,
+    SSHTimeoutError,
+    TestCaseVerifyError,
+)
 from .logger import DTSLOG, getLogger
 from .settings import SETTINGS
-from .test_result import BuildTargetResult, Result, TestCaseResult, 
TestSuiteResult
+from .test_result import (
+    BuildTargetResult,
+    DTSResult,
+    Result,
+    TestCaseResult,
+    TestSuiteResult,
+)
 from .testbed_model import SutNode
 
 
@@ -37,10 +48,12 @@ class TestSuite(object):
     """
 
     sut_node: SutNode
+    is_blocking = False
     _logger: DTSLOG
     _test_cases_to_run: list[str]
     _func: bool
     _result: TestSuiteResult
+    _dts_result: DTSResult
 
     def __init__(
         self,
@@ -48,6 +61,7 @@ def __init__(
         test_cases: list[str],
         func: bool,
         build_target_result: BuildTargetResult,
+        dts_result: DTSResult,
     ):
         self.sut_node = sut_node
         self._logger = getLogger(self.__class__.__name__)
@@ -55,6 +69,7 @@ def __init__(
         self._test_cases_to_run.extend(SETTINGS.test_cases)
         self._func = func
         self._result = 
build_target_result.add_test_suite(self.__class__.__name__)
+        self._dts_result = dts_result
 
     def set_up_suite(self) -> None:
         """
@@ -118,6 +133,8 @@ def run(self) -> None:
                     f"the next test suite may be affected."
                 )
                 self._result.update_setup(Result.ERROR, e)
+            if len(self._result.get_errors()) > 0 and self.is_blocking:
+                raise BlockingTestSuiteError(test_suite_name)
 
     def _execute_test_suite(self) -> None:
         """
diff --git a/dts/framework/testbed_model/node.py 
b/dts/framework/testbed_model/node.py
index d48fafe6..c5147e0e 100644
--- a/dts/framework/testbed_model/node.py
+++ b/dts/framework/testbed_model/node.py
@@ -40,6 +40,7 @@ class Node(object):
     lcores: list[LogicalCore]
     _logger: DTSLOG
     _other_sessions: list[OSSession]
+    _execution_config: ExecutionConfiguration
 
     def __init__(self, node_config: NodeConfiguration):
         self.config = node_config
@@ -64,6 +65,7 @@ def set_up_execution(self, execution_config: 
ExecutionConfiguration) -> None:
         """
         self._setup_hugepages()
         self._set_up_execution(execution_config)
+        self._execution_config = execution_config
 
     def _set_up_execution(self, execution_config: ExecutionConfiguration) -> 
None:
         """
diff --git a/dts/framework/testbed_model/sut_node.py 
b/dts/framework/testbed_model/sut_node.py
index 2b2b50d9..9b17ac3d 100644
--- a/dts/framework/testbed_model/sut_node.py
+++ b/dts/framework/testbed_model/sut_node.py
@@ -1,14 +1,27 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2010-2014 Intel Corporation
 # Copyright(c) 2023 PANTHEON.tech s.r.o.
+# Copyright(c) 2022-2023 University of New Hampshire
 
 import os
 import tarfile
 import time
 from pathlib import PurePath
-
-from framework.config import BuildTargetConfiguration, NodeConfiguration
-from framework.remote_session import CommandResult, OSSession
+from typing import Union
+
+from framework.config import (
+    BuildTargetConfiguration,
+    BuildTargetInfo,
+    InteractiveApp,
+    NodeConfiguration,
+    NodeInfo,
+)
+from framework.remote_session import (
+    CommandResult,
+    InteractiveShell,
+    OSSession,
+    TestPmdShell,
+)
 from framework.settings import SETTINGS
 from framework.utils import EnvVarsDict, MesonArgs
 
@@ -16,6 +29,52 @@
 from .node import Node
 
 
+class EalParameters(object):
+    def __init__(
+        self,
+        lcore_list: LogicalCoreList,
+        memory_channels: int,
+        prefix: str,
+        no_pci: bool,
+        vdevs: list[VirtualDevice],
+        other_eal_param: str,
+    ):
+        """
+        Generate eal parameters character string;
+        :param lcore_list: the list of logical cores to use.
+        :param memory_channels: the number of memory channels to use.
+        :param prefix: set file prefix string, eg:
+                        prefix='vf'
+        :param no_pci: switch of disable PCI bus eg:
+                        no_pci=True
+        :param vdevs: virtual device list, eg:
+                        vdevs=[
+                            VirtualDevice('net_ring0'),
+                            VirtualDevice('net_ring1')
+                        ]
+        :param other_eal_param: user defined DPDK eal parameters, eg:
+                        other_eal_param='--single-file-segments'
+        """
+        self._lcore_list = f"-l {lcore_list}"
+        self._memory_channels = f"-n {memory_channels}"
+        self._prefix = prefix
+        if prefix:
+            self._prefix = f"--file-prefix={prefix}"
+        self._no_pci = "--no-pci" if no_pci else ""
+        self._vdevs = " ".join(f"--vdev {vdev}" for vdev in vdevs)
+        self._other_eal_param = other_eal_param
+
+    def __str__(self) -> str:
+        return (
+            f"{self._lcore_list} "
+            f"{self._memory_channels} "
+            f"{self._prefix} "
+            f"{self._no_pci} "
+            f"{self._vdevs} "
+            f"{self._other_eal_param}"
+        )
+
+
 class SutNode(Node):
     """
     A class for managing connections to the System under Test, providing
@@ -30,9 +89,11 @@ class SutNode(Node):
     _env_vars: EnvVarsDict
     _remote_tmp_dir: PurePath
     __remote_dpdk_dir: PurePath | None
-    _dpdk_version: str | None
     _app_compile_timeout: float
     _dpdk_kill_session: OSSession | None
+    _dpdk_version: str | None
+    _node_info: NodeInfo | None
+    _compiler_version: str | None
 
     def __init__(self, node_config: NodeConfiguration):
         super(SutNode, self).__init__(node_config)
@@ -41,12 +102,14 @@ def __init__(self, node_config: NodeConfiguration):
         self._env_vars = EnvVarsDict()
         self._remote_tmp_dir = self.main_session.get_remote_tmp_dir()
         self.__remote_dpdk_dir = None
-        self._dpdk_version = None
         self._app_compile_timeout = 90
         self._dpdk_kill_session = None
         self._dpdk_timestamp = (
             f"{str(os.getpid())}_{time.strftime('%Y%m%d%H%M%S', 
time.localtime())}"
         )
+        self._dpdk_version = None
+        self._node_info = None
+        self._compiler_version = None
 
     @property
     def _remote_dpdk_dir(self) -> PurePath:
@@ -75,6 +138,32 @@ def dpdk_version(self) -> str:
             )
         return self._dpdk_version
 
+    @property
+    def node_info(self) -> NodeInfo:
+        if self._node_info is None:
+            self._node_info = self.main_session.get_node_info()
+        return self._node_info
+
+    @property
+    def compiler_version(self) -> str:
+        if self._compiler_version is None:
+            if self._build_target_config is not None:
+                self._compiler_version = 
self.main_session.get_compiler_version(
+                    self._build_target_config.compiler.name
+                )
+            else:
+                self._logger.warning(
+                    "Failed to get compiler version because"
+                    "_build_target_config is None."
+                )
+                return ""
+        return self._compiler_version
+
+    def get_build_target_info(self) -> BuildTargetInfo:
+        return BuildTargetInfo(
+            dpdk_version=self.dpdk_version, 
compiler_version=self.compiler_version
+        )
+
     def _guess_dpdk_remote_dir(self) -> PurePath:
         return self.main_session.guess_dpdk_remote_dir(self._remote_tmp_dir)
 
@@ -84,6 +173,10 @@ def _set_up_build_target(
         """
         Setup DPDK on the SUT node.
         """
+        # we want to ensure that dpdk_version and compiler_version is reset 
for new
+        # build targets
+        self._dpdk_version = None
+        self._compiler_version = None
         self._configure_build_target(build_target_config)
         self._copy_dpdk_tarball()
         self._build_dpdk()
@@ -262,48 +355,43 @@ def run_dpdk_app(
             f"{app_path} {eal_args}", timeout, verify=True
         )
 
-
-class EalParameters(object):
-    def __init__(
+    def create_interactive_shell(
         self,
-        lcore_list: LogicalCoreList,
-        memory_channels: int,
-        prefix: str,
-        no_pci: bool,
-        vdevs: list[VirtualDevice],
-        other_eal_param: str,
-    ):
+        shell_type: InteractiveApp,
+        path_to_app: PurePath | None = None,
+        timeout: float = SETTINGS.timeout,
+        eal_parameters: EalParameters | None = None,
+    ) -> Union[InteractiveShell, TestPmdShell]:
+        """Create a handler for an interactive session.
+
+        This method is a factory that calls a method in OSSession to create 
shells for
+        different DPDK applications.
+
+        Args:
+            shell_type: Enum value representing the desired application.
+            path_to_app: Represents a path to the application you are 
attempting to
+                launch. This path will be executed at the start of the app
+                initialization. If one isn't provided, the default specified 
in the
+                enumeration will be used.
+            timeout: Timeout for reading output from the SSH channel. If you 
are
+                reading from the buffer and don't receive any data within the 
timeout
+                it will throw an error.
+            eal_parameters: List of EAL parameters to use to launch the app. 
This is
+                ignored for base "shell" types.
+        Returns:
+            Instance of the desired interactive application.
         """
-        Generate eal parameters character string;
-        :param lcore_list: the list of logical cores to use.
-        :param memory_channels: the number of memory channels to use.
-        :param prefix: set file prefix string, eg:
-                        prefix='vf'
-        :param no_pci: switch of disable PCI bus eg:
-                        no_pci=True
-        :param vdevs: virtual device list, eg:
-                        vdevs=[
-                            VirtualDevice('net_ring0'),
-                            VirtualDevice('net_ring1')
-                        ]
-        :param other_eal_param: user defined DPDK eal parameters, eg:
-                        other_eal_param='--single-file-segments'
-        """
-        self._lcore_list = f"-l {lcore_list}"
-        self._memory_channels = f"-n {memory_channels}"
-        self._prefix = prefix
-        if prefix:
-            self._prefix = f"--file-prefix={prefix}"
-        self._no_pci = "--no-pci" if no_pci else ""
-        self._vdevs = " ".join(f"--vdev {vdev}" for vdev in vdevs)
-        self._other_eal_param = other_eal_param
+        # if we just want a default shell, there is no need to append the DPDK 
build
+        # directory to the path
+        default_path = shell_type.get_path()
 
-    def __str__(self) -> str:
-        return (
-            f"{self._lcore_list} "
-            f"{self._memory_channels} "
-            f"{self._prefix} "
-            f"{self._no_pci} "
-            f"{self._vdevs} "
-            f"{self._other_eal_param}"
+        if shell_type != InteractiveApp.shell:
+            default_path = self.main_session.join_remote_path(
+                self.remote_dpdk_build_dir, shell_type.get_path()
+            )
+        return self.main_session.interactive_session.create_interactive_shell(
+            shell_type,
+            path_to_app if path_to_app else default_path,
+            str(eal_parameters) if eal_parameters else "",
+            timeout,
         )
diff --git a/dts/tests/TestSuite_smoke_tests.py 
b/dts/tests/TestSuite_smoke_tests.py
new file mode 100644
index 00000000..b7e70ee1
--- /dev/null
+++ b/dts/tests/TestSuite_smoke_tests.py
@@ -0,0 +1,118 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2023 University of New Hampshire
+
+import re
+
+from framework.config import InteractiveApp
+from framework.remote_session import TestPmdShell
+from framework.settings import SETTINGS
+from framework.test_suite import TestSuite
+
+
+class SmokeTests(TestSuite):
+    is_blocking = True
+    # dicts in this list are expected to have two keys:
+    # "pci_address" and "current_driver"
+    nics_in_node: list[dict[str, str]] = []
+
+    def set_up_suite(self) -> None:
+        """
+        Setup:
+            Set the build directory path and generate a list of NICs in the 
SUT node.
+        """
+        self.dpdk_build_dir_path = self.sut_node.remote_dpdk_build_dir
+        for nic in self.sut_node.config.ports:
+            new_dict = {
+                "pci_address": nic.pci,
+                "current_driver": nic.os_driver.strip(),
+            }
+            self.nics_in_node.append(new_dict)
+
+    def test_unit_tests(self) -> None:
+        """
+        Test:
+            Run the fast-test unit-test suite through meson.
+        """
+        self.sut_node.main_session.send_command(
+            f"meson test -C {self.dpdk_build_dir_path} --suite fast-tests",
+            300,
+            verify=True,
+        )
+
+    def test_driver_tests(self) -> None:
+        """
+        Test:
+            Run the driver-test unit-test suite through meson.
+        """
+        list_of_vdevs = ""
+        for dev in self.sut_node._execution_config.vdevs:
+            list_of_vdevs += f"--vdev {dev} "
+        list_of_vdevs = list_of_vdevs[:-1]
+        if list_of_vdevs:
+            self._logger.info(
+                "Running driver tests with the following virtual "
+                f"devices: {list_of_vdevs}"
+            )
+            self.sut_node.main_session.send_command(
+                f"meson test -C {self.dpdk_build_dir_path} --suite 
driver-tests "
+                f'--test-args "{list_of_vdevs}"',
+                300,
+                verify=True,
+            )
+        else:
+            self.sut_node.main_session.send_command(
+                f"meson test -C {self.dpdk_build_dir_path} --suite 
driver-tests",
+                300,
+                verify=True,
+            )
+
+    def test_devices_listed_in_testpmd(self) -> None:
+        """
+        Test:
+            Uses testpmd driver to verify that devices have been found by 
testpmd.
+        """
+        testpmd_driver = 
self.sut_node.create_interactive_shell(InteractiveApp.testpmd)
+        # We know it should always be a TestPmdShell but mypy doesn't
+        assert isinstance(testpmd_driver, TestPmdShell)
+        dev_list: list[str] = testpmd_driver.get_devices()
+        for nic in self.nics_in_node:
+            self.verify(
+                nic["pci_address"] in dev_list,
+                f"Device {nic['pci_address']} was not listed in testpmd's 
available devices, "
+                "please check your configuration",
+            )
+
+    def test_device_bound_to_driver(self) -> None:
+        """
+        Test:
+            Ensure that all drivers listed in the config are bound to the 
correct driver.
+        """
+        path_to_devbind = self.sut_node.main_session.join_remote_path(
+            self.sut_node._remote_dpdk_dir, "usertools", "dpdk-devbind.py"
+        )
+
+        regex_for_pci_address = "/[0-9]{4}:[0-9]{2}:[0-9]{2}.[0-9]{1}/"
+        all_nics_in_dpdk_devbind = self.sut_node.main_session.send_command(
+            f"{path_to_devbind} --status | awk '{regex_for_pci_address}'",
+            SETTINGS.timeout,
+        ).stdout
+
+        for nic in self.nics_in_node:
+            # This regular expression finds the line in the above string that 
starts
+            # with the address for the nic we are on in the loop and then 
captures the
+            # name of the driver in a group
+            devbind_info_for_nic = re.search(
+                f"{nic['pci_address']}[^\\n]*drv=([\\d\\w]*) [^\\n]*",
+                all_nics_in_dpdk_devbind,
+            )
+            self.verify(
+                devbind_info_for_nic is not None,
+                f"Failed to find configured device ({nic['pci_address']}) 
using dpdk-devbind.py",
+            )
+            # We know this isn't None, but mypy doesn't
+            assert devbind_info_for_nic is not None
+            self.verify(
+                devbind_info_for_nic.group(1) == nic["current_driver"],
+                f"Driver for device {nic['pci_address']} does not match driver 
listed in "
+                f"configuration (bound to {devbind_info_for_nic.group(1)})",
+            )
-- 
2.41.0

Reply via email to