Main Content

Assess Requirements-Based Testing Quality by Using the Model Testing Dashboard

You can assess the status of your model testing activities by using the metrics in the Model Testing Dashboard. When you test your models against requirements, you maintain traceability between the requirements, models, tests, and results. The dashboard helps you to track the status of these artifacts and the traceability relationships between them. Each metric in the dashboard measures a different aspect of the quality of the testing artifacts and reflects guidelines in industry-recognized software development standards, such as ISO 26262 and DO-178C. For more information, see Assess Requirements-Based Testing for ISO 26262.

From the dashboard, you can identify and fix testing issues. Update the dashboard metrics to track your progress toward testing compliance.

Open the Project and Model Testing Dashboard

The Model Testing Dashboard shows data on the traceability and testing status of each unit in your project. A unit is a functional entity in your software architecture that you can execute and test independently or as part of larger system tests. You can label models as units in the Model Testing Dashboard. If you do not specify the models that are considered units, then the dashboard considers a model to be a unit if it does not reference other models. The dashboard considers each model in your project to represent a unit because you use models to design and test the algorithms.

  1. Open the project that contains the models and testing artifacts. For this example, in the MATLAB® Command Window, enter:


  2. Open the Model Testing Dashboard by using one of these approaches:

    • On the Project tab, click Model Testing Dashboard.

    • At the MATLAB command line, enter modelTestingDashboard.

    The first time that you open the dashboard for the project, the dashboard must identify the artifacts in the project and collect traceability information.

    The Project panel shows the architecture of the software units and components in the current project. The Artifacts panel shows the artifacts that the dashboard traced to the current unit or component selected in the Project panel.

    The dashboard performs an initial traceability analysis and collects metric results for the metrics available in your installation. Collecting results for each of the MathWorks® metrics requires licenses for Simulink® Check™, Requirements Toolbox™, and Simulink Test™. If metric results have been collected, viewing the results requires only a Simulink Check license.

  3. In the Project panel, select the unit cc_DriverSwRequest.

    The dashboard analyzes the traceability links from the artifacts to the models in the project and populates the widgets with metric results for the unit that is selected in the Project panel.

Model Testing Dashboard showing results for unit cc_DriverSwRequest

Assess Traceability of Artifacts

When the dashboard collects and reports metric data, it scopes the results to the artifacts in one unit in the project. Use the Artifacts panel to view the artifacts that trace to the unit that you selected in the Project panel.

  1. In the Artifacts panel, expand the Functional Requirements section and then expand the Implemented and Upstream sections. This unit implements the requirements in the file cc_SoftwareReqs.slreqx and links to the upstream, system-level requirements in das_SystemReqs.slreqx.

  2. Click the arrow to the left of a file name to see the individual requirements that trace to the model.

  3. To see the artifact type and the path to the artifact, click the three dots next to the artifact name to open a tooltip.

You can explore the sections in the Artifacts panel to see which requirements, tests, and test results trace to each unit in the project. For more information on how the dashboard analyzes this traceability, see Trace Artifacts to Units and Components.

Explore Metric Results for a Unit

  1. In the Project panel, click the unit cc_DriverSwRequest to view the Model Testing results. The dashboard widgets populate with the metric results for the unit.

  2. In the Test Analysis section of the dashboard, locate the Tests with Requirements widget. To view tooltips with details about the results, point to the sections of the gauge or to the percentage result.

    Tests with Requirements widget with tooltip indicating the number of tests that have links to requirements

  3. To explore the metric data in more detail, click an individual metric widget. For example, click the Tests with Requirements widget to view the Metric Details section for the metric.

    Table in Metric Details section with tests and statuses of whether each test is linked to requirements

    The table shows each test for the unit, the test file containing each test, and whether the test is linked to requirements.

  4. The table shows that the test Detect long decrement is missing linked requirements. You can use the Model Testing Dashboard to open the test in the Test Manager. In the Artifact column, click Detect long decrement to open the test case in Test Manager.

  5. At the top of the Model Testing Dashboard, there is a breadcrumb trail from the Metric Details back to the Model Testing results.

    Breadcrumb trail buttons at the top of the dashboard tab

    Click the breadcrumb button for cc_DriverSwRequest to return to the Model Testing results for the unit.

You can click on any of the widgets in the dashboard to view the details of their metric results. Use the hyperlinks in the tables to open the artifacts and address testing gaps. For more information on using the data in the dashboard, see Explore Status and Quality of Testing Activities Using Model Testing Dashboard.

Track Testing Status of a Project Using the Model Testing Dashboard

To use the Model Testing Dashboard to track your testing activities, set up and maintain your project using the best practices described in Manage Project Artifacts for Analysis in Dashboard. As you develop and test your models, use the dashboard to identify testing gaps, fix the underlying artifacts, and track your progress towards model testing completion. For more information on finding and addressing gaps in your model testing, see Fix Requirements-Based Testing Issues.

See Also

Related Topics