Towards Defect-based Testing for Safety-critical ML Components

Authors

  • Carmen Carlan
  • Amit Sahu

Keywords:

Verification, Machine Learning, Defect-based Testing

Abstract

The input space of Machine Learning (ML) components used in safety-critical applications is complex.  Testing such components on exponentially large datasets that cover all potential real-world situations to meaningfully measure performance metrics, such as the false negative rate, is infeasible due to practical restrictions.  Consequently, we must limit the test data while adequately covering critical input data points.  Inspired by defect-based software testing, a method for specifying adequate test cases for software components, we propose a process for collecting adequate test data for ML components.  Concretely, we systematically employ different existing ML data quality metrics, and methods for enhancing the test data, to uncover critical scenarios where the ML component may be less performant.  We exemplify the usage of our process in two case studies, each involving an ML component implementing different functionalities, i.e. stop sign recognition, and railway track segmentation.

Argument Explaining the Collection of Test Data Using NNDK Metrics

Downloads

Published

2024-08-22