Towards Defect-based Testing for Safety-critical ML Components
Keywords:
Verification, Machine Learning, Defect-based TestingAbstract
The input space of Machine Learning (ML) components used in safety-critical applications is complex. Testing such components on exponentially large datasets that cover all potential real-world situations to meaningfully measure performance metrics, such as the false negative rate, is infeasible due to practical restrictions. Consequently, we must limit the test data while adequately covering critical input data points. Inspired by defect-based software testing, a method for specifying adequate test cases for software components, we propose a process for collecting adequate test data for ML components. Concretely, we systematically employ different existing ML data quality metrics, and methods for enhancing the test data, to uncover critical scenarios where the ML component may be less performant. We exemplify the usage of our process in two case studies, each involving an ML component implementing different functionalities, i.e. stop sign recognition, and railway track segmentation.
Downloads
Published
Issue
Section
License
Copyright (c) 2024 Edge Case Research Inc.You may use this work providing you clearly acknowledge the Author (or their copyright holder) and the Safety-Critical Systems eJournal.