2021
DOI: 10.1007/s10664-021-09982-4
|View full text |Cite|
|
Sign up to set email alerts
|

Can Offline Testing of Deep Neural Networks Replace Their Online Testing?

Abstract: We distinguish two general modes of testing for Deep Neural Networks (DNNs): Offline testing where DNNs are tested as individual units based on test datasets obtained without involving the DNNs under test, and online testing where DNNs are embedded into a specific application environment and tested in a closed-loop mode in interaction with the application environment. Typically, DNNs are subjected to both types of testing during their development life cycle where offline testing is applied immediately after DN… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
20
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 27 publications
(22 citation statements)
references
References 26 publications
2
20
0
Order By: Relevance
“…If any of the objects within the red bounding boxes were on a collision course with the ego car, commencing PAEB would indeed be the right action for SMIRK and thus not violate SYS-SAF-REQ1. This observation corroborates the position by (Haq et al, 2021), i.e., system level testing that goes beyond model testing on single frames is critically needed. All results from running ML model testing, i.e., ML Verification Results [Z], are documented in the Protocols folder.…”
Section: Model Testing [Aa]supporting
confidence: 86%
See 2 more Smart Citations
“…If any of the objects within the red bounding boxes were on a collision course with the ego car, commencing PAEB would indeed be the right action for SMIRK and thus not violate SYS-SAF-REQ1. This observation corroborates the position by (Haq et al, 2021), i.e., system level testing that goes beyond model testing on single frames is critically needed. All results from running ML model testing, i.e., ML Verification Results [Z], are documented in the Protocols folder.…”
Section: Model Testing [Aa]supporting
confidence: 86%
“…Second, SMIRK could be used as a realistic test benchmark for automotive ML testing. The testing community has largely worked on offline testing of single frames, but we know that this is insufficient (Haq et al, 2021). Third, we recommend the community to port SMIRK to other simulators beyond ESI Pro-SiVIC.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…If False, the Rule Engine performs a sanity check based on laws of physics. (9) If UM remains confident that collision with a pedestrian is imminent, the signal to perform PAEB propagates to ego car.…”
Section: Smirk Architecturementioning
confidence: 99%
“…SMIRK allows researchers to explore data testing, ML model testing, integration testing, and system testing since data sets, ML model architectures, and the source code are publicly available. For example, offline model testing can be compared to online system testing, as Haq et al recently proved important [9]. Concrete test techniques that could be evaluated using SMIRK include search-based software testing, metamorphic testing, fuzz testing, neural network test adequacy assessments, and testing for explainable AI.…”
Section: Impact Overviewmentioning
confidence: 99%