Friday, April 15, 2016 - 02:50 pm
2A27 Swearingen Engineering Center
Gregory Gay, University of South Carolina
Abstract: There are two key artifacts necessary to test software, the test data - inputs given to the system under test (SUT) - and the oracle - which judges the correctness of the resulting execution. Substantial research efforts have been devoted towards the creation of effective test inputs, but relatively little attention has been paid to the creation of oracles. Specifying test oracles remains challenging for many domains, such as real-time embedded systems, where small changes in timing or sensory input may cause large behavioral differences. Models of such systems, often built for analysis and simulation before the development of the final system, are appealing for reuse as oracles. These models, however, typically represent an idealized system, abstracting away certain considerations such as non-deterministic timing behavior and sensor noise. Thus, even with the same test data, the model’s behavior may fail to match an acceptable behavior of the SUT, leading to many false positives reported by the oracle.
This talk will present an automated framework that can adjust, or steer, the behavior of the model to better match the behavior of the SUT in order to reduce the rate of false positives. This model steering is limited by a set of constraints (defining acceptable differences in behavior) and is based on a search process attempting to minimize a numeric dissimilarity metric. This framework allows non-deterministic, but bounded, behavior differences, while preventing future mismatches, by guiding the oracle—within limits—to match the execution of the SUT. Results show that steering significantly increases SUT-oracle conformance with minimal masking of real faults and, thus, has significant potential for reducing false positives and, consequently, development costs.
Bio: Gregory Gay is an Assistant Professor of Computer Science & Engineering at the University of South Carolina. His research interests include automated testing and analysis—with an emphasis on test oracle construction—and search-based software engineering. Greg received his Ph.D. from the University of Minnesota, working with the Critical Systems research group, and an M.S. from West Virginia University. He has previously worked with NASA Ames Research Center and the Chinese Academy of Sciences.
This seminar is open to anyone who is interested, not just students enrolled in the CSCE 791 class. Please consider attending.