Benchmark frameworks and τbench

Stephen W. Thomas, Richard T. Snodgrass, Rui Zhang

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Software engineering frameworks tame the complexity of large collections of classes by identifying structural invariants, regularizing interfaces, and increasing sharing across the collection. We wish to appropriate these benefits for families of closely related benchmarks, say for evaluating query engine implementation strategies. We introduce the notion of a benchmark framework, an ecosystem of benchmarks that are related in semantically rich ways and enabled by organizing principles. A benchmark framework is realized by iteratively changing one individual benchmark into another, say by modifying the data format, adding schema constraints, or instantiating a different workload. Paramount to our notion of benchmark frameworks are the ease of describing the differences between individual benchmarks and the utility of methods to validate the correctness of each benchmark component by exploiting the overarching ecosystem. As a detailed case study, we introduce τBench, a benchmark framework consisting of ten individual benchmarks, spanning XML, XQuery, XML Schema, and PSM, along with temporal extensions to each. The second case study examines the Mining Unstructured Data benchmark framework, and the third examines the potential benefits of rendering the TPC family as a benchmark framework.

Original languageEnglish (US)
Pages (from-to)1047-1075
Number of pages29
JournalSoftware - Practice and Experience
Volume44
Issue number9
DOIs
StatePublished - Sep 2014
Externally publishedYes

Keywords

  • XML
  • benchmarks
  • temporal databases

ASJC Scopus subject areas

  • Software

Fingerprint

Dive into the research topics of 'Benchmark frameworks and τbench'. Together they form a unique fingerprint.

Cite this