Generated by GPT-5-mini| TPC-DS | |
|---|---|
| Name | TPC-DS |
| Developer | Transaction Processing Performance Council |
| Release | 2008 |
| Genre | Decision support benchmark |
| Website | Transaction Processing Performance Council |
TPC-DS
TPC-DS is a decision-support benchmark designed to evaluate the performance of complex data warehouse systems. It models a retail supply chain and provides a comprehensive set of schemas, data generation tools, query workloads, and performance metrics for comparing database engines and hardware platforms. The benchmark is maintained by the Transaction Processing Performance Council and has been widely used by vendors such as IBM, Oracle Corporation, Microsoft Corporation, Amazon Web Services, and Google to publish comparative performance results.
TPC-DS emulates analytic workloads common to large retailers and service providers, encompassing sales, returns, inventory, promotions, web sales, and catalog operations. It produces a multi-table schema with star and snowflake patterns and scales data from gigabytes to petabytes using a scale factor parameter. Major adopters include Hewlett-Packard Enterprise, Dell Technologies, SAP SE, Snowflake Inc., and Teradata, while research groups at Massachusetts Institute of Technology, Stanford University, and University of California, Berkeley have used it for academic evaluations.
The specification defines a logical schema with 24 tables including dimension and fact tables, a data generator called DSGen, and a query template set. Components specified in the standard include the data model, data generation scripts, query templates, and compliance rules. Governance is overseen by the Transaction Processing Performance Council, with formal documentation and audit procedures similar to other standards published by bodies like ISO and influenced by benchmarking traditions from TPC-C and TPC-H. Industry partners such as Intel Corporation and NVIDIA have used the specification to tune hardware and software stacks.
The workload comprises 99 query templates that cover reporting, ad-hoc, iterative OLAP, and extraction tasks; these templates expand into hundreds of executable SQL queries via parameterization and randomization. Queries exercise features such as complex joins, window functions, subqueries, analytic functions, and large aggregations across tables like STORE_SALES and CATALOG_SALES. Vendors often publish results demonstrating performance on platforms from Amazon Redshift and Google BigQuery to on-premises systems like IBM Db2 and Oracle Exadata. Research prototypes such as Apache Spark and Presto have also been evaluated using the query suite.
The methodology prescribes data loading, run rules, and reporting formats to ensure repeatability and comparability. Key metrics include the Composite Query-per-Hour (QphDS) metric and single-stream and multi-stream measurements, emphasizing throughput and price-performance measures. Results often report peak throughput, power consumption, and total cost of ownership using hardware vendors like Supermicro and Lenovo as examples. Independent auditing by accredited firms such as Deloitte and Ernst & Young is required for public submissions to the council, mirroring practices found in other standardized benchmarks used by SPEC and SAP SD Benchmark participants.
Implementations span commercial data warehouses, cloud analytics services, and open-source engines. Published result sets include entries from IBM Db2 Warehouse, Oracle Exadata Database Machine, Microsoft Azure Synapse Analytics, and Snowflake Data Cloud, with hardware configurations using processors from AMD or Intel Xeon families and storage subsystems from NetApp or Pure Storage. Academic and industry white papers from groups at Carnegie Mellon University and ETH Zurich have analyzed performance characteristics and scaling behavior. Vendor press releases often cite QphDS figures, while third-party benchmarking labs provide comparative analyses between platforms such as Cloudera and Hortonworks.
Critics argue that the benchmark’s retail-centric schema may not generalize to domains like finance or telecommunications, and that parameterized query templates can be tuned by vendors to exploit system-specific optimizations. Concerns mirror debates around other benchmarks such as TPC-C and SPEC CPU where workload representativeness and result transparency are questioned. Additional limitations include the complexity of audit requirements, the cost of producing certified results for smaller vendors, and potential mismatches with real-world mixed OLTP/OLAP environments exemplified by systems at Walmart or eBay.
Work on the benchmark began in the early 2000s within the Transaction Processing Performance Council as an evolution beyond earlier benchmarks like TPC-H. The specification was released in 2008 after collaboration among database vendors, hardware manufacturers, and academic contributors. Subsequent revisions and clarifications have been influenced by advances in columnar storage, massively parallel processing, and cloud-native architectures from companies such as Amazon Web Services and Google Cloud Platform. The benchmark continues to be maintained by the council with ongoing community engagement through vendor submissions and academic evaluations.
Category:Database benchmarks