Navigation auf uzh.ch
This is the companion web site for preprint titled “Patterns in the Chaos – a Structured Large-Scale Analysis of Variations in the Performance of Public IaaS Providers”. Please see the e-print of the submission on arXiv. Here, you will find additional information, the raw data we collected, the benchmarks we used to collect this data, and further analysis that we did not find space for in the publication. Should we later on find errors in our paper, we will also collect errata here.
Benchmarking the performance of public cloud providers is a common research topic. Previous research has already extensively evaluated the performance of different cloud platforms for different use cases, and under different constraints and experiment setups. In this paper, we present a principled, large-scale literature review to collect and codify existing research regarding the predictability of performance in public Infrastructure-as-a-Service (IaaS) clouds. We formulate 15 hypotheses relating to the nature of performance variations in IaaS systems, to the factors of influence of performance variations, and how to compare different instance types. In a second step, we conduct extensive real-life experimentation on Amazon EC2 and Google Compute Engine to empirically validate those hypotheses. At the time of our research, performance in EC2 was substantially less predictable than in GCE. Further, we show that hardware heterogeneity is in practice less prevalent than anticipated by earlier research,
while multi-tenancy has a dramatic impact on performance and predictability.
For the paper, we looked at performance predictability in Google Compute Engine, AWS Elastic Compute Cloud, Microsoft Azure, and IBM Softlayer.
Full Data Set (ZIP (ZIP, 640 KB))
R Script used for Data Cleaning (R code (R, 4 KB))
To generate the data sets, we used our tool Cloud Workbench (CWB) to install and run the benchmarks in the cloud instances we wanted to evaluate. CWB is freely available as open source software (however the installation procedure still needs work – send us a mail if you have troubles). CWB uses Chef to actually provision the benchmark code. Below, we provide as ZIP archives of Chef recipes all benchmarks we used for this paper.
Benchmarks (ZIP (ZIP, 68 KB))
The recipes depend on a special Chef cookbook, which installs the CWB client library, so maybe you will need to adapt the recipes to install them successfully outside of CWB. If you are only interested in how we actually execute the benchmarks, you may investigate the files in files/default in each ZIP archive. The configurations in these files are inherited from attributes/default.rb.
The hypotheses formulated in the paper are based on a principled study of existing research in cloud benchmarking. Below, we give full references for all 56 studies that we considered in our work, in alphabetical order of the last name of the first author.