-
Notifications
You must be signed in to change notification settings - Fork 14
Results and analysis
The code can be found here.
The publications which describe the results of OSPBench have been listed in the page on References, Publications and Talks.
The charts that were used here can all be regenerated by checking out the results-analysis
folder.
The results data has been added there as well.
This article will lead you through the notebooks that are provided to generate visualizations and do further analysis.
This folder contains Jupyter notebooks with PySpark to do the analysis.
You need Jupyter and Spark installed. Follow one of the many installation guides online.
Our packages have gotten slightly old. We used:
- Jupyter 4.3.0
- Python 3.6.6
- Spark 2.3.0
It is possible that some syntax changes are required for the latest versions of these tools.
Run Jupyter on the root of the result-analysis folder to have the paths correct in the notebooks. For us this meant:
cd result-analysis
pyspark
Then open the notebooks at localhost:8888
.
Initially, we used the structure for the input data which is used in the data
folder. The latency, periodic burst and single burst notebooks make use of this.
Later on, we switched to a different structure: the one used in the folder scalability-data
. The scalability and failure notebooks make use of this.
There is one notebook per workload. The names of the notebook are similar to the names of the corresponding workloads.
The data which we used in our papers on OSPBench and the notebooks used to visualize them are both available here.
This work has been made possible by Klarrio