Difference between revisions of "Cloud benchmarks"

From SimpleWiki
Jump to navigationJump to search
 
(2 intermediate revisions by the same user not shown)
Line 11: Line 11:
 
== Benchmarks Scripts ==
 
== Benchmarks Scripts ==
  
* [http://traces.simpleweb.org/cloud_benchmarks/capture.tar.gz Download the scripts for running the benchmarks]
+
* [http://traces.simpleweb.org/cloud_benchmarks/benchmark_tools.tar.gz Download the scripts for running the benchmarks and post-process the data]
  
The scripts are written in python. All scripts require: netifaces, pcapy
+
The scripts are written in python. See the README for instructions on how to execute the benchmarks
 
 
=== How to execute the benchmarks ===
 
 
 
  ./delta_encoding.py -i wlan0 --seed 123134 --bytes 10000 --test 3 -o /tmp/output/ --ftp 1.1.1.1 --port 2121 --user "user_name" --passwd "password" --folder="."
 
 
 
Important remarks:
 
 
 
1 - The folder <nowiki>ftp://user:pass@server/folder/</nowiki> '''must''' be in a synchronized folder of the storage tool.
 
 
 
2 - The file delta_encoding.py '''must not''' be in a synchronized folder, otherwise the .pyc files created at run-time will disturb the experiment.
 
 
 
3 - The folder /tmp/output/ '''must not''' be in a synchronized folder, for the same reasons as above.
 
 
 
4 - Disable as much processes as possible in the benchmarking machine. This will minimize external interference on the test.
 
 
 
5 - If the storage system is running on a virtual machine, make sure the host machine is powerful enough to support the load. Check also whether the virtual machine limit or shape the network traffic.
 
  
 
== Traces ==
 
== Traces ==
  
Some traffic traces that generated the results in the paper can be downloaded from these links:
+
Some traffic traces that generated the results in the IMC paper can be downloaded from these links:
  
 
{| class="wikitable" style="text-align: center; width: 400px; height: 100px;"
 
{| class="wikitable" style="text-align: center; width: 400px; height: 100px;"
Line 56: Line 40:
 
|}
 
|}
  
These traces, together with the scripts below, produce the results in Figure 7 of the paper. More details about this dataset can also be obtained in Chapter 5 of:
+
These traces produce the results in Figure 7 of the paper. More details about this dataset can also be obtained in Chapter 5 of:
  
 
* Drago, I. (2013) [http://eprints.eemcs.utwente.nl/24136/ "Understanding and Monitoring Cloud Services"]. PhD thesis, University of Twente. CTIT Ph.D. thesis Series No. 13-279. ISBN 978-90-365-3577-9.
 
* Drago, I. (2013) [http://eprints.eemcs.utwente.nl/24136/ "Understanding and Monitoring Cloud Services"]. PhD thesis, University of Twente. CTIT Ph.D. thesis Series No. 13-279. ISBN 978-90-365-3577-9.
  
=== Post-processing the data ===
+
== Traffic Identification Example ==
  
The benchmark scripts generate a pcap file per benchmark step. In order to produce the figures presented in the paper, the pcap files need to be post-processed. The scripts in the following are examples to generate Figure 7. The typical flows of each tool are isolated by means of lists of server IP addresses, and statistics are calculated according to heuristics to determine the start and the end of synchronization steps during the benchmarks.  
+
The identification of the cloud storage traffic in the testbed can be done using the FQDNs that cloud storage clients use to contact to servers. This can be achieved, for example, by means of the methodology implemented in Tstat DN-Hunter. [http://www.simpleweb.org/w/images/3/3c/Cloud_storage_fqdn.txt.zip Here we post a list of FQDNs used by three popular providers as an example].
 
 
* [http://traces.simpleweb.org/cloud_benchmarks/post-process.tar.gz Download the scripts for post-processing the pcap files]
 
  
 
== Acceptable Use Policy ==
 
== Acceptable Use Policy ==
Line 78: Line 60:
 
     year          = {2013}
 
     year          = {2013}
 
   }
 
   }
 
== Paper abstract ==
 
 
Personal cloud storage services are data-intensive applications already producing a significant share of Internet traffic. Several solutions offered by different companies attract more and more people. However, little is known about each service capabilities, architecture and - most of all - performance implications of design choices. This paper presents a methodology to study cloud storage services. We apply our methodology to compare 5 popular offers, revealing different system architectures and capabilities. The implications on performance of different designs are assessed executing a series of benchmarks. Our results show no clear winner, with all services suffering from some limitations or having potential for improvement. In some scenarios, the upload of the same file set can take seven times more, wasting twice as much capacity. Our methodology and results are useful thus as both benchmark and guideline for system design.
 
  
 
== Acknowledgments ==
 
== Acknowledgments ==
  
 
This work was partly funded by the Network of Excellence project Flamingo (ICT-318488) and the EU-IP project mPlane (n-318627). Both projects are supported by the European Commission under its Seventh Framework Programme.
 
This work was partly funded by the Network of Excellence project Flamingo (ICT-318488) and the EU-IP project mPlane (n-318627). Both projects are supported by the European Commission under its Seventh Framework Programme.

Latest revision as of 08:30, 21 May 2015

This page contains the software and data presented in:

  • Preliminary work: "Benchmarking Personal Cloud Storage" by Idilio Drago, Enrico Bocchi, Marco Mellia, Herman Slatman and Aiko Pras. In Proceedings of the 13th ACM Internet Measurement Conference. IMC 2013.

More results of our work on personal cloud storage can be found on this page and on this page.

The slides of the presentation can be downloaded from here.

Benchmarks Scripts

The scripts are written in python. See the README for instructions on how to execute the benchmarks

Traces

Some traffic traces that generated the results in the IMC paper can be downloaded from these links:

Provider File Size
Amazon Cloud Drive 197M
Dropbox 88M
Google Drive 70M
Microsoft SkyDrive 69M
LaCie Wuala 63M

These traces produce the results in Figure 7 of the paper. More details about this dataset can also be obtained in Chapter 5 of:

Traffic Identification Example

The identification of the cloud storage traffic in the testbed can be done using the FQDNs that cloud storage clients use to contact to servers. This can be achieved, for example, by means of the methodology implemented in Tstat DN-Hunter. Here we post a list of FQDNs used by three popular providers as an example.

Acceptable Use Policy

  • When writing a paper using software or data from this page, please cite:
 @inproceedings{drago2013_imc,
   author        = {Idilio Drago and Enrico Bocchi and Marco Mellia and Herman Slatman and Aiko Pras},
   title         = {Benchmarking Personal Cloud Storage},
   booktitle     = {Proceedings of the 13th ACM Internet Measurement Conference},
   series        = {IMC'13},
   pages         = {205--212},
   year          = {2013}
 }

Acknowledgments

This work was partly funded by the Network of Excellence project Flamingo (ICT-318488) and the EU-IP project mPlane (n-318627). Both projects are supported by the European Commission under its Seventh Framework Programme.