https://www.mdu.se/

mdu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Methodological Principles for Reproducible Performance Evaluation in Cloud Computing
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems.ORCID iD: 0000-0002-1364-8127
Vrije Universiteit Amsterdam, Amsterdam, Netherlands.
Vrije Universiteit Amsterdam, Amsterdam, Netherlands.
Vrije Universiteit Amsterdam, Amsterdam, Netherlands.
Show others and affiliations
2021 (English)In: IEEE Transactions on Software Engineering, ISSN 0098-5589, E-ISSN 1939-3520, Vol. 47, no 8, p. 1528-1543, article id 8758926Article in journal (Refereed) Published
Abstract [en]

The rapid adoption and the diversification of cloud computing technology exacerbate the importance of a sound experimental methodology for this domain. This work investigates how to measure and report performance in the cloud, and how well the cloud research community is already doing it. We propose a set of eight important methodological principles that combine best-practices from nearby fields with concepts applicable only to clouds, and with new ideas about the time-accuracy trade-off. We show how these principles are applicable using a practical use-case experiment. To this end, we analyze the ability of the newly released SPEC Cloud IaaS benchmark to follow the principles, and showcase real-world experimental studies in common cloud environments that meet the principles. Last, we report on a systematic literature review including top conferences and journals in the field, from 2012 to 2017, analyzing if the practice of reporting cloud performance measurements follows the proposed eight principles. Worryingly, this systematic survey and the subsequent two-round human reviews, reveal that few of the published studies follow the eight experimental principles. We conclude that, although these important principles are simple and basic, the cloud community is yet to adopt them broadly to deliver sound measurement of cloud environments.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers Inc. , 2021. Vol. 47, no 8, p. 1528-1543, article id 8758926
Keywords [en]
Experimental evaluation, experimentation, observation study, Architectural acoustics, Cloud computing, Economic and social effects, Engineering research, Software engineering, Software testing, Benchmark testing, Computer performance, Performance evaluation, Systematics, Benchmarking
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:mdh:diva-60081DOI: 10.1109/TSE.2019.2927908ISI: 000684687400001Scopus ID: 2-s2.0-85069901345OAI: oai:DiVA.org:mdh-60081DiVA, id: diva2:1701182
Available from: 2022-10-05 Created: 2022-10-05 Last updated: 2022-11-18Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Papadopoulos, Alessandro

Search in DiVA

By author/editor
Papadopoulos, Alessandro
By organisation
Embedded Systems
In the same journal
IEEE Transactions on Software Engineering
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 23 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf