Evaluation of Software Visualization Tools: Lessons Learned
Loading...
Date
2009
Journal Title
Journal ISSN
Volume Title
Publisher
IEEE
Abstract
Many software visualization (SoftVis) tools are continuously being
developed by both researchers as well as software development
companies. In order to determine if the developed tools are effective
in helping their target users, it is desirable that they are exposed
to a proper evaluation.
Despite this, there is still lack of a general guideline on how these
evaluations should be carried out and many of the tool developers
perform very limited or no evaluation of their tools. Each person
that carries out one evaluation, however, has experiences which, if
shared, can guide future evaluators. This paper presents the lessons
learned from evaluating over 20 SoftVis tools with over 90 users
in five different studies spread on a period of over two years. The
lessons covered include the selection of the tools, tasks, as well as
evaluation participants. Other discussed points are related to the
duration of the evaluation experiment, its location, the procedure
followed when carrying out the experiment, as well as motivation
of the participants. Finally, an analysis of the lessons learned is
shown with the hope that these lessons will be of some assistance
to future SoftVis tool evaluators.
Description
Keywords
Software Visualization Tools, Lessons
Citation
Sensalire, M., Ogao, P., & Telea, A. (2009). Evaluation of Software Visualization Tools: Lessons Learned. In EPRINTS-BOOK-TITLE University of Groningen, Johann Bernoulli Institute for Mathematics and Computer Science.