Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Towards a Test Automation Improvement Model (TAIM)
Karlstad University, Faculty of Health, Science and Technology (starting 2013), Department of Mathematics and Computer Science.
2014 (English)In: Proceedings - IEEE 7th International Conference on Software Testing, Verification and Validation Workshops, ICSTW 2014, p. 337-342Article in journal (Refereed) Published
Resource type
Text
Abstract [en]

In agile software development, industries are becoming more dependent on automated test suites. Thus, the test code quality is an important factor for the overall system quality and maintainability. We propose a Test Automation Improvement Model (TAIM) defining ten key areas and one general area. Each area should be based on measurements, to fill the gap of existing assessments models. The main contribution of this paper is to provide the outline of TAIM and present our intermediate results and some initial metrics to support our model. Our initial target has been the key area targeting implementation and structure of test code. We have used common static measurements to compare the test code and the source code of a unit test automation suite being part of a large complex telecom subsystem. Our intermediate results show that it is possible to outline such an improvement model and our metrics approach seems promising. However, to get a generic useful model to aid test automation evolution and provide for comparable measurements, many problems still remain to be solved. TAIM can as such be viewed as a framework to guide the research on metrics for test automation artifacts. 

Place, publisher, year, edition, pages
IEEE Press, 2014. p. 337-342
Keyword [en]
Automation; Computer software; Industry; Measurements; Software engineering; Testing, Agile software development; Automated test; Intermediate results; Source codes; Static measurements; System quality; Test Automation; Test code, Software testing
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kau:diva-43224DOI: 10.1109/ICSTW.2014.38Scopus ID: 2-s2.0-84903647674OAI: oai:DiVA.org:kau-43224DiVA, id: diva2:937731
Conference
7th IEEE International Conference on Software Testing, Verification and Validation Workshops, ICSTW 2014 ; Conference Date: 31 March 2014 Through 4 April 2014
Available from: 2016-06-15 Created: 2016-06-15 Last updated: 2018-01-10Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records BETA

Eldh, Sigrid

Search in DiVA

By author/editor
Eldh, Sigrid
By organisation
Department of Mathematics and Computer Science
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 28 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf