AUT LibraryAUT
View Item 
  •   Open Research
  • AUT Research Institutes, Centres and Networks
  • SERL - Software Engineering Research Laboratory
  • View Item
  •   Open Research
  • AUT Research Institutes, Centres and Networks
  • SERL - Software Engineering Research Laboratory
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Data accumulation and software effort prediction

MacDonell, SG; Shepperd, M
Thumbnail
View/Open
MacDonell and Shepperd (2010) ESEM.pdf (60.52Kb)
Permanent link
http://hdl.handle.net/10292/2441
Metadata
Show full metadata
Abstract
BACKGROUND: In reality project managers are constrained by the incremental nature of data collection. Specifically, project observations are accumulated one project at a time. Likewise within-project data are accumulated one stage or phase at a time. However, empirical researchers have given limited attention to this perspective.

PROBLEM: Consequently, our analyses may be biased. On the one hand, our predictions may be optimistic due to the availability of the entire data set, but on the other hand pessimistic due to the failure to capitalize upon the temporal nature of the data. Our goals are (i) to explore the impact of ignoring time when building cost prediction models and (ii) to show the benefits of re-estimating using completed phase data during a project.

METHOD: Using a small industrial data set of sixteen software projects from a single organization we compare predictive models developed using a time-aware approach with a more traditional leave-one-out analysis. We then investigate the impact of using requirements, design and implementation phase data on estimating subsequent phase effort.

RESULTS: First, we find that failure to take the temporal nature of data into account leads to unreliable estimates of their predictive efficacy. Second, for this organization, prior-phase effort data could be used to improve the management of subsequent process tasks.

CONCLUSION: We should collect time-related data and use it in our analyses. Failure to do so may lead to incorrect conclusions being drawn, and may also inhibit industrial take up of our research work.
Date
September 16, 2010
Source
Presentation at the 4th International Symposium on Empirical Software Engineering and Measurement, Bolzano-Bozen, Italy and published in Proceeding ESEM '10 Proceedings of the 2010 ACM-IEEE International Symposium on Empirical Software Engineering and Measurement ACM New York, NY, USA
Item Type
Conference Contribution
Publisher
IEEE Computer Society Press
DOI
10.1145/1852786.1852828
Publisher's Version
http://dx.doi.org/10.1145/1852786.1852828
http://esem2010.case.unibz.it/program.php
Rights Statement
Copyright © 2010 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

Contact Us
  • Admin

Hosted by Tuwhera, an initiative of the Auckland University of Technology Library

 

 

Browse

Open ResearchTitlesAuthorsDateSERL - Software Engineering Research LaboratoryTitlesAuthorsDate

Alternative metrics

 

Statistics

For this itemFor all Open Research

Share

 
Follow @AUT_SC

Contact Us
  • Admin

Hosted by Tuwhera, an initiative of the Auckland University of Technology Library