skip to main content
10.1145/2365324.2365330acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicseConference Proceedingsconference-collections
research-article

Web effort estimation: the value of cross-company data set compared to single-company data set

Published:21 September 2012Publication History

ABSTRACT

This study investigates to what extent Web effort estimation models built using cross-company data sets can provide suitable effort estimates for Web projects belonging to another company, when compared to Web effort estimates obtained using that company's own data on their past projects (single-company data set). It extends a previous study (S3) where these same research questions were investigated using data on 67 Web projects from the Tukutuku database. Since S3 was carried out, data on other 128 Web projects was added to Tukutuku; therefore this study uses the entire set of 195 projects from the Tukutuku database, which now also includes new data from other single-company data sets. Predictions between cross-company and single-company models are compared using Manual Stepwise Regression+Linear Regression and Case-Based Reasoning. In addition, we also investigated to what extent applying a filtering mechanism to cross-company datasets prior to building prediction models can affect the accuracy of the effort estimates they provide. The present study corroborates the conclusions of S3 since the cross-company models provided much worse predictions than the single-company models. Moreover, the use of the filtering mechanism significantly improved the prediction accuracy of cross-company models when estimating single-company projects, making it comparable to that using single-company datasets.

References

  1. Conte, S. D., Dunsmore, H. E., Shen, V. Y., 1986. Software Engineering Metrics and Models. Benjamin-Cummins. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Turhan, B., Menzies, T., Bener, A., Di Stefano, J., 2009. On the relative value of cross-company and within-company data for defect prediction. ESE, 14: 540--578. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Cohen, J., 1988. Statistical power analysis for the behavioral sciences. Lawrence Earlbaum Ass., 2nd edition.Google ScholarGoogle Scholar
  4. Cook, R. D., 1977. Detection of influential observations in linear regression. Technometrics, 19, 15--18.Google ScholarGoogle Scholar
  5. Kitchenham B., 1998. A Procedure for Analyzing Unbalanced Datasets. IEEE TSE, 278--301. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Kitchenham, B. A., Pickard, L., Peeger, S. 1995. Case studies for method and tool evaluation. IEEE Software 12(4): 52--62. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Kitchenham, B., Pickard, L. M., MacDonell, S. G., Shepperd, M. J, 2001. What accuracy statistics really measure. IEE Proceedings Software 148 (3) 81--85.Google ScholarGoogle ScholarCross RefCross Ref
  8. Kitchenham, B., Mendes, E., 2004. A Comparison of Cross-company and Single-company Effort Estimation Models for Web Applications, in Procs. of EASE, 47--55.Google ScholarGoogle Scholar
  9. Kitchenham B. A., Mendes, M., Travassos, G. H., 2007. Cross versus Within-Company Cost Estimation Studies: A Systematic Review. IEEE TSE 33(5): 316--329. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Kocaguneli, E., Gay, G., Menzies, T., Yang, Y., Keung, J. W., 2010. When to use data from other projects for effort estimation. In Procs. of ASE, 321--324 Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Kocaguneli, E., Menzies, T., 2011. How to Find Relevant Data for Effort Estimation? In Procs of ESEM, 255--264. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Kocaguneli, E., Menzies, T., Bener, A. B., Keung, J. W., 2012. Exploiting the Essential Assumptions of Analogy-Based Effort Estimation. IEEE TSE 425--438. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Li, Y. F., Xie, M., Goh, T. N., 2009. A Study of Project Selection and Feature Weighting for Analogy based Software Cost Estimation. JSS, 82(2), 241--252. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Maxwell, K., 2002. Applied Statistics for Software Managers. Software Quality Institute Series, Prentice Hall.Google ScholarGoogle Scholar
  15. Mendes, E., 2008. Web Cost Estimation and Productivity Benchmarking. ISSSE: 194--222.Google ScholarGoogle Scholar
  16. Mendes, E., Mosley, N., Counsell, S., 2002. Comparison of Length, complexity and functionality as size measures for predicting Web design and authoring effort, IEE Procs. Software 149 (3), 86--92.Google ScholarGoogle ScholarCross RefCross Ref
  17. Mendes, E., Mosley, N., Counsell, S., 2003. Investigating early web size measures for web cost estimation. In Procs. Evaluation and Assessment in Software Engineering, 1--22.Google ScholarGoogle Scholar
  18. Mendes, E., Mosley, N. and Counsell, S., 2005. The Need for Web Engineering: An Introduction. Web Engineering, Springer-Verlag, Mendes, E. and Mosley, N. (Eds.)Google ScholarGoogle Scholar
  19. Mendes, E., Mosley, N., Counsell, S. 2005. Investigating Web Size Metrics for Early Web Cost Estimation. JSS, 77(2), 157--172. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Mendes, E., Kitchenham B. A., 2004. Further Comparison of Cross-Company and Within Company Effort Estimation Models for Web Applications. In Procs of METRICS, IEEE Computer Society, 348--357. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Mendes, E., Di Martino, S., Ferrucci, F., Gravino, C., 2007. Effort Estimation: How Valuable is it for a Web company to Use a Cross-company Data Set, Compared to Using Its Own Single-company Data Set?. In Procs of WWW07, 963--972. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Mendes, E., Di Martino, S., Ferrucci, F., Gravino, C., 2008. Cross-company vs. single-company web effort models using the Tukutuku database: An extended study. JSS 81, 673--690. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Menzies, T., Butcher, A., Marcus, A., Zimmermann, T., Cok, D. R., 2011. Local vs. global models for effort estimation and defect prediction. In Procs. of ASE, 343--351. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Shepperd, M. J., and G. Kadoda, 2001. Using Simulation to Evaluate Prediction Techniques, in Proceedings IEEE 7th Intl Software Metrics Symposium, London, UK, 349--358. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Web effort estimation: the value of cross-company data set compared to single-company data set

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          PROMISE '12: Proceedings of the 8th International Conference on Predictive Models in Software Engineering
          September 2012
          126 pages
          ISBN:9781450312417
          DOI:10.1145/2365324

          Copyright © 2012 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 21 September 2012

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          PROMISE '12 Paper Acceptance Rate12of24submissions,50%Overall Acceptance Rate64of125submissions,51%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader