1 / 25

Institutional Effects on Software Metrics Programs: A Structural Equation Model

Institutional Effects on Software Metrics Programs: A Structural Equation Model. Anand Gopal Robert H. Smith School of Business University of Maryland – College Park. Software Metrics Programs. Software Development Teams. Program of Research.

lilah
Download Presentation

Institutional Effects on Software Metrics Programs: A Structural Equation Model

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Institutional Effects on Software Metrics Programs: A Structural Equation Model Anand Gopal Robert H. Smith School of Business University of Maryland – College Park

  2. Software Metrics Programs Software Development Teams Program of Research • “Communications and Processes in Offshore Software Development”, Communications of the ACM • “Contracts in Offshore Software Development: An Empirical Analysis”, Forthcoming, Management Science • “Contracts and Project Profitability in Offshore Software Development: An Endogenous Switching Regression Model”, Working Paper OffshoreSoftware Development • “Determinants of Metrics Programs Success in Software Development”, IEEE Transactions of Software Engineering • “Institutional Effects on Software Metrics Programs: A Structural Equation Model”, Revise & Resubmit, MIS Quarterly • “Behavioral and Technical Factors Influencing Software • Development Productivity: A Field Study”, Working Paper • “Organizational Control Systems and Software Quality: A Cross- National Investigation”, ICIS 2003, Seattle, Research-in-Progress Software Quality

  3. What are software metrics programs? • Antecedents to measurement-based process improvement initiatives • Primary objective – quantitatively determine the extent to which a software process, product or project possesses a certain attribute • Anecdotal evidence – 2 out of 3 metrics programs fail within the first 2 years • Organizations in the early 1990s did not follow well-defined standard processes for metrics collection and feedback [Humphrey, 1995] • Need to understand factors affecting adoption and acceptance of metrics programs in organizations

  4. Treating metrics programs as an administrative innovation • Administrative innovations exist in highly complex organizational structures • Mere adoption of metrics programs inadequate • Organizations need to ensure adaptation of work-processes through to infusion • Benefits of metrics-based decision-making -> routinization and infusion of metrics into organization • Important to study factors that go beyond just adoption of an innovation • Stage-based approach to innovation diffusion • Apply to both administrative and technical innovations

  5. Stages of Innovation Diffusion in Organizations [Kwon and Zmud, 1987] • Six stage model of innovation diffusion • Initiation • Adoption • Adaptation • Acceptance • Routinization • Infusion • Prior work has studied factors influencing diffusion of innovations in organizations • User, environmental, organizational, technical and task characteristics • Need to consider the institutional aspects [King et al, 1994], especially in the IT / IS context

  6. Institutional Theory • Institutional forces - drive organizations to adopt practices and policies to gain legitimacy • Institutional isomorphism [DiMaggio & Powell, 1983] • Innovation adoption – seen through an institutional lens • Westphal et al [1997], Tan and Fichman [2002] • Software industry – increasing role of institutional forces • Move towards an “engineering” focus • Formal programs in CS/ IS/ Software Engineering • Institutions such as the ACM • Organizations such as the Software Engineering Institute • Understand the role of institutional forces in process innovation infusion into organizations

  7. Research Questions • What factors determine the extent of metrics programs adaptation within an organization? • How is adaptation measured? • How do the institutional forces in the software industry influence the level of adaptation of metrics programs? • Does adaptation lead to acceptance of metrics programs in software organizations? • Does adaptation mediate the relationship between the institutional forces and acceptance of metrics programs?

  8. Background Theory • Metrics Programs – Anecdotal and case literature • Pfleeger [1993] • Daskalantonakis [1992] • Case studies – Eastman Kodak [Seddio, 1993], US Army [Fenick, 1990] • Innovation Diffusion • Kwon and Zmud [1987] • King et al [1994] • Saga and Zmud [1994] • Institutional Theory • DiMaggio and Powell [1983], Meyer and Rowan [1977] • Westphal et al [1977] • Teo et al [2003]

  9. Research Hypotheses • Adaptation – stage in which the innovation is developed, installed and maintained • Org. procedures are revised or created around innovation • New work-practices are developed for the innovative practice • Organizational members are trained both in procedures and use • Hypothesis 1 - The extent of metrics programs adaptation is determined by the following work-processes • Regularity of metrics collection • Seamless and efficient data collection • Use of sophisticated data analysis techniques • Use of suitable communication mechanisms • Presence of automated data collection tools

  10. Research Hypotheses • Hypothesis 2 - Higher levels of institutional forces are associated with higher levels of adaptation • Hypothesis 3 - Management commitment in software organizations is associated with higher levels of adaptation • Hypothesis 4 - Greater levels of adaptation in software organizations are associated with increased acceptance • Acceptance – efforts taken by organizational members to commit to use of innovation in decision-making [Saga and Zmud, 1994]

  11. Structural Model of Metrics Adaptation

  12. Research Methods • Online survey for data collection • Potential respondents sent login and passwords • Data collection through survey questionnaire • Sample from three sources • Private organization conducting tutorials and conferences on metrics • US Department of Defense – organization that coordinated metrics activities for contractors and software divisions • Attendees of the SEI’s training programs in metrics programs • Response rate ~ 59% → final sample size of 214 • 130 from defense contractor or DOD organization • 84 from commercial sector • Average respondent – 8 years experience

  13. Research Variables • Adaptation – measured through individual work-processes • Metrics Regularity – 4 items, Pressman [1997] • Data Collection – 3 items, Daskalantonakis [1992] • Quality of Data Analysis – 4 items, Briand et al [1996] • Communication – 4 items, Kraut and Streeter [1995] • Presence of automated tools – 3 items, Hall and Fenton [1997] • Exploratory factor analysis – each individual work-process loads well on items • Discriminant validity – factor analysis on all questionnaire items show the presence of 5 factors • Reliability – above 0.70 Cronbach’s alpha • Confirmatory factor analysis using Lisrel • Use factor scores in subsequent analysis

  14. Exploratory Factor Analysis – Adaptation of Work-processes

  15. Confirmatory Factor Analysis – Adaptation of Work Processes

  16. Research Variables • Institutional Forces – measured using 5 items • Little prior work in capturing these concepts in the IS literature • Exploratory in nature • Good reliability (alpha=0.81), load well on one factor • Management Commitment – measured using 4 items • Adapted from Igbaria [1990] • Demonstrated support and allocation of resources • Metrics Acceptance – measured using 4 items • Frequency with which members use metrics-related information in decision-making • Good reliability (alpha=0.76, load well on one factor

  17. Data Analysis • Structural model estimated using Lisrel • Use factor scores for Metrics Adaptation rather than original items • Assumption of multivariate normality not rejected • Multivariate skewness = 1.089 • Univariate skewness < 2, kurtosis < 7 [Curran et al, 1996] • Estimation performed using variance-covariance matrix using Maximum Likelihood • Measurement model strongly significant • Structural model significant at GFI = 0.88 • Comparative fit index = 0.90, root mean square residual = 0.05

  18. Structural Model - Results Degrees of Freedom = 130 Minimum Fit Function Chi-Square = 290.09 (p=0.00) Satorra-Bentler Scaled Chi-Square = 265.01 (p=0.00) Standardized Root Mean Square Residual = 0.052 Goodness of Fit Index = 0.87 Comparative Fit Index = 0.91

  19. Mediation of Adaptation on Acceptance • Structural model tested with direct path from Institutional Forces to Acceptance • Other paths remain the same • Insignificant path from Institutional Forces to Acceptance • Change in chi-square not significant • Results indicate that Adaptation fully mediates the relationship between Institutional Forces and Acceptance • Although organizational mandate can cause orgns to adopt metrics, acceptance requires adaptation of work-processes

  20. Summary of Results • All four hypotheses strongly supported by structural model • Institutional forces influence the level of adaptation and indirectly the level of acceptance of metrics-based decision-making • Management commitment key in adaptation • Adaptation leads to acceptance – support for the six-stage model of innovation diffusion • Measurement of adaptation – confirmatory factor analysis • Five individual work-practices provide strong measure of adaptation

  21. Limitations • Most of the data is perceptual • Respondent bias • Common method variance • List of work-processes for adaptation not exhaustive • Several other factors mentioned in case literature • Some common control variables missing • Organizational size • Organizational slack

  22. Future Work • Augmenting survey data with objective data from organizations • Clearly show the benefits / costs of metrics programs • Why do metrics programs fail? • The role of institutions in the software industry • The effects on standards • Influence on software development methodologies • Institutional forces and their influences on software industries in different countries • Measurement issues

  23. Institutional Effects on Software Metrics Programs: A Structural Equation Model Anand Gopal Robert H. Smith School of Business University of Maryland – College Park

  24. Correlation Table

  25. Measurement Model Results

More Related