Complexity of Information Systems Development Projects: Conceptualization and Measurement Development Weidong Xia Department of Information and Decision Sciences Carlson School of Management University of Minnesota 321 19th Avenue South Minneapolis, MN 55455 Phone: (612) 626-9766 Emails: [email protected] Gwanhoo Lee Department of Information Technology Kogod School of Business American University 4400 Massachusetts Avenue, NW Washington, DC 20016-8044 Phone: (202) 885-1991 Email: [email protected] Acknowledgments: This study was supported by research grants provided by the University of Minnesota and by the Juran Center for Leadership in Quality. The Information Systems Special Interest Group of the Project Management Institute sponsored the collection of the survey data. We thank Carl Adams, Shawn Curley, Gordon Davis, Paul Johnson, Rob Kauffman and research workshop participants at the University of Minnesota for helpful comments on earlier versions of the paper. Complexity of Information Systems Development Projects: Conceptualization and Measurement Development Abstract This paper conceptualizes and develops valid measurements of the key dimensions of information systems development project (ISDP) complexity. A conceptual framework is proposed to define four components of ISDP complexity: structural organizational complexity, structural IT complexity, dynamic organizational complexity, and dynamic IT complexity. Measures of ISDP complexity are generated based on literature review, field interviews, focus group discussions and two pilot tests with 76 IS managers. The measures are then tested using both exploratory and confirmatory data analyses with survey responses from managers of 541 ISDPs. Results from both the exploratory and confirmatory analyses support the fourcomponent conceptualization of ISDP complexity. The final 20-item measurements of ISDP complexity are shown to adequately satisfy the criteria for unidimensionality, convergent validity, discriminant validity, reliability, factorial invariance across different types of ISDPs, and nomological validity. Implications of the study results to theory development and practice as well as future research directions are discussed. Keywords and phrases: Complexity; Information Systems Development Projects; Conceptual Framework; Scale Development; Exploratory Factor Analysis; Confirmatory Factor Analysis 1 Complexity of Information Systems Development Projects: Conceptualization and Measurement Development Introduction Information systems (IS) development is inherently complex because it must deal with not only technological issues but also organizational factors that by and large are outside of the project team’s control [13, 70, 71]. As organizations increasingly depend on IS for one-stop customer service capabilities and cross-selling opportunities, any new development efforts must be seamlessly integrated with other existing systems, introducing a system-level complexity that is often constrained by the existing technology architecture and infrastructure [86]. In addition, as both information technology and business environments are fast changing, it becomes increasingly difficult to determine business requirements and to freeze system specifications, making system development a progressively more dynamic and complex process [83, 92]. The complexity of IS development is manifested by the historically high failure rate of information systems development projects (ISDPs). In the last four decades, there have been many reports about ISDP failures in various organizations across industries (e.g., [42, 57]). The Standish Group reports that US companies spent more than $250 billion each year in the early 1990s on IS projects, with only 16.2 percent considered successful [78]. The Standish Group’s 2001 report indicates that US companies invested four times more money in IS projects in 2000 than they did annually in the 1990s; however, only 28 percent of the projects could be considered successful [79]. The results suggest that, although much improved, the success rate of IS projects is still very low. As organizations increasingly invest in IS with the intention to enhance top-line growth and bottom-line savings, IS project failures have significant organizational 2 consequences, in terms of both wasted critical resources and lost business opportunities. Therefore, there is a strong economic incentive for companies to improve IS project performance. ISDP failures are not isolated incidents, but rather they recur with a great deal of regularities in organizations of all types and sizes [28, 30, 44, 79]. Many experts believe that ISDPs are uniquely more complex than other types of projects such as construction and product development projects. Edward W. Felten, a computer science professor at Princeton University was recently quoted: “A corporate computer system is one of the most complex things that humans have ever built” ([4], p. 118). Brooks [13], in his famous book, Mythical Man-Month, states that “software entities are more complex for their size than perhaps any other human construct” and that “software systems differ profoundly from computers, buildings, or automobiles” (p. 182). He contends that the complexity of software is an essential property, not an accidental one. Such essential complexity is unique to software development because software is invisible and unvisualizable, and is subject to conformity and continuous changes. Brooks concludes, “many of the classical problems of developing software products derive from this essential complexity” (p. 183). Projects are temporary organizations within organizations [22]. Different types of projects demonstrate different contingency characteristics that require different management approaches. The literature on project complexity in general is still in early development. Some researchers argue that one of the difficulties in advancing theory development on project complexity may stem from the conventional approach of developing “one-size-fits-all” theories (e.g., [5, 27, 77]). In a recent study, Shenhar [75] shows “one-size-does-not-fit-all” and calls for a more contingent approach to the study of projects. Given the call for contingency theories for 3 different types of projects and the unique characteristics of ISDPs, it is necessary to develop theories for ISDP complexity, as opposed to theories for general project complexity. By utilizing the general project complexity concepts, research on ISDP complexity may provide new insights that will contribute to the general project management literature. Managing ISDP complexity has become one of the most critical responsibilities of IS managers and executives [38, 72]. However, IS organizations have displayed great difficulties in coping with the progressively increasing ISDP complexity [9, 61]. Before we can develop effective strategies to control and manage ISDP complexity, it is necessary that we understand the project characteristics that constitute ISDP complexity and are able to use those characteristics to assess the complexity of an ISDP. However, research on ISDP complexity has mostly been conceptual and anecdotal. The ISDP complexity construct has often been used without precise definitions and appropriate operationalizations. To our knowledge, no systematic frameworks and validated measurements of ISDP complexity have been reported. This research represents a first step towards conceptualizing and developing measures of the key dimensions of ISDP complexity. The research is conducted using a systematic fourphase measurement development process involving multiple data sources and research methods. The conceptual framework and the initial measures of ISDP complexity are developed based on literature reviews, field interviews, focus group discussions and two pilot tests with IS managers. The conceptual framework and measures are then systematically tested using survey data provided by 541 ISDP managers. The four-component conceptualization of the ISDP complexity construct is tested through analyzing the factor structures underlying the measures. The measures are tested by both exploratory and confirmatory analysis techniques, each with a random split-half sub-sample of the overall sample. A comprehensive set of measurement tests 4 are conducted on the measures, including reliability, unidimensionality, convergent validity, discriminant validity, factorial invariance across three types of ISDPs, and nomological validity. Using multiple data sources, multiple methods and multiple measurement validation criteria help triangulate the analyses and enhance the quality and creditability of the results. The development of such conceptual frameworks and measurement makes significant contributions to both theory development and practice. For researchers, well-developed conceptual frameworks of ISDP complexity provide the basis for consistently defining the construct across studies so that results of different studies can be meaningfully compared. Empirically developed theories involving ISDP complexity are not possible without valid measures of ISDP. In addition, theoretical development in ISDP complexity will contribute to the general project management literature by providing insights about project complexity under a specific contingent context. For practitioners, sound conceptual frameworks provide a critical language and lens through which managers can describe and communicate the key dimensions of ISDP complexity. Operational indicators provide managers with the necessary tools to assess the complexity of specific ISDPs. The ability to assess ISDP complexity is a prerequisite for developing effective strategies to control and manage ISDP complexity. The paper is organized as follows. The next section reviews relevant prior literature and proposes a conceptual framework that defines the key dimensions and components of ISDP complexity. The following section describes the four-phase measurement development process through which the research is conducted. The results of the measurement development and testing are then discussed. The paper concludes with discussions of the contributions and limitations of the research as well as directions for future research. 5 Theoretical Background In this article, IS development refers to the analysis, design and implementation of IS applications/systems to support business activities in an organizational context. ISDPs are temporary organizations that are formed to perform IS development work including new applications/systems yet to be installed, as well as enhancement of existing applications/systems [80]. New applications/systems include both in-house application/systems and off-the-shelf packaged software. By building on the existing literature and incorporating insights gained from ISDP managers through field interviews and focus group discussions, we attempt to define the key dimensions of ISDP complexity and develop operational indicators that capture the most critical characteristics of the project that reflect the complexity of the system development process. Recognizing that it is impossible to capture all dimensions and characteristics of ISDP complexity, our intention is to develop a parsimonious set of measures that provide a starting point for developing measurement theories of ISDP complexity. The literatures on task complexity, project complexity, software complexity, and software project risk factors are particularly relevant to our study and are thus used as the bases for developing our conceptual framework and measures. Task complexity ISDPs are complex organizational tasks. The task complexity literature has identified a variety of dimensions and task characteristics that constitute task complexity. For example, Campbell [16] viewed complexity as a function of four task characteristics. He proposes that task complexity increases when (1) only one path leads to goal attainment while multiple paths 6 exist, (2) multiple desired outcomes are required, (3) there exists conflicting interdependence among paths, and (4) the connection between path activities and desired outcomes cannot be established with certainty. Wood [93] defines three types of task complexity: component, coordinative, and dynamic. Component complexity of a task is a function of the number of distinct acts that need to be executed in the performance of the task and the number of distinct information cues that must be processed in the performance of those acts. As the number of acts increases, the knowledge and skill requirements for a task also increase simply because there are more activities that an individual needs to be aware of and is able to perform. Coordinative complexity refers to the nature of the relationships between task inputs and task outputs. The form and strength of the relationships between information cues, acts, and products are aspects of coordinative complexity. As the requirements for timing, frequency, intensity, and location of acts become more complex, the difficulty for coordination increases. Dynamic complexity is caused by changes in the states of the task environments. Although not explicitly acknowledged by Wood [93], there is a hierarchical relationship between the three types of task complexity. The overall task complexity is first determined by the component-level complexity. Then, it is affected by the system-level complexity involving coordinations among components. While both component complexity and coordinative complexity describe the structural configurations of the task that are relatively static, dynamic complexity describes the uncertainties caused by changes in the task environments. Performance of a dynamically complex task requires knowledge about how the component and coordinative complexities change over time. To some extent, one may view component complexity as a first order complexity, coordinative complexity as a second order and dynamic complexity as a third order complexity. 7 Project complexity There seem to be no validated measures for assessing general project complexity. Measures used in the empirical studies have mostly been project specific. It is difficult to develop general “one-size-fits-all” complexity concept and measures that are applicable to all types of projects. Most of the commonly identified project complexity dimensions are consistent with those identified in the task complexity literature, albeit for different contexts and levels of analysis. Based on a review of the project complexity literature, Baccarini [2] defines project complexity in terms of the number of varied elements and the interdependency between those elements. Following this definition, he proposes two types of project complexity: organizational complexity and technological complexity. Organizational complexity refers to the number of, and relationships between, hierarchical levels, formal organizational units, and specializations. Technological complexity refers to the number of, and relationships between, inputs, outputs, tasks, and technologies. Turner and Cochrane [84] propose uncertainty as another dimension of project complexity, which is the extent to which the project goals and means are ill defined and are subject to future changes. Uncertainty in systems requirements/scope and uncertainty in new information technologies are examples of goal and mean uncertainties, respectively. By integrating the dimensions proposed by Baccarini [2] and Turner and Cochrane [84], Williams [91] defines two distinct aspects of project complexity: structural complexity (the underlying structure of the project) and uncertainty-based complexity (the uncertain or changing nature of the project). He contends that uncertainty adds to the complexity of a project so it can be viewed as a constituent dimension of project complexity. Dvir et al. [26] propose four levels of task technological uncertainty (low, medium, high and super-high) and three levels of 8 complexity (an assembly project, a system project and an array project or program). While the uncertainty-based complexity dimension is based on the level of technological uncertainty at the initiation stage of the project, the structural complexity dimension is based on the scope or a hierarchical framework of systems and subsystems. Software complexity Software is one of the major outcomes of an ISDP. Numerous frameworks and measures of software complexity have been proposed in the last two decades. Among the most frequently cited measures are the number of program statements [88], McCabe’s cyclomatic number [56], and Halstead’s programming effort [36]. Banker and Slaughter [6] use the number of data elements per unit of application functionality to measure the total data complexity of an application. Tait and Vessey [81] measure system complexity in terms of the difficulty in determining the information requirements of the system, the complexity of processing, and the overall complexity of the system design. Meyer and Curley [58] define technology complexity of an expert system as a composite measure of diversity of technologies, database intensity, and systems integration effort. Based on the basic functions of an application software, the function point analysis literature defines a complexity index that takes into considerations of the specific complexity of each software function [31]. In addition, the complexity of the general development environment of the application software is assessed using fourteen characteristics of the environment. Example characteristics include complexity levels with regard to data communication, distributed data processing, transaction rate, online update, complex processing, multiple sites, and installation ease. 9 Our review reveals that there is a rich body of research literature on software complexity. However, software complexity cannot be directly applied to project level complexity for the following reasons. First, software complexity and ISDP complexity are at two different levels of analysis. Software complexity is at software code and structure level, which is too limited to capture the broader organizational and technological factors that constitute IS development complexity at the project level. Second, a prerequisite for assessing software complexity is the existence or detailed knowledge about the specific software. In the initial stage of an ISDP, the software often does not exist and systems requirements are often poorly defined. It is difficult to assess software complexity in the initial stage of an ISDP. However, the software complexity literature is useful to derive some measures of the technological dimensions of ISDP complexity for this research. Software project risk factors According to McFarlan [57], failure to assess and manage individual IS project risk is a major source of the software development problems. In the risk management literature, a risk is defined as the product of (1) the probability associated with an undesirable event and (2) the consequences of the occurrence of this event [34]. Consistent with this general definition, IS development risk is commonly defined as the product of the probability of the occurrence of negative conditions or events and their consequences if they do occur [10, 18, 65, 76, 90]. Recognizing the difficulties associated with accurately estimating the probabilities of occurrence of negative conditions/events and the consequences of project losses, researchers have suggested alternative approaches to defining, assessing and managing IS development risk. In lieu of assessing the probabilities of undesirable conditions/events, for example, Boehm [10] 10 recommends identifying and assessing risk factors that influence the occurrence of those negative conditions and events. Barki et al. [7] define IS project development risk as the product of the uncertainties associated with project risk factors and the magnitude of potential loss due to project failure. Accordingly, once the risk factors are identified, one can then estimate IS development risk by assessing the probabilities of their occurrence and the likely loss associated with project failure. Following this alternative approach, researchers have generated a number of checklists of IS project risk factors. For example, Boehm and Ross [11] provide a “top-ten” checklist of IS project risk factors. Barki et al. [7] develop an instrument for assessing various risk factors that can affect IT projects. Schmidt et al. [70] develop a ranked list of IS project risk factors. ISDP risk and ISDP complexity are two related but different concepts. IS project risk factors are related to the probabilities of the negative conditions or events and the magnitude of losses associated with project failure. ISDP complexity, on the other hand, refers to the characteristics of the project that constitute difficulties for the development effort. In the context of ISDPs, although risk factors and complexity both represent some negative connotations, complexity tends to capture project characteristics that are inherently present in projects whereas risk factors represent possible events or conditions that cause project failures. The two constructs are obviously related. As the inherent project characteristics, indicators of ISDP complexity represent the underlying factors that may drive project risks. Recognizing the conceptual difference and relatedness between the two constructs, in this research, we use the IS project risk factor literature as a basis for developing our frameworks and measures. In summary, our review of the IS literature suggests that, although the ISDP complexity construct has been frequently mentioned, there exist no well-developed frameworks that can be 11 used to delineate its conceptual meanings. No systematically validated measures for ISDP complexity have been reported. In the general task and project literature, most studies have used measures that are specific to the types and contexts of the tasks/projects, no general measures of task complexity or project complexity have been reported. In the software complexity literature, a number of frameworks and measurements have been developed, representing one of the most advanced areas among the literatures we reviewed. The IS project risk literature has primarily focused on identifying ranked lists of project risk factors. The various literatures provide a useful basis for developing our frameworks and measures of ISDP complexity. Our literature review suggests that there are a few commonly suggested dimensions and characteristics of complexity that can be adapted for conceptualizing and measuring ISDP complexity. For example, at the task level, complexity may consist of three hierarchical levels: component complexity, coordinative complexity, and dynamic complexity. At the project level, complexity can be described as either organizational or technological complexity, either structural or dynamic complexity, and either component or system complexity. A Conceptual Framework for ISDP Complexity In addition to the literature review, in the last three years, we have worked closely with the CIOs and IT project managers from 14 large US companies to gain insights on the dimensions and measures of ISDP complexity. We also used focus group discussions with ISDP managers to identify the most important dimensions and project characteristics that can be used to conceptualize and assess ISDP complexity. Based on the literature review and on the insights that we obtained from working with the CIOs, interviews and focus group discussions with ISDP managers, we propose a conceptual framework of ISDP complexity (shown in Figure 1). The 12 framework is composed of two dimensions. The first dimension captures whether the complexity is about the structural setup or the dynamic aspects of the project; the second dimension captures whether the complexity is about the organizational aspects or the technological aspects of the project. In this framework, each dimension suggests two distinct aspects of ISDP complexity rather than being a continuum-based variable. ============================ Figure 1 about here ============================ The first dimension, structural versus dynamic complexity, is consistent with those proposed by Wood [93], Turner and Cochrane [84], Williams [91], and Dvir [26]. Typical ISDPs involve a number of components including the existing systems, infrastructure, new technology, user units, stakeholders, the project team, vendors, and external service providers. As the number of components increases, it becomes more difficult to monitor and control the project. Relationships among these components make it even more difficult to predict the project’s process and outcome. According to Leveson [51], the problems in building complex systems today often arise in the interfaces between the various components such as hardware, software, or human components. Dynamic complexity refers to the complexity caused by changes in project components and in their relationships. Changes may result from either the stochastic nature of the environment or a lack of information and knowledge about the project environment. As these changes occur, the cause and effect relationship becomes ambiguous and nonlinear. Dynamic complexity becomes particularly relevant and critical for ISDPs because their environments, both business and IT, are constantly changing. Conventional management methods are not adequate 13 in dealing with dynamic complexity, although they can handle structural complexity relatively well [73]. The second dimension, organizational versus technological complexity, has been widely accepted in the general project management literature (e.g., [2, 91]) and in the IS software project risk factor literature [48]. Organizational factors of an ISDP include organizational structure, business processes, organizational information needs, user involvement, top management support, and project personnel capabilities. IT factors include not only “hard” technological components such as hardware, software, and network, but also “soft” technological components such as project staff’s knowledge, skills, and experiences with technologies. Meyer and Curley [58] proposes that the technological complexity in the context of expert systems consists of such technological variables as diversity of platforms, diversity of technologies, database intensity, and systems integration effort. The distinction between the organizational and technological complexity is important because they require different project capabilities, thus have different implications for project management. Based on the two dimensions, we define four components of ISDP complexity: Structural Organizational Complexity (SORG), Structural IT Complexity (SIT), Dynamic Organizational Complexity (DORG), and Dynamic IT Complexity (DIT). The SORG component reflects the nature and the strength of the relationships between the project elements and the organizational supporting environment, e.g., project resources, support from top management and users, project staffing, and the skill proficiency levels of the project personnel. The SIT component captures the coordinative complexity among the IT elements, reflecting the diversity of user units, software environments, nature of data processing, variety of technology platform, need for integration, and the diversity of external vendors and contractors. The DORG 14 component captures the rate and pattern of changes in the ISDP organizational environments, including changes in user information needs, business processes, and organizational structures. It also reflects the dynamic nature of the project’s impact on the organizational environment. The DIT component measures the pattern and rate of changes in the IT environment of the ISDP, including changes in IT infrastructure, architecture and software development tools. In the following sections, we develop measures of ISDP complexity based on this conceptual framework and validate the framework and the measures using empirical data collected from ISDP managers. Measurement Development Process We used a systematic four-phase process involving a variety of methods to develop and validate the measurement of ISDP complexity. This four-phase process is developed based on Churchill [20] and Sethi and King [74]. As shown in Figure 2, the four phases are (1) conceptual development and initial item generation, (2) conceptual refinement and item modification, (3) survey data collection, and (4) data analysis and measurement validation. ============================ Figure 2 about here ============================ Phase 1 - Conceptual development and initial item generation In this phase, the conceptual framework and an initial pool of measurement items were first developed through literature reviews, field interviews and focus group discussions with ISDP managers. In developing the measures, whenever possible, we adapted relevant measures in the literature. A focus group discussion with 45 IS managers using nominal group techniques was conducted to independently generate a ranked list of ISDP complexity items. The combined 15 pools of items were then verified and further modified through interviews with a number of IS senior managers and ISDP project managers. By combining literature reviews with focus group discussions and a number rounds of interviews with ISDP managers, we attempt to ensure the face and content validity of the measures, i.e., to ensure that the measures cover the appropriate scope/domain of ISDP complexity. A total of 30 items were generated as the initial pool of measures. To save space, below we discuss the rationales and sources of the final 20 measures used to capture the four components of ISDP complexity. Appendix A lists the sources of the final 20 items. Measures of the first complexity component, structural organizational complexity, are generally associated with the roles of and relationships among the various stakeholders of the project. Prior literature has identified important stakeholders in information systems development, including top management, end users, project managers, and project staff [12, 28, 49, 66]. Five items are used to capture complexity that is related to those stakeholders. Top management support has been considered one of the most critical success factors in information system development [53, 82]. As ISDPs often involve conflict of interests across user units and organizational changes, lack of sustainable commitment and support from top management makes it difficult to achieve the project objectives. As ISDPs depend on users for requirement determination and for effective implementation and use, user involvement is critical for avoiding poorly defined system scope and requirements and for avoiding user resistance in making changes that are necessary for implementing the system [37, 41, 54]. Lack of a project manager’s control over the project is a significant aspect of project complexity that may increase project risk [48]. Appropriately staffing projects with personnel who have the right skills and backgrounds is a significant factor that reflects the complexity of an ISDP [7, 68, 71]. 16 Measures of the second component of ISDP complexity, structural IT complexity, captures the complexity associated with (1) the number of the components that are directly or indirectly related to technologies and (2) the cross-functional coordination of these components. Seven items were used to measure SIT. Two items were used to capture the complexity related to the multiplicity or heterogeneity of the software development environments and technological environment [58]. One item was used to assess the complexity associated with the coordination requirements with multiple external vendors [7, 43, 70, 94]. Three items were used to describe the complexity related to coordinations among the various components, specifically with managing cross-functional project teams [29], coordinating multiple business units that are involved in the ISDP [7, 43, 70] and integrating the external systems that interface with the system under development [58]. One item was used to capture the complexity related to realtime data processing involved in the new system [31, 62, 69]. Measures of the third component of ISDP complexity, dynamic organizational complexity, captures the complexity associated with changes in the organization. The interactions between information systems development and organizational changes are bidirectional [64]. As such, three items are used to capture changes in the organizational structure and business processes that affect business requirements. Two items are used to capture the business changes that are caused by the information systems delivered by the ISDP [8, 71]. In today’s hypercompetitive business environment, users’ information needs change frequently, which require substantial coordination and rework [10, 68, 70]. Information systems and business processes are closely interweaved in today’s organizations [17, 40]. Business process changes increase ISDP complexity because they necessitate dynamic alignment between business processes and information systems under development. In addition, changes in 17 organizational structure causes changes in the information flows and systems scope of the project, which in turn increase the complexity of the ISDP. Measures of the last component of ISDP complexity, dynamic IT complexity, captures the complexity associated with changes in the technological environments and development tools. Three items are used to assess DIT. One item was used to measure complexity related to changes in IT architecture. IT architecture refers to the overall structure of the corporate information system and consists of the applications and databases for various levels of the organization [24, 67]. The second item measures changes in IT infrastructure. As application systems are developed under the constraints of the existing IT infrastructure of the organization [14, 25, 86], changes in IT infrastructure cause significant uncertainty in ISDPs. The third item captures changes in systems development tools. Adoption of new development tools during an ongoing ISDP causes interruptions because the developers must take time to learn the new tools and adjust their initial analysis and design to suite the new development environments. Phase 2 – Conceptual refinement and item modification The framework and initial pool of 30 measures resulted from Phase 1 were refined and modified through a sorting procedure and two pilot tests. The sorting procedure was used to qualitatively assess the face validity and the construct validity of the initial items. Four IS researchers with an average of eight years of IS work experience participated in the sorting procedure. Details of the sorting procedure are provided in Appendix B. Overall, 26 measures were retained after the sorting procedure. To further validate the relevance, coverage, and clarity of the measurement items, we conducted two pilot tests. The first pilot test was conducted through one-hour individual 18 interviews with four IS project managers and three IS researchers. In the interview, the participant first filled out a questionnaire regarding the importance and relevance of each item to ISDP complexity. They were then asked to identify items that appeared to be inappropriate or irrelevant to ISDP complexity. Participants also made suggestions for improving the relevance, coverage, understandability, and clarity of the items. Five items were dropped based on the results of the pilot test. Two items were combined into a single item because of their similarity. After refining the items, we created an online survey questionnaire using the remaining 20 items. To reveal any potential problems or issues with web-based online survey, a second pilot test was conducted with 15 ISDP managers who had an average of seven years of experience in ISDP management. The ISDP managers logged on to the web survey using their individually assigned IDs and filled out the questionnaire based on their experience with the most recently completed ISDPs they were involved in. After finishing the survey, the manager provided suggestions for improving the content and the format of the online survey. Overall, only minor editorial issues related to the format and wordings of the questionnaire were reported and resolved. Phase 3 – Survey data collections The web-based online survey resulted from the second phase of the research process was used to collected the large-scale data for validating the conceptual framework and the measures of ISDP complexity. The items were randomized to minimize any bias from the survey method. Seven-point Likert scales were used for the items measuring ISDP complexity. The source of the survey respondents was the Information Systems Specific Interest Group of the Project Management Institute (PMI-ISSIG) which is an international organization with about 15,000 19 members of IS project professionals. We used three criteria to select our target respondents: (1) North American PMI-ISSIG members who (2) were project managers (not specialists such as programmers or systems analysts), and (3) had managed a recently completed ISDP. The reason for choosing North American members was to avoid bias and problems that might be caused by language barriers that the non-English speaking members in the other regions might have. The PMI-ISSIG sponsored this research by providing their membership information and promoting the survey to their members. A PMI-ISSIG-sponsored email letter with a hyperlink to the web-based survey was sent to the target group. To encourage participation, survey participants were entered into a drawing to receive ten awards of a PMI-ISSIG official shirt and forty awards of a $25 gift certificate from a well-known online store. A reminder was sent two weeks after the initial PMI-ISSIG-sponsored email was sent out. A second email reminder was sent two weeks later. The total number of potential respondents was 1,740. In total, 565 responses were received, representing a response rate of 32.5%. Twenty-four incomplete responses were dropped, resulting in a usable sample size of 541 and a final response rate of 31.1%. Given the nature of the survey, this response rate is relatively high. Table 1 illustrates the characteristics of the study sample. The sample represents various industry sectors, ranging from manufacturing, financial services, software, consulting, retailing, transportation, healthcare, to utility. On average, companies in the sample had annual sales of $2.55 billion with 14,800 employees. Three types of ISDPs – in-house new development, packaged software implementation, and enhancement of existing software – were evenly represented in the sample. On average, projects in the sample had a budget of $2.1 million, team size of 34 members, and duration of 12 months. 20 Since the sample represented a broad range of companies and projects, the findings are unlikely to be biased by the sample. ============================ Table 1 about here ============================ Phase 4 – Data analysis and measurement validation 4.1. Data screening and descriptive analysis The survey data were carefully screened for unusual patterns, non-response bias, and outliers. A careful screening of the responses did not reveal any unusual patterns or careless responses, indicating that the questionnaire’s design was appropriate and the respondents were serious and careful in completing the questionnaires. To examine non-response bias, we recorded the dates on which the responses were received. Comparisons of early responses and later responses on key demographic and complexity item scores did not identify any significant differences, indicating that response bias is unlikely to be a problem. In addition, using three standard deviations from the mean as a benchmark, we did not find any outliers on the complexity item scores. 4.2. Sample split Since the framework and measures of ISDP developed in this study were new, we used a combination of exploratory and confirmatory factor analysis methods to take advantage of the strengths of both methods and to facilitate cross-validation. The sample was first split into two equally-sized sub-samples using random numbers. The first sub-sample was used in the exploratory factor analysis and the second in the confirmatory factor analysis. 21 To ensure that the two sub-samples were comparable and unbiased, we used t-tests to examine the equality between the two sub-samples’ means on the demographic data. As shown in Table 2, the results indicate that there were no significant differences between the two subsamples in company size indicated by the number of employees and annual sales, project size indicated by the number of project members and project budget, and project duration. ============================ Table 2 about here ============================ 4.3. Exploratory factor analysis Exploratory factor analysis was conducted to examine the factor structure of the measures and to validate the reliability and construct validity of the measures. A common factor analysis using principle component methods with varimax rotations was conducted to determine the factor structure of the measures. The number of factors was determined based on two criteria: eigenvalue above 1.0 and a scree plot. The reliability of the measures is indicated by the Cronbach’s alpha. The convergent validity and discriminant validity are indicated by the factor structure and factor loadings of the measures. 4.4. Confirmatory factor analysis Confirmatory factor analysis with LISREL was used to test the measures that were resulted from the exploratory factor analysis. First, in order to investigate the appropriateness of the measurement model structure, five alternative models were generated and compared based on the overall goodness of model fit indexes. The five models were: (a) a null model with all measures uncorrelated to each other, (b) all measures were loaded onto a single first-order factor, (c) the measures were loaded onto four uncorrelated first-order factors, (d) the measures were 22 loaded onto four correlated first-order factors, and (e) there existed a second-order factor name ISDP complexity above the four first-order factor. These five alternative models are illustrated in Figure 3. The four first-order factors correspond to the four components of ISDP complexity as defined in the conceptual framework. The alternative models were compared using two groups of goodness of fit indexes. The first group of indexes, including the p-value of χ2 statistic, the ratio of χ2 to degrees of freedom, the Goodness of Fit Index (GFI), the Adjusted Goodness of Fit Index (AGFI), the root mean square error of approximation (RMSEA), and the standardized Root Mean Square Residual (RMR), are absolute indexes because they are sensitive to sample size. The second group of index, including the Comparative Fit Index (CFI), and the Normed Fit Index (NFI), are relative indexes that are less sensitive to sample size [32, 46]. The existence of a second-order factor is justified by applying the target coefficient [55]. ============================ Figure 3 about here ============================ As a result of the model comparisons, the “best-fitted” measurement model was chosen for further measurement validation. The unidimensionality and convergent validity of the four latent components were assessed by specifying a single factor model for each latent variable. The reliability was assessed by the composite reliability index that was calculated based on factor loadings and variances [87]. The discriminant validity of the first-order factors was assessed using the techniques suggested by Venkatraman [85] and Sethi and King [74]. 4.5. Factorial invariance analysis Factorial invariance analysis was conducted to establish the generalizability of the measures across three types of ISDPs. After eliminating thirty data cases with missing data on 23 project type, the remaining data of the overall sample were segmented by project type. The sample size was 195 for in-house new development projects, 173 for packaged software implementation projects, and 143 for major enhancement of existing software, respectively. In conducting the factorial invariance analysis, a baseline model was first established and tested for model-data fit. A measurement model with invariance of factor loadings was then specified and tested. If the difference in model fits between the baseline model and the model with invariance constraints was not significant, invariance of the factorial structure of the measurement across the three ISDP types was supported. 4.6. Nomological validity Finally, the nomological (or predictive) validity of the ISDP complexity measure was examined. A positive association between ISDP complexity and project duration was predicted. In order to test this predicted relationship, a composite score was calculated for each of the four factors based on their corresponding items. In addition, an overall ISDP complexity score was obtained by averaging the four factor scores. Since projects in the sample were completed at the time of data collection, project duration data was available. A path analysis was used to test whether the relationship between ISDP complexity and project duration was positive as predicted, if yes, it indicates the nomological or predictive validity of the ISDP complexity measures. Results Exploratory factor analysis Construct validity Exploratory factor analysis using principle component method with varimax rotation was used to test the factor structure of the measurement. As shown in Table 3, four factors with 24 eigen values greater than one emerged from the analysis, which can be interpreted as corresponding to the four components of ISDP complexity. A scree plot test also indicated that the four-factor structure was reasonable. These four factors collectively explained 55.2 percent of the variance. There were no cross-loaded items with loadings greater than 0.30. Overall, these results provide the initial empirical support to the conceptual framework and to the convergent and discriminant validity of the measures of the ISDP complexity. ============================ Table 3 about here ============================ Reliability Reliability refers to the internal consistency of measurement items within each construct. Table 4 reports the corrected item-total correlations for individual items and the Cronbach’s alphas for the factors. All four factors had Cronbach’s alphas higher than 0.70, indicating adequate levels of reliabilities [63]. According to Hair, et al. [35], an item is considered to have an acceptable level of internal consistency if its corrected item-total correlation is equal to or greater than 0.33. All measures of ISDP complexity demonstrated adequate levels of corrected item-total correlation. Therefore, no item was eliminated based on the internal consistency criteria. ============================ Table 4 about here ============================ Confirmatory factor analysis Model-data fits of alternative models The measurement of ISDP complexity was specified as a second-order model with four first-order factors (Model 5). As discussed before, four alternative models were tested: (a) a null 25 model (Model 1), (b) a model with one first-order factor (Model 2), (c) a model with four uncorrelated first-order factors (Model 3), and (d) a model with four correlated first-order factors (Model 4). A model is considered to have good model-data fit if the p-value is above .05, the χ2 to degrees of freedom is smaller than 3, the GFI is above .90, the AGFI is above .80, the RMSEA is less than .08, the standardized RMR is less than .10, the CFI is above .90, and the NFI is above .90 [15, 19, 39, 45, 89]. As shown in Table 5, all model-fit indices of Model 4 indicate better fit than those of the first three competing models. The significance test results shown in Table 6 indicate that Model 4 was significantly better than the first three alternative models. Therefore, the results support the measurement model structure with four correlated first-order factors. ============================ Table 5 about here ============================ ============================ Table 6 about here ============================ Table 6 indicates that the difference between Model 4 and Model 5 was not significant. Table 5 suggests that most of the model-fit criteria of Model 5 (with a second-order model) were as good as those of Model 4 (with four correlated first-order factors). These results warrant further investigation of the existence of a second-order factor using the target coefficient (T) [55]. The T coefficient can be calculated using the following formula: T = χ2 (baseline model) / χ2 (alternative model) A high T coefficient implies that the second-order factor does not significantly increase χ2. Since the T coefficient between Model 4 and Model 5 is 0.98, we concluded that there existed a second-order factor and that the second-order model best represented the data in a more 26 parsimonious way. Therefore, Model 5 was chosen as the “best-fitted” measurement model for further validation. The second-order factor can be interpreted as an overall trait of ISDP complexity. Figure 4 shows the results of the parameter estimations of the second-order model. ============================ Figure 4 about here ============================ Unidimensionality and convergent validity Unidimensionality and convergent validity require that one single latent variable underlies a set of measures [1]. To test unidimensionality and convergent validity, we generated four first-order factor models with each corresponding to one component of ISDP complexity. The results shown in Table 7 suggest that all four latent variables demonstrated adequate levels of model fit. Overall, the results indicate that the measures of each of the four ISDP components satisfy the unidimensionality and convergent validity requirements. ============================ Table 7 about here ============================ Internal consistency reliability The composite reliability (ρc) which represents the proportion of measure variance attributable to the underlying latent variable was calculated to assess the reliability of the measure [87]. One of the advantages of this reliability index is that it is free from the restricted assumption of equal importance of all indicators on which the Cronbach α is based. Following Werts et al [87], Venkatraman [85], and Sethi and King [74], the composite reliability was calculated from the factor loadings of each indicator and error variances using the following formula: ρc=(Σ λi)2 Variance(A) / ((Σ λi)2 Variance(A) + Σ θδ) 27 Values of ρc in excess of .50 indicate that the variance captured by the measures is greater than that captured by error components, thus suggesting satisfactory levels of reliability [3]. The results in Table 8 show that the composite reliability estimates were .74 for SORG, .78 for SIT, .81 for DORG, and .87 for DIT, respectively, suggesting that all four latent variables have adequate levels of reliability. ============================ Table 8 about here ============================ Discriminant validity Discriminant validity assesses the degree to which measures of different components of ISDP complexity are unique from each other. The results of the pair-wise tests are shown in Table 9. The results suggest that all six pairs were statistically different; indicating that the four components of ISDP complexity demonstrated adequate levels of discriminant validity. ============================ Table 9 about here ============================ Analysis of factorial invariance Tests of factorial invariance examines if a measure operates equivalently across different sub-populations [15]. Analysis of factorial invariance is important for establishing the generalizability of a measurement. The value of a measurement model is greatly enhanced if the same factorial structure and properties can be replicated across various subpopulations [55]. ISDPs were defined in this study to include three types of system development: in-house new development, packaged software implementation, and major enhancement of existing software. As such, it is important to examine if the factorial structure and properties of the measure is invariant across the three types of ISDPs. 28 Tests of factorial invariance take a hierarchical approach. First, a baseline model (Model A) was established and tested. This baseline model was essentially the second-order model as specified in Figure 4. However, the difference was that the second-order model was established separately for each group. This baseline model did not have any invariance constraints across the three types of ISDPs. Once the baseline model was established, invariance of first-order factor loadings was tested (Model B). In this step, the second-order factor loadings were not constrained to be invariant. The rationale behind this approach was that tests of higher order invariance would make sense only when there was reasonable invariance among the first-order factors [55]. Model B was compared with Model A to examine if the first-order factor loadings were invariant. Difference in the χ2 between the two models was tested. However, χ2 tests can be so powerful that trivial difference may lead to significant χ2 values. Therefore, other criteria such as the ratio of χ2 to degrees of freedom and the target coefficient (T) should be also considered. The χ2 of the baseline model serves as a target for optimum fit. If the first-order factor loadings turned out to be invariant, a more restricted model with invariant first- and second-order factor loadings (Model C) was tested. Again, this model was compared with the baseline model to examine if the second-order factor loadings were invariant. If the χ2 tests and other criteria do not indicate invariant factorial structure, it suggests that the second-order factor loadings cannot be invariant because the first-order factor loadings are invariant. The factorial invariance analysis results shown in Table 10 suggest that the first-order factor loadings were invariant because the χ2 difference was not significant. In addition, the ratio of χ2 to degrees of freedom was reasonable and the target coefficient was very high, suggesting 29 good overall model-data fit. Similarly, the second-order factor loadings appeared to be invariant because the χ2 difference between Model C and the baseline model was not significant. The target coefficient was also very high, indicating good fit. In sum, we concluded that the factorial structure of the second-order measurement model of ISDP complexity was invariant across the three types of ISDPs. Therefore, the results provide the initial empirical evidence of the generalizability of the measurement model across the three types of ISDPs. ============================ Table 10 about here ============================ Nomological validity Nomological (or predictive) validity assesses if a construct measured by the new measures is associated with other constructs whose measures are known to be valid, as the theory would predict. In this study, we attempted to analyze the predictive validity of the ISDP complexity measures by testing a hypothesized positive relationship between ISDP complexity and project duration. Since our purpose was not testing theory, we provided only the necessary justification for the hypothesized relationship without considering other constructs. Our proposed positive relationship between ISDP complexity and project duration is justified by the argument that project complexity imposes more workload and thus causes longer project duration [21, 33, 59]. For example, Meyer and Utterback [59] found that technological complexity as measured by the number of technologies in the development effort was positively associated with absolute the development time. Table 11 shows the results of the path analyses of the impacts of (1) overall ISDP complexity and (2) the four components of ISDP complexity on project duration, respectively. The results indicate that all four factors of ISDP complexity as well as the overall ISDP 30 complexity positively affected project duration. Therefore, the prediction was supported by the data. We concluded that the measurement of ISDP complexity demonstrated adequate nomological validity in predicting project duration. ============================ Table 11 about here ============================ Discussions and Conclusions The results of both the exploratory and confirmatory data analyses suggest that the 20item measure of ISDP complexity developed in this research exhibited adequate levels of measurement properties. The exploratory factor analysis produced a factor structure as we hypothesized, providing an initial empirical support to our conceptualization of ISDP complexity as a four component construct. In addition, the confirmatory factor analysis results suggest that the hypothesized measurement model had adequate levels of goodness of fit. It also suggested the existence of a second-order factor, which can be interpreted as the overall ISDP complexity. The measures were shown to satisfy criteria related to unidimensionality, convergent validity, discriminant validity, internal consistency reliability, factorial invariance across three types of projects, and nomological validity. Contributions to theory development and methodology This research makes significant contributions to theoretical development. We believe that ISDP complexity will be an important construct in the vocabulary of IS researchers and practitioners for the following two reasons. First, an increasing portion of IS activities in organizations are organized around projects. Therefore, projects constitute an important context 31 and a unit of analysis for research. Second, the constantly changing information technology and business environments coupled with the growing needs for IT application integrations will cause the level of ISDP complexity to continue to increase. Therefore, managing complexity appears to be critical to IS success. Given this increasing significance of ISDP complexity, it is timely and important to develop a conceptual framework and a valid, reliable measure of the construct. Although there are other related constructs such as software complexity, general project complexity, and task complexity, they are not substitutes for ISDP complexity. Task complexity and project complexity are too general to tap into the unique context of ISDPs. Software complexity is too limited and narrow to assess the various aspects of ISDP complexity. The new measure developed in this research overcomes these limitations and covers a wide range of the domain of the ISDP complexity construct with enhanced specificity. In addition, by defining four distinct components of ISDP complexity, this research enables researchers to theorize the construct more precisely. As Baccarini [2] argues, since complexity is multi-dimensional, when referring to project complexity, it is important to state clearly the type of complexity being dealt with. The conceptual framework and the measure of ISDP complexity developed in this research will enable researchers to use these measures to build and test theories that explain the determinants and impacts of ISDP complexity. Depending on their study purposes, researchers can select either the second-order factor or the first-order factors of ISDP complexity as focal constructs to develop theories related to ISDP complexity. This research employed a combination of exploratory data analysis and confirmatory data analysis in developing and testing the measure. Such a research lifecycle consisting of both 32 exploratory and confirmatory methods ensures both the relevance and the rigor of the instrument development process, which in turn enhances the validity of the measurement. Practical implications The results of our study also have important practical implications. Although the importance of assessing and managing complexity of ISDPs has been widely recognized, organizations are not well equipped to cope with these challenges. As Kotter [50] suggests, managing structural and dynamic complexities has become a key responsibility of managers and executives. As such, this research provides a much needed language and measurement tool that managers can use to describe and communicate ISDP complexity. First, the empirically validated four-component framework of ISDP complexity serves as a useful language for defining and communicating ISDP complexity. Using this framework, project managers can clearly define the specific aspects and components of ISDP complexity that they must consider and manage. Second, the measures developed in this study can be used to assess and manage the complexity of ISDPs in the early planning stages and during implementation. Without such an assessment tool, it would be difficult for project managers to identify areas of concerns and take appropriate measures. It has been found that complexity influences the selection of project inputs including project organizational form, budget, manpower, expertise, and experience requirements of the project team [2, 47]. Therefore, being able to accurately assess ISDP complexity enables organizations to better plan, coordinate, and control their projects. In addition, a valid and reliable measurement tool would allow organizations to learn from past experiences and establish a knowledge base of organizational techniques that have been proven to be effective in dealing with different aspects of ISDP complexity. Used together, 33 the assessment tool and the knowledge base enable organizations to develop critical capabilities that are needed for planning and controlling their ISDPs. The second-order factor and the first-order factors (or the four components) can serve different purposes in practice. The second-order factor is useful for communicating the overall level of ISDP complexity with users and business unit managers. It is also useful for overall project planning in the early stages of project lifecycle. In contrast, the four first-factors (or the four components) of ISDP complexity can be used to facilitate detailed assessments and communications within the project team. They are useful for identifying specific problem areas, thus enable the managers to strategically manage and control the most important aspects or components of ISDP complexity during project implementations. Limitations of the study Some cautions should be taken when interpreting the study findings and applying the measure developed in this research. In testing the nomological or predictive validity of the measurement, the same respondent provided information about both the independent and the dependent variables, which might cause potential common source biases. Since the projects in the sample were all recently completed projects, the performance measures such as delivery time, cost, and functionality were known and thus might have been less subjective. Future research is needed to further test the nomological validity of the measurement using different sources for information about the independent and dependent variables. If independent sample sources had been used for the exploratory factor analysis and the confirmatory factor analysis, the value of cross-validation of the measure would have been even greater. However, split-halves from the same sample source also have advantages in that they 34 eliminate sampling errors resulted from different sample sources. In addition, the random assignment of the data cases to the two sub-samples minimizes potential biases that might be caused by the differences between the two sub-samples. Nevertheless, future research using different sample sources is needed to overcome the limitations caused by the use of the same sample source. Directions for future research This research represents the first step toward building theories that provide insights about the conceptualization and measurement of ISDP complexity. The framework and the measures developed in this study can help organizations better understand ISDP complexity and can provide the initial tools for assessing and managing the complexity of their ISDPs. Based on this study, future research may investigate the organizational determinants of ISDP complexity. Organizations can then minimize unnecessary complexity and effectively manage necessary complexity to enhance the success rate of their ISDPs, by creating effective strategies, methods, and coping mechanisms to control and manage those organizational factors that influence ISDP complexity. In addition, future research may investigate the patterns through which the four components of ISDP complexity affect such dependant variables as project success and organizational performance. Another promising future research direction is to conduct longitudinal studies of ISDP complexity. The levels and the impacts of ISDP complexity may vary between different stages of a project’s lifecycle. Understanding the dynamics of ISDP complexity can help managers cope with complexity at different points in time during their project implementation. 35 Finally, it would be important to examine project portfolio complexity. Organizations are most likely to run more than one ISDP at a time. Optimizing one project may create global suboptimization that hinders the performance of the overall project portfolio. Therefore, understanding complexity at the portfolio level enables IT organizations to achieve efficiency and effectiveness at the portfolio level in addition to individual project level. Our hope is that this research serves as a starting point for stimulating researchers to develop theories for understanding and managing ISDP complexity, and ultimately stopping the dollar drain of ISDP failures. 36 REFERENCES 1. Anderson, J.C., and Gerbing, D.W. Structural Equation Modeling in Practice: A Review and Recommened Two-Step Approach. Psychological Bulletin, 103, 3 (1988), 411-423. 2. Baccarini, D. The Concept of Project Complexity-A Review. International Journal of Project Management, 14, 4 (1996), 201-204. 3. Bagozzi, R.P. An Examination of the Validity of Two Models of Attitude. Multivariate Behavioral Research, 16, (1981), 323-359. 4. Baker, S. Where Danger Lurks: Spam, Complexity and Piracy Could Hinder Tech's Recovery. Business Week (August 25 2003), 114-118. 5. Balachandra, R., and Friar, J.H. Factors for Success in R&D Project and New Product Innovation: A Contextual Framework. IEEE Transactions on Engineering Management, 44, (1997), 276-287. 6. Banker, R.D., and Slaughter, S.A. The Moderating Effects of Structure on Volatility and Complexity in Software Enhancement. Information Systems Research, 11, 3 (2000), 219-240. 7. Barki, H.; Rivard, S.; and Talbot, J. Toward an Assessment of Software Development Risk. Journal of Management Information Systems, 10, 2 (1993), 203-225. 8. Barki, H.; Rivard, S.; and Talbot, J. An Integrative Contingency Model of Software Project Risk Management. Journal of Management Information Systems, 17, 4 (2001), 37-69. 9. Benamati, J., and Lederer, A.L. Coping With Rapid Changes in IT. Communication of the ACM, 44, 8 (2001), 83-88. 10. Boehm, B.W. Software Risk Management: Principles and Practices. IEEE Software, 8, 1 (1991), 32-41. 11. Boehm, B.W., and Ross, R. Theory-W Software Project Management: Principles and Examples. IEEE Transactions on Software Engineering, 15, 7 (1989), 902-916. 12. Briggs, R.O.; De Vreede, G.J.; Nunamaker Jr., J.F.; and Sprague Jr., R.H. Special Issue: Information Systems Success. Journal of Management Information Systems, 19, 4 (2003), 58. 13. Brooks, F., P., Jr. The Mythical Man-Month. Reading, MA: Addison-Wesley, 1995. 14. Byrd, T.A., and Turner, D.E. Measuring the Flexibility of Information Technology Infrastructure: Exploratory Analysis of a Construct. Journal of Management Information Systems, 17, 1 (2000), 167-208. 15. Byrne, B.M. Structural Equation Modeling with LISREL, PRELIS, and SIMPLIS: Basic Concepts, Applications, and Programming. Mahwah, N.J.: Lawrence Erlbaum Associates, 1998. 16. Campbell, D.J. Task Complexity: A Review and Analysis. Academy of Management Review, 13, 1 (1988), 40-52. 17. Caron, J.R. Business Reengineering at CIGNA Corporation: Experiences and Lessons Learned from the First Five Years. MIS Quarterly, 18, 3 (1994), 233-250. 18. Charette, R.N. Software Engineering Risk Analysis and Management. New York: Inter-text, 1989. 19. Chin, W.W., and Todd, P.A. On the Use, Usefulness, and Ease of Use of Structural Equation Modeling in MIS Research: A Note of Caution. MIS Quarterly, 19, 2 (1995), 237-246. 20. Churchill, G.A., Jr. A Paradigm for Developing Better Measures of Marketing Constructs. Journal of Marketing Research, 16, 1 (1979), 64-73. 37 21. Clark, K.B. Project Scope and Project Performance: The Effect of Parts Strategy and Supplier Involvement on Product Development. Management Science, 35, 10 (1989), 12471263. 22. Cleland, D.I., and King, W.R. Systems Analysis and Project Management. New York: McGraw-Hill, 1983. 23. Davis, F.D. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Quarterly, 13, 3 (1989), 319-339. 24. Dickson, G.W., and Wetherbe, J.C. The Management of Information Systems. New York, NY: McGraw-Hill, 1985. 25. Duncan, N.B. Capturing Flexibility of Information Technology Infrastructure: A Study of Resource Characteristics and Their Measure. Journal of Management Information Systems, 12, 2 (1995), 37-57. 26. Dvir, D.; Lipovetsky, S.; Shenhar, A.; and Tishler, A. In Search Of Project Classification: A Non-Universal Approach to Project Success Factors. Research Policy, 27, 915-935 (1998). 27. Eisenhardt, K.M., and Tabrizi, B.N. Accelerating Adaptive Processes: Product Innovation in the Global Computer Industry. Administrative Science Quarterly, 40, (1995), 84-110. 28. Ewusi-Mensah, K. Critical Issues in Abandoned Information Systems Development Projects. Communication of the ACM, 40, 9 (1997), 74-80. 29. Faraj, S., and Sproull, L. Coordinating Expertise in Software Development Teams. Management Science, 46, 12 (2000), 1554-1568. 30. Field, T. When Bad Thing Happen to Good Projects. CIO Magazine (Oct. 15 1997). 31. Garmus, D., and Herron, D. Function Point Analysis: Measurement Practices for Successful Software Projects: Addison-Wesley, 2001. 32. Gefen, D. It is Not Enough to be Responsive: The Role of Cooperative Intentions in MRP II Adoption. The DATA BASE for Advances in Information Systems, 31, 2 (2000), 65-79. 33. Griffin, A. The Effect of Project and Process Characteristics on Product Development Cycle Time. Journal of Marketing Research, 34, (1997), 24-35. 34. Haimes, Y.Y. Total Risk Management. Risk Analysis, 11, 2 (1991), 167-171. 35. Hair, J.F.; Anderson, R.E.; Tatham, R.L.; and Black, W.C. Multivariate Data Analysis with Readings. Sydney, Australia: Prentice Hall, 1995. 36. Halstead, M.H. Elements of Software Science. New York: Elsevier North-Holand, 1977. 37. Hartwick, J., and Barki, H. Communication as a Dimension of User Participation. IEEE Transactions on Professional Communication, 44, 1 (2001), 21-36. 38. Hopper, M.D. Complexity: The Weed that Could Choke IS. Computerworld, 30, 28 (July 8 1996), 37. 39. Hu, L.T., and Bentler, P.M. Evaluating Model Fit, In Hoyle, R.H. (ed.), Structural Equation Modeling: Concepts, Issues, and Applications, Thousand Oaks, CA: Sage Publications, 1995, pp. 76-99. 40. Huizing, A., and Koster, E. Balance in Business Reengineering: An Empirical Study of Fit and Performance. Journal of Management Information Systems, 14, 1 (1997), 93-118. 41. Ives, B., and Olson, M.H. User Involvement and MIS Success: A Review of Research. Management Science, 30, 5 (1984), 586-603. 42. Jesitus, J. Broken Promises? Foxmeyer's Project Was A Disaster. Was The Company Too Aggressive or Was It Misled? Industry Week (Nov. 3 1997), 31-37. 38 43. Jiang, J.J., and Klein, G. Information System Success as Impacted by Risks and Development Strategies. IEEE Transactions on Engineering Management, 48, 1 (2001), 4655. 44. Johnson, J. Chaos: the dollar drain of IT project failures. Application Development Trends, 2, 1 (1995), 41-47. 45. Jöreskog, K.G., and Sörbom, D. LISREL7: A Guide to the Program and Applications. Chicago, IL: SPSS Inc., 1989. 46. Jöreskog, K.G., and Sörbom, D. LISREL VIII User's Guide. Mooresville, IN: Scientific Software, Inc., 1993. 47. Kearney, J.K.; Sedlmeyer, R.L.; Thompson, W.B.; Gray, M.A.; and Adler, M.A. Software Complexity Measurement. Communication of the ACM, 29, 11 (1986), 1044-1050. 48. Keil, M.; Cule, P.E.; Lyytinen, K.; and Schmidt, R.C. A Framework for Identifying Software Project Risks. Communication of the ACM, 41, 11 (1998), 76-83. 49. Keil, M.; Tiwana, A.; and Bush, A. Reconciling user and project manager perceptions of IT project risk: a Delphi study. Information Systems Journal, 12, 2 (2002), 103-119. 50. Kotter, J.P. What Leaders Really Do. Harvard Business Review, 68, 3 (1990), 103-111. 51. Leveson, N.G. Software Engineering: Stretching the Limits of Complexity. Communication of the ACM, 40, 2 (1997), 129-131. 52. Lyytinen, K.; Mathiassen, L.; and Ropponen, J. Attention Shaping and Software Risk - A Categorical Analysis of Four Classical Risk Management Approaches. Information Systems Research, 9, 3 (1998), 233-255. 53. Markus, M.L. Power, Politics, and MIS Implementation. Communication of the ACM, 26, 6 (1983), 430-444. 54. Markus, M.L., and Keil, M. If We Build It, They Will Come: Design Information Systems that People Want to Use. Sloan Management Review (1994), 11-25. 55. Marsh, H.W., and Hocevar, D. Application of Confirmatory Factor Analysis to the Study of Self-Concept: First and Higher Order Factor Models and Their Invariance Across Groups. Psychological Bulletin, 97, 3 (1985), 562-582. 56. McCabe, T.J. A Complexity Measure. IEEE Transactions on Software Engineering, SE-2, 4 (1976), 308-320. 57. McFarlan, F.W. Portfolio Approach to Information Systems. Harvard Business Review (September-October 1981), 142-150. 58. Meyer, M.H., and Curley, K.F. An Applied Framework for Classifying the Complexity of Knowledge-Based Systems. MIS Quarterly, 15, 4 (1991), 455-472. 59. Meyer, M.H., and Utterback, J.M. Product Development Cycle Time and Commercial Success. IEEE Transactions on Engineering Management, 42, November (1995), 297-304. 60. Moore, G.C., and Benbasat, I. Development of an Instrument to Measure the Perceptions of Adopting an Information Technology Innovation. Information Systems Research, 2, 3 (1991), 192-222. 61.Murray, J.P. Reducing IT Project Complexity. Information Strategy: The Executive's Journal, 16, 3 (2000), 30-38. 62. Nilsen, K. Adding Real-time Capabilities to Java. Communications of the ACM, 41, 6 (1998), 49-56. 63. Nunnally, J.C. Psychometric Theory. New York: McGraw Hill, 1967. 64. Orlikowski, W.J., and Robey, D. Information Technology and the Structuring of Organizations. Information Systems Research, 2, 2 (1991), 143-169. 39 65. Rainer, R.K.; Snyder, C.A.; and Carr, H.H. Risk Analysis for Information Technology. Journal of Management Information Systems, 8, 1 (1991), 129-147. 66. Ravichandran, T., and Rai, A. Quality Management in Systems Development: An Organizational System Perspective. MIS Quarterly, 24, 3 (2000), 381-415. 67. Richardson, G.L., and Jackson, B.M. A Principles-Based Enterprise Architecture: Lessons from Texaco and Star Enterprise. MIS Quarterly, 14, 4 (1990), 385-403. 68. Ropponen, J., and Lyytinen, K. Components of Software Development Risk: How to Address Them? A Project Manager Survey. IEEE Transactions on Software Engineering, 26, 2 (2000), 98-112. 69. Schmidt, D.C. Middleware for Real-time and Embedded Systems. Communications of the ACM, 45, 6 (2002), 43-46. 70. Schmidt, R.; Lyytinen, K.; Keil, M.; and Cule, P. Identifying Software Project Risks: An International Delphi Study. Journal of Management Information Systems, 17, 4 (2001), 5-36. 71. Scott, J.E., and Vessey, I. Managing Risks in Enterprise Systems Implementations. Communication of the ACM, 45, 4 (2002), 74-81. 72. Scott, K. Battle Complexity to Add Profitability. InformationWeek 700 (September 14 1998), 18ER. 73. Senge, P.M. The Fifth Discipline. New York: NY: Doubleday, 1990. 74. Sethi, V., and King, W.R. Development of Measures to Assess the Extent to which an Information Technology Application Provides Competitive Advantage. Management Science, 40, 12 (1994), 1601-1627. 75.Shenhar, A.J. One Size Does Not Fit All Projects: Exploring Classical Contingency Domains. Management Science, 47, 3 (2001), 394-414. 76. Sherer, S.A. Measuring the Risk of Software Failure: A Financial Application, In Proceedings of the Tenth International Conference on Information Systems, Boston, 1989, pp. 237-245. 77. Souder, W.E., and Song, X.M. Contingency Product Design and Marketing Strategies Influencing New Product Success and Failure in U.S. and Japanese Electronics Firms. Journal of Product Innovation Management, 14, (1997), 21-34. 78. Standish Group. The Chaos Report, 1994. 79. Standish Group. The Chaos Report, 2001. 80. Swanson, E.B., and Beath, C.M. Reconstructing the Systems Development Organization. MIS Quarterly, 13, 3 (1989), 293-307. 81. Tait, P., and Vessey, I. The Effect of User Involvement on System Success: A Contingency Approach. MIS Quarterly, 12, 1 (1988), 91-108. 82. Thong, J.Y.L.; Yap, C.S.; and Raman, K.S. Top Management Support, External Expertise and Information Systems Implementation in Small Businesses. Information Systems Research, 7, 2 (1996), 248-266. 83. Truex, D.P.; Baskerville, R.; and Klein, H. Growing Systems in Emergent Organizations. Communication of the ACM, 42, 8 (1999), 117-123. 84. Turner, J.R., and Cochrane, R.A. Goals-and-Methods Matrix: Coping with Projects with ill Defined Goals and/or Methods of Achieving Them. International Journal of Project Management, 11, (1993), 93-102. 85. Venkatraman, N. Strategic Orientation of Business Enterprises: The Construct, Dimensionality, and Measurement. Management Science, 35, 8 (1989), 942-962. 40 86. Weill, P., and Broadbent, M. Leveraging the New Infrastructure: How Market Leaders Capitalize on Information Technology. Boston, MA: Harvard Business School Press, 1998. 87. Werts, C.E.; Linn, R.L.; and Joreskog, K.G. Intraclass Reliability Estimates: Testing Structural Assumptions. Educational and Psychological Measurement, 34, (1974), 25-33. 88. Weyuker, E.J. Evaluating Software Complexity Measures. IEEE Transactions on Engineering Management, 14, 9 (1988), 1357-1365. 89. Wheaton, B.; Muthen, B.; Alwin, D.; and Summers, G. Assessing Reliability and Stability in Panel Models, In Heise, D. (ed.), Sociological Methodology, San Francisco, CA: Jossey-Bass, 1977, pp. 84-136. 90. Wideman, R.M. Risk Management. Project Management Journal, 14, 4 (1986), 20-26. 91. Williams, T.M. The Need for New Paradigms for Complex Projects. International Journal of Project Management, 17, 5 (1999), 269-273. 92. Winklhofer, H. Information Systems Project Management During Organizational Changes. Engineering Management Journal, 14, 2 (2002), 33-37. 93. Wood, R.E. Task Complexity: Definition of the Construct. Organizational Behavior and Human Decision Processes, 37, 1 (1986), 60-82. 94. Wozniak, T.M. Significance vs. Capability: "Fit for Use" Project Controls, In American Association of Cost Engineers International, Dearborn, MI, 1993, pp. A.2.1-8. 41 Tables and Figures Table 1. Characteristics of the Overall Study Sample (n=541) Characteristics of projects Characteristics of organizations Industry Consulting Finance/Insurance Government Healthcare Manufacturing Retail Software Telecom/network Transportation Utility Other 6.3% 20.6% 9.2% 5.9% 13.7% 5.3% 9.7% 5.0% 4.0% 7.4% 12.9% Company annual sales Less than $100 million $100 million - $1 billion Over $1 billion 26.0% 31.2% 42.8% Number of employees Less than 1,000 1,000 – 10,000 Over 10,000 26.6% 40.5% 32.9% Type of project In-house new development Packaged software implementation Enhancement of existing software 38.1% 33.9% 28.0% Number of project members Less than 10 10 – 50 Over 50 25.0% 55.4% 19.6% Project budget Less than $100,000 $100,000 – 1 million Over $1 million 17.5% 41.8% 40.7% Project duration Less than 6 months 6 – 12 months Over 12 months 24.8% 40.9% 34.3% Table 2. Comparability Test of the Two Randomly-split Sub-samples (n1=270, n2=271) Subsample 1 2 16,041 13,604 0.932 0.352 Annual sales (dollars) 1 2 2.48 billion 2.63 billion -0.291 0.771 Number of project members 1 2 33 35 -0.501 0.616 Project budget (dollars) 1 2 2.19 million 2.06 million 0.301 0.764 Project duration (days) 1 2 363 351 0.517 0.605 Variable Number of employees t-test for Equality of Means t p Mean 42 Table 3. Exploratory Analysis – Factor Structure and Loadings (n1=270) Item Factor 1 (SIT) ISDPC16 .787 ISDPC 25 .729 ISDPC 30 .693 ISDPC 23 .618 ISDPC 19 .583 ISDPC 11 .566 ISDPC 4 .538 Factor 2 (DORG) ISDPC 12 .823 ISDPC 1 .766 ISDPC 26 .747 ISDPC 24 .692 ISDPC 14 .602 Factor 3 (DIT) ISDPC 7 .881 ISDPC 17 .873 ISDPC 22 .844 Factor 4 (SORG) ISDPC 27 .763 ISDPC 6 .701 ISDPC 29 .629 ISDPC 9 .623 ISDPC 21 .619 Eigenvalue 4.421 2.761 2.084 1.772 % of variance 22.106 13.806 10.422 8.860 Note: 1. Items in the questionnaire were randomly ordered to avoid biases. The item labels in this table reflect the order in which they appeared in the questionnaire. 2. Factor loadings less than 0.30 are not shown. The results indicate that there were no items with cross-factor loadings greater than 0.30. 43 Table 4. Exploratory Analysis - Reliability (n1=270) Factor Item SORG SIT DORG DIT Corrected itemtotal correlation Cronbach’s α ISDPC 29 .445 0.71 ISDPC 6 .458 ISDPC 21 .436 ISDPC 27 .555 ISDPC 9 .424 ISDPC 19 .442 ISDPC 11 .425 ISDPC 16 .640 ISDPC 23 .486 ISDPC 4 .374 ISDPC 25 .588 ISDPC 30 .574 ISDPC 12 .667 ISDPC 1 .642 ISDPC 26 .576 ISDPC 14 .496 ISDPC 24 .572 ISDPC 22 ISDPC 7 ISDPC 17 .728 .846 .833 0.78 0.81 0.90 Table 5. Confirmatory Analysis - Model Fits of Alternative Models (n2 =271) Criteria Model 1 Threshold Null model Model 2 One-factor Model 3 Four first-order factors (Uncorrelated) Model 4 Four first-order factors (Correlated) Model 5 Second-order model with four first-order factors χ2 2653.01 1802.06 395.39 335.45 341.31 d.f. 190 170 170 164 166 χ2 / d.f. (<5.0) 13.96 10.60 2.33 2.05 2.06 GFI (>.90) 0.50 0.60 0.87 0.89 0.89 AGFI (>.80) 0.45 0.51 0.84 0.86 0.86 RMSEA (<.08) 0.219 0.189 0.070 0.062 0.063 RMR (<.10) 0.22 0.16 0.11 0.067 0.071 CFI (>.90) 0.00 0.34 0.87 0.90 0.90 NFI (>.90) 0.00 0.31 0.80 0.82 0.82 44 Table 6. Confirmatory Analysis - Differences between Alternative Models (n2 =271) Model 2 & Model 1 Model 3 & Model 2 Model 4 & Model 3 Model 5 & Model 4 Difference in χ2 850.95 1406.67 59.94 5.86 Difference in d.f. 20 0 6 2 Significance level p <.01 p <.01 p <.01 n.s. Table 7. Confirmatory Analysis - Unidimensionality/Convergent Validity (n2 =271) Factor No. of indicators χ2 d.f χ2/d.f. GFI AGFI RMR CFI SORG 5 16.34 5 3.27 0.98 0.93 0.040 0.95 SIT 7 82.84 14 11.83 0.92 0.84 0.065 0.88 DORG 5 36.41 5 5.28 0.95 0.85 0.054 0.92 DIT a 3 33.16 19 1.75 0.97 0.94 0.041 0.98 Note: (a) This model is saturated because the number of indicators is 3. Therefore, the fit indexes are not available. Fit indexes for this factor were produced from a two-factor model including DIT and SORG. Table 8. Confirmatory Analysis - Composite Reliability ρc (n2 =271) Factor No. of indicators ρc SORG 5 0.736 SIT 7 0.784 DORG 5 0.812 DIT 3 2 0.869 2 Note: ρc=(Σ λi) Variance(A) / ((Σ λi) Variance(A) + Σ θδ) 45 Table 9. Confirmatory Analysis - Discriminant Validity of Four First-Order Factors (n2 =271) χ2 Statistic ML Estimate Ø t-value SORG - SIT 0.02 0.20 183.73 (54) 148.22 (53) 35.51** SORG - DORG 0.11 1.59 129.11 (35) 95.25 (34) 33.86** SORG - DIT 0.37 3.67** 49.74 (20) 33.16 (19) 16.58** SIT - DORG 0.20 2.47* 190.83 (54) 159.63 (53) 31.20** SIT - DIT 0.22 2.84** 135.81 (35) 115.43 (34) 20.38** DORG - DIT 0.23 3.30** 91.35 (20) 65.34 (19) 26.01** Construct Pair Note: * -- p<.05, Constrained model (df) Unconstrained Difference model (df) ** -- p<.01 Table 10. Confirmatory Analysis - Invariance of the Second-Order Model across Three Types of ISDP (n =511) Model description χ2 (d.f.) χ2 / d.f. χ2 difference with Model A T Model A: Baseline model with no invariance constraints across three types of ISDPs 914.20 (498) 1.84 – 1.00 Model B: Model A with invariant items loadings to the first-order factors 950.01 (530) 1.79 35.81 (32), n.s. 0.96 Model C: Model B with invariant structural coefficients between first- and second-order factor 950.01 (538) 1.77 35.81 (40), n.s. 0.96 46 Table 11. Nomological Validity – Regression Analysis Results (n = 467) Adjusted R2 Relationship Beta 0.128 Model I Project duration = Overall ISDP complexity 0.361** 0.126 Model II Project duration = Structural organizational complexity + 0.155** Structural IT complexity + 0.156** Dynamic organizational complexity + 0.122** Dynamic IT complexity 0.142** Note: * -- p<.05, ** -- p <.01 Table B1. Results of the Sorting Procedures ACTUAL CATEGORY TARGET CATEGORY SORG SORG 33 SIT 8 DORG 3 DIT Total item placement: 120 SIT DORG DIT 36 17 12 Hits: 98 47 Unclear 11 Total Target % 44 75 44 82 20 85 12 100 Overall Hit Ratio: 82% Table C1. Covariance matrix of the (total) sample ISDPC 19 ISDPC 14 ISDPC 29 ISDPC 6 ISDPC 11 ISDPC 16 ISDPC 23 ISDPC 21 ISDPC 24 5.293 ISDPC 4 ISDPC 25 ISDPC 27 ISDPC 9 ISDPC 30 ISDPC 12 ISDPC 1 ISDPC 26 ISDPC 22 ISDPC 7 ISDPC 17 0.481 2.969 0.582 0.267 3.179 0.531 0.359 1.065 4.561 1.063 0.594 0.224 0.000 4.184 1.472 0.560 0.124 0.164 1.288 0.994 0.561 0.047 0.220 1.187 1.240 3.117 0.198 0.021 1.014 1.357 -0.012 -0.312 0.010 3.639 0.966 1.577 0.605 0.380 0.422 0.522 0.597 0.406 4.471 0.710 0.505 -0.342 0.058 0.655 0.833 1.179 -0.195 0.548 1.492 0.587 0.275 0.217 1.108 2.605 1.032 -0.120 0.627 0.863 3.800 0.571 0.093 1.456 1.407 0.111 0.039 0.120 1.280 0.296 -0.257 0.197 3.184 0.125 -0.042 0.980 0.818 0.000 -0.134 -0.062 1.330 0.331 -0.551 -0.090 1.063 3.027 1.617 0.381 0.084 0.253 1.192 1.636 1.335 0.009 0.597 0.796 1.681 0.106 -0.125 3.377 0.228 1.442 0.362 0.314 0.349 0.222 0.603 0.241 1.651 0.171 0.369 0.187 0.185 0.350 3.567 0.204 0.743 0.262 0.262 0.060 0.154 0.178 0.152 1.942 0.175 0.461 0.264 0.223 0.343 1.921 2.832 0.252 1.011 0.337 0.532 0.304 0.136 0.558 0.420 1.450 0.119 0.290 0.501 0.302 0.458 2.096 1.543 3.212 0.692 0.114 0.409 0.128 0.473 0.414 0.266 0.225 0.571 -0.055 0.580 0.214 0.210 0.594 0.639 0.658 0.688 2.690 0.796 0.052 0.711 0.480 0.444 0.372 0.316 0.578 0.695 -0.043 0.664 0.644 0.561 0.775 0.570 0.737 0.582 2.009 3.219 0.768 0.109 0.694 0.605 0.344 0.430 0.450 0.499 0.616 -0.057 0.616 0.582 0.463 0.767 0.658 0.705 0.671 1.892 2.635 3.728 2.478 3.259 ISDPC19 ISDPC14 ISDPC29 ISDPC6 ISDPC11 ISDPC16 ISDPC23 ISDPC21 ISDPC24 ISDPC4 ISDPC25 ISDPC27 ISDPC9 ISDPC30 ISDPC12 ISDPC1 ISDPC26 ISDPC22 ISDPC7 ISDPC17 48 Organizational Structural Organizational Complexity (SORG) Dynamic Organizational Complexity (DORG) Technological Structural IT Complexity (SIT) Dynamic IT Complexity (DIT) Structural Dynamic Figure 1. A Conceptual Framework of ISDP Complexity 49 Phase 1 – Conceptual development and initial item generation Literature review Field interviews (12) - Frameworks - New measures - Existing measures - Insights Focus groups (45) - Nominal group process - New measures Sorting procedure Qualitative assessment of construct validity (4) Phase 2 – Conceptual refinement and item modification Pilot test 1 Assessment of content validity (7) Pilot test 2 Refinement and test of online survey (15) Phase 3 – Online survey of select PMI-ISSIG members (541 valid responses) Survey data collection Phase 4 – Data analysis and measurement validation 4.1 Data screening and descriptive analysis Assessment of non-response bias Screening for outliers 4.2 Sample split Split the sample into two random sub-samples (n1=270 and n2=271) Test comparability of the sub-samples (t-test on key sample characteristics) 4.3. Exploratory analysis Exploratory factor analysis - Convergent validity - Discriminant validity Reliability (Cronbach’s alpha) MTMM correlation analysis 4.4. Confirmatory analysis Test model-data fits of alternative models Parameter estimates of selected “best” model Convergent validity/unidimensionality Discriminant validity Reliability 4.5. Factorial invariance analysis Test invariance/equivalence of measurement model (factorial structure and loadings) across three types of ISDPs 4.6. Test of nomological validity Test relationships between ISDP size, complexity and performance measures Note: Numbers in parentheses indicate number of project managers involved. Figure 2. A Four-Phase Process of Measure Development and Validation 50 Item 1 Factor1 Item 1 Item 2 Factor2 Item 2 Item 3 Factor3 Item 3 Item 4 Factor4 Item 4 Item 5 Factor5 Item 5 Item 6 Factor6 Item 6 …. …. (b) Model 2: One First-Order Factor Model (a) Model 1: The Null Model Item 1 Item 1 Item 2 Item 2 SIT Item 6 Item 7 Item 14 Item 13 DORG Item 14 Item 19 DORG ... ... Item 18 SIT ... ... Item 13 SORG ... SORG ... Item 6 Item 7 ISDP Complexity Item 18 DIT Item 19 DIT ... ... (c) Model 3: Four First-Order Factor Model (uncorrelated) (d) Model 4: Four First-Order Factor Model (correlated) Item 1 Item 2 SORG ... SIT ISDP Complexity ... Item 6 Item 7 Item 13 Item 14 DORG ... Item 18 Item 19 DIT ... (e) Model 5: The Second-Order Factor Model Figure 3. Alternative Models Tested in the Confirmatory Analysis 51 0.63 0.80 0.58 0.52 0.67 0.80 0.83 0.39 0.70 0.79 0.39 0.59 0.77 0.63 0.32 0.47 0.45 0.51 0.12 0.30 ISDPC29 0.61 ISDPC6 0.45 0.65 0.70 0.57 ISDPC21 ISDPC27 SORG ISDPC9 ISDPC19 ISDPC11 ISDPC16 ISDPC23 ISDPC4 ISDPC25 0.37 0.44 0.41 0.78 0.55 0.46 0.78 0.64 SIT 0.30 ISDPC30 ISDP Complexity ISDPC14 0.48 ISDPC24 0.61 0.82 0.73 0.74 DORG 0.70 0.94 0.84 DIT ISDPC12 ISDPC1 0.37 0.76 ISDPC26 ISDPC22 ISDPC7 ISDPC17 Figure 4. Parameter Estimates of the Second-Order Model (Model 5, n2 =271) 52 Appendix A. Measures and their Reference Sources ISDP Components SORG Item Item description (reference) ISDPC6 ISDPC9 ISDPC21 The project manager did not have direct control over project resources Business users provided insufficient support and involvement [43, 48, 70] There was no sufficient commitment/support from the top management [7, 43, 48, 52, 70] There was no sufficient/appropriate staffing for the project [10, 48, 52, 68, 70] The project personnel did not have required knowledge/skills [7, 8, 43, 48, 52, 68, 70, 71] ISDPC27 ISDPC29 SIT DORG ISDPC4 ISDPC11 ISDPC16 ISDPC23 ISDPC25 ISDPC30 ISDPC19 The project team was cross-functional [7, 8] The system involved real-time data processing [31] The project involved multiple software environments [58] The project involved coordinating multiple user units [8, 43, 70] The project involved multiple technology platforms [58] The project involved a lot of integration with other systems [7, 8, 58] The project involved multiple external contractors and vendors [7, 43, 70, 94] ISDPC1 ISDPC12 ISDPC14 ISDPC26 The end-users' organizational structure changed rapidly The end-users' business processes changed rapidly Implementing the project caused changes in the users' business processes [8, 71] Implementing the project caused changes in the users' organizational structure [8, 71] The end-users' information needs changed rapidly [10, 68, 70] ISDPC7 ISDPC17 ISDPC22 IT architecture that the project depended on changed rapidly [70] IT infrastructure that the project depended on changed rapidly [71] Software development tools that the project depended on changed rapidly ISDPC24 DIT Note: The items without references were obtained from field interviews and focus group discussions with IS project managers 53 Appendix B. Sorting Procedure and Results A sorting procedure was used to qualitatively assess the face validity and the construct validity of the initial 30 items that were generated in Phase 1 of the research process. Four IS researchers with an average of eight years of IS work experience participated in the sorting procedure. Each item in the initial pool was printed on a 3¯5-inch index card. In the sorting procedure, each judge was asked to carefully read the card and place it in one of the four components of ISDP complexity. An additional category, “too ambiguous/unclear” was included so that the judges could put a card into that category if they felt the card did not seem to belong to any of the four pre-defined categories. Prior to actually sorting the cards, the judges read a standard set of instructions. To make sure that the judges understood the sorting procedure, they did a sorting exercise with the well-known 12-item ease of use and usefulness instrument [23]. All judges completed this sorting exercise successfully, indicating they had clear understanding of the sorting procedure and were able to do it appropriately. Each judge then individually sorted the ISDP complexity item cards. After completing the sorting procedure, they explained why they sorted cards into the “too ambiguous/unclear” category, if any. Following Moore and Benbasat [60], we calculated the overall item placement ratio. This ratio represents how well the judges were able to sort the items into the target constructs. In total, the judges classified 98 items into the target categories and 22 items into other categories (shown in Table B1), resulting in an overall placement ratio of 82%. This indicates that the items were, in general, being placed as they were intended to be. Four items in the SORG category were commonly placed in the “ambiguous/unclear” category and were dropped. As a result, 26 items remained after the sorting procedure. 54 ============================ Table B1 about here ============================ Appendix C. Covariance matrix of the (total) sample ============================ Table C1 about here ============================ 55
© Copyright 2026 Paperzz