Journal of Operations Management 32 (2014) 313–336 Contents lists available at ScienceDirect Journal of Operations Management journal homepage: www.elsevier.com/locate/jom The effect of performance measurement systems on firm performance: A cross-sectional and a longitudinal study Xenophon Koufteros a,∗ , Anto (John) Verghese a , Lorenzo Lucianetti b a Business Administration Department of Information & Operations Management Mays Business School, Texas A & M University Department of Information & Operations Management, College Station, TX 77843-4217, United States b Department of Management and Business Administration University of Chieti and Pescara, Viale Pindaro 42 - 65127, Pescara, Italy a r t i c l e i n f o Article history: Received 11 December 2013 Received in revised form 13 June 2014 Accepted 17 June 2014 Keywords: Performance Measurement Systems Operations management Accounting Financial performance Empirical Panel data a b s t r a c t Performance measurement (PM) systems have been popularized over the last 20 years and the operations management literature is replete with discussion of metrics and measurement systems. Yet, a comprehensive nomological network relating types of PM system uses to organizational capabilities and performance is lacking. Furthermore, there is scant empirical evidence attesting to the explanatory efficacy of PM systems as it relates to organizational performance. We view PM system uses through the lenses of the Resource Orchestration Theory (ROT) and explore specific relationships of underlying variables by relying on the Organizational Information Processing Theory (OIPT). Resting on the extant literature, we identify two types of uses which include Diagnostic Use (the review of critical performance variables in order to maintain, alter, or justify patterns in an organizational activity) and interactive use (a forward-looking activity exemplified by active and frequent involvement of top management envisioning new ways to orchestrate organizational resources for competitive advantage) and relate them along with their interaction (i.e., dynamic tension) to organizational capabilities. We further link capabilities to target performance, which subsequently impacts organizational performance (operationalized through both perceptual and objective financial performance measures). The nomological network is tested via a cross sectional study (386 Italian firms) while the efficacy of PM systems to explain organizational performance is examined by using longitudinal panel data approaches over a 10 year period. There is sufficient evidence to suggest that the use of PM systems leads to improved capabilities, which then impact performance. Contrary to the extant literature, however, we discovered that Diagnostic Use appears to be the most constructive explanatory variable for capabilities. On the other hand, in light of a longitudinal study, we also uncovered that Diagnostic Use experienced depreciating returns as far as objective financial measures are concerned. Also, when high levels of Diagnostic Use were coupled with low levels of Interactive Use, they produced the lowest levels of organizational capabilities. Conversely, high levels of both types of PM system use generated extraordinary high levels of capabilities. There is sufficient evidence to suggest that organizations cannot rely merely on Diagnostic Use of PM systems. We also learned that the effects of PM systems (measured via adaptation) fade unless high learning rates are applied. We offer detailed recommendations for future research which have theoretical as well as empirical implications. © 2014 Elsevier B.V. All rights reserved. 1. Introduction The operations management literature has been a strong proponent of metrics and respective PM systems for quite some time. For instance, Bititci et al. (1997) offer a developmental guide to ∗ Corresponding author. Tel.: +1 979 845 2254/fax: +1 979 845 5653. E-mail addresses: [email protected] (X. Koufteros), [email protected] (A.J. Verghese), [email protected] (L. Lucianetti). http://dx.doi.org/10.1016/j.jom.2014.06.003 0272-6963/© 2014 Elsevier B.V. All rights reserved. construct integrated PM systems and Gunasekaran et al. (2004) proposed a framework for supply chain performance measurement. Melnyk et al. (2004) discuss metrics and performance measurement while Neely (1999; 2005) furnished a treatise on the evolution of performance measurement research in operations management. Later on, Gunasekaran and Kobu (2007) provided a literature review of performance measures and metrics in logistics and supply chain management. While the topic is popular, what is vividly missing from the literature is a judicious examination of how companies actually use PM systems to orchestrate their responses to 314 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 organizational challenges and whether such uses do in fact enhance operational, strategic, and external stakeholder related capabilities and performance over time. PM systems are integral to resource orchestration processes and over the last three decades many organizations have invested enormous amounts of capital, time, and effort developing and implementing such systems. Undoubtedly, one of the most popular paradigms is the Balanced Scorecard, first documented and articulated by Kaplan and Norton (1992). There are nevertheless numerous other measurement frameworks (Bititci and Turner, 2000) that have been proposed and implemented–e.g. the Performance Measurement Matrix (Keegan et al., 1989), the Result and Determinants Model (Fitzgerald et al., 1991), the Performance Pyramid (Lynch & Cross, 1992), and the Performance Prism (Neely et al., 2002). These formal performance systems are utilized as mechanisms that enable organizations to orchestrate their resources more effectively. The Resource-Based Theory (RBT) has long argued that possessing valuable, rare, inimitable, and non-substitutable resources is vital to firm sustained advantage (Hitt et al., 2011; Wowak et al., 2013). But as Hansen et al. (2004, p. 1280) state “what a firm does with its resources is at least as important as which resources it possesses.” Sirmon et al. (2011) similarly note that while possession of resources is essential, the ability of a firm to “orchestrate” its resources is more fundamental as the firm bids to prosecute its strategic objectives. Orchestration of resources however is also subject to top management who mobilizes the vision to direct and use firm resources to achieve objectives (Chirico et al., 2011; Crook et al., 2008). Resource Orchestration Theory (ROT) is an emerging theoretical stream of work which rests on the conceptual work of Sirmon et al. (2007) and Helfat et al. (2007). Hitt et al. (2011) argue that “Resources orchestration is concerned with the actions leaders take to facilitate efforts to effectively manage the firm’s resources” (p. 64). Such actions include for instance the structuring of the firm’s resource portfolio, bundling resources into capabilities, and leveraging the capabilities to create value to customers (Hitt et al., 2011; Sirmon and Hitt, 2003; Sirmon et al., 2007). According to Hitt et al. (2011) and Holcomb et al. (2009), while each action and its particular nuances are vital, the synchronization of actions can contribute positively towards performance. To manage each action and to synchronize the orchestration of resources, leaders rely on mechanisms such as PM systems which yield information regarding the functioning of their resource portfolio and bundle of capabilities (Hitt et al., 2011). Such information is critical for leaders because it enables them to make crucial adjustments to their resources and mobilize requisite resources as conditions change. Melnyk et al. (2004) recognized the orchestrating role of PM systems in operations management and assert that the “performance measurement system is ultimately responsible for maintaining alignment and coordination” (p. 213). Operations management leaders, for instance, need information regarding inventory performance to decide whether additional space to house inventory is necessary in order to pursue a new “same-day delivery” strategy that demands high service levels, such as the expansive Amazon.com Local Express Delivery strategy. Via the diagnostic attributes of a PM system, managers can focus attention on issues of strategic significance, monitor performance, and detect whether the desired service level can be achieved given the current level and blend of resources. In addition, the active and personal engagement of the leadership with performance measurement processes can serve as a catalyst in orchestrating the acquisition and bundling of essential resources and capabilities to meet delivery targets. Melnyk et al. (2004) highlight that metrics and respective PM systems serve “as essential links between strategy, execution, and ultimate value creation” (p. 209). A PM system can also be characterized as a management control system that incorporates a structured framework specifying key financial and non-financial performance metrics. From a theoretical point of view, a PM system can be described as an ambidextrous system because it incorporates both mechanistic and organic elements. As an orchestration mechanism, the organization can use the PM system to control organizational behavior (alike to a mechanistic use) but on the other hand it can use it to promote organizational innovation and strategic renewal (resembling an organic use). The literature however tends to focus more on the “mechanistic” use of PM systems while the more “organic” use is in general neglected. The mechanistic use is coined as the diagnostic use in the literature and it is primarily responsible for furnishing information. From an organizational information processing theory (OIPT) perspective (Galbraith, 1974), diagnostic use is liable to reduce uncertainty. On the other hand, the organic use can be described as interactive in nature and is deployed by top management to enact debate and reduce equivocality. Simons (1995) acknowledges the complementary nature of the two systems, but only few studies explicitly test relationships between types of uses of PM systems (e.g., Widener, 2007) or specify and account for their interactions (e.g., Henri, 2006a). However, these authors operationalize organizational performance only via perceptual measures rather than objective and longitudinal financial performance measures. Underscoring the importance of PM systems, Kaplan and Norton (2008) made a rather revealing statement suggesting that “We believe that if you don’t measure progress toward an objective, you cannot manage and improve it” (p. 7). From an ROT perspective, firms that deploy their PM systems should be capable of shaping capabilities to meet or exceed target performance. Highlighting and motivating interest in adopting a PM system is, therefore, the claim that organizations with a PM system outperform their counterparts without a PM system (Davis & Albright, 2004; Crabtree & DeBusk, 2008). Unfortunately, this claim is debatable in part because there is only a handful of published empirical research studies investigating the claim (e.g., Ahn, 2001; Chenhall & Langfield-Smith, 1998; Henri, 2006a; Hoque & James, 2000; Ittner & Larcker, 1998; Chenhall, 2005; Widener, 2007) and in part because the extant empirical literature has reported mixed results regarding the effects of PM system usage on organizational performance (Chenhall & Langfield-Smith, 1998; Ittner et al., 2003a). Henri (2006a) surmises that prior work examined the role of PM systems toward strategy implementation and strategy formulation, but concedes there is scant empirical evidence attesting to the professed virtues of such systems. The operations management literature is especially devoid of studies in this respect and much of what is currently published can be ascribed as contributions from the accounting discipline. Henri (2006a) submits that the specific relationship between PM systems and strategy is ambiguous and at times contradictory and attributes such results to differences in definitions and operationalizations of the variables. Pavlov and Bourne (2011) add that the mechanism relating a PM system to performance is poorly understood. Henri (2006a) notes that prior research has specified and tested direct links between PM system usage and performance but the effects may actually be reflected instead by the capabilities that PM systems incite as orchestration mechanisms. Pavlov and Bourne (2011) argue that it has not been demonstrated exactly how PM systems are linked to performance, thus leaving the gap between PM systems and performance still unresolved. As Pavlov and Bourne state, the power of a PM system can be seen as significant and yet somewhat opaque. In other words, there is still a “black box.” Overall, there are three gaps in the literature which merit investigation. The first gap relates to the specific types of uses of PM systems: For what purposes do organizations use PM systems? The second and related gap pertains to the modality by which different types of uses of PM systems impact performance: How do specific X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 types of uses of PM systems actually affect performance? Stated otherwise, there is a lack of a nomological network that relates specific types of uses of PM systems to performance. Exploring the effectiveness of each type of use is important because organizations can direct resources and effort on specific types of uses in order to maximize returns. The third gap concerns the efficacy of PM systems to explain variation in performance: Does the implementation of PM systems benefit the organization financially over time? Both the operations management and the accounting literatures are lacking of empirical studies that demonstrate the financial impact of PM system implementations from a longitudinal perspective. Davis and Albright (2004) examined the impact of PM systems on performance (i.e., key financial measures pertinent only to financial institutions) but the study was constrained to a single firm (i.e., a bank) and several of its branches, and assessed change in performance only across two periods. This manuscript tackles the first two gaps in the literature by identifying salient types of uses and by constructing a nomological network of constructs. Then, based on literature (e.g., Henri, 2006a) that posits that the effects of PM systems on performance are communicated via the capabilities they engender, it examines how each type of use of a PM system impacts organizational capabilities across three dimensions which include operational capability, strategic management capability, and external stakeholder relations capability. It articulates simple effects as well as interactions amongst types of PM system usage, while controlling for an extensive list of demographics and tangible and intangible resources. In essence, we are able to disentangle the orchestrating role of PM systems from the mere possession of resources. Furthermore, the relationship between organizational capabilities and target performance is tested. Finally, the link between target performance and financial performance is tested utilizing both subjective and objective performance measures. To tackle the third gap in the literature, the impact on four financial ratios is studied over time (before and after implementation of a respective PM system) using several methods. Analyses rely on primary data and secondary data over a ten year time horizon. The longitudinal effects of PM system usage are assessed using panel data analysis approaches and Stata12. 2. Literature review & theory development A review of the operations and supply chain management literature suggests that there is no shortage of discussion regarding metrics and PM systems (e.g., Neely, 1999; Bourne et al., 2000; Bullinger et al., 2002; Melnyk et al., 2004; Neely, 2005; Hult et al., 2008; Chia et al., 2009; Lehtinen & Ahola, 2010; Taylor & Taylor, 2013; Garengo & Sharma, 2014). Neither there is scarcity of frameworks for measuring strategic, tactical, and operational performance at the plant level or in a supply chain (Bititci et al., 1997; Bititci & Turner, 2000; Neely et al., 2000; Gunasekaran et al., 2001; Gunasekaran et al., 2004; Gunasekaran and Kobu, 2007; Estampe et al., 2013). The implied assumption is made however that PM systems are by their very nature conducive towards better performance because they can be used to benchmark operations against targets and abet the organization orchestrate resources to enhance future performance. Metrics and performance measurement are seen as critical elements in translating an organization’s mission, or strategy, into reality and for maintaining alignment and coordination (Melnyk et al., 2004). Thus, PM systems are typically ascribed with positive evaluations. The majority of studies that have surfaced in the literature assess PM system effectiveness in relation to overall organizational performance (Braam & Nijssen, 2004; Crabtree & DeBusk, 2008; Davis & Albright, 2004; Ittner, Larcker, & Randall, 2003b), thereby assuming a direct association between the PM system and 315 performance. In this respect, Stede et al. (2006) found that regardless of strategy, organizations with more extensive PM systems, especially those that included objective and subjective nonfinancial measures, have better overall performance. However, the literature suggests that the findings are inconsistent (Bourne et al., 2013; de Leeuw & van den Berg, 2011; Hamilton & Chervany, 1981). At the organizational level, Henri (2006a) and Widener (2007) advocate that a more in-depth understanding of the engendering role of PM systems on organizational capabilities may help to resolve some of the ambiguous findings. Recently, Grafton et al. (2010) explored the process by which PM systems impact the way managers orchestrate their resources via mobilization to achieve competitive advantage and contribute to performance outcomes. Relying on RBT, Grafton et al. (2010) argue that the use of performance measurement information for feedback and feed-forward control influences the extent to which an organization is able to exploit and identify its strategic capabilities, respectively. Modelling capabilities as a mediating variable would allow researchers to explore the processes by which PM systems enhance organizational outcomes. Also, the extant literature describes several PM system types of uses (e.g., coordination, monitoring, diagnosis, problem solving, legitimization, scorekeeping, focusing attention). Simons (1994) work describing types of uses or purposes of PM systems has been distinctively influential in the literature. Simons (1994) effectively covers the domain of PM system types or styles of uses (Bisbe & Otley, 2004), yet he encapsulates them efficiently via only two categories (i.e., diagnostic and interactive uses), which rely on respective diagnostic systems and interactive systems. According to Simons (1994), diagnostic systems are “formal feedback systems used to monitor organizational outcomes and correct deviations from pre-set standards of performance” (p. 170). Diagnostic systems are in essence information systems that reduce uncertainty and thus aid strategy implementation (Bisbe & Otley, 2004). Interactive systems are used “by top managers to regularly and personally involve themselves in the decision activities of subordinates” (Simons, 1994, p. 171). Bisbe and Otley (2004), resting on the work of Simons (1994), suggest that interactive systems help organizations adapt to changes in the environment by orchestrating organizational responses. The modality by which PM systems enable resource orchestration can be explained via the theoretical lenses of the organizational information processing theory (OIPT) (Daft & Lengel, 1986; Galbraith, 1974). As Melnyk et al. (2004) state, “An information processing perspective offers yet another potentially rewarding way to look at metrics” (p. 214). Choi et al. (2013) state that PM systems “function as a framework that organizes the firm’s information environment around its strategy” (p. 106). According to Grosswiele et al. (2012), “Performance measurement aims to provide decision makers with information that enables them to take effective actions and evaluate whether a company is progressing in line with its strategy” (p. 1017). Bisbe and Malagueño (2012) suggest that PM systems “. . .support the decision-making processes of an organization by gathering, processing, and analyzing quantified information about its performance, and presenting it in the form of a succinct overview” (p. 297). Organizations however do not always have the requisite information and thus they may face uncertainty and, sometimes, even when they do possess information it may have multiple and conflicting interpretations which creates equivocality (Daft & Lengel, 1986). In order to tackle facets of both uncertainty and equivocality, organizations turned to PM systems which can furnish relevant information and engender debate and enactment of organizational action. This study aims to examine how different types of uses of PM system can fuel organizational capabilities leading to improved organizational performance. Specifically, this research focuses on 316 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Diagnosc Use Subjecve Financial Performance Strategic Management Capability ROTA Dynamic Tension (Diagnosc X Interacve) Operaonal Capability Target Performance ROA ROE Interacve Use External Stakeholder Relaons Capability EBIT Fig. 1. Research Framework Note: Effects on ultimate dependent variables are controlled for industry, firm size, geographic diversification, R&D expenditures, Industrial Patents & Intellectual Property Rights, Plant & Machinery, and Industrial & Commercial Equipment. These effects are displayed in Table 4. the orchestrating roles of PM systems: “diagnostic use” and “interactive use.” According to Henri (2006a), “These two types of use work simultaneously but for different purposes. Collectively, however, their power lies in the tension generated by their balanced use, which simultaneously reflects a notion of competition and complementarity” (p. 531). Fig. 1 displays the research framework. 2.1. Diagnostic PM system use Diagnostic Use can be defined as a set of formalized procedures that use information to maintain or alter patterns in an organizational activity (Henri, 2006a). These procedures may include reporting systems, monitoring tools, and performance results being communicated to all staff. Also, such procedures may describe a conceptual model for understanding drivers of organizational success or how excellence may be attained (i.e., success map or strategy map), or evaluate critical success factors to monitor performance and specific targets. Widener (2007) states that diagnostic use motivates employees not only to perform but also to align their behavior with organizational goals. Diagnostic use comes in many flavors however. In this paper, we focus on three rather comprehensive diagnostic uses which are tasked to furnish information and aid resource orchestration: “monitoring”, “focusing attention”, and “legitimization.” Monitoring enables managers to track performance against a plan and can identify what, if any, is going awry (and needs correction) (Henri, 2006a,b). The monitoring element of diagnostic use enables managers to benchmark against targets (Widener, 2007). For instance, top management may set goals for customer service level. Monitoring can help them establish whether goals are met and pinpoint where performance is lacking. Focusing attention generates information which is pertinent in directing organizational members on issues of strategic significance. It draws interest on common issues, critical success factors, and integration. Widener (2007) further notes that focusing attention sets boundaries or constraints on employee behavior, alike to the boundary system advocated by Simons et al. (2000). By focusing attention on critical success factors, such as customer service level top management can promote unity of effort across organizational units such as purchasing, operations, and distribution since many constituents influence the capability of the organization to obtain and sustain high levels of customer service. Using a PM system for legitimization has the intention of justifying and qualifying actions and ensuring they are aligned with strategy. In this sense, the legitimization process generates information to enhance the rightfulness and credibility of decisions (Henri, 2006b) to heighten the legitimacy of current and future organizational action. Legitimization can be seen as a political tool (Henri, 2006b) that management can use to assess whether a specific action was judicious. One of the managers we interviewed for this study commented that the methodological validity or the legitimization of the planning process can only be ascertained if it is verifiable ex post. Demonstrating that decisions are credible can positively impact stakeholders’ support and heighten trust in management, which should positively impact the ability to meet or exceed performance targets. Top management can use the PM system to obtain data regarding customer service level to legitimize strategic choices and garner employee support and commitment. The three specific purposes are in essence the nuances that reflect the diagnostic component of PM system use. Diagnostic Use can be thought of as a higher-order variable composed of monitoring, focusing attention, and legitimization. Monitoring ensures that managers know when standards are not being met and thus can make corrections; focusing attention warrants that all levels of the organization remain on the same page and in communication about where the organization is headed while legitimization attests that those decisions are justified and are well-grounded in the organization’s business strategy. Diagnostic use can be used as a tool to quantify actions (Neely et al., 1995) across a variety of metrics. Henri (2006a) suggests that these metrics can assume a variety of orientations such as being financial or non-financial, internal or external, short or long term as well as ex post or ex ante (Franco-Santos et al., 2012). The diagnostic use of PM system also “reflects two important features associated with mechanistic controls: (i) tight control of operations and strategies, and (ii) highly structured channels of communication and restricted flows of information” (Henri, 2006a, p.7). This formal use of PM system provides a mechanistic approach to decision making which furnishes signals when issues such as productivity and efficiency have fallen (Henri, 2006a) outside expectations. However, as a mechanistic control mechanism, diagnostic use, as Henri (2006a) states, has been associated with several dysfunctional behaviors in terms of smoothing, biasing, focusing, filtering, and illegal acts (Birnberg et al., 1983). “These distortions constitute defensive routines that aim to reduce potential embarrassment or threat, or to improve personal interest” (Henri, 2006a, p.7). Given some of the potential for adverse effects, Henri (2006a) in fact labels diagnostic use as a negative force and even hypothesizes a negative relationship between Diagnostic Use and organizational capabilities. It is noted however that Henri (2006a) operationalized Diagnostic Use by measures reflecting only monitoring. Unlike Henri (2006a), we suspect diagnostic use may have a net positive contribution especially in the presence of high interactive use. Diagnostic use is critical because it furnishes information regarding deviations, positive or negative, against expectations. While firms tend to focus on negative deviations rather than X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 positive deviations, the PM system itself is quite adept and furnishes valuable information that can reduce uncertainty and thus improve organizational capabilities. A negative deviation prompts action to steer the organizational energy and resources towards the targets while a positive deviation can be utilized to reinforce organizational efforts. PM systems via diagnostic use can also furnish information that can be consumed to improve relationships with stakeholders. Information generated through PM systems can be pertinent to customers who evaluate suppliers for future business opportunities. Such information may also be germane to the market, which determines the valuation of the firm and its attractiveness to potential investors. Agreeably, diagnostic use may somewhat limit the role of a PM system to a measurement tool (Henri, 2006a), but even as a tool it is nevertheless essential because it plays a significant role in orchestrating resources, especially when deployed concurrently with interactive use. 2.2. Interactive PM system use The interactive use of PM systems tends to have a more developmental role and represents a positive force (Henri, 2006a) employed to expand opportunity seeking and learning throughout the organization. Using a PM system interactively forces dialogue and stimulates the development of new ideas and initiatives (Henri, 2006a; Widener, 2007). According to Henri (2006a), when a PM system is used interactively, the information generated is a recurrent and important agenda item for senior leadership; frequent and regular interest is fostered throughout the organization; data are discussed and interpreted among managers from different hierarchical levels; and “continual challenge and debate occur concerning data, assumptions, and action plans” (Henri, 2006a, p.5). Bisbe and Otley (2004) state that senior managers, through their personal involvement and commitment, can use interactive systems to orchestrate responses to environmental perturbations and craft subsequent organizational renewal. Employing a PM system from an interactive use perspective can generate pressure across organizational levels and motivate further information gathering while stimulating debate. Interactive use embraces institutional dialogue and debate and stimulates information exchange and communication (Henri, 2006a), which play a vital role in reducing equivocality. In addition, interactive PM system use equips the organization with two attributes. It expands the organization’s information processing capacity and engenders interaction amongst constituents (Henri, 2006a), leading to more eloquent orchestration of resources. For example, top management can use the PM system to critically evaluate discrepancies between targeted and actual customer service levels. Top management gets intimately involved with the use of the PM system to analyse root causes and orchestrate necessary resources to achieve targets. Henri (2006a) labels interactive use as a “positive” force suggesting that interactive use promotes opportunity seeking behavior and learning across organizational levels. According to Henri (2006a), the interactive use of a PM system expands its role to a strategic management tool. 2.3. Dynamic tension – diagnostic and interactive PM system roles PM systems have two complementary and interdependent roles (Mundy, 2010) which are information based: (1) to exert control over the attainment of organizational goals and (2) to enable employees to search for opportunities and solve emergent problems via top management commitment and involvement (Ahrens & Chapman, 2004; Chenhall & Morris, 1995; Simons, 1995; Zimmerman, 2000). These roles necessitate a balance between taking actions congruent with the organization’s goals while also 317 giving employees sufficient autonomy to make decisions (Roberts, 1990; Sprinkle, 2003). When combined, controlling and enabling uses of a PM system create dynamic tension (Simons, 1994) which harvests unique organizational capabilities and competitive advantages through organizational dialogue, mutual understanding and better direction (Henri, 2006a). An organization’s inability to balance different uses of PM systems is associated with slower decision making, wasted resources, instability and, ultimately, lower performance (Bisbe et al., 2007; Henri, 2006a). Our theoretical model investigates whether the joint use of PM systems for diagnostic and interactive purposes creates dynamic tension that reflects complementarities (focusing on intended and emergent strategies). According to Henri (2006a, p. 533), “. . . the diagnostic use of PM systems may represent a mechanistic control tool used to track, review, draw attention, and support the achievement of predictable goals. . .” via the provision of information. On the other hand, “. . .interactive use may be considered as an organic control system supporting the emergence of communication processes and the mutual adjustment of organizational actors. . .” which are used to reduce equivocality and stimulate change. Also, diagnostic use of a PM system can be seen as a pure measurement instrument, while interactive PM system use can elevate its role to a strategic management tool (Kaplan & Norton, 2001). Henri’s (2006a) empirical work concludes that there is evidence to suggest that using the two levers of control generates tension which then positively affects performance. Henri tested simple effects and modelled tension as an interaction effect but Henri’s operationalization of interactive use does not adhere to Simons (1994) description. Henri (2006a) operationalizes interactive use via measures of focusing attention while Simons (1994) states that interactive use is described by nuances such as top management commitment and engagement. In contrast to Henri, Widener (2007) operationalized diagnostic use via measures of monitoring and focusing attention while she operationalized interactive use via measures of top management engagement. Widener also suggests that the two levers are complementary and based on empirical evidence claims that the full benefits from a PM system can only be derived when they are used from both diagnostic and interactive perspectives. Widener (2007) however neither specified nor tested interactions between the two systems. 2.4. Hypotheses: PM system USE→ ORGANIZATIONAL CAPABILITIES While much of the PM systems literature posits direct effects on performance, this perspective has been challenged and empirical evidence is scarce. In essence, recent literature supports the notion that diagnostic and interactive uses of PM systems advance organizational capabilities which subsequently help the organization meet its targets. Organizational capabilities “refer to a firm’s capacity to deploy Resources, usually in combination, using organizational processes, to effect a desired end. . ..Capabilities are based on developing, carrying, and exchanging information through the firm’s human capital” (Amit & Schoemaker, 1993, p. 35). The PM system may affect organizational capabilities by increasing communication between employees or encouraging employees to analyze how resources should be combined to solve problems. A PM system can encourage improvements across several capabilities such as in the areas of strategic management, operations, and stakeholder relationships. Hence, varied PM system usage (i.e., diagnostic and interactive uses) may have an effect on firm performance via its impact on organizational capabilities and meeting organizational targets. Effective strategic management involves necessarily a variety of critical orchestrating activities. Strategic management 318 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 encompasses activities such as the prioritization of strategic actions and objectives, obtaining information and feedback regarding the existing strategy and forthcoming strategy, the alignment of organizational processes with the intended strategy, and the allocation of budgets and capital expenditures among others (Porter, 1991; Simons, 1991). These activities rely on feedback obtained from diagnostic systems and direction from top management managers who interactively obtain information, and discuss, challenge, and debate courses of action with other organizational members across levels. The information obtained from diagnostic and interactive PM system uses can help organizations reduce uncertainty and ease equivocality, enhancing strategic management capability. Without information derived from performance systems, organizations cannot assess their current performance vis-à-vis their intended strategy and may have a difficult time charting a new course of action or direction without knowing the facts and without significant debate that can emerge from interactive use. The use of both systems is necessary as the organization has to balance the needs for control and renewal. We expect that PM system uses will have a positive effect on strategic management capability but we also posit that the interaction between the two types of uses will explain additional variation, above and beyond what the individual uses may contribute. – H1a : Diagnostic Use will have a positive impact on Strategic Management Capability. – H1b : Interactive Use will have a positive impact on Strategic Management Capability. – H1c :The Dynamic Tension between Diagnostic Use and Interactive Use will have a positive impact on Strategic Management Capability. Organizations also rely on their operational capabilities to garner advantage against competitors. Competitive advantage can emerge for instance from more productive, innovative, and integrated processes, and higher employee productivity. To enhance such processes, however, the organization relies on information furnished by both diagnostic and interactive systems and the tension created when both systems are used concurrently. For instance, information regarding productivity or quality has to be collected and shared in order for organizational constituents to take corrective measures when necessary or for promotion of processes that prove to be exemplary. Also, in order to integrate processes and orchestrate resources, organizations need to discuss the value proposition, identify costs and benefits, and set direction, all of which need information and deliberation across organizational members along with top management involvement. We expect that diagnostic and interactive PM system uses along with the interaction between the two uses will have a boosting effect on operational capabilities. H2a :Diagnostic Use will have a positive impact on Operational Capability. H2b : Interactive Use will have a positive impact on Operational Capability. H2c : The Dynamic Tension between Diagnostic Use and Interactive Use will have a positive impact on Operational Capability. Organizations rely on external stakeholders for business, affirmation, and resources (Freeman, 1984; Donaldson & Preston, 1995, Agle et al., 1999; Berman et al., 1999). Relationships with such stakeholders depend however on information. For instance, relationships with customers depend on timely and accurate information generated via performance monitoring systems. Customers can demand information regarding quality defects or costs. Similarly, suppliers seek feedback information furnished by monitoring systems concerning their performance. Challenges and opportunities regarding customers, suppliers, and regulators require debate and guidance by top management, which can interactively take an interest to generate solutions and resolve equivocality. However, when the two systems are used together, there is even greater potential to improve external stakeholder relations because the two systems can tackle uncertainty and equivocality only when used jointly. Thus, we hypothesize that the two systems and their interaction will have a positive effect on external stakeholder relations. H3a : Diagnostic Use will have a positive impact on External Stakeholder Relations Capability. H3b : Interactive Use will have a positive impact on External Stakeholder Relations Capability. H3c : The Dynamic Tension between Diagnostic Use and Interactive Use will have a positive impact on External Stakeholder Relations Capability. 2.5. Hypotheses: ORGANIZATIONAL CAPABILITIES→TARGET PERFORMANCE Organizations capable of effective strategic management can direct resources towards the attainment of organizational targets. First, managing feedback regarding existing strategy can help the organization align its operations with the respective strategy and thus properly equip the organization to meet its goals. Second, strategic management efforts can help the organization decide on its strategic objectives and allocate resources to achieve those objectives. In this form, strategic management is a protagonist but it also plays an enabling role. Effective strategic management should impact the attainment of organizational goals positively. H4 : Strategic Management Capability has a positive effect on Target Performance. Organizations depend on their processes to add value and derive healthy returns. These processes have to bolster the organizational operational capabilities which are responsible to propel the organization to meet or exceed its targets. Operational capabilities can be reflected by a variety of nuances such as enhanced productive processes, improved employee performance, integrated processes, and innovation in working processes (Simons, 1991). For example, improved employee performance, more integrated processes, and more innovation in working processes can help customers derive better value from the relationship with the firm and thus the organization can reap such benefits to meet its targets regarding customer relations. Similarly, delivery performance can be enhanced through operational capabilities and thus the organization can subsequently meet its operational performance targets. Consequently, we expect that higher levels of operational capabilities will enable the organization to meet its targets more effectively. H5 : Operational Capability has a positive effect on Target Performance. Organizations depend on external stakeholders for business opportunities, support, resources, acknowledgement, validation, and rulings (Donaldson & Preston, 1995; Agle et al., 1999; Berman et al., 1999). External stakeholders include customers, suppliers, regulators or other governmental institutions as well as the market at large. Customers for instance provide business opportunities while suppliers are assuming an ever growing importance in assuring the firm’s value chain. Organizations can improve profitability by increasing revenues and/or cutting costs. Improving relationships with customers can potentially increase revenues while working closer with suppliers can trim costs. Organizations cannot ignore external stakeholders because they can help the organization meet or exceed its targets (Berman et al., 1999; Margolis & Walsh, 2003). X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 H6 : External Stakeholder Relations Capability has a positive effect on Target Performance. 2.6. Hypotheses: MEETING TARGET PERFORMANCE→FIRM PERFORMANCE When organizations meet their customer relations targets, this generally manifests an ability to keep customers happy and improve market share. Ultimately, this translates into higher returns in terms of sales growth and profit growth, among other financial measures. Similarly, when organizations meet operational performance targets, they enjoy productivity improvements or shorter lead times which then lead to better returns on investment and higher profit to sales ratios. We expect that these effects will be robust across both subjective as well as objective financial ratios. We thus hypothesize that firms that meet their performance targets will demonstrate better performance. H7 : Target Performance will have a positive effect on firm performance. 2.7. Hypotheses: LONGITUDINAL EFFECTS PM systems are expensive to implement and maintain. The premise, however, is that the use of a PM system generates positive financial returns over time. PM systems furnish information and engender more effective resource orchestration which is critical for the organization’s current well-being and future prosperity. Organizations obtain and process information for operational, relational, and strategic use. We expect that firms that use a PM system would outperform firms that do not deploy a PM system. They should have better returns on assets and enjoy higher earnings. From a static perspective, the performance of the firm can be assessed by examining financial ratios before and after implementation. From a dynamic view, performance improves over time as organizations learn to adapt to a new system. The dynamic perspective acknowledges that time plays a vital role. H8 : PM system use will have a positive effect on firm financial performance. 3. Cross sectional study Our first study can be described as cross sectional in nature. We illustrate below the operationalization of constructs, research design and research methods, and provide results based on subjective and objective data obtained for 2010. Objective data for both the cross-sectional and longitudinal analyses were obtained from Aida - Bureau van Dijk (https://aida.bvdinfo.com/), a source of financial performance data and other information pertaining to firms operating in Italy and many other countries. 3.1. Operationalization of constructs In order to operationalize the variables, the researchers sought to identify existing scales that can be employed or scales that can be adapted to the specific context. Where existing scales were not identified, the researchers relied on the extant literature to specify the content domain of each construct. The review also included an investigation of contextual and organizational factors that may affect performance. Contextual factors may involve primary industry affiliation, firm size, and geographic ownership diversification while organizational factors may embrace measures of the resources the firm possesses. To assure that the questions could be correctly understood by respondents and easily answered by them, the initial survey questionnaire was carefully pre-tested. Thus, a preliminary draft of the 319 questionnaire was discussed with academic scholars to assess clarity, simplicity, and content validity. Afterwards, a pilot study was also conducted with a group of accounting managers and controllers of six large organizations, where their inputs were used to improve the clarity, comprehensiveness, and relevance of the survey questions. Finally, the emerged survey was reviewed once more by a panel of three academics. The final survey items are discussed below and appear in Table 2. Diagnostic use was specified as a second-order construct and was operationalized via three first-order variables which include monitoring, focusing attention, and legitimization. All three rely on measures developed by Henri (2006a,b) which he adapted from Vandenbosch (1999). Monitoring is reflected via four manifest variables. Focusing Attention is operationalized by seven observed variables. Legitimization includes nine observed items. Interactive use relies for content on the work of Widener (2007). Strategic Management Capability is a new scale and consists of six measures that address whether the PM system in the organization supports strategic management activities. These measures draw on the work of Simons (1991) and his work on management control systems. Operational Capability is also a new scale and was captured via five indicators addressing productivity, employee performance, innovation of working processes, and development of integrated solutions. External Stakeholder Relations Capability is a multi-item scale and relies on the work of Cousins et al. (2008) and Mahama (2006). We operationalized Target Performance via a five item scale which drew items from Ittner et al. (2003b). It addresses the extent to which the organization meets the performance targets across a variety of categories. Finally, Subjective Financial Performance is measured using four financial indicators relative to industry average. 3.2. Research design The sample population consisted of 1000 organizations operating in Italy and randomly selected from Amadeus - Bureau Van Dijk (a database of public and private firms which include Italian firms as well as multinational firms filing separately for their Italian operations). An appropriate sample of organizations covering different industry sectors and organizational characteristics has been selected to maximize the generalizability of the findings. We contacted each firm’s management directly to select a list of companies prepared to cooperate with the research. We then identified the key figures in each company who would be competent to fill in the survey. Names and e-mail addresses of targeted respondents were obtained during telephone contacts. The research design enabled the researchers to target respondents that possess substantive organizational level knowledge and specific knowledge as it pertains to PM systems. We targeted high level executives (i.e., CEOs, CFOs) as well as controllers and managers (e.g., managing directors, operations managers). Given the positional levels of these individuals, we are confident that the target population can serve as excellent respondents both in reference to explanatory variables as well as performance variables. The research has been carried out through a survey questionnaire using fax and e-mail during 2010. To increase the survey response rate, we pre-notified specific respondents by phone to seek their input regarding their availability to work on the survey. This approach probably led to more involvement and commitment by the respondents from the beginning of the project (Baldauf et al., 1999). The survey instrument was preceded by an introductory letter clarifying the purposes and objectives of the entire research project. Managers were also promised an overall PM system benchmark study to compare their responses to those of other participating organizations. Respondents were alerted to respond to the survey on three subsequent occasions. 320 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 1 Demographics. Respondent’s job title CEO CFO Controller Managing Director Operations Manager HR Manager Finance Manager Others Frequency Percentage Cumulative percent 19 57 185 34 58 17 6 10 4.90 14.80 47.90 8.80 15.00 4.40 1.60 2.60 4.90 19.70 67.60 76.40 91.40 95.80 97.40 100.00 246 140 63.70 36.30 63.70 100.00 202 184 52.30 47.70 52.30 100.00 221 69 18 13 25 28 12 57.30 17.90 4.70 3.40 6.50 7.30 3.10 57.30 75.10 79.80 83.20 89.60 96.90 100.00 12 22 51 59 61 53 120 8 3.20 5.80 13.50 15.60 16.10 14.10 30.00 2.10 3.20 9.00 22.50 38.10 54.20 68.30 100.00 Industry Manufacturing Service Geographic Diversification Italian Multinational Employee Size n < 500 500 ≤ n ≤ 999 1000 ≤ n ≤ 1499 1500 ≤ n ≤ ≤1999 2000 ≤ n ≤ 2999 3000 ≤ n ≤ 9999 n ≥ 10000 Organizational Age Age ≤ 5 6 ≤ Age ≤ 10 11 ≤ Age ≤ 20 21 ≤ Age ≤ 30 31 ≤ Age ≤ 40 41 ≤ Age ≤ 50 Age ≥ 51 Not reported To address the research questions and hypotheses, we obtained useful responses from a cross-section of 386 firms operating in Italy, which represents close to a 39% effective response rate. We note that we obtained surveys from a total of 401 firms but responses from 15 firms were discarded due to excessive missing data. There were very rare instances of missing values for the retained firms and those appeared to be missing at random; missing values were mean replaced. Demographics appear in Table 1. Roughly 20% of the respondents are C-level executives while about 48% of respondents are controllers. Operations managers and managing directors accounted for another 23% of participants. As far as firm size is concerned, small, medium, and large organizations are represented and reflect general industrial demographics in Italy. Around 57.30% of the firms are considered small with employment levels below 500 while 17% of the firms have a size greater than 2,000. About 64% of the sample represents manufacturing firms and 36% of the responding firms are service oriented. There is almost equal representation between pure Italian firms (52%) vis-à-vis multinationals (48%). The firms tend to be older with relatively few firms (9%) dating less than 10 years old. Overall, the demographics suggest a broad representation of firms operating in Italy. In order to identify whether the responding firms differ from the population of firms operating in Italy, we randomly selected a blend of another 581 firms operating in Italy which shared a similar proportion of manufacturing and service firms as our sample, as well as domestic and multinational firms operating in Italy, and compared the two groups across many dimensions. Based on t-tests (a = .05), the analysis suggests no differences in terms of employee size, sales, profitability, and total assets. Thus, there is no evidence to suggest that the sample is biased against the population of firms operating in Italy. Our research design included a variety of statistical and separation approaches (Craighead et al., 2011) to minimize the risk of common method variance (CMV). In terms of statistical remedies, Harman’s single factor test using CFA failed to support a single factor. We also used a marker variable technique (Malhotra et al., 2006). According to Craighead et al. (2011), the marker variable technique calls for the inclusion of an additional variable in the study which is theoretically unrelated to at least one other variable of interest (Harrison et al., 1996). Specifically, the marker variable (i.e., My company has an operating top management philosophy of a growth strategy primarily through external financing [borrowing, capital issues, etc.]; rated on a seven point scale where 1 = to a very great extent and 7 = not at all) exhibited low correlations with all focal variables: Diagnostic Use (−.008), Interactive Use (.081), Strategic Management Capability (.010), Operational Capability (.033), External Stakeholder Relations Capability (.094) Target Performance (−.105), and Subjective Financial Performance (−.094). We also lessened the risk of CMV via methodological separation (Craighead et al., 2011). The independent and dependent perceptual variables were measured using different scales and different formats. Subjective financial performance was operationalized using a different scale than target performance and all other variables in the model. Importantly, the perceptual data set is relying on a cross-sectional survey while objective financial performance was obtained from secondary sources for a ten year period, including time periods both before and after the collection of perceptual data. Given the applied remedies, CMV is not expected to impact our specified relationships. 3.3. Research methods and results To address the measurement model through confirmatory factor analysis, we used a covariance matrix as input and tested it via Mplus (Table 2). The measurement model includes first-order factors as well as a second-order factor which operationalizes Diagnostic Use. To ascertain the efficacy of the second-order model for Diagnostic Use, we examined four models within the context of the entire measurement model, relying on the guidance of Koufteros et al. (2009). Details about the methods of examining for the second-order model specification appear in Appendix A. The overall CFA model, which includes the second-order specification for Diagnostic Use, has acceptable model fit as indicated by the fit statistics (2 = 1985.576, df = 1090, 2 /df = 1.82, CFI = 0.93, NNFI = 0.92, RMSEA = 0.046, and 90% CI of RMSEA = (.043, .049). All item-factor loadings are greater than 0.50, and are statistically significant at the 0.01 level. We assessed discriminant validity by comparing the squared correlation between each pair of firstorder constructs and their respective average variance extracted (AVE) (Table 2) (Koufteros, 1999). A squared correlation greater than individual AVEs provides evidence to suggest the absence of discriminant validity. The highest squared correlation, which can be derived from Table 3, is 0.40 and it is significantly lower than the respective AVEs of the two specific first-order constructs, i.e., Strategic Management Capability (AVE = 0.57) and Operational Capability (AVE = 0.63). The composite reliabilities (Table 2) for the focal constructs are greater than the threshold of 0.70. Overall, the constructs appear to be reliable and valid. We subsequently specified a structural model to partially examine the hypotheses advanced earlier. Before testing the structural model, however, we examined the distribution of each variable via measures of kurtosis and skewness, along with visual inspections. Each variable appeared to have an approximately normal distribution. Also, the scores for Diagnostic Use and Interactive Use were mean-centered before creating the interaction term which X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 321 Table 2 Measurement model and descriptive statistics. Latent Variables and Respective Manifest Items Mean/item Standard Deviation Standardized Coefficient T-value Response format for the following three constructs: Using the scale where 1 = To a very great extent, 2 = To a great extent, 3 = To some extent, 4 = To a little extent, 5 = To a very little extent, and 6 = Not at all, please rate the extent to which your top management team currently uses performance measures to: Monitoring, Composite Reliability (CR) = .82, AVE = .54 Track progress towards goals Review key measures Monitor results Compare outcomes to expectations 2.16 2.73 1.83 1.96 .923 1.007 .793 .888 .86 .72 .68 .67 39.39 24.69 21.48 20.57 3.03 2.86 2.44 2.98 2.76 2.92 2.55 1.002 .98 .971 1,088 1.030 1.018 1.016 .75 .80 .78 .67 .75 .66 .74 29.69 35.57 33.20 21.40 29.83 19.94 28.08 2.67 2.72 2.72 2.81 2.57 2.78 2.51 2.72 2.92 .928 .989 .901 .920 .866 .936 .978 .953 .970 .71 .57 .63 .75 .80 .76 .69 .69 .70 24.65 15.25 18.58 29.63 37.54 30.76 23.56 23.28 24.01 Focusing Attention, CR = .89, AVE = .54 Integrate the organisation- i.e. tie the organisation together Enable the organisation to focus on common issues Enable the organisation to focus on your critical success factors Develop a common vocabulary in the organisation Provide a common view of the organisation Enable discussion in meetings of superiors, subordinates and peers Enable continual challenge and debate underlying results, assumptions and action plans Legitimization, CR = .90, AVE = .50 Increase your understanding of the business Justify decisions Verify assumptions Maintain your perspectives Support your actions Reinforce your beliefs Stay close to the business Increase your focus Validate your point of view Response format for the following construct: In my organisation, senior managers have strong commitment to. . .Use the scale where 1 = To a very great extent, 2 = To a great extent, 3 = To some extent, 4 = To a little extent, 5= To a very little extent, and 6 = Not at all. Interactive Use, CR = .90, AVE = .64 Reviewing organisational performance based on our performance measurement systems Using performance measures to analyse root problems Using performance measures to achieve targets Using performance measures to make decisions Using performance measures to get information to support decision making 2.81 2.50 2.31 2.40 2.40 .977 .915 .871 .878 .865 .74 .87 .76 .82 .82 27.91 51.89 30.48 38.87 38.97 Response format for the following three constructs: The performance measurement systems in my organisation . . .Use the scale where 1 = To a very great extent, 2 = To a great extent, 3 = To some extent, 4 = To a little extent, 5 = To a very little extent, and 6 = Not at all. Strategic Management Capability, CR = .89, AVE = .57 Support the achievement of key strategic objectives Improve the prioritisation of actions, projects and objectives Give feedback on the company strategy & its strategic direction Give feedback on operational processes Improve the alignment of strategy and operations Enhance negotiation of capital expenditure, budget allocation and financial support to projects 2.36 2.45 2.48 2.61 2.69 2.80 .861 .911 .889 .906 .906 .940 .75 .80 .77 .72 .84 .62 30.01 37.10 32.09 26.49 45.09 17.97 2.91 2.89 2.66 2.76 2.97 .880 .907 .838 .816 .876 .85 .79 .84 .75 .73 45.78 34.91 45.20 27.15 27.84 2.95 3.17 2.80 3.32 .972 1.028 1.036 1.190 .80 .72 .71 .50 27.92 20.99 20.87 11.27 Operational Capability, CR = .89, AVE = .63 Increase the innovation of working practices Enhance the development of integrated solutions Promote operational improvements Increase productivity Improve employee performance in their operations External Stakeholder Relations Capability, CR = .78, AVE = .49 Improve the overall company’s leadership in the market Improve our relationship with suppliers Improve our relationship with customers Improve our relationship with regulators or government institutions Response format for the following construct: To what extent does your organisation meet the targets set for the following categories of measures: Use the scale where 1 = To a very great extent, 2 = To a great extent, 3 = To some extent, 4 = To a little extent, 5 = To a very little extent, and 6 = Not at all. Target Performance, CR = .80, AVE = .49 Short term financial results – e.g. operating income, sales growth, etc. Customer relations- e.g. market share, customer satisfaction, etc. Employee relations- e.g. employee satisfaction, safety, etc. Operational performance – e.g. productivity, lead times, etc. Environmental performance – environmental compliances, etc. 2.25 2.59 3.06 2.41 2.94 .870 1.016 1.038 .911 1.276 .50 .69 .75 .71 .67 10.94 21.34 25.38 22.35 19.37 322 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 2 (Continued). Latent Variables and Respective Manifest Items Mean/item Standard Deviation Standardized Coefficient T-value Response format for the following construct: In comparison with the industry average, how would you rate the performance of your company over the last two years in terms of the following indicators? Use the scale where: 1: Well Above Average, 2: Above Average, 3: Short of Above Average, 4: Average, 5: Short of Below Average, 6: Below Average, 7: Well Below Average Subjective Financial Performance, CR = .93, AVE = .77 Rate of sales growth Rate of profit growth Return on investment (ROI) Profit/sales ratio 3.11 3.28 3.35 3.30 1.257 1.358 1.289 1.266 .78 .91 .90 .92 35.03 75.31 72.86 87.89 .82 .87 .85 31.67 33.22 35.96 Diagnostic Use-Second Order Factor, CR = .88, AVE = .72 Monitoring Focus Attention Legitimization Fit Indices: Chi-Square (df) = 1985.576 (1090), Chi-Square/df = 1.82, NNFI = .92, CFI = .93, RMSEA = .046, 90% CI RMSEA = (.043,.049) operationalizes Dynamic Tension to mitigate potential multicollinearity. The structural model fit was evaluated using Mplus via several criteria and structural paths were examined for statistical significance based on t-tests and respective p-values. Correlations with objective financial performance measures appear negative due to the specific scaling (see Table 2) of all perceptual measures. Based on ROT and RBT, we control for the possession of a variety of both intangible and tangible resources. This was necessitous in order to isolate the unique contributions of asset orchestration in contrast to asset possession. The intangible assets we control for include Research & Development Expenditures and Industrial Patents & Intellectual Property Rights while the tangible assets encompass Plant & Machinery and Industrial & Commercial Equipment. We also control for contextual variables such as firm size (number of employees in current year), industry (manufacturing versus service) and whether the firm is multinational. Table 4 presents standardized coefficients along with respective significance levels and t-values. The fit indices indicate that the structural model exhibits a good model-to-data fit (2 = 2515.635, df = 1666, 2 /df = 1.509, CFI = 0.91, NNFI = 0.90, RMSEA = 0.043, 90% CI of RMSEA = [0.039, 0.046]). H1a and H1b predict respectively that diagnostic use and interactive use have a positive effect on strategic management capability. There is ample statistical evidence to support the hypotheses (p < .000 and p < .000 respectively) although the effect size for Diagnostic Use ( = 0.625) is significantly larger than the effect size of Interactive Use ( = 0.246). Furthermore, H1c suggests that strategic management capability is subject to the interaction between diagnostic use and interactive use, termed as Dynamic Tension. Dynamic Tension did produce a positive relationship ( = 0.084) with Strategic Management Capability though the effect was statistically moderate (p < .05). Similarly, H2a , H2b , and H2c posit respective positive effects on operational capability. All three standardized effects were positive but clearly the effect of diagnostic use (as opposed to interactive use and dynamic tension) is more prominent from a substantive ( = 0.577 vs = 0.233 vs = 0.113) and statistical perspective (p < .000 vs p < .002 vs p < .010). As far as external stakeholder relations capability is concerned, diagnostic use (H3a ) and interactive use (H3b ) play a significant role. Diagnostic Use and Interactive Use have statistically significant positive effects ( = 0.535, p < .000 and = 0.320, p < .000 respectively) on External Stakeholder Relations Capability while the impact of Dynamic Tension (H3c ) was negative, albeit minimal ( = −0.032, p < .280). Collectively, all three explanatory variables contribute towards organizational capabilities. While we expected that to be the case, we were surprised that diagnostic use had such a strong positive showing given Henri’s (2006a) conceptual and empirically based designation of diagnostic use as a “negative force.” We note however that our operationalization of diagnostic use and our choices for capabilities differ and thus direct comparisons may not be fruitful. Henri operationalized diagnostic use only via measures of monitoring, which may be perceived as surveillance instruments that can be used merely to exercise managerial control. Instead, when monitoring is coupled with focusing attention and legitimization, employees can more clearly see the purpose of diagnostic use and verify that prior actions were credible, commanding higher levels of buy-in. We found that diagnostic use is a mechanism organizations can rely on to improve a wide range of capabilities. Organizations strive to meet their target performance using several means. The data analysis here illustrates that these capabilities can have a statistically significant and positive effect on the ability of the organization to meet its target performance. Interestingly, the effect of external stakeholder relations capability was more pronounced ( = 0.431, p < .000) than the effects of strategic management capability ( = 0.174, p < .024) and operational capability ( = 0.163, p < .027). This was surprising as we expected that operational capability, being more proximal to performance, would probably be the most salient factor. However, we found some corroborating theoretical and empirical evidence in the literature that can perhaps explain the prominence of stakeholder relations capability as an explanatory variable. For instance, Sarkis et al. (2010) underscore the importance of stakeholders and according to Franco et al. (2012) organizations are under a relentless pressure to deliver value not only to shareholders but increasingly to other stakeholders. The first consequence is perhaps the inclusion of specific organizational performance targets which are salient to stakeholders while the second consequence is the development of a top leadership compensation structure which is tied to stakeholder relations/corporate social performance. The stakeholder theory (Donaldson & Preston, 1995) is considered to be rather instrumental for performance. As Donaldson and Preston (1995) articulate, “. . .the principal focus of interest here has been the proposition that corporations practicing stakeholder management will, other things being equal, be relatively successful in conventional performance terms.” (p. 66–67). Berman et al. (1999) present a strategic stakeholder management model under which “. . .the nature and extent of managerial concern for a stakeholder group is viewed as determined solely by the perceived ability of such concern to improve firm financial performance” (p. 488). Berman et al. (1999) empirically demonstrated that after controlling for the strategy and operational environment, stakeholder efforts are positively related to performance. A review of the literature that examined the relationship between corporate Table 3 Correlation matrix. 1 2 3 4 5 6 7 8 1. Monitoring 2. Focus Attention 3. Legitimization 4. Diagnostic Use1 5. Interactive Use 6. Dynamic Tension2 7. Strg. Mgmt Cap. 8. Operational Cap. 9. External Stakehol. Relations Cap. 10. Target Perf. 11. Subjective Financial Perf. 12. ROTA3 13. ROA4 14. ROE5 15. EBITM6 16. Employee Size7 17. Industry 18. Geographic Div. 19. R&DE8 20. IP&IPR9 21. P&M10 22. ICE11 1 .633** .604** .796** .525** .286** .599** .448** 9 10 1 .630** .883** .477** .318** .554** .554** 1 .897** .513** .387** .583** .511** 1 .575** .391** .660** .590** 1 .408** .644** .616** 1 .404** .375** 1 .635** 1 .374** .463** .468** .465** .410** .485** .487** .543** .522** .448** .219** .257** .473** .458** .578** .492** 1 .453** 1 .216** −.113* −.107* −.081 −.135** −.048 −.068 −.220** .008 .019 .093 −.107 .225** −.088 −.080 −.066 −.082 −.099 −.045 −.117* −.021 −.053 .104 −.139* .191** −.102* −.094 −.101* −.121* −.025 −.056 −.185** .024 −.016 .224 −.096 .239** −.113* −.105* −.096 −.125* −.066 −.062 −.191** .004 −.026 .095 −.131* .241** −.073 −.069 −.085 −.114* −.066 −.074 −.156** −.036 −.020 .747 −.115* .101* −.036 −.034 −.128* −.048 −.039 −.120* −.057 −.017 −.028 .779 −.011 .273** −.048 −.040 −.035 −.066 −.098 −.027 −.101* −.004 −.005 −.104 −.102 .246** −.070 −.058 −.058 −.083 −.021 −.081 −.117* −.047 −.034 −.008 −.132* .247** −.002 .002 −.074 .006 −.063 −.041 −.131** −.084 −.055 −.013 −.106 .360** −.124* −.116* −.178** −.130* −.083 −.099 −.262** −.029 .002 −.078 −.103 2 11 12 13 14 15 16 17 18 19 20 21 22 1 −.094 −.085 −.168** −.109* −.111* −.021 −.054 .009 .003 −.127* −.051 1 .994** .619** .962** −.130* .233** .228** .009 .045 .048 .033 1 .618** .957** −.128* .229** .222** .008 .037 .048 .015 1 .588** −.041 .217** .144** .020 .035 .048 −.031 1 −.123* .213** .227** .011 .180** .430** .202** 1 −.145** .155** .185** .256** .255** .467** 1 .181** .067 −.109* −.134* .001 1 .071 −.034 .043 .134* 1 .020 −.011 .297** 1 .153** .566** 1 .377** 1 Diagnostic Use is a second-order Construct that includes Monitoring, Focus Attention, & Legitimization, Dynamic Tension= Diagnostic Use * Interactive Use, ROTA = Return on Total Assets = EBIT/Total Net Assets where EBIT= Revenue–COGS - Operating Expenses - Depreciation & Amortization ROA = Return on Assets = Net Income/Total Assets 5 ROE = Return on Equity = Net Income/Shareholders Equity 6 EBITM = Earnings before Interest and Taxes Margin = Last Four Quarters of Operating Earnings/Last Four Quarters of Sales 7 Employee size was measured on a scale where: 1: n < 500, 2: 500 < n < 999, 3: 999 < n < 1499, 4: 1500 < n < 1999, 5: 2000 < n < 2999, 6: 3000 < n < 9999, 7: n ≥ 10,000 8 Research & Development Expenditures (R&DE - in Euros): Investments in Research & Development (Based on the Balance Sheet). Count as Intangible Assets. 9 Industrial Patents and Intellectual Property Rights (IP&IPR - in Euros): Count as Intangible Assets. 10 Plant and Machinery (P&M - in Euros): Investments in Tangible Assets. 11 Industrial and Commercial Equipment (ICE - in Euros): Investments in Tangible Assets. 3 4 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 1 Variable 323 324 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 4 Hypotheses testing. Hypothesis Effect Standardized Coefficient t-Value Significance H1a H1b H1c H2a H2b H2c H3a H3b H3c H4 H5 H6 Diagnostic Use→ Strategic Management Capability Interactive Use→ Strategic Management Capability Dynamic Tension→ Strategic Management Capability Diagnostic Use→ Operational Capability Interactive Use→ Operational Capability Dynamic Tension→ Operational Capability Diagnostic Use→ External Stakeholder Capability Interactive Use→ External Stakeholder Capability Dynamic Tension→ External Stakeholder Capability Strategic Management Capability → Target Performance Operational Capability → Target Performance External Stakeholder Capability → Target Performance Target Performance→ Subjective Financial Performance Target Performance→ROTA Target Performance→ROA Target Performance→ROE Target Performance→EBITM Firm Size → Subj. Fin. Perf. Industry → Subj. Fin. Perf. Geog. Diversification → Subj. Fin. Perf. R&D Expenditures → Subj. Fin. Perf. Industrial Patents & Intell. Prop. Rights → Subj. Fin. Perf. Plant & Machinery → Subj. Fin. Perf. Industrial & Commercial Equipment → Subj. Fin. Perf. Firm Size → ROTA Industry → ROTA Geog. Diversification → ROTA R&D Expenditures → ROTA Industrial Patents & Intell. Prop. Rights → ROTA Plant & Machinery → ROTA Industrial & Commercial Equipment → ROTA Firm Size → ROA Industry → ROA Geog. Diversification → ROA R&D Expenditures → ROA Industrial Patents & Intell. Prop. Rights → ROA Plant & Machinery → ROA Industrial & Commercial Equipment → ROA Firm Size → ROE Industry → ROE Geog. Diversification → ROE R&D Expenditures → ROE Industrial Patents & Intell. Prop. Rights → ROE Plant & Machinery → ROE Industrial & Commercial Equipment → ROE Firm Size → EBITM Industry → EBITM Geog. Diversification → EBITM R&D Expenditures → EBITM Industrial Patents & Intell. Prop. Rights → EBITM Plant & Machinery → EBITM Industrial & Commercial Equipment → EBITM .625 .246 .084 .577 .233 .113 .535 .320 −.032 .174 .163 .431 .414 −.199 −.169 −.225 −.115 −.090 −.010 .070 .026 −.031 −.102 .053 .088 .026 .007 .005 −.011 .054 −.080 .153 .001 −.016 .019 −.038 .073 −.127 .058 .093 −.096 .059 −.009 .105 −.147 .070 −.107 −.045 .025 −.097 .492 −.015 8.878 3.224 1.712 7.308 2.820 2.324 6.200 3.578 −.587 1.974 1.922 4.156 6.952 −3.035 −2.584 −3.526 −1.954 −1.345 −.174 1.152 .418 −.447 −1.393 .686 1.272 .435 .110 .084 -.149 .712 −1.020 2.232 .018 −.258 .291 -.531 .976 −1.623 .846 1.581 −1.563 .927 −.128 1.422 −1.904 1.141 −2.015 −.809 .429 −1.530 7.943 −.222 .000 .000 .043 .000 .002 .010 .000 .000 .280 .024 .027 .000 .000 .001 .005 .000 .025 .090 .431 .125 .338 .328 .082 .243 .102 .332 .456 .465 .441 .438 .154 .013 .493 .398 .355 .298 .165 .052 .198 .057 .059 .354 .449 .078 .028 .127 .022 .209 .334 .063 .000 .412 H7 Controls Fit Indices: Chi-Square (df) = 2515.635 (1666), Chi-Square/df = 1.509, NNFI = .91, CFI = .90, RMSEA = .043, 90% CI RMSEA = (.039,.046). Statistically significant effects are noted in italics. social performance and corporate financial performance over 1972-2002 identified 109 empirical studies, out of which 54 exhibited a positive association while only 7 suggest a negative association (Margolis & Walsh, 2003). In the context of PM systems, Cousins et al. (2008) and Mahama (2006) empirically found that PM systems enhance perceived interfirm financial and nonfinancial performance indirectly by first improving cooperation and socialization among firms. Mahama’s (2006) research suggests that PM systems help to ensure that performance information is distributed fairly among participants in the supply chain relationship, which promotes learning and problem solving. Furthermore, the sharing of information helps to align the interests amongst relationship constituents, motivating them to adapt to changes without the need for coercive measures. Cousins et al. (2008) demonstrate that the use of PM systems enhances communication, which improves socialization, an essential element that can facilitate orchestration of resources across organizational boundaries. Given the perceived contributions that relations with external stakeholders can beget, companies have incentivised their top leadership in order to improve relations with stakeholders (Deckop et al., 2006). Towards this end, Deckop et al. (2006) suggest that CEO pay structure can be used as an incentive to promote better relations with stakeholders. They found that CEO long-term pay focus did have an impact on corporate social performance. We conjecture that given the incentive regime where top leadership is rewarded based on relations with stakeholders, the organization will be more likely to measure performance against respective targets and then direct resources toward the improvement of stakeholder relations, which subsequently fuels better performance. Considering that this is a relatively nascent topic, we expect that there is significant variance on how organizations use their resources to improve X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Diagnosc Use 325 Subjecve Financial Performance Strategic Management Capability ROTA Dynamic Tension (Diagnosc X Interacve) Operaonal Capability Target Performance ROA ROE Interacve Use External Stakeholder Relaons Capability EBIT P - Values p<.01 0.01 ≤ p<.05 Non - stascally Significant Fig. 2. Test Results Note: Effects on ultimate dependent variables are controlled for industry, firm size, geographic diversification, R&D expenditures, Industrial Patents & Intellectual Property Rights, Plant & Machinery, and Industrial & Commercial Equipment. These effects are displayed in Table 4. stakeholder relations capabilities and this rationale explains the striking positive effects of external stakeholder relations capabilities on target performance. We next examined whether firms that meet their target performance actually do perform well financially, while controlling for intangible and tangible firm resources as well as firm demographics. We demonstrate that meeting target performance does have a significant impact on both subjective financial performance as well as objective financial measures of performance. Subjective financial performance was operationalized via several manifest variables while objective financial performance was reflected via four measures: ROTA, ROA, ROE, and EBITM (details about formulations appear at the bottom of Table 3). As expected, the effect of Target Performance on Subjective Financial Performance was substantive (ˇ = .414, p < .000, R2 = .195). Though the effects on objective financial performance measures were not as strong, they are sizable given the plethora of other variables that can impact the specific dimensions of financial performance. The coefficients assume a negative sign due to the reverse scale coding of Target Performance (see Table 2). The respective impact was: ROA (ˇ = −.169, p < .005, R2 = .050), ROTA (ˇ = −.199, p < .001, R2 = .048), ROE (ˇ = −.225, p < .000, R2 = .079), and EBITM (=−.115, p<.025, R2 = .252). Fig. 2 presents the hypothesized structural model and respective significance levels. Although we suspected that the relationship between target performance and subjective financial performance will be statistically and substantively significant, we were surprised that the effects on the objective financial measures of performance were all relatively sizable and statistically significant. There is always the concern that results grounded on primary data may differ from the results which are resting on secondary data. In this particular case, the findings are consistent and thus more credible. 3.3.1. Further review of interaction effects The data analysis revealed two statistically significant interaction terms relating Dynamic Tension to Strategic Management Capability and Operational Capability. In order to analyze and interpret the statistically significant interaction effects (between diagnostic and interactive use), the next step involved the construction of two separate factorial designs (one for each respective statistically significant interaction effect). The differences in means and associated p-values emerged from Univariate General Linear Models and post hoc procedures using Fisher’s Least Significant Distance (LSD). Details about specific methodology and results appear in Appendix B. ANOVA results appear in Tables 5 and 6. The first issue is the ANOVAs’ overall significance. As described by the model R2 s, the models have significant explanatory power since the minimum R2 -adjusted was .365 (for Operational Capability). The next issue concerns the significance of the factors and interactions. The results demonstrate that for each of the baseline model paths: (1) Diagnostic Use as main effect was positive at p < .001; (2) Interactive Use as a main effect was positive at p < .001 (confirming the baseline model SEM results); and (3) their interaction was positive and statistically significant, at least at p < .002. Partial 2 for main or interaction effects in each table reveal that substantial variation for each dependent variable is explained. It is clear from Fig. 3a that at a low level of Diagnostic Use, a low level of Interactive Use produces the lowest mean score for Strategic Management Capability. At a low level of Diagnostic Use, the means produced by moderate and high Interactive Use are similar but substantively higher than for low Interactive Use. At a moderate level of Diagnostic Use, the means produced by moderate and high Interactive Use are still substantially higher than the mean produced by low Interactive Use but there is a divergence in the sense that high Interactive Use produces a mean score which is appreciatively higher than the mean score generated by moderate Interactive Use. The results for high Diagnostic Use are similar to those for moderate Diagnostic Use except that the mean produced by high Interactive Use is now substantially higher than the mean produced by moderate Interactive Use. Turning our attention on Interactive Use, at low levels of Interactive Use, higher levels of Diagnostic Use are associated with higher means of Strategic Management Capability. The slope is positive and the line appears to be rather linear. Similarly, at moderate levels of Interactive Use, higher levels of Diagnostic Use produce higher Strategic Management Capability means. At high levels of Interactive Use, the pattern changes significantly. Moving from low to moderate levels of Diagnostic Use engenders higher mean levels for strategic management capability but moving from moderate to high levels produces an extraordinary response where the mean value generated by high Diagnostic Use is substantially higher. It is apparent from Fig. 3a that the low and moderate lines for Interactive Use are practically parallel and relatively linear albeit that the moderate line is at a higher level. However, the high Interactive Use line is non-linear and the slope is substantially greater, especially when moving from a moderate to a high level of Interactive Use. In essence, the returns from the combination of high Interactive Use 326 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 5 Factorial tests for strategic management capability. Source Type III Sum of Squares df Mean Square F Sig. Partial Eta Squared Noncent. Parameter Observed Powerb Corrected Model Intercept Interactive Use Diagnostic Use Interactive Use* Diagnostic Use Error Total Corrected Total 3564.591a 35603.970 403.850 573.164 239.637 3549.339 98553.000 7113.930 8 1 2 2 4 377 386 385 445.574 35603.970 201.925 286.582 59.909 9.415 47.328 3781.746 21.448 30.440 6.363 .000 .000 .000 .000 .000 .501 .909 .102 .139 .063 378.620 3781.746 42.896 60.880 25.454 1.000 1.000 1.000 1.000 .990 a b R2 = .501 (Adjusted R2 = .490). Computed using alpha = .05. Fig. 3. (a) Estimated Marginal Means of strategic Management capacity, (b) Estimated Marginal Means of Operational capability. and high Diagnostic Use appear to be more like exponential than linear. The results for Operational Capability are similar to the findings for Strategic Management Capability in some respects but divergent in other. Fig. 3b does demonstrate that at higher levels of Diagnostic or Interactive Use the means for Operational Capability are always higher. At a low level of Diagnostic Use, high Interactive Use produces the highest mean for Operational Capability but it does not appear to be substantively different from the mean generated by moderate Interactive Use. At a moderate level of Diagnostic Use, the mean produced by high Interactive Use is still the highest but the mean produced by a moderate level of Interactive Use begins to converge to the mean produced by low Interactive Use. This phenomenon is more pronounced at a high level of Diagnostic Use where the means generated by moderate and low levels of Interactive Use converge while the mean for high Interactive Use diverges upwards significantly. The convergence implies that at high levels of Diagnostic Use, there is no difference between the mean scores for Operational Capability whether there is low or moderate level of Interactive Use. Focusing on the levels of Interactive Use, unlike the graph for Strategic Management Capability, the low and moderate lines are not parallel though both appear to be practically linear. The two lines converge when Diagnostic Use is high. However, the line associated with high Interactive Use is non-linear and the slope changes substantively above a moderate level of Diagnostic Use. Collectively, diverse combinations of Diagnostic and Interactive Use produce significantly different means for Strategic Management Capability and Operational Capability. In general, high levels of both Diagnostic and Interactive use engender high mean scores across both dependent variables. However, the combination of high Diagnostic Use and high Interactive Use produces mean scores which are significantly higher than all other combinations. These findings suggest that companies should be using PMS from both diagnostic and interactive logics. Furthermore, the positive slopes observed for all lines in Fig. 3 do imply that the effects on capabilities appear to be synergistic rather than compensatory. Practically, Table 6 Factorial tests for operational capability. Source Type III Sum of Squares df Mean Square F Sig. Partial Eta Squared Noncent. Parameter Observed Powerb Corrected Model Intercept Interactive Use Diagnostic Use Interactive Use* Diagnostic Use Error Total Corrected Total 1868.033a 31007.481 201.572 314.641 145.586 3076.394 82715.000 4944.427 8 1 2 2 4 377 386 385 233.504 31007.481 100.786 157.321 36.397 8.160 28.615 3799.845 12.351 19.279 4.460 .000 .000 .000 .000 .002 .378 .910 .061 .093 .045 228.920 3799.845 24.702 38.558 17.841 1.000 1.000 .996 1.000 .938 a. R2 = .378 (Adjusted R2 = .365), b. Computed using alpha = .05 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 if managers would like to attain the highest levels of capabilities, they should invest in both types of PM systems. 4. Longitudinal analysis To examine the longitudinal effects of PM systems on firm financial performance, we used several approaches. These approaches involved the use of panel data over a ten year time horizon. 4.1. Model 1 – specification Our first approach utilizes a panel data design using two time periods. The first time period is the pre-implementation period and is set as the year before the implementation of PM systems in order to avoid contamination from potential implementation activities. The second time period is set as the year 2010 which represents the year we collected our primary data through a survey. The respondents to the survey furnished detailed information regarding the number of years since their firm has been using a formal PM system. The intent of this approach is to examine the impact of the adoption of PM systems on the change in firm performance while controlling for time invariant factors and other variables. Towards this end, two competing models could be adopted: a fixed effects (FE) model or a random effects (RE) model. We opted for the random effects model because fixed effects models remove time invariant factors. Furthermore, the Hausman test which compares a FE model and a RE model was statistically insignificant, favoring a RE model for our analysis. 327 been using a PM system at the time of primary data collection (i.e., 2010). Research development expendituresit represents the expenditures for a given firm for research and development and is representative of the intangible resources possessed by the firm. Similarly, Industrial pattents intellectual property rightsit also represents investments in intangible resources. Plant machineryit and Industrial commercial machineryit reflect resources which are tangible in nature. The mean of the unobserved heterogeneity is a constant term labeled as “˛.” The unobserved random firm specific effects are labeled as ui while εit is the disturbance term. The analysis was performed with STATA12 using the xtreg (re) command. 4.2. Model 1 – results The approach we followed was iterative in nature. The model was first tested with all variables included and then iteratively identified and excluded variables which do not contribute statistically toward each of the respective dependent variables, i.e., the four financial ratios. Several control variables were retained though variable retention varies by model. Table 7 displays the coefficients for PM system adoption, the R-square values, the Wald’s 2 statistic values, and the p-values accompanied by their respective significance level for the final model we obtained. Wald’s test is an omnibus test and it is used to examine whether the variables collectively have a significant impact on each dependent variable. Based on Wald’s 2 we find Perfit = ˇ1 PMsystem diagostic useit + ˇ2 PMsystem interactive useit + ˇ3 PMsystem diagnostic useit ∗ PM system interactive useit + ˇ4 PM Years of system useit + ˇ5 firm sizei +ˇ6 Geog Diversificationi + j=16 ˇj indj + ˇ7 Research developement exp enditureit j=1 +ˇ8 Industrial pattents intellectual property rightsit +ˇ9 Plant machinery exp endituresit + +ˇ10 Industrial commercial equipmentit + ˛ + ui +εit The variable Perfit essentially represents the difference in financial performance of a firm i between two time periods. The first period is calculated as the year at which the firm begun using the formal PMS minus one. For instance, if a firm implemented and begun using the PMS reportedly in 2007, the first period is set as 2006. We set the second time as 2010, the year for which we collected primary/survey data for diagnostic and interactive types of usage. The financial performance is either measured in terms of changes in ROA, ROTA, ROE, or EBITM and thus four models were estimated. The variables included in the model consist of the manner in which PM systems are used (i.e., diagnostic use and interactive use) along with their interaction. The values for diagnostic and interactive use of the PM system for the second time period capture the extent and manner in which a PM system is deployed within each firm at the time of cross-sectional data collection. The values for the diagnostic and interactive use of the PM system are set at ‘0 in the first time period when firms had not yet implemented a PM system. Covariates also include controls that we identified as variables that can potentially influence our dependent variables. These include firm size (firm sizei , industry (indj ), and geographical ownership diversification (Geog Diversificationi ). Firm size is captured using a seven point Likert type scale, while the geographical diversification is depicted using a binary variable that denotes whether a firm has multinational presence or not. Specific industry affiliation was reported in the survey. Furthermore, we control for the duration of PM system use. PMYears of system useit assumes a value of 0 for the first period while for the second time period it is set as the number of years since the firm has that all variables, including the time invariant factors, jointly had an influence on each dependent variable (Wald’s 2 is statistically significant at least at 0.001).The results presented in Table 7 suggest that Diagnostic Use has a statistically significant negative impact on the change in ROA (p < .05) and the change in EBITM (p < .05) (notice that based on variable scaling for Diagnostic and Interactive Use, a positive coefficient is indicative of negative effects and a negative coefficient is reflective of positive effects), while the interactive use of PM systems has a positive moderate impact on the change in EBITM (p < .10). Dynamic Tension appears to have a strong positive impact on the change in ROA (p < 0.05) and ROE (p < 0.01) and a moderate positive impact on the change in ROTA (p < 0.10). The findings suggest that the Diagnostic and Interactive uses along with Dynamic Tension appear to explain changes in financial performance between the pre-implementation period and 2010. What the data suggests is rather interesting. Diagnostic use does contribute to changes in financial performance (for ROA and EBITM) but those companies reporting larger changes in diagnostic use between the two periods experience smaller changes in financial performance across the two periods. This does not necessarily intimate that diagnostic use has a negative effect on financial performance. It suggests instead that as companies add to their diagnostic use portfolio over time, they will experience smaller or less dramatic changes in ROA and EBITM. In essence, the effects on the two performance variables may still be positive but they get smaller in magnitude for firms that use diagnostic systems more intensely over time. On the other hand, companies that add to their interactive use over the years engender larger changes in EBITM. 328 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 7 Longitudinal effects –model 1. Dependent Variables Independent Variables Diagnostic Use Interactive Use Dynamic Tension Years of PMS Use Finance and Insurance Manufacturing Mining Other Services Real Estate, Rental & Leasing Geog. Diversification R&DEc IP&IPRd ICEe Constant Observations Groups Wald’s 2 R2 a b c d e f † * ** *** ROA .0148a (.029b )* −.021(.190) −.002(.014)* .041(.096)† ,f −.639(.063)† −.417(.006)** −1.654(.000)*** ROTA .004(.314) −.005(.416) −.001(.063)† ROE .008(.167) .024(.176) −.002(.015)* .616(.075)† −.402(.009)** −1.620(.000)*** −1.582(.000)*** −.276(.026)* -.370(.005)** .119(.068)† .151(.029)* .097(.097)† .504(.002)** 250 137 32.79(.000)*** .154 .606(.000)*** 250 137 28.86(.000)*** .144 .097(0.097)† 245 137 30.89(.000)*** .124 −1.202(.001)*** EBITM .014(.028)* −.034(.077)† −.000(.389) −.555(.068)† .556(.009)** 1.892(.000)*** −.157(.023)* .136(.035)* .261(.003)** −.040(.289) 237 131 46.76(0.000)*** .224 Coefficient. p-value. R&DE: Research & Development Expenditures. IP&IPR: Industrial Patents & Intellectual Property Rights. ICE: Industrial & Commercial Equipment. Only statistically significant effects at p < .10 are shown for control variables. p-value < 0.1. p-value < 0.05. p-value < 0.01. p-value < 0.001. Similarly, firms that escalate both diagnostic use and interaction use over time do experience larger changes in ROA, ROE, and ROTA in contrast to firms that exhibit lower levels of both diagnostic and interaction use. 4.3. Model 2 – specification Model 2 utilizes a panel data design but includes several observations per individual firm. Again, we opted for the RE model because the use of FE model tends to ignore time invariant factors. The technique followed here resembles the approach used by Benner and Veloso (2008), but we use a RE model instead of using a FE model. We use the following specification to further examine whether the use of a PM system does lead to better firm performance over time. PM system while after adoption they received values 1, 2, 3 . . . for each year until the final time period in our analysis (i.e., 2012). For example, if a particular firm adopted a PM system in 2010, it received a value of zero for the years 2003 until 2009, and subsequently it received values 1, 2, and 3 for the years 2010, 2011, and 2012 respectively. A similar approach was carried to capture ISO9000 adaptation trend by Benner and Veloso (2008). However, we recognize that learning may diminish over time and thus we tested the model also using three learning rates (80%, 60%, and 40%) in order to have more robust results and respective conclusions. The variables included in the model consist of PM system adaptation, and the covariates that we identified and specified in earlier analysis. We also included time dummies (Year dumk ) to control for the yearly shocks that can potentially influence the performances of all firms. We operationalize the mean of the unobserved Perfit = ˇ1 PMsystem adaptationit + ˇ2 Geog Diversificationi + ˇ3 firm sizei + j=16 k=10 j indj + j=1 k Year dumk k=1 +ˇ4 Research development exp endituresit +ˇ5 Industrial pattents intellectual property rightsit + ˇ6 Plant machinery exp endituresit +ˇ7 Industrial commercial equipmentit + ˛ + ui εit Due to learning effects, we theorize that firms benefit while adapting to a new system. The variable Perfit represents the financial performance of a firm i during time period t. The financial performance is either measured in terms of ROA, ROTA, ROE, or EBITM. Data from 2003–2012 are used in data analysis. Prior empirical work on windows of opportunity after process changes suggests that firms tend to adapt to a new system or technology over a period of time (Heim & Ketzenberg, 2011; Tyre & Orlikowski, 1994). Thus, we created a variable related to longitudinal PM system adoption which is used to measure the trend of PM system adaptation (PMsystem adaptationit ). This variable is built to capture the improvement in performance as firms adapt to the new system after implementing it. In our coding scheme, firms received a value of zero for PMsystem adaptationit until such time they adopted a heterogeneity as a constant term which we label as ␣, while ui captures the unobserved firm specific random effects and it is the disturbance term. The analysis was performed with STATA12 using the xtreg (re) command. 4.4. Model 2 – results Table 8 (which includes one panel for each measure of financial performance) produces the coefficients of PM system adaptation, R-square values, Wald’s 2 statistics, and the p-values accompanied by their respective significance level. Across all four dependent variables, the Wald’s 2 statistic is highly statistically significant for the basic model as well across learning rates, except of two occasions for ROA (i.e., at a learning rate of 60% and 40%). This suggests X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 329 Table 8a Longitudinal effects –model 2: Panel A (ROA). Trend Since Adoption Independent Variables Adaptation Geog. Diversification Firm Size Industry Dummiesd Time Dummies R&DE IP&IPR P&M ICE Constant Observations Groups Wald’s 2 R2 a b c d † * ** *** ROA (Trend = +1/year) .087a (.000b )*** .029(.376) .053(.027)* Included Included −.018(.220) .018(.279) .017(.309) −.049(.075)† −.769(.001)** 2361 315 50.88(.000)*** .067 ROA (Trend = 80% LRc ) .116(.000)*** .037(.346) .054(.024)* Included Included −.020(.191) .016(.298) .016(.319) −.047(.088)† −.685(.004)** 2361 315 45.04(.000)*** .065 ROA (Trend = 60% LR) .081(.020)* .065(.242) .058(.018)* Included Included −.022(.171) .014(.320) .017(.312) −.042(.111) −.362(.071)† 2361 315 34.77(.125) 0.054 ROA (Trend = 40% LR) .017(.351) .083(.182) .061(.014)* Included Included −.021(.177) .014(.328) .018(.297) −.041(.119) −.136(.276) 2361 315 30.61(.217) 0.042 Coefficient p-value LR = Learning Rate Industry and Time dummies were included in the model but not displayed here. p-value < 0.1. p-value < 0.05. p-value < 0.01. p-value < 0.001. that all variables, including time invariant factors and the time dummies, jointly had a strong influence on the dependent variables. The individual coefficients for industry and time dummies are not displayed in Table 8 for simplicity of exposition. We find that the adaptation variable has a positive influence on firm performance across all four performance measures for the basic model (see Table 8-all four panels). The lowest value for R2 is roughly 6% while the highest value is around 20% and both are rather substantive given the overabundance of variables than can impact financial performance over time. At an 80% learning rate, we observe similar results where all the coefficients are statistically significant for the four financial measures. At a 60% learning rate, the adaptation coefficient for ROE appears (see panel C) to be non-statistically significant. Nevertheless, the coefficients for ROA, ROTA, and EBITM remain statistically significant and although Rsquare values sustain some losses, they are still substantive. Finally, at a 40% learning rate, only the coefficients for ROTA and EBITM (see panels B and D) remain statistically significant. As expected the R-squares saw some decline. The data analysis suggests here that the expansive use of PM systems over time is indeed conducive towards firm performance. We conjecture that the longer an organization uses its PM system the more effectively it becomes at orchestrating its resources. However, we also demonstrated that the impressions of this adaptation over time are subject to learning rate effects. When higher learning rates are applied, the effects of adaptation become more salient towards performance. The practical consequences can be significant; if a firm does not maintain the momentum engendered by the implementation of a PM system and continue adapting over time at a high rate, it may see diminishing effects on performance. Thus, management is urged to continue investing in PM systems and advocate programs that can uphold high learning rates. Table 8b Longitudinal Effects–Model 2: Panel B (ROTA). Trend Since Adoption Independent Variables Adaptation Geog. Diversification Firm Size Industry Dummiesd Time Dummies R&DE IP&IPR P&M ICE Constant Observations Groups Wald’s 2 R2 a b c d † * ** *** ROTA (Trend = +1/year) .068a (.000b )*** .024(.399) .062(.013)* Included Included −.022(.187) −.002(.481) .022(.272) −.027(.227) −.526(.022)* 2336 315 45.95(.016)* .069 ROTA (Trend = 80% LRc ) .104(.000)*** .024(.379) .062(.013)* Included Included −.024(.169) −.003(.467) .021(.281) −.026(.241) −.524(.023)* 2336 315 45.16(.019)* .069 Coefficient p-value LR = Learning Rate Industry and Time dummies were included in the model but not displayed here. p-value < 0.1. p-value < 0.05. p-value < 0.01. p-value < 0.001. ROTA (Trend = 60% LR) .106(.005)** .040(.331) .064(.011)* Included Included −.025(.150) −.004(.447) .021(.280) −.022(.273) −.344(.085)† 2336 315 40.72(.046)* 0.063 ROTA (Trend = 40% LR) .079(.047)* .056(.271) .066(.009)** Included Included −.026(.145) −.005(.437) .022(.272) −.020(.293) −.162(.243) 2336 315 38.36(.091)† 0.054 330 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 8c Longitudinal Effects–Model 2: Panel C (ROE). Trend since adoption Independent Variables Adaptation Geog. Diversification Firm Size Industry Dummiesd Time Dummies R&DE IP&IPR P&M ICE Constant Observations Groups Wald’s 2 R2 a b c d † * ** *** ROE (Trend = +1/year) .030a (.057b )† −.059(.239) .042(.048)* Included Included −.027(.153) .056(.048)* .032(.190) −.089(.009)** −.113(.321) 2317 315 66.30(.000)*** .063 ROE (Trend = 80% LRc ) .038(.100)† −.056(.252) .043(.046)* Included Included −.028(.146) .055(.050)* .032(.190) −.088(.011)* −.071(.385) 2317 315 65.36(.000)*** .062 ROE (Trend = 60% LR) .029(.240) −.048(.283) .044(.042)* Included Included −.028(.140) .054(.053)† .033(.188) −.086(.011)* −.024(.459) 2317 315 64.33(.000)*** 0.059 ROE (Trend = 40% LR) .007(.442) −.041(.309) .044(.039)* Included Included −.028(.140) .054(.054)† .033(.184) −.085(.012)* .103(.318) 2317 315 63.92(.000)*** 0.057 Coefficient. p-value. LR = Learning Rate. Industry and Time dummies were included in the model but not displayed here. p-value < 0.1. p-value < 0.05. p-value < 0.01. p-value < 0.001. 4.5. Model 2 robustness check –specification We also performed a robustness check for Model 2 using a Propensity Score Matching (PSM) technique. Ideally, it would be optimal to assess whether the performance of the adopters would have been different if they had not adopted PM systems. However, it is not practically feasible to explicitly compare the performance of adopters after adoption and under the assumption that they had not adopted. One approach to overcome this challenge is to identify non-adopters that are similar to adopters based on certain characteristics, and compare the performance of the adopters to these similar firms that have not yet adopted PM systems. In essence, a Propensity Score Matching technique enables us to compare the treatment firms (i.e., early adopters of PM systems) and control groups (i.e., firms that did not adopt a PM system) which share similar characteristics. PSM allows us to evaluate firms based on a single propensity score, which here is determined based on industry (Corbett et al., 2005), firm size, the geographic diversification of firms (Peikes et al., 2008), as well as the four resource-based variables. Since adoption can be modelled as a dichotomous variable, we use a logit model shown below to determine the propensity score in STATA12. The propensity score for the observations in our sample is the probability estimate obtained from our logistic regression model. j=16 Logit(p(Adoptiont )) = ˛ + ˇ1 Geog Diversificationi + ˇ2 firm sizei + indj + j=1 j ˇ3Research development exp endituresit + ˇ4Industrial pattents intellectual property rightsit ˇ5plant machinery exp endituresit + ˇ6Industrial commercial equipment it + ε Table 8d Longitudinal effects –model 2: panel D (EBITM). Trend Since Adoption Independent Variables Adaptation Geog. Diversification Firm Size Industry Dummiesd Time Dummies R&DE IP&IPR P&M ICE Constant Observations Groups Wald’s 2 R2 a b c d † * ** *** EBITM (Trend = +1/year) .044a (.009b )** −.101(.124) .066(.007)** Included Included −.029(.093)† −.016(.297) .345(.000)*** −.021(.268) −.543(.013)* 2311 312 176.28(.000)*** .195 EBITM (Trend = 80% LRc ) .081a (.002b )** −.107(.112) .065(.007)** Included Included −.030(.086)† −.016(.295) .344(.000)*** −.021(.268) −.619(.006)** 2311 312 178.87(.000)*** .197 Coefficient. p-value. LR = Learning Rate. Industry and Time dummies were included in the model but not displayed here. p-value < 0.1. p-value < 0.05. p-value < 0.01. p-value < 0.001. EBITM (Trend = 60% LR) .108a (.002b )** −.101(.124) .066(.006)** Included Included −.032(.074)† −.017(.286) .344(.000)*** −.019(.289) −.557(.008)** 2311 312 179.38(.000)*** 0.195 EBITM (Trend = 40% LR) .111a (.004b )** −.090(.150) .068(.005)** Included Included −.033(.066)† −.018(.278) .345(.000)*** −.016(.311) −.434(.022)* 2311 312 178.12(.000)*** 0.191 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 9 Longitudinal effects –robustness check for model 2. Dependent Variable Adoption Type ROA ROTA ROE EBITM 5. Summary and discussion Performance 2012 PM System Adopters 175 175 175 175 PM System Non-Adopters 54 49 51 49 Average treatment effect for treated (ATT) (95% C.I) .607 (.143, .940) .377 (.038, .817) .466 (.022, .793) .251 (.001, .631) In our PSM method we employ the nearest neighbor matching technique which in turn uses the computed propensity score to match the treatment firms with the control group (Heckman et al., 1997). In essence, the interest is to examine the average effect of PM systems on the adopters by comparing them directly against the most comparable non-adopters. This effect is termed as the average treatment effect on the treated (ATT). The empirical form for estimating ATT in the case of the nearest neighbor approach is shown below. ATT = 1 n1 y1,i − y0,j 331 i ∈ {D=1) The outcome for the treated observation i which belongs to the treated group D is represented by y1,i while y0,j denotes the outcome for the matched observation j in the control group and n1 is the number of observations in the treated group. The matched observation from the control group is determined using the criteria min||pi -pj || where Pi is the propensity score for a treated observation and Pj is the propensity score for the non-treated observation. The nearest neighbor matching approach enables us to compute the effect of PM systems adoption on firm performance by controlling for the propensity of adoption between adopters and non-adopters. We choose 2005 to be the year we splice the data to examine the difference between adopters and non-adopters of PM systems. The choice of the specific year was governed by the principle that we need a sufficiently large sample of firms that had not yet adopted a PM system in order to identify a suitable control group while retaining a sufficiently large number of firms that had adopted a PM system (>50%). We carried out PSM using the nearest neighbor method on outcome variables in 2012 (Guo & Fraser, 2010). Recall that we demonstrated earlier that performance improves with adaptation, rendering support for performance improvement over time across the whole study sample. We label firms that have adopted PM systems by 2005 as early adopters. 4.6. Model 2 robustness check results Table 9 provides the results for the average treatment effect on the treated group. The significance of the results is assessed based on the 95% confidence interval (C.I) values for the ATT obtained by bootstrapping the standard errors. We find significant evidence to suggest that the early adopters of PM systems have better financial performance when compared to the control group. We specifically find that treated firms, when compared to others, have a significantly higher performance measured by all financial measures. We find that this result corroborates our findings presented earlier where we conclude that the adaptation variable has a significant and positive impact on performance. The data analysis furnishes additional evidence to suggest that longer usage of PM systems, ceteris paribus, is rather conducive for firm performance. Over the last two decades, firms have invested heavily in implementing PM systems under the premise that such systems will deliver financially. Practitioners, consultants, societies, and academics alike ascribe positive evaluations to PM systems and tout the benefits. However, there is scant empirical evidence to validate such claims. First, the literature is devoid of a comprehensive framework that identifies specific uses of PM systems and examines their respective relationships to organizational capabilities and firm performance. In essence, the extant literature lacks a theoretically based comprehensive nomological network. Second, there is no material evidence attesting that PM systems do actually contribute positively to financial performance over time. We summarize our key findings and recommendations for future research in Table 10. As far the extant literature is concerned, our empirical findings are rather interesting. First, we found that interactive use is a positive force just as Henri (2006a) articulated. However, diagnostic use is also contributing positively towards organizational capabilities and performance. This was manifested by the instrumental role of diagnostic PM system use in impacting positively all three capabilities examined in the cross-sectional study. In fact, diagnostic use exhibited the most potent effects on capabilities from both a statistical and a substantive perspective. Although we did not hypothesize and test for indirect effects, there is significant evidence here to suggest that diagnostic use is also indirectly related with the subjective measure of financial performance (p < .000) as well as with the four objective measures of financial performance (p-value for any indirect coefficient at least < 031). Furthermore, the positive role of diagnostic use can be surmised when the impact of dynamic tension is considered, as it is generated from the joint forces of interactive and diagnostic use. Dynamic tension was crucial for explaining variability in strategic management capability and operational capability. Also, from a longitudinal perspective, dynamic tension demonstrated statistically significant positive effects on three out of four measures of change in financial performance. On the aggregate, the evidence contradicts Henri’s (2006a) claim that diagnostic use acts as negative force. As explained earlier in the results section however, Henri’s operationalization of diagnostic use and choice of capabilities may have contributed to the disparity between our findings. Also, it is noteworthy to point that the correlations between diagnostic use, capabilities, and performance, as reported in Henri (2006a, p. 541), were all positive, albeit one correlation with a capability was not statistically significant. In order to elicit a more thorough interpretation regarding the use of PM systems, we conducted retrospective interviews with four managers that participated in the cross-sectional study. All managers suggested that it would be most beneficial if a PMS is used from both a diagnostic and interactive logic concurrently. In one of the companies where diagnostic use dominated, a manager labeled his top management as “authoritarian” and “rigid” based on how it deployed diagnostic use. He also suggested that diagnostic use is akin to a “corporate thermometer.” But, while top management could see the temperature rising, it could not interact effectively with constituents to find out what was going awry. Top management communicated with functional managers via a top–down approach but this communication had the ulterior motive of control. Lower level managers were totally cut-off from information exchange and thus a lot of opportunities for identifying the root causes of challenges were relinquished. At another company, a manager pointed that diagnostic use and interactive use work best when deployed in unison. For example, she noted that actions undertaken by management via interactive use can only be evaluated for their soundness by an ex post assessment afforded by the diagnostic system. Similarly, she stated that 332 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Table 10 Going forward: new research questions and implications. Key Finding/Limitation Diagnostic Use, Interactive Use, & Dynamic Tension were found to impact a set of competitive capabilities positively. Diagnostic Use exhibited the most potent effects in the cross-sectional study. Diagnostic Use demonstrated diminishing financial returns over time when examined via a longitudinal study. PM system use does improve financial performance over time but learning rates do matter. How firms adapt over time is important. We did not examine the behavior of specific types of PM use over time; rather we assessed whether the presence or absence of PM systems makes a difference in financial performance overall. We examined the impact of PM systems through the theoretical lenses of the Resource Orchestration Theory. We found that resource orchestration, after accounting for resource possession, does matter financially. We did not examine how PM system uses affect specific resource orchestration nuances: structuring the resource portfolio, bundling resources, and leveraging resources. We studied the impact of PM system uses on capabilities and performance variables. These outcomes pertain to the measurement of capabilities and performance of a firm. Given the level of inter-firm relationships, it would be fruitful to examine the saliency of specific PM system uses on supply chain capabilities and performance. Our financial performance measures were inclusive of cost considerations but we did not specifically examine the implementation costs of PM systems and subsequent usage costs. We produced evidence suggesting that diverse uses of PM systems engender positive performance. However, it is possible that PM systems also produce negative effects which went unaccounted here. Based on our interviews with practitioners, we suspect that mere use of Diagnostic Use will produce negative outcomes. We studied organizational level effects; we did not examine the impact of different types of PM system uses on the individual. It is not known whether PM system use impacts for instance the motivation or creativity of individuals. The use of PM systems may impact individual and organizational attributes. We do not know whether the effects on the individual can explain more or less variance in performance than the effects on organizational attributes. Also, we do not know whether the effects on the individual and organization vary by the cultural environment where the firm operates. We studied firms operating in Italy. Although a number of multinationals were included, we do not know whether the findings stand in light of differences in cultural environments. Research Question Discussion & Potential Contribution Diagnostic versus Interactive PM system use What is the idiosyncratic contribution of the two types of use and dynamic tension on capabilities and performance over time? Is the effect of Diagnostic Use significantly more potent? What are the effects on capabilities and performance metrics which are most salient to operations and supply chain management? Do firms that merely rely on Diagnostic Use perform well across a variety of capabilities and performance measures? Are the effects of Diagnostic Use and Interactive Use sustainable over time? What activities can the organizations engaged in to maintain high learning rates? PM Systems & resource orchestration Do the effects of diverse PM system uses vary across the three resource orchestration nuances? Which specific PM system uses are more salient for each nuance? What is the relative efficacy of resource orchestration vis-à-vis resource possession in explaining performance over time? Is the interaction between resource orchestration and resource possession statistically significant? Do the different types of PM system use foster resource orchestration at the supply chain level? PM system implementation costs & adverse effects Do the costs of implementation and usage vary by type of PM system use? What type of negative effects does PM system use produce? Are the effects of Diagnostic Uses and Interactive Uses equally salient when negative effects are considered? Effects on the organization vis-à-vis the individual What is the impact of the different types of PM system uses on the individual attributes (e.g., motivation, creativity)? Do the effects on the individual explain more or less variance in performance than the effects on organizational attributes? Do the effects on the individual and organization vary by the cultural environment where the firm operates? Generalizability across cultural environments Do the relationships studied here vary by cultural environments? diagnostic use can be seen as a mere futile exercise if actions via interactive use are not carried out. Considering the response from the first manager, the organization may observe the temperature rising on the “thermometer” but if no action is initiated to Organizations possess limited resources and these resources need to be orchestrated as effectively as possible. Firms need to focus on systems that can produce the best returns for the money and effort. Examining the individual and collective role of each type of PM system use over time can help organizations properly apportion resources across diagnostic and interactive uses of PM systems. The organizations also need to know where specifically to invest their resources in order to retain high learning rates, which were found to be crucial for sustaining high levels of financial performance. Supply chain management professionals and scholars also need to know whether PM systems matter when respective metrics are considered. Resource orchestration necessarily involves a lot of activities. Some deal for instance with the acquisition of resources while others involve the mobilization of resources. Organizations need to know which type of a PM system use is most efficacious for each activity and invest fittingly. Furthermore, organizations need to know whether it is best to accumulate resources or exploit existing resources or whether there needs to be a balance in budget allocation across the two. Also, considering the level of outsourcing that exists today, organizations need to know whether their specific PM systems and respective uses can actually help them orchestrate resources across the value chain. If adjustments are needed, they need to know where those are needed. Before organizations invest in PM systems, they need to have an accurate estimate about the costs of implementation and usage. They also need to know the relative costs generated by diagnostic vs interactive PM systems so they can make better investments. Although there are some conjectures and cursory evidence, we do not know much about the negative consequences prompted by the use of PM systems and which type of use is most problematic (and thus needs to be addressed). PM system uses may impact organizational attributes but it is likely that they have a pervasive impact on the individual attributes as well. Organizations need to know what individual attributes get affected and which specific type of PM system use stimulates it. In order to invest wisely, organizations need to know also whether it is best to promote capabilities at the individual level, organizational level, or both. Furthermore, organizations need to know whether the effects are sustainable across different environments so they can make proper adjustements. Organizations need to know whether the saliency of specific types of PM system use is invariant across environments. They need to identify which type of PM system use works best for each environment. uncover the true causes of the surge in temperature, the organization’s performance may be at risk. Based on our analysis and our interviews, it appears that diagnostic use is indispensable but management should be warned that if a PM system is X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 used solely from a diagnostic use motive, performance may be at peril. Second, the extant operations management or accounting literatures have produced no longitudinal studies regarding the effects of PM systems that involve multiple companies over a considerable time. We initially utilized a panel data approach and specified a random effects model to examine and contrast financial performance across two different periods (one of the periods was pre-implementation and one after). We found that some of the rates of change in performance are positively related with interactive use and dynamic tension while some of the rates of change in performance are negatively related with diagnostic use, suggesting perhaps diminishing returns. We also examined a learning effect due to adaptation by applying a variety of learning rates over a ten year period. There is strong evidence to suggest that the effects of adaptation on all four financial measures of performance are positive across diverse learning rates. However, the effects of adaptation on performance wane down when lower learning rates are applied. Finally, a robustness check compared PM system early adopters and PM system non-adopters using a propensity score matching approach and produced evidence to suggest that PM system early adopters outperform PM system late and non-adopters across all four financial measures. Admittedly the Propensity Score Method is rather demanding because of the difficulty of identifying “similar” firms, yet it still identified that early adopters perform the best as far as financial performance is concerned. When we view all the results jointly, we can assert that PM systems do positively impact performance but organizations need to maintain high levels of learning. From a theoretical perspective, ROT suggests that organizations need not only amass resources but they also need to orchestrate those resources in order to garner competitive advantage (Hitt et al., 2011; Sirmon et al., 2011). We position PM systems as mechanisms that can help organizations manage and synchronize resource orchestration (Hitt et al., 2011). Effective synchronization demands information regarding the firm’s environment and a framework that can be used to process that information. A PM system, through its capacity as an information and control system, can facilitate top management to evaluate prior actions and craft a new strategy through the acquisition, bundling, and mobilization of resources. Hitt et al. (2011) along with Sirmon et al. (2007,2011) are credited with the advancement of the ROT tenets. They claim that the primary actions needed for resource orchestration include structuring the resource portfolio (i.e., acquiring, accumulating, and divesting resources), bundling resources (i.e., stabilizing existing capabilities, enriching current capabilities, and pioneering new capabilities) and leveraging resources (i.e., mobilizing capabilities to form capability configurations, coordinating capability configurations, and deploying the configurations). Although we conceptually linked resource orchestration at large with PM systems, a more in depth study is needed to tie the three specific nuances of resource orchestration with PM systems. In other words, further research is needed to understand how a PM system facilitates the constituent elements of structuring, bundling, and leveraging of resources. For instance, how does diagnostic use of a PM system affect mobilization? Sirmon et al. (2007), resting on Hamel and Prahalad (1994), assert that “The intent of mobilizing is to identify the capabilities needed and to design the capability configurations necessary to exploit opportunities in the market and gain competitive advantage” (p. 284). Diagnostic use of PM systems may be more conducive towards the identification of capabilities needed through its monitoring attributes, which furnish information regarding deviations from targets. On the other hand, design of capabilities may be facilitated more effectively by the interactive use of PM systems because the design of capabilities necessarily involves top management involvement. Similarly, diagnostic use of PM systems may 333 facilitate the enrichment of current capabilities while interactive use may be more salient for pioneering new capabilities. We expect that the different types of PM system uses and the tension they create will have pervasive effects on the three primary resource orchestration actions but their resonance may vary. Our study adds to the body of literature by providing empirical evidence attesting to the value of resource orchestration. We used four measures of resource endowments (both tangible and intangible) as control variables in our empirical inquiry in order to isolate the effects of resource orchestration from the mere possession of resources. In this respect, we provide empirical support attesting to the efficacy of resource orchestration to explain performance. From both a theoretical and an empirical perspective, it would be interesting to examine the relative role of resource possession vis-à-vis resource orchestration. Their relative contribution towards performance can be examined via variance partitioning. While both may be valuable however, we believe that the best performance level can be attained when valuable resources are used effectively. In other words, we posit that the interaction between resource possession and resource orchestration can explain performance variance more effectively. Organizations cannot orchestrate resources they do not have or control, and possessing resources without leveraging them can be futile. The operations management literature and the supply chain management literature have relied extensively on the resourcebased theory but there is scant research that examines the orchestration of resources. The importance of PM systems is taken for granted but more is needed to ascertain how specific PMS uses can be deployed to orchestrate resources and add on firm capabilities. A recent exception is Ketchen et al. (2014) who examine resource orchestration in the context of product returns. Given, however, that firms now also rely on the resources bestowed by their supply chain partners, it would be useful to examine how PM systems facilitate the structuring, bundling, and mobilization of resources across organizational boundaries. We are arguing in essence that resource orchestration has to be considered in light of supply chains that span multiple firms and multiple locations. Although diagnostic PM system use can furnish valuable information pertaining to the organization, its efficacy may be curtailed as far as the supply chain is concerned because the organization may lack access to network-level information. Such information may reside with suppliers and customers across the supply chain. Obtaining relevant information and orchestrating action across the network may require top management involvement. Top management can communicate with its counterparts to obtain information, stimulate the development of new ideas, and deliberate on possible courses of action. In essence, interactive PM system use may be more productive than diagnostic use. Future research could also benefit from measuring the costs of implementation and maintenance of PM systems and use these costs to evaluate whether the benefits outweigh the costs. Our measures of financial performance were net of costs but it would be instrumental for both managers and academics to ascertain what it specifically costs to implement and run PM systems. Widener (2007) used management attention as a surrogate for cost and hypothesized that diagnostic control systems make more efficient use of management attention vis-à-vis interactive control systems. Based on responses from 122 Chief Financial Officers, Widener (2007) found support for her hypotheses. Widener explains that challenges or concerns that are more routine in nature have smaller information deficits and thus can be handled efficiently via diagnostic systems, without much intervention from management. On the other hand, Widener credits the additional cost for interactive systems on the larger information deficits and more equivocal tasks interactive systems are purported to tackle. Since interactive systems by definition personally involve individuals at the top 334 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 managerial ranks, they can devour managerial attention. While diagnostic systems may not demand constant managerial attention, there is a compelling need however to invest in tangible and intangible assets that foster PM diagnostic use; i.e., monitoring, focusing attention, and legitimization. For instance, the organization may need to invest in information/surveillance systems (hardware, software, and human capital) that can furnish information to monitor critical success factors. As Simons (1994) states, “Any diagnostic control system can be made interactive by continuing and frequent top management attention and interest” (p. 171). Once the diagnostic system is in place, managers can use it interactively. We assert that the implementation cost may be higher for diagnostic systems than interactive systems but this is an empirical question that future research can address. It would be worthwhile to evaluate whether the costs of the two PM system uses do vary and identify which one is the costlier. The conception of cost has to be more encompassing however and include elements beyond managerial attention. Future research should also account for the potential negative effects that may emerge from using PM systems. Some individuals envision PM systems as bureaucratic, time consuming, and unnecessary. Voelpel et al. (2006) suggest that Balance Score Cards (BSCs), which represent an exemplar of PM systems, are rigid and internally focused and are credited with static-ism. These “tyrannical” effects, as Voelpel et al. coin them, may have adverse effects on innovation. However, Voelpel et al. (2006) were assessing the role of BSCs from a pure diagnostic PM system use perspective. Future research should include variables connoting negative attributions, such as rigidity and static-ism, as dependent variables to measure and account for the role of diagnostic and interactive PM system uses and dynamic tension as we evaluate performance. We suspect that if an organization purely uses PM systems from a diagnostic use perspective the effects on innovation may be indeed negative, while the effects of interactive use and dynamic tension may be positive. PM systems can also be seen as translating a firm’s strategy to its employees (Burney & Widener, 2013). Thus, the impact on the individual actors should also be examined. For instance, how does the use of PM systems impact employee motivation, effort, and performance? Burney and Widener (2013) furnish empirical evidence suggesting individual level effects but more research is due. Specifically, Burney and Widener (2013) used responses from 242 financial services employees to demonstrate that the extent to which a strategic performance measurement system is coupled with strategy impacts employee performance via perceived self-efficacy and perceived psychological contract while controlling for the effects of organizational citizenship behavior. Burney and Widener relied on data from a single company, which generates generalizability concerns, and they focused only on a few pertinent individual level variables. Furthermore, they did not study the individual or collective role of PM system use (i.e., diagnostic vs interactive vs dynamic tension). It would be rather conducive to assess the effects of PM system use on the individual by examining such effects across many organizations and cultures as well as including more variables. For instance, we do not know how diagnostic PM use or interactive PM use or their collective use impact creativity at the individual level. Our conjecture is that mere diagnostic use will stifle creativity as it suppresses experimentation and elicits performance which is aligned to pre-established targets. Furthermore, we do not know whether individual (employee) performance, fuelled by PM system use, does impact capabilities and firm performance. Rothaermel and Hess (2007) studied the effects of individual, firm, and network level variables on innovation output in the pharmaceutical industry and hypothesized direct effects as well as two competing interaction hypotheses (i.e., substitute versus complementary effects). They found that a significant amount of variance in innovation was explained by individual-level factors but they also uncovered that individual-level variables can serve as substitutes for firm- and network-level antecedents to innovation. Within the realm of PM system use, there are no studies that examine the relative or combined effect of individual and firm level effects on organizational performance. While we expect individual and firm level factors to contribute to organizational performance, it is likely that the effects are complementary. Creativity and motivation reside with the individual and firms rely on human capital to build their capabilities. At the same time, individuals rely on the organizational capabilities to perform more effectively. Thus, we postulate that individual (employee) performance and organizational level capabilities interact in a complementary fashion to impact firm performance. Empirical research is however necessary to address the research question. The generalizability of our findings is somewhat limited in the sense that data was obtained only from firms operating in Italy, although a sizable number of multinationals were included. Both the sociology and international business literatures have examined cross-regional differences and their potential effects on organizational life (Deal and Kennedy, 1982; Hofstede, 1980). Performance measurement systems are not immune to these effects; regional differences in culture, political systems, and economic development can have pervasive effects on the organization (Rhody & Tang, 1995). We suspect that three cultural variables, power distance (where authority is stratified and concentrated), collectivism (promoting group goals over individual goals), and uncertainty avoidance (comfort with ambiguity and change), exert influence on the relationships between PM types of use and capabilities. Our conjecture is that these variables may serve as moderators. Cultural attributes derive from both formal and informal societal structures and are reflective of belief and boundary systems. Belief and boundary systems may act as catalysts or as hurtles when organizations use PM systems (Simons, 1994). Future research should sample from other countries and respective geographical regions in order to identify whether the effects of specific PM system uses vary across cultural dimensions. Appendix A. Appendix A Examination of Competing Measurement Model Configurations In order to examine the efficacy of a second-order model specification to explain variation for Diagnostic Use, a series of competing models were examined. The first model specified one first-order latent variable of Diagnostic Use, the second model specified three uncorrelated first-order latent variables (i.e., monitoring, focus attention, and legitimization) while the third model specified three correlated first-order latent variables. Finally, the last model specified one second-order latent variable with three first-order latent variables. Iteratively, the models are compared examining whether there is a statistically significant difference in 2 values, where lower 2 values are favored. Model 1 produced a 2 = 2470.76 (1094 df) while Model 2 had an 2 = 2402.00 (1082 df). The diff 2 is 68.76 and is statistically significant (P < .01) based on 12 df. Model 3 produced a significantly lower 2 = 1964.489 (1079 df) than Model 2 and the difference of 437.511 (3 df) is statistically significant (p < .01). Finally, Model 4 generated a 2 = 1985.576 (1090 df) and diff 2 = 21.087 (21 df), which is not statistically significant, suggesting model equivalence based on model fit. In fact, a comparison of CFI and NNFI illustrates practically identical model fit between Models 3 and 4. Statistically, first-order correlated models, such as Model 3, will always generate a better model fit than secondorder models, such as Model 4, but a second-order model with comparative model fit offers an attractive alternative when theory can support such model. Given the underlying characteristics X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 of a diagnostic performance measurement system, consideration for parsimony, and based on statistical findings, the best model to represent Diagnostic Use appears to be at the second-order level of abstraction. Appendix B. Appendix B Interaction Effects Explanations For interpretability, the main effect constructs were recoded based on quartiles as follows: (1) Interactive Use into “Low” (below 25% percentile) if Interactive Use < = 10 (N = 122), “Moderate” (between 25% and 75% percentiles) if Interactive Use = 10.01–15.00 (N = 202), and “High” (above 75% percentile) if Interactive Use > = 15.01 (N = 62); and (2) Diagnostic Use into “Low” (below 25% percentile) if Diagnostic Use < = 44 (N = 101), “Moderate” (between 25% and 75% percentiles) if Diagnostic Use = 44.01–59.00 (N = 203), and “High” (above 75% percentile) if Diagnostic Use > = 59.01 (N = 82). Using Factorial ANOVA, the main and interaction effects were examined using p-values and partial 2 (analogous to R2 , reflecting explanatory strength). Fig. 3a and b display three lines representing Low/ Moderate/High Interactive use where Diagnostic Use is represented on the X-axis. The Y-axes represent Strategic Management Capability and Operational Capability respectively. The graphs can be interpreted in two ways: (1) given a particular level of Diagnostic Use, do the Strategic Management Capability (or the Operational Capability) means differ depending on the level of Interactive Use? The answers to these types of questions can be visualized in the graph by choosing a level of Diagnostic Use (e.g., low on the x-axis) and comparing the three Interactive Use means graphed at that level of Diagnostic Use. (2) The second way to interpret the results answers the question: given a particular level of Interactive Use, do the Strategic Management Capability or the Operational Capability means differ depending on the level of Diagnostic Use? For example, given low Interactive Use, do the Strategic Management Capability means differ depending on whether Diagnostic Use is low vs. moderate? The answers to these types of questions can be visualized in the graph by choosing a level of Interactive Use (ex., the low Interactive Use line) and comparing the three Diagnostic Use means graphed (e.g., Fig. 3a) at the corresponding three levels of Diagnostic Use (low vs. moderate vs. high on the Diagnostic Use x-axis). References Agle, B.R., Mitchell, R.K., Sonnenfeld, J.A., 1999. What matters to CEOs? An investigation of stakeholder attributes and salience, corporate performance and CEO values. Acad. Manage. J. 42 (5), 507–525. Ahn, H., 2001. Applying the balanced scorecard concept: an experience report. Long Range Plann. 34 (4), 441–461. Ahrens, T., Chapman, C.S., 2004. Accounting for flexibility and efficiency: a field study of management control systems in a restaurant chain. Contemp. Account. Res. 21 (2), 271–301. Amit, R., Schoemaker, P.J., 1993. Strategic assets and organizational rent. Strategic Manage. J. 14 (1), 33–46. Baldauf, A., Reisinger, H., Moncrief, W.C., 1999. Examining motivations to refuse in industrial mail surveys. J. Market Res. Soc. 41 (3), 345–353. Benner, M.J., Veloso, F.M., 2008. ISO 9000 practices and financial performance: a technology coherence perspective. J. Oper. Manage. 26 (5), 611–629. Berman, S.L., Wicks, A.C., Kotha, S., Jones, T.M., 1999. Does stakeholder orientation matter? The relationship between stakeholder management models and firm financial performance. Acad. Manage. J. 42 (5), 488–506. Birnberg, J.G., Turopolec, L., Young, S.M., 1983. The organizational context of accounting. Account. Org. Soc. 8 (2), 111–129. Bisbe, J., Batista-Foguet, J.M., Chenhall, R., 2007. Defining management accounting constructs: a methodological note on the risks of conceptual misspecification. Account. Org. Soc. 32 (7), 789–820. Bisbe, J., Malagueño, R., 2012. Using strategic performance measurement systems for strategy formulation: does it work in dynamic environments? Manage. Account. Res. 23 (4), 296–311. Bisbe, J., Otley, D., 2004. The effects of the interactive use of management control systems on product innovation. Account. Org. Soci. 29 (8), 709–737. 335 Bititci, U.S., Carrie, A.S., McDevitt, L., 1997. Integrated performance measurement systems: a development guide. Int. J. Oper. Prod. Manage. 17 (5), 522–534. Bititci, U.S., Turner, T., 2000. Dynamics of performance measurement systems. Int. J. Oper. Prod. Manage. 20 (6), 692–704. Bourne, M., Mills, J., Wilcox, M., Neely, A., Platts, K., 2000. Designing, implementing and updating performance measurement systems. Int. J. Oper. Prod. Manage. 20 (7), 754–771. Bourne, M., Pavlov, A., Franco, M., Lucianetti, L., Mura, M., 2013. Generating organisational performance: the contributing effects of performance measurement and human resource management practices. Int. J. Oper. Prod. Manage. 33 (11/12), 1599–1622. Braam, G.J., Nijssen, E.J., 2004. Performance effects of using the balanced scorecard: a note on the Dutch experience. Long Range Plann. 37 (4), 335–349. Bullinger, H.J., Kuhner, M., Van Hoof, A., 2002. Analyzing supply chain peformance using a balanced measurement method. Int. J. Prod. Res. 40 (15), 3533–3543. Burney, L., Widener, S.K., 2013. Behavioral work outcomes of a strategic performance measurement system-based incentive plan. Behav. Res. Account. 25 (2), 115–143. Chenhall, R.H., 2005. Integrative strategic performance measurement systems, strategic alignment of manufacturing, learning and strategic outcomes: an exploratory study. Account. Org. Soc. 30 (5), 395–422. Chenhall, R.H., Langfield-Smith, K., 1998. The relationship between strategic priorities, management techniques and management accounting: an empirical investigation using a systems approach. Account. Org. Soc. 23 (3), 243–264. Chenhall, R.H., Morris, D., 1995. Organic decision and communication processes and management accounting systems in entrepreneurial and conservative business organizations. Omega 23 (5), 485–497. Chia, A., Goh, M., Hum, S.H., 2009. Performance measurement in supply chain entities: balanced scorecard perspective. Benchmark Int. J. 16 (5), 605–620. Chirico, F., Sirmon, D.G., Sciascia, S., Mazzola, P., 2011. Resource orchestration in family firms: investigating how entrepreneurial orientation, generational involvement, and participative strategy affect performance. Strat. Entr. J. 5 (4), 307–326. Choi, J.W., Hecht, G.W., Tayler, W.B., 2013. Strategy selection, surrogation, and strategic performance measurement systems. J. Account. Res. 51 (1), 105–133. Corbett, C.J., Montes-Sancho, M.J., Kirsch, D.A., 2005. The financial impact of ISO 9000 certification in the United States: an empirical analysis. Manage. Sci. 51 (7), 1046–1059. Cousins, P.D., Lawson, B., Squire, B., 2008. Performance measurement in strategic buyer-supplier relationships: the mediating role of socialization mechanisms. Int. J. Oper. Prod. Manage. 28 (3), 238–258. Crabtree, A.D., DeBusk, G.K., 2008. The effects of adopting the balanced scorecard on shareholder returns. Adv. Account. 24 (1), 8–15. Craighead, C.W., Ketchen, D., Dunn, K.S., Hult, G., 2011. Addressing common method variance: guidelines for survey research on information technology, operations, and supply chain management. Eng. Manage. IEEE Trans. 58 (3), 578–588. Crook, T.R., Ketchen, D.J., Combs, J.G., Todd, S.Y., 2008. Strategic resources and performance: a meta- analysis. Strategic Manage. J. 29 (11), 1141–1154. Daft, R.L., Lengel, R.H., 1986. Organizational information requirements, media richness and structural design. Manage. Sci. 32 (5), 554–571. Davis, S., Albright, T., 2004. An investigation of the effect of balanced scorecard implementation on financial performance. Manage. Account. Res. 15 (2), 135–153. Deckop, J.R., Merriman, K.K., Gupta, S., 2006. The effects of CEO pay structure on corporate social performance. J. Manage. 32 (3), 329–342. Deal, T.E., Kennedy, A.A., 1982. Corporate Cultures. Reading MA:. Addison-Wesley. de Leeuw, S., van den Berg, J.P., 2011. Improving operational performance by influencing shopfloor behavior via performance management practices. J. Oper. Manage. 29 (3), 224–235. Donaldson, T., Preston, L.E., 1995. The stakeholder theory of the corporation: concept, evidence, and implications. Acad. Manage. Rev. 20 (1), 65–91. Estampe, D., Lamouri, S., Paris, J.L., Brahim-Djelloul, S., 2013. A framework for analysing supply chain performance evaluation models. Int. J. Prod. Econ. 142 (2), 247–258. Fitzgerald, L., Brignall, S., Silvestro, R., Voss, C., Robert, J., 1991. Performance Measurement in Service Businesses: Chartered Institute of Management Accountants London. Franco-Santos, M., Lucianetti, L., Bourne, M., 2012. Contemporary performance measurement systems: a review of their consequences and a framework for research. Manage. Account. Res. 23 (2), 79–119. Freeman, R.E., 1984. Strategic Management: A Stakeholder Approach. Pitman, Boston. Galbraith, J.R., 1974. Organization design: an information processing view. Interfaces 4 (3), 28–36. Garengo, P., Sharma, M.K., 2014. Performance measurement system contingency factors: a cross analysis of Italian and Indian SMEs. Prod. Plan. Control 25 (3), 220–240. Grafton, J., Lillis, A.M., Widener, S.K., 2010. The role of performance measurement and evaluation in building organizational capabilities and performance. Account. Org. Soc. 35 (7), 689–706. Grosswiele, L., Röglinger, M., Friedl, B., 2012. A decision framework for the consolidation of performance measurement systems. Decis. Support Syst. 54 (2), 1016–1029. Gunasekaran, A., Patel, C., Tirtiroglu, E., 2001. Performance measures and metrics in a supply chain environment. Int. J. Oper. Prod. Manage. 21 (1/2), 71–87. 336 X. Koufteros et al. / Journal of Operations Management 32 (2014) 313–336 Gunasekaran, A., Patel, C., McGaughey, R.E., 2004. A framework for supply chain performance measurement. Int. J. Prod. Econ. 87 (3), 333–347. Gunasekaran, A., Kobu, B., 2007. Performance measures and metrics in logistics and supply chain management: a review of recent literature (1995–2004) for research and applications. Int. J. Prod. Res. 45 (12), 2819–2840. Guo, S., Fraser, M.W., 2010. Propensity score analysis. Stat. Methods Appl.. Hamel, G., Prahalad, C.K., 1994. Competing for the Future: Breakthrough Strategies for Seizing Control of Your Industry and Creating the Markets of Tomorrow. Harvard Business School Press. Hamilton, S., Chervany, N.L., 1981. Evaluating information system effectiveness-Part I: Comparing evaluation approaches. MIS Quart. 5 (3), 55–69. Hansen, M.H., Perry, L.T., Reese, C.S., 2004. A Bayesian operationalization of the resource-based view. Strategic Manage. J. 25 (13), 1279–1295. Harrison, D.A., McLaughlin, M.E., Coalter, T.M., 1996. Context, cognition, and common method variance: psychometric and verbal protocol evidence. Organ. Behav. Hum. Dec. 68 (3), 246–261. Heckman, J.J., Ichimura, H., Todd, P.E., 1997. Matching as an econometric evaluation estimator: Evidence from evaluating a job training programme. Rev. Econ. Stud. 64 (4), 605–654. Heim, G.R., Ketzenberg, M.E., 2011. Learning and relearning effects with innovative service designs: an empirical analysis of top golf courses. J. Oper. Manage. 29 (5), 449–461. Helfat, C.E., Finkelstein, S., Mitchell, W., Peteraf, M., Singh, H., Teece, D., Winter, S.G., 2007. Dynamic Capabilities: Understanding Strategic Change in Organizations. Malden. MA: Blackwell. Henri, J.F., 2006a. Management control systems and strategy: A resource-based perspective. Account. Org. Soc. 31 (6), 529–558. Henri, J.-F., 2006b. Organizational culture and performance measurement systems. Account. Org. Soc. 31 (1), 77–103. Hitt, M.A., Ireland, D.R., Sirmon, D.G., Trahms, C.A., 2011. Strategic entrepreneurship: creating value for individuals, organizations, and society. Acad. Manage. Perspect. 25 (1), 57–75. Hofstede, G., 1980. Culture’s Consequences. Beverly Hills. Sage Publications. Holcomb, T.R., Holmes, R.M., Connelly, B.L., 2009. Making the most of what you have: managerial ability as a source of resource value creation. Strat. Manage. J. 30 (5), 457–485. Hoque, Z., James, W., 2000. Linking balanced scorecard measures to size and market factors: Impact on organizational performance. J. Manage. Account. Res. 12 (1), 1–17. Hult, G.T.M., Ketchen, D.J., Adams, G.L., Mena, J.A., 2008. Supply chain orientation and balanced scorecard performance. J. Manage. Issues 20 (4), 526–544. Ittner, C.D., Larcker, D.F., 1998. Are nonfinancial measures leading indicators of financial performance? An analysis of customer satisfaction. J. Account. Res. 36 (1), 1–35. Ittner, C.D., Larcker, D.F., Meyer, M.W., 2003a. Subjectivity and the weighting of performance measures: evidence from a balanced scorecard. Account. Rev. 78 (3), 725–758. Ittner, C.D., Larcker, D.F., Randall, T., 2003b. Performance implications of strategic performance measurement in financial services firms. Account. Org. Soc. 28 (7), 715–741. Kaplan, R.S., Norton, D.P., 1992. The balanced scorecard – measures that drive performance. Harvard Bus. Rev. 70 (1), 71–79. Kaplan, R.S., Norton, D.P., 2001. Transforming the balanced scorecard from performance measurement to strategic management: Part I. Accounting Horizons 15 (1), 87–104. Kaplan, R.S., Norton, D.P., 2008. Mastering the management system. Harvard Bus. Rev. 86 (1), 1–16. Keegan, D.P., Eiler, R.G., Jones, C.R., 1989. Are your performance measures obsolete? Manage. Account. 70 (12), 45–50. Ketchen, D.J., Wowak, K., Craighead, C.W., 2014. Resource gaps and resource orchestration shortfalls in supply chain management: The case of product recalls. J. Sup. Chain Manage. 50 (3), 6–15. Koufteros, X., 1999. Testing a model of pull production: a paradigm for manufacturing research using structural equation modeling. J. Oper. Manage. 17 (4), 467–488. Koufteros, X., Babbar, S., Kaighobadi, M., 2009. A paradigm for examining secondorder factor models employing structural equation modeling. Int. J. Prod. Econ. 120 (2), 633–652. Lehtinen, J., Ahola, T., 2010. Is performance measurement suitable for an extended enterprise? Int. J. Oper. Prod. Manage. 30 (2), 181–204. Lynch, R.L., Cross, K.F., 1992. Measure up!: The Essential Guide to Measuring Business Performance: Mandarin. Mahama, H., 2006. Management control systems, cooperation and performance in strategic supply relationships: a survey in the mines. Manage. Account. Res. 17 (3), 315–339. Malhotra, N.K., Kim, S.S., Patil, A., 2006. Common method variance in IS research: a comparison of alternative approaches and a reanalysis of past research. Manage. Sci. 52 (12), 1865–1883. Margolis, J.D., Walsh, J.P., 2003. Misery loves companies: rethinking social initiatives by business. Admin. Sci. Quart. 48, 268–305. Melnyk, S.A., Stewart, D.M., Swink, M., 2004. Metrics and performance measurement in operations management: dealing with the metrics maze. J. Oper. Manage. 22 (3), 209–218. Mundy, J., 2010. Creating dynamic tensions through a balanced use of management control systems. Account. Org. Soc. 35 (5), 499–523. Neely, A.D., 1999. The performance measurement revolution: why now and what next? Int. J. Oper. Prod. Manage. 19 (2), 205–228. Neely, A.D., 2005. The evolution of performance measurement research. Int. J. Oper. Prod. Manage. 25 (12), 1264–1277. Neely, A.D., Gregory, M., Platts, K., 1995. Performance measurement system design: a literature review and research agenda. Int. J. Oper. Prod. Manage. 15 (4), 80–116. Neely, A.D., Mills, J., Platts, K., Richards, H., Gregory, M., Bourne, M., Kennerley, M., 2000. Performance measurement system design: developing and testing a process-based approach. Int. J. Oper. Prod. Manage. 20 (10), 1119–1145. Neely, A.D., Adams, C., Kennerley, M., 2002. The Performance Prism: The Scorecard for Measuring and Managing Business Success. Prentice Hall Financial Times London. Pavlov, A., Bourne, M., 2011. Explaining the effects of performance measurement on performance: n organizational routines perspective. Int. J. Oper. Prod. Manage. 31 (1), 101–122. Peikes, D.N., Moreno, L., Orzol, S.M., 2008. Propensity score matching. Am. Stat. 62 (3), 222–231. Porter, M.E., 1991. Towards a dynamic theory of strategy. Strat. Manage. J. 12 (2), 95–117. Rhody, J.D., Tang, T.L.P., 1995. Learning from Japanese transplants and American corporations. Public Pers. Manage. 24 (1), 19–32. Roberts, J., 1990. Strategy and accounting in a UK conglomerate. Account. Org. Soc. 15 (1), 107–126. Rothaermel, F.T., Hess, A.M., 2007. Building dynamic capabilities: innovation driven by individual-, firm-, and network-level effects. Organ. Sci. 18 (6), 898–921. Sarkis, J., Gonzalez-Torre, P., Adenso-Diaz, B., 2010. Stakeholder pressure and the adoption of environmental practices: the mediating effect of training. J. Oper. Manage. 28 (2), 163–176. Simons, R., 1991. Strategic orientation and top management attention to control systems. Strat. Manage. J. 12 (1), 49–62. Simons, R., 1994. How new top managers use control systems as levers of strategic renewal. Strat. Manage. J. 15 (3), 169–189. Simons, R., 1995. Control in an age of empowerment. Harvard Bus. Rev. 73 (2), 80–88. Simons, R., Dávila, A., Kaplan, R.S., 2000. Performance Measurement & Control Systems for Implementing Strategy: Text & Cases. Prentice Hall Upper Saddle River, NJ. Sirmon, D.G., Hitt, M.A., Ireland, D.R., Gilbert, B.A., 2011. Resource orchestration to create competitive advantage: breadth, depth, and life cycle effects. J. Manage. 37 (5), 1390–1412. Sirmon, D.G., Hitt, M.A., 2003. Managing resources: Linking unique resources, management and wealth creation in family firms. Entrepreneurship Theory and Practice 27 (4), 339–358. Sirmon, D.G., Hitt, M.A., Ireland, D.R., 2007. Managing firm resources in dynamic environments to create value: looking inside the black box. Acad. Manage. Rev. 32 (1), 273–292. Sprinkle, G.B., 2003. Perspectives on experimental research in managerial accounting. Account. Org. Soc. 28 (2), 287–318. Stede, W.A.V., Chow, d., Lin, C.W.T.W, 2006. Strategy, choice of performance measures, and performance. Behav. Res. Account. 18 (1), 185–205. Taylor, A., Taylor, M., 2013. Antecedents of effective performance measurement system implementation: an empirical study of UK manufacturing firms. Int. J. Prod. Res. 51 (18), 1–14. Tyre, M.J., Orlikowski, W.J., 1994. Windows of opportunity: temporal patterns of technological adaptation in organizations. Organ. Sci. 5 (1), 98–118. Vandenbosch, B., 1999. An empirical analysis of the association between the use of executive support systems and perceived organizational competitiveness. Account. Org. Soc. 24 (1), 77–92. Voelpel, S.C., Leibold, M., Eckhoff, R.A., 2006. The tyranny of the balanced scorecard in the innovation economy. J. Intellect. Cap. 7 (1), 43–60. Wowak, K.D., Craighead, C.W., Ketchen, D.J., Hult, G.T.M., 2013. Supply chain knowledge and performance: a meta-analysis. Decision Sci. 44 (5), 843–875. Widener, S.K., 2007. An empirical analysis of the levers of control framework. Account. Org. Soc. 32 (7), 757–788. Zimmerman, J.L., 2000. Accounting for Decision Making and Control:. Irwin/McGraw-Hill, Boston.
© Copyright 2026 Paperzz