Treating Words as Data with Error: Estimating Uncertainty in the Comparative Manifesto Project Measures∗ Kenneth Benoit Trinity College Dublin [email protected] Michael Laver New York University [email protected] Slava Mikhaylov Trinity College Dublin [email protected] September 13, 2007 Abstract Spatial models of party competition are central to political science, but can only be tested empirically if there exist reliable and valid measurements of agents’ positions on salient policy dimensions. To date, the main source of cross-national, time-series data on party policy positions comes from from the content analysis of party manifestos by the Comparative Manifesto Project (CMP), cited in hundreds of published studies since the first publication of its results in 1987. Despite the very widespread use of these policy position estimates for studies published in the best journals in our discipline, however, and despite the fact that estimates in these data arise from documents coded only once by one of a handful of CMP coders, the level of uncertainty associated with each CMP estimate has never been characterized. This greatly undermines the value of the CMP dataset as a scientific resource, since even the CMP’s creators agree that their dataset contains several types of error. We propose a remedy to this situation. First, we outline the process by which the CMP data are generated, characterizing the error that arises both from documented tests of coder unreliability, as well as from the inherently stochastic process by which latent party positions are recorded by manifesto authors into observable texts. Second, using previously ignored coding-unit frequencies reported by the CMP project, we reproduce the error-generating process by simulating coder unreliability and bootstrapping analyses of coded quasi-sentences to reproduce both forms of error. This allows us to estimate precise levels of non-systematic error for every category and scale reported by the CMP for its entire set of 3,000+ manifestos. Finally, using our estimates of these errors, we demonstrate ways to correct otherwise biased inferences from statistical analyses of the CMP data, and issue recommendations for future research based on CMP measures. Key Words: Comparative Manifesto Project, mapping party positions, party policy, error estimates, measurement error. ∗ A previous version of this paper was prepared for the 2007 Annual Meeting of the American Political Science Association, Hyatt Regency and Sheraton Chicago, Chicago, Illinois, August 30–September 2, 2007. This research was partly supported by the European Commission Fifth Framework (project number SERD-2002-00061), and by the Irish Research Council for Humanities and the Social Sciences. We thank Andrea Volkens for generously sharing her experience and data regarding the CMP, and Thomas Daubler for research assistance. We also thank James Adams, Garrett Glasgow, Simon Hix, Abdoul Noury, and Sona Golder for providing and assisting with their replication datasets and code. An outline form of this paper was drawn up for the Conference on the Dynamics of Party Position Taking, March 23-24, 2007, University of Binghamton. 1 Using text to measure policy positions Political text, properly analyzed, offers extraordinary potential as a source of information about the policies, preferences and positions of political actors, information vital to many models at the heart of modern political science. The authors of political text may be single individuals such as candidates, prime ministers or judges; they may be collective entities such as political parties, courts or cabinets.1 Research designed to measure policy positions using political text has, until now, largely been limited by a lack of suitable methods for systematically extracting information from the vast amounts of suitable text available for analysis. Recent methodological advances have made some progress in this area, perhaps most importantly by breaking from traditional forms of content analysis to treat text, not as an object for subjective interpretation, but very explicitly as objective data from which information about the author can be estimated in a rigorous and replicable way (e.g. Slapin & Proksch 2007, Monroe & Maeda 2004, Laver, Benoit & Garry 2003, Laver & Garry 2000). Treating words as data opens up the use of well-understood methods of statistical analysis to allow inferences to be drawn from the observable content of authored text about the fundamentally unobservable underlying characteristics (for example policy positions) of its author. The main advantage of this statistical approach is that, relying on wellestablished procedures for rigorously characterizing patterns in data, it eliminates both subjectivity and the propensity for human error, while making the results of text-based analysis easily replicable. One particularly important consequence of treating texts as data and basing inferences on solid statistical procedures, furthermore, is that measures of uncertainty accompany any resulting estimates—a feature now recognized not just as a bonus, but as a sine qua non, for serious empirical research in the social sciences (King, Keohane & Verba 1994, 9). While our methodological focus in this paper is on the analysis of text in general, we do have a more specific type of text in mind. This is the text contained in party manifestos, an enormous volume of which has been analyzed, using human coders, by the Comparative Manifestos Project (CMP), in a twenty-year effort representing one of the most long-running and ambitious data projects in the history of political science. The result of this truly Herculean effort is a dataset reporting estimates of the policy positions of a large number of national political parties in a large number of countries over the entire postwar period.2 First reported in 1987 (Budge, Robertson & Hearl 1987), a hugely expanded version of this 1 Of course there are many alternative ways to measure political positions, including but not limited to the analysis of: legislative roll calls; survey data on preferences and perceptions of political elites; survey data on preferences and perceptions of voters; surveys of experts familiar with the political system under investigation; the analysis of political texts generated by political agents of interest. Benoit and Laver (2006) review and evaluate these different approaches. 2 We also note, however, that the CMP is not the only text-based measure that is based on party manifestos: Laver & Garry 2 dataset was reported in the project’s core publication, Mapping Policy Preferences (Budge, Klingemann, Volkens, Bara & Tanenbaum 2001, hereafter MPP), to have covered thousands of policy programs, issued by 288 parties, in 25 countries over the course of 364 elections during the period 1945-1998. The dataset has recently been extended, as reported in the project’s most recent publication Mapping Policy Preferences II (Klingemann, Volkens, Bara, Budge & McDonald 2006, hereafter MPP2) to incorporate 1,314 cases generated by 651 parties in 51 countries in the OECD and central and eastern Europe (CEE). For scholars seeking long time series of party policy positions in many different countries, the CMP dataset is effectively the only show in town. Since, commendably, these data are freely available as a public resource for the profession, researchers have indeed used them very widely, as evidenced by over 800 Google Scholar citations by third-party researchers of core CMP publications.3 However, despite near-universal recognition within the profession of vital need to measure and report uncertainty in all social scientific data analysis, the CMP data, as used in every one of the very many pieces of published research to date, come with no measures whatsoever of uncertainty. Substantively, the range of research applications using the CMP data is vast, encompassing four major areas of political science: descriptive analyses of party systems (e.g. Bartolini & Mair 1990, Evans & Norris 1999, Mair 1987, Strom & Lejpart 1989, Webb 2000); empirically grounded analyses of party competition (e.g. Adams 2001, Janda, Harmel, Edens & Goff 1995, Meguid 2005, Van Der Brug, Fennema & Tillie 2005); models of coalition building and government formation (e.g. Baron 1991, Schofield 1993); and measuring responsiveness of representative democracy in linkages between government programs and governmental policy implementation (e.g. Petry 1991, Petry 1988, Petry 1995). Additional applications of the CMP data include the analysis of American political behavior (e.g. Erikson, Mackuen & Stimson 2002); evaluation of partisan effects on government expenditure (e.g. Bräuninger 2005); identification of structure and dimensionality of the political space of European Parliament, European Commission and Council of Ministers (e.g. Thomson, Boerefijn & Stokman 2004); evaluation of the issue convergence in US presidential campaigns (e.g. Sigelman & Emmett 2004); analysis of the relationship between budgetary cycles and political polarization and transparency (e.g. Alt & Lassen 2006); evaluation of partisan effects on trade policy (e.g. Milner & Judkins 2004); and establishing the effect of endogenous deregulation on productivity in OECD (e.g. Duso & Röller 2003); The CMP data has also been used as a means to validate other measures of parties’ policy positions, such expert surveys and (2000), Laver, Benoit & Garry (2003), and Slapin & Proksch (2007) are also examples. 3 As of August 25, 2007. The precise number of third-party citations is hard to calculate because third-party users are likely to cite several CMP sources in the same paper. 3 automated content analysis (e.g. Laver, Benoit & Garry 2003, Ray 1999). Notwithstanding this range and volume of applications, the CMP measures of party policy positions come with no measure of the uncertainty surrounding their point estimates, which lie on very precise continuous scales (0–100 or -100–100). While this problem has long been acknowledged by the project and its critics (e.g. Benoit & Laver 2007, Klingemann et al. 2006, Ch. 5), no conclusive solution has yet been offered to the problem of providing uncertainty measures for any of the policy scores reported by the CMP. Reliable and valid use of the CMP data, however, mandates to use of such measures. Without them, users of the data cannot distinguish between “signal” and “noise”, making it impossible to tell the difference between measurement error and “real” differences in policy positions. Compounding this problem, the overwhelming use of CMP estimates of party policy positions is as explanatory variables. Measurement error in these CMP variables, if not taken into account, will yield biased inferences about key causal relationships and thus misleading research findings. In short, the unknown level of non-systematic error contained in the CMP dataset drastically undermines its primary value for the profession, as a reliable and valid set of estimates of party policy positions across a wide range of years, countries and policy dimensions. In this paper, we set out to address this problem head on, and thereby to enhance the value of the CMP data for the profession as a whole. We do this by decomposing stochastic elements in the data generation process that leads to the reported CMP estimates and simulating the effects of these. This enables us to obtain precise estimates of the non-systematic error associated with every reported CMP measure of party policy positions. The result of our analyses of these stochastic processes is a set of standard errors for the CMP data that allow scholars to make more informed judgments, when they compare two CMP estimates of different parties at the same time or of the same party at different times, about whether these differences are likely to be “real”, or simply the result of stochastic features of the data generation process. Our analysis proceeds as follows. First, we briefly describe the CMP dataset and the processes that led to its generation. We then analyze the two main stochastic processes that are involved: manifesto writing and manifesto coding. Based on these analyses, we show how to calculate standard errors for each estimate in the CMP dataset. Analyzing these error estimates, we find that many CMP quantities should be associated with substantial uncertainty. Exploring the effects of this uncertainty on substantive work, we use show how the error estimates can be brought to bear in judging substantive change versus measurement error in both time-series and cross-sectional comparisons of party positions. Finally, we 4 suggest ways to use our error estimates to correct analyses that do use CMP data as covariates, re-running and correcting some prominent analyses reported in recent literature. 2 The CMP coding process In order to characterize the error in the CMP dataset, we must begin by understanding the process which gives rise to the observed estimates. First, researchers who spoke the language of the country under investigation were asked to identify a “text” for a particular party at a particular election, which was either a formal party “manifesto” or, when this was not available, some valid manifesto surrogate. Second, the selected text was read by a trained human coder and broken down into what the CMP calls “quasi sentences”. These are defined as self-contained text units, each expressing a distinct policy message, which could have been written as stand-alone sentences had the manifesto been written by a different author with different stylistic proclivities. “Language is...complex, and different forms or styles may be used to express the same political ideas” (Klingemann et al. 2006, 166). The identification of natural sentences only requires a coder be able to recognize a simple punctuation mark when s/he sees it; but the identification of quasi-sentences: (a) fundamentally recognizes the inherently stochastic nature of text-generation; (b) requires a fundamentally subjective judgment that might well vary between coders. In practice, as we see below, the total number of quasi-sentences identified for a given manifesto does indeed vary substantially between different coders. Despite this potential variance in even the basic text units to be coded, all estimates reported in the CMP dataset are based on manifestos coded once, and only once, by a single coder. Having broken the text under analysis into quasi-sentences, the coder then assigns each quasisentence to one of the 56 policy categories that form the CMP coding scheme. A quasi-sentence such as “...incentives for employee share ownership,...” is assigned to category 202, “Democracy: Positive” (example taken from Klingemann et al. 2006, 177).4 If no suitable category can be identified, the quasisentence is marked as “uncoded.” The proportion of uncoded quasi-sentences in each manifesto in the CMP dataset is typically in the range 0-5 percent. Once a manifesto has been coded in this way, the number of quasi-sentences coded into each content category is counted, and reported as a percentage of 4 This quasi-sentence is taken from the sample coded manifesto from the British Liberal/SDP Alliance in 1983 (Klingemann et al. 2006, 177). The CMP coding scheme is based on a 56-category coding scheme developed in the mid-1980s, building on earlier work by David Robertson (1976). The coding of party manifestos by CMP researchers has been very extensively described in the CMP’s core publications (MPP and MPP2). Note that here we do not evaluate the substantive validity of this scheme, noting only that the particular 56-category CMP coding scheme that generated these data is one of many possible coding schemes that might have been devised. 5 the total number of quasi-sentences in the manifesto. These percentages, reported by coding category, are the fundamental quantities in the CMP data that are used by third-party researchers. In practice, most scholars do not use the full 56-category set of CMP coding categories; rather, they seek a low-dimensional representation of party policy positions. It is very common practice, therefore, for scholars to use the composite “left-right” scale developed and tested by the CMP. Based on subjective judgments about the meaning of particular categories, supplemented by exploratory factor analyses, this simple additive scale combines information from 26 CMP policy variables, 13 referring to left wing positions and 13 referring to right wing positions. The CMP refers to the measurement of party movement on this left-right positional scale as its “crowning achievement” (MPP, 20) and has relied extensively on such left-right movements when establishing the face validity of its estimates. We do not examine the substantive validity of this scale here. We do, however, devote considerable attention to an analysis of potential errors in the estimation of this left-right scale, since it by far the CMP measure most widely used by third party researchers. 3 The stochastic process generating CMP data The reason to analyze manifestos is that they provide authoritative statements of party policy at the time of elections. If we treat these texts as data, however, we must understand the processes giving rise to these data, especially whether these processes generate systematic and/or non-systematic error. Thinking about the analysis of party manifestos, consider the stochastic relationship between: the unobservable party position(s) of a manifesto’s author(s); the observed text deposited in the party manifesto; and the coded dataset generated from this text. We view this relationship as having two main elements. The first involves manifesto authorship, the stochastic generation of observed data, in the form of text deposits in manifestos, from non-stochastic unobserved party positions. Put simply, two different authors, or one author at two different times (perhaps after hard disk crash), are almost certain to generate two somewhat different manifestos, even when doing their utmost to express an identical party position. The observed manifesto is one of many different manifestos that could have been written to express precisely the same position. A second stochastic process involves manifesto coding, given the particular manifesto text deposited on the record. Two different coders, or the same coder working at two different times, are very likely to code the same manifesto in somewhat different ways. To begin with, the guidelines for determining what constitutes the basic text unit for the CMP coding scheme—the “quasisentence”—are quite difficult for different coders to follow consistently, leading to a stochastic process 6 in the dissection of manifestos into text units, before any coding takes place.5 As a consequence of the number of quasi-sentences in any given manifesto being a matter of subjective judgment (by a single coder), the basic number of coded units observed for any manifesto forms a random variable. Following this, the particular allocation of a particular text unit to a particular coding category is also subject to variation between two coders setting out to do precisely the same thing. In what follows, we elaborate on these two different aspects of the stochastic process generating the observed CMP data. 3.1 From unobserved policy position to written text Our characterization of the text generation process begins by assuming there is indeed some underlying policy position that text authors set out to communicate. This is shown on the left hand side of Figure 1. We need not concern ourselves here with whether this position is “sincere” or “strategic”, nor with whether or not there is internal party conflict over what the party position should be. We are interested solely in the position that text authors set out to communicate, for whatever reason. Given this position— fundamentally unobservable by anyone else but common knowledge to the authors of the text—a text is generated to do this job. Our point is that many different texts could be generated to communicate the same position. There is nothing “golden” about the actual published text, although the generation of policy texts is a mysterious and idiosyncratic process, about which our best information is largely anecdotal. Thus, for example, we do know that drafts of eventually-published policy texts are typically amended many times, often right up the last minutes before publication. This means there are typically many different actual texts, stored in word processors, filing cabinets and trash cans; but only one of these texts is eventually published. To this we add the very many conceivable policy texts that could have been drafted by authors earnestly setting out to communicate precisely the same position. Thus if you and I set out to draft the same party manifesto containing precisely the same policy messages, it would be eerie beyond belief, vanishingly improbable, for us somehow to generate two texts that were word-for-word identical. Looking at this another way, imagine the universe of possible 10,000word texts that could be generated using a 5,000-word political lexicon. There is a gigantic number of 5 If we were to go further, we could also identify the problem of category “seepage”, which happens when a text unit is related to several categories and therefore prone to a degree of inherent misclassification, a process likely to cause bias whereby some true categories πi j are over-represented by empirically observed categories pi j . While this problem is widely acknowledged in CMP work (e.g. MPP2, 116; Volkens 2001a, 38) and certainly adds additional error to the CMP dataset, here we stick with the assumption of Equation 1 which states that whatever misclassification problems might exist in the coding process, they do not result in systematic bias. Additional forms of systematic error, which we do not consider here, arise from the potential for which arises because coders know, before they start coding, which party generated the manifesto under investigation, or because certain coders tend to use certain categories more than other coders. 7 possible different texts in this universe and it would clearly be ridiculous to claim that every single one of these communicates a unique policy position. [FIGURE 1 ABOUT HERE] Working within the CMP’s implicit measurement assumption that the universe of possible policy positions is spanned by a set of fixed positions 56 different issue dimensions—where differences in position between parties “consist of contrasting emphases placed on different policy areas” (Budge et al. 2001, 82), we characterize the fixed policy positions as a set of proportions in each coding category (rescaled by the CMP as percentages) that directly reflect positions. This is works within the CMP view of policy positioning, grounded in a “saliency” theory of party competition (see Budge et al. 2001, Ch. 3). For a given policy category j, we define πi j as the true, unobservable policy position of countryparty-date unit i. The j categories in this case are the 56 policy categories in the CMP coding scheme, plus an additional category for “uncoded”, giving a total of k = 57 categories. Since, according to the CMP’s measurement model, these true policy positions are represented by relative or “contrasting” emphases on different policy categories within the manifesto, these policy positions are relative proportions, with ∑kj=1 π j = 1.6 For example, party i’s emphasis, for a given election, on the 20th issue category in the CMP coding scheme (401: Free Enterprise), is represented as πi20 . Although we never directly observe the πi j “true” policy positions of manifesto authors, it is possible, using the CMP’s coded manifesto data, to measure the relative emphasis that each πi j is given in party i’s manifesto, measured as p j , . . . pk , where p j ≥ 0 for j = 1, . . . , k and ∑kj=1 p j = 1. In the absence of systematic error (bias): E(pi j ) = πi j (1) In other words, the relative emphasis given in a party’s manifesto will on average reflect the true, fixed, and unobservable underlying position πi j . Its particular manifestation in any given manifesto, however, reflects the imperfect and stochastic process of text authorship, yielding the observed proportions pi j . Put another way, every time a manifesto is written with the intention of expressing precisely the same underlying positions πi j , we expect to observe a slightly different values pi j . Given the way we characterize both observed and unobservable policy positions—which directly 6 In what follows, we refer to these quantities as policy “positions.” The CMP’s saliency theory of party competition is neither widely accepted nor indeed taken into any account by most third-party users of CMP data. However, inspection of the definitions of the CMP’s coding categories reveals that all categories but one of the 56 are very explicitly positional in their definitions, which refer to “favorable mentions of...”, “need for...,” etc. The sole exception is PER408 “Economic goals”, a category which is (quite possibly for this reason) almost never used by third-party researchers. For this reason, we do not regard it as in any way problematic that third-party users almost invariably interpret the CMP’s “saliency” codings as “positional.” 8 follows the CMP’s own assumptions—we can go further and derive a statistical distribution for the observed policy positions. If we consider that each quasi-sentence’s category is determined independently from that of other quasi-sentences, then we can characterize the CMP’s manifesto coding scheme as corresponding to the well-known multinomial distribution with parameters ni and πi j , where ni refers to the total number of quasi-sentences in manifesto i. The probability for any manifesto i of observing counts of quasi-sentences xi j from given categories j is then described by the multinomial formula: Pr(X j = x j , . . . , Xk = xk ) = xk x1 n! x j !···xk ! π1 · · · πk 0 when ∑kj=1 x j = n (2) otherwise Translating into the CMP coding scheme, (for a given manifesto) each xk represents the observed number of quasi-sentences that occur in a given category j. In terms of the “PER” or percentage categories reported by the CMP for each manifesto, what is actually reported is xi j /ni j 100, or the estimate of manifesto i’s “true” percentage (πi j 100) of the quasi-sentences from category j. For our purposes, we have no additional information to suggest any reason why there would be a systematic function mapping (in a biased way) the true position to a different expected observed position—already expressed by Equation 1.7 Our concern here is with non-systematic (unbiased) error, which is the extent to which Var(pi j ) > 0, even though πi j is fixed at a single, unvarying point. So far we have considered only the case of a “given” manifesto, but of course the combined CMP dataset set deals with many such units—a total of 3,018 separate units representing different combinations of country, election date, and political parties for the combined (MPP + MPP2) datasets.8 If we are to fully characterize the error from the stochastic process whereby texts are generated, then this will mean estimating Var(pi j ) for every manifesto i for all k = 57 categories. 3.2 From written manifesto to CMP-coded text units The second stochastic process which we identify here concerns the way that a manifesto, once authored, published, and effectively “fixed”, is then translated into the CMP text units and assigned a code. Here 7 Despite the clearly observable and striking variations in manifesto length and content, in other words, we know of no systematic expectation, based on empirical evidence or even pure theory, about the nature of the stochastic process that maps actors’ (unobservable) political positions to observed manifesto content. The process whereby the unobserved parameters πi j are generated, in other words, is something of a mystery that we simply assume to be revealed by the coded manifesto dataset itself. 8 It is not quite accurate to state that the dataset represents 3,018 separate manifestos, since some of these countryelection-party units share the same manifesto with other parties (progtype=2) or have been “estimated” from adjacent parties (progtype=3). See Appendix 3, MPP. The full CMP dataset also failed to provide figures on either total quasi-sentences or the percentage of uncoded sentences for 141 manifesto units, limiting the sample analyzed here to 2,877 units. 9 we focus on one particular aspect of this process, the identification of the “quasi-sentences” that are coded and counted as the basis of the policy estimates reported. The right hand side of Figure 1 shows what happens between the generation of an observed party manifesto and the eventual characterization of this document in the CMP dataset. [TABLE 1 ABOUT HERE] The task of parsing a manifesto into quasi-sentences is significant because policy texts not only vary widely in length, but also express separate, politically relevant ideas (or “arguments” in CMP parlance) within single natural sentences. One thing clear from any investigation of the CMP dataset, for example, is that the lengths (or ni ) of the coded manifestos vary significantly, as shown in Table 1. Although the total number of quasi-sentences identified in each manifesto is reported the CMP dataset, (and summarized in Table 1), this quantity is almost never referred to by subsequent users of the data. However, about 30 percent of all manifestos had less than 100 quasi-sentences, coded into one of 56 categories. Some manifestos had less than 20 quasi-sentences; some had more than 2000 quasi-sentences. Despite this wide variation in the amount of policy information in different manifestos, all estimated party policy positions are almost invariably treated in the same way, regardless of whether they are derived from a coding of 20 text units, or 2000. The total number of quasi-sentences found in any one manifesto appears to be—at least for lack of any systematic information or prior expectations on this topic—unrelated to any of the political variables of interest. Yet assuming that the proportions πi j remain the same irrespective of document length, then increasing the length of a manifesto should also increase our confidence about the estimates of these proportions. This reflects one of the most fundamental concepts in statistical measurement: the notion that uncertainty about an estimate should decrease as we add information to that estimate. [TABLE 2 ABOUT HERE] The problem of short manifestos and thus of sparse data is compounded by the fact that not all quasi-sentences can be coded by the CMP scheme. For instance, while immigration is self-evidently an important subject of political debate these days in many different countries, there is no unambiguous coding category in the CMP scheme in which to record statements about immigration, asylum-seekers, or labor market restrictions to foreigners. Table 2 describes the frequencies of manifestos with varying levels of uncodable content, as a percentage of the total number of quasi- sentences. While the median percentage of uncoded content is quite low, at 2.1%, the top quarter of all manifestos contained 8% or more of uncoded content, and 10 percent contained 21% or more uncoded content. The saliency theory 10 of party competition that motivated the original CMP project led to the methodological assumption that party policy in particular areas can be measured in terms of the percentage of total manifesto content coded into each category—in effect assuming that ∑ j πi j = 1. This means that increasing the proportion of uncodable content will necessarily decrease the estimated proportions of coded content. This problem has been noted previously (e.g. Kim & Fording 1998) in the context of the left-right scale in particular, but affects all policy categories. Since (unpredictably variable) uncoded content affects the policy estimates that are calculated from coded content, it does introduce a further stochastic element in the manifesto generation process.9 CMP researchers have long been aware that different coders can code the same quasi-sentence in different ways, quite apart from finding a different set of quasi-sentences in the same manifesto. The ideal way to drive towards more reliable estimates, assuming no coder is perfectly reliable, is of course to ask many different human coders to code each text and then combine this information in a conventional way to produce a set of estimates with associated uncertainty. The resource implications of doing this are so enormous, however, that it is simply never going to happen; the reality is that nearly all of the 3,000+ manifestos that form the basis of the CMP dataset were coded once, and once only, by a single coder. A second-best solution to minimizing coder unreliability, adopted by the CMP, has been to devote increasing levels of care and attention to the training of hired coders.10 The objective of this training has been to reduce the variance in human coders’ application of the same coding scheme to the same “ master” text, selected by the CMP and reproduced in (Klingemann et al. 2006, Appendix 2). This training process does also generate limited empirical evidence about levels of variation in the qualitative judgments made about the same text by different coders. This is reported in various publications by Andrea Volkens, who supervised this process over a long period of time (Volkens 2001a, Volkens 2001b, Volkens 2007). Even when coders are given a second chance to code the training document, following comparison to the gold standard “master” coded document, the best average correlation of coders is still only 0.88; before correction, this proportion can be as low as 0.72 (Volkens 2001a, 39). Because our concern here is limited to non-systematic forms of error in the coding process, we focus only on the stochastic process whereby a written manifesto is (subjectively) parsed into quasi9 When uncoded content consists of purely non-political sentences, then this still affects the political sentence codings because of the additive assumption that position is determined by the proportion of mentions. In the context of discrete choice models such as the multinomial logit, this problem is known as a violation of the IIA principle. When uncoded content consists of political content that the CMP scheme cannot capture, the problem is likely to be a more serious problem of bias, although we do not investigate that possibility here. 10 Despite the considerable attention now devoted by the CMP to the training of hired coders, it is also the case that manifestos from early CMP publications, and much of the current CMP data associated with this, were coded by senior researchers in the original project who were, to all intents and purposes, completely untrained. 11 sentences by the human coder. An analysis of the results from repeated codings of the training document used by the CMP to initiate new coders by Volkens (2001b) reveals important insight into the deviation of coders from the “correct” quasi-sentence structure. Volkens reports that average deviation from correct quasi-sentence structure in reliability tests of thirty-nine coders employed in the CMP was around ten percent. These results suggest that, even using the best estimates from coders who have been corrected and given a second chance to get it right, different coders nonetheless make very different judgments made during the first step of parsing a document into “quasi”-sentences. In the tests we have analyzed ourselves, involving 59 different CMP coders, we found that these identified between 127 and 211 quasisentences in the same training document, with a SD of 19.17 and an IQR of (148, 173).11 As mentioned earlier, an inter-coder variance in counting quasi-sentences in the tests is typically +/-10%, a conclusion also reached by Volkens (2001b, 38). Because these correlations refer only to totals of quasi-sentences identified, there is no measure of the correspondence of the actual parsed sentences themselves. This means that there may well be many more differences between the quasi-sentence structures identified by different researchers than those we report here. Nonetheless on the basis of these reliability test results, we can make the assumption that E(n̂i ) = ni (3) Var(n̂i ) = (0.10 ni )2 (4) ni ∼ N(n̂i , 0.10n̂i ) (5) where n̂i refers to the to the total number of quasi-sentences reported for each manifesto i in the CMP dataset, and (as before) ni refers to the “true” or correct number of quasi-sentences in manifesto i. Using this characterization of variance, combined with the the multinomial characterization of the stochastic manifesto authorship process, we are now in a position to estimate the errors for each policy category for each manifesto. 11 We thank Andrea Volkens for kindly sharing not only this dataset with us, but also her vast experience in dealing with different coders over the many years of the CMP. In reporting the variance of the quasi-sentence length of the training document, we have deliberately not reported the “correct” text length since this document is still used by the CMP for training coders. 12 4 4.1 Estimating Error in Manifesto Generation Analytical error estimation One way to assess the error variance of estimated percentages of text units any of the CMP’s 56 coding categories is using analytic calculations of variance for the two error distributions we have stipulated, the multinomial and normal distributions. For the normal distribution which describes the stochastic distribution of total quasi-sentences, calculating the variance for any manifesto length is trivial, since (by Equations 4) this is a direct proportion of the observed (by Equation 3) total quasi-sentences. The discussion here focuses on the slightly more complicated case of computing the analytic variance for the 56 policy categories. The target is to determine the variance of the percentage (“PER”) categories reported by the CMP, which in the language described above represent π̂i j 100 for each category j and each manifesto i. Here we are assuming no bias (i.e. Equation 1), where each πi j represents the true but unobservable policy position of country-party-date unit i on issue j. Returning to the definition of the multinomial distribution in Equation 2, for any multinomial count Xi j , the variance is defined as Var(Xi j ) = ni pi j (1 − pi j ) (6) With a bit of algebraic manipulation (detailed in Appendix A) we can express the variance of the proportion pi j , and the rescaled percentage (used by the CMP as): 1 pi j (1 − pi j ) ni (7) SD(pi j 100) = 100 √ ni (8) SD(pi j ) ∝ 1 √ ni Var(pi j ) = q pi j (1 − pi j ) In part, then, the error will depend on the size of the true percentage of mentions pi j 100 for each “ PER” category j. Assuming this quantity is fixed for each party-election unit i, however, what is variable as a result of the data generating process is the length ni of the manifesto that the party associated with i deposits. This aspect of the error in the CMP estimates, therefore, is inversely proportional to the (square root of the) length of the manifesto. This should be reassuring, since it means that longer manifestos will reduce the error in any CMP category, irrespective of its proportion p j . Longer manifestos provide 13 more information, in other words, and therefore we can be more confident about codes constructed from them. The situation is slightly more complicated for additive measures such as the pro-/anti-EU scale (PER108-PER110) or for RILE, the additive category obtained by summing 13 categories on the “ left” and subtracting 13 categories on the “right”. This is because for summed multinomial counts, the covariances between categories must also be estimated, since it is a property of variance that Var(aX + bY ) = a2 Var(X)+b2 Var(Y )+2abCov(X,Y ). There are several strong reasons—including coder practice as well as innate relationships between mentions of related categories—to suspect that these covariances will be non-zero. For these reasons, we do not recommend using the analytically derived errors in practice, since the simulated errors developed in the next section will allow us to confront the unknown covariances directly, as well as to combine the multinomial and normal variances into a single measure of estimated error. 4.2 Estimating Error Through Simulation Given the potential analytical problems we identify at the end of the previous sections, we also suggest a way to assess the extent of error in the CMP estimates using simulations that recreate the stochastic processes that led to the generation of each manifesto, based on our belief that there are many different possible manifestos that could have been written to communicate the same underlying policy position. This is done by bootstrapping the analysis of each coded manifesto, based on resampling from the set of quasi-sentences in each manifesto reported by the CMP. Bootstrapping is a method for estimating the sampling distribution of an estimator by resampling with replacement from the original sample. Bootstrapping has three principal advantages over the analytic derivation of CMP error from the previous section. First, it does not require any assumption about the distribution of the data being bootstrapped and can be used effectively with small sample sizes (N < 20) (Efron 1979, Efron & Tibshirani 1994). Second, bootstrapping can be used directly on additive indexes such as the CMP “right-left” scale without making the assumptions about the covariances of these categories that would be required to derive an analytic variance. Since the exact covariances of these categories are unknown, highly sample dependent, and very probably also influenced by coder errors, it would be speculative to make any assumptions about these covariances for use in an analytically computed variance for the combined scales. Finally, simulation allows us to mix error distributions, a key requirement in our case if we are to incorporate the two types of error we have identified here, consisting of both multinomial and normal distribution 14 parameters. For all of these reasons, we always recommend the bootstrapped error variances over an analytic solution for additive CMP measures such as the left-right scale. The bootstrapping procedure is straightforward. Since the CMP dataset contains percentages of total manifesto sentences coded into each category, as well as the raw total number of quasi-sentences observed, we can convert reported percentages in each category back into raw numbers. This gives us a new dataset in which each manifesto is described in terms of the number of sentences allocated to each CMP coding category. We then bootstrap each manifesto by drawing 100 different random samples from the multinomial distribution, using the pi as given from the reported PER categories, and a new n̂i drawn from a normal distribution as per Equation 5 for each sample. Each (re)sampled manifesto looks somewhat like the original manifesto and has the same length, except that some sentences will have been dropped and replaced with other sentences that are repeated. We feel this is a fairly realistic simulation of the stochastic text generation process. The nature of the bootstrapping method applied to texts in this way, furthermore, will strongly tend to reflect the intuition that longer (unbiased) texts contain more information than shorter ones. One problem that is not addressed by bootstrapping the CMP manifesto codings is that, as anyone who has a close acquaintance with this dataset will know, it is typical for many CMP coding categories to be empty for any given manifesto—resulting in zero scores for the variable concerned. No matter how large the number we multiply by zero, we still get zero. Thus a user of the CMP data dealing with a 20sentence manifesto that populates only 10 coding categories out of 56 must in effect assume that, had the manifesto been 20,000 sentences long, it would still have populated only 10 categories.12 In extremis, if some manifesto populated only a single CMP coding category, then every sampled manifesto would be identical. We cannot get around this problem with the CMP data by bootstrapping, unless we make some arbitrary assumptions about probability distributions for non-observed categories, and we are very reluctant to do this. Aside from the problem of empty CMP coding categories, the great benefit of bootstrapping the CMP estimates in this way, to simulate the stochastic process of text generation, is that we can generate the standard errors and confidence intervals associated with the estimated value for each coding category, for scales generated by combining these categories, such as the CMP additive left-right scale, as well as more specific scales measuring party positions on social, liberalism or economic policy. 12 In this context note that any manifesto with fewer than 56 quasi-sentences is constrained to yield some empty policy categories in the CMP’s 56-category scheme. More generally, even with more that 56 quasi-sentences, the probability of empty coding categories increases as the number of quasi-sentences declines. 15 [FIGURE 2 ABOUT HERE] The results of using this bootstrapping procedure are reported in Figure 2, where we show the relationship between manifesto length, measured in quasi-sentences, and the bootstrapped standard error of both the additive CMP left-right scale, as well as the error estimated for PER501, the “pro-environment” category.13 The first panel (a) of Figure 2 shows the effects of the bootstrapped measure for PER501, the single category indicating a pro-environmental stance. As predicted in the previous section, its error variance declines directly with (logged) manifesto length. The wide differences (indicated by the thick band) reflect the very different proportions of piPER501 for different manifestos i. In panel (c), we compare the bootstrapped error variance to the variance computed analytically (per Equation 8) for the Environment (PER501) category. The near equivalence of these two methods of obtaining the standard errors serves to validate both the analytical derivation of the CMP error variance, as well as showing that bootstrapping the manifestos from their quasi-sentences is a valid way to simulate error variance in the categories. The bootstrapped method, in other words, faithfully reproduces the error we have identified using the analytic variance based on the multinomial distribution. [FIGURE 3 ABOUT HERE] When we add in the additional step of letting the number of quasi-sentences vary (according to Equation 5), then we observe additional noise in the results of the bootstrapping procedure, arising from varying the number of quasi-sentences by approximately +/-10%. The results are shown in Figure 3, which plots the bootstrapped error without randomly drawing the simulated ni vector, versus those where it has been randomly drawn. As can be seen by the clustering about the unit axis, this procedure introduces additional noise but no evident bias, since the number of quasi-sentences is not systematically related to the policy quantities being estimated. In the analyses in the next section that apply error estimates to specific empirical research problems, we will always use this combined bootstrapped error with the randomized ni as our best approximation of overall non-systematic error in the CMP’s reported estimates. 5 Using CMP Error Estimates in Applied Research There are two main research settings in which scholars need to estimate the policy positions of political actors. The first is cross-sectional: a map of some policy space is needed, based on estimates of different 13 Given the distribution of the data, both axes of Figure 2 uses logarithmic scales and the figure demonstrates clearly that the level of bootstrapped error in this scale error is very strongly related to manifesto length—short manifestos have much more potential for error of this type than the long ones. 16 agent positions at the same point in time. The second is longitudinal: a time series of policy positions is needed, based on estimates of the same agent’s policy positions at different points in time. As we have noted, many alternative techniques are available for estimating cross-sectional policy spaces. The signal virtue of the CMP data is that it purports to offer time series estimates of the policy positions of key political agents. Clearly, either cross sectional or time series estimates of policy positions convey little or no rigorously usable information if they do not come with associated measures of uncertainty. Absent any such measure, estimates of “different” policy positions may either be different noisy estimates of the same underlying signal, or accurate estimates of different signals. We have no way to know which is the case; yet we need to know this if we are to draw valid statistical inferences based on variation in our estimates. Drawing such inferences is, of course, the entire purpose of conducting empirical research in the first place. 5.1 Estimating valid differences A substantial part of the discussion found in MPP and MPP2 of the face validity of the CMP data comes in early chapters of each book, during which the policy positions of specific parties are plotted over time. Sequences of estimated party policy movements are discussed in detail and held to be substantively plausible, with this substantive plausibility taken as evidence for the face validity of the data. But are such estimated changes in party policy “real”, or merely measurement noise? We illustrate how to answer this question with a specific example taken from Germany, the country in which the CMP has been based for many years. Figure 4 plots the time series of the estimated environmental policy positions of the CDU-CSU (PER 501 in the CMP coding scheme). The dashed line shows the CMP estimates, the error bars show the bootstrapped 95 percent confidence intervals around these estimates, calculated as described above. [FIGURE 4 about here] We see clearly from this figure that the error bands around the CMP estimates are large and thus that most of the estimated “changes” in the CDU-CSU’s environmental policy over time could well be noise. Statistically speaking, we can conclude that the CDU-CSU was more pro-environmental in the early 1990s than it was either in the early 1980s or the early 2000s, but that every other observed “movement” on this policy dimension could well be caused by measurement error. [TABLE 3 about here] Table 3 reports the result of extending this anecdotal discussion in a much more comprehensive 17 way. It deals with observed “changes” of party positions on the CMP’s widely-used left-right scale (RILE) and thus systematically summarizes all of the information about policy movements that is used anecdotally, in the early chapters of MPP and MPP2, to justify the face validity of the CMP data. The table reports, considering all situations in the CMP data in which the same party has an estimated position for two adjacent elections, the proportion of cases in which the estimated policy “change” from one election to the next is statistically significant. These results should be of considerable interest to all third-party researchers who use the CMP data to generate a time series of party positions and show that observed policy “changes” are statistically significant in only about 28 percent of the relevant cases, or slightly more than one in four. We do not of course conclude from this that the CMP estimates are invalid. But we do conclude that many of the policy “changes” that have hitherto been used to justify their validity are not statistically significant and may well just be noise. More generally, we argue that, if valid statistical (and hence logical) inferences are to be drawn from “changes” over time in party policy positions, estimated from the CMP data, then it is essential that these inferences are based on valid measures of the uncertainty in the CMP estimates, which have not until now been available. [FIGURE 5 about here] The other type of estimated policy difference that interests scholars analyzing party competition, as we have seen, concerns cross-sectional differences between different parties at the same point in time. Considering some static spatial model of party competition that is realized by estimating the positions of actual political parties at some particular time, many of the model’s implications will depend on differences in the policy positions of the various parties. It is crucial, therefore, when estimating a cross-section of party policy positions, to know whether the estimated positions of different parties do indeed differ from each other in a statistical sense. Figure 5 illustrates this problem for recent elections in Germany (top panel) and France (bottom panel), giving estimates for 2001 of party positions on the CMP left-right scale. These show us quite clearly, for the case of Germany, that the five political parties under investigation took up only three statistically distinguishable positions. Taking account of the uncertainty of the estimates, the Social Democrats and Free Democrats have statistically indistinguishable estimated positions on the CMP left-right scale, as do the Greens and PDS. The situation is even “muddier” in France where four parties—Communists, Socialists, Greens and UMP—have statistically indistinguishable estimated positions on the CMP left-right scale in 2001. Only the far-right FN had an estimated left-right position that clearly distinguished it from all other parties. To conclude this discussion, we can see that deriving measures of uncertainty for the CMP esti- 18 mates is no small matter. On one hand it considerably extends our confidence in the inference that two party positions differ—cross-sectionally or longitudinally—if this difference can be shown to be significant in a statistical sense. On the other hand we see that many of the policy “differences” on which third-party authors have hitherto relied in their empirical work may well be just noise. This stands to reason. It has never been a secret that some of the manifestos on which the CMP estimates are based are very short indeed, and thus convey rather little information, while other manifestos are quite long, and thus tend to convey much more information. Notwithstanding this, users of CMP data have hitherto treated all estimates as if they conveyed precisely the same amount of information. As we have just seen, using the measures of uncertainty we develop in this paper makes a big difference to the statistical inferences we draw from the CMP estimates, whether in a longitudinal or cross-sectional setting. 5.2 Correcting estimates in linear models When covariates measured with known error are used in linear regression models, the result is bias and inefficiency when estimating coefficients on the error-laden covariates (Carroll, Ruppert, Stefanski & Crainiceanu. 2006). The usual assumption is that coefficients will suffer from “attenuation bias”: they are likely to be biased towards zero, thus underestimating the relevant variables’ true effect.14 The CMP estimates of party policy positions clearly have some error, even on the CMP’s own published accounts of inter-coder reliability. As we have shown above, it is possible to derive measures of the effect of some of the error processes involved in generating the CMP data. Nearly all empirical research that relies on CMP estimates as measures of party policy positions uses these estimates as covariates in linear regression models. Of all of the studies that have used CMP data as covariates in linear regression models, however, to our knowledge not a single one has explicitly taken account of the likelihood of error in the CMP estimates, or even used the length of the underlying manifesto as a crude indication of potential error. As a result, we expect many of the reported coefficients in the studies that use CMP estimates as covariates to be biased and inconsistent. We address this issue by replicating and correcting three recent high-profile studies that use the CMP data, each published in a major political science journal: Adams, Clark, Ezrow & Glasgow (2006, 14 The assumption of attenuation bias in models with error-contaminated covariates has even been called the “Iron Law of Econometrics” (Hausman 2001, 58). However, Carroll et al. (2006, 41) suggest that conclusion of attenuation bias has to be qualified, because it depends on the relationship between true covariate, its observed proxy, and possibly other covariates in the model. Moreover, in case of multiple variables containing measurement error it will also depend on the relationship between measurement error components of these error-contaminated variables. Thus, in linear regression depending on whether it is a simple or multiple regression, single or multiple covariates measured with error, and whether measurement error contains bias, the results will range from simple attenuation bias to cases where real effects are hidden, or relationships found in the observed data are nonexistent in true data, or the reversal of the signs of coefficients compared to the case with true data. 19 AJPS), Hix, Noury & Roland (2006, AJPS) , and Golder (2006, BJPS). Each study uses CMP-based estimates of policy as explanatory variables, and each uses the measures in a slightly different fashion. In all three cases we obtained the datasets (and replication code) from the authors and replicated the analyses correcting for measurement error in the CMP-derived variables. We do this using a simple error correction model known as simulation-extrapolation (SIMEX) that allows generalized linear models to be estimated with correction for known or assumed error variance for error-prone covariates (Stefanski & Cook 1995, Carroll et al. 2006). SIMEX is a method to estimate and reduce the effect of measurement error on an estimator experimentally via simulations. The basic idea of the method is fairly straightforward: If a causal relationship is biased by measurement error, then adding more measurement error should increase the degree of this bias. By adding (or simulating) successive levels of measurement error, it is possible to estimate the relationship between the expected value of the coefficient and its true value, and then extrapolate back to the unbiased estimate.15 Given a relatively robust set of assumptions, the SIMEX procedure has been verified to be exactly consistent in linear and loglinear mean models, and approximately consistent in general (Carroll et al. 2006, 151), and to provide a good approximation to the unbiased coefficient estimates in a variety of cases, including non-linear models. SIMEX is particularly useful for models with additive measurement error. More complicated error corrections are of course possible, but here we deliberately chose a method that is simple, applicable to a wide class of general linear models, and for which freely available software is available that can be used with popular statistical packages.16 5.2.1 Adams, Clark, Ezrow and Glasgow (2006) Adams et al. (2006) are concerned with whether “niche” parties (for example Communists, Greens or extreme-right) are fundamentally different from mainstream parties. The first model investigated in the article focuses on the question whether such “niche” parties adjust their policies in response to public opinion shifts. The authors operationalize party policy shifts between elections as changes in a party’s CMP left-right scale (“Rile”) in adjacent elections. This measure is then regressed on public opinion shifts, a dummy variable for niche party status, and the interaction of these two variables, plus 15 More precisely, if β is the true value of the coefficient, and b̂ is the estimated (biased) coefficient, then E(b̂) = β 1 + σ2m /σ2p (9) where σm is the standard deviation (measurement error) of the predictor, and σ p is the standard deviation of the true predictor. As σm → 0, E(b̂) → β. See Carroll et al. (2006) for full details. 16 For R, the simex package is available from CRAN. Information on SIMEX implementation in STATA can be found at http://www.stata.com/merror/. 20 lagged policy shifts, lagged vote share change, the interaction of these two terms, and a set of country dummies.17 In our replication of Adams et. al. Table 1, we focus on the effect of measurement error in lagged policy shift and its interaction with lagged vote share change, since error effects in covariates are what is usually considered in the classical measurement error model. Our estimate of error for lagged policy shift is based on the bootstrapped standard error. For the interaction term of lagged policy shift and lagged vote share change, we assume that change in vote share is measured without error. In this and in the replications that follow, our error estimates for each error-prone covariate is the mean of the in-sample average error variance from the bootstrapping procedure with randomized text lengths (and specified in the note to each table). The second model in Adams et al. (2006) tests whether policy adjustments tend to enhance parties’ electoral support. The authors model change in party’s vote share as a function of parties’ policy shifts towards the center of the voter distribution and away from it, and how it differs between mainstream and niche parties. The key explanatory variables are derived directly from CMP and thus are expected to be error-prone: Centrist policy shift, Noncentrist policy shift, Niche Party * Centrist policy shift, Niche Party * Noncentrist policy shift. The first variable measures the shift in party’s policy position towards the mean voter position and is calculated as the absolute value of the change in party’s position on the CMP left-right scale when a leftist party shifts right or rightist party shifts left, and zero otherwise. Non-centrist shift concerns the shift away from the center from election t − 1 to t, and it is calculated as the absolute value of the change in party’s position on the CMP left-right scale when a leftist party shifts left or rightist party shifts right and zero otherwise. The next two variables pick up the differences in the electoral effects for niche and mainstream parties in relation to centrist and non-centrist policy shifts. Two additional more control variables are based on CMP measures: Party policy convergence and Party policy convergence * Peripheral party. The former variable is constructed to capture the idea that parties’ electoral success may be affected by policy shifts of other parties. It is operationalized as the sum of all centrist policy moves by all parties in the system. The latter variable is an interaction of Party policy convergence with a dummy variable for peripheral parties. In addition to these six error17 We expect to find error contamination in both dependent variable on the left-hand side, its lagged value on the right-hand side, and an interaction of lagged dependent variable and lagged change in vote share. In order to remain within the classical measurement error (CME) domain we assume that error in first-differences in the dependent variable is uncorrelated with error in second-differences in its lagged value. Within CME it is known that measurement error in the dependent variable, if uncorrelated with other covariates, will only inflate standard error of the regression (Wooldridge 2002, 711-72). Moreover the effect of measurement error in first-difference estimation in panel data models is much higher than in level models, which may somewhat explain low reported R2 . 21 prone covariates, Model 2 in Adams et al. (2006) contains dummy variables for niche parties, governing parties, coalition governments, previous change in vote share, and several economic control variables: changes in unemployment and GDP rates and their interaction with governing party dummy. [TABLE 4 about here] Table 4 presents results for two models, taken from the two regression tables from Adams et al. (2006). For each model, we report our replication of the published results, plus the SIMEX errorcorrected results.18 The SIMEX correction of Model 1 strengthens the coefficient on Previous policy shift variable by 40%, and quadruples the coefficient on the interaction term of Previous policy shift and Previous change in vote share. Although error correction does not alter the remaining results greatly it also serves to illustrate a generally a turbid nature of error processes when the coefficient on a key explanatory variable in Model 1 (Public opinion shift) weakens by 7%, and Previous change in vote share becomes statistically significant. Neither Public opinion shift or Previous change in vote share are derived from the CMP both are affected due to their intra-model relationship with error contaminated covariates. In Model 2 a quarter of all variables are measurement error prone. Due to high levels of collinearity we should expect the SIMEX error correction to have a much more pronounced effect on the results than in Model 1. The key explanatory variable Niche Party * Centrist policy shift does not change significantly (weakens only by 5%) and still suggests that niche parties are punished at the polls for moving towards the center of voter distribution. The SIMEX error corrections unmasks several results that do not appear in the original analysis. Thus the coefficient on Previous change in vote share gains slightly and becomes barely statistically significant. However—and more importantly—the coefficient on Party policy convergence grows by 65% and becomes statistically significant, suggesting the substantive conclusion that when party policy positions move to the center of the policy space centrist parties are electorally penalized more than peripheral parties. Another interesting result concerns the effect of a policy shift away from the center of voter distribution on mainstream parties. The negative and statistically significant coefficient on Noncentrist policy shift is expected within the theoretical framework of Adams et al. (2006) but couldn’t be shown empirically. Here after the SIMEX error correction the coefficient on Noncentrist policy shift almost doubles and becomes statistically significant. This result of the SIMEX error correction on Noncentrist policy shift is shown graphically in Figure 6. 18 In this replication and the ones that follow, we report our replicated estimates instead of the published estimates, which are in each case slightly but not significantly different. This differences arose due to slight errors in data preparation in each published analysis which we uncovered in the course of replicating their results. For full details including our own replication code are available from http://kenbenoit.net/cmp/. 22 [FIGURE 6 about here] Correcting for bias induced by measurement error in Noncentrist policy shift is a good illustration of general principles behind the SIMEX method. In Figure 6, the y-axis represents the size of the coefficient. The amount of noise added to the variance of error-prone covariate is on the x-axis. It is in the form (1 − λ), where λ is a set of discrete values, usually: {−1, −.5, 0, .5, 1, 1.5, 2}. When λ = 0 the model is estimated without any error correction or error inflation—often called naive model. Each node in Figure 6 represents a mean of 100 simulations of the model with each successive increase in λ, viz. an increase in the amount of measurement error added to the model. Once a pattern is established SIMEX extrapolates back to the case of λ = −1, i.e. no measurement error. Among several extrapolants a quadratic function has been shown to have good numerical properties. As becomes clear from Figure 6 while the naive model produced a coefficient on Noncentrist policy shift of around -2.3 (standard error=1.8) and increasing amounts of measurement error brings the size of the coefficient closer to 0, however, the SIMEX correction shows the coefficient to be around -4.3 (standard error=2.3). This is a 84% increase compared to the naive model. Overall, measurement error in linear models either in one or several covariates cannot be simply dismissed as it affects the estimates produced in standard normal-linear models, sometimes in a substantively significant way. 5.2.2 Hix, Noury, and Roland (2006) Hix, Noury & Roland (2006) are concerned with the content and character of political dimensions in the European Parliament (EP). Following an inductive scaling analysis of all roll-call votes in the EP from 1979 and 2001, Hix, Noury & Roland (2006) attempt to validate their interpretations of these inductively derived dimensions by mapping variation in the dimensional measures to exogenous data on nation party policy positions. The authors conclude that the first dimension is the classic left-right dimension, while the second dimension primarily concerns party positions on European integration as well as government-opposition conflicts. Substantive interpretation of dimensions is achieved by regressing the mean position of each national party’s delegation of MEPs on each dimension in each parliament on two sets of independent variables. The first set of variables includes exogenous measures of national party positions on the left-right, social and economic left-right, and pro-/anti-EU dimensions. The second set relates to government-opposition dynamics and consists of categorical variables whether a national party was in government and whether the party had a European Commissioner, as well as dummy variables for each European party group, each EU member state, and each (session of) European parliament. 23 The measures of national party positions of interest to us here are either taken directly (like leftright) or constructed from the CMP data (EU, economic and social left-right). National party positions on the EU dimension are taken as the difference between positive mentions of EU in party manifesto (category PER108) and negative mentions (category PER110). Party positions on economic and social dimensions follow the suggestion in Laver & Garry (2000, 628-629), where the economic dimension is constructed as the difference between “Capitalist Economics” and “State Intervention” clusters, and social dimension is constructed as the difference between “Social Conservatism” and “Anti-establishment views” clusters from Laver & Budge (1992, 24).19 We expect to find error contamination in all models containing variables derived from the CMP data. [TABLE 5 about here] Table 5 reports our replicated estimations of four models in Hix, Noury & Roland (2006) along with error-corrected measurements based on our bootstrapped variances with randomized text length. In each error-corrected estimation, we used simulation-extrapolation with an error variance estimated as the mean of the in-sample average error variance from the bootstrapping procedure with randomized text lengths. Although Hix, Noury & Roland (2006) present numerous estimations, here we examine only those that used CMP-based policy measures as covariates in their regressions on the first dimension from their inductive scaling procedure. Model 2 is the most basic model regressing their first estimated dimension on several covariates: general left-right and EU dimensions taken from the CMP, plus categorical variables whether a national party was in government and whether the party had a European Commissioner, and as dummy variables for each (session of) European parliament. Model 3 replaces the general leftright with two more specific left-right dimensions drawn from the CMP: economic left-right and social left-right. Model 6 extends Model 3 to include dummy variables for each European party group. Model 9 extends Model 6 to include a set of country dummies. The first thing that is apparent from the Table 5 is that the SIMEX error correction had only minimal effect on most coefficients with all error-prone covariates increasing only slightly and most other covariates decreasing slightly. By far the biggest correction is achieved on the coefficient on national party position on EU dimension. This is to be expected, since the “pro vs anti EU” variable constructed from the CMP data uses only two of the 56 substantive coding categories and is relatively sparsely populated with coded data. The coefficient on Pro-/Anti-EU variable with the SIMEX error 19 Thus, economic dimension = (per401 + per402 + per407 + per414 + per505) - (per403 + per404 + per406 + per412 + per413); social dimension = (per203 + per305 + per601 + per603 + per605 + per606)- (per204+per304+per602+per604) 24 correction increased by 78% in Model 2, doubled in Model 3, increased by 2.5 times in Model 6, and by 2.2 times in Model 9. In the latter two models, these large increases in the size of the estimated coefficients means that the Pro-/Anti-EU variable becomes statistically significant, where it failed to do so in the naive models. For example, in Model 3 it would suggest that for the mean of each national party’s delegation on the first dimension to move from the center to extreme right position, ceteris paribus, its position on the EU dimension would have to change by 23 and not by 45 points as in the naive estimates. Party policy positions on EU dimension have an in-sample range of 33. Thus while such a whole-sale move from one end of political spectrum to another seems highly unlikely, it is nevertheless possible under error corrected results. Substantively, for an explanation of the mean position of national party’s delegation its national party’s position on the EU dimension is more important than its position on general left-right or substantive economic and social left-right. Moreover, the SIMEX error correction of Pro-/Anti-EU variable and its subsequent regaining of importance in Models 6 and 9 forces rethinking of some of the conclusions in the article. Hix, Noury & Roland (2006) expect to find the first dimension to be explained by policy positions of political parties on general (domestic) left-right issues, while the second dimension should be explained by policy positions of political parties on the EU dimension.20 Indeed, the estimates from naı̈ve models, viz. not taking account of measurement error bias in CMP derived variables, provide ample evidence to support original theoretical expectations. Once party groups are controlled for, party positions on EU dimension have no role to play in interpreting the first dimension. However, applying the SIMEX error correction to the error-prone Pro-/Anti-EU covariate, not only more than doubles it in magnitude but also makes it statistically significant. Thus, the first dimension still contains Pro-/Anti-EU and not only left-right as claimed by the authors. This calls into question the conclusion of the article that political parties’ positions in EU policy making process are shaped predominantly by important socioeconomic positions in domestic politics rather than preferences about European integration. Instead, the error-corrected results suggest that there is no clear distinction between general left-right positioning and positioning on European integration when it comes to the primary axis of contestation in the European Parliament, and that if anything, positioning on European Union appears to be even more important than general left-right in explaining this primary axis. 20 In the words of Hix, Noury & Roland (2006) interpreting their results from the naive model: EU policies of national parties and national party participation in government are only significant without the European party group dummies. This means that once one controls for European party group positions these variables are not relevant explanatory factors on the first dimension. (502) 25 5.2.3 Golder (2006) Golder (2006) applied non-linear probit models to investigate the formation of pre-electoral coalitions between parties and the forces that may influence the likelihood of such coalitions. Pre-electoral coalitions between dyads of parties in a system are modeled as functions of Ideological incompatibility, Polarization and Electoral threshold (plus the interaction between these two variables), Coalition size and Coalition size squared, Asymmetry, and an interaction between Coalition size and Asymmetry. We can expect three variables to be particularly error-prone, since they are based on the CMP left-right measure: Ideological incompatibility, Polarization and Polarization * Electoral threshold. Ideological incompatibility measures the ideological distance between parties in a dyad. It is directly computed as the absolute value of the difference of the CMP “rile” score for parties in a dyad. We compute error variance for the incompatibility measure combining error variances for individual parties in a dyad utilizing the analytic definition of difference between two uncorrelated variances (see Appendix A). Polarization measures ideological dispersion in a system and is calculated as the difference of the CMP “ rile” scores for the biggest left-wing and the biggest right-wing parties in political system. We compute error variance for Polarization following the same procedure as for Ideological incompatibility. [TABLE 6 about here] Table 6 applies the SIMEX correction to Golder’s probit model predicting the probability of preelectoral coalition formation.21 The majority of covariates experience only minimal changes after correcting for measurement error. As a result of the SIMEX error correction, the coefficient for Polarization changes sign from positive to negative, but remains statistically indistinguishable from zero. Error correction on Incompatibility results in a 18% coefficient increase, thus making the creation of pre-electoral coalition even less probable with the increase in ideological incompatibility of potential partners. Although none of the substantive results changed with the SIMEX error correction, our implementation of error-correction for this model demonstrates that error correction using our estimates of CMP error variances is straightforward to apply to non-linear models. 21 We depart from original estimation in two ways here. First, we use newer CMP dataset made available with the publication of MPP2, which results in some very slight differences in the dataset. Second, we use estimate only a fixed effects probit model, rather than the additional random effects probit model as in the original piece to avoid complicating the computation. Neither change resulted in any substantive differences in our replications versus the original results. 26 6 Concluding Remarks and Recommendations When we treat text as data that are prone to measurement error, we can analyze bodies of text using wellknown statistical tools. The implications of this are very general, although in this paper we focus on the text found in party manifestos and analyzed by the CMP; this is by far the largest and most significant dataset in the profession that is based on systematic text analysis. We show how manifestos provide systematic information about their authors, in the form of random variables deposited as text through process of authorship that is inherently stochastic, even when the author’s underlying position is fixed. In addition, we characterize one important aspect of the CMP coding process as stochastic: non-systematic variation in identifying coding units, or quasi-sentences, on which all CMP measures are based. Putting together these two stochastic processes and resampling from quasi-sentence counts reconstructed from the CMP dataset, we show how to compute errors in a practical and theoretically supported way for the entire set of CMP measures. We also show how these errors affect descriptive and causal inferences based on CMP measures. Building on this method, we offer a “corrected” version of the CMP dataset with bootstrapped standard errors for all key estimates. A full version of these corrections is available on our website http://kenbenoit.net/cmp/. Our task in this paper has been fundamentally constructive. The CMP dataset is potentially a very valuable, indeed currently unique, resource dealing with party positions on many dimensions in a cross-national time series context. We therefore regard it as essential that these data be used in a statistically rigorous way. Hitherto, third-party authors have either ignored or swept under the carpet the self-evident fact that the CMP data are error-prone, yet come with no estimate of associated error. The huge volume of CMP source documents, written in many languages, were coded once only by a single human coder and will likely never all be recoded many times by different coders. (Some of them may even be lost). We offer a way around this problem with a method for simulating the level of error in key CMP variables and a way to make use of these simulated errors when deploying CMP variables as covariates in statistical analyses—by far their most common use. When applied to descriptive analyses, we have shown that for many of the apparent “differences” in CMP estimates of party policy positions— differences over time in the position of one party or differences between parties at one point in time—are probably attributable to measurement error rather than to real differences in policy position. Thus only about one quarter of all CMP-estimated “movements” in parties’ left-right policy positions over time were assessed on the basis of our simulations to be statistically significant. Error in the measurement of covariates is likely to bias regression coefficients, yet none of the 27 large volume of published empirical research using CMP data takes any account whatsoever of this. We replicate three prominent recently-published articles in which error-prone CMP variables are used as covariates, and show how to correct these using a SIMEX error correction model, parameterized by our bootstrapped estimates of likely error. The likely systematic effect of error contaminated covariates is attenuation bias, and this is what we tend to find in each case. Some error-corrected effects are stronger, and more significant, than those estimated in models taking no account of error in the covariates. This can cause substantively important reinterpretation of results in cases where inferences have relied on the lack of significance of some covariates. A good example is what emerges as the potentially flawed inference that national party policy positions on the EU have no influence in their EP roll-call voting behavior, an inference that is reversed once account is taken of error-contamination in the CMP dataset’s sparsely-populated variables measuring EU policy. While we have taken an important first step towards providing a practical and theoretically supported means to estimate non-systematic measurement error in CMP estimates, the solution we provide here is hardly the last word on this topic. Analyses of coder differences and/or coder error, for instance, could uncover systematic error leading to bias, which we have not considered in this paper; this possibility deserves investigation in future work. Additional experiments with coder reliability using additional texts would also provide important information on the second type of error we identify. Finally, other means of implementing error correction models are certainly possible, including Bayesian-MCMC methods that can take into account the unit-specific nature of error in our error estimates. Indeed, we hope that our focus of attention on error in the widely used CMP estimates will stimulate a broader dialogue on measurement error in many of our most commonly used measures in political science, such as opinion survey results, expert survey measures, or other computed quantities. Given our knowledge of measurement error and the wide availability of techniques for dealing with such error, there is no longer any excuse for future scholars to use error-prone measures as if these were error free. 28 A Computing the Analytic Variance of CMP “PER” Categories Assume that E(pi j ) = πi j , where each πi j represents the true and unobservable policy position of country-party-date unit i on issue j. For any given Xi j from the multinomial distribution, Var(Xi j ) = ni pi j (1 − pi j ). To derive the analytic variance for the percentage quantity Var(pi j 100), we have (dropping the manifesto index i for simplicity): E(X j ) = np j x j = np j Var xj = pj n 1 xj n = Var(p j ) 1 Var(x j ) = Var(p j ) n2 1 np j (1 − p j ) = Var(p j ) n2 1 p j (1 − p j ) = Var(p j ) n Translating into the CMP’s percentage metric (p j ∗ 100): 10, 000Var(p j ) = Var(p j 100) = SD(p j 100) = 10, 000 p j (1 − p j ) n 10, 000 p j (1 − p j ) n q 100 √ p j (1 − p j ) n To compute the variance for additive categories like the pro-/anti-EU scale (PER108 minus PER110) or the 26-category “right-left” scale, the covariances between categories must also be estimated, since it is a property of variance that Var(aX +bY ) = a2 Var(X)+b2 Var(Y )+2abCov(X,Y ). Rather than attempt to estimate these unknown—but certainly non-zero—covariances from finite samples, we recommend the bootstrapping procedure instead of the analytic computation. 29 References Adams, James. 2001. “A Theory of Spatial Competition with Biased Voters: Party Policies Viewed Temporally and Comparatively.” British Journal of Political Science 31(1):121–158. Adams, James, M. Clark, L. Ezrow & G. Glasgow. 2006. “Are Niche Parties Fundamentally Different from Mainstream Parties? The Causes and the Electoral Consequences of Western European Parties’ Policy Shifts, 1976–1998.” American Journal of Political Science 50(3):513–529. Alt, James & David Lassen. 2006. “Transparency, Political Polarization, and Political Budget Cycles in OECD Countries.” American Journal of Political Science 50(3):530–550. Baron, D.P. 1991. “A Spatial Bargaining Theory of Government Formation in Parliamentary Systems.” The American Political Science Review 85(1):137–164. Bartolini, S. & P. Mair. 1990. Identity, Competition, and Electoral Availability: The Stabilisation of European Electorates 1885-1985. Cambridge University Press. Benoit, Kenneth & Michael Laver. 2006. Party Policy in Modern Democracies. London: Routledge. Benoit, Kenneth & Michael Laver. 2007. “Benchmarks for Text Analysis: A Reply to Budge and Pennings.” Electoral Studies 26:130–135. Bräuninger, T. 2005. “A partisan model of government expenditure.” Public Choice 125(3):409–429. Budge, Ian, David Robertson & Derek Hearl. 1987. Ideology, Strategy and Party Change: Spatial Analyses of Post-War Election Programmes in 19 Democracies. Cambridge: Cambridge University Press. Budge, Ian, Hans-Dieter Klingemann, Andrea Volkens, Judith Bara & Eric Tanenbaum. 2001. Mapping Policy Preferences: Estimates for Parties, Electors, and Governments 1945–1998. Oxford: Oxford University Press. Carroll, Raymond J., David Ruppert, Leonard A. Stefanski & Ciprian M. Crainiceanu. 2006. Measurement Error in Nonlinear Models: A Modern Perspective. Number 105 in “Monographs on Statistics and Applied Probability” 2nd ed. Boca Raton, FL: Chapman and Hall/CRC. Duso, T. & L.H. Röller. 2003. “Endogenous Deregulation: Evidence from OECD Countries.” Economics Letters 81(1):67–71. Efron, Bradley. 1979. “Bootstrap Methods: Another Look at the Jackknife.” The Annals of Statistics 7(1):1–26. Efron, Bradley & Robert Tibshirani. 1994. An Introduction to the Bootstrap. New York: Chapman and Hall/CRC Hall. Erikson, R.S., M.B. Mackuen & J.A. Stimson. 2002. The Macro Polity. Cambridge University Press. Evans, G. & P. Norris. 1999. Critical Elections: British Parties & Voters. Sage Publications. Golder, Sona N. 2006. “Pre-Electoral Coalition Formation in Parliamentary Democracies.” British Journal of Political Science 36(2):193–212. Hausman, J. 2001. “Mismeasured Variables in Econometric Analysis: Problems from the Right and Problems from the Left.” The Journal of Economic Perspectives 15(4):57–67. 30 Hix, Simon, Abdul Noury & Gérard Roland. 2006. “Dimensions of Politics in the European Parliament.” American Journal of Political Science 50(2, April):494–511. Janda, K., R. Harmel, C. Edens & P. Goff. 1995. “Changes in Party Identity: Evidence from Party Manifestos.” Party Politics 1(2):171–196. Kim, Heemin & Richard C. Fording. 1998. “Voter Ideology in Western Democracies, 1946-1989.” European Journal of Political Research 33(1):73–97. King, Gary, Robert Keohane & Sidney Verba. 1994. Designing Social Inquiry. Princeton: Princeton University Press. Klingemann, Hans-Dieter, Andrea Volkens, Judith Bara, Ian Budge & Michael McDonald. 2006. Mapping Policy Preferences II: Estimates for Parties, Electors, and Governments in Eastern Europe, European Union and OECD 1990-2003. Oxford: Oxford University Press. Laver, Michael & Ian Budge. 1992. Party policy and government coalitions. New York: St. Martin’s Press. Laver, Michael & John Garry. 2000. “Estimating Policy Positions from Political Texts.” American Journal of Political Science 44(3):619–634. Laver, Michael, Kenneth Benoit & John Garry. 2003. “Estimating the policy positions of political actors using words as data.” American Political Science Review 97(2):311–331. Mair, P. 1987. The Changing Irish Party System: Organisation, Ideology and Electoral Competition. Pinter. Meguid, B. 2005. “Competition Between Unequals: The Role of Mainstream Party Strategy in Niche Party Success.” American Political Science Review 99(3):347–359. Milner, H.V. & B. Judkins. 2004. “Partisanship, Trade Policy, and Globalization: Is There a Left-Right Divide on Trade Policy?” International Studies Quarterly 48(1):95–120. Monroe, Burt & Ko Maeda. 2004. “Talk’s cheap: Text-based estimation of rhetorical ideal-points.” Working paper: Michigan State University. Petry, F. 1988. “The Policy Impact of Canadian Party Programs: Public Expenditure Growth and Contagion from the Left.” Canadian Public Policy/Analyse de Politiques 14(4):376–389. Petry, F. 1991. “Fragile Mandate: Party Programmes and Public Expenditures in the French Fifth Republic.” European Journal of Political Research 20(2):149–171. Petry, F. 1995. “The Party Agenda Model: Election Programmes and Government Spending in Canada.” Canadian Journal of Political Science/Revue canadienne de science politique 28(1):51–84. Ray, L. 1999. “Measuring party orientations towards European integration: Results from an expert survey.” European Journal of Political Research 36(2):283–306. Robertson, David. 1976. A Theory of Party Competition. London and New York: Wiley. Schofield, Norman. 1993. “Political competitition and multiparty coalition governments.” European Journal of Political Research 23(1):1–33. Sigelman, L. & H.B. Emmett. 2004. “Avoidance or Engagement? Issue Convergence in U.S. Presidential Campaigns, 1960-2000.” American Journal of Political Science 48(4):650–661. 31 Slapin, Jonathan & Sven-Oliver Proksch. 2007. “A Scaling Model for Estimating Time-Series Policy Positions from Texts.” Paper presented at the annual meeting of the Midwest Political Science Association, April 12. Stefanski, L. A. & J. R. Cook. 1995. “Simulation-Extrapolation: The Measurement Error Jackknife.” Journal of the American Statistical Association 90(432, December):1247–1256. Strom, K. & J.Y. Lejpart. 1989. “Ideology, Strategy, and Party Competition in Postwar Norway.” European Journal of Political Research 17(3):263–288. Thomson, R., J. Boerefijn & F. Stokman. 2004. “Actor alignments in European Union decision making.” European Journal of Political Research 43(2):237–261. Van Der Brug, W., M. Fennema & J. Tillie. 2005. “Why Some Anti-Immigrant Parties Fail and Others Succeed: A Two-Step Model of Aggregate Electoral Support.” Comparative Political Studies 38(5):537–573. Volkens, Andrea. 2001a. Manifesto Research Since 1979. From Reliability to Validity. In Estimating the Policy Positions of Political Actors, ed. Michael Laver. London: Routledge pp. 33–49. Volkens, Andrea. 2001b. Quantifying the Election Programmes: Coding Procedures and Controls. In Mapping Policy Preferences: Parties, Electors and Governments: 1945-1998: Estimates for Parties, Electors and Governments 1945-1998, ed. Ian Budge, Hans-Dieter Klingemann, Andrea Volkens, Judith Bara, Eric Tannenbaum, Richard Fording, Derek Hearl, Hee Min Kim, Michael McDonald & Silvia Mendes. Oxford: Oxford University Press. Volkens, Andrea. 2007. “Strengths and Weaknesses of Approaches to Measuring Policy Positions of Parties.” Electoral Studies 26(1):108–120. Webb, P.D. 2000. The Modern British Party System. Sage Publications. Wooldridge, J. 2002. Econometric Analysis of Cross-sectional and Panel Data. The MIT Press, Cambridge, MA. 32 Total Quasi-sentences (0,5] (5,10] (10,15] (15,20] (20,25] (25,50] (50,75] (75,100] (100,200] (200,250] (250,500] (500,1000] (1000,2000] (2000,5000] > 5000 Total N 0 14 14 29 41 277 303 251 711 208 521 343 216 79 4 Cumulative 0.0% 0.5% 0.9% 1.9% 3.3% 12.5% 22.5% 30.9% 54.5% 61.4% 78.7% 90.1% 97.2% 99.9% 100.0% 3,011 Table 1: Total quasi-sentences in CMP dataset, for manifestos where total is reported. Percentage Uncoded 0 (0,5] (5,10] (10,15] (15,20] (20,30] (30,40] (40,50] (50,65] (65,80] (80,95] (95,100] N 864 1,121 375 193 119 129 87 50 30 9 7 - Total 2,984 Cumulative 29.0% 66.5% 79.1% 85.6% 89.5% 93.9% 96.8% 98.5% 99.5% 99.8% 100.0% 100.0% Table 2: Percentage of uncoded quasi-sentences in the CMP dataset, for manifestos where percentage uncoded is reported. 33 Statistically Significant Change? No Yes N/A Total Elections % total % applicable 1,509 590 778 52.5% 20.5% 27.0% 71.9% 28.1% – 2,877 100.0% 100.0% Table 3: Comparative over-time mapping of policy movement on Left-Right measure, taking into account statistical significance of shifts. 34 Model 1 Model 2 OLS Replication 0.0591 (0.1469) OLS with SIMEX Correction 0.0836 (0.146) OLS Replication 2.1212 (2.0939) OLS with SIMEX Correction 3.7403 (2.4541) Previous policy shift -0.5127 (0.0772) -0.7158 (0.1036) Previous policy shift * Previous change in vote share -0.0044 (0.0184) -0.0171 (0.0189) 1.089 (1.3607) 0.5123 (1.4056) Noncentrist policy shift -2.3335 (1.837) -4.2818 (2.3389) Niche party * Centrist policy shift -5.4069 (2.823) -5.1522 (2.8178) Niche party * Noncentrist policy shift 2.6163 (2.6946) 3.6774 (2.8245) Party policy convergence -1.2513 (0.8416) -2.0689 (1.058) Party policy convergence * Peripheral party 1.3238 (1.1628) 0.9724 (1.1732) -0.1548 (0.0943) -1.4181 (1.8592) -0.166 (0.0939) -1.856 (1.8926) 5.4118 (1.7649) -3.0767 (1.7271) 0.9764 (1.1872) -0.8533 (0.7263) -0.3958 (0.3977) 0.1045 (1.0674) -0.021 (0.4996) -0.1594 (1.8179) 6.0491 0.2575 0.0832 122 6.0475 (1.8046) -3.114 (1.7206) 1.2103 (1.2048) -0.9707 (0.729) -0.5087 (0.4063) 0.0079 (1.0657) -0.0351 (0.4974) 0.4034 (1.8334) 4.4781 0.2338 0.054 122 Variable Intercept Centrist policy shift Previous change in vote share Niche party Public opinion shift Public opinion shift * Niche party Directional public opinion shift 0.0176 (0.0099) 0.1059 (0.1341) 0.8447 (0.2304) -1.5508 (0.4433) 0.0165 (0.0102) 0.1063 (0.133) 0.9008 (0.2338) -1.5675 (0.4488) Governing party Governing in coalition Change in unemployment rate Change in GDP Governing party * Change in unemployment rate Governing party * Change in GDP Peripheral party RMSE R2 Adj R2 N 0.5908 0.3578 0.2982 154 0.6057 0.3247 0.262 154 Table 4: Results of SIMEX error correction in Adams et. al. (2006). Country dummies are included in the estimations but not reported here. Italicized and grayed-out variables are error-corrected as follows: (Model 1) Previous policy shift=.43, interaction of Previous policy shift and Previous change in vote share=.43; (Model 2) Centrist policy shift=0.228, Noncentrist policy shift=0.181, Party policy convergence=0.847, interaction of Niche party and Centrist policy shift=0.072, interaction of Niche party and Noncentrist policy shift=0.048, interaction of Party policy convergence and Peripheral party=0.390. Coefficients in bold are statistically significant at the p ≤ .1 level; SIMEX standard errors are based on jacknife estimation. 35 Variable Model 2 OLS OLS Replication SIMEX Model 3 OLS OLS Replication SIMEX Model 6 OLS OLS Replication SIMEX Model 9 OLS OLS Replication SIMEX Intercept 0.045 (0.0581) 0.044 (0.0584) -0.149 (0.0596) -0.181 (0.0604) 0.364 (0.0440) 0.339 (0.0459) 0.332 (0.0667) 0.296 (0.0693) Pro-/Anti-EU 0.018 (0.0062) 0.032 (0.0076) 0.019 (0.0062) 0.038 (0.0079) 0.004 (0.0037) 0.010 (0.0048) 0.005 (0.0038) 0.011 (0.0049) Left-Right 0.012 (0.0009) 0.012 (0.0010) Social L-R 0.011 (0.0023) 0.014 (0.0025) 0.004 (0.0013) 0.005 (0.0015) 0.003 (0.0014) 0.004 (0.0015) Economic L-R 0.024 (0.0022) 0.024 (0.0022) 0.006 (0.0015) 0.007 (0.0015) 0.007 (0.0016) 0.008 (0.0017) 0.022 (0.0304) 0.061 (0.0259) -0.560 (0.0351) -0.641 (0.2115) -0.168 (0.0359) -1.003 (0.0510) 0.077 (0.0993) -0.820 (0.0496) 0.099 (0.0623) -0.230 (0.0539) -0.785 (0.0568) 0.447 (0.1244) 0.2013 0.8136 0.8029 349 0.022 (0.0304) 0.056 (0.0259) -0.548 (0.0355) -0.630 (0.2114) -0.164 (0.0359) -0.978 (0.0532) 0.071 (0.0991) -0.786 (0.0527) 0.110 (0.0622) -0.225 (0.0539) -0.765 (0.0577) 0.453 (0.1242) 0.2022 0.8119 0.8011 349 0.030 (0.0315) 0.034 (0.0271) -0.556 (0.0350) -0.693 (0.2097) -0.178 (0.0360) -1.017 (0.0524) 0.119 (0.1018) -0.797 (0.0511) 0.079 (0.0644) -0.254 (0.0551) -0.804 (0.0572) 0.393 (0.1245) 0.1923 0.83 0.8122 349 0.031 (0.0314) 0.030 (0.0271) -0.544 (0.0357) -0.671 (0.2104) -0.173 (0.0360) -0.990 (0.0546) 0.107 (0.1019) -0.760 (0.0548) 0.091 (0.0646) -0.245 (0.0552) -0.781 (0.0582) 0.409 (0.1243) 0.1935 0.8279 0.8099 349 Commissioner In Government 0.104 (0.0495) 0.107 (0.0434) 0.095 (0.0492) 0.086 (0.0440) 0.079 (0.0495) 0.103 (0.0434) 0.070 (0.0492) 0.077 (0.0439) 0.3585 0.4089 0.395 349 0.3621 0.3971 0.3829 349 0.3578 0.4112 0.3956 349 0.3645 0.3892 0.373 349 Socialists Italian Communists and allies Liberals Greens British Conservatives and allies Radical left French Gaullists and allies Non-attached members Regionalists Radical right RMSE R2 Adj R2 N Table 5: Results of SIMEX error correction in Hix, Noury & Roland (2006, 503–504, Table 4). All models include dummies for parliament, and Model 9 includes country dummies but these are not shown. Italicized and grayed-out variables are error-corrected as follows: Left-Right=4.48, Social L-R=2.27, Economic LR=2.06, Pro-/anti-EU=1.83. Coefficients in bold are statistically significant at the p ≤ .1 level; SIMEX standard errors are based on jacknife estimation. 36 Probit Replication Probit with SIMEX Correction Intercept -2.0000 (0.2121) -1.8220 (0.2267) Incompatibility -0.0146 (0.0024) 0.0014 (0.0028) 0.0003 (0.0002) -0.0173 (0.0029) -0.0026 (0.0034) 0.0005 (0.0002) 0.0170 (0.0067) 0.0409 (0.0088) -0.0004 (0.0001) 0.0243 (0.2455) -0.0226 (0.0071) 1250 3385 0.0114 (0.0072) 0.0408 (0.0088) -0.0004 (0.0001) 0.0113 (0.2467) -0.0223 (0.0071) Variable Polarization Polarization * Threshold Threshold Seat Share Seat Share2 Asymmetry Asymmetry * Seat Share -2*Log-Likelihood N 3385 Table 6: Results of SIMEX error correction in Golder (2006, 203, Table 1, Model 2). Italicized and grayed-out variables are error-corrected with standard deviations of 8.54, 8.25, and 8.25 respectively. Coefficients in bold are statistically significant at the p ≤ .1 level; SIMEX standard errors are based on jacknife estimation. 37 Underlying position to be communicated by text author; fundamentally unobservable by others Stochastic text generation process Observed text used as indicator author’s position Stochastic text coding process Data characterizing unobservable underlying position of author in terms of coded observed text Figure 1: Summary of stochastic processes involved in the generation of policy texts 38 (b) ● ● ● ● ● ●● ●●● ●● ●● ● ●●●● ● ● ●●● ●● ●●●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ●● ●●● ● ● ● ● ●●● ● ● ● ● ● ●● ●● ● ● ● ● ● ●● ● ● ● ● ●● ●● ● ● ● ● ●●●● ● ●● ● ●● ●● ●● ● ● ● ●● ●● ●● ●● ● ●● ● ●●● ●● ●● ● ●●● ● ● ● ● ● ●● ● ● ●● ● ●● ●● ● ● ● ●● ● ● ●● ● ● ●● ● ●● ●● ● ● ●● ● ● ●● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●●● ●● ● ● ●● ● ●● ●●● ● ● ● ● ● ● ●●●● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ●●● ●● ● ● ●● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●●● ●● ● ●●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ●●● ●● ● ● ●● ● ●● ●● ●● ● ●● ●●● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ●●● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●● ● ●● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ●● ●●● ● ●● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ●● ●● ● ● ● ● ● ●● ● ● ● ● ●●●●● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ●●● ● ●●● ● ● ● ●●● ●● ●● ● ●● ●● ●● ●● ●● ●● ● ● ● ● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●●● ● ● ● ● ● ●●● ●● ● ●●● ● ●● ●●● ●● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ●● ●● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ●● ● ●● ● ●● ● ● ●● ● ●●●● ●●● ●● ● ●● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●●●● ● ● ● ● ● ● ●●● ● ●● ●● ●●● ● ●● ●● ●● ● ●● ●●● ● ● ●● ●●● ● ● ● ● ● ●●● ● ● ●●● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ●● ● ● ●●● ●● ● ●●● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ●●●● ● ● ●● ● ●●●●● ● ● ●● ● ● ●●●● ●● ●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ● ● ●● ●●● ● ●● ● ● ● ●●● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ●●● ● ● ● ● ●●● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ●● ●● ● ●● ● ● ●● ● ●●●●● ●● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ● 20.0 10.0 1.0 2.0 5.0 RILE Standard Error ● ● ● ● ● 20 50 100 ● ● ●● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ●●●● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ●● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●●●●● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●●●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ●●● ● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ●● ● ●● ● ● ● ● ●●● ● ● ●● ●● 0.5 0.2 0.5 1.0 2.0 5.0 ● ● 0.1 PER501 (Environment) Standard Error 10.0 (a) 500 ● 2000 5000 10 20 Number of Quasi−Sentences 50 200 500 2000 Number of Quasi−Sentences ● 0.2 0.5 1.0 2.0 5.0 ●● ● ● ●●●● ● ●●●●● ●● ● ●● ●● ● ● ● ●● ●● ●● ● ● ●●●●●●●● ● ● ● ●●● ●● ● ● ● ●● ●●● ●●● ●●● ● ● ●● ●●● ● ● ● ●● ●● ● ●●● ● ●● ●●● ● ●●● ● ●● ●● ● ● ●●●● ●●●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●●●● ●● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ●●●●● ● ●● ● ● ● ●● ●●● ● ●●● ●● ● ●● ● ● ● ● ● ●● ●● ●● ● ●● ● ● ●● ●● ● ● ●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ●● ●●● ● ● ●● ● ● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●●●● ● ●●● ● ● ● ● ● ● ●● ● ●●● ● ●● ●● ● ● ● ●● ● ● ● ● ●●● ● ● ● ● ●● ● ● ● ● ●● ● ● ●● ● ●●● ●● ● ● ● ● ● ● ●● ● ● ●●● ● ●●● ●● ● ● ● ●● ● ● ●● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ●●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ●● ●● ● ● ● ● ●● ● ● ●● ● ●● ● ● ● ● ●●● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ●● ● ● ●● ●● ● ●● ● ● ●● ● ●● ● ● ● ●●●● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●●● ●● ● ●●● ● ● ● ● ●●● ● ●● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ●●● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●●● ● ●●● ●●● ● ● ● ●● ● ●● ● ● ● ●● ● ● ●● ● ●●● ● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ●● ● ● ●●● ●● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●●● ● ● ● ● ● ● ● ●● ●● ● ● ●● ●● ●●●● ● ●● ● ●● ● ●● ● ● ●● ● ●● ●● ●● ●●● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ●● ● ●● ● ● ●● ● ●● ● ● ●●● ●● ● ●●● ●●●● ●●●● ● ● ● ● ●● ●● ●● ●● ●●●●● ● ● ● ● ● ●● ● ●●●● ●●●● ●● ●●● ● ● ● ●● ● ●● ●● ● 0.1 Bootstrapped PER501 (Environment) Standard Errors 10.0 (c) ● ● 0.1 0.2 0.5 1.0 2.0 5.0 10.0 Analytical PER501 (Environment) Standard Errors Figure 2: Comparing estimated error variances for Environment (PER501) and the CMP Left-Right scale (RILE). Axes use log scales. 39 10.0 (a) Environment (PER501) 2.0 1.0 0.5 0.2 ● 0.1 S.E. with random N 5.0 ● ● ● ● ● ● ●● ●● ● ●●● ● ●● ●●● ● ● ● ● ● ● ● ●●●●● ● ● ● ● ● ● ● ●● ●●●●● ● ● ● ●● ●●●●●● ● ● ●● ● ● ●● ●●● ●● ●● ● ●● ● ● ● ● ● ● ●●●●● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●●● ●● ● ●● ● ● ●● ● ●●● ●●● ●● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ●●●●● ● ● ● ● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●●● ● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ●●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ●● ● ● ● ●●●●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●●● ● ● ● ● ●●● ●● ● ● ● ● ●● ● ● ● ● ● ●● ●● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●●● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ●● ● ● ● ●●●● ● ● ●● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●● ● ● ● ●● ●● ● ● ●● ●●● ● ● ● ● ● ● ● ● ●●●● ● ●● ● ●● ●● ● ●●● ●● ●● ● ● ● ● ●● ● ● ●●● ●● ● ● ● ● ● ●●● ● ●● ● ● ●● ● ●● ●●● ● ● ●● ● ● ● ● ● ● ● ● ● ●●● ●●● ● ●● ● ● ● ● ● ● ● 0.1 0.2 0.5 1.0 2.0 5.0 10.0 S.E. with fixed N (b) RILE 10.0 5.0 2.0 ● ● ● ● ● ●●● ● ● ●● ●● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ●● ●● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●●● ● ● ● ● ● ●● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●●●● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●●● ● ●●●● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ●● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●●● ● ● ●●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●●● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ● ● ● ● ●● ● ●●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ●●● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ● ● ● ●● ● ●● ● ● ●●● ●●● ●● ● ● ● ●●● ●● ●● ● ●● ● ● ●● ● ● ● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●●● ●● ●● ● ●● ● ● ● ● ●●● ● ●●●● ● ● ● ● ● ●● ● ●● ● ● ● ● ●●● ● ●● ●● ●● ●● ● ●● ● ● ● ● ● ●● ● ● 0.5 1.0 S.E. with random N 20.0 ● ● ● 0.5 1.0 2.0 5.0 10.0 20.0 S.E. with fixed N Figure 3: Comparing bootstrapped standard errors when n is fixed to when n is randomly determined, for Environment (PER501) and the CMP left-right scale (RILE). Axes use log scales. 40 25 20 15 10 ● ● ● 5 ● ● ● ● ● ● ● ● ● ● ● −5 0 CMP Environment Dimension (PER501) ● 1950 1960 1970 1980 1990 2000 German CDU−CSU Figure 4: Movement on environmental policy of German CDU-CSU over time. Movement of Dashed line is % environment with 95% CI; dotted line is the number of quasi-sentences per manifesto coded PER501. 41 (a) Germany 2002 CDU−CSU ● SPD ● FDP ● Greens ● PDS ● −40 −30 −20 −10 0 10 20 30 ● CMP Right−Left Dimension (b) France 2002 FN ● UDF ● UMP ● Verts ● PS ● PCF ● −30 −20 −10 0● 10 20 30 40 CMP Right−Left Dimension Figure 5: Left-Right placement of German and French parties in 2002. Bars indicate 95% confidence intervals. 42 −1.0 ● ● ● 2.5 3.0 −2.0 −3.0 ● −4.0 Noncentrist policy shift ● ● 0.0 0.5 1.0 1.5 2.0 (1 + λ) Figure 6: SIMEX plot for Noncentrist policy shift in Table 2 in Adams et. al. (2006). λ is the amount of noise added to variance of error-prone covariate. (1 + λ) = 1 is the naive estimate with no extra noise added. Measurement error is added in the increments of .5 to establish a pattern, and then extrapolated back (quadratically) to the case of no measurement error (λ = −1). 43
© Copyright 2026 Paperzz