WIDER Working Paper No. 2013/023 Measuring

econstor
A Service of
zbw
Make Your Publications Visible.
Leibniz-Informationszentrum
Wirtschaft
Leibniz Information Centre
for Economics
Bratton, Michael
Working Paper
Measuring government performance in public opinion
surveys in Africa: Towards experiments?
WIDER Working Paper, No. 2013/023
Provided in Cooperation with:
United Nations University (UNU), World Institute for Development
Economics Research (WIDER)
Suggested Citation: Bratton, Michael (2013) : Measuring government performance in public
opinion surveys in Africa: Towards experiments?, WIDER Working Paper, No. 2013/023, ISBN
978-92-9230-600-7
This Version is available at:
http://hdl.handle.net/10419/81051
Standard-Nutzungsbedingungen:
Terms of use:
Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen
Zwecken und zum Privatgebrauch gespeichert und kopiert werden.
Documents in EconStor may be saved and copied for your
personal and scholarly purposes.
Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle
Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich
machen, vertreiben oder anderweitig nutzen.
You are not to copy documents for public or commercial
purposes, to exhibit the documents publicly, to make them
publicly available on the internet, or to distribute or otherwise
use the documents in public.
Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen
(insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten,
gelten abweichend von diesen Nutzungsbedingungen die in der dort
genannten Lizenz gewährten Nutzungsrechte.
www.econstor.eu
If the documents have been made available under an Open
Content Licence (especially Creative Commons Licences), you
may exercise further usage rights as specified in the indicated
licence.
WIDER Working Paper No. 2013/023
Measuring government performance in public opinion
surveys in Africa
Towards experiments?
Michael Bratton*
March 2013
Abstract
In examining the study of government performance, this paper asks whether field
experiments can improve the explanatory precision of results generated by public opinion
surveys. Survey research on basic health and education services sub-Saharan Africa shows
that the perceived ‘user friendliness’ (or ease of use) of services drives popular evaluations of
government performance. For the reliable attribution of causality, however, surveys and field
experiments, combined in a variety of mixed research designs, are more rigorous and reliable
than either method alone. The paper proposes a menu of such designs.
Keywords: government performance, surveys, experiments, health, education, Africa
JEL classification: H4, I00, I18, I25
Copyright © UNU-WIDER 2013
*Michigan State University, USA, email: [email protected]
This study has been prepared within the UNU-WIDER project ‘ReCom—Research and
Communication on Foreign Aid’, directed by Tony Addison and Finn Tarp.
UNU-WIDER gratefully acknowledges specific programme contributions from the
governments of Denmark (Ministry of Foreign Affairs, Danida) and Sweden (Swedish
International Development Cooperation Agency—Sida) for ReCom. UNU-WIDER also
gratefully acknowledges core financial support to its work programme from the governments of
Denmark, Finland, Sweden, and the United Kingdom.
ISSN 1798-7237
ISBN 978-92-9230-600-7
The World Institute for Development Economics Research (WIDER) was established by the
United Nations University (UNU) as its first research and training centre and started work
in Helsinki, Finland in 1985. The Institute undertakes applied research and policy analysis
on structural changes affecting the developing and transitional economies, provides a forum
for the advocacy of policies leading to robust, equitable and environmentally sustainable
growth, and promotes capacity strengthening and training in the field of economic and
social policy making. Work is carried out by staff researchers and visiting scholars in
Helsinki and through networks of collaborating scholars and institutions around the world.
www.wider.unu.edu
[email protected]
UNU World Institute for Development Economics Research (UNU-WIDER)
Katajanokanlaituri 6 B, 00160 Helsinki, Finland
Typescript prepared by Janis Vehmaan-Kreula at UNU-WIDER.
The views expressed in this publication are those of the author(s). Publication does not imply endorsement by
the Institute or the United Nations University, nor by the programme/project sponsors, of any of the views
expressed.
1
Introduction
Citizens have a stake in the quality of government. As consumers of public goods and
services, the general public is uniquely situated to judge how governments are handling
various policy tasks—from the delivery of basic services to the creation of jobs—that directly
affect popular welfare. Moreover, if people live in a democracy—the preferred form of
political regime around the world—they enjoy rights to express opinions about government
performance.
This paper contends that public opinion surveys based on nationally representative
probability samples are a suitable method for describing and analysing public reactions to
government performance. Well-designed surveys can capture patterns in popular sentiment
and behaviour across different geographic regions or social groups. Comparative surveys can
track trends in public opinion over time, juxtapose satisfaction with services across policy
sectors, and contrast one government’s performance with another. If the right questions are
asked, survey research can even reveal the foundations—cognitive, affective, or evaluative—
on which people form opinions about government performance.
But there are drawbacks to any kind of observational research. A typical problem is
endogeneity, which refers to the circularity or reciprocity embedded in social relationships.
Does A cause B? Or, vice-versa, does B cause A? For example, do people think governments
perform well because they enjoy easy access to desired public services? Or do they judge
service accessibility in terms of their general impression of how well governments are
performing? It is insufficient to simply observe that a relationship exists; we also need to
unravel which comes first, preferably by specifying some kind of causal mechanism. In a
related problem, public policy practitioners face the issue of attribution. They want to know
whether a given policy, programme or project leads to intended social outcomes. The
challenge is to distinguish the consequences of a policy intervention from a welter of other
contributory factors. Where a planned policy intervention is only one variable in a
kaleidoscopic social setting, policy impact is difficult to discern.
The twin features of multi-causality and endogeneity are endemic to social reality and thus
intrinsic to social research (Franzese 2007). Public opinion has multiple roots, being based
simultaneously on a person’s social background, exposure to government, and deeply held
identities, interests and ideas. In short, any explanation of how people judge government
performance is bound to be multi-causal. Moreover, to avoid cognitive dissonance, people
endeavor to think and act consistently, thus raising the prospect of endogeneity. Untangling
the Gordian knot of social causation is a perennial challenge for researchers.
According to scientific norms, the laboratory experiment is the best method—a ‘gold
standard’ if you will—for addressing issues of causality (Druckman et al. 2006).
Experimental research designs allow researchers to maximize control over objects of interest.
But many of the most compelling questions in social science do not lend themselves to
inquiry under the artificial conditions of a laboratory. Thus researchers endeavour to extend
the logic of experimental research to the collection and analysis of field data. Randomized
controlled trials (RCTs), designed ex ante and rigorously conducted, seek to simulate
laboratory controls. Even where RCTs are not feasible, quasi-experimental research designs,
for example conducted ex post on extant data, can provide a measure of insight into causality.
1
Experiments and quasi-experiments have long informed research in development economics
(e.g., Banerjee and Duflo 2009; Barrett and Carter 2010; Duflo 2006) and recently in
comparative politics (e.g., Fearon et al. 2009; Humphreys et al. 2006; Humphreys and
Weinstein 2009; Moehler 2010). These studies often employ social surveys as measurement
tools to generate baseline indicators or response data from experimental groups. Currently,
researchers are beginning to apply the logic of experiments to the structure of surveys
through panel designs (in which the same respondents are interviewed over time in order to
trace and attribute change), via vignettes embedded in questionnaires (that expose survey
respondents to manipulated stimuli in order to assess differential effects) or behavioural
games run parallel to surveys (to test theories about the effects of economic, social or policy
incentives).
This paper examines the study of government performance through public opinion surveys in
Africa and asks whether the logic of scientific experimentation can improve explanatory
precision. The central argument is that surveys and field experiments, combined in a variety
of mixed research designs, are more rigorous and reliable than either method alone.
The first section examines the concept of government performance and introduces points of
convergence between survey and experimental methods. Data from the Afrobarometer are
then employed to make operational measurements of government performance in two
sectors—primary education and basic health care—and to explain performance outcomes as
seen by citizens. Addressing multi-causality, I argue that the perceived accessibility (or ‘userfriendliness’) of social services drives popular satisfaction with government performance.
Recognizing that a degree of endogeneity remains in this result, however, the paper discusses
a menu of econometric and experimental methods to address it. A conclusion makes the case
that, far from being rival methods, survey and experimental research designs are
complementary.
2
Framework
2.1 Government performance
The object of interest in this paper is government performance, that is, the capacity of
governing authorities to provide public goods and services. Public goods and services are
benefits intended for shared consumption from which nobody can be legally or effectively
excluded. These benefits may be general political guarantees (such as national defense, law
and order, or civil rights) or particular socio-economic commodities (such as transport
infrastructure, electricity supply, or health and education services). Either way, government
performance concerns the track record of state agencies at addressing collective social needs.
In studying government performance, researchers refer to best practices in the delivery of
public goods and services, which are usually measured in terms of the outputs, outcomes or
impacts of public programmes, including citizen assessments thereof. At the heart of the
enterprise is performance management: the planning, programming, implementation,
measurement, monitoring, evaluation, and reporting of programme results. In a highperforming government, each agency defines problems and goals, identifies programmes to
address goals, names leaders, and outlines strategies for implementation, along with key
2
performance indicators. Ideally, agency heads are held accountable for ensuring the
performance of the institutions and policies for which they are responsible (GPS 2012).
As researchers, we need to be clear about the specific points of reference for government
performance. Are we measuring attributes of government institutions, government policies,
or government personnel? A well-designed research instrument would explicitly distinguish
among these alternative objects. Using classic principles of public administration, we may
differentiate institutions by level of government—central, sub-national, or local—or by
function—such as specialized technical or service activities (Fesler 1949). As far as public
policies are concerned, we may distinguish extractive, regulatory or distributive policies or
make precise reference to specific policy sectors such as health or education. When it comes
to government personnel, we want to know how citizens evaluate the performance of various
officials, especially front-line agents who interact directly with the general public.
2.2 Surveys versus experiments?
The question arises as to the most appropriate method for assessing government performance,
particularly as experienced by citizens. Whatever method is chosen—whether a broad
national survey of mass opinion or a field experiment focused on a local policy
intervention—researchers immediately face the issue of case selection. Simply stated, which
units should be chosen for observation and analysis? Because it is difficult (and anyway
unnecessary) to measure the characteristics of all relevant units, most evaluation research is
designed around the logic of sampling. In surveys and experiments alike, researchers face the
challenge of drawing samples that represent relevant populations of eligible citizens. For both
surveys and experiments, randomization is the preferred selection mechanism, meaning that
every appropriate person has an equal probability of being selected into the sample.
Sampling strategies differ mainly in terms of the scope of the population under consideration.
For national surveys of government performance, researchers wish to refer to the citizenry as
a whole, often meaning adult individuals of voting age. The latest official census, updated
where necessary with projections for population growth and movement, affords a suitable
sampling frame. Although a national cross-section may be constructed in several stages to
take into account various demographic, administrative or cultural strata in the population, the
final product is one sample representing the country as a whole. For experiments on
government performance, researchers require more fine-grained information about citizen
exposure to particular government programmes. As such, they assign respondents to
treatment and control groups, each on a random basis. The procedure produces at least two
distinct but comparable samples, or more if the researcher wishes to measure different levels
of treatment. Because survey and experimental methods share the underlying logic of
randomization, however, there is no insuperable barrier to mixing these methods in hybrid
research designs.
3
Measuring government performance
This paper uses data from the Afrobarometer, an independent, large-scale, comparative public
opinion research project (Afrobarometer 2013). To be valid and reliable, indicators of
government performance must refer to specific objects of measurement and be stated in terms
that ordinary people can comprehend.
3
3.1 Institutional performance
The structure of modern government resembles a vertical pyramid composed of central,
regional and local tiers. And within each of these tiers, distinctive institutions perform
various executive, legislative, judicial, service delivery or other functions. Recognizing this
diversity, but being unable to exhaustively cover every kind of government structure and
function, the designers of the Afrobarometer make strategic choices.
The first addresses central government. The survey instrument measures the public’s trust in
a list of specific apex institutions such as the presidency, the parliament, the electoral
commission, the tax department, the police, the army and the national courts of law. The
second tier is local government. In recent Afrobarometer surveys, respondents are asked to
evaluate how well or badly their municipal or district council is performing quotidian
functions such as the maintenance of roads, the regulation of land use, and sanitation (Bratton
2012).
3.2 Personnel performance
Popular views about the performance of government derive not only from individual
experiences with institutions; they also depend on daily face-to-face interactions. In practice,
the way that agents of the state routinely relate to citizens may decide whether the
government gets high or low marks for performance. Are the public’s experiences with
officials marked by respect and responsiveness? Or are they marred by rudeness and neglect?
Critical too is whether rent seeking and corruption creep into official behaviour.
To tap these aspects of government performance, the Afrobarometer deploys batteries of
questions on the perceived job performance and honesty of particular government officials,
including police officers, tax collectors, customs agents and elected local government
councillors. Beyond general perceptions of official corruption (which can be inflated by
gossip and rumor) other items ask about individual experience with demands for bribes.
3.3 Policy performance
To capture popular perceptions of policy performance, the Afrobarometer applies a root
question—‘how well or badly is the government handling the following matters?’—to
policies in more than a dozen sectors. These range from macro-economic policies—such as
job creation and inflation control—to micro-social policies—like delivering household water
and controlling crime.
As will be shown, the Afrobarometer has devoted more effort to measuring policy
performance than any other aspect of government performance. The survey has delved in
detail into two fundamental government services: primary education, and basic health care.
The choice is appropriate because many people encounter these services as they try to
improve family welfare. Health and education rank as the top social priorities among the
most important problems that the government should address (Bratton 2009). And both
sectors have received prior attention from researchers using experimental methods (e.g.,
Banerjee et al. 2010a; Glewwe et al. 2000).
4
4
Government performance in Africa
This paper employs public opinion data for six African countries in 2012—Benin, Cape
Verde, Kenya, Mauritius, South Africa, and Uganda—to describe profiles of government
performance. The starting point is popular satisfaction with the delivery of public goods and
services across a variety of policy sectors. Figure 1 suggests that Africans distinguish among
these domains and arrive at separate judgments about each. In the social sectors, people
consider that governments are performing quite well. About two-thirds approve of
government performance in the delivery of education and health services (72 and 65 per cent
respectively). Perceived government performance on every social service, including domestic
water supply and crime control, exceeds the Afrobarometer mean (40 per cent) for
government performance.
Not so with economic policies; only 40 per cent thought that the governments in their
countries were doing well at managing the national economy. Similarly, more people scored
governments as doing ‘badly’ rather than ‘well’ on a range of activities from ensuring food
security to controlling corruption. Moreover, performance at all economic tasks was
evaluated as falling at or below the Afrobarometer mean for government performance. At the
extreme, only about one quarter of respondents gave a positive rating to performance at job
creation, inflation control and closing the gap between rich and poor.
Figure 1: Popular Satisfaction with Government Performance,
6 African Countries, 2012
100
90
72
80
65
70
60
49
46
50
40
40
40
35
33
26
30
21
16
20
10
0
Ed
uc
a
H
tio
n
ea
lth
W
at
er
C
rim
e
Ec
on
om
y
Fo
od
Co
rr
up
tio
Jo
bs
n
In
co
m
Pr
ee
ic
es
qu
al
ity
A
B
M
ea
n
Perceived positive performance on social policies
Perceived positive performance on economic policies
Afrobarometer mean
Question: “How well or badly would you say the current government is handling the following matters?”
Percentage of respondents who think government is performing well or very well in this service sector.
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
A factor analysis on all responses about satisfaction with government performance generates
two components that validate the sharp distinction in perceived performance between social
and economic sectors. Reliability analysis also confirms, however, that an overarching
concept bridges this distinction, allowing an overall index of ‘government performance’. The
country mean scores for this index along a four-point scale from ‘very badly’ to ‘very well’
have face validity. On average, citizens rate government performance relatively well in
Mauritius, a more developed country than Uganda. Importantly, though, since scores for all
5
countries fall below the mean on the scale, the Africans we interviewed are apparently more
dissatisfied than satisfied with overall government performance.
The analysis below employs three indicators of government performance: ‘satisfaction with
health services’, ‘satisfaction with education services’, and ‘satisfaction with basic social
services’, which is an average construct of both (health and education) indicators. The
construct is permissible because the people who are satisfied with health services tend to also
be satisfied with education services, and vice versa. But what are the main determinants of
satisfaction with these aspects of government policy performance?
4.1 Accessibility of services
Infrastructure
One possible source of public satisfaction is the physical proximity of service infrastructure
in the locations where people live. After all, an individual’s prospect of gaining access to
services starts from the convenient availability of a nearby service outlet (World Bank 2004).
The Afrobarometer measures the proximity of service infrastructure by the observed presence
or absence of primary schools and health clinics in every primary unit in the survey sample.
In the six countries studied, the density of service infrastructure is everywhere greater for
schools than clinics. On average, more than four-fifths of adults live in areas with access to a
local primary school compared to less than three-fifths who possess ready access to a local
health clinic.
User-friendliness
Quite apart from physical proximity, the accessibility of services depends upon the
organizational feature of ‘user-friendliness’. From a user’s perspective, services may be
simple, transparent and inclusive or they may be formal, complex and exclusionary. For poor
or illiterate people, the approachability of a bureaucratic state may be a prime consideration.
The relevant survey questions are direct: ‘In your experience, how easy or difficult is it to
obtain the following services: A place in a primary school for a child? How about medical
treatment at a nearby clinic? Or do you never try to get these services from the government?’.
An initial examination of the data shows that people find it easier to get a child into school
than to get medical attention (Figure 2). Whereas, across six African countries in 2012, 72 per
cent reported that it is easy to gain access to a basic educational service, some 63 per cent
said the same about medical care.
6
Figure 2: Ease of Access to Education and Health Services:
Popular Estimates, 6 African Countries, 2012
100
90
80
70
60
48
49
50
40
24
29
23
30
20
5
15
8
10
0
Very Difficult
Difficult
Easy
Education
Very Easy
Health
“In your experience, how easy or difficult is it to obtain the following services:
A place in a primary school for a child? Medical treatment at a nearby clinic?
Or do you never try to get these services from government?”
Figures exclude those who never tried to use these services.
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
4.2 Experiences with education services
Which obstacles—of service availability, quality, and cost—arise most frequently? For
education, the survey asked, ‘Have you encountered any of these problems with your local
public schools during the past 12 months?’. A list of six problems was then read out, ranging
from ‘overcrowded classrooms’ to ‘unable to pay’. Respondents were also asked separately if
they had ever, in the previous year, had to ‘pay a bribe, give a gift, or do a favour’ to a
government official in order to ‘get a place in a primary school for a child’.
Figure 3 compares the reported frequency of problems arising with education services. In this
case, only those persons are counted who have had contact with primary schools during the
previous 12 months.
Because popular demand for education exceeds the supply of school facilities, overcrowded
classrooms are the most common problem, reported by 62 per cent of users. It is cited most
often in countries like Kenya and Uganda with universal free primary education (UPE). The
related problem of teacher absenteeism arises almost as frequently (56 per cent). About half
of all African users register objections to a trio of core issues: shortages of textbooks and
other teaching materials, substandard school buildings and facilities, and the low quality of
instruction.
7
Figure 3: Experience with Education Services,
Specific Problems Encountered, 6African Countries, 2012
100
90
80
70
60
50
40
30
20
10
0
62
56
53
53
50
43
12
Ov
e
rc
r
ow
Ab
se
nt
de
d
cl
as
sr
oo
m
Sh
te
ac
he
rs
or
t
Su
ag
e
of
te
xt
bo
bs
ta
n
Po
or
da
ok
s
rd
fa
c
ili
tie
Fe
te
ac
h
in
es
Pa
to
oe
g
xp
en
s
y
br
ib
e
iv
e
s
s
Percentage encountering this problem*
“Have you encountered any of these problems with your local public schools during the past 12 months?”
* Excludes those who say they don’t know or have no recent experience with a public school.
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
Is public schooling too expensive? Are users unable to pay? In the litany of client problems,
the costs of primary schooling assume low priority. Fewer than half of all respondents say
that the expense of required fees inhibits them from sending children to school. In this
instance, the provision of UPE unexpectedly fails to reduce popular concern about the costs
of schooling: indeed, in 2012, Kenyans and Ugandans are more likely than other Africans
surveyed to complain about the high levels of school fees. Even in these countries, parents
still face a bevy of unofficial charges and expenses.
Finally, few users (12 per cent) say they confront demands for illegal payments from teachers
or school administrators, for example in return for school placement. Such corruption
reportedly hardly ever happens in Mauritius or Cape Verde (so say under 2 per cent), but it is
said to be common in Kenya and Uganda (over 20 per cent). Perhaps because teachers and
administrators feel overstretched by the influx of waves of new pupils under UPE, they feel
justified in seeking illicit rents.
For the record, it is worth noting that specific problems of access to education form a
composite whole. In other words, people who perceive one obstacle in gaining access to
education are likely to see other obstacles too. An average index of this factor—labelled as
‘service experiences (education)’—is expected to predict popular satisfaction with education
services. In order to tease out the relative influence of specific problems encountered,
however, each problem will first be entered separately in the explanatory models that follow.
4.3 Experiences with health care services
A parallel set of questions was asked about health care: ‘Have you encountered any of these
problems with your local public clinic or hospital during the past 12 months?’. A list of six
problems was offered, ranging from ‘long waiting times’ to ‘lack of attention or respect from
8
staff’. Respondents were also asked separately if they had ever had to ‘pay a bribe, give a
gift, or do a favour’ to a government official in order to get treatment.
Figure 4 breaks down the recent experiences of persons who attempted to use clinics and
hospitals. On average, slightly more users report a specific problem with health services (51
per cent) than with education services (47 per cent).
Figure 4: Experience with Health Services,
Specific Problems Encountered, 6 African Countries, 2012
100
90
80
77
63
58
70
53
48
60
45
50
40
30
20
16
10
0
Lo
ng
wa
i ti
ng
Di
Sh
Fe
Ab
La
rty
es
ck
or
se
ta
nt
to
fa
of
ge
o
c il
d
r
ex
oc
es
so
it i
pe
pe
tor
fm
es
ns
ct
tim
s
ed
iv
es
e
ic i
ne
s
Pa
yb
r ib
e
Percentage encountering this problem*
“Have you encountered any of these problems with your local public clinic or hospital during the past 12 mo nths?”
* Excludes those who say they don’t know or have no recent experience with a public medical facility.
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
This discrepancy is most evident in relation to overcrowded facilities, where three quarters
(77 per cent) of clinic users complain about ‘long waiting times’. By a clear margin, delays in
delivery at the point of service are the biggest problem. On any given day, urban hospitals are
typically unable to accommodate all patients; long lines of applicants regularly assemble
outside rural clinics; and, too often, some people are turned away at the end of the day
without consultation or treatment. Relative to effective client demand, which is higher for
healthcare than schooling, health services are in even shorter supply than education services.
Nearly two thirds of healthcare clients also report shortages of medicines and medical
supplies. Regardless of whether a sound health infrastructure has been built, local clinics may
lack basic commodities needed for routine preventive care. Huge cross-country disparities are
evident. Whereas over 80 per cent or more of Kenyans and Ugandans express concern about
the under-provisioning of health facilities, the same proportion in Mauritius does not.
About half of all users of health services also find fault with other common problems: staff
who behave disrespectfully toward patients (58 per cent), doctors who fail to arrive at work
(53 per cent), the high cost of consultations or medicines (48 per cent), or dirty or unsanitary
facilities (43 per cent). The data show that these sorts of complaints are least common in the
relatively richer countries of Mauritius and South Africa and more prevalent in relatively
poorer countries like Benin and Uganda.
9
Across all countries, however, health service problems are significantly more common in
rural than urban areas. Waiting times are longer at rural clinics in part because of the sparser
coverage of health infrastructure in remote areas; medicines are less readily available due
partly to long supply lines from the capital city; and doctors are more often absent, in part
because professionals are reluctant to serve at distant outposts. In the only exception to this
general tendency, clients claim that medical staff—nurses, technicians, and clerks—are more
likely to treat them rudely at urban hospitals and clinics.
Finally, few Africans (16 per cent) say they encounter demands from health care workers for
illegal payments (‘bribes’). Demands for bribes tend to be more common at rural clinics,
perhaps because Ministry officials find it difficult to supervise field staff in outlying areas.
As in the education sector, specific experiences of health access cohere into a single reliable
factor. This convergence enables the construction of an average index called ‘service
experiences (health)’. Moreover, there is a marked coincidence between ‘service experiences
(education)’ and ‘service experiences (health)’. On these grounds, a meta-index of ‘service
experiences (combined)’ is employed in analysis below.
4.4 Corruption
Popular encounters with official corruption corrode service satisfaction (Barr et al. 2009;
Olken 2007). As a key component of bad governance, the multifaceted concept of corruption
is worth measuring from various angles. One would expect that, if ordinary people perceive
official corruption to be prevalent, negative effects on service satisfaction would follow. The
Afrobarometer asks, ‘how many government officials do you think are involved in
corruption?’. This question taps the popular reputation of service providers independently of
whether an individual respondent has ever actually paid a bribe. In absolute terms, public
workers in Africa have yet to win reputations for honesty; on average, 39 per cent of the
Africans we interviewed perceive that ‘most’ or ‘almost all’ are corrupt.
A gulf exists between perceptions of corruption and the act of handing over an inducement.
As reported earlier, an average of only 14 per cent said they did the latter in the past year in
return for education or health services. As such, people either over-estimate the extent to
which corruption pervades their society, or they under-report their own involvement in the
socially disapproved act of paying a bribe. Or, most likely, both these biases are present in
the data. As such, the real level of illegal exchanges of public goods for private gain probably
lies somewhere between these extreme estimates (i.e. between 14 and 39 per cent).
For purposes of analysis, we will want to explore whether perceptions or experiences of
corruption have the bigger impact on service satisfaction. And we will wish to confirm that
such effects always run in the expected, negative direction.
5
Explaining service satisfaction
This section of the paper addresses these issues, among others. The goal is a comprehensive,
multivariate explanation of popular satisfaction with public services. As discussed, the object
of study is represented by several versions of the dependent variable, namely satisfaction with
education services, satisfaction with health services, and an average index of both services.
10
The proposed explanatory factors are the independent variables surveyed above under the
rubrics of social structure, service accessibility, service experiences, and corruption.
What, then, are the sources of public satisfaction (and dissatisfaction) with service delivery in
a sample of six African countries? The OLS regression models in Table 1, which control for
country fixed effects (not shown), tell a similar story regardless of the way that satisfaction is
measured. The strongest and most statistically significant relationships are highlighted in bold
and their explanatory power is ranked in parentheses.
The most important consideration—always ranked first—is service accessibility. What
matters most is whether clients consider services to be ‘easy to use’. This feature of ‘userfriendliness’ captures whether citizens regard public services as being uncomplicated to
operate. It addresses the extent to which clients wish to avoid entanglements in administrative
red tape and formal interactions with officials whom they feel to be socially distant. If
ordinary people can get a child into school or receive primary health care with a minimum of
such hassles, they are likely to be satisfied with service delivery.
Table 1: Sources of Service Satisfaction,
6 African Countries, 2012
EDUCATION
SERVICES
HEALTH
SERVICES
BOTH
SERVICES
-.073*** (4)
-.079*** (4)
Social Structure
Poverty
-.077*** (3)
Service Accessibility
Availability of service infrastructure
.064*** (4)
.050***
.063***
Perceived ease of service access
.181*** (1)
.232*** (1)
. 266*** (1)
Service Experiences
Fees too expensive
-.058**
-.078*** (3)
Shortages of supplies
-.004
-.045**
Poor quality teaching /treatment
-.133*** (2)
-.007***
Poor/dirty facilities
-.029
-.086*** (2)
Service experiences (education)
-.113*** (2)
Service experiences (health)
-.082*** (3)
Corruption
Perception of corruption
-.029*
-.022
-.023
Experience of corruption
-.006
-.051***
-.044***
Constant
2.415***
2.299***
2.060***
Cell entries are standardized regression coefficients = beta. (Explanatory rank in parentheses). Constant = B.
Country fixed effects included but not shown. Weak predictors (beta <.015) not shown. ***p =<.001, **p =<.01, *p =<.05
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
It is important to note that the subjective element of service accessibility is more critical than
the objective one. As Table 1 shows, physical infrastructure—whether there is a school or
clinic in the locality—remains less important to service satisfaction. The relative superiority
of subjective over objective criteria raises a challenge for government ministries responsible
for health and education in Africa. Success at service delivery is not simply a matter of
building more clinics and schools. Instead, it requires an organizational commitment to an
ethic of customer service by which the client comes to feel that his or her needs are being
considered and addressed.
11
As for social structure, the poverty status of users—measured in terms of shortages of basic
human needs—is the key social consideration. The connection of poverty to service
(dis)satisfaction is at least three times as strong as the average for other societal influences
like gender, age or residential location (not shown). And poverty’s negative impact is
consistent for both health and education services. The poorer a person, the less likely is he or
she to be satisfied with government performance at social delivery.
This robust result indicates that existing services embody an anti-poor bias, at least in the
opinion of the poor themselves. To the extent that poverty is more prevalent in the rural areas
of every country surveyed, a bias in accessibility is offset and obscured by the apparent
willingness of rural residents to accept lower quality services. This combination of facts
suggests that, African governments that seek to improve popular perceptions of performance,
would be well advised to design pro-poor health and education policies and to direct these
services initially to urban populations.
Service experiences—that is, the quality of users’ encounters with service providers—are
also part of a complete explanation. With regard to education services, poor quality of
teaching matters most to respondents, many of whom are parents of school age children. By
contrast, two other factors loom large in shaping popular satisfaction with healthcare
services: the expense of service fees and substandard facilities (‘dirty clinics’).
In other words, clients seem willing reluctantly to tolerate the inconvenience of excess of
demand for services above the available supply (as manifest in overcrowded classrooms and
long waiting times at clinics). They seemingly value education and health care so highly that
they resign themselves to congestion as an unavoidable cost of accessing scarce services. But
they nonetheless insist that teachers provide high quality education and that the government
equip medical staff with the necessary infrastructure and supplies to do their jobs properly.
Finally, what is the impact of corruption? Table 1 indicates that, if people enter into corrupt
transactions with health clinic officials, their satisfaction with services drops by significant
margins. In other words, actual encounters with bribery have negative effects on service
satisfaction that are independent of, and larger than, mere perceptions of official graft. It is
worth noting that this result departs from an original analysis of 18 African countries in 2005,
which showed that paying bribes to get services marginally increased service satisfaction in
the health sector (Bratton 2009). The adjudication of this discrepancy awaits the availability
of 2012 data from a larger number of countries.
6
Explaining government performance
To summarize, citizens in several African countries appraise the delivery of social services
principally in terms of ‘user-friendliness’, that is, whether these services are easy to use. This
feature has less to do with the physical proximity of schools and clinics and more to do with
customer relations, that is, whether citizens find front-line workers to be approachable and
responsive. Citizens want to know that government officials understand their needs and are
committed to public service.
But how far does ‘user-friendliness’ travel? While it may be an essential feature of social
services, does it also apply to government performance writ large? After all, the overall
performance of the government encompasses not only the delivery of health and education
12
services (which citizens rate in a mostly positive light), but also to the management of the
economy (which citizens rate negatively) (Figure 1). One might expect, therefore, that a
government with relatively strong performance in the social sectors could still find difficulty
in obtaining public satisfaction across the board, including with regard to its economic
policies.
But this is not what we find when we apply the same regression model to a broad index of
government performance (derived from all items in Figure 1). Rather, a customer-friendly set
of social services appears to create a halo of goodwill that enhances public perceptions of the
government’s overall performance. Table 2 shows that the perceived ease of access to health
and education services remains the principal determinant of citizen assessments of
government performance as a whole. To be sure, such positive assessments are offset if
citizens experience unhelpful encounters with staffs at government clinics and schools. But
this result only serves to underline the importance of the treatment handed out by front-line
public workers. In this case, perceptions (not experience) of corruption loom much larger in
the public’s assessment of overall government performance, perhaps because respondents are
suspicious of elite behaviour in policy arenas, like economic management, that are
technically complex and opaque to public scrutiny.
Table 2: Sources of Perceived Government Performance,
6 African Countries, 2012
GOVERNMENT
PERFORMANCE
Social Structure
Poverty
-.095***
Service Accessibility
Availability of service infrastructure
.037*
Perceived ease of service access
.172*** (1)
Service Experiences
Service experiences (education)
-.119*** (3)
Service experiences (health)
-.107*** (4)
Corruption
Perception of corruption
-.139*** (2)
Experience of corruption
-.035**
Constant
2.013***
Cell entries are standardized regression coefficients = beta. (Explanatory rank in parentheses). Constant = B.
Country fixed effects included but not shown. Weak predictors (beta <.015) not shown. ***p =<.001, **p =<.01, *p =<.05
Source: Afrobarometer, Round 5, 2012, www.afrobarometer.org
7
Towards experiments?
So far, the empirical results in this paper have been based on observational data. Issues of
multi-causality were addressed by modelling government performance with multivariate
regression under ceteris paribus conditions. In so doing, we learned that public opinion about
government performance depends more on the quality of service provision than on the
physical quantity of service outlets.
13
But what about causal direction? Given that the ‘user-friendliness’ of services is the best
explanation of popular satisfaction, we cannot dismiss a potential loop of endogeneity. Might
the causal arrow run in a reverse direction? Citizens who are broadly satisfied with the
performance of the incumbent government might be led to conclude that such services are
easy to use. Indeed, it could be argued that ‘ease of use’ is simply a proxy for (that is, an
indirect way of measuring) policy or institutional performance. This concern is heightened by
an inherent weakness of survey data: both factors are perceptions offered by the same
informant, and thus likely to be mutually constructed.
So how can we distinguish the chicken from the egg? How can we be certain that we have
attributed performance outcomes correctly to the quality of face-to-face interactions between
service providers and their clients? This section of the paper distinguishes econometric or
statistical approaches to addressing the causal problem from experimental approaches.
A simple test for causality—sometimes called ‘poor man’s exogeneity’—is the passage of
time. Actions that occur in the past can be assumed to be independent of events that occur
later. As a starting point, it is reasonable to assume that people form evaluations of
government performance after an encounter with a service provider. Thus experience of userfriendliness is causally prior to service satisfaction. But social preferences can confound this
seeming causal connection; citizens bring expectations to their interactions with
governments. Supporters of opposition political parties, for example, may be predisposed to
judge performance harshly regardless of ease of service access. Conversely, loyalists of an
incumbent government may be overly generous in overlooking deficits in the responsiveness
of state agents. And other contaminating orientations may remain even if partisan
identification is controlled for.
To tease out reciprocal feedback loops, researchers may use simultaneous equation models
(Bollen 1989; Greene 2002; Barretto and Howland 2005). This statistical technique employs
regression analysis in two stages as a means of detecting causation within observational data.
Unlike a single-equation model in which a dependent variable (y) is a function of independent
variables (x), the y variable is used simultaneously as a predictor in its own right. To control
for endogeneity (x causes y causes x), simultaneous equations include instrumental variables
(IVs) that are chosen because they are uncorrelated with (exogenous to) the error term in the
explanation. If the IV turns out to correlate with x when it is used as a dependent variable,
however, then we can conclude that we are in the presence of reverse (or bidirectional)
causation.
Instrumental variables, however, are difficult to identify, are often conceptually vague, and
are easily misapplied. Moreover, non-specialists find simultaneous equations hard to execute
and interpret. Thus the technique is rarely used outside econometrics, for example in analysis
aimed at policy practitioners. To employ it here, we would have to find an IV that predicts
both government performance and the user-friendliness of services, but which is independent
of both. One is hard pressed to conceive of such a factor, let alone to find an operational
indicator for all the ten thousand cases of survey respondents in the present study. The only
conceivable source is another item in the same survey, but we already know that survey
respondents adjust their attitudes by avoiding cognitive dissonance, defaulting to political
partisanship, or relying on informational shortcuts, thus inviting endogeneity.
This is not to say that two-stage least squares models with instrumental variables can never
be used with survey data. A separate Afrobarometer study asked which comes first: state
14
building or democracy? Having established that citizens associate the rule of law with an
effective state, the authors demonstrate that this public good is produced not just by effective
states but also by democratic regimes (Bratton and Chang 2006). As such, while democracy
is most easily built on the foundations of an established state, state building is also a product
of democratization.
Given the complexity of applying ex-post econometric fixes to survey observations, however,
researchers tend to prefer ex-ante research designs based on experimental logic as a possible
way forward. Random assignment of treatments in controlled field trails would seem to
improve the odds of reliably attributing causality. Is it therefore possible to design field
experiments on government performance at social service delivery? A limited literature has
begun to find positive effects of participatory monitoring of service providers by citizens in
the developing world, including in sub-Saharan Africa (SSA) (Paul 2000; Bjorkmann and
Svennson 2009; Olken 2010). An experiment would be most pertinent to the concerns of this
paper if the focus were on education or health services at the grassroots level and if the
treatment of interest were the user-friendliness of service delivery. Such an experiment would
contrast an accessible delivery mechanism with one that citizens found difficult to use; the
experiment would manipulate the way that staff treats clients.
The relevant services are a primary school place for a child and medical treatment for oneself
or one’s family. An experiment in the educational sector could therefore examine the effects
on popular satisfaction of a generous school admissions policy. The treatment would throw
open the school doors to any child who wished to attend by doing away with the need for
documentary proof of the child’s identity, age and place of residence or requirements that
parents purchase school uniforms or school supplies. The control group would be admissions
procedures as currently constituted. An experiment in the health sector could follow similar
lines. Clinic staff would be required to admit all-comers without any requirement for
verification, registration or payment. All clients, not only the indigent, would receive free
consultation and treatment without any fee or co-payment. Again, the control group would be
exposed to extant procedures.
At issue would be the impact of these treatments on popular satisfaction with service
delivery. The main hypothesis would be that, relative to existing procedures, open and
accessible treatments would induce citizens to express greater satisfaction. Providing the
sample of participants were large enough, the experiment could be refined to include different
packages of treatments as a way of discerning, for example, whether identity documents, fee
payments, corruption or other hurdles to service accessibility were blocking satisfaction.
The Afrobarometer data suggest that client experiences are distinctive in each service sector.
In the African countries we studied, citizens drew most attention to poor quality teaching in
schools as opposed sub-standard (dirty) facilities in the health clinics. Experiments could be
designed to randomly vary these treatments in schools and clinics. Perhaps policy initiatives
could improve teacher quality in randomly selected primary schools, whether through
retraining or improved salaries, or upgrade the medical supply chain or building maintenance
schedules for randomly selected health clinics. The treatment-control logic would then be
applied to a comparative analysis of service satisfaction among clients of service outlets ‘with
and without’ policy improvements.
An experimental study on the use of insecticide-treated bed nets to combat malaria in Kenya
confirmed the importance of the ‘user friendliness’ of health services (Cohen and Dupas
15
2008). It questioned whether the imposition of a nominal charge for a service ensured that
clients value the good provided, thereby avoiding wastage. The study showed unexpectedly
that pregnant women who received free bed nets were no less likely to use them than others
who paid a subsidized price. Because cost sharing dampened demand, the study concluded
that free distribution (the more accessible option) would save lives, especially among young
children.
Under field conditions, mundane and moral considerations often disrupt the abstract purity of
experimental research designs. Government officials, even if open to evidence-based policy
research (a big ‘if’), commonly resist treating target communities differentially because they
wish to avoid charges of political favouritism. Moreover, the ethics of human development
demand that no group be deprived of opportunity to improve wellbeing simply for research
purposes. For this reason, opportunities for access to services would have to be added to
treatment groups rather than subtracted from control groups. But these sorts of honourable
considerations constrain the ability of researchers to manipulate the full range of variance in
the objects of their interest.
Most importantly, the results of field experiments cannot be assessed without data. The
question arises therefore as to where such facts and figures come from. Experimental
researchers always have to rely on information from research subjects including, very
commonly, data elicited through public opinion and other social surveys. The final section of
this paper discusses a menu of options for taking advantage of complementarities in
experimental and survey research methods.
8
A menu of mixed methods
We first consider three integral permutations in which surveys and experiments are combined
(ex-ante) into a single research design: (a) surveys around experiments; (b) surveys within
experiments, and; (c) experiments within surveys. We then turn to two permutations in which
separate surveys or experiments are deployed (ex-post) in sequential fashion; (d) experiments
beyond surveys and (e) surveys beyond experiments. The pros and cons of each admixture are
discussed in the context of interventions to improve the delivery of basic public services in
sub-Saharan Africa.
8.1 Surveys around experiments
A quasi-experimental ‘before-and-after’ research design is a straightforward way to combine
surveys and experiments. This design brackets a project intervention between two surveys
administered at different times. As a first step, a ‘before’ survey provides baseline data about
the beliefs, attitudes and behaviours of a randomly selected sample of prospective clients.
The ‘before’ survey precedes the project intervention, which then takes place. At some
subsequent time, when project effects and impacts are expected to have appeared, an ‘after’
survey is administered to a random sample of project participants in order to take a second
measurement of the variables of interest. Researchers then compare ‘before’ and ‘after’
results and attribute any differences to the project.
The key assumption of before-and-after experimental designs is that changes in outcomes
from time t to time t+1 are likely to have a causal connection. But challenges remain. First,
attribution: is the project the only variable factor in the period between surveys? Or can
16
outcomes be ascribed to other, uncontrolled contextual changes? Second, how can
researchers contain research costs and obtain timely results when the design requires
mounting field teams on at least two separate occasions?
Third, even if baseline and participant samples are randomly drawn, do they amount to wellmatched cross-sections? After all, certain participants—often persons who are more
proximate, educated or well connected—may select themselves into a project. If these
individuals bring distinctive values, attitudes and behaviours into the project, they may cause
selection bias. One possible solution is to employ a panel design—meaning that the same
people are interviewed before and after the project. But a panel design has its own
shortcomings, the most common of which is the inevitable attrition of respondents over time,
sometimes to the point that the second or subsequent samples become too small to support
reliable generalizations.
As an omnibus instrument, the Afrobarometer lends itself to use as a source of baseline data
when surveys are required around experiments. It offers cost savings and convenience to
policy analysts who are unable to mount customized surveys of field conditions prior to
planned interventions. Due to limitations of sample size, however, these data can only be
disaggregated reliably to the region, provincial or state level, and cannot be used to track
projects at the district level or below (subject to exceptions described in the next section).
But, because this survey employs a national probability sample, it can be used for monitoring
and evaluating planned government interventions on a national scale.
For example, a donor agency commissioned an evaluation of the impact of civic education
programmes in Zambia based on pilot Afrobarometer surveys. The first survey followed the
emerging Afrobarometer standard of a randomized national probability sample; the second
covered a smaller snowball sample of programme participants identified by project leaders.
Initial results suggested that civic education had positive impacts on political knowledge and
occasionally changed popular values like political tolerance. These results survived statistical
controls for citizens’ education levels and exposure to mass media. The problem of selfselection was addressed by imposing statistical controls for political activism, measured in
terms of these individuals’ political efficacy and organizational skills. As a result, the study
concluded that persons exposed to civic education remained slightly more likely than the
average Zambian to register to vote (Bratton et al. 1999; see also Finkel 2002, Finkel et al.
2012).
8.2 Surveys within experiments
An alternative approach embeds surveys within experiments as instruments to measure the
differential effects of treatments and controls. This ‘with-and-without’ design does not
require a baseline observation from an earlier period. Instead, researchers randomly assign
participants in a service project to Treatment Groups (T1…Tk, depending on levels of service)
and a Control Group (C). When project outcomes have become visible, researchers
administer one common survey questionnaire to all participants. They expect respondents in
Groups T1…Tk (the groups with various project treatments) to show greater and more
positive project effects than respondents in Group C (the standard control group, that is,
without treatment).
17
A ‘with-and-without’ design has the advantage of roughly approximating laboratory
conditions in the field. The quality of the simulation depends on the researcher’s success at
isolating presumed causal factors (treatments) while holding other conditions constant
(controls). Much also depends on whether subjects are assigned to treatment and control
groups on a random or matched basis, and are therefore comparable.
But field settings pose inconvenient obstacles. First, it is difficult to perfectly match treatment
and control groups. One common way to limit environmental variance is to locate the study
in neighbouring communities; in this case, one community receives the public service and the
other does not. But the advantages of geographic proximity may be offset if interactions
among neighbours contaminate the purity of the treatment-control distinction. Second, the
intense scrutiny of researchers in a confined locality may have an unintended Hawthorne
Effect, whereby the very act of observing research subjects changes their attitudes or
behaviours. Third and finally, field experiments, especially if conducted on a pilot basis
within strict boundaries, suffer a similar problem to laboratory experiments. One can never be
certain that results are externally valid, that is, that they apply outside the experimental
setting, for example, when service programmes are scaled up to wider populations (Barabas
and Jerit 2010).
Omnibus public opinion surveys like the Afrobarometer survey are not designed for use
within a field experiment. To serve this purpose, modifications are required to the survey’s
sampling protocol and questionnaire instrument. A purposive over-sample must be added that
is large enough to encompass treatment and control groups, both randomly selected, within a
specific project area. In a second requirement, questionnaires must include items that explore
how respondents react to particular project goals, procedures and outcomes. For this reason,
service providers who wish to conduct a ‘with-and-without’ experiment are advised to design
their own survey sample and instrument.
Note, however, that the Afrobarometer has assisted an international development assistance
agency with this task in Uganda, Mali, and Senegal. The subject matter was local government
performance. The treatment groups were local councils that received donor-funded
institution-building inputs; control groups were councils that did not. In Uganda, the project
focused on strengthening local government councils in 16 of the country’s 80 districts, with
the remaining 62 districts serving as a control. The donor sponsored an over-sample of 100
cases in each of 16 treatment districts and added several questions about local government
performance. This collaboration offered project managers the opportunity to test for impact at
both the collective level (i.e., 16 districts combined) and, at the level of individual districts
within a wider margin of error (Logan 2011). Unfortunately, to our knowledge, the donor
agency never made full use of the data in order to derive policy relevant results about the
efficacy of their local government strengthening programmes. This shortcoming suggests
that, where sponsors lack in-house research capacity, analysis should be subcontracted to the
survey agencies that collect the data.
8.3 Experiments within surveys
Given the state of the art in public opinion studies, experiments within surveys lie at a
research frontier. This approach embeds randomized controlled experiments inside survey
instruments.
18
Researchers use vignettes—or capsule stories—as a means to expose survey respondents to
varying stimuli (‘treatments’) within an interview. A sub-set of interviewees, who serve as a
control group, receives no variations of treatment. The stimuli are hypothetical scenarios in
which the researcher systematically manipulates a variable that is expected to influence
outcomes of interest. For example, South African survey respondents, when exposed to
contrasting stories about the eviction of a squatter, were found to be more likely to see the act
as legitimate if they believed that it occurred by means of due process of law (Gibson 2008).
Other researchers varied the names of candidates for public office in order to reveal that
perceived candidate ethnicity influences voters’ expressed voting intentions (Conroy-Krutz
2012; Dunning and Harrison 2010).
On the positive side, experiments within surveys allow researchers to move beyond passive
observation by actively varying an object of interest. As such, the method offers a means to
increase leverage over causal inference. On the negative side, hypothetical vignettes are just
that: hypothetical. The researcher is left wondering whether any observed impacts are
accompanied by behavioural changes in the real world. Moreover, vignettes take up a lot of
scarce space and time in survey questionnaires. And random assignment of respondents is
logistically challenging under the far-flung field conditions of a social survey, thus inviting
error.
8.4 Experiments beyond surveys
Two final approaches allow researchers to triangulate on a given research problem, even if
multiple methods were not envisaged as part of an original research design. For example,
researchers who wish to test for theorized causal mechanisms may choose to add experiments
based on economic games as supplements to surveys (Camerer 2003).
Take ethnicity, a factor relevant to government performance in Africa. Existing studies have
established that ethnic diversity, especially if reinforced by inter-group inequality, inhibits
public goods provision (Miguel 2004; Baldwin and Huber 2010). A survey in urban
Kampala, Uganda both confirmed this negative relationship and provided a frame for
randomly selecting participants for a field experiment (Habyarimana et al. 2007). The
researchers asked participants to play economic games and provided cash incentives to join
others in community projects to provide clean drains, garbage collection, and security from
crime. Participants rewarded or punished potential collaborators (shown in photographs)
according to whether or not they saw them as co-ethnics. Analysis of these games showed
that ethnic co-operation was strategic; it did not grow out of shared cultural preferences but
rather from rational expectations of reciprocity or sanctions.
If experiments conducted after surveys can reveal why empirical relationships exist, then they
promise theoretical progress. Game theory is certainly a useful device for exploring
behavioural outcomes that respond to economic incentives. But, even if conducted in the
field, behavioural games do not fully replicate real-world contexts and thus do not entirely
vitiate the problem of external validity associated with laboratory experiments. Moreover, as
applied in the literature so far, games seem better suited to modelling client behaviour than
the performance of government service providers. Finally, while scholars may wish to
excavate the theoretical foundations of behaviour, it is far from clear that practitioners need
to do so; the latter may be content to know ‘what works’ in policy terms without necessarily
delving deeply into why it does so.
19
8.5 Surveys beyond experiments
In a final combination of methods, researchers may use the results of field experiments to
improve the construction survey instruments. The purpose here is to test the external validity
of field experiments on large populations. For example, several experimental studies from
various parts of the world suggest that public information campaigns can increase popular
satisfaction and induce participatory behaviour (Boeri and Tabellini 2010; Collier and
Vicente 2011; Gerber and Green 2000). The external validity of such results can be tested
with questionnaire items about the availability, nature and spread of public information in
national or cross-national surveys.
For example, experiments that monitor the effects of public report cards on the performance
of elected political representatives would seem especially well suited to the study of social
service delivery (Humphreys and Weinstein 2012; Banerjee et al. 2010b). There is no logical
barrier to using report cards to inform the general public about the quality of services at
schools and clinics, especially in the urban areas of Africa where respondents have a measure
of choice in the consumption of education or health services (e.g., public or private,
traditional or modern). To my knowledge, no such experiment has yet been conducted. But, if
it were, its results would be helpful in designing questionnaires items that tap respondents’
levels of knowledge about service provision, including from report cards. In this way,
experiments would strengthen survey instruments that are used to test theory and praxis about
government performance.
9
Conclusion
This paper has argued that, as tools for explaining government performance, public opinion
surveys and field experiments are potentially compatible research methods. Both methods
employ random selection of cases, either to match treatment and control groups or to make
inferences to wider populations. By combining surveys and experiments in mixed research
designs, social scientists are able to generate more dependable results than by using either
method in isolation. To take the example used here, mixed methods may help sort out
whether the ‘user friendliness’ of social services is a precursor or a product of government
performance.
At minimum, surveys are well suited to discover the attitudes and behaviour of a
government’s clientele: the citizens. This kind of raw data can provide an informative
foundation for experimental research. Moreover, the large numbers of cases in survey data
enable statistical consideration of multi-causality, that is, the diversity and relative weight of
rival explanations of government performance. Public opinions tend to be mutually
endogenous, however, which limits their utility in discerning the direction of underlying
causes. This is where field experiments come in. Experimental controls allow analysts to
manipulate a selected aspect of government performance such as the ease with which clients
are able to access social services. The inference is that any observable difference between
treatment and control groups is causally attributable to the experimental treatment.
Among various options, the most promising approach is to apply mixed survey and
experimental methods to the field performance of government agencies, especially in
developing countries where knowledge is often uncertain about ‘what works’. The researcher
may choose to embed surveys and experiments within a single, integral design or opt instead
to employ sequential combinations of these methods on an as-needed basis. Much depends on
20
the evolving state of theoretical knowledge and on practical considerations such as the size of
research budgets and the availability of existing data. But, all told, the future looks promising
for the continued use of surveys and experiments as complementary devices for the
measurement of the impacts of government performance in the field.
References
Afrobarometer (2013). ‘Survey and Methods’. www.afrobarometer.org.
Baldwin, K. and J. Huber (2010). ‘Economic versus Cultural Differences: Forms of Ethnic
Diversity and Public Goods Provision’. American Political Science Review, 104(4): 64462.
Banarjee, A. and E. Duflo (2009). ‘The Experimental Approach to Development Economics’.
Annual Review of Economics, 1(1): 151-78.
Banarjee, A., E. Duflo, and S. Khemani (2010a). ‘Pitfalls of Participatory Programs:
Evidence from a Randomized Evaluation of Education in India’. American Economic
Journal: Economic Policy, 2(1): 1-30.
Banarjee, A., S. Kumar, R. Pande, and F. Su (2010b). ‘Do Informed Voters Make Better
Choices? Experimental Evidence from Urban India’. Unpublished paper, Massachusetts
Institute of Technology, October.
Barabas, J. and J. Jerit (2010). ‘Are Survey Experiments Externally Valid?’. American
Political Science Review, 104(2): 226-43.
Barr, A., M. Lindelow, and P. Serneels (2009). ‘Corruption in Public Service Delivery: An
Experimental Analysis’. Journal of Economic Behaviour and Organization, 72(1): 22539.
Barrett, C. and M. Carter (2010). ‘The Power and Pitfalls of Experiments in Development
Economics: Some Non-Random Reflections’. Applied Economic Perspectives and
Policy, 32(4): 515-48.
Barretto, H. and F. Howland (2005). Introductory Econometrics. New York: Cambridge
University Press.
Bjorkman, M. and J. Svensson (2009). ‘Power to the People: Evidence from a Randomized
Experiment of a Citizen Report Card Project in Uganda’. Quarterly Journal of
Economics, 124: 735-70.
Boeri, T. and G. Tabellini (2010). ‘Does Information Increase Political Support for Pension
Reform?’. Unpublished paper, Bocconi University, April.
Bollen, K. (1989). Structural Equations with Latent Variables. New York: Wiley.
Bratton, M. (2009). ‘Are You Being Served? Popular Satisfaction with Health and Education
Services in Africa’. In G. Bland and C. Arnson (eds), Democratic Deficits: Addressing
Deficits of Sustainability and Consolidation Around the World. Washington, DC:
Woodrow Wilson Center.
Bratton, M. (2012). ‘Citizen Perceptions of Local Government Responsiveness in SubSaharan Africa’. World Development, 40(3): 516-27.
Bratton, M., P. Alderfer, G. Bowser, and J. Temba (1999). ‘The Effects of Civic Education
on Political Culture: Evidence from Zambia’. World Development, 27(5): 807-24.
21
Bratton, M. and E. Chang (2006). ‘State Building and Democratization: Forwards,
Backwards, or Together?’. Comparative Political Studies, 39(9): 1059-83.
Camerer, C. (2003). Behavioural Game Theory: Experiments in Strategic Interaction.
Princeton: Princeton University Press.
Cohen, J. and P. Dupas (2008). ‘Free Distribution or Cost-Sharing? Evidence from a Malaria
Prevention Experiment’. National Bureau of Economic Research Working Paper No.
14406.
Collier, P. and P. Vicente (2011). ‘Votes and Violence: Evidence from a Field Experiment in
Nigeria’. Working Paper, Centre for the Study of African Economies, University of
Oxford.
Conroy-Krutz, J. (2012). ‘Information and Ethnic Politics in Africa’. British Journal of
Political Science, First View online, September.
Druckman, J., D. Green, J. Kuklinski, and A. Lupia (2006). ‘The Growth and Development
of Experimental Research in Political Science’. American Political Science Review,
100(4): 627-36.
Duflo, E. (2006). ‘Field Experiments in Development Economics’: Paper prepared for the
World Congress of the Econometric Society, January.
Dunning, T. and L. Harrison (2010). ‘Crosscutting Cleavages and Ethnic Voting: An
Experimental Study of Cousinage in Mali’. American Political Science Review, 104(1):
21-39.
Fearon, J., M. Humphreys, and J. Weinstein (2009). ‘Can Development Aid Contribute to
Social Cohesion after Civil War? Evidence from a Field Experiment in Post-Conflict
Liberia’. American Economic Review: Papers and Proceedings, 99(2): 287-91.
Fesler, J. (1949). Area and Administration. Birmingham: University of Alabama Press.
Finkel, S. (2002). ‘Civic Education and the Mobilization of Political Participation in
Developing Democracies’. Journal of Politics, 64(4): 994-1020.
Finkel, S., J. Horowitz, and R. Rojo-Mendoza (2012). ‘Civic Education and Democratic
Backsliding in the Wake of Kenya’s Post-2007 Election Violence’. Journal of Politics,
74(1): 52-65.
Franzese, R. (2007). ‘Multicausality, Context Conditionality and Endogeneity’. In C. Boix
and S. Stokes (eds), The Oxford Handbook of Comparative Politics. Oxford: Oxford
University Press.
Gerber, A. and D. Green (2000). ‘The Effects of Canvassing Phone Calls and Direct Mail on
Voter Turnout: A Field Experiment’. American Political Science Review, 94(3): 653-63.
Gibson, J. (2008). ‘Group Identities and Theories of Justice: An Experimental Investigation
Into the Justice and Injustice of Land Squatting in South Africa’. Journal of Politics, 70:
700-16.
Glewwe, P., M. Kremer, S. Moulin, and E. Zitzewitz (2000). ‘Retrospective versus
Prospective Analyses of School Inputs: The Case of Flip Charts in Kenya’. National
Bureau of Economic Research Working Paper No. 8018.
Government
Performance
Summit
http://governmentperformance.org/index.htm.
(GPS)
Greene, W. (2002). Econometric Analysis. New York: Macmillan, fifth edition.
22
(2012).
Habyarimana, J., M. Humphreys, D. Posner, and J. Weinstein (2007). ‘Why Does Ethnic
Diversity Undermine Public Goods Provision?’. American Political Science Review,
101(4): 709-45.
Humphreys, M., W. Masters, and M. Sandbu (2006). ‘The Role of Leaders in Democratic
Deliberations’. World Politics, 58(2): 583-622.
Humphreys, M. and J. Weinstein (2009). ‘Field Experiments and the Political Economy of
Development’. Annual Review of Political Science, 12: 367-78.
Humphreys, M. and J. Weinstein (2012). ‘Policing Politicians: Citizen Empowerment and
Political Accountability in Uganda’. Columbia University, draft paper, April.
Logan C. (2011). ‘The Uses of the Afrobarometer in Policy Planning, Program Design and
Evaluation’. Afrobarometer Briefing Paper No. 98. www.afrobarometer.org.
Miguel E. (2004). ‘Tribe or Nation? Nation Building and Public Goods in Kenya versus
Tanzania’. World Politics, 56(3): 327-62.
Moehler, D. (2010). ‘Democracy, Governance and Randomized Development Assistance’.
The Annals of the American Academy of Political and Social Science, 628: 30-46.
Olken, B. (2007). ‘Monitoring Corruption: Evidence from a Field Experiment in Indonesia’.
Journal of Political Economy, 115(2): 200-49.
Olken, B. (2010). ‘Direct Democracy and Local Public Goods: Evidence from a Field
Experiment in Indonesia’. American Political Science Review, 104(2): 243-67.
Paul, S. (2000). ‘Making Voice Work: the Report Card on Bangalore’s Public Services’.
Paper No. RPG 68207M for a project on Effectiveness of Client Surveys in Improving
Service Delivery, World Bank.
World Bank (2004). World Development Report, 2004: Making Services Work for Poor
People. Washington, DC: World Bank.
23