A Meta-Analysis of Role Ambiguity and Role Conflict on IS

Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
A Meta-Analysis of Role Ambiguity and Role Conflict on IS Professional Job
Satisfaction
Yide Shen
CIS Department
Robinson College of Business
Georgia State University
Atlanta, GA, USA 30302
[email protected]
Abstract
There have been numerous IS studies on the topic of role
ambiguity (RA) and role conflict (RC). The need for
resolving the disagreement in understanding the impact of
RA and RC on job satisfaction within an IS context
motivated the current research. Employing Hunter and
Schmidt’s (1990) meta-analysis method, this study
synthesizes empirical IS study results from twelve
independent samples. The results revealed that both RA
and RC are negatively correlated with IS professional’s
job satisfaction across diverse studies. Compared with
non-IS related jobs, IS professionals are not necessarily
more- or less-susceptible to the negative impacts of RA
and RC on job satisfaction. Furthermore, the unexplained
variance in the study effect sizes indicates the existence of
potential moderating variables. Finally, an unexpected
but noteworthy contribution of the current research is to
raise the issue of misuse of the term meta-analysis within
the IS research community.
1. Introduction
With the proliferation of computers and information
systems in organizations, information technology (IT)
professionals are playing increasingly critical roles in
supporting organizational operations and attaining
strategic goals. Hence, understanding the factors that
influence IS professionals’ job satisfaction is an area
which concerns both IS managers and IS researchers. The
importance of this issue is reinforced by research showing
that job satisfaction is the most important antecedent of IS
employees’ intention to stay in an organization and thus,
the most important determinant of IS personnel turnover
[3]. Additionally, decreased job satisfaction often leads to
increased absenteeism, ill health, and grievances [40],
which impairs the quality of work life of IS professionals.
Among the variables expected to influence job
satisfaction, two role stressors – role ambiguity (RA) and
role conflict (RC) – have captured the attention of
numerous researchers. The results from both the
organizational behavior and IS literatures suggest that RA
and RC are generally negatively related to job
satisfaction, however the strength and significance level
of these relationships are not always consistent.
Especially in IS research, we lack a definitive conclusion
about the relative importance of RA and RC.
The need for resolving the disagreement in
understanding the impact of RA and RC on job
satisfaction within an IS context motivated the current
research. Given the broad range of empirical IS research
on this topic, a formal meta-analysis is useful to
accumulate results across many independent studies and
synthesize the statistical data from multiple sources. In
addition, due to the uniqueness and rapid changes
occurring in the IT environment, it is worthwhile to
empirically examine the applicability of role stressor-job
satisfaction relationships in IS settings. This is the
objective of this paper: to accumulate empirical results
regarding the relationship between RA/RC and IS
professionals’ job satisfaction across independent studies,
to examine the relative importance of the two role
stressors, to compare them with those from a broad range
of job occupations in the organizational behavior
literature, and investigate the variables that may moderate
these relationships.
The remainder of the paper is structured as follows.
Section 2 reviews the concepts of RA and RC as they
have been described in the management literature and
develops hypotheses for this study. Sections 3 and 4
describe the meta-analysis method employed in this study.
Section 5 presents the result of our analysis. The paper
concludes by discussing the results from the quantitative
analysis and their implications for IS personnel
researchers and IS scholars, more broadly.
2. Literature Review and Hypotheses
“A role is most typically defined as a set of
expectations about behavior for a position in a social
structure” [47, p. 155]. RA refers to the degree to which a
given job is lacking in terms of 1) the predictability of
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
1
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
responses to one’s behavior, and 2) the existence or
clarity of behavior requirements [47]. RC is defined as
“the degree of incongruity or incompatibility in the
expectations or requirements communicated to a focal
person” [4, p. 93]. It is generally believed that RA and RC
result in lower job satisfaction. Consistent with the
theoretical basis, most of the empirical results reported
medium to high negative correlations between either RA
or RC and job satisfaction, while only two studies have
shown positive correlations [25, 32]. In line with the
goals of this research – to accumulate empirical results
regarding the relationship between RA/RC and IS
professionals’ job satisfaction across independent studies
– we propose the first two hypotheses:
Hypothesis 1: RA and job satisfaction will be
negatively correlated among IS professionals.
Hypothesis 2: RC and job satisfaction will be
negatively correlated among IS professionals.
Another inconsistency that has been observed is the
relative impact of either RA or RC on job satisfaction. For
instance, by studying a sample of 229 IS professionals
from nine companies in various industries (e.g.,
insurance, banking, manufacturing, etc.), Baroudi [2]
found that RA is a more influential variable, accounting
for 22% of the variance in job satisfaction, compared to
RC, which accounted for only 7% of the variance in job
satisfaction. In contrast, based on their study of 109 IS
managers from a wide variety of industries (e.g., banking,
insurance, manufacturing, etc.), Li and Shani [37]
concluded that:
Role ambiguity, having the lowest mean composite score
[2.9] among the four stressors perceived by the IS
managers [work overload (5.0), role conflict (4.5), jobinduced anxiety (3.2), and role ambiguity (2.9)], is
significantly influenced by many organizational
contextual as well as job satisfaction factors. Ivancevich,
Napier, and Wetherbe (1983) had a similar finding that
role ambiguity was not a major stressor perceived by the
IS personnel; it was the fifth highest mean score among
the seven stressor scales used in their study [37, p. 123].
However, a closer review of these two studies reveals
some differences in terms of the research design. First,
Baroudi’s sample consisted of IS professionals from a
range of job types (such as applications programmers,
programmer/analysts, analysts, and project leaders), while
the subjects in Li and Shani’s sample were all IS
managers. Second, the two studies used the same measure
for RA and RC [47], but they used different measures for
job satisfaction (Job Descriptive Index [50] vs. a homegrown scale [16]). Thus, it is impossible to conclude
whether the different impacts of RA and RC on job
satisfaction are due to the nature of these two constructs,
or whether the contradictory results are introduced by
these other artifacts.
Based on above discussion, hypothesis 3 compares the
magnitude of the negative effects of RA and RC on job
satisfaction:
Hypothesis 3: The correlation between RA and job
satisfaction will differ from the correlation between
RC and job satisfaction among IS professionals.
Two meta-analysis studies from the organizational
behavior literature have identified a generally negative
relationship between both RA and RC and job satisfaction
in many different types of occupations, although the
strength of the relationships varied across different
occupational samples [10, 28]. Kahn and colleagues’ [30]
study of various jobs revealed that the level of RC and
RA were high among people whose jobs involved
innovation or boundary spanning (i.e., coping with people
outside the person’s work group).
This is exactly the nature of most types of job in the IS
work environment. For example, rapid changes in
technological innovation require IS employees to update
their knowledge and skills frequently. System analysts
need to work with cross- departmental and even crossorganizational stakeholders. These characteristics indicate
that the problem of high RC and RA especially pertains to
IS professionals:
The evidence is highly suggestive, therefore, that systems
professional with their high … need for achievement,
autonomy, and cognition, and low need for change are
particularly prone to the adverse effects of role conflict
and role ambiguity [2, p. 342].
On the other hand, it is possible that because role
stressors are inherent within IS work environments,
workers that have been selected for and socialized to
work within the IS field have a relatively high tolerance
for RA and RC [23]. King and Sethi [31] described how
the process of socialization may make IS professionals
accustomed to accepting such challenges in their jobs.
Thus, even though the levels of RA and RC may be high
for IS professionals, their high levels of tolerance for role
ambiguity and role conflict may help to moderate the
negative effects that would otherwise occur, in terms of
low job satisfaction.
Based on above discussion, hypothesis 4 compares the
negative effects of RA and RC between IS professionals
and other non-IS job types:
Hypothesis 4a: The strength of the relationship
between RA and job satisfaction will be similar among
IS professionals and other job types.
Hypothesis 4b: The strength of the relationship
between RC and job satisfaction will be similar among
IS professionals and other job types.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
2
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
3.2. Coding of Study Information
3. Meta-Analysis Method
The term meta-analysis was coined by Glass [11, p. 3]
to refer to “the analysis of analyses…the statistical
analysis of a large collection of analysis results from
individual studies for the purpose of integrating the
findings.” The main thrust of meta-analysis methods is to
accumulate empirical results across independent studies
and provide a more accurate representation of population
characteristics. Over the years, meta-analysis has become
a legitimate statistical tool to integrate empirical research
findings in many disciplines, such as medicine, education,
and psychology [19].
Hunter and Schmidt’s [18] psychometric meta-analysis
is the primary method employed in this study. The focus
of Hunter and Schmidt’s method is on effect sizes, which
represent correlational measures of effect. The advantage
of Hunter and Schmidt’s approach is that it provides
techniques to correct for artifactual and methodological
problems that may distort effect sizes from the true score
values. In survey research, correlations (Pearson’s r) are
the most frequently-used measure of effect size, and the
focus of the current research. (In contrast, controlled
laboratory studies employ other types of effect sizes to
characterize the differences between experimental and
control group effects, such as Cohen’s d.) Following the
processes suggested by Hunter and Schmidt [18], our
meta-analysis consists of three major phases: literature
search, coding of study information (both methods and
results), and data analysis. Each phase is discussed in
detail in the following sections.
3.1. Literature Search
A literature research was performed primarily on
electronic sources (ABI/Inform, EBSCO Business Source
Premier, ACM Digital Library, and ScienceDirect), as
well as several conference proceedings (ICIS, HICSS,
AMCIS, and ACM SIG CPR) using the keywords “role
ambiguity”, “role conflict”, “job satisfaction”, “stress”,
“role stress,” “burnout”, and “work exhaustion”. Since the
focus of this research is on IS professionals, the term
“information systems” and “information technology”
were used in combination with the above keywords to
restrict the search results. We also reviewed the reference
list of the articles retrieved to identify other related
papers. The search yielded 40 papers related to RA, RC,
and job satisfaction in an IS context. Methodological
details and correlational results were then coded based on
these papers.
A total of 25 papers were excluded from preliminary
screening for various reasons. Since the focus of this
study is on IS professionals, four papers examining end
users or other non-IS occupational types were excluded.
Meta-analysis requires data regarding correlations and
sample sizes in order to compute a weighted correlation
between independent and dependent variables, thus,
another eighteen papers were excluded because they
either were theoretical papers without any empirical data
(two papers) or didn’t report correlations for either of the
role stressor variables and job satisfaction (sixteen
papers). In addition, three other papers that did not study
job satisfaction were also excluded.
Following Viswesvaran and Ones’s [51] advice,
methodological weaknesses were not treated as grounds
to eliminate any studies. For example, if a paper reported
a weak result for internal reliability (Cronbach alpha) for
RA or RC, rather than excluding the study, we
incorporated the information about the (weak) reliability
metrics, and used Hunter and Schmidt’s procedures to
correct for this methodological weakness (and others).
The remaining fifteen papers were then coded for the
following information: sample size, Pearson’s correlations
between the role stressors and job satisfaction, and
reliability statistics (Cronbach’s alpha). During the coding
process, we found that five data samples were each used
in two different papers, respectively. To avoid double
counting of the same sample, only one paper for each
sample was included in our analysis.
Furthermore, two papers separated their sample into
two sub-groups. One of these studies [25] split the sample
randomly, using one sub-sample to validate their
instruments and another sub-sample to test their research
model. The two subgroups from this study were thus
treated as independent samples in our analysis. The other
study [26] divided the sample according to the mean
value of a specific personality variable (Type A versus
Type B personality behavior), and found a moderating
effect of this variable. Hunter and Schmidt [18] suggested
using subgroup correlations for the two subgroups
separately if there is evidence of a moderating effect. The
two subgroups from the latter study were also treated as
independent samples. This left us with a final sample size
of ten papers consisting of twelve independent samples
(Table 1). Of the twelve samples, all of them reported
correlations between RA and job satisfaction, but only ten
of the twelve samples reported correlations between RC
and job satisfaction. The meta-analysis for each set of
variables is discussed in the following section.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
3
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
Table 1. Final sample for meta-analysis
r
Study
Sample # Sample N
1 Goldstein Rockart 1984
2 Bostrom 1981
Ivancevich, Napier, and
3 Wetherbe 1985
4 Li Shani 1991
Igbaria, Parasuraman, and
5 Badawy 1994
Cronbach's Alpha
Role
Role
Job
RA Æ JS RC Æ JS
Ambiguity Conflict
Satisfaction
-0.57
-0.40
0.81
0.79
0.80
1
118
2
3
75
221
-0.45
-0.38
-0.34
~
4
225
-0.02
~
~
~
~
0.85
~
0.93
5
109
-0.63
-0.41
0.91
0.85
0.84
6
224
0.39
0.43
0.86
0.90
0.86
7
214
0.54
0.46
0.88
0.89
0.88
6 Guimaraes Igbaria 1992
7 Igbaria Chidambaram 1997
8
209
-0.46
-0.36
0.84
0.78
0.82
9
348
-0.50
-0.41
0.82
0.87
0.70
8 Baroudi 1985
9 Lee 2000
10
229
-0.51
-0.39
0.85
0.79
0.91
11
273
-0.44
-0.17
0.81
0.82
0.83
10 King and Xia 2001
12
187
0.15
-0.50
0.71
0.70
0.86
* RA Æ JS: correlation between RA and Job Satisfaction;
RC Æ JS: correlation between RC and Job Satisfaction.
4. Data Analysis
There are four steps to conducting a meta-analysis:
correcting for any study artifacts, calculating effect sizes,
computing the Fail-Safe N, and evaluating the homogeneity or heterogeneity of sample effects.
The first step is to correct for measurement error. With
the reliability statistics for both independent and
dependent variables, measurement error was corrected
using the formula provided by Hunter and Schmidt [18].
In cases where the reliability values were missing in the
samples statistics, the averaged Cronbach’s alpha value
was used instead. After correcting for measurement error,
the sampling error was also corrected. The second step is
to calculate the effect size by averaging the reported
correlations, weighted by sample size [18, p. 150]. The
formula for effect size is
r=
¦ [N r ] , where r is the
¦N
i i
i
i
correlation after correcting for measurement error in study
i and Ni is the number of subjects in study i.
The third step is to calculate the Fail-Safe N, which is
a way to address the “file-drawer” problem – the
sampling bias in meta-analysis which assumes that an
unknown number of studies with effect sizes of zero were
not included in the analysis because they have either not
been submitted for publication or have been rejected [48].
That is to say, the actual mean effect size may not be as
high as the one generated by meta-analysis, due to the
fact any studies reporting weak or zero effect sizes are
less likely to be published, and thus less likely to be
included in the meta-analysis. However, “the combined
study results can be highly significant statistically even
though the mean effect size is small or even tiny” [18, p.
512]. Hunter and Schmidt [18] further provided the
formula to calculate Fail-Safe N, which indicates the
number of missing studies with zero effect size that would
have to exist to bring the mean effect size down to some
specific level (0.1 in this study). The lower the value of
the “Fail-Safe N” (e.g., a value = 2), then the more
uncertainty that exists with regard to whether the mean
effect size may be biased by the number of studies are
“buried in the file drawer”, due to weak, non-existent, or
contrary effects.
The fourth step examines whether the results are
consistent (homogeneous) across studies, or whether there
are large variations in the pattern of correlations. The
meta-analysis results are meaningful only if the various
samples are homogeneous. The purpose of the
homogeneity test is then to inspect whether any variations
in results across studies are due to sampling error, or due
to possible moderator variables [18]. Two criteria were
applied to evaluate homogeneity of the sample –
unexplained variance and the Q test of homogeneity. As a
rule of thumb, Hunter and Schmidt [18] suggested that if
the fraction of unexplained variance is less than 25%, this
variance can be neglected (i.e., due to some unmeasurable
errors). Otherwise, if the percentage of unexplained
variance is greater than 25%, there might be potential
moderators that affect the correlation. The second
criterion is the Q test of homogeneity. If the Q test value
is not significant, then there is no true variation across
studies; but if it is significant, it is reasonable to conclude
that there are moderating variables influencing the
correlations found from various studies.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
4
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
Table 2. Results of meta-analysis
Correlate
k
Total N
Mean Effect
Size
95% C.I.
Lower
Higher
% Unexplained
Fail-Safe N Variance in sample Homogeneity Test
correlations
Q
p
RA* --> JS
12
2432
-0.271
-0.311
-0.232
21
97.96%
502.737
0.0000
RC* --> JS
10
1986
-0.243
-0.287
-0.199
15
97.39%
339.646
0.0000
* RA Æ JS: correlation between RA and Job Satisfaction;
RC Æ JS: correlation between RC and Job Satisfaction.
Safe N values (21 and 15 for RA and RC, respectively)
are a sign of considerably strong results.
5. Results and Discussion
5.1. Hypotheses Testing
5.1.1. Hypothesis 1 and 2 – Supported. The results of
the meta-analysis are summarized in Table 2. Table 2
shows that the mean effect size for the correlation of RA
and job satisfaction is -0.271, and that between RC and
job satisfaction is -0.243. In general, both role stressor
variables – RA and RC – are negatively correlated with
job satisfaction. Thus, hypotheses 1 and 2 are both
supported. Based on Cohen’s [7] standard effect size
levels of 0.10, 0.30, and 0.50 as small, medium, and large,
respectively, the strength of the relationship between RA
and job satisfaction (-0.271) is close to medium, while
that between RC and job satisfaction (-0.243) is
somewhat weaker than medium.
Another principle to judge the significance of the mean
effect size is based on confidence intervals. According to
[39], if the confidence interval does not include the value
of zero, then the mean effect size is significant at the level
specified by the confidence intervals. The confidence
interval shown in Table 2 indicate that both of the mean
effect sizes are significant at the p<0.05 level, because
zero is not included in the confidence intervals.
Furthermore, according to Hunter and Schmidt [18],
even when the computed mean effect size is relatively
small, the cumulated study results can still be statistically
significant. When we examine the Fail-Safe N for each
relationship, we would need to locate 21 additional
studies with zero-effect sizes in order to reduce the
calculated mean correlation of 0.271 between RA and job
satisfaction to 0.1 (the threshold for weak correlation)
when our result is calculated based on the 12 studies that
we did find and code. Moreover, 15 studies with zeroeffect sizes would be required to reduce the mean effect
size from 0.243 correlation across studies to 0.1 when we
calculated the mean correlation between RC and job
satisfaction based on 10 studies.
In summary, both RA and RC are negatively correlated
with job satisfaction with medium strength (for RA) and
somewhat less than medium strength (for RC). The
correlations are significant at 0.05 level. The large Fail-
5.1.2. Hypothesis 3 – Not Supported. We hypothesized
that the magnitude of the negative effects of RA and RC
on job satisfaction would be different. To test this
hypothesis, the non-parametric Wilcoxon paired signed
rank test was conducted based on the corrected correlation
coefficient multiplied by the sample sizes for each
individual study. The result indicates that the correlations
of RA and RC with job satisfaction (after correcting for
sampling error and measurement error), did not differ
significantly from each other (Z=-1.078, p=0.139). Thus,
RA and RC have a statistically indistinguishable negative
effect on job satisfaction. Hypothesis 3 is disconfirmed.
5.1.3. Hypothesis 4 – Supported. The two meta-analysis
studies [10, 28] from the organizational behavior
literature provide a comparative basis for Hypothesis 4.
Jackson and Schuler’s [28] meta-analysis study included
96 papers none of which featured IS professionals. Fisher
and Gitelson’s [10] study reviewed 43 studies and 59
independent samples. Even though they didn’t provide
population information for the samples, a review of the
studies’ titles included in their bibliography did not
indicate the presence of any IS-related samples.
Therefore, the synthesized effect sizes from these two
studies represent the relationship between RA/RC and job
satisfaction across various types of occupations (which
appear not to have included IS employees).
The mean effect sizes in the current study are, by and
large, comparable to these two prior meta-analysis
studies. More specifically, the correlation between RA
and job satisfaction from this study (-0.271) falls between
the effect sizes obtained from the two prior meta-analyses
published in the organizational behavior literature (-0.25
and -0.30, respectively). In contrast, the correlation
between RC and job satisfaction in our study (-0.243) is
weaker than the effect sizes from both of these prior
studies (-0.35 and -0.31, respectively).
To further test the statistical significance of above
comparisons, a non-parametric Mann-Whitney U test was
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
5
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
performed. Since only one of the organizational behavior
studies [28] provided the original correlations for each
individual study in its meta-analysis, the correlations from
the current study are compared only with those from
Jackson and Schuler [28].1
The results of the Mann-Whitney U test indicated that
the correlations between RA and job satisfaction among
IS professionals are not statistically different from other
job types (Z = -0.049, p=0.961). Similarly, correlations
between RC and job satisfaction did not differ statistically
between IS professionals and other job types (Z = -0.895,
p=0.371). Thus, both Hypothesis 4a and 4b are supported.
While the mean effect size for the correlation between RC
and job satisfaction was slightly weaker (-0.243),
compared to the two prior meta-analyses for other job
types (-0.35 and -0.31)., this difference was not
statistically significant. .
These results indicate that even though most IS jobs
require both innovation and boundary spanning [2],
which, in turn, introduce highs level of RA and RC,
respectively [30], the negative impacts of RA and RC are
not necessarily stronger among IS professionals than for
other occupations. One potential explanation of this
phenomenon is that IS professionals are selected for (or
have learned to have) high tolerance for ambiguity and
role conflict [23, 31].
5.2. Moderating Effect
The unexplained variances in both of the correlations
are higher than the recommended threshold of 25%. This
indicates that sampling error only accounts for a small
percentage of the variance in sample correlations (2.92%
for RA, 3.51% for RC). Thus, there may be one or more
moderators having influence on the correlations found
from various studies. The highly significant homogeneity
Q test results also corroborates this conclusion. It is also
consistent with that of the two organizational behavior
meta-analyses [10, 28], namely, theoretically meaningful
moderators should be included to explain the variance of
both correlations across the samples.
On the other hand, the unexplained variance and the
significant Q test result from this study are both even
higher than the results obtained from general job settings.
However, considering that we conducted this analysis
only among IS professional and excluded the samples of
IT end users , we have , to some extent, controlled for
variations in job type. The significant results from the
homogeneity test suggest that large variations among the
broad category of IS professionals may exist. This may
indicate that a further specification of the job types of IS
1
In addition, because the reliability statistics (Cronbach’s
alpha) for each individual study are not shown in Jackson and
Schuler’s (1985) paper, our analysis is based on the original
correlations, instead of the weighted correlations after correcting
for sampling error and measurement error.
professionals are required, rather than simply aggregating
all employees together under the general category of IS
professionals.
According to Hunter and Schmidt [18], subgroup
analysis can be performed to detect possible moderator
variables. The hypothesized moderator variable can be
used to split the sample into two or more subsamples. The
same meta-analysis procedures described above then can
be carried out for each subset independently. For
example, the data could be analyzed separately for all
samples that consisted of “systems analysts,” and another
analysis for samples consisting of “project managers”. If
the variance of the effect sizes in each subset is lower
than that for the data as a whole, it shows that the variable
used to split the samples is a moderator variable which
causes the differences in average effect size.
Some IS researchers have inspected potential
moderators (job involvement, self-esteem, gender) for the
relationship between role stressors and job satisfaction.
Job involvement is defined as “the psychological
identification of an individual with a specific job” [25, pp.
177]. Based on a sample of 464 IS professionals holding
various job positions, Igbaria and colleagues [24] found
that the impact of role stressors on IS employees differed
significantly among different levels of job involvement
(low, moderate, and high). Particularly, the negative
consequences of both RA and RC are stronger among the
moderate and high job involvement subgroups than for
the low-involvement subgroup. The authors concluded
that:
these findings are analogous to those documented in the
work-family literature that high levels of job involvement
and family involvement heighten the level of workfamily conflict experienced by individuals and its
aversive effects on individuals’ well-being” [24, p. 193].
Prior research has also found self-esteem to moderate
the relationship between role stressors and psychological
and physical strain [29]. One IS study based on a sample
of 124 system developers revealed that self-esteem
moderates the relationship between role stress and job
satisfaction [44]. Nevertheless, the authors didn’t mention
the direction of the moderating effect in either their
hypotheses or results.
Igbaria and Chidambaram [23] investigated the impact
of gender on various aspects of IS professionals among
348 members of the Data Processing Management
Association (DPMA) professional association. Their
results showed that women and men reported no
significant differences in role stressors. After controlling
for gender, they found that both RA and RC were still
significant predictors of job satisfaction.
Another potential moderator that may differentiate
among the observed results is the diversity in job
satisfaction measurements used for different studies.
Among the ten studies included in our meta-analysis, job
satisfaction was measured by five different scales:
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
6
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
Hackman and Oldham's (1976) Job Diagnostic Survey
(JDS), Smith, Kendall and Hulin's (1969) Job Descriptive
Index (JDI), the Minnesota Satisfaction Questionnaire
Short Form (1967), measures from Hoy [16] , and revised
measures from Hoppock (1935). Some of these
instruments feature various dimensions of job satisfaction
(e.g., satisfaction with pay, promotional opportunities, coworkers, and the job itself) Yes only four studies reported
the correlation between role stressors and the specific subdimensions of their job satisfaction scales [50]. Such a
lack of specific results may inhibit the comparison
between general job satisfaction and these facets of job
satisfaction, which have been shown to differ in the
earlier meta-analysis studies [10, 28]. For example,
Jackson and Schuler’s study showed that it is necessary to
explore moderators for the relationship between RC and
general job satisfaction and also for RC and “satisfaction
with advancement opportunities,” while there was no
evidence of moderators for other sub-types of satisfaction
(satisfaction with supervision, work itself, co-workers,
and pay), because the percentage of unexplained variance
was close to zero.
Despite our identification of several potential
moderator variables, due to the minimal sample size
required in each sub-sample for conducting a moderator
analysis, and to the lack of necessary information for each
potential moderator, we were unable to conduct further
moderator analyses with the available data.
6. Conclusion
6.1. Summary of Research Findings
This paper presents the use of meta-analysis as a
statistical technique to integrate empirical IS research
results regarding the relationship between RA/RC and IS
professionals’ job satisfaction across independent studies.
In general, the results of this study found that both of the
role stressor variables – RA and RC – are negatively
correlated with IS professional’s job satisfaction across
diverse studies. The mean effect size for RA is more
strongly correlated with job satisfaction (-0.271) than the
analogous correlation between RC and job satisfaction (0.243), although the difference between them was not
statistically significant. Compared with non-IS related
jobs, IS professionals are not necessarily more- or lesssusceptible to the negative impacts of RA and RC on job
satisfaction. Also, the unexplained variance in the study
effect sizes indicates the existence of potential moderating
variables.
To understand the various demographic and contextual
factors that may differentiate various populations, future
studies should examine other, theoretically meaningful
moderators to explain the variations of both correlation
coefficients across samples of IS employees.
6.2. Limitations
It is important to clarify the limits or bounds of this
study. First, because of the lack of individual reliability
statistics from prior research [28], Hypothesis 4 is
statistically tested based on the original correlations,
instead of the weighted correlations after correcting for
sampling error and measurement error. Second, due to the
relatively small number of studies in the IS literature,
compared with the minimum sample size required for
moderating analysis, we were unable to conduct subgroup
moderating analysis in our study.
6.3. Implications and Contributions
This study is the first attempt in IS literature to
synthesize empirical study results for the correlations of
RA/RC with job satisfaction . The meta-analysis method
allows us to take every available empirical study into
consideration, overcoming the limitations of traditional
narrative reviews and provides a quantitative measure of
the relationship in question. Nevertheless, as advocated
by Hwang [19], the purpose of meta-analysis is by no
means to “determine the final word”, but “to take stock
and provide direction for future research” (p. 44). Despite
the limitations mentioned above, our research seeks to
provide future research directions in the following areas.
6.3.1. Research Directions for RA/RC Area. Past IS
studies have reported different levels of the relationship
between RA/RC and job satisfaction. The fact that our
meta-analysis supports both Hypotheses 1 and 2 in this
study helps to reduce this confusion. The lack of support
for Hypothesis 3 indicates that both RA and RC have
similar levels of negative impact on job satisfaction.
Thus, these two role stressors are both important
predictors of IS professional’s job satisfaction.
This meta-analysis study also contributes to IS
research by identifying the need to explore theoretically
based moderating variables to account for the different
levels of relationship between RA/RC and general job
satisfaction. The large unexplained variance in the mean
effect sizes is consistent with the two organizational
behavior studies, which also advocated the need for
identifying moderators for the relationship between
RA/RC and other dependent variables (including selfrated and superior-rated job performance, tension/anxiety,
and turnover intentions [10, 28]. Even though our study
only examined one dependent variable (job satisfaction),
considering the support for Hypotheses 4a and 4b (IS
professionals are not necessarily less susceptible to the
negative impacts of RA and RC on job satisfaction than
other occupations), it is possible that the correlations
between RA/RC with other outcome variables in IS
setting are also sensitive to certain moderators. Thus, IS
researchers should not limit their future research
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
7
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
regarding RA and RC to simple, bivariate relationships,
but should examine possible moderators such as gender,
job duties, self-efficacy, tolerance of ambiguity, need for
structure and other personal attributes.
Another future research direction is the investigation
of the correlation between role stressors and the specific
sub-dimensions of the job satisfaction scales (e.g.,
satisfaction with co-workers, pay, promotional
opportunities, and the job task itself), since the
organizational behavior meta-analyses [10, 28] showed
that RA/RC differ in terms of their effect sizes with
general job satisfaction versus other facets of job
satisfaction.
6.3.2. Application of Meta-analysis in IS Field. The
fourth issue is related to the broader application of metaanalysis method throughout the IS field than with the
specific research area of RA and RC.
6.3.2.1 Misuse of the Term Meta-Analysis. It is not the
focus of this research to provide a comprehensive
methodological review of the use of meta-analysis in IS
research, which has been covered quite thoroughly by
Hwang [19]. Nonetheless, there is one problem that has
emerged recently (from the mid-to-late 1990s) in IS
research – the misuse of the term meta-analysis. We used
the terms “meta-analysis” or “meta analysis” as keywords
to search for IS-related, published meta-analysis research
in several online databases and conference proceedings
(as describe in the “Literature Search” section) and
located 21 studies. All 21 studies had “meta-analysis” or
“meta analysis” in their article titles or abstracts, however
2
nine of these studies did not actually perform a true
meta-analysis.
Before we mention these studies, a restatement of the
meta-analysis definition will help in clarifying the
criterion used to judge whether a study constitutes a true
meta-analysis. The term meta-analysis refers to “the
analysis of analyses … the statistical analysis of a large
collection of analysis results from individual studies for
the purpose of integrating the findings” [11, p. 3]. This
definition indicates that in order to be considered a metaanalysis study, the study should: 1) collect quantitative
results from multiple prior studies; 2) perform statistical
analysis on these quantitative data; and 3) focus on
analysis of effect size, an indicator of the strength of the
relationship between the relevant research variables [11].
The most common measures of effect sizes are Pearson’s
r (for survey research) and Cohen’s d (for laboratory
research).
2
Among the 21 studies, three of them are review papers for
meta-analysis itself, nine of them didn’t employ true metaanalysis procedures, the other nine of these studies (which are
denoted with ** in the reference) did actually perform a true
meta-analysis.
The three criteria that we identified are used here to
evaluate the nine papers that do not meet Glass’s [11]
definition of meta-analysis, defined above. Among the
nine papers, two of them first qualitatively reviewed
existing research to either develop a research framework
[11] or to identify information system success (ISS)
factors [36], and then collected first-hand data to either
evaluate the framework or study the rank order of the ISS
factors. Similarly, three other papers [15, 33, and 43] only
conducted a qualitative review of prior research, with no
quantitative data analysis involved. These five studies
satisfied none of the three meta-analysis criteria. In all
cases, the authors’ use of the term “meta-analysis” in their
paper titles or abstracts actually referred to the traditional
qualitative literature review – which Hunter and Schmidt
[18] label as “narrative review.”
The other four papers met the first two criteria, in that
they gathered quantitative data from prior research and
performed some statistical analysis on these data.
However, none of these studies considered effect sizes as
the focus of their study. That is, none of them actually
integrated statistical results from prior studies’ research
variables. Instead, they examined other characteristics of
the secondary data. Palvia and his colleagues [45]
collected secondary data about the ranking of key IT
issues, and then used cluster analysis to regroup them and
used t-tests to verify the categories. Chau [6] investigated
the relationship between construct reliability and research
design characteristics using ANOVA to identify which
design characteristics affect reliability. Esteves and
Ramos [9] provide a descriptive statistical overview of IS
studies conducted in Portugal. Kohli and Devaraj [34]
employed logistic regression and discriminant analysis to
examine the structural variables that affect IT payoff.
While the last study seems closest to a true meta-analysis,
the authors did not re-analyze the original effect sizes
(correlation coefficients) from the original studies; rather,
they simply counted the number of prior studies with
positive and significant, zero, or non-significant findings
as their form of analysis and conducted some analysis on
the numbers of studies in each category. They did not
examine the actual strength of the correlation, regression,
or Lisrel paths in the prior studies and thus, their study
neglected to analyze true effect sizes.
These examples of studies that mis-use the term metaanalysis show that this method is not yet well understood
by the IS research community. This is unfortunate, since
there are meta-analyses in the IS literature going back
over a dozen years [1]. An unexpected but noteworthy
contribution of our study is to raise this issue and
emphasize the correct use of the term, meta-analysis,
which is elaborated in the following section.
6.3.2.2. Correct Use of the Term Meta-Analysis. The
evaluation of the papers that do not meet Glass’s [11]
definition of meta-analysis indicates that the meaning of
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
8
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
meta-analysis has been diluted, so that some researchers
tend to mis-label two other types of studies as metaanalysis. The first type is the traditional qualitative
literature review or narrative review (e.g., [11], [36], [15],
[33], and [43]). This is by no means intended to de-value
the contributions that narrative reviews can make, but
there are certain circumstances under which a qualitative
review is unable to provide an accurate representation of
population characteristics. One such situation is when the
number of prior studies is large and the studies are not
comparable in terms of sampling, measures, and so forth.
This limitation can, to certain degree, be well handled by
meta-analysis. For example, Hunter and Schmidt
recommended procedures that can “deal with variations in
study effect size due to sampling error and other artifacts
and with attenuation of effect-size estimates due to
measurement unreliability and range restriction” [18, p.
485]. Thus, qualitative literature review and meta-analysis
are two different methods which are suited for different
objectives – the term meta-analysis is not a synonym for
qualitative literature reviews.
The second type of study that often being mis-labeled as
meta-analysis is research that does include quantitative
analyses based on existing literature, however, the
quantitative analysis does not focused on prior studies’
research variables (e.g., [45], [6], [9], and [34]). Again,
we have no intent to discredit the value of this type of
study. Nevertheless, the emphasis of these studies is
substantially different from the objective of metaanalysis. Therefore, the term meta-analysis is not
interchangeable with this type of research.
In conclusion, we advocate the term meta-analysis should
only be used for those studies that: 1) collect quantitative
results from multiple prior studies; 2) perform statistical
analysis on these quantitative data; and 3) focus on
analysis of effect size, an indicator of the strength of the
relationship between the actual research variables from
those prior studies[11].
7. References3
[1] Alavi, M. and Joachimsthaler, E. A. “Revisiting DSS
Implementation Research: A Meta-Analysis of the
Literature and Suggestions for Researchers,” MIS
Quarterly, 16:1, March 1992, pp. 95-116.
[2]*Baroudi, J.J. “The Impact of Role Variables on IS
personnel Work Attitudes and Intentions,” MIS Quarterly,
December 1985, pp. 341-356.
[3] Bartol, K.M., and Martin, D. "Managing Information
Systems Personnel: A Review of the Literature and
3
The papers denoted with * were included in our quantitative
meta-analysis. The papers denoted with ** were the prior
studies in the IS literature that did employ true meta-analysis
procedures.
Managerial Implications," MIS Quarterly (6), December
1982, pp. 49-70.
[4]*Bostrom, R. P. "Role Conflict and Ambiguity:
Critical Variables in the User-Designer Relationship,"
Proceedings of the Seventeenth Annual Computer
Personnel Research Conference, Association for Computing
Machinery Press, New York, 1981, pp. 88-112.
[5]**Cao, Q., Nah, F.F., and Siau, K. “A Meta-Analysis
on Relationship Modeling Accuracy: Comparing
Relational and Semantic Models”, Proceedings of AMCIS
Conference, 2000, Long Beach, CA, 470-473.
[6] Chau, P.Y.K. “On the Use of Construct Reliability in
MIS Research: A Meta-Analysis,” Information &
Management (35), 1999, pp. 217-227.
[7] Cohen, T.D. Statistical Power Analysis for the
Behavioral Sciences, (2nd ed.). Jillsdale, NJ: Lawrence
Erlbaum Associates, 1988.
[8]**Dennis, A.R., Haley, B.J., and Vandenberg, R.J. “A
Meta-Analysis of Effectiveness, Efficiency, and
Participant Satisfaction in Group Support Systems
Research”, Proceedings of ICIS Conference, 1996,
Cleveland, OH, 278-289.
[9] Esteves, J. and Ramos, I. “A Meta-Analysis of
Information Systems Research Diversity in Portugal”,
Proceedings of AMCIS Conference, 2003, Tampa, FL,
966-974.
[10] Fisher, C.D. and Gitelson, R. “A Meta-Analysis of
the Correlates of Role Conflict and Ambiguity,” Journal
of Applied Psychology (68:2), 1983, pp. 320-333.
[11] Glass, G.V. “Primary, secondary, and metaanalysis,” Educational Researcher (5), 1976, pp. 3-8.
[12]*Goldstein, D. K., and Rockart, J. F. "An
Examination of Work-Related Correlates of Job
Satisfaction in Programmer/Analysts," MIS Quarterly (8),
1984, pp. 103-115.
[13] Greenhaus, J.H. and Beutell, N.J. “Sources of
Conflict Between Work and Family Roles,” Academy of
Management Review (10:1), 1985, pp. 76-88.
[14]*Guimaraes, T. and Igbaria, M. “Determinants of
Turnover Intentions: Comparing IC and IS personnel,”
Information Systems Research (3:3), 1992, pp. 273-303.
[15] Heine, M.L., Grover, V. and Malhotra, M. K. “The
Relationship between Technology and Performance: A
Meta-Analysis of Technology Models,” Omega (31:3),
June 2003, pp. 189-204.
[16] Hoy, F. Small Business Survey Questionnaire.
Athens, GA: Small business Development Center
Publications, The University of Georgia, 1983.
[17] Huarng, A. S. “Burnout syndrome among
information system professionals,” Information Systems
Management (18:2), Spring 2001, pp. 15-xx.
[18] Hunter, J.E. and Schmidt, F.L. Methods of MetaAnalysis, Correcting Error and Bias in Research
Findings. Thousand Oaks: Sage Publication, 1990.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
9
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
[19] Hwang, M.I. “The Use of Meta-Analysis in MIS
Research: Promises and Problems,” The DATA BASE for
Advances in Information Systems (27:3), 1996, pp. 35-48.
[20]**Hwang, M. “Did Task Type Matter in the Use of
Decision Room GSS? A Critical Review and a MetaAnalysis,” Omega (26: 1), 1998, pp. 1-15.
[21]**Hwang, M.I. and Wu, B.J.P. “The Effectiveness of
Computer Graphics for Decision Support: A MetaAnalysis Integration of Research Findings,” The DATA
BASE for Advances in Information Systems (21:3), 1990,
pp. 11-20.
[22]**Hwang, M.I., Windsor, J.C., and Pryor, A.
“Building a Knowledge Base for MIS Research: A MetaAnalysis of a Systems Success Model,” Information
resources Management Journal (13:2), 2000, pp. 26-32.
[23]*Igbaria, M. and Chidambaram, L. “The Impact of
Gender on career success of information systems
professionals:
A
Human-Captial
Perspective,”
Information Technology & People (10:1), 1997, pp. 6386.
[24] Igbaria, M., Meredith, G., and Smith, D.C.,
“Predictors of Intention of IS Professionals to Stay With
the Organization in South African,” Information &
Management (26), 1994, 245-256.
[25]*Igbaria, M., Parasuraman, S. and Badawy, M.K.
“Work Experiences, Job Involvement, and Quality of
Work Life Among Information Systems Personnel,” MIS
Quarterly, June 1994, 175-201.
[26] Ivancevich, M., Napier, A. H., and Wetherbe, J. C.
"Occupational Stress, Attitudes, and Health Problems in
the Information Systems Professional," Communications
of the ACM (26:10), 1983, pp. 800-806.
[27]*Ivancevich, M., Napier, A. H., and Wetherbe, J. C.
“An Empirical Study of Occupatoinal Stress, Attitudes
and Health Among Information Systems Personnel,”
Information & Management (9), 1985, pp. 77-85.
[28] Jackson, S.E. and Schuler, R.S. “A Meta-Analysis
and Conceptual Critique of Research on Role Ambiguity
and Role Conflict in Work Settings,” Organizational
Behavior and Human Decision Processes (36), 1985, pp.
16-78.
[29] Jex, S. and Elacqua, T. “Self-esteem as a Moderator:
A Comparison of Global and Organization-based
Measures,” Journal of Occupational & Organizational
Psychology (72:1), 1999, pp. 71-81.
of Organizational & End User Computing (16:1), 2004,
pp. 59-72.
[42]**Mahmood, M.A., Hall, L., and Swanberg, D.L.
“Factors Affecting Information Technology Usage: A
Meta-Analysis of the Empirical Literature,” Journal of
Organizational Computing and Electronic Commerce
(11:2), 2001, pp. 107-130.
[43] Marshall, C., Shipman, F. M., and McCall, R. J.
“Making Large-Scale Information Resources Serve
[30] Kahn, R.L., Wolfe, D., Quinn, R., Snoek, J.D., and
Rosenthal, R. Organizational Stress: Studies in Role
Conflict and Role Ambiguity, John Wiley and Sons, New
York, New York, 1964.
[31] King, R.C. and Sethi, V., “The Impact of
Socialization on the Role Adjustment of Information
Systems Professionals,” Journal of Management
Information Systems (14:4), Spring 1998, pp. 195-217.
[32]*King, R.C. and Xia, W.D., “Retaining IS Talents in
the New Millennium: Effects of Socialization on IS
Professionals’ Role Adjustment and Organizational
Attachment,” Proceedings of the ACM SIGCPR, San
Diego, CA., 2001, 144-156.
[33] Klaus, H., Rosemann, M. and Gable, G.G. “What is
ERP?” Information Systems Frontiers (2:2), 2000, pp.
141-162.
[34] Kohli, R. and Devaraj, S., “Measuring Information
Technology Payoff: A Meta-Analysis of Structural
Variables in Firm-Level
Empirical Research,”
Information Systems Research (14:2), 2003, pp. 127-145.
[35]*Lee, P.C.B., “Turnover of Information Technology
Professionals: a contextual model,” Accounting,
Management and Information Technology (10),, 2000,
101-124.
[36] Li, E.Y. “Perceived Importance of Information
System Success Factors: A Meta Analysis of Group
Differences,” Information & Management (32), 1997, pp.
15-28.
[37]*Li, E.Y., and Shani, A. B. "Stress Dynamics of
Information Systems Managers: A Contingency Model,"
Journal of Management Information Systems (7:4), 1991,
pp. 107-130.
[38]**Lim, J., Richardson, V.J., and Roberts, T.L.
“Information Technology Investment and Firm
Performance: A Meta-Analysis,” Proceedings of the 37th
HICSS, 2004, HI, pp. 1-10.
[39] Lipsey, M.W., & Wilson, D.B. (2001). Practical
Meta-analysis. Thousand Oaks: Sage.
[40] Locke, E. “The Nature and causes of Job
Satisfaction,” in Handbook of Industrial and
Organizational Psychology, M.D. Dunnette (ed.), Rand,
Chicago, Illinois, 1976, pp. 1297-1349.
[41]**Ma, Q. and Liu L. “The Technology Acceptance
Model: A Meta-Analysis of Empirical Findings,” Journal
Communities of Practice,” Journal of Management
Information Systems (11: 4), Spring 1995, pp. 65-86.
[44] Nelson, A.C. and LeRouge, C., “Self-Esteem:
Moderator Between Role Stress Fit and Satisfaction and
Commitment?,” Proceedings of the ACM SIGCPR,
2001, San Diego, CA, pp. 74-77.
[45] Palvia, P.C., Palvia, S.C.J., and Whitworth, J.E.,
“Global Information Technology: A Meta-Analysis of
Key Issues,” Information & Management (39), 2002, pp.
403-414.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
10
Proceedings of the 38th Hawaii International Conference on System Sciences - 2005
[46] Premkumar, G., “A Meta-Analysis of Research on
Information Technology Implementation in Small
Business,” Journal of Organizational Computing &
Electronic Commerce (13:2), 2003, pp. 91-121.
[47] Rizzo, J.R., House, R.J., and Lirtzman, S.I., “Role
Conflict and Ambiguity in Complex Organizations,”
Administrative Science Quarterly (15:2), 1970, pp. 150163.
[48] Rosenthal, R. “The “file drawer problem” and
tolerance for null results,” Psychological Bulletin (86),
1979, pp. 638-641.
[49]**Sharma, R. and Yetton, P. “A Contingent Model
for the Role of Management Support in IS
Implementation: A Meta-Analysis”, Proceedings of
ICIS Conference, 1997, Atlanta, GA, 492-493.
[50] Smith, P., Kendall, L. and Huliln, C., The
Measurement of Satisfaction in Work and Retirement,
Rand, Chicago, Illinois, 1969.
[51] Viswesvaran, C., and Ones D.S., “Theory Testing:
Combining Psychometric Meta-Analysis and Structural
Equations Modeling,” Personnel Psychology (48), 1995,
865-885.
0-7695-2268-8/05/$20.00 (C) 2005 IEEE
11