A Comparison of Business Conununication Quality between

The present study compared handwritten reports with those written on computers for
ten precisely defined quality factors. Reports written under both conditions by 160
upper division college students in a business communication course were submitted to
the Grammatik 11 program, for quality analysis and the results were subsequently
analyzed for significant differences using ANOVA. More similarities than differences
were found as six ofthe ten tests did not show significant differences. The findings of
the study do not support a general conclusion that basic writing qualify will improve
when documents are written by computer as compared to writing them by hand. The
importance ofthe present findings as they relate to earlier research and future research directions are discussed.
A Comparison of Business Conununication
Quality between Computer Written and
Handwritten Samples
C. Glenn Pearce & Randolph T. Barker
Virginia Commonwealth University
he computer is becoming an accepted organizational and pedagogical
T
tool in business communication, but does it improve the quality of
the written output? Much effort has been placed on developing word
processing packages and courses, designing electronic mail services,
encouraging (and in some cases, requiring) students to own a computer
for coursework, and using computer laboratories as part of business
commimication classes. Yet, the question remains as to whether this
acceptance and implementation of the technology has produced the
anticipated results.
LITERATURE REVIEW
In a critical review ofthe literature on writing with word processors.
Hooper (1987) noted that most studies of h i ^ school and college English
writing courses support computer use because "revision is made easier"
(p. 5). Nevertheless, research to date regarding the effect of computer
use on writing quality and quantity has produced conflicting findings.
Several authors have found that using the word processor for writing
activities enhances communication. Maik (1987) noted increases in
revision and editing in college business writing courses. Tone and
Winchester's (1988) ERIC digest also noted that in some cases using the
computer had an effect "if not a dramatic impact on both the quality and
quantity of writing" (p. 2). Tone and Winchester looked at studies of
English and writing composition conducted in elementary, secondary.
141
142
The Joumal af Business Cammunicatian
28:2:Spriiig 1991
and college courses. They also found, however, that in many studies,
quantity did not increase, nor did quality improve.
McCallister and Louth (1988) examined the effects of word processing
on the quality of revisions. The findings revealed that the quality of
college basic English student compositions was improved when using the
word processor as compared to pen and pencil. Also focusing on college
basic writing, McCuthen, Hull, and Smith (1987) found that using an
interactive editor, supplemented by a simple word processing package,
resulted in more corrections, particularly for those errors requiring a
consulting strategy to apply grammatical rules.
Comparing the performance of junior high school students using the
word processor with students using pen and paper, Kurth (1987) found
that the word processing group wrote more substantial compositions,
although the quality of global revisions was not improved. Hawisher
(1987) found similar results upon investigating the effects of word
processing on the revision strategies of college freshmen in EngUsh
classes. Essays produced by pen and typewritor received quality ratings
comparable to those produced with the computer.
Haas (1989) found that experienced and student writors planned less
when writing on the computor than when writing by hand or when using
a combination of the two methods. Also studying the work of experienced
writers, Lutz (1987) found no clear distinctions for either professional
writers or Ph.D. writing theory students. The author did noto a trend,
though, for both groups to make more changes on the computor when
revising and editing.
Regarding quality issues, Bemhardt, Eldwards, and Wojahn (1989)
found that computor writton college compositions were of bettor quality
than those writton by hand. On the other hand. Collier (1983b) found
that the quality of writing did not improve when nursing students used
the computor compared with their using conventional methods; however,
the quantity of text writton was greator. In another study by Collier
(1983a), college student texts completod hy hand, typewritor, and computor were compared for creativity and writing skills. Analysis revealed
that both writing and creativity were enhanced by using either the
typewritor or the computor.
Focusing on syntax and quality, Etohison (1985) conductod a comparative study of handwriting and word processing. The results showed that
the quality of college student compositions increased for those subjects
using word processing. However, some syntactic variables were not
affectod by the mode of composition. Daiuto (1986), investigating draft
editing and revision quality, found that high school student writors
Business Communication Quality • Pearce/Barker
143
corrected more errors when using a word processor than when using pen
and paper. Quality ratings on the draft texts were h i ^ e r for the pen and
paper compositions, but the computer texts contained more revisions.
To summarize the current status of the research findings, diverse
designs and methodologies have been used, so it is difficult to compare
the various fmdings. Further, definitions of quality have varied greatly,
very small sample sizes have been frequent, and measurements of
writing quality have been primarily holistic ratings or ratings based on
analytical scales. For the most part, subjects have been high school or
college basic English students. Other researchers have su^iested that a
more complex array of factors beyond writing method may affect writing
quality, such as the complex interaction of physical and cognitive processes (Aschauer & White, 1984; Daiute, 1986; Herrmann, 1987; Hooper,
1987).
The research question directing this study was: Is there a quality
difference between handwritten reports and those written on computers
for precisely defined and measured quality factors using upper division
college students as subjects?
METHODOLOGY
All 266 students enrolled in 10 sections of a required undergraduate
business communication course at a msgor mid-Atlantic university were
target participants for the study. Only juniors and seniors may take Hie
course, prerequisites for which are junior standing and completion of two
introductory written composition courses.
Subjects who were present on the first two consecutive class days of
the fall semester comprised the sample. First day participants numbered
198, while second day participants numbered 204. Of these, 182 attended
both days and could therefore provide comparative writing samples. Of
this number, 22 were eliminated fi-om the study for failing to follow
directions or submitting incomplete writing samples. As a result, returns
from 160 subjects, 60.15 percent ofthe population, were analyzed. This
percentage is not deemed unusual for the first week of classes since
students actively add and drop classes at ihis time. Additionally, some
students assume that introductory information with minimal course
content will be covered during the initial classes and therefore do not
attend. Despite these constraints, for the purpose of this study, it was
necessary to collect the data during the first week, before any course
content was delivered.
144
The Journal of Business Communication
28:2:Spring 1991
Fifty-six percent (n = 89) of the subjects were females and 44 percent
(n = 71) were males. The majority (71 percent, n = 114) of the subjects
were ages 18 to 22, 15 percent (n = 25) were ages 23 to 28, 7 percent (n
= 11) were 29 to 34, 3 percent (n = 5) were 35 to 40, and 3 percent (n =
5) were over 40. Eighty percent (n = 129) of the subjects were white, 14
percent (n = 23) were black, 4 percent (n = 7) were Asian or Pacific
Islanders, and 0.6 percent (n = 1) were American Indian or Alaskan
Native. All subjects were upper division students, with 82 percent (n =
132) being juniors and 17 percent (n = 28) being seniors. Fi%-seven
percent (n = 92) had some experience using a computer, 21 percent (n =
34) had little to no experience, and 21 percent (n = 34) had much to a
high level of experience. In terms of experience using word processing
programs, 21 percent (n = 31) had no experience, 28 percent (n = 42) had
a httle experience, 36 percent (n = 53) had some experience, and 15
percent (n = 21) had much to a high level of experience. As to keyboarding
or typing experience, 10 percent (n = 17) indicated no experience, 11
percent (n = 19) had little experience, 39 percent (n = 63) had some
experience, and 39 percent (n = 63) had much to a high level of experience.
Data Collection
Two business report writing cases developed by one of the researchers
provided the framework for the data collection. Each case consisted ofa
business scenario in response to which subjects were to compose short
reports, within parameters given in a set of written instructions. Students were to write the text for a short report based on facts provided in
each case without regard for headings or format considerations. Subjects
were instructed to add relevant information to the case as needed; and
to accommodate differences in preknowledge of the subject matter of the
case, subjects were instructed that the document produced would be
evaluated for writing quality but not for factual accuracy. Also, the
subjects were told to correct their language (writing) and grammatical
errors as they saw fit. Both cases were analyzed for appropriateness and
parallel difficulty by six university professors ofbusiness communication
working independently, and revisions were made accordingly.
Research Design
A cross-over design was used in this investigation of the effect of
writing methods on writing quality. Since the order in which participants
completed the two case studies might have had an effect on writing
quality, order was included as an independent variable. The treatment
Business Communication Quality • Pearce/Barker
145
variable, writing method, has two levels: computer written and handwritten. The dependent variable, quality, has 10 levels addressing both
v/riting and grammatical errors. These levels are readability, passive
construction, use of prepositions, trite expressions, wordy expressions,
and redundant expressions, all of which represent writing errors; and
capitalization, punctuation, spelling, and vague adverbs, which represent grammatical errors.
Procedures
This study was completed during the first two class sessions, and
students were told that this project was part of the standard course
requirements and would count as part of their course grade. Each subject
was randomly assigned to one of the two groups. One group completed
the handwritten case first, followed by the computer written case the
following day. The other group completed the computer written case first,
followed by the handwritten case the following class day. Handwritten
reports were completed in the classroom, while computer written reports
were completed in a university computer laboratory. All subjects used
WordPerfect software on IBM compatible microcomputers and were
instructed not to use dictionaries or any other writing or grammatical
aids.
A management professor, one of the researchers, administered the
written cases to both groups. Although the possibility of experimenter
bias exists, the use of written administrative procedures minimized that
possibility. An information systems professor administered the computer written cases to both groups, also following written instructions.
Only questions regarding computer operating mechanics were
answered. Students did not ask either case administrator any questions
about the cases.
Handwritten reports of <he cases were entered into the computer for
analysis. Both the handwritten and computer written reports were
scored for errors using the Grammatik II (1986) software program. While
Grammatik provides 14 quality measures, the analysis was limited to
10 measures deemed most important by the previously mentioned
professors. This action was necessary because the Grammatik program
will review no more than ten factors in a document
Analysis of variance with repeated measures was performed to investigate the effect of writing method and order of case completion on the
ten indicators of quality.
146
The Journal (^Business Communication
28:2:Spring 1991
RESULTS
Table 1 shows the means for the four types of grammatical errors
measured. A statistically significant difference (p < .05) was found
between computer written and handwritten reports for punctuation,
with handwritten reports containing more punctuation errors (M =
.1500) than those written on the computer (M = .0500). No significant
differences were found between writing methods for errors in capitalization, spelling, and use of vague adverbs.
Table 1
Grammatical Error Means for Writing Method
and Case Completion Order
Error
Writing Method
Computer Hand
Completion Order
First
Second
Capitalization
Punctuation
Spelling
Vague Adverbs
.1062
.0500*
.0000
.0937
.1625
.0312*
.0062
.1125
.1687
.1500*
.0125
.1125
.1125
.1687*
.0062
.0937
Note, n = 160.
*p < .05
A statistically significant difference (p < .05) was found for punctuation errors between reports completed first and those completed second.
More punctuation errors were made on reports completed second (Af =
.1687) than on the initial reports (M = .0312). No significant differences
were found in order of completion for errors in capitalization, spelling,
and use of vague adverbs.
Table 2 shows the means for the six types of writing errors measured.
A statistically significant difference (p < .05) was found between computer written and handwritten reports for readability, passive construction, and trite expressions. The computer written reports were easier to
read (M = 56.4125) but contained more passive construction (Af =
24.0060) and trite expressions (Af = .7000) than the handwritten reports
(readability, M = 53.0750; passive construction, M = 16.3690; trite
expressions, M = .2125). No statistically significant differences between
writing methods were found for errors in use of prepositions, wordy
expressions, or redundant expressions.
Statistically significant differences between reports completed first
and second were found for readability, passive construction, and use of
prepositions. Reports completed first were more difficult to read (Af =
Business Communication Quality • Pearce/Barker
147
Table 2
Writing Error Means for Writing Method and Case Completion Order
Error
Writing Method
Computer
Hand
Completion Order
First
Second
Readability
Passive
Preposition
Trite
Wordy
Redundant
56.4125*
24.0060*
12.9063
.7000*
.1562
.0062
53.0750*
17.8000*
11.8375*
.4750
.1000
.0000
53.0750*
16.3690*
12.1250
.2125*
.1062
.0000
52.9812*
22.5750*
13.1937*
.4375
.0062
.0062
Note, n = 160.
*p< .05
53.0750) than reports completod second (Af = 52.9812). However, initial
reports contained fewer instances of passive construction (Af = 17.8000)
and fewer prepositions (Af = 11.8375) than did the reports completod
second (passive construction, M = 22.5750; prepositions, Af = 13.1937).
No significant differences were found in order of report completion for
trite, wordy, or redimdant expressions.
Since the cross-over design was used in this analysis, tosting for
intoraction effects was not necessary (Cochran & Cox, 1954).
DISCUSSION
The data revealed two general findings regarding differences in
quality between handwritton and computor writton reports. First, the
ten variables imder scrutiny showed more similarities than dissimilarities as six of the tosts were not significant. Likewise, six of the
ton tosts for differences in the order in which subjects wroto the cases
were not significant Second, the results were mixed to the extont that
no trends were revealed. Mixed results were also reported by Hooper
(1987) and Tone and Winchestor (1988) in their critical reviews. And
although Bemhardt et al. (1989), Collier (1983a), Etohison (1985), Maik
(1987), McCalUstor and Louth (1988), and McCuthen et al. (1987)
reportod findings favoring computor writton aver handwritton documents, Daiuto (1986) found handwritton texts superior, while Collier
(1983b) and Hawisher (1987) reported no differences in quality.
Two m^jor points should be considered when evaluating the definitiveness of the reported studies. First, the subjective nature of the
measurements and weaknesses in the design or execution render the
148
The Joumal of Business Communication
28:2:Spring 1991
results questionable (Bemhardt et a l , 1989; Collier, 1983a, 1983b;
Etchison, 1985; Maik, 1987; McCallister & Louth, 1988). Even in instances where raters were trained and acceptable inter-rater reliability
coefficients were established (Bemhardt et al., 1989; Collier 1983a,
1983b; Daiute, 1986; Etchison, 1985; Hawisher, 1986; McCallister &
Louth, 1988), scoring consistency does not necessarily indicate valid
assessment of the target variables. Evidence of this effect is indicated by
the confounding variables extant in studies by Collier (1983a), Etchison
(1985), Daiute (1986), and Hawisher (1987). The second limiting point is
that often quality appeared to be but a secondary measure (Daiute, 1986;
Kurth, 1987; Maik, 1987; Lutz, 1987; Haas, 1989)
The finding that computer written reports contained fewer punctuation errors is partially corroborated by McCuthen et al. (1987), whose
subjects consulted an interactive editor for grammatical rules (which
included punctuation) as they wrote. Further indirect support is
evidenced in the Bemhardt et al. (1989) finding that computer written
compositions were superior; following conventions, which included
punctuation, was one of the minor variables rated. In contrast, Hawisher
(1987) found that pen and typewriter produced essays were comparable
to those written by computer; punctuation was a minor variable rated in
this study as well.
Computer written reports were found to be easier to read according
to Flesch's readability formula (Flesch, 1949), but handwritten reports
contained fewer instances of passive construction and fewer trite expressions. As far as can be determined, only Sterkel et al. (1986) have
measured readability, and that was for the work of experimental groups
before revisions were made.
The data showed that reports written first contained fewer pimctuation errors, were easier to read, and contained fewer instances of passive
construction and fewer prepositions than those written second. However,
the meaning of these results is difficult to determine given the available
data. Fatigue could not have played a part, for example, because a full
day lapsed between writing the first and second reports. Also, had
experience played a part, it would have favored the second writing, not
the first. Possibly, the subjects wrote the first report more deliberately,
but that would conflict with Hawisher's (1987) assertion that qixaUty
improved from the first to final drafts.
In several ways, the present study appears to be an initial investigation into certain areas of computer writing research. For example, except
for Sterkel et al. (1986), this appears to be the first specific study of
business writing qualities. To distinguish business writing from other
Business Communication Quality • Pearce/Barker
149
types of transactional writing may be important because the contexts
can differ. For example, business writing is often done under time
pressures that make it difficult to prepare successive revisions of documents (Gilsdorf, 1987), an activity ofton encouraged when students writo
on a computor, and required in process-centored instruction. Therefore,
the context in which business managers work appears to make it
desirable for business communication students who are preparing for
managerial positions to practice preparing initial drafts for evaluation
with no opportunity for revision, as was done in the present study.
The litorature review did not reveal any other studies using passive
construction or trito expressions as measvu-es of quality, or using upper
division undergraduatos as subjects. As a result, no comparisons on
these factors are possible.
Rather than using raters, this study employed a computor toxt
analysis program to measure quality after the subjects wroto their
documents. While Storkel et al. (1986) used a similar program, the
quality measure was grades received in the course. Further, the present
study is one of the first to measure basic quality variables as independent
measures, as opposed to holistic measures or measures using analytical
scales. While Etohison (1985) did measure syntax variables independently, they are unlike the basic writing variables measured in the
present study.
CONCLUSION
In the larger view, the results of the present study do not support a
conclusion that basic writing quality will improve when a computor is
used. Rather, the evidence appears to support Hooper's (1987) report of
conflicting results and Hawisher's (1987) conclusion that the writing tool
is not the variable that influences success.
Given the mixed results concerning the effects of computor writing on
quality, the findings of Bemhardt et al. (1989) that the toacher has the
greatost effect on performance perhaps bears further scrutiny. Were that
the case, toacher selection and training could be productive avenues for
further investigation.
The current research on computer writing strongly supports
Hawisher's (1986) observation that poorly conceived and incomparable
research designs make it difficult to generalize. In fact, possibly current
research has produced confounding results at least partly because the
measuring devices are subjective and imprecise. If so, studies of writing
quality that focus on such lower order variables as syntax, grammatical
150
The Journal of Business Communication
28:2:Spring 1991
errors, and writing errors, where objective measures can be taken, may
prove to be productive. If these variables are the building blocks of
writing quality, study of lower order variables could lead to a better
understanding ofmiddle order effects such as those measured by analytical scales and holistic ratings.
Another promising avenue for further investigation is to evaluate
writers themselves along with evaluating the documents they write.
Little research thus far has focused on demographic variables that may
explain differences in writing quality. In fact, a limitation ofthe present
study is the variability in computer experience and typing ability. Future
studies might focus on the effects of such factors.
REFERENCES
Aschauer, M., & White, F. (1984, February). Towards a marriage of two minds:
The word processor and natural habits of thought in the discovery stage of
composing. Pajier presented at the Spring Conference of the Delaware
Valley Writing Council and ^^llanova Universit3^s English Department,
Villanova, PA.
Bauer, R. (1986). Gram.matik H: The writing analyst. San Francisco: Reference
Software, Inc.
Bemhardt, S. A , Edwards, P., & Wojahn, P. (1989). Iteaching college cranposition with computers. Written Communication, 6(1), 108-133.
Cochran, D. S., Giallourakis, M. C , & Shim, J. P. (1986). PCs in the BC
classroom: Present and future perception. The Bulletin ofthe Association
for Business Communication, 49(1), 7-11.
Cochran, W. G., & Cox, G. M. (1954). Experimental designs (2nd ed.). New York:
John Wiley & Sons, Inc.
Collier, R. (1983a). The effects of a computer-based text-editor on the creativity
and writing skills of community college composition students: A pilot study.
Alberta, Canada. (ERIC Document Reproduction Service No. ED 246 430)
Collier, R. (1983b). The word processor and revision strategies. College Composition and Communication, 35,149-155.
Cortese, J. G. (1989). Computer applications in teaching business and professional writing courses in four-year colleges: Research suggesting curricular changes. The Bulletin of the Association for Business
Communication, 42(4), 3-5.
Daiute, C. (1986). Physical and cognitive factors in revising: Insights from
studies with computers. Research in the Teaching ofEnglish, 20(2), 141-159.
Etchison, C. (1985). A comparative study ofthe quality and syntax of compositions by first year college students using handwriting and word processing.
Glenville, WV: Glenville State College. (ERIC Document Reproduction
Service No. ED 282 215)
Business Communication Quality • Pearce/Barker
151
Flesch, R. (1949). The art of readable writing. New York: Harper and Brothers.
Gilsdorf, J. W. (1987). More "process": Aluxury? an illusion? a millstone? The
Journal of Business Communication, 24(1), 27-28.
Haas, C. (1989). How the writing medium shapes the writing process: Effects
of word processing on planning. Research in the Teaching of English, 23(2),
181-207.
Hawisher, G. (1986). Studies in word processing. Computers and Composition,
4,6-31.
Hawisher, G. (1987). The effects of word processing on the revision strategies
of college freshmen. Research in the Teaching of English, 21(2), 145-159.
Herrmann, A. W. (1987, December). Research into writing and computers:
Viewing the gestalt. Paper presented at the Modem Language Association
Conference, San Francisco, CA.
Hooper, S. C. (1987). Using word processors in high school and college writing
instruction: A critical review ofcurrent literature. Tbwson, MD: Calvert Hall
College Prep High School. (ERIC Document Reproduction Service No. ED
286 203)
Kurth, R. J. (1987, April). Word processing and composition revision strategies.
Papier presented at the Annual Meeting of the American Educational
Research Association, Washington, DC.
Lutz, J. A. (1987). A study of professional and experienced writers revising
and editing at the computer and with pen and paper. Research in ttie
Teaching of English, 21(4), 398-421.
Maik, T. A. (1987). Word processing in the business writing classroom: Applications and reactions. The Bulletin of the Association for Business Communication, 50(4), 4-6.
McCallister, C, & Louth, R. (1988). The effect of word processing on the quality
of basic writers' revisions. Research in the Teaching of English, 22(4),
417-427.
McCuthen, D., Hull, G. A , & Smith, W. L. (1987). Editing strategies and error
correction in basic writing. ]^itten Communication, 4(2), 139-154.
Munter, M. (1986). Using the computer in business communication courses.
The Journal of Business Communication, 23(1), 31-42.
Sterkel, K., Johnson, M., & Sjogren, D. (1986). Textual analysis with computers
to improve the writing skills of business communication students. The
Journal of Business Comm.unication, 23(1), 43-61.
Tbne, B., & Winchester, D. (1988). Computer-assisted writing instruction.
ERIC DIGEST: Clearinghouse on Reading and Communication Skills
(Report No. 2). (ERIC Document Reproduction Service No. ED 293 130)
—Accepted by NLR, 3/26/90
Final version received 5/21/90