Verbal description benefits for faces when description conditions are

This article was downloaded by: [Royal Holloway, University of London]
On: 14 March 2013, At: 07:38
Publisher: Routledge
Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office:
Mortimer House, 37-41 Mortimer Street, London W1T 3JH, UK
The Quarterly Journal of Experimental
Psychology
Publication details, including instructions for authors and subscription
information:
http://www.tandfonline.com/loi/pqje20
Verbal description benefits for faces when
description conditions are unknown a priori
a
a
a
a
Todd C. Jones , Ruth Armstrong , Allanah Casey , Rebecca A. Burson &
Amina Memon
b
a
School of Psychology, Victoria University of Wellington, Wellington, New
Zealand
b
Department of Psychology, Royal Holloway University of London, London,
UK
Accepted author version posted online: 07 Feb 2013.Version of record first
published: 12 Mar 2013.
To cite this article: Todd C. Jones , Ruth Armstrong , Allanah Casey , Rebecca A. Burson & Amina Memon
(2013): Verbal description benefits for faces when description conditions are unknown a priori, The Quarterly
Journal of Experimental Psychology, DOI:10.1080/17470218.2013.771688
To link to this article: http://dx.doi.org/10.1080/17470218.2013.771688
PLEASE SCROLL DOWN FOR ARTICLE
Full terms and conditions of use: http://www.tandfonline.com/page/terms-and-conditions
This article may be used for research, teaching, and private study purposes. Any substantial
or systematic reproduction, redistribution, reselling, loan, sub-licensing, systematic supply, or
distribution in any form to anyone is expressly forbidden.
The publisher does not give any warranty express or implied or make any representation that the
contents will be complete or accurate or up to date. The accuracy of any instructions, formulae,
and drug doses should be independently verified with primary sources. The publisher shall not
be liable for any loss, actions, claims, proceedings, demand, or costs or damages whatsoever or
howsoever caused arising directly or indirectly in connection with or arising out of the use of this
material.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013
http://dx.doi.org/10.1080/17470218.2013.771688
Verbal description benefits for faces when description
conditions are unknown a priori
Todd C. Jones1, Ruth Armstrong1, Allanah Casey1, Rebecca A. Burson1,
and Amina Memon2
1
School of Psychology, Victoria University of Wellington, Wellington, New Zealand
Department of Psychology, Royal Holloway University of London, London, UK
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
2
Some prior research has shown a benefit for describing nonverbal study stimuli, particularly faces, on a
later recognition test relative to a control (no description) condition. In such studies, participants have
known a priori whether a stimulus will need to be described, meaning that encoding differences other
than the description could account for the effect. In Experiment 1, a description benefit was obtained
for faces that could not be attributed to encoding differences. A direct manipulation of description duration, thus allowing more time to produce descriptors, did not influence the description effect. In
Experiment 2, visual rehearsal instructions (without any verbal descriptions) failed to produce a rehearsal benefit. The experiments provide direct evidence against an account where the description or rehearsal enhances the featural information of nonverbal representations. For the present results, a benefit
stemming from the encoding and retrieval of descriptors appears to be an attractive theoretical alternative over one that posits an enhancement or alteration of featural or holistic information.
Keywords: Recognition memory; Description benefit; Verbal facilitation; Episodic face recognition;
Conjunction errors.
Researchers have been interested in the benefit of
additional verbal information on memory for nonverbal materials (e.g., Bartlett, Till, & Levy, 1980;
Daniel, 1972; Kurtz & Hovland, 1953), and much
of this research has specifically targeted face stimuli
(e.g., Brown & Lloyd-Jones, 2005, 2006; Kerr &
Winograd, 1982; McKelvie, 1976; Nakabayashi,
Lloyd-Jones, Butcher, & Liu, 2012; Winograd,
1981; also see Lloyd-Jones, Brandimonte, &
Bäuml, 2008; Meissner, Sporer, & Susa, 2008;
Sporer, 1991). The present work was designed to
investigate the benefit of generating descriptions
to faces presented during a study phase on a later
face recognition memory test. In some respects,
our method closely followed procedures from
recent studies that have obtained benefits for generating written descriptions (Brown, Gehrke, &
Lloyd-Jones, 2010; Brown & Lloyd-Jones, 2005,
2006; Nakabayashi, Lloyd-Jones, et al., 2012; also
see Bartlett et al., 1980). However, our methods
departed from those methods in a couple of important ways. In prior research, the manipulation of
descriptions has been done either between groups
or within participants, but in separate blocks, and
participants have known before the presentation
of a face whether they would need to describe it
Correspondence should be addressed to Todd C. Jones, School of Psychology, Victoria University of Wellington, Box 600,
Wellington, New Zealand. E-mail:[email protected]
We thank James C. Bartlett for providing helpful comments on this version of the manuscript, Kazuyo Nakabayashi for participating in a good discussion about her work and the topic, and Siegfried L. Sporer and anonymous reviewers for giving helpful comments
on an earlier version of the manuscript.
# 2013 The Experimental Psychology Society
1
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
(e.g., Brown et al., 2010; Brown & Lloyd-Jones,
2005, 2006; Nakabayashi, Lloyd-Jones, et al.,
2012; Wickham & Lander, 2008; also see Sporer,
1991). In contrast, our description manipulation
was accomplished within participants on a trialby-trial basis where participants did not know
whether they would be required to generate a
description or perform a filler task until after each
face had been presented.
The present method removes potential confounds, such as differences in encoding of the
faces themselves or differences in the timing of
descriptions and filler tasks—in previous studies a
description could be mentally generated, at least
in part, while the faces were present. We also
used a conjunction paradigm (e.g., Reinitz,
Lammers, & Cochran, 1992; Underwood &
Zimmerman, 1973), which has been utilized to
test theories of recognition memory (e.g., Bartlett,
Searcy, & Abdi, 2003; Edmonds, Glisky, Bartlett,
& Rapscak, 2012; Jones & Bartlett, 2009; Jones
& Jacoby, 2001; Kroll, Knight, Metcalfe, Wolf, &
Tulving, 1996; McKone & Peh, 2006; Reinitz
et al., 1992; Underwood & Zimmerman, 1973).
These methods offer further advantages, which
we describe later.
Work on episodic face recognition has been
pursued from an applied interest in eyewitness situations and from a theoretical interest in the mechanisms that underlie performance (e.g., Shapiro &
Penrod, 1986). Much of the work on the topic of
descriptions for faces has focused on the difference
between a description condition and a nodescription condition (i.e., a description effect)
but has given less attention to how participants
manage to do the recognition memory task overall
(cf. Brown & Lloyd-Jones, 2005; Lloyd-Jones
et al., 2008; though see Clare & Lewandowsky,
2004; Sauerland, Holub, & Sporer, 2008). In contrast, our starting point is to acknowledge that the
memory task is a recognition memory task, and
we approach the topic from a view that emphasizes
recognition memory processes and how the recognition memory decision is impacted by having provided an earlier description. In the remainder of
this introduction, we devote subsections to cover
research and theory related to the present work.
2
Face processing
To briefly summarize, basic perceptual information
such as, but not limited to, luminance/reflectance
(e.g., Longmore, Liu, & Young, 2008; O’Toole,
Vetter, & Blanz, 1999), shape (O’Toole et al.,
1999) and pigmentation (O’Toole et al., 1999)
has been shown to be important for episodic face
recognition (for brief reviews, see O’Toole, Jiang,
Roark, & Abdi, 2006; Sinha, Balas, Ostrovsky, &
Russell, 2006). For unfamiliar faces, a change in
the pose has been shown to decrease recognition
memory performance (e.g., Bruce, 1982; Davies
& Milne, 1982; Longmore et al., 2008; O’Toole,
Edelman, & Bülthoff, 1998; O’Toole et al., 1999;
also see Sporer & Horry, 2011a). Although presenting the same face stimulus at study and test
has had its detractors where there is more of an
applied interest at stake (e.g., Bruce, 1982; Davies
& Milne, 1982), such a condition represents a
type of ideal situation and is quite useful for theoretical work (e.g., Bartlett et al., 2003; Busey,
Tunnicliff, Loftus, & Loftus, 2000; Edmonds
et al., 2012; Maurer, Le Grand, & Mondloch,
2002; McKone & Peh, 2006; O’Toole et al.,
1998, 1999; Tanaka & Farah, 1993; and see the
interesting theoretical perspective of Longmore
et al., 2008, and comments by Nakabayashi,
Burton, Brandimonte, & Lloyd-Jones, 2012).
Beyond more basic information, the type of
information that is extracted from a face stimulus
may reflect individual features (Mondloch, Le
Grand, & Maurer, 2002) or different types of configural information: the general, appropriate position of features (e.g., eyes above nose, which is
above the mouth; called first-order relations),
local spatial configurations of features (i.e., distances between features; called second-order
relations), or holistic information (i.e., the features
bound into a whole; Maurer et al., 2002; also see
Searcy & Bartlett, 1996). There are findings that
support the use of these different types of information, and the extraction of one type of information does not preclude the extraction of
another type.
Eye-tracking data show that the internal facial
features (eyes, eyebrows, nose, and mouth) are
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
prioritized (e.g., Nakabayashi, Lloyd-Jones, et al.,
2012; Walker-Smith, Gale, & Findlay, 1977),
but the external features (hair, jaw, chin, ears, and
general head shape) certainly can contribute to recognition and identification memory performance.
For example, early research on episodic recognition
of unfamiliar faces showed that, after presenting
whole faces (with hair) at study, presenting only
the inner features or only the outer features at test
can produce performance clearly above chance
(though not as good as presenting the whole face
again), without an advantage for inner features or
outer features (Ellis, Shepherd, & Davies, 1979;
also see Sporer & Horry, 2011b). In contrast, findings from a matching task show that viewpoint
changes (full view to three-quarter view) and orientation changes (e.g., upright to inverted) can impact
the ability to match internal features more than
external features over very short intervals
(Meinhardt-Injac, Meinhardt, & Schwaninger,
2009). For known faces, Sinha and Poggio (1996,
2002) provided a powerful demonstration by conjoining the U.S. President’s inner features (either
W. Clinton or G. Bush Jr.) with the U.S. Vice
President’s outer features (A. Gore or D. Cheney,
respectively); the images of the President and the
President/Vice President conjunction are presented
side by side. Naïve observers readily misidentify the
conjunction faces as the (now former) Vice
Presidents even though the internal features are
not theirs (and appear adjacent to the image of
the President).
Conjunction faces are thought to leave featural
information and some local configural information
intact while providing a study–test mismatch in
holistic information (e.g., Bartlett et al., 2003;
Reinitz, Morrissey, & Demb, 1994; also see
Cabeza & Kato, 2000; McKone & Peh, 2006).
We utilized the conjunction procedure for naturalistic, previously unknown faces (e.g., Bartlett
et al., 2003; Jones, Bartlett, & Wade, 2006) and followed the approach of Bartlett and colleagues
(2003), where the focus is on featural and holistic
information.1 (For a neuroscience finding in
support of this position, see Harris & Aguirre,
2008.)
Recognition memory theory
Theories on recognition memory can be categorized as single-process (familiarity only) theories
(Banks, 1970; T. E. Parks, 1966) or dual-process
(familiarity and recollection) theories (for a
review, see Yonelinas, 2002; also see James,
1890). In a familiarity only approach, test items
are evaluated in terms of familiarity strength, and
if the familiarity strength surpasses a criterion,
then an “old” response will be given (for some
mathematical models, see Clark & Gronlund,
1996, and for faces, a single-dimensional confidence–accuracy model, Busey et al., 2000; the
general abstract processing system—GAPS—
model, Bartlett et al., 2003; and an exemplarbased generalization model, Nosofsky, 1988). For
dual-process theories, familiarity is typically
described as automatic (or relatively automatic),
and recollection is described as slow and controlled.
In Jacoby’s (1991, 1999) view, conditions can be
devised where familiarity or recollection work independently but in concert to accept a test target as
“old” or where familiarity and recollection oppose
one another. In the opposition case, familiarity
pushes one toward incorrectly accepting a test lure
(as “old”), whereas recollection can potentially be
used to correctly reject a test lure (as “old”). To be
clear, recollection can potentially be used to
accept targets or reject lures (Jacoby, 1991), sometimes referred to as recall-to-accept and recallto-reject, respectively (e.g., Rotello, Macmillan, &
Van Tassel, 2000). There is currently agreement
1
Two other procedures that have been used to get at issues regarding the use of featural or configural information include the
manipulation of face orientation (i.e., upright vs. inverted, e.g., Carey & Diamond, 1977; Searcy & Bartlett, 1996; Valentine,
1988; Yin, 1969) and creating mathematical averages of faces (morphs; e.g., Busey & Tunnicliff, 1999; Cabeza & Kato, 2000; also
see Galton, 1879). Inverting faces is thought to decrease or eliminate the extraction of holistic information (e.g., Diamond &
Carey, 1986; Friere, Lee, & Symons, 2000; Searcy & Bartlett, 1996); creating morphs is thought to achieve a high level of study–
test similarity while altering featural information, second-order configural information, and holistic information configuration (e.g.,
Busey & Tunnicliff, 1999; Cabeza & Kato, 2000).
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
3
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
that familiarity runs on a continuum that conforms
to a signal detection-like manner but disagreement
about whether recollection is better described as
all-or-none (Yonelinas, 1994; C. M. Parks &
Yonelinas, 2007) or continuous (Ingram, Mickes,
& Wixted, 2012; Mickes, Wais, & Wixted,
2009; Wixted, 2007).
Although the basis of familiarity—that is, the
type of information that is tapped—may vary
depending on the materials (for verbal materials,
see Jones, Brown, & Atchley, 2007; for faces, see
Jones & Bartlett, 2009; Jones et al., 2006), Jones
and Jacoby (2001) proposed that, regardless of
the type of material, conjunction errors for episodic recognition reflect the influence of familiarity
in the absence of (successful) recollection for
parent items. Work on the conjunction paradigm
with compound word materials (study: blackmail,
jailbird; test lure: blackbird) has provided support
for dual-process theories by showing that, in
some cases, participants can correctly reject compound word conjunction lures via recollection
(e.g., Arndt & Jones, 2008; Jones & Atchley,
2006; Jones & Jacoby 2001; Lampinen,
Odegard, & Neuschatz, 2004). Bartlett and colleagues (2003) incorporated Jones and Jacoby’s
(2001) theory on conjunction performance in
their research with face conjunction materials
and suggested that face conjunctions for novel
(unknown) faces might not be able to be rejected
via recollection. They concluded that there are
two routes to face recognition. One route capitalizes on featural information (or processing) and
leads to perceived familiarity, which accounts for
incorrect recognition of conjunction lures as “old”
above a baseline. The other route is holistic,
which is indicated by the ability to discriminate
targets from conjunction faces. They also concluded that the subjective experience of recollection is strongly influenced by holistic processing,
with holistic processing providing considerable
support for such an experience.
Additional work with conjunction faces led
Jones and his colleagues (Jones & Bartlett, 2009;
Jones et al., 2006) to the conclusion that a singleprocess (familiarity only) model can account for
recognition memory performance of novel faces
(also see Busey & Tunnicliff, 1999; Busey et al.,
2000). Jones and Bartlett (2009) used conditions
(several study repetitions or a very short retention
interval) that have produced evidence for recollection-based rejections of compound word conjunctions. However, they found no evidence that
recollection could be used to reject a conjunction
face lure. In fact, the highest conjunction error
rate in their experiments was for a condition
where a conjunction face lure immediately followed
a parent face in a continuous recognition memory
procedure (i.e., no trials intervened between the
second of the two parent faces and the conjunction
lure; e.g., Shepard & Tehgtsoonian, 1961).2
Description effects
The research on the influence of descriptions for
faces is mixed. Sometimes, a cost has been reported
(called verbal overshadowing, e.g., Chin &
Schooler, 2008; Meissner & Brigham, 2001;
Schooler & Engstler-Schooler, 1990), and we
save comments on description costs for the
General Discussion. Some recent studies have
obtained verbal benefits for face stimuli (Brown
et al., 2010; Brown & Lloyd-Jones, 2005, 2006;
Nakabayashi, Lloyd-Jones, et al., 2012; Wickham
& Lander, 2008; also see Nakabayashi, Burton,
et al., 2012). Such benefits are generally consistent
with prior research showing the benefits of accurate
labels of objects (Kurtz & Hovland, 1953), descriptions of picture scenes (Bartlett et al., 1980; also see
Loftus & Kallman, 1979), labels of faces
(McKelvie, 1976), a “deeper” level of processing
for faces (Bloom & Mudd, 1991; Bower &
Karlin, 1974; also see Sporer, 1991; Wells, 1985),
and elaborative processing for faces (Kerr &
Winograd, 1982; Winograd, 1981) relative to
control conditions. In several recent studies, multiple faces were shown in a study phase, and, for
description conditions, a description immediately
2
Although recollection might be used correctly to accept target faces as “old”, or perhaps incorrectly to exacerbate conjunction
errors (via recollection of a feature or locally configural information), their data provided no compelling basis for such arguments.
4
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
followed each face (Brown & Lloyd-Jones, 2005,
2006; Nakabayashi, Lloyd-Jones, et al., 2012;
Wickham & Lander, 2008; also see Bartlett et al.,
1980; Kerr & Winograd, 1982; Sporer, 1991;
Winograd, 1981). The immediacy and short,
unforced nature of the descriptions arguably
produce accurate descriptions, which should
provide an opportunity to aid later performance
(Brown et al., 2010). 3
Description benefits have been obtained for
description instructions that focused on features
only (eyes, hair, nose, etc.; Brown & Lloyd-Jones,
2005; though see Wickham & Lander, 2008),
general characterizations only (age, occupation,
resemblance to another person; referred to as configural/holistic descriptors; Brown & Lloyd-Jones,
2005; also see Wickham & Lander, 2008), and
open (free) instructions (Brown et al., 2010;
Brown & Lloyd-Jones, 2005, 2006; Nakabayashi,
Lloyd-Jones, et al., 2012). Sporer (1989) suggested
that only distinctive faces might benefit from a
description, though Brown and Lloyd-Jones
(2006) found a description benefit regardless of
prerated distinctiveness. Following the study of
faces exclusively in an upright orientation, participants have shown description benefits for upright
test faces, as well as inverted test faces, and for
famous faces and nonfamous faces (Brown et al.,
2010). Similarly, studying upright faces produced
a description benefit when those faces were presented intact or scrambled into four vertical sections
(top of head, eyes, nose, mouth–chin–jaw) on the
test (Nakabayashi, Lloyd-Jones, et al., 2012).
Finally, description benefits have been obtained
for same-race faces and other-race faces
(Nakabayashi, Lloyd-Jones, et al., 2012). We note
that in all of these studies, the stimuli have
changed from study to test, going from either a
full-frontal pose to a three-quarter view on the
test (Brown et al., 2010; Brown & Lloyd-Jones,
2005, 2006; Nakabayashi, Lloyd-Jones, et al.,
2012) or from a still image from a video clip to a
different photo (though in the same pose;
Wickham & Lander, 2008). Thus, any description
effect has been obtained when the conditions would
make it more challenging to match internal features
from study to test (Meinhardt-Injac et al., 2009;
also see Nakabayashi, Burton, et al., 2012). (We
are not opposed to changes in appearance from
study to test, but we do think that study-to-test
changes, including pose, are best accomplished
and understood through direct manipulation, e.g.,
Longmore et al., 2008.)
Across their studies, Brown and colleagues
(Brown et al., 2010; Brown & Lloyd-Jones, 2005,
2006) have suggested a number of potential
reasons for a verbal description benefit: (a) the
enhancement of featural information by the
description, particularly from a perspective that
describing more features is better (e.g., Bloom &
Mudd, 1991; Bower & Karlin, 1974); (b) the
enhancement of holistic information (e.g., Wells
& Hryciw, 1984) via alteration of a holistic representation; (c) enhancement of holistic information via an alteration to encoding processes (a
shift to more holistic processing, Brown et al.,
2010); (d) the benefit of semantic information, in
terms of richer associations (Anderson & Reder,
1979; Bruce & Young, 1986; also see Klatzky,
Martin, & Kane, 1982); (e) the alteration of retrieval for the nonverbal stimulus, where the description acts as an action tag, guiding processing of
the test face (Brown & Lloyd-Jones, 2005, 2006);
(f) the involvement of recollection and, perhaps,
familiarity (Brown & Lloyd-Jones, 2006).
Brown and colleagues (Brown et al., 2010;
Brown & Lloyd-Jones, 2005, 2006) have consistently favoured the idea that one locus of the effect
is semantic, along the lines of a richer set of associations. Other aspects of their conclusions have been
less consistent. For example, Brown and LloydJones (2006) stated, “… we argue that the
3
We examined the list of studies covered by Meissner et al. (2008) for their meta-analyses. In support of Brown and colleagues’
(2010) contention, studies that used multiple study faces with a description after each face produced either a benefit or a null effect
(sometimes with a nonsignificant advantage of a description condition; e.g., Chance & Goldstein, 1976; Wogalter, 1991). We also
note that the vast majority of studies in their meta-analyses used only a single study stimulus, which tends to be a procedure more
common in the eyewitness/applied literature.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
5
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
facilitative effects of verbalization arise from an
alteration to the memory representation of a particular face rather than some generalized strategy”
(p. 283). (Additional comments appear to indicate
that they mean an alteration of a holistic representation.) In contrast, Brown et al. (2010) make no
note of their earlier position but state, “… we
propose that the process of generating a description
encourages a more global visual encoding strategy
that is particularly useful for visually differentiating
between a large number of highly visually similar
stimuli” (p. 62). Nakabayashi, Lloyd-Jones, et al.
(2012) concluded that describing faces changes
the configural encoding of them, specifically
through more sampling of the inner features.
Although Brown and Lloyd-Jones (2005, 2006)
appear to express some preference for the action
tag hypothesis (also see Nakabayashi, Burton,
et al., 2012), this idea is not mentioned in Brown
et al. (2010). These latter papers do not mention
the retrieval processes, recollection and familiarity.
EXPERIMENT 1
Our interest was to employ methods that would
allow us to pinpoint the basis for a description
benefit. In prior work, the description instructions
have been manipulated between groups (Brown
et al., 2010; Brown & Lloyd-Jones, 2005, 2006;
Wickham & Lander, 2008; also see Sporer, 1991)
or between blocks of trials in a within-participants
design (Brown & Lloyd-Jones, 2005; Nakabayashi,
Lloyd-Jones, et al., 2012). In these prior cases, participants knew before a face was presented whether
it would be described. Thus, the differences
between the description and no-description conditions included not just the descriptions themselves but potentially how participants encoded a
face during its presentation (Brown et al., 2010;
Nakabayashi, Lloyd-Jones, et al., 2012; also see
Loftus & Kallman, 1979). In addition, the onset
of the description and filler task was likely to be
confounded because participants could mentally
generate a description of a face, at least in part,
while a face was present instead of waiting until it
disappeared from view. This timing issue is critical
6
for separating whether generated information influences the encoding of a nonverbal stimulus or
whether it alters aspects of a nonverbal representation (ex post facto; e.g., Brown & Lloyd-Jones,
2006). Our entirely within-participants design
with conditions mixed on a trial-by-trial basis
meant that treatment of the faces themselves and
the timing of the descriptions and filler task
should be constant across conditions. Any benefit
could then be attributed to alteration of nonverbal
representations or generation and retrieval of
descriptors instead of differences in encoding the
nonverbal stimulus.
Description effects have been obtained for
instructions that emphasize facial features and
instructions that have been characterized as emphasizing holistic information. Thus, one could argue,
first, that a description enhances aspects of the nonverbal (face) representation and, second, that an
enhancement of either featural or holistic information can lead to a benefit. Stated explicitly, this
approach means that some form of nonverbal representation is altered. The form of the nonverbal
representation that gets altered depends on one’s
theoretical viewpoint. To elaborate, one could
posit that separate nonverbal representations for
featural and holistic information are altered.
Alternatively, one could posit that information
within a single, holistic representation is altered.
For example, McKone and Yovel (2009) proposed
that a holistic representation contains all types of
information. In this case, alteration of featural
information would, by necessity, alter the holistic
representation. In contrast to these accounts that
propose an alteration of a nonverbal representation,
one could focus on the generation and retrieval of
descriptors. In this alternative, part or all of a
verbal description may be retrieved during the
test, helping to verify a test face as “old”.
Earlier in this introduction, we described the
conjunction paradigm and the idea that errors for
conjunction lures reflect the influence of familiarity
engendered by featural information. Featural information also provides one basis for recognizing a
target face as “old”. If generating descriptions
enhances featural information (only), then generating descriptions should increase the conjunction
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
error rate and hit rate relative to no-description
control conditions. In this case, old–conjunction
discrimination should not change across description conditions. We also indicated that the ability
to discriminate old (target) faces from conjunction
(lure) faces can be taken to reflect the influence of
holistic information. If holistic information (only)
is enhanced by the description, then generating
descriptions should increase the hit rate relative to
a no-description control but have no influence on
the conjunction error rate. If both featural and configural information are enhanced, either as separate
representations or as a single, holistic representation containing all types of information, then
generating a description should increase conjunction error rates and hit rates relative to a nodescription control, but the hit rate should increase
more than the conjunction error rate.
In contrast to these ideas concerning alteration
of nonverbal representations, we were specifically
interested in the alternative noted above that
focuses on generating and retrieving the description. As a starting point, one could argue that the
face stimuli are processed quickly and provide the
foundation for the recognition decision. For
example, Jones and Bartlett (2009) concluded that
recognition of faces could be accomplished via a
fast, single (familiarity) process (also see Busey
et al., 2000). Similarly, Hsiao and Cottrell (2008)
showed that participants could successfully recognize novel faces above chance in one fixation and
that, with two fixations, participants recognized
faces as well as they did with three or four fixations.
The time duration per fixation in their study was
short (less than 300 ms, on average). From our
point of view, descriptions from the study phase
simply represent an additional bit of information
that can contribute to the recognition decision
(positively or negatively, depending on the
accuracy).
Bartlett and colleagues (2003) proposed that
holistic information is particularly important with
regard to the subjective experience of remembering
and, more importantly here, to its ability to act as a
retrieval cue for other information. In its strongest
form, their line of thinking means that only faces
that represent a holistic match from study to test
should cue additional information—in the present
case the verbal description. Thus, their suggestion
means that a description benefit should occur for
old (target) faces but not conjunction faces. To be
clear, the idea here is not that the description influences retrieval of the nonverbal information (e.g.,
Brown & Lloyd-Jones, 2005, 2006). Instead, the
proposal is that the nonverbal information influences retrieval of the verbal information. Bartlett
and colleagues’ (2003) proposal has roots in
Tulving’s (1983) principle of encoding specificity
and his idea of synergistic ecphory. In short, the
stronger the match between a study face stimulus
and the test face stimulus, the better retrieval cue
the target face should be for the verbal information
(also see Shapiro & Penrod, 1986).
In the experiment, we used open description
instructions, and we expected the bulk of participants’ descriptions to focus on features (e.g.,
Brown et al., 2010; Brown and Lloyd-Jones,
2005). A minor aim was to assess whether the quantity of feature descriptors boosts a description benefit
(a more-is-better hypothesis; see Bloom & Mudd’s,
1991, and Bower & Karlin’s, 1974, conclusions for
level-of-processing effects for faces).4 To address
this aim, we directly manipulated the amount of
time to write a description (15 s vs. 60 s), with the
reasonable assumption that a longer period of time
would result in more descriptors. Basically, according to these earlier ideas, a greater number of featural
descriptors, assuming that those descriptors are
accurate, should further increase the hit rate and
conjunction error rate. Therefore, this manipulation
also provided a particularly strong test of the featural
enhancement idea.
Method
Participants
Victoria University of Wellington undergraduates
(N = 56) participated in return for credit toward a
4
A different perspective was given by Winograd (1981). He suggested that a greater number of descriptors is beneficial because it
increases the likelihood of describing one distinctive feature.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
7
JONES ET AL.
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
class research requirement. The criteria for
inclusion in the data set were that the instructions
for the experiment were followed (e.g., verbal
descriptions were given only when the instructions
indicated to do so, filler maths problems were
attempted to be solved) and that overall recognition
memory performance was above chance (i.e., overall
hit rate false-alarm rate for entirely new faces).
Apparatus
All of the stimuli—greyscale bitmaps of male,
Caucasian faces without facial hair or spectacles—
came from Jones et al. (2006). There were 54 original faces (48 critical faces and 6 extra faces for
primacy and recency buffers), and 48 construction
faces created by swapping inner features (eyes, eyebrows, nose, and mouth) and outer features (hair,
forehead, ears, cheeks, jaw, and chin) across pairs
of faces. For example, those construction faces
were created by taking the inner features of Face
A and putting them within the external features
of Face B to create Face AB and by taking the
inner features of Face B and putting them within
the external features of Face A to create Face BA.
(By construction, we only mean that those faces
were created digitally.) Four faces, two originals
and two constructions, made up a face set. Sets
were assigned to eight lists matched for distinctiveness. (The distinctiveness ratings were based on
two sets of 60 subjects. All mean distinctiveness
ratings were less than 6.0 on a 0–9.0 scale, i.e.,
not very distinctive.) The experiment was run
using E-prime software (Schneider, Eschman, &
Zuccolotto, 2002a, 2002b) on PC computers with
15′′ CRT monitors. Separate booklets were provided for writing descriptions and solving maths
problems.
Design
The design was accomplished entirely within participants. There were two within-participant
factors: verbal description condition (15 s, 60 s,
no description) and test condition (old, conjunction, new). After a correction for the false-alarm
rate to new faces, the design simplified to a 3
(description condition) × 2 (test condition) design.
8
Procedure
Conjunction stimuli are test stimuli that contain
parts of faces (parents) from the study phase, and
there are different ways of making this happen.
We set up the experiment so that constructions
(Face AB and Face BA) and originals (Face C
and Face D) were presented during the study
phase, and only originals (Face A, Face B, Face
C, Face D) were presented on the test. Thus, all
of the faces on the test were yearbook originals,
but the test faces represented different conditions
based on what was presented during the study
phase. In this example, Face A and Face B would
be conjunction lures.
There were three study–test blocks, with 14
faces (1 primacy buffer, 12 faces for the later test,
and 1 recency buffer) presented in each study
phase and 16 faces (6 old, 6 conjunction, and 4
new) presented on each test. The 12 critical faces
in each study phase were 6 originals and 6 constructions; the faces in each test were always originals.
On each test, there were two old and two conjunction faces from each of the three description conditions. For the conjunction conditions, two
constructed faces presented in the study phase
had their inner features swapped (back to the originals). Two lists were used for the condition
(instead of one) to help keep performance off the
floor. The eight lists were rotated through the conditions: three lists for the old conditions, three lists
for the conjunction conditions, and two lists for the
new condition for a given participant (see Table 1).
Participants were run in small groups (1–5) at individual computer stations (approximately 50 cm
from a monitor). For the study sets, the faces
were presented one at a time in the centre of the
screen for 2 s (the same duration as that used by
Brown & Lloyd-Jones, 2005, 2006). Each face
was preceded by focal point (+) for 5 s. The
study instructions explained that there would be a
60-s period following each face. For some faces, a
verbal description would be required for some
amount of time (15 or 60 s). For the 15-s condition, maths problems were to be solved in the
remainder of the 60 s. For some faces, maths problems would be solved for the full 60 s. For the
15-s description condition, after 15 s the screen
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
VERBAL DESCRIPTION BENEFIT
Table 1. Schematic of assignment of eight lists to conditions across eight participants
Old
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
Participant
P1
P2
P3
P4
P5
P6
P7
P8
Conjunction
15 s
60 s
None
15 s
60 s
None
A
B
C
D
E
F
G
H
B
C
D
E
F
G
H
A
C
D
E
F
G
H
A
B
D
E
F
G
H
A
B
C
E
F
G
H
A
B
C
D
F
G
H
A
B
C
D
E
New
G
H
A
B
C
D
E
F
H
A
B
C
D
E
F
G
Note: Lists: A–H. Participants: P1–P8. Duration of description period during study: 15 s, 60 s, and none.
flickered between a black-and-white screen 5 times
in 1 s to alert the subject to shift to the maths problems. For all three conditions, at the end of 60 s
for descriptions/maths problems, the screen flickered between a black-and-white screen 5 times in
1 second to signal the subject to stop the current
task and to prepare for the next trial. To be clear,
subjects did not know whether a face would be
described until after its presentation was complete.
The order of the study faces was mixed with
regard to description conditions and (later) test
conditions (old or conjunction); buffer faces were
not described. Parent faces for later conjunction
lures were separated by at least one study trial.
The study and test procedures were explained so
that subjects could continue through the experiment without interruption. For the descriptions,
subjects were told, “Please be as complete in your
description as possible so that another person
seeing only your description could get as accurate
an idea as possible of what the face is like” (e.g.,
Brown & Lloyd-Jones, 2005, p. 1444) and “to do
the best you can in the time allotted” (also see
Bartlett et al., 1980). We were not interested in
creating a situation where subjects might be more
likely to write down inaccurate descriptors.
Participants knew how much time was available
to write down a description. However, we did not
tell them that they had to use the full period, and
we did not enforce usage of the full period. The
yes–no (i.e., old–new) format, the types of test
items (old, conjunction, new), and the correct
responses for the test items were explained.
Recognition responses were made using the F and
J keys, with a 10-s response time (RT) limit per
trial. When a response was entered, the program
continued to the next test trial.
Results and discussion
Before turning to the key data, we note that the
descriptions met our expectations. Our first key
point regarding the descriptions is that the longer
description period increased the number of featural
descriptions. A simple glance at the descriptions
clearly demonstrated that the longer description
period resulted in longer descriptions. To verify
this observation in more concrete terms, we
scored a sample of 16 participants’ descriptions,
which included 2 participants from each counterbalance. The mean number of feature descriptions for
the 60-s condition (5.31, SD = 1.66) was about
double that for the 15-s condition (2.65, SD =
0.64), t(15) = 8.91, SE = 0.30, and the mean
total number of descriptions for the 60-s condition
(6.47, SD = 1.83) was also about double that for
the 15-s condition (3.31, SD = 0.67), t(15) =
10.02, SE = 0.33.5 (For all statistical analyses,
Because of our use of conjunction stimuli, head shape was considered a “feature”. Excluding head shape from the feature scores did
not change our conclusions. The first and fourth author coded the descriptions independently; the authors produced similar coded
scores, and the first author’s scores were used for analyses.
5
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
9
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
alpha was set at .05.) Our second point for the
descriptions is that we would be hard pressed to
identify very many (if any) of the descriptors for
features (e.g., bushy eyebrows, intense eyes, etc.)
as inaccurate. Finally, we note that the proportion
of descriptors for features was fairly high and
similar for the two conditions (M = .83,
SD = .13, and M = .81, SD = .10, for the 15-s
condition and 60-s condition, respectively).
The use of the conjunction faces in the experiment provided the strongest leverage for the tests
of the featural and holistic enhancement hypotheses. The “old” response rates showed the typical
pattern: old face (M = .71, SD = .11) conjunction
face (M = .48, SD = .13) new face (M = .15,
SD = .13). A repeated measures analysis of variance
(ANOVA) with a repeated contrast showed that
the false-alarm rate for conjunction faces was significantly higher than that for new faces, F(1,
55) = 339.46, MSE = .018, η2p = .86, and that the
hit rate for old faces was significantly higher than
the false-alarm rate for conjunction faces, F(1,
55) = 148.83, MSE = .020, η2p = .73. These
results can be interpreted as reflecting the influence
of featural information (only) in the conjunction
condition but reflecting the influence of featural
and holistic information for the old condition.
Given this platform, one can then assess the
success of the featural and holistic enhancement
hypotheses.
As can be seen in Figure 1, verbal descriptions
increased hit rates but did not influence conjunction error rates. A 3 × (verbal description condition) × 2 (test condition: old, conjunction)
repeated measures ANOVA on the corrected probabilities shown in the figure produced a significant
main effect of verbal description, F(2, 55) = 4.12,
MSE = .027, η2p = .07, and a significant Verbal
Description × Test Condition interaction, F(2,
110) = 3.58, MSE = .037, η2p = .06. Our first
question was whether descriptions boosted performance in old and conjunction conditions relative
to the control condition, and our second question
was whether the 60-s description increased performance above that for the 15-s description
period. Given our specific questions, we ran separate repeated measures ANOVAs with Helmert
10
contrasts for the old and conjunction scores: The
no-description control was compared to the 15-s
and 60-s verbal description conditions (collapsed)
to assess the verbal description effect, and the
15-s condition was compared to the 60-s condition
to assess a quantity-of-features hypothesis. The
specific contrasts showed that, for old faces, there
was a verbal description benefit (no description,
M = .49, SD = .20, vs. description, M = .59,
SD = .18), F(1, 55) = 11.97, MSE = .054,
η2p = .18, but the 15-s and 60-s conditions were
not significantly different (15-s description,
M = .62, SD = .21, vs. 60-s description, M = .57,
SD = .23), F(1, 55) = 1.88, MSE = .067,
η2p = .03. Neither of the specific contrasts for conjunction scores was significant (no description,
M = .33, SD = .14, vs. description, M = .33,
SD = .21, Fs = 0.03; and 15-s description,
M = .33, SD = .20, vs. 60-s description, M = .33,
SD = .17, F = 0.01).
The key findings are as follows. First, a description benefit occurred for old faces. This effect
cannot be attributed to encoding differences of
the nonverbal stimuli themselves because the cue
to describe a face was given only after the face
was removed from view. Second, the manipulation
involving the quantity of descriptors (15-s vs. 60-s
description) provided no evidence for an additional
benefit with more descriptors. Third, there was no
evidence for a verbal benefit effect on conjunction
scores.
These findings provide strong evidence against
a feature enhancement hypothesis and a “more
features is better” hypothesis (e.g., Bloom &
Mudd, 1991; Winograd, 1978) when participants
do not know a priori whether a stimulus will be
described. The outcome is consistent with a holistic enhancement prediction where the holistic
representation is altered by the description. The
findings are also consistent with Bartlett and colleagues’ (2003) idea that holistic information can
cue retrieval of additional information. In this
latter hypothesis, the ability of a test probe to
cue additional information depends on the
study–test match in the nonverbal stimulus.
From this viewpoint, the holistic match, but not
the conjunction, cued retrieval of the description,
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
Figure 1. Results for Experiment 1. Mean corrected probability of an “old” response for old and conjunction conditions (i.e., with the new-face
false-alarm rate subtracted) across verbal description condition. Error bars reflect the standard error within a condition. Going from left to right
the means are .49 (SD = .20), .33 (SD = .21), .62 (SD = .21), .33 (SD = .20), .57 (SD = .23), and .33 (SD = .17). The corresponding
A′ scores are .83 (SD = .10), .76 (SD = .14), .88 (SD = .09), .77 (SD = .10), .86 (SD = .10), and .77 (SD = .11).
and this description provided a benefit in the recognition decision.
EXPERIMENT 2
A desirable situation would allow one to argue an
advantage for one of the viable explanations in
Experiment 1. To support the holistic alteration
account, one might be tempted to could focus on
nonfeatural descriptors, even though they were
relatively infrequent, with the presumption that
some descriptors, such as “looks older” or “rugby
player” alter holistic information but not featural
information (cf. Kerr & Winograd, 1982; also see
Brown & Lloyd-Jones, 2005). In this case, one
would expect no effect to occur for conditions
that use only featural descriptors. However,
Brown and Lloyd-Jones (2005) obtained an effect
for a feature (only) description condition. We
note that Wickham and Lander (2008) did not
obtain a benefit for feature-only descriptions, but
their description period was 2 minutes long and
enforced. A longer, enforced description period
has been shown to produce overshadowing
effects, presumably due to description inaccuracies
introduced by the participant (Meissner,
Brigham, & Kelley, 2001). Thus, the difference
may simply reflect that some of the descriptors concerning features in Wickham and Lander’s studies
were inaccurate.
Another avenue would be to suggest that
describing features or giving a general characterization leads to an enhancement of holistic information. In this case, a single holistic
representation might be altered (e.g., McKone &
Yovel, 2009). Because evidence against the featural
enhancement hypothesis was obtained in our first
experiment, a peculiar caveat of this idea is that
the holistic representation would have to be
altered without enhancement of featural information within that representation. This idea
seems puzzling. However, a potential solution to
this puzzle is that, during the description period,
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
11
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
participants hold in mind or retrieve a holistic
mental face image (from short-term/working
memory or long-term memory). This maintenance
or rehearsal could be the basis of the benefit instead
of what appears to be a benefit from the description.
We are not the first to report a version of this idea.
For example, Nakabayashi, Burton, et al. (2012)
suggested that Brown and colleagues’ (Brown
et al., 2010; Brown & Lloyd-Jones, 2005, 2006)
results reflected that their “paradigm focused on
verbalisation during the retrieval and rehearsal of
faces in memory” (p. 382). This idea is consistent
with some findings on face rehearsal (Read, 1979;
Read, Hammersley, Cross-Calvert, & McFadzen,
1989; also see Sporer, 1988, 1989; Wogalter,
1991), and it rests on the important premise that
a complex holistic representation can be held in
mind or retrieved.
Evidence from prior studies on image rehearsal
is mixed. In a rigorous set of experiments, Graefe
and Watkins (1980) instructed participants to
“‘imagine,’ ‘visualize,’ or otherwise ‘think about’”
(p. 157) a cued item from pairs of nonverbal
stimuli during encoding. Importantly, the cue was
provided after each stimulus pair was presented.
Overall, they obtained fairly small but consistent
rehearsal benefits for different types of nonverbal
stimuli, including faces, but they explicitly cautioned against interpretation of what occurred in
the covert rehearsals of participants, noting that
the benefit could be based on rehearsal of verbal
instead of nonverbal information. In contrast,
Shaffer and Shiffrin (1972) presented visual pictures of objects (primarily) during a study phase
and manipulated the stimulus duration and the
poststimulus interval (1, 2, or 4 s). Participants
were told “when a slide leaves the screen, think
about it and try to remember it during the period
before the next slide appears” (p. 293). They
obtained no difference in “old–new” recognition
memory performance for poststimulus (rehearsal)
periods of 1, 2, or 4 s and concluded that visual
rehearsal is ineffective, if it occurs at all, for relatively complex stimuli.
Evidence from Jones and Bartlett’s (2009) continuous recognition memory experiment with conjunction faces noted in the introduction fits with
12
Shaffer and Shiffrin’s (1972) conclusion, as well
as recent work that indicates limits in the resolution
of more complex stimuli in working memory (Awh,
Barton, & Vogel, 2007) or that suggests severe
limits for the number of complex stimuli that can
be held in visual working memory (Alvarez &
Cavanagh, 2004). In Jones and Bartlett’s
Experiments 2 and 3, if a mental image of a
parent face could be maintained in working
memory or retrieved from long-term memory on
the immediately following trial, then a conjunction
error could be avoided. There was no clear evidence
that participants were able to reject conjunction
lures by holding or retrieving an image of a
parent item from the immediately preceding trial.
(In fact, the conjunction error rate, as well as the
confidence rating for that error rate, was at its
peak on the trial immediately following a parent
face.)
In summary, there is the possibility that the
effect in the first experiment might have been
driven by visual rehearsal/retrieval of holistic face
representations, which could enhance holistic
information, but there is research that goes
against this possibility. We essentially conducted
the first experiment again but altered the study
instructions to encourage visual rehearsal of the
faces without any verbal description instructions.
If participants are able to hold or retrieve an
image of a whole face in mind and rehearse it
during the description period, then holistic information of the face should be enhanced (e.g.,
Read, 1979). If the finding in the first experiment
was due to enhancement of holistic information,
then the same pattern of results should be obtained.
In contrast, if there is no effect of rehearsal (or
image retrieval), then a role of the verbal description itself gains additional legitimacy. Finally,
although our concentration was on the possibility
of holistic enhancement, the other predictions
from Experiment 1 applied. If visual rehearsal
enhances featural information in the representation, then visual rehearsal should increase the
hit rate and the conjunction error rate.
Alternatively, if visual rehearsal enhances both featural and configural information, then rehearsal
should increase the hit rate and the conjunction
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
VERBAL DESCRIPTION BENEFIT
error rate, but the hit rate should be boosted more
than the conjunction error rate.
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
Method
Participants
Forty Victoria University of Wellington undergraduates participated and earned credit toward
an introductory psychology class. With the exception of the verbal descriptions, the same criteria as
those from Experiment 1 were used for inclusion
in the dataset.
Apparatus
The apparatus were the same as those in
Experiment 1.
Design
There were two within-subjects factors: visual
rehearsal condition (15 s, 60 s, no rehearsal) and
test condition (old, conjunction, new). After a correction for the false-alarm rate to new faces, the
design simplified to a 3 (rehearsal condition) × 2
(test condition) design.
Procedure
The procedure was the same as that in Experiment
1 with the following exceptions. Instead of giving
instructions to write verbal descriptions of the
faces when prompted to do so, we instructed participants “to picture the face in your mind with as
much detail as possible for 15/60 seconds”.
Results and discussion
As in the first experiment, the “old” response rates
followed the typical pattern: old face (M = .72,
SD = .12) conjunction face (M = .51, SD = .13)
new face (M = .17, SD = .14). Again, a repeated
measures ANOVA on the three conditions with
specific contrasts showed that the false-alarm
rate for conjunction faces was significantly
higher than that for new faces, F(1, 39) =
169.89, MSE = .027, η2p = .81, and that the hit
rate for old faces was significantly higher than
the false-alarm rate for conjunction faces, F(1,
39) = 62.56, MSE = .028, η2p = .62. As can be
seen in Figure 2, there was perhaps a small hint
of a benefit of rehearsal for the 15-s rehearsal condition, which produced a slightly higher hit score
than the other conditions and a slightly lower conjunction score than the other conditions.
However, as a whole, the image rehearsal instructions produced no clear benefit on either the hit
scores or the conjunction error scores. A 3
(visual rehearsal condition) × 2 (test condition:
old, conjunction) repeated measures ANOVA on
the corrected scores (i.e., with the new face falsealarm rate subtracted) produced nonsignificant
outcomes for the rehearsal factor, F(2, 78) =
0.05, MSE = .032, η2p = .001, and the Visual
Rehearsal Condition × Test Condition interaction, F(2, 78) = 1.89, MSE = .034, η2p = .03.
The same contrasts as those for Experiment 1 on
old faces and conjunction faces did not produce
any significant outcomes [for old faces: no rehearsal M = .53, SD = .19 vs. rehearsal, M = .56,
SD = .26, F(1, 39) = 0.81, MSE = .04; 15-s
rehearsal M = .58, SD = .20 vs. 60-s rehearsal,
M = .54, SD = .18, F(1, 39) = 1.67, MSE = .06;
for conjunction faces, no rehearsal M = .35,
SD = .24 vs. rehearsal, M = .33, SD = .18, F(1,
39) = 0.18, MSE = .01; 15-s rehearsal M = .31,
SD = .23 vs. 60-s rehearsal, M = .35, SD = .23,
F(1, 39) = 0.55, MSE = .04]. To sum up, we
removed the verbal description instructions but
retained the opportunity for the enhancement of
holistic or featural information in the nonverbal
representation(s) via image rehearsal. Unlike
Experiment 1, no benefit was obtained.
These findings are consistent with those
obtained by Shaffer and Shiffrin (1972) and Jones
and Bartlett (2009) but not those of Graefe and
Watkins (1980, also see Read, 1979; Read et al.,
1989; Sporer, 1988; Wogalter, 1991). However,
our image rehearsal instructions were narrower
than those of Graefe and Watkins. One might
argue, in a manner consistent with Shaffer and
Shiffrin, that participants were unable to maintain
or rehearse a visual image of a face (also see
Alvarez & Cavanagh, 2004). Alternatively, participants might have been able to form some sort of
image (Read & Peterson, 1975; Yarmey, 1975),
but this image may have lacked sufficient resolution
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
13
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
Figure 2. Results for Experiment 2. Mean corrected probability of an “old” response for old and conjunction conditions (i.e., with the new-face
false-alarm rate subtracted) across visual rehearsal condition. Error bars reflect the standard error within a condition. Going from left to right,
the means are .53 (SD = .19), .35 (SD = .24), .58 (SD = .20), .31 (SD = .23), .54 (SD = .20), .35 (SD = .23). The corresponding A′
scores are .85 (SD = .08), .75 (SD = .17), .87 (SD = .08), .74 (SD = .15), .85 (SD = .08), and .76 (SD = .16).
(Awh et al., 2007) to be of later value. In either
case, one comes to the straightforward conclusion
that holistic information was unlikely to have
been enhanced much, if at all, by image maintenance or retrieval in Experiment 1.
GENERAL DISCUSSION
The results for Experiment 1 showed that generating descriptions increased the hit rate but did not
influence the conjunction error rate relative to
control conditions. The generation of more
descriptors of features did not boost the effect.
Again, the description benefit cannot be attributed
to differences in encoding of the study faces or to
mentally describing a face when a face was presented because our within-participant, trial-bytrial procedure rules out these possibilities.
The feature enhancement hypothesis predicts
that a description effect should occur for old faces
and conjunction faces. The more is better
14
hypothesis predicts a larger description benefit
when more features are described (i.e., the long
description period) than when fewer features are
described (i.e., the short description period). This
hypothesis predicts a higher hit rate and a higher
conjunction error rate for the long description
period than for the short duration period.
However, there was only an effect for old faces,
with the longer description period producing a
numerically, but not significantly, lower hit rate
than that for the short description period.
Therefore, these results provide unambiguous evidence against a featural enhancement hypothesis,
as well as a more is better hypothesis (e.g., Bloom
& Mudd, 1991). Because our experiment used
the same features in the same orientation (pose)
from study to test, our experiment provides a very
rigorous test of these hypotheses.
The evidence concerning an alteration of a
holistic representation is not as strong as the evidence against a featural enhancement hypothesis.
However, when both experiments are considered
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
with additional research, our results put pressure on
the holistic enhancement hypothesis. Prior research
has been vague on precisely how a holistic representation might be altered (e.g., Brown & LloydJones, 2005), but it seems fair to assume that the
holistic representation would need to be accessed
in some manner in order for it to be altered.
Experiment 1 showed a description effect that
could be explained by enhancement of holistic
information (emphasis in the descriptions for features notwithstanding). The visual rehearsal
instructions of Experiment 2 were intended to
test whether direct (conscious) access to and rehearsal of a holistic image would benefit later recognition, which Nakabayashi, Burton, et al. (2012)
pointed to as a reason for Brown et al.’s (2010)
description benefits. However, there was no effect
of visual rehearsal. Thus, the findings of
Experiment 1 do not appear to be based on direct
access of holistic information. This conclusion is
consistent with the conclusions from other studies
that participants are unable to hold and rehearse
high-resolution mental images of complex stimuli
(Alvarez & Cavanagh, 2004; Awh et al., 2007;
Jones & Bartlett, 2009; Shaffer & Shiffrin, 1972;
though see Read, 1979; Sporer, 1988, 1989). To
be clear, our results do not make any bold claims
with regard to face processing. Instead, the
benefit of descriptions with the present method
does not appear to be based on enhancement/
alteration of nonverbal representations that
support face processing/recognition. We encourage
researchers who advocate a holistic enhancement
(alteration) account to more clearly indicate how
the holistic information might be accessed and
altered.
The results support our proposal that the face
stimuli provide the primary basis for the recognition decision and that a description generated
during the study phase represents additional information that can contribute to the recognition
decision. Bartlett et al. (2003) suggested that a
match in holistic information may be required to
cue retrieval of additional information, and our
results support this contention. When the descriptive information is accurate and is retrieved during
the test, it can help one to identify a study face as
“old” to a small but significant degree. Evidence
of subjective judgements obtained by Brown and
Lloyd-Jones (2006) suggests that the verbal
description, in part or whole, may be accessed via
recollection (e.g., Brown & Lloyd-Jones, 2006).
Lockhart’s (1975) rationale on the benefit of
recall on later recognition (also see Jones &
Roediger, 1995) is applicable for the benefits of
generating descriptions on later recognition. For
some faces, the familiarity strength would be sufficient to pass an “old–new” decision threshold
regardless of whether they were described. In such
cases, verbal information might be retrieved but
not change the binary “old–new” decision. The
description benefit thus concerns those items that
lack sufficient strength to be recognized without
being described. Put simply, for a small proportion
of faces the retrieval of the description could nudge
the evidence for an “old” judgement over a decision
threshold. A reasonable expectation from this
approach is that, although the binary response
might not be affected by retrieval of the description
in some cases, the retrieval of additional information should make one more confident in the recognition
decision.
Therefore,
confidence
judgements should be higher in a description condition than in a no-description condition. Prior
work with descriptions provided by the experimenter supports this expectation (Kerr &
Winograd, 1982).
In the remainder of this discussion, we relate our
findings to other findings and theories on description benefits, as well as costs. The findings reproduced recent findings of a benefit of generating
descriptions of faces on a later recognition test
(Brown et al., 2010; Brown & Lloyd-Jones, 2005,
2006; Nakabayashi, Lloyd-Jones, et al., 2012;
Wickham & Lander, 2008). The findings are also
consistent with those on the benefit of descriptive
labels for faces (McKelvie, 1976) and the benefit
of descriptive information provided by the experimenter (Kerr & Winograd, 1982; Winograd,
1981). Our conclusions are generally consistent
with Winograd’s (1981) and Kerr and Winograd’s
(1982) ideas on the benefits of elaboration and
Paivio’s (1971) idea that a verbal code and an
image code can contribute to memory performance.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
15
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
Our conclusion regarding the generation and
retrieval of verbal information in the description
is concordant with Brown and colleagues’ (Brown
et al., 2010; Brown & Lloyd-Jones, 2005, 2006)
consistent conclusion that the verbal (semantic)
information is a locus of the description benefit,
but we have provided a proposal that incorporates
the effect into the bigger picture of how the recognition memory task is accomplished. Although
Brown and Lloyd-Jones (2005, 2006) suggested
that the description may serve as an action tag
during retrieval (cf., Loftus & Kallman, 1979),
there is nothing in our results or their results that
would require the verbal information to necessarily
guide processing of a test face. That is, the action
part of action tag is not necessary to explain the
present results.
Our results rule out hypotheses on the benefit
for the present experiment, but they do not rule
out any of the variety of possibilities for the
benefit in studies where the description condition
is known prior to stimulus presentation. The
present results more clearly suggest, however, that
retrieval of the verbal description itself may be the
basis for at least part of the effect in those studies.
For findings where description conditions are
known a priori, it is very difficult to strongly rule
out hypotheses on the benefit in the description
condition. Nakabayashi, Lloyd-Jones, and colleagues (2012, Experiment 1) manipulated
whether same-race or different-race faces were presented intact or scrambled at study and intact or
scrambled at test. A description benefit occurred
for faces that were presented intact at study but
not for faces that were presented scrambled at
study, suggesting to the authors that configural
information was enhanced by the description.
One oddity is that, for the control conditions,
faces that were studied as scrambled produced
better recognition performance than faces studied
as intact. (However, for faces studied as scrambled,
descriptions produced lower scores than those for
the control condition.) The pose of the faces in
all conditions changed from study to test, so there
is some ambiguity on what information participants
used in the different conditions.
16
In a second experiment, Nakabayashi, LloydJones, and colleagues (2012) obtained eye-tracking
data during the study and test phases. They concluded that the description condition encouraged
greater sampling of the internal features, which
resulted in better configural processing of those
inner features. For our experiment, if descriptions
enhanced an inner configuration, then the conjunction error rate should have been higher for the
description condition than for the control condition. Instead, there was no effect. Of course,
there is the possibility that knowing a face was to
be described a priori would increase the conjunction
error rate. Thus, the conjunction paradigm would
provide considerable leverage to address this
question.
Finally, for their first experiment, Nakabayashi,
Burton, and colleagues (2012) found a benefit for
a description condition over an articulatory suppression condition but not over a finger-tapping
condition, with the finger-tapping condition producing superior performance to the articulatory
suppression condition. Their study stimuli were
presented for 7 seconds (each), and each description was produced orally while a stimulus was in
full view. Nakabayashi, Burton, et al. did not
make any strong claims about what the fingertapping control participants were doing during
the presentation of the study stimuli, though participants knew that their memory for the study
stimuli would be tested. The most straightforward
possibility is that in the finger-tapping condition,
participants covertly noted characteristics of the
faces that they would have noted aloud in the
description condition. If so, then generating
descriptive information covertly or overtly (orally
or in written form) may produce a description
benefit. If participants have a tendency to make
covert verbal notes in a control condition, then
this encoding strategy creates an obstacle for
obtaining description benefits. From this standpoint, the relatively brief (2-s) study stimulus presentation in our Experiment 1 and other
experiments (e.g., Brown et al., 2010; Brown &
Lloyd-Jones, 2005, 2006; Nakabayashi, LloydJones, et al., 2012) may have created a better
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
VERBAL DESCRIPTION BENEFIT
opportunity to observe a description benefit by limiting the time for any covert description.
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
Description costs
Neither experiment was designed to obtain a
description detriment (i.e., a verbal overshadowing
effect). We think that description accuracy
(Meissner et al., 2008; also see Kurtz & Hovland,
1953) and conditions that influence description
accuracy (immediate vs. delayed description;
Meissner et al., 2008; Nakabayashi & Burton,
2008; Nakabayashi, Burton, et al., 2012; the enforcement of producing descriptive information for
the full description period, Meissner et al., 2001)
are very important to whether a benefit or cost
will be obtained for the stimuli that are described.
Nevertheless, there are three theories for why a
cost may occur (e.g., Chin & Schooler, 2008).
One idea is that descriptions induce an emphasis
on featural processing during the test, which is
inconsistent with an emphasis on holistic processing assumed to occur for the original encoding of
a study face. According to this idea, our description
conditions should have lowered the hit rate but
raised the conjunction error rate relative to the
no-description control conditions. (The opposite
pattern occurred.) A second idea is that providing
a description simply induces a criterion shift, with
participants in a description group being less
willing to endorse targets or lures as “old”. Such evidence comes from the manipulation of description
conditions between groups, with each group
getting just one study item (e.g., Clare &
Lewandowsky, 2004; Sauerland et al., 2008).
Both of our experiments used an entirely withinparticipants design with study and test conditions
mixed on a trial-by-trial basis. Thus, although a
bias account may sometimes capture what appears
to be a description cost, the description benefit in
our first experiment cannot reasonably be attributed
to a decision threshold difference. A third idea is
that the verbal description content interferes with
retrieval of the corresponding visual representation.
Although we obtained no evidence of retrieval
interference, this idea is somewhat hard to discount
for studies that produce a cost. Alternatively
though, as forgetting occurs, it will be harder to
identify a test probe as a match, leaving one to
seek whatever additional information might be
retrieved. If the content of the description deviates
from the study face, then the likelihood of a cost
increases (Meissner et al., 2008). We suspect that
the weighting of this additional information
might vary based on a participant’s expectation
that it is accurate or diagnostic (cf. Bartlett et al.,
1980). To summarize, there is a considerable
amount of research that focuses on costs of descriptions, the inaccuracies of the descriptions, and the
relationship of the inaccuracies and final recognition performance. Our experiments, in contrast,
focused on a circumstance where the descriptions
were short and quite accurate. Apart from the
idea that participants may rely on the retrieval of
verbal information (i.e., in the description) in the
absence of the retrieval of a high-resolution
image, the ideas on why costs occur do not
account for our results.
CONCLUSION
The present results provide evidence against the
alteration (enhancement) of featural or holistic
information as potential bases of the description
benefit when participants do not know a priori
whether a description will need to be generated.
We have treated the description as an additional
source of information that helps one make a recognition decision, and the data indicate that the better
the match between the study stimulus and test
probe, the more likely this verbal information is
to be retrieved and aid in the recognition decision.
This conclusion is indirect but is bolstered by Kerr
and Winograd’s (1982) findings where recognition
confidence and description recall were directly
measured. When considered from an overall
study–test match, the present findings complement
verbal context effects for faces (Watkins, Ho, &
Tulving, 1976) and visual context effects for faces
(Watkins et al., 1976; Winograd & RiversBulkeley, 1977) and are consistent with the conclusion that a stronger match between the study
episode and the target episode will confer a
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
17
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
benefit over conditions with a weaker match
(Watkins & Tulving, 1975). Winograd and colleagues’ (Kerr & Winograd, 1982; Winograd,
1981) work showed that a benefit could be obtained
with descriptions provided by the experimenter.
Thus, actual generation is not necessary to obtaining a benefit, but it may have provided some of the
benefit obtained in the present experiment (also see
Brown et al., 2010; Sporer, 1991). Future work
might focus on creating stronger tests of hypotheses
when description instructions are known a priori or
attempt to provide a better understanding of the
benefit of actually generating the description or
assess what and how verbal information is retrieved
during test trials (cf., Chance & Goldstein, 1976;
Kerr & Winograd, 1982).
Original manuscript received 20 November 2011
Accepted revision received 26 November 2012
First published online 13 March 2013
REFERENCES
Alvarez, G. A., & Cavanagh, P. (2004). The capacity
of visual short term memory is set both by information
load and by number of objects. Psychological Science, 15,
106–111. doi:10.1111/j.0963-7214.2004.01502006.x
Anderson, J. R., & Reder, L. M. (1979). An elaborative
processing explanation of depth of processing. In
L. S. Cermak & F. I. M. Craik (Eds.), Levels of processing in human memory (pp. 385–403). Oxford, UK:
Lawrence Erlbaum Associates.
Arndt, J., & Jones, T. C. (2008). Elaborative processing
and conjunction errors in memory. Memory &
Cognition, 36, 899–912. doi:10.3758/MC.36.5.899
Awh, E., Barton, B., & Vogel, E. K. (2007). Visual
working memory represents a fixed number of items
regardless of complexity. Psychological Science, 18,
622–628. doi:10.1111/j.1467-9280.2007.01949.x
Banks, W. P. (1970). Signal detection theory and human
memory. Psychological Bulletin, 74, 81–99.
Bartlett, J. C., Searcy, J. H., & Abdi, H. (2003). What
are the routes to face recognition? In M. A.
Peterson & G. Rhodes (Eds.), Perception of faces,
objects, and scenes: Analytic and holistic processes.
Advances in visual cognition (pp. 21–47). Oxford:
Oxford University Press.
18
Bartlett, J. C., Till, R. E., & Levy, J. C. (1980). Retrieval
characteristics of complex pictures: Effects of verbal
encoding. Journal of Verbal Learning and Verbal
Behavior, 19, 430–449. doi:10.1016/S0022-5371
(80)90303-5
Bloom, L. C., & Mudd, S. A. (1991). Depth of processing approach to face recognition: A test of two theories. Journal of Experimental Psychology: Learning,
Memory, and Cognition, 17, 556–565. doi:10.1037/
0278-7393.17.3.556
Bower, G. H., & Karlin, M. B. (1974). Depth of processing pictures of faces and recognition memory. Journal
of Experimental Psychology, 103, 751–757.
doi:10.1037/h0037190
Brown, C., Gehrke, J., & Lloyd-Jones, T. J. (2010). A
visual and semantic locus to beneficial effects of verbalization on face memory. American Journal of
Psychology, 123, 51–69.
Brown, C., & Lloyd-Jones, T. J. (2005). Verbal facilitation of face recognition. Memory & Cognition, 33,
1442–1456. doi:10.3758/BF03193377
Brown, C., & Lloyd-Jones, T. J. (2006). Beneficial
effects of verbalization and visual distinctiveness on
remembering and knowing faces. Memory &
Cognition, 34, 277–286. doi:10.3758/BF03193377
Bruce, V. (1982). Changing faces: Visual and non-visual
coding processes in face recognition. British Journal of
Psychology, 73, 105–116. doi:10.1111/j.2044-8295.
1982.tb01795.x
Bruce, V., & Young, A. (1986). Understanding face recognition. British Journal of Psychology, 77, 305–327.
doi:10.1111/j.2044-8295.1986.tb02199.x
Busey, T. A., & Tunnicliff, J. L. (1999). Accounts of
blending, distinctiveness, and typicality in the false
recognition of faces. Journal of Experimental
Psychology: Learning, Memory, and Cognition, 25,
1210–1235. doi:10.1037/0278-7393.25.5.1210
Busey, T. A., Tunnicliff, J. L., Loftus, G. R., & Loftus,
E. (2000). Accounts of confidence–accuracy relation
in recognition memory. Psychonomic Bulletin &
Review, 7, 26–48. doi:10.3758/BF03210724
Cabeza, R., & Kato, T. (2000). Features are also important: Contributions of featural and configural processing to face recognition. Psychological Science, 11,
429–433. doi:10.1111/1467-9280.00283
Carey, S., & Diamond, R. (1977). From piecemeal to
configurational representation of faces. Science, 195,
312–314. doi:10.1126/science.831281
Chance, J., & Goldstein, A. G. (1976). Recognition of
faces and verbal labels. Bulletin of the Psychonomic
Society, 7, 384–386.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
Chin, J. M., & Schooler, J. W. (2008). Why do words
hurt? Content, process, and criterion shifts in verbal
overshadowing. European Journal of Cognitive
Psychology, 20, 396–413. doi:10.1080/0954144070
1728623
Clare, J., & Lewandowsky, S. (2004). Verbalizing facial
memory: Criterion effects in verbal overshadowing.
Journal of Experimental Psychology: Learning,
Memory, and Cognition, 30, 739–755. doi:10.1037/
0278-7393.30.4.739
Clark, S. E., & Gronlund, S. D. (1996). Global matching models of recognition memory: How the models
match the data. Psychonomic Bulletin & Review, 3,
37–60. doi:10.3758/BF03210740
Daniel, T. C. (1972). Nature of the effect of verbal labels
on recognition memory for form. Journal of
Experimental Psychology, 96, 152–157. doi:10.1037/
h0033486
Davies, G., & Milne, A. (1982). Recognizing faces in
and out of context. Current Psychological Research, 2,
235–246. doi:10.1007/BF02684516
Diamond, R., & Carey, S. (1986). Why faces are and are
not special. Journal of Experimental Psychology:
General, 115, 107–117.
Edmonds, E. C., Glisky, E. L., Bartlett, J. C., &
Rapscak, S. Z. (2012). Cognitive mechanisms of
false facial recognition in older adults. Psychology &
Aging, 27, 54–60. doi:10.1037/a0024582
Ellis, H. D., Shepherd, J. W., & Davies, G. M. (1979).
Identification of familiar and unfamiliar faces from
internal and external features: Some implications for
theories of face recognition. Perception, 8, 431–439.
doi:10.1068/p080431
Friere, A., Lee, K., & Symons, L. A. (2000). The faceinversion effect as a deficit in the encoding of configural information: Direct evidence. Perception, 29,
159–170. doi:10.1068/p3012
Galton, F. (1879). Composite portraits, made by combining those of many different persons into a single,
resultant figure. The Journal of the Anthropological
Institute of Great Britain and Ireland, 8, 132–144.
Graefe, T. M., & Watkins, M. J. (1980). Picture rehearsal: An effect of selectively attending to pictures no
longer in view. Journal of Experimental Psychology:
Human Learning and Memory, 6, 156–162.
doi:10.1037/0278-7393.6.2.156
Harris, A., & Aguirre, G. K. (2008). The representation
of parts and wholes in face-selective cortex. Journal of
Cognitive Neuroscience, 20, 863–878. doi:10.1162/
jocn.2008.20509
The Hawkeye (1963). Iowa City. IA: State University of
Iowa.
Hsiao, J. H., & Cottrell, G. (2008). Two fixations suffice
in face recognition. Psychological Science, 19,
998–1006. doi:10.1111/j.1467-9280.2008.02191.x
Ingram, K. M., Mickes, L., & Wixted, J. T. (2012).
Recollection can be weak and familiarity can be
strong. Journal of Experimental Psychology: Learning,
Memory, and Cognition, 38, 325–339. doi:10.1037/
a0025483
Jacoby, L. L. (1991). A process dissociation framework:
Separating automatic from intentional uses of
memory. Journal of Memory & Language, 30,
513–541. doi:10.1016/0749-596X(91)90025-F
Jacoby, L. L. (1999). Ironic effects of repetition:
Measuring age-related differences in memory.
Journal of Experimental Psychology: Learning,
Memory, and Cognition, 25, 3–22. doi:10.1037/
0278-7393.25.1.3
James, W. (1890). Principles of psychology (Vol. 1).
London: Macmillan.
Jones, T. C., & Atchley, P. (2006). Conjunction errors,
recollection-based rejections, and forgetting in a continuous recognition task. Journal of Experimental
Psychology: Learning, Memory, and Cognition, 32,
70–78. doi:10.1037/0278-7393.32.1.70
Jones, T. C., & Bartlett, J. C. (2009). When false recognition is out of control: The case of facial conjunctions. Memory & Cognition, 37, 143–157.
doi:10.3758/MC.37.2.143
Jones, T. C., Bartlett, J. C., & Wade, K. A. (2006).
Nonverbal conjunction errors in recognition
memory: Support for familiarity but not for feature
bundling. Journal of Memory and Language, 55,
138–155. doi:10.1016/j.jml.2006.01.002
Jones, T. C., Brown, A. S., & Atchley, P. (2007). Feature
and conjunction effects in recognition memory:
Toward specifying familiarity for compound words.
Memory & Cognition, 35, 984–998. doi:10.3758/
BF03193471
Jones, T. C., & Jacoby, L. L. (2001). Feature and conjunction errors in recognition memory: Support for
a dual-process theory. Journal of Memory and
Language, 45, 82–102. doi:10.1006/jmla.2000.2761
Jones, T. C., & Roediger, H. L. (1995). The experiential
basis of serial position effects. European Journal of
Cognitive Psychology, 7, 65–80.
Kerr, N. H., & Winograd, E. (1982). Effects of contextual elaboration on face recognition. Memory &
Cognition, 10, 603–609. doi:10.3758/BF03202443
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
19
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
Klatzky, R. L., Martin, G. L., & Kane, R. A. (1982).
Semantic interpretation effects on memory for faces.
Memory & Cognition, 10, 195–206.
Kroll, N. E. A., Knight, R. T., Metcalfe, J., Wolf, E. S.,
& Tulving, E. (1996). Cohesion failure as a source of
memory illusions. Journal of Memory & Language, 35,
176–196. doi:10.1006/jmla.1996.0010
Kurtz, K. H., & Hovland, C. I. (1953). The effect of verbalization during observation of stimulus objects upon
accuracy of recognition and recall. Journal of
Experimental Psychology, 45, 157–164. doi:10.1037/
h0061022
Lampinen, J. M., Odegard, T. N., & Neuschatz, J. S.
(2004). Robust recollection rejection in the memory
conjunction paradigm. Journal of Experimental
Psychology: Learning, Memory, and Cognition, 30,
332–342. doi:10.1037/0278-7393.30.2.332
Lloyd-Jones, T. J., Brandimonte, M. A., & Bäuml,
K.-H. (2008). Verbalising visual memories.
European Journal of Cognitive Psychology, 20,
387–395. doi:10.1080/09541440701755543
Lockhart, R. S. (1975). The facilitation of recognition by
recall. Journal of Verbal Learning and Verbal Behavior,
14, 253–258. doi:10.1016/S0022-5371(75)80068-5
Loftus, G., & Bell, S. M. (1975). Two types of information in picture memory. Journal of Experimental
Psychology: Human Learning and Memory, 1,
103–113. doi:10.1037/0278-7393.1.2.103
Loftus, G. R., & Kallman, H. J. (1979). Encoding and use
of detail information in picture recognition. Journal of
Experimental Psychology: Human Learning and
Memory, 5, 197–211. doi:10.1037/0278-7393.5.3.197
Longmore, C. A., Liu, C. H., & Young, A. W. (2008).
Learning faces from photographs. Journal of
Experimental Psychology: Human Perception and
Performance, 34, 77–100. doi:10.1037/00961523.34.1.77
Maurer, D., Le Grand, R., & Mondloch, C. J. (2002).
The many faces of configural processing. Trends in
Cognitive Sciences, 6, 255–260. doi:10.1016/S13646613(02)01903-4
McKelvie, S. J. (1976). The effects of labelling on recognition memory for schematic faces. The Quarterly
Journal of Experimental Psychology, 28, 459–474.
doi:10.1080/14640747608400573
McKone, E., & Peh, Y. X. (2006). Memory conjunction
errors for realistic faces are consistent with configural
processing. Psychonomic Bulletin & Review, 13,
106–111. doi:10.3758/BF03193820
McKone, E., & Yovel, G. (2009). Why does pictureplane inversion sometimes dissociate perception of
20
features and spacing in faces, and sometimes not?
Toward a new theory of holistic processing.
Psychonomic Bulletin & Review, 16, 778–797.
doi:10.3758/PBR.16.5.778
Meinhardt-Injac, B., Meinhardt, G., & Schwaninger, A.
(2009). Does matching of internal and external facial
features depend on orientation and viewpoint?
Acta Psychologica, 132, 267–278. doi:10.1016/j.
actpsy.2009.07.011
Meissner, C. A., & Brigham, J. C. (2001). A meta-analysis of the verbal overshadowing effect in face identification. Applied Cognitive Psychology, 15, 603–616.
doi:10.1002/acp.728
Meissner, C. A., Brigham, J. C., & Kelley, C. M. (2001).
The influence of retrieval processes in verbal overshadowing. Memory & Cognition, 29, 179–186.
doi:10.3758/BF03195751
Meissner, C. A., Sporer, S. L., & Susa, K. J. (2008). A
theoretical review and meta-analysis of the description–identification relationship in memory for faces.
European Journal of Cognitive Psychology, 20,
414–455. doi:10.1080/09541440701728581
Mickes, L., Wais, P. E., & Wixted, J. T. (2009).
Recollection is a continuous process: Implications
for dual process theories of recognition memory.
Psychological Science, 20, 509–515. doi:10.1111/
j.1467-9280.2009.02324.x
Mondloch, C., Le Grand, R., & Maurer, D. (2002).
Configural face processing develops more slowly
than featural face processing. Perception, 31,
553–566. doi:10.1068/p3339
Nakabayashi, K., & Burton, A. M. (2008). The role of verbal
processing at different stages of recognition memory for
faces. European Journal of Cognitive Psychology, 20,
478–496. doi:10.1080/09541440801946174
Nakabayashi, K., Burton, A. M., Brandimonte, M. A.,
& Lloyd-Jones, T. J. (2012). Dissociating positive
and negative influences of verbal processing on
the recognition of pictures of faces and objects.
Journal of Experimental Psychology: Learning,
Memory, and Cognition, 38, 376–390. doi:10.1037/
a0025782
Nakabayashi, K., Lloyd-Jones, T. J., Butcher, N., & Liu, C.
H. (2012). Independent influences of verbalization and
race on the configural and featural processing of faces: A
behavioural and eye movement study. Journal of
Experimental Psychology: Learning, Memory, and
Cognition, 38, 61–77. doi:10.1037/a0024853
Nosofsky, R. M. (1988). Exemplar-based accounts of
relations between classification, recognition, and typicality. Journal of Experimental Psychology: Learning,
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
VERBAL DESCRIPTION BENEFIT
Memory, and Cognition, 14, 700–708. doi:10.1037/
0278-7393.14.4.700
O’Toole, A. J., Edelman, S., & Bülthoff, H. H. (1998).
Stimulus-specific effects in face recognition over
changes in viewpoint. Vision Research, 38,
2351–2363. doi:10.1016/S0042-6989(98)00042-X
O’Toole, A. J., Jiang, F., Roark, D., & Abdi, H. (2006).
Predicting human performance for face recognition.
Face recognition by humans. In W. Zhao &
R. Chellappa (Eds.), Face processing: Advanced modeling and methods (pp. 293–319). New York, NY:
Academic Press.
O’Toole, A. J., Vetter, T., & Blanz, V. (1999). Threedimensional shape and two-dimensional surface
reflectance contributions to face recognition: An
application of three-dimensional morphing. Vision
Research, 39, 3145–3155. doi:10.1016/S0042-6989
(99)00034-6
Paivio, A. (1971). Imagery and verbal processes. New York,
NY: Holt, Rinehart, & Winston.
Parks, C. M., & Yonelinas, A. P. (2007). Moving
beyond pure signal detection models: Comment on
Wixted (2007). Psychological Review, 114, 188–201.
doi:10.1037/0033-295X.114.1.188
Parks, T. E. (1966). Signal detectability theory of recognition memory performance. Psychological Review, 73,
44–58. doi:10.1037/h0022662
Read, J. D. (1979). Rehearsal and recognition of human
faces. American Journal of Psychology, 92, 71–85.
doi:10.2307/1421480
Read, J. D., Hammersley, R., Cross-Calvert, S., &
McFadzen, E. (1989). Rehearsal of faces and details
in action events. Applied Cognitive Psychology, 3,
295–311. doi:10.1002/acp.2350030403
Read, J. D., & Peterson, R. H. (1975). Individual differences in the ease of imagining the faces of others.
Bulletin of the Psychonomic Society, 5, 347–349.
Reinitz, M. T., Lammers, W. J., & Cochran, B. P.
(1992).
Memory
conjunction
errors:
Miscombination of stored features can produce illusions of memory. Memory & Cognition, 20, 1–11.
doi:10.3758/BF03208247
Reinitz, M. T., Morrissey, J., & Demb, J. (1994). Role of
attention in face encoding. Journal of Experimental
Psychology: Learning, Memory, and Cognition, 20,
161–168. doi:10.1037/0278-7393.20.1.161
Rotello, C. M., Macmillan, N. A., & Van Tassel, G.
(2000). Recall-to-reject in recognition: Evidence
from ROC curves. Journal of Memory & Language,
43, 67–88. doi:10.1006/jmla.1999.2701
Sauerland, M., Holub, F. E., & Sporer, S. L. (2008).
Person descriptions and person identifications:
Verbal overshadowing or recognition criterion shift?
European Journal of Cognitive Psychology, 20,
497–528. doi:10.1080/09541440701728417
Schneider, W., Eschman, A., & Zuccolotto, A. (2002a).
E-Prime reference guide. Pittsburgh, PA: Psychology
Software Tools.
Schneider, W., Eschman, A., & Zuccolotto, A. (2002b).
E-Prime user’s guide. Pittsburgh, PA: Psychology
Software Tools.
Schooler, J. W., & Engstler-Schooler, T. Y. (1990).
Verbal overshadowing of visual memories: Some
things are better left unsaid. Cognitive
Psychology, 22, 36–71. doi:10.1016/0010-0285
(90)90003-M
Searcy, J. H., & Bartlett, J. C. (1996). Inversion and processing of component and spatial–relational information in faces. Journal of Experimental Psychology:
Human Perception and Performance, 22, 904–915.
doi:10.1037/0096-1523.22.4.904
Shaffer, W., & Shiffrin, R. M. (1972). Rehearsal and
storage of visual information. Journal of Experimental
Psychology, 92, 292–296. doi:10.1037/h0032076
Shapiro, P. N., & Penrod, S. (1986). Meta-analysis of
facial identification studies. Psychological Bulletin,
100, 139–156. doi:10.1037/0033-2909.100.2.139
Shepard, R. N., & Teghtsoonian, M. (1961). Retention
of information under conditions approaching a steady
state. Journal of Experimental Psychology, 62, 302–309.
doi:10.1037/h0048606
Sinha, P., Balas, B. J., Ostrovsky, Y., & Russell, R.
(2006). Face recognition by humans. In W. Zhao &
R. Chellappa (Eds.), Face processing: Advanced modeling and methods (pp. 257–291). New York, NY:
Academic Press.
Sinha, P., & Poggio, T. (1996). I think I know that face.
…. Nature, 384, 404. doi:10.1038/384404a0
Sinha, P., & Poggio, T. (2002). “United” we stand.
Perception, 31, 133. doi:10.1068/p3101no
Sporer, S. L. (1988). Long-term improvement of facial
recognition through visual rehearsal. In M. M.
Gruneberg, P. E. Morris, & R. N. Sykes (Eds.),
Practical aspects of memory: Current research and issues
(pp. 182–188). New York, NY: John Wiley & Sons.
Sporer, S. L. (1989). Verbal and visual processes in
person identification. In H. Wegener, F. Lsel, &
J. Haisch (Eds.), Criminal behavior and the justice
system: Psychological perspectives (pp. 303–324).
New York, NY: Springer-Verlag.
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)
21
Downloaded by [Royal Holloway, University of London] at 07:38 14 March 2013
JONES ET AL.
Sporer, S. L. (1991). Deep—deeper—deepest? Encoding
strategies and the recognition of human faces. Journal
of Experimental Psychology: Learning, Memory, and
Cognition, 17, 323–333. doi:10.1037/0278-7393.
17.2.323
Sporer, S. L., & Horry, R. (2011a). Pictorial versus structural representations of ingroup and outgroup faces.
Journal of Cognitive Psychology, 1, 1–11.
doi:10.1080/20445911.2011.594434
Sporer, S. L., & Horry, R. (2011b). Recognizing faces
from ethnic in-groups and out-groups: Importance
of outer face features and effects of retention interval.
Applied Cognitive Psychology, 25, 424–431.
doi:10.1002/acp.1709
Tanaka, J. W., & Farah, M. J. (1993). Parts and wholes
in face recognition. Quarterly Journal of Experimental
Psychology, 46A, 225–245. doi:10.1080/14640749
308401045
Tulving, E. (1983). Elements of episodic memory, Oxford:
Clarendon Press.
Underwood, B. J., & Zimmerman, J. (1973). The syllable
as a source of error in multisyllable word recognition.
Journal of Verbal Learning and Verbal Behavior, 12,
701–706. doi:10.1016/S0022-5371(73)80050-7
Valentine, T. (1988). Upside-down faces: A review of the
effect of inversion on face recognition. British Journal
of Psychology, 79, 471–491. doi:10.1111/j.20448295.1988.tb02747.x
Walker-Smith, G. J., Gale, A. G., & Findlay, J. M.
(1977). Eye movement strategies involved in face perception. Perception, 6, 313–326. doi:10.1068/p060313
Watkins, M. J., Ho, E., & Tulving, E. (1976). Context
effects in recognition memory for faces. Journal of
Verbal Learning and Verbal Behavior, 15, 505–517.
doi:10.1016/0022-5371(76)90045-1
Watkins, M. J., & Tulving, E. (1975). Episodic memory:
When recognition fails. Journal of Experimental
Psychology: General, 1, 5–29. doi:10.1037/00963445.104.1.5
Wells, G. L. (1985). Verbal descriptions of faces from
memory: Are they diagnostic of identification accuracy? Journal of Applied Psychology, 70, 619–626.
doi:10.1037/0021-9010.70.4.619
22
Wells, G. L., & Hryciw, B. (1984). Memory for faces:
Encoding and retrieval operations. Memory &
Cognition, 12, 338–344. doi:10.3758/BF03198293
Wickham, L. H. V., & Lander, K. (2008). The effect of
description and processing type on face identification.
European Journal of Cognitive Psychology, 20,
577–586. doi:10.1080/09541440701728417
Winograd, E. (1978). Encoding operations which facilitate memory for faces across the life span. In M. M.
Gruneberg, P. E. Morris, & R. N. Sykes (Eds.),
Practical aspects of memory. New York, NY:
Academic Press.
Winograd, E. (1981). Elaboration and distinctiveness in
memory for faces. Journal of Experimental Psychology:
Human Learning and Memory, 7, 181–190.
doi:10.1037/0278-7393.7.3.181
Winograd, E., & Rivers-Bulkeley, N. T. (1977). Effects
of changing context on remembering faces. Journal of
Experimental Psychology: Human Learning and
Memory, 3, 397–405. doi:10.1037/0278-7393.3.4.397
Wixted, J. T. (2007). Dual-process theory and
signal-detection theory of recognition memory.
Psychological Review, 114, 152–176. doi:10.1037/
0033-295X.114.1.152
Wogalter, M. S. (1991). Effects of post-exposure
description and imaging on subsequent face recognition performance. Proceedings of the Human Factors
Society, 35, 575–579.
Yarmey, A. D. (1975). Introspection and imagery reports
of human faces. Perceptual and Motor Skills, 41,
711–719. doi:10.2466/pms.1975.41.3.711
Yin, R. K. (1969). Looking at upside-down faces. Journal
of Experimental Psychology: Human Learning and
Memory, 7, 181–190. doi:10.1037/h0027474
Yonelinas, A. P. (1994). Receiver-operating characteristics in recognition memory: Evidence for a dualprocess model. Journal of Experimental Psychology:
Learning, Memory, and Cognition, 20, 1341–1354.
doi:10.1037/0278-7393.20.6.1341
Yonelinas, A. P. (2002). The nature of recollection
and familiarity: A review of 30 years of research.
Journal of Memory and Language, 46, 441–517.
doi:10.1006/jmla.2002.2864
THE QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2013, 00 (0)