Substantive bias in the learning of harmony patterns

Substantive bias in the learning of harmony patterns
Allison Shapp
New York University
December 5, 2012
1
Introduction
Not all possible linguistic patterns are learned with the same ease, and both effects of the
formal structure (analytic bias) and effects of the phonetic content (substantive bias) of
the patterns have been shown to exist. (Moreton 2008, Carpenter 2010). In regards to
the substantive content of phonological patterns, some patterns can be shown to be better
phonetically grounded than others. Often, phonetically grounded patterns are also widely
attested, but this is not always the case. The experiments in this paper compare anteriority
harmony, which is phonetically grounded and well attested, with voicing harmony, which
is less well grounded and also typologically rare. The results of these experiments show
that the anteriority harmony is better learned than the voicing harmony. In this case, the
learnability, phonetic grounding, and typological attestation of these two patterns are all in
line with each other. These two formally identical patterns differ only in the feature being
harmonized, yet one of them is easier to learn. This asymmetry supports a substantive bias
in learnability.
2
Previous Research on Learnability
The learnability of linguistic patterns, as well as analytic bias and substantive bias specifically, have been fruitfully explored through the use of artificial grammar learning experiments, where participants are taught linguistic patterns in a made-up language. Previous research has shown that phonological patterns can be successfully taught through this method,
and results which show both analytic and substantive bias have been found. Here I will summarize several of these previous results and use them to set the stage for the presentation
of my experiments. In section 2.1 I discuss learning studies more generally, then in 2.2 I
specifically look at a few studies that deal with substantive bias.
2.1
Learning Linguistic Patterns in Artificial Grammar
In Finley and Badecker (2009), English speakers were taught a backness vowel harmony
pattern. The stimuli were stems of the form CVCV with an additional -CV syllable as a
suffix. The suffix was either -mi if the C2 of the stem was a front vowel (/i/, /e/, or /æ/), and
-mu if the preceding vowel in the stem was back (/u/, /o/, or /a/). The method they used for
training the participants was an alternation. Participants were presented with a stem, such as
[p ædi] or [muto], and then stem+suffix pairs where the vowel in the suffix matched the vowel
1
in the stem on the feature of backness ([pædimi] or [mutomu], respectively). Participants
were part of either a mid-holdout group or a low-holdout group. The mid-holdout group
was trained on harmonic stimuli consisting of only high and low vowels, and the low-holdout
group was trained on only high and mid vowels. Then all participants were asked in a forced
choice test to choose between a harmonic suffixed form and a disharmonic suffixed form.
They were tested on old stems, new stems with the same consonant and vowel inventory
that they saw in training, and finally, new stems that contained the vowel class that they
were not exposed to during training.
Finley and Badecker found that participants were able to learn the vowel harmony and
generalize it to novel segments (those in the height class that they were not trained on),
suggesting that it was feature-based learning that was happening rather than segment level
learning. If the learning had been segment level, the participants would not have been able
to produce the same backness harmony for the vowel segments they had not been exposed
to in training. Finley and Badecker (2009) had very robust findings that this pattern was
learned well by their participants, and that feature based generalizations were being formed.
Finley (2011) ran a similar experiment, teaching a pattern of harmony of sibilant consonants. Participants were again given alternations of stem+suffix pairs, where if the last
consonant in the stem was a [S] then the suffix was [-Su], and if the last consonant in the
stem was [s], then the suffix was [-su]. Stimuli stems were only of the form CVsV or CVSV,
so the target consonant was always identical to the trigger consonant in harmonic stimuli.
In testing, participants were given a two-alternative forced-choice between a harmonic and a
disharmonic form that differed only between the two suffixes (e.g. [bisisu] vs. [bisiSu]), and
asked to choose which was part of the language that they just learned in training.
Finley’s results show that subjects favored harmonic forms more in the training condition
than in the control condition, but the experiment only tested cases of strict segmental identity
(only one anterior segment [s] and one posterior segment [S] were used). I test the same
anteriority harmony pattern in the experiments in this paper, but in both training and
testing there are three anterior segments [s,z,ts] and three posterior segments [S, Z, tS]. My
results show that anteriority harmony can be learned at the featural level among non-identical
segments.
Moreton (2008) investigated how well different assimilation patterns were learned based
on both analytic structures and different substantive structures. He isolates the effects of
analytic bias, as opposed to substantive bias, by choosing two patterns that are equally well
phonetically grounded but differ in formal representation. He looked at patterns where vowel
height matched vowel height (HH) and patterns where there was a dependency between vowel
height and consonant voicing (HV). Moreton identifies two phonetic precursors that make
HV patterns phonetically grounded: the tendency for vocalic articulations to be exaggerated
before voiceless obstruents, and the pharyngeal-cavity expansion which occurs during the
production of voiced obstruents (Thomas 2000). Both precursors lead to a slightly lower
vowel F1 before a voiceless obstruent than before a voiced one (Moreton 2008). Moreton did
a survey of studies that measured vowel F1 in the relevant contexts for HH and HV patterns,
and found the phonetic precursors of both types of patterns to be comparable in size.
In Moreton’s study, the two patterns, HH and HV, are typologically asymmetrical. In a
typological survey, Moreton found HH patterns to exist in many language families, but only
a few marginal cases of HV patterns were found. An example of an HV pattern is found in
Polish, where /O/ raises to [o] before underlyingly voiced non-nasal codas. The productivity
2
of this pattern is doubtful, though (Sanders 2003).
Moreton then tested how well the two patterns, HH and HV, were learned in an artificial
grammar experiment. Moreton used phonotactic training rather than the alternation training
used by Finley (2010) and Finley and Badecker (2009). He trained the participants by having
them hear and produce a set of forms that conformed to the relevant phonotactic pattern.
All training stimuli had the form CVCV. Participants were trained on stimuli that exhibited
either HH harmony (/titu/, /titi/) or HV harmony (/tidO/, /tudæ/). They then heard novel
forms and had to decide if they were part of the language they had trained on. Moreton found
that the HH patterns were learned at higher rates than the HV pattern, despite the fact that
the two patterns are equally phonetically grounded. Moreton attributes this difference in
learnability to analytic bias based on the difference that HH patterns are agreement in the
same feature of the same type of segment, while HV patterns are assimilation of different
features between different segment types, one consonant and one vowel.
Moreton then conducted a second experiment that tested HV harmony against VoiceVoice (VV) harmony patterns. This comparison involves a substantive difference, because
the HV pattern is phonetically grounded while the VV pattern is not, as well as an analytic
difference, but both patterns are typologically rare. Moreton’s VV harmony is the same pattern that is used in the experiments described in this paper (see Section 3.1.2 for typological
examples). The stimuli were again CVCV and constructed from the same inventory. The
results of this second experiment showed that the VV patterns were learned better than the
HV patterns, but worse than the HH patterns from the first experiment.
2.2
Substantive Bias in Artificial Grammar Learning
The experiments in Carpenter (2010) are designed specifically to test substantive bias in the
learning of phonological patterns, by conducting an artificial grammar experiment teaching
two different stress patterns that were identical except for their phonetic grounding. Carpenter characterizes the following pattern as phonetically natural: stress leftmost low vowel,
else stress leftmost vowel. To create a formally similar but phonetically unnatural pattern,
she takes the opposite of her natural pattern: stress leftmost high vowel, else stress leftmost
vowel. The phonetic grounding of the natural pattern is based on the fact that when stress
is sensitive to vowel quality, it favors low vowels. Stress has been observed to seek high
sonority segments (de Lacy 2002), and low vowels are more sonorous than mid vowels, which
are more sonorous than high vowels (Clements 1990). Her two patterns are formally parallel
but differ in phonetic content, allowing the results of the experiment to isolate the effects of
substantive bias. Additionally, her natural pattern is rare but typologically attested, while
her unnatural pattern is completely unattested.
Participants were trained by listening to nonce words that exhibited the stress patter.
Then they were tested on choosing a correctly stressed word over an incorrectly stressed version of the same nonce word in a two-alternative forced-choice task. Carpenter’s participants
were both English speakers, a language with lexical stress, and French speakers, a language
with predictable stress. Carpenter added additional pre-training for French speaking participants in order to accommodate their “stress deafness” (Dupoux et al. 1997) and ensure
they were able to learn new stress patterns. Both English and French speakers learned the
phonetically grounded pattern significantly better than the ungrounded one, supporting the
involvement of a substantive bias in the learning of phonological patterns.
3
Wilson (2006) also aims to specifically test the effects of substantive bias in linguistic
pattern learning. In this experiment, participants were trained on a phonological pattern
and it was observed that they extended this pattern to novel contexts and sounds in ways
that accord with linguistic typology. Wilson taught a pattern of velar palatalization before
a vowel which was either a mid vowel or a high vowel. Velar stops preceding a vowel
are articulated further forward on the palate the further front the vowel is (Keating &
Lahiri 1993), making them more similar to palatoalveolar affricates before [i] than before
[e]. This difference in phonetic grounding between the two patterns that Wilson tested
correlates with linguistic typology in that palatalization before mid vowels asymmetrically
implies palatalization before high vowels in most attested languages. If participants were to
exhibit substantive bias in the learning of these patterns in the artificial grammar learning
experiment, they would be expected to generalize a pattern of palatalization before midvowels to high-vowels, but not the other way around.
The stimuli in this experiment were CVCV nonce words. Participants were asked to play
a language game where they were presented with two stimulus items, and then repeated the
second member of the pair in training. There were two experimental groups: participants
either were trained on velar palatalization only before mid-vowels (e.g. were given a pair
[ken@] [tSen@]) or only before high vowels (e.g. [gip@] [dZip@]). In testing, the participants
were asked to supply the second member of the pair themselves. In testing, the velar could
appear before any one of the vowels [i, e, A], but the participant would have only previously
encountered velarization in training before one of either [i] or [e].
Wilson’s results showed that participants in the high condition palatalized more often
before high vowels in testing than before mid and low vowels, and participants in the mid
condition palatalized equally often in all vocalic environments. Participants thus were able
to generalize a learned pattern of palatalization before a mid-vowel to palatalization before
a high vowel, but did not generalize from high to mid vowels. The generalizations go beyond
the training data, and they do so in a way that reflects the phonetic grounding of the
patterns. Wilson’s results thus support a substantive learning bias of the patterns.
The research summarized here has shown that participants in artificial grammar learning
settings can learn phonological patterns (Finley & Badecker 2009, Finley 2011). Research
has been done on what types of patterns can be learned and which are easier to learn than
others. It has been shown that patterns with certain analytic properties can be easier to
learn (Moreton 2008), and other research has identified different substantive patterns that
are better learned (Carpenter 2010, Wilson 2006). My experiments look at two patterns that
have not been looked at together closely before, anteriority harmony and voicing harmony.
These two patterns differ in phonetic content (only the feature being harmonized is different)
but are otherwise formally identical. Anteriority harmony has plausible phonetic grounding,
unlike voicing harmony. It is also widely typologically attested, whereas cases of voicing
harmony are rare or dubious. My results show that the anteriority harmony pattern was
learned better than the voicing harmony pattern, and thus support a substantive bias in the
learning of these patterns.
3
Anteriority vs. Voicing
In this section I lay out the typological attestation of anteriority harmony and voicing harmony, the two patterns that are compared in the experiments in this paper, and discuss their
4
phonetic grounding. Anteriority harmony is widely attested in the world’s languages, even
seen as the“prototypical” example of consonant harmony, while voicing harmony is very rare
and the few identified cases are marginal. The typological survey here is based on Hansson (2001)’s database documenting 120 languages with consonant harmony of various types.
Discussion of the phonetic grounding of the two patterns shows that anteriority harmony
across a vowel can be articulatory local, whereas voicelessness harmony cannot.
3.1
3.1.1
Typological Attestation
Anteriority Harmony
In Hansson’s database, coronal harmony is by far the most common type of harmony, and
Gafos (1996 [1999]) makes the even stronger claim that coronal harmony is the only possible
type of consonant harmony. Coronal harmony refers to assimilatory interactions between
coronals where the property involved is ‘minor place of articulation’ and has to do with the
configuration of the tongue tip and blade. Coronal harmony is attested for a wide range of
segment types including stops, affricates, fricatives, nasals, and liquids. The most common
type of coronal harmony is the one used in the below experiments, sibilant harmony.
Hansson calls sibilant harmony the“prototypical consonant harmony.” Sibilant harmony
systems make up about 1/3 of the entries in Hansson’s database, and the languages involved
belong to about 15 different language families distributed over four continents. The most
common contrast to participate in sibilant harmony is alveolar vs. post-alveolar, i.e. /s, z,
ts, dz/ vs. /S, Z, tS, dZ/.
Navajo sibilant harmony is one of the best known examples of consonant harmony. In
Navajo, sibilant harmony applies from right to left in an alternation involving prefixes and
roots. A sibilant in the root triggers assimilation to a target sibilant in the prefix, which will
then trigger assimilation in any additional preceding prefixes.
(1) Sibilant harmony in Navajo (data from McDonough 1991)
a. /j-iS-mas/
jismas
“I’m rolling along.”
b. /dz-iS-l-ta:l/
dZiSta:l
“I kick him [below the belt]”
Similar sibilant harmony can be found in other Southern Athapaskan languages (e.g.
Chiricahua Apache, Kiowa Apache), as well as in the Northern and Pacific Coast branches
of Athapaskan. Sibilant harmony as a root-internal cooccurrence restriction can be reconstructed as far back as Proto-Athapaskan-Eyak (Krauss 1964). In these languages, the
sibilant harmony is almost always anticipatory, occurring from right to left. In many cases
this is epiphenominal of the morphological structure. Sibilant harmony often occurs “inside
out,” or from the root to affixes, and the Athapaskan languages have primarily prefixing
morphology. Alveolar vs. post-alveolar sibilant harmony, both root-internal and involving
affixes, can also be found in many other native North American languages.
Sibilant harmony involving the alveolar vs. postalveolar distinction is also found in many
African languages. In the Afro-Asiatic family, it is independently attested in at least three
branches, and it also is found in various branches of the Niger-Congo family. The language
Koyra of the Omotic subgrouping of Afro-asiatic languages exhibits this harmony both root
internally as well as in alternation in affixes, where /s, z/ -> [S, Z].
5
(2) Harmony in Koyra Suffixes /-us/, /-os:o/, and /-es:e/
“cause to fear”
dZaS-uS
patS:-oS:o
“it became less”
dZaS-uS-eS:e
“let him/them frighten (s.o.)!”
In both the root internal and alternation examples of the sibilant harmony pattern in
Koyra, the two consonants can be separated by at most a vowel; harmony does not hold at
greater distances:
(3)
Sod-us
Sod:-os:o
“cause to uproot”
“he uprooted”
(*Sod-uS)
(*Sodd-oS:o)
But in other languages like Aari and Gimira there is no limit on the distances between
the trigger and the target consonants in sibilant harmony. (Hayward 1988, Breeze 1990).
‘Coronal harmony’ refers to interactions that involve ‘minor place of articulation’ of the
tongue tip and blade. Parameters that can be contrasted include tongue posture (apical vs.
laminal) and target region (dental vs. alveolar vs. postalveolar) (Hansson 2001). Several
combinations of these contrasts are attested in languages, including lamino-dental vs. alveolar, retroflex vs. lamino- alveolar, alveolar vs. retroflex, and alveolar vs. lamino-postalveolar.
Three-way anteriority contrasts also exist, with, for example dental vs. alveolar vs. postalveolar.
Anteriority harmony, especially harmony in the alveolar vs. postalveolar distinction, is
thus robustly attested in the world’s languages. It is independently attested in many different language families in different parts of the world, and makes up a large portion of all
cases of consonant harmony in Hansson’s database.
3.1.2
Voicing Harmony
While harmony of laryngeal features is relatively well attested, most of the cases involve
pulmonic vs. glottalic distinctions, while very few involve voicing. In addition, laryngeal
harmony where it does exist rarely extends across morpheme boundaries, so it is not involved in alternations. It most commonly manifests itself as a static root-level cooccurrence
restriction, such as in Ngizim, the one solid example of voicing harmony in Hansson (2001).
Hansson cites only two cases of laryngeal harmony resulting in alternations and they are both
voicing harmony, Yabem (Oceanic) and Kera, but both these cases turn out to be suspect.
In Ngizim, a West Chadic language, obstruents in a root are either both voiced or both
voiceless. The harmony is not sensitive to differences in stricture, so fricatives and stops are
both affected.
(4) Root-level laryngeal harmony in Ngizim (data from Schuh 1997)
kùt´@r
“tail”
tàsáu
“find”
d´@bâ
“woven tray”
z`@dù
“six”
6
The restriction in Ngizim is asymmetric in two ways, in that it is strictly anticipatory,
applying only from right to left, and it does not cause voiced obstruents to ever become
voiceless. Thus, examples with the form D...T exist, but not *T...D.
(5) Asymmetric character of Ngizim voicing harmony (data from Schuh 1997)
“roast”
bàkú
gùmtSí
“chin”
Yabem and Kera are the two languages given as examples by Hansson (2001) of laryngeal
harmony in alternation, but both are disputed as actual cases of voicing harmony. Yabem
is an Austronesian language of Papua New Guinea (Dempwolff 1939). In Yabem, tone and
voicing are directly correlated in obstruents: in high-toned syllables, stops are voiceless,
whereas in low-toned syllables, they are voiced. This is illustrated in the following minimal
pairs:
(6) a. minimal pair without obstruents
áwé
“outside”
àwè
“woman”
b. minimal pair containing obstruents
típ
“shell”
dìb
“speech”
In Yabem, when a prefix containing an obstruent combines with a root containing obstruents, the obstruents agree in voicing, but whether they are all voiced or all voiceless depends
on the tone in the root. For example:
(7) 1st Sg prefix /ka-/
a. Before high-toned root:
ká-táN
“I weep” (realis)
b. Before low-toned root:
gà-dèN
“I move towards” (realis)
However, this only holds with monosyllabic roots. Looking only at voicing, it would
appear the cooccurrence restriction is enforced only within a word final foot, and does not
reach across foot boundaries. Hansson concludes that there is no need to describe this pattern
in Yabem as voicing harmony at all. Since obstruent voicing is completely predicable based
on tone, the facts in Yabem could be described as exhibiting foot-bounded tone harmony
instead. This is a preferable analysis because tone spread and tonal harmony are well-attested
phenomena (Goldsmith 1976).
In Kera, if a root contains a plosive stop or affricate, then plosives in any affix (both
prefixes and suffixes) will agree in voicing with the root plosive.
(8) Laryngeal harmony alternations in Kera (Ebert 1979)
a. nominal prefix /k-/
“cooking pot (plur.)”
k@-taatá-w
g@-dàar`@
“friend”
7
b. feminine suffix /-ka/
“black (fem.)”
sár-ká
dZàr-gá
“colorful (fem.)”
The laryngeal harmony in Kera is parasitic on identity in stricture, i.e. plosive voicing
does not interact with the voicing of fricatives or sonorants. The harmony thus can be nonlocal across a vowel or other segment that does not participate in the harmony. The voicing
harmony in Kera is asymmetric in that voiceless plosives become voiced but voiced plosives
will not become voiceless.
Pearce (2006) gives evidence against this pattern in Kera being voicing spread at all. She
presents exceptions to the voicing agreement, and argues that the data can be accounted
for by tone spreading and the resultant change in VOT. In a statistical analysis of a lexicon
constructed from her corpus of collected Kera speech, Pearce finds that 9% of all words in
the lexicon have a voicing mismatch. She gives examples such as the following:
(9) Kera nominal prefix /k-/
k-ágày
“hoe (plur.)”
k-ágàmlà
“bull (plur.)”
The words in (9) have voiced obstruents in their roots, so the voicing spread account
predicts that the prefixes should also be voiced, as seen in (8a). However, a tone-spreading
account would predict that since the first syllable in the root has a high tone, the prefix
as the onset of that syllable will have long VOT, making it perceived as voiceless. The
facts in (9) support the second account, the explanation of tone spreading rather than voice
spreading.
Thus, there are very few cases of voicing harmony attested in the world’s languages.
Ngizim is the only example Hansson gives that remains considered a voicing harmony pattern.
It only applies to voiceless consonants becoming voiced and is also strictly anticipatory, while
there are several cases of anteriority harmony that are totally symmetric and exceptionless.
It is clear that while anteriority harmony is extremely robustly attested, cases of voicing
harmony are vanishingly rare.
3.2
Phonetic Grounding
While phonological grammar carries out formal manipulations of abstract symbols, phonetics
is grounded in human capabilities for speech production and speech perception. Evaluating
phonological concepts based on facts about the physical world gives us an idea of the phonetic grounding of a given phonological pattern. Depending on the phonological theory (for
example, taking a feature matching view of harmony), the patterns of anteriority harmony
and voicing harmony may be formally identical and a difference in their implementation
would not be expected. However, anteriority and voicing are not the same with regards to
physical production in the vocal tract, and so they are not both phonetically grounded to
the same extent.
For a pattern to be consonant harmony, at minimum, an intervening vowel must be
present and must not be audibly affected by the assimilating feature. How this is possible
for anteriority harmony is clear. Anteriority is articulated using the tongue tip and blade,
and this gesture can be carried through a vowel without changing its perception. In the
following figure, representing a simplified gestural representation of the utterance [sas], the
8
tongue tip can have a constriction location of alveolar throughout the utterance, because
this gesture does not affect the percept of the vowel. If the utterance were [SaS], then the
constriction location of the tongue tip would be postalveolar, but the concept would be the
same
Figure 1: Simplified gestural score of an example of anteriority harmony
Gafos (1996 [1999])’s theory of articulatory locality says that consonant harmony can
only involve articulatory specifications controlling the shape of and orientation of the tongue
tip and blade, since only those can pass through intervening vowels and consonants without
interfering with their acoustic properties or articulation. In Gafos’ view, all harmony is
driven by the desire to minimize the number of gestures required to produce an utterance.
Thus, Gafos predicts that coronal harmony with distinctions such as dental vs. alveolar,
alveolar vs. postalveolar, and apical vs. laminal, is the only type of consonant harmony
possible at all.
It is impossible, however, to produce an output such as [pap], with voiceless consonants on
either side of a voiced vowel, with only one gesture producing the voicing values. Voicing is
represented by opening and closing gestures of the glottis. At the beginning of the utterance,
the glottis is open, and then from /p/ to /a/ the glottis closes. From /a/ to the second /p/
there is a second opening gesture of the glottis. In this case one gesture cannot be held
through the whole utterance without affecting the percept of the vowel.
Figure 2: Simplified gestural score of an example of voicing harmony
The main difference between the articulation of patterns like those in Figure 1 and Figure
2, is that it is possible to produce anteriority harmony with one gesture across a vowel, but it
9
is necessary to have two separate opening gestures of the glottis to produce the utterance with
harmony of [−voice]. In this way, anteriority harmony minimizes the number of articulatory
gestures necessary and thus is more phonetically motivated than a voicing harmony pattern,
which requires more gestures.
4
The Experiments
In this section I will describe the artificial grammar learning experiments that were conducted
to test and compare the learnability of anteriority harmony and voicing harmony. Then
I will present and discuss the results of the experiments. The two experiments exposed
adult learners to an artificial language exhibiting a consonant harmony pattern involving
harmony in either voicing or anteriority. I used the alternation training method of Finley
and Badecker (2009) because it has yielded stronger effect sizes in past research than has
phonotactic training like that in Moreton’s (2008) experiments. Participants were then tested
on whether they had learned this pattern.
4.1
Stimuli
The stimuli were stem items with the structure CVCV that each had 2 corresponding suffixed
forms with the structure CVCVCV. All stimuli were nonce words created by combining a
set of 3 consonants in the C1 slot and 6 consonants in the C2 slot, with 4 vowel patterns.
C1 and C2 were never identical, and V1 and V2 were never identical. There was a total of
72 stems that could take each of 2 different suffixes, resulting in 216 total stimulus items for
each experiment.
For the anteriority experiment, the C2 , or trigger of the harmony in the suffix, was one
of 6 sibilants: 3 anterior sibilants [s, z, ts] and 3 posterior sibilants [S, Z, tS]. The C1 in
all items was outside the harmony paradigm, meaning they were chosen so as not to have
a counterpart in English phonology that would exhibit the feature contrast salient to the
harmony being tested. C1 s for the anteriority experiment were voiceless stops: [p,t,k]. The
suffix was either -sa or -Sa. The stimuli were thus items like [pasi], which when suffixed would
be [pasisa], or [tiSu] which would be [tiSuSa] when given a harmonic suffix.
For the voicing experiment, C2 was one of 6 stops: 3 voiced stops [b, d, g] and 3 voiceless
stops [p,t,k]. The C1 for the voicing experiment was one of [m, n, l]. The suffix was either
-pa or -ba. Both experiments used the vowels [a, i, o, u] in four configurations so that each
vowel was V1 and V2 in one configuration. These vowel patterns were: [a, i], [i, u], [o, a],
and [u, o]. Stimuli for the voicing experiment were thus items like [madi] which when given
a harmonic suffix would be [madiba], or [laki] which would become [lakipa] when suffixed
harmonically.
Each experiment also had 24 filler stems that also took both applicable suffixes for the
paradigm. The fillers for both experiments used the same 4 vowel configurations and C1
consonants as the experimental stimuli, but the C2 were replaced with consonants that could
not participate in the relevant harmony. The anteriority experiment fillers used [p,t,k,m,n,l]
as C2 , producing filler stems like [pami] and [toka]. The voicing fillers used [m, n, l, w, j] for
C2 , producing filler stems such as [nali] and [mowa].
All stimuli were recorded in a sound booth on a Zoom H4 recorder, by a native Hebrew
speaker who was unaware of the purpose of the experiment. A Hebrew speaker was chosen
10
Posterior Trigger
Anterior Trigger
because Hebrew has a canonical five vowel inventory with no vowel reduction and a contrast
between fully voiced and voiceless unaspirated stops. The speaker read each word 3 times
from a printout, and was instructed to put stress on the first syllable of each item. Filler
words were added to the beginning and end of each column, so as to avoid prosodic effects
associated with list reading affecting experimental stimuli.
Table 1 contains all the anteriority experiment stimuli for the vowel pattern of [a,i] with all
combinations of consonants and suffixes. Table 2 shows the same for the voicing experiment.
Stem
pazi
tazi
kazi
pasi
tasi
kasi
patsi
tatsi
katsi
Stem
paZi
taZi
kaZi
paSi
taSi
kaSi
patSi
tatSi
katSi
Harmonic (-sa)
pazisa
tazisa
kazisa
pasisa
tasisa
kasisa
patsisa
tatsisa
katsisa
Harmonic (-Sa)
paZiSa
taZiSa
kaZiSa
paSiSa
taSiSa
kaSiSa
patSiSa
tatSiSa
katSiSa
Disharmonic (-Sa)
paziSa
taziSa
kaziSa
pasiSa
tasiSa
kasiSa
patsiSa
tatsiSa
katsiSa
Disharmonic (-sa)
paZisa
taZisa
kaZisa
paSisa
taSisa
kaSisa
patSisa
tatSisa
katSisa
Table 1: Examples of stimuli in the anteriority experiment
4.2
Design
Both experiments had one control group and one critical group. Participants were taught
how to make the plural form of a word in a made-up language by being shown a picture
of a single item with a bare stem and then two of that item with a suffixed form. In the
critical groups, the suffixed forms were always harmonic; in the control groups they were
half harmonic and half disharmonic.
The experiment consisted of a training phase and then a testing phase. The critical
groups were trained on stems and stem+suffix pairs exhibiting consonant harmony. For
example, in the anteriority experiment, if a critical participant were shown the stem [pazi],
then they would also be shown the suffixed form [pazisa], where the C2 and C3 share the
same value for the feature of anteriority, but not the suffixed form [paziSa], in which C2
and C3 do not share the same value for anteriority. In the voicing experiment, a critical
participant would be shown the stem [madi] with the suffixed form [madiba], where C2 and
C3 have the same value for voicing, but never [madipa]. The control groups were not shown
11
Voiceless Trigger
Voiced Trigger
Stem
mapi
napi
lapi
mati
nati
lati
maki
naki
laki
Stem
mabi
nabi
labi
madi
nadi
ladi
magi
nagi
lagi
Harmonic (-pa)
mapipa
napipa
lapipa
matipa
natipa
latipa
makipa
nakipa
lakipa
Harmonic (-ba)
mabiba
nabiba
labiba
madiba
nadiba
ladiba
magiba
nagiba
lagiba
Disharmonic (-ba)
mapiba
napiba
lapiba
matiba
natiba
latiba
makiba
nakiba
lakiba
Disharmonic (-pa)
mabipa
nabipa
labipa
madipa
nadipa
ladipa
magipa
nagipa
lagipa
Table 2: Examples of stimuli in the voicing experiment
any pattern in the items they were exposed to, that is, a stem like [pasi] had an equal chance
of being presented with either [pazisa] or [paziSa]. The critical groups in each experiment
were presented with all harmonic suffixed forms, while the control group saw 50/50 harmonic
and disharmonic pairs. The control training was also balanced for how many -sa endings and
how many -Sa endings a participant saw in the anteriority experiment, or how many -pa vs.
-ba endings a control participant saw in the voicing experiment. See Table 3 for an example
of the distribution of stimuli seen in each condition of the voicing experiment.
In the testing phase, each participant in all four conditions was asked to choose which
form of an item is correct in the language they were trained on. They were given a forced
choice between a harmonic form and a disharmonic form. All testing items were new to the
participants; none had been seen before in the training phase, but they contained the same
segments as the training items.
4.3
Procedure
The experiment was run using Experigen, an online platform created by Becker and Levine
(2010). Participants were told that they were going to learn words from an alien language,
and how to form the plural of those words. They were told that they need not memorize
words, but to pay attention to how the plural is formed, because later they would have to
do it themselves.
12
Critical Condition
Control Condition
Stem
mapi
napi
lapi
mati
nati
lati
maki
naki
laki
mabi
nabi
labi
madi
nadi
ladi
magi
nagi
lagi
mapi
napi
lapi
mati
nati
lati
maki
naki
laki
mabi
nabi
labi
madi
nadi
ladi
magi
nagi
lagi
Voicing
Plural (Suffixed)
mapipa
napipa
lapipa
matipa
natipa
latipa
makipa
nakipa
lakipa
mabiba
nabiba
labiba
madiba
nadiba
ladiba
magiba
nagiba
lagiba
mapiba
napipa
lapiba
matipa
natiba
latipa
makiba
nakipa
lakiba
mabiba
nabipa
labiba
madipa
nadiba
ladipa
magiba
nagipa
lagiba
Table 3: Examples of which stimuli are presented during training to participants in the critical and
control conditions of the voicing experiment. Bold represents harmonic forms.
4.3.1
Training Phase
During the training phase, participants were shown pictures with associated lexical items.
The stimuli were presented in audio form as well as orthographically. I used standard English
13
spellings for the phonemes that have an unambiguous letter associated with them, and to
represent [Z], the only phoneme without a canonical English spelling, I used “zh.”
First, a picture of a single item appeared with the phrase “This is a
” above it. A
stem form of a stimulus item was in the blank, (e.g. “tucho”) (See Figure 3). An audio play
button accompanied each stimulus item, and it was necessary to press the button and listen
to the audio file before proceeding.
Figure 3: Training Trial
Next, the participants were shown a picture of two of that same item with the phrase
“These are
” above it. A suffixed form of the stimulus item was in the blank (e.g.“tuchosha”).
The goal was to teach the participant how to form the plural of a word in this language by
adding one of the two possible suffixes (See Figure 4). The pictures were paired randomly
with the stimulus items, and the same words and picture were not associated with each other
for all participants.
14
Figure 4: Training Trial
Each participant was shown 26 such pairs of pictures, 20 including environments where
harmony could occur, and 6 fillers. The set of stimuli that each participant in each group
saw was randomly selected from the total set of stimuli, so each participant saw them in a
different order, and a different breakdown between which stimuli were seen in training and
which in testing.
4.3.2
Testing Phase
Once the training phase was over, the participant was told that now it was their turn to
provide the plural forms. 42 trials followed, 36 with potential harmony environments, and
6 fillers. During each trial, participants were first shown a picture of a single object and
its associated lexical item, as in the training phase. But then the participant was shown
the picture of two of that item and asked to provide the plural form of the lexical item.
15
They were given a two-alternative forced choice test. During each trial, the participant was
presented a harmonic form and a disharmonic form (e.g. “Are these pasisa or pasisha?”)
and asked to choose which was correct in the language they had just learned. Again, the
participant had to listen to the audio file of each stimulus before buttons appeared at the
bottom to make their choice (See Figure 5).
Figure 5: Testing Trial
16
4.4
Participants
The participants were 400 adult native English speakers (100 participants per each of the
4 experimental conditions), recruited through Amazon Mechanical Turk (AMT), Amazon’s
crowdsourcing website (www.mturk.com) and paid $0.75 to complete the experiment, which
took about 10 minutes. AMT has been shown to produce results to a linguistic study that are
indistinguishable from an identical study conducted in a laboratory setting (Sprouse 2010).
The participants were told that they must be native speakers of English to do the task,
and were asked at the end in a brief demographic survey to report their native language.
Additionally, participants were restricted to those which AMT detected as having an IP
address in the United States. Responses were discarded if the participant self-reported as a
native speaker of any language other than English. English does not exhibit a categorical
pattern of consonant harmony across a vowel, and any other bias that English speakers may
have from other subcategorical properties of the language are controlled for by the control
conditions. 5 participants were excluded before data analysis because they had answered
with the same ending for 100% of their testing trials. This leaves the total N=395 for the
results presented below.
5
Results
The responses from each participant’s testing phase were coded for whether they were harmonic or not, i.e. whether the ending chosen for the stem agreed with the trigger (C2 ) for
the feature at hand, either voicing or anteriority. For each participant, there were 36 binary
data points.
A Mixed Logit Model was fit1 using the lmer() function in the lme4 package (Bates &
Maechler 2010) for the R software (http://www.R-project.org.) The dependent variable was
whether the response was harmonic or not. The predictors in the model were feature (anteriority or voicing) and condition (critical or control), and the interaction between condition
and feature. Random intercepts were included for participant and stimulus stem. Random
slopes were not included for subjects, because subjects only participated in one treatment
group each. Random slopes by stimulus item were also not included since the same items
were not used in all four conditions (though the same items were used for the control and
critical training conditions for a single feature). There is a significant main effect of condition
and a significant interaction between condition and feature. The full model is given in Table
4.
intercept
condition
feature
condition:feature
Estimate
−0.253
2.135
0.170
−1.332
Standard Error
0.1257
0.184
0.263
0.322
Wald’s z
−2.013
11.626
0.648
−4.136
p
0.044
< .0001
0.517
< .0001
Table 4: Results of a Mixed Logit Model with dependent variable of harmonic responses, testing for
main effects of condition (baseline=control) and feature (baseline=anteriority) and an interaction
between condition and feature.
1
lmer(correct∼condition+feature+condition:feature+(1—userCode)+(1+condition—root), family=“binomial”)
17
The results are shown graphically in Figure 6. The main effect of condition is significant,
because in both experiments, the participants in the critical group gave a higher proportion
of harmonic answers than did those in the control groups (p < .0001, 76.4% vs. 44.6% in the
anteriority experiment, and 63.0% vs. 48.0% in the voicing experiment). The main effect
of feature was not significant, showing that the proportion of overall harmonic responses
was not affected by the harmonic feature. The significant interaction between condition and
feature (p < .0001) results from the larger difference between the performance of the control
and critical groups in the anteriority experiment compared to the voicing experiment. The
anteriority experiment has a difference between the two conditions of 31.8% while the voicing
experiment has a difference of 15%. Participants learned the anteriority pattern better than
the voicing pattern, and the improvement in harmonic responses from the control group to
the critical group was twice as big in the anteriority experiment as in the voicing experiment.
Figure 6: Proportion of harmonic responses given in each experiment, by condition. Error bars
indicate the standard error of the mean.
The data from the critical groups of both experiments were also analyzed by feature value
of the trigger consonant, comparing the proportion of correct harmonic responses that were
[−feature] and [+feature]. For example, in the anteriority experiment, answering [pazisa] to
the given stem [pazi] is an instance of correctly providing harmony of the [+anterior] feature,
while giving the plural [patSiSa] to the given stem [patSi] is also a harmonic response, but
it matches on the value of [−anterior]. Analogously, [padiba] is harmonic on the value of
[+voice], while [patipa] is harmonic on the value of [−voice].
To test for any differences between feature values, another Mixed Logit Model was fit
18
to just the data from the critical conditions, not the control data. The dependent variable
was whether the response was harmonic or not. The predictors in the model were feature
(anteriority or voicing) and feature value of the trigger consonant ([+feature] or [−feature]),
and the interaction between feature and feature value. Random intercepts were included for
participant. There is a significant main effect of feature, but no significant effect of feature
value. The interaction between feature and feature value was significant. The full model is
given in Table 5.
intercept
feature value
feature
feature value:feature
Estimate
2.254
0.030
−1.10
−0.710
Standard Error
0.212
0.098
0.290
0.124
Wald’s z
10.625
0.308
−3.779
−5.712
p
< .0001***
0.758
.0002 ∗ ∗
< .0001***
Table 5: Results of a Mixed Logit Model with dependent variable of harmonic responses, testing
for main effects of feature value (baseline=[−feature]) and feature (baseline=anteriority) and an
interaction between feature value and feature.
The results are shown graphically in Figure 7. As expected, in this analysis the main effect
of feature is significant, because overall participants in the critical anteriority experiment gave
more harmonic responses than those in the critical voicing group, across both [+feature] and
[−feature] cases. This is consistent with the main analysis presented above. The main effect
of feature value was not significant, and there is no reason to believe that [+voice] and
[+anteriority] would pattern together or that [−voice] and [−anteriority] would do so either.
Both anteriority and voicing are included in the same model here strictly to compare the
trend between plus and minus feature between the two experiments. The interaction between
feature and feature value was significant, because there was no difference between [+anterior]
and [−anterior] but there was a significant difference between [+voice] and [−voice]. Of all
the trials where the trigger consonant was [−voice], 69.2% were answered correctly, but only
56.8% of the [+voice] trials were answered correctly.
19
Figure 7: Proportion of harmonic responses given based on whether the trigger consonant is [+feature] or [−feature].
Data from the critical training conditions were further examined for an effect of specific
trigger consonant. The data from the critical training condition of the anteriority experiment
was analyzed separately from that of the critical training condition of the voicing experiment.
This analysis was done to see if any aggregate results were due to particular triggers behaving
differently from the others, particularly to see if the learnability could have been attributed
to identity effects. It might have been easier for participants to choose the harmonic response
when the trigger consonant was identical to the target consonant (e.g. [pasisa]) as opposed
to when the target was harmonic with the trigger but not identical (e.g. [pazisa]). Many
languages distinguish between identical and non-identical segments in phonotactic patterns,
and identity has previously been assumed to be a property that phonological constraints
can reference (MacEachern 1999; Coetzee & Pater 2008). If there were identity effects, then
the triggers [s] and [S] in the anteriority experiment, and triggers [b] and [p] in the voicing
experiment would show significantly higher rates of correct answers than the other triggers
in their respective experiments, as participants would be matching identical segments better
than ones that only agree in the feature.
This analysis also illustrated whether the trigger [ts] behaved differently than other consonants, which might be expected because it only appears in English at the end of words
and across morpheme boundaries and thus is not a phoneme of English, and also whether
[Z] behaved differently because it is an uncommon phoneme in English.
Table 6 shows the results of a Mixed Logit Model with the dependent variable being
20
harmonic responses and the independent variable being trigger consonants from the anteriority experiment [s, z, ts, S, Z, tS]. The six level factor was dummy coded, with [tS] as the
baseline, and the other five consonants are compared to it. Random intercepts were included
for participants.
intercept ([tS])
[s]
[S]
[ts]
[z]
[Z]
Estimate
2.437
0.528
0.315
0.163
0.263
0.540
Standard Error
0.327
0.175
0.171
0.173
0.173
0.175
Wald’s z
7.452
3.024
1.841
0.945
1.522
3.080
p
< .0001***
0.003*
0.066
0.345
0.128
0.002*
Table 6: Results of a Mixed Logit Model with dependent variable of harmonic responses, testing
for main effect of trigger consonant.
There was a significant difference (p< .01) between [tS], which had the lowest rate of
harmonic responses, and the two consonants that had the highest rates of harmonic responses
[s] and [Z]. The model found no other significant results. The results for the anteriority trigger
consonants are presented graphically in Figure 8.
Figure 8: Harmonic responses by trigger consonant for the anteriority experiment
The results show that matching on identity of the segment does not across the board
make it easier to choose a harmonic response, as here [s] as a trigger consonant got slightly
21
higher rates of harmonic responses (80%) than did the other two anterior consonants, but the
differences were not significant per a Tukey’s posthoc test. [S] actually did worse (76%) than
[Z] (80%) and not significantly different from [tS] (73%). The graph in Figure 8 also shows
that the trigger [ts] does not behave significantly differently than the other five triggers
in the experiment, which could be expected due to evidence that [ts] is not treated as a
single affricate phoneme by native English speakers. Similarly, there is no aberrant behavior
observed for the consonant [Z] either (there is only a significant difference between [Z] and
[tS], but [Z] actually performs better, p=.025), even though it is an uncommon sound in
English. This serves as a check that it was valid to use [ts] and [Z] in the experiment as part
of the sibilant series. The two affricates [ts] and [tS] have the lowest percentages of harmonic
responses given.
Table 7 shows the results of a Mixed Logit Model with the dependent variable being
harmonic responses and the independent variable being trigger consonants from the voicing
experiment [b, d, g, p, t, k]. The six level factor was dummy coded, with [b] chosen by the
model as the baseline, and the other five consonants compared to it. Random intercepts
were included for participants.
intercept ([b])
[d]
[g]
[k]
[p]
[t]
Estimate
0.725
−0.525
−0.466
0.262
0.389
0.317
Standard Error
0.144
0.128
0.129
0.133
0.133
0.133
Wald’s z
5.029
−4.095
−3.623
1.980
2.925
2.389
p
< .0001***
< .0001***
< .001**
0.048
0.003*
0.017
Table 7: Results of a Mixed Logit Model with dependent variable of harmonic responses, testing
for main effect of trigger consonant.
The model found that [b] had significantly more harmonic responses than [d] with a
p < .0001 and significantly more than [g] with a p < .001. The trigger consonant [b] was also
found to have significantly fewer harmonic responses than the trigger consonant [p] with a
p < .01. The results for the voicing trigger consonants are presented graphically in Figure 9.
22
Figure 9: Harmonic responses by trigger consonant for the voicing experiment
[b] got significantly more harmonic responses (63%) as a trigger than did [d] (54%) or
[g] (55%), which is consistent with an identity advantage, but the same is not true for the
voiceless consonants. [p] did not get more correct responses than [t] or [k] (all three were
around the same, between 68-70% correct), which is not consistent with an identity effect.
All three voiced consonants did worse than the voiceless consonants, so the effect in Figure
7 above, where the [−voice] triggers have significantly higher rates of harmonic responses
than the [+voice] triggers, is not due to disproportionate effects of only one trigger in the
category.
6
Discussion
The results of the experiments show that participants in the critical training groups learned
both anteriority harmony as well as voicing harmony, but that the anteriority pattern was
learned better. As seen in Figure 7, there was no significant difference between [+anterior]
and [−anterior] triggers, but [−voice] triggers got significantly more harmonic responses than
did [+voice] triggers. The difference in results here between [+voice] and [−voice] triggers
is actually contra the phonetic grounding, as harmony of [+voice] is phonetically grounded
because the intervening vowel is also voiced, but harmony of [−voice] is not grounded. The
typological facts reflect this, with patterns like the one in Ngizim being asymmetrical, only
applying to harmony of the [+voice] vowel.
The results broken down by specific trigger consonant show that [Z], which is an un23
common segment in English, and [ts], which is not a phoneme in English, do not behave
significantly differently than the other consonants. It also does not seem to be easier to
choose a harmonic response when the target consonant is identical to the trigger consonant
than when they are non-identical but share a feature. Place of articulation did not have a
consistent affect on harmonic responses in the voicing experiment.
English does have at least two phonological patterns that involve voicing assimilation,
the plural -s morpheme which takes forms /s/ and /z/, and the past tense -ed morpheme,
which takes forms /t/ and /d/. In both cases, the underlying consonant assimilates to
the preceding consonant in voicing, i.e./ kæts/ vs /dOgz/ and /klæpt/ vs /taImd/. This
pattern does not qualify as harmony in the sense that Hansson (2001) and Gafos (1999)
are dealing with, because it does not have, at minimum, an intervening vowel between the
two harmonizing consonants. However, it’s still possible to imagine that this familiarity
with voicing agreement could affect English speaker’s performance in the experiments, since
there is no comparable pattern in English involving anteriority. However, my results do not
show any effect of bias towards voicing harmony in the control condition, and the poorer
performance on the voicing condition goes in the opposite direction as predicted if native
phonology is a main factor.
Another way that the features of anteriority and voicing are different is that anteriority
is clearly a binary feature with [+anterior] and [−anterior] values, but it has been argued
that VOICE is privative, and that [−voice] does not actually exist. This view could be
corroborated by the typological evidence that in the few patterns that have been described
as voicing harmony, it is common for the system to be asymmetric such that only voiceless
consonants become voiced and not the other way around, such as in Ngizim and Kera
described above. This would make sense if [−voice] didn’t exist and it was just the presence
or absence of a privative VOICE participating in the assimilation. What this difference in
feature binarity would mean for the comparison of the learnability of anteriority or voicing
harmony is unclear, and the effects could be confounded with the lack of phonetic grounding
of [−voice] harmony because of the articulatory gestures involved. As discussed above, in
my voicing harmony experiment, the participants actually did better on choosing harmonic
forms when the trigger was [−voice] than when it was [+voice], which is not inline with
the typological data, the phonetic grounding, or what predicted effects of feature privativity
might be.
7
Conclusion
The results of the experiments in this paper show a clear difference in how well participants
learned anteriority harmony and voicing harmony. Both patterns were learned, but anteriority harmony was learned significantly better. This is in line with the phonetic grounding
of the two patterns. Anteriority is a feature produced by the tongue tip and tongue body,
and this gesture can be held through a vowel without affecting the percept of the vowel.
The glottal opening and closing gestures that produce voicing do affect the production of
the vowel, and an open glottis (required for voiceless consonants) cannot be held through a
vowel without making it voiceless too. Because of this, anteriority is phonetically grounded
with respect to consonant harmony across an intervening vowel, while voicing harmony is not.
In addition, the typological facts show that anteriority harmony is robustly attested cross
linguistically, and there are many cases of fully symmetric and nearly exceptionless patterns.
24
Voicing harmony, on the other hand, is scantily attested and the few existing examples are
marginal. While my results do not suggest any causal link in any direction between these
three facts, it remains that the learnability of anteriority harmony over voicing harmony
found in my experiments is in line with both the phonetic grounding and typological facts.
The results in Figure 6 show that it is not the case that voicing harmony wasn’t learned at
all. Both patterns were learned at significantly higher rates in the critical condition than in
the control condition, but anteriority harmony was learned much better. In her paper about
naturalness bias, Carpenter posits two separate cognitive-mechanisms for pattern leaning,
one that is general and one that is language-specific. In this view, the general mechanism is
used for all pattern learning, but when natural linguistic substance is involved, the languagespecific learning mechanism is also engaged, doing the additional work that exhibits itself in
higher rates of successful learning of the pattern. My results would fit with such a model,
where the general pattern learning mechanism allowed voicing harmony to be learned to an
extent, but a specific language learning mechanism privileged anteriority harmony (whether
because of its phonetic grounding or some other reason) as a linguistic pattern rather than a
non-linguistic abstract pattern, and allowed the pattern to be learned more quickly or better
by the participants.
This paper shows that adult native English speakers show a bias towards anteriority
harmony, but there are several explanations for why this might be. An analysis of the English
lexicon could be performed to find out if there are any statistical biases in the language that
could influence speakers to be more familiar with, or do better at learning, the anteriority
harmony patterns. Since the quality of the headphones that participants (recruited through
Amazon Mechanical Turk) used could not be controlled, the experiment could be duplicated
in a laboratory environment to see if the same learning asymmetry would occur with better
quality equipment. The experiment could also be repeated with the stimuli recorded by an
English speaker, to see if participants were influenced by the non-native voicing distinctions
produced by the Hebrew speaker used in this study. In these experiments, participants were
asked to listen to stimuli and perform a forced-choice task. For further research in this area,
a production study could be undertaken to measure the learnability of the patterns when
participants are asked to produce the harmonic forms. Additionally, the study could be
duplicated using a phonotactic training method to see if the patterns are still learned to the
same degree. Different amounts of training can be provided to find at what point the learner
has enough exposure to learn the patterns.
25
References
Bates, Doug and Martin Maechler (2010). lme4: Linear mixed effects models using S4
classes. Available at http://cran.r-project.org/web/packages/lme4/index.html
Becker, Michael & Jonathan Levine (2010). Experigen: an online experiment platform.
Available at https://github.com/tlozoot/experigen
Breeze, Mary J. (1990). A sketch of the phonology and grammar of Gimira (Benchnon). In
Hayward (1990c), pp. 1-67
Carpenter, Angela C. (2010). A naturalness bias in learning stress. Phonology 27: 345-392.
Clements, G. N. (1990). The role of the sonority cycle in core syllabification. In John
Kingston & Mary E. Beckman (eds.) Papers in laboratory phonology I: between the grammar and physics of speech. Cambridge : Cambridge University Press. pp 283-333.
Coetzee, Andries and Joe Pater (2008). Weighted constraints and gradient restrictions on
place cooccurrence in Muna and Arabic. Natural Language and Linguistic Theory 26: 289337.
de Lacy, Paul (2002). The formal expression of markedness. PhD dissertation. University
of Massachusetts, Amherst.
Dempwolff, Otto (1939). Grammatik der Jabm-Sprache auf Neuguinea. Hamburg: Friederichsen, de Gruyter & Co.
Dupoux, Emmanuel, Christophe Pallier, Núria Sebastián and Jacques Mehler (1997). A
destressing deafness in French? Journal of Memory and Language 36: 406-421.
Ebert, Karen (1979). Sprache und Tradition der Kera (Tschad), Teil III: Grammatik. (Marburger Studien zur Afrika- und Asienkunde, Serie A: Afrika, 15.) Berlin: Reimer.
Finley, Sara and William Badecker (2009). Artificial language learning and feature-based
generalization. Journal of Memory and Language 61: 423-437.
Finley, Sara (2011). The privileged status of locality in consonant harmony. Journal of
Memory and Language 65: 74-83.
Gafos, Adamantios I. (1996). The articulatory basis of locality in phonology. Doctoral dissertation. Johns Hopkins University. [Published 1999, New York: Garland.]
Goldsmith, John (1976). Autosegmental Phonology. Doctoral dissertation. MIT. [Published
1979, New York: Garland.]
26
Hansson, Gunnar (2001). Theoretical and Typological Issues in Consonant Harmony. Doctoral Dissertation. University of California, Berkeley.
Hayward, Richard J. (1988). Remarks on the Omotic sibilants. Cushitic-Omotic: Papers
from the International Symposium on Cushitic and Omotic languages, Cologne January
6-9, 1986, ed. by M. Bechhaus-Gerst & F. Serzisko, pp. 263-99. Hamburg: Helmut Buske.
Keating, P. and A. Lahiri (1993). Fronted velars, palatalized velars, and palatals. Phonetica
50: 73-101.
Krauss, Michael (1964). Proto-Athapaskan-Eyak and the problem of Na-Dene: The phonology. International Journal of American Linguistics 30: 118-31.
McDonough, Joyce (1991). On the representation of consonant harmony in Navajo. Proceedings of WCCFL 10: 319-35.
MacEachern, Margaret (1999). Laryngeal co-occurrence restrictions. New York: Garland.
Moreton, Elliott (2008). Analytic bias and phonological typology. Phonology 25: 83-127.
Pearce, M (2006). Evidence against voicing spread in Kera. UCL Working Papers in Linguistics 18: 345-355.
Sanders, Nathan (2003). Opacity and sound change in the Polish lexicon. PhD Dissertation.
UCSC.
Schuh, Russell G. (1997). Changes in obstruent voicing in Bade/Ngizim. Ms. Thesis University of California, Los Angeles.
Sprouse, Jon (2011). A validation of Amazon Mechanical Turk for the collection of acceptability judgments in linguistic theory. Behav Res 43: 115-167.
Thomas, Erik R. (2000). Spectral differences in /ai/ offsets conditioned by voicing of the
following consonant. Journal of Phonetics 28: 1-25.
Wilson, Colin (2006). Learning Phonology With Substantive Bias: An Experimental and
Computational Study of Velar Palatalization. Cognitive Science 30: 945-982.
27