PDF 749k - Signata

Signata
Annales des sémiotiques / Annals of Semiotics
6 | 2015
Sémiotique de la musique
Melodic Primitives and Semiosis
José Roberto do Carmo Jr. and Thiago Corrêa de Freitas
Publisher
Presses universitaires de Liège (PULg)
Electronic version
URL: http://signata.revues.org/1054
ISSN: 2565-7097
Printed version
Date of publication: 31 December 2015
Number of pages: 23-41
ISBN: 978-2-87562-087-3
ISSN: 2032-9806
Electronic reference
José Roberto do Carmo Jr. and Thiago Corrêa de Freitas, « Melodic Primitives and Semiosis », Signata
[Online], 6 | 2015, Online since 31 December 2016, connection on 29 March 2017. URL : http://
signata.revues.org/1054
Signata - PULg
PrimitiveS
Melodic Primitives and Semiosis
José Roberto do Carmo Jr.
hiago Corrêa de Freitas
Curso Superior de Luteria - UFPR
1. Primitives
Some of the biggest unanswered questions about the meaning of music concern
the nature of musical primitives: what they are, how many they are, what their role
in semiosis is. Obviously, this is very far from being a simple problem and would
require a signiicant efort even to begin to igure out the nature of its solution,
if there is one. For this reason, the discussion proposed in this essay addresses a
much more restricted subject. We will look at melody, particularly at tonal melody
— which is a quite concrete and traceable object — through the lenses of linguistics
and psychoacoustics. here are two assumptions behind this approach. First, we
assume that the kinship between music and language is deeper than is usually
thought. hus, the comparative method is not just the best one, it is the single
method through which we can advance in these questions. Second, we assume that
no matter how meaningful music may be, it has a concrete aspect whose relevance
in semiosis is underestimated. Paradoxically, the deeper we go in musical meaning,
the more we need to look at the physical aspects of sound production and sound
perception.
Ever since Helmholtz’s seminal work, On the sensations of tone as a physiological
basis for the theory of music, 1 it is known that pitch, intensity, duration and timbre
are the main attributes of tone, which in turn is the basic unit of melody. From
a semiotic point of view it may be better to say that these attributes are minimal
features of tone, emphasizing the fact that two diferent tones can be fully
characterized on the basis of diferences in pitch, intensity, duration and timbre.
1.
Helmholtz (1895).
24
Primitives
To say that these features are minimal is to say that they are at the very endpoint
of musical analysis, in the same way as phonetic features are at the endpoint of
linguistic analysis. To say that these features are melodic primitives is equivalent
to saying that any aspect of melody could be described as a composition, that is, a
particular arrangement of them.
Such “atomistic” approach is not viewed favorably in semiotics, at least within
the Greimasian framework, and hence the question of musical primitives has been
relegated to a minor role. Several reasons contribute to this neglect, the main one
being the holistic principle already formulated by Aristotle, according to which
“the whole is more than the sum of its parts”. Accordingly, the properties of a
complex object cannot be understood only by the properties of its parts, rather
the object’s complexity afects the structure of its parts. Greimas’ criticism of the
sort of componential analysis carried out in the 60s is particularly clear about this
point:
On peut imaginer, théoriquement, qu’une vingtaine de catégories
sémiques binaires, considérées comme base taxinomique d’une combinatoire,
est susceptible de produire quelques millions de combinaisons sémémiques,
nombre largement suisant, à première vue, pour recouvrir l’univers sémantique
coextensif à une langue naturelle donnée. Sans parler de la diiculté pratique
d’établir une telle base d’universaux sémantiques, un autre problème - non
moins ardu - se pose lorsqu’il s’agit de préciser les règles de compatibilité et
d’incompatibilité sémantiques, qui président non seulement à la construction
des sémèmes, mais aussi à celles d’unités syntagmatiques plus larges (énoncé,
discours). Aussi voit-on que l’analyse sémique (ou componentielle) n’obtient de
résultats satisfaisants qu’en pratiquant des descriptions taxinomiques limitées…,
et que l’idée de pouvoir disposer, pour l’interprétation sémantique, de matrices
comparables à celles que la phonologie est susceptible de fournir pour sa propre
interprétation, doit être abandonnée… Ainsi, la grande illusion des années
1960 — qui croyait possible de doter la linguistique des moyens nécessaires
pour l’analyse exhaustive du plan du contenu des langues naturelles — a-t-elle
dû être abandonnée, car la linguistique s’était ainsi engagée, sans toujours s’en
rendre bien compte, dans le projet extraordinaire d’une description complète
de l’ensemble des cultures, aux dimensions mêmes de l’humanité. Greimas &
Courtés (1979, p. 327)
However, the fact that it is very hard to reduce the whole to its parts does
not imply that we will not improve our understanding of musical semiosis if we
consider those parts separately and study their speciic properties. Indeed, it is not
necessary to choose between “atomism” and “holism”; rather, semiotic analysis
comprises both a holistic approach, in which the object (a musical piece) is taken
as a whole, and an atomistic approach, in which the properties of primitives
become the focus of the investigation. his kind of research leads us to questions
like: What are the semiotic properties of tone per se? What are the meaningful
efects of dynamics and tempo? Why is there no speciic notation for timbre? his
Melodic Primitives and Semiosis
25
point is illustrated by the two fragments of Mozart’s clarinet concerto shown in
igure 1. he irst fragment (1) contains the motif played on the violin (bars 1-2);
the second (2) contains the motif played on the clarinet (bars 57-58).
Fig. 1. Mozart’s clarinet concerto in A (K622), Allegro, violin (1) and clarinet (2) parts.
In some respects, these two fragments may be viewed as exactly the same; in
some others, they may not. Actually they are almost indistinguishable from each
other, the diference being the key signature — 3 sharps in (1) and no sharps or
lats in (2) — which is just a notational convention. However, considering the way
these fragments are actually played, they are quite distinct. he actual sound we
hear when musicians play these fragments is somehow independent from their
form, and therefore should be investigated by themselves. Evidently, we do not ask
the reasons why Mozart decided to have the strings (1) introduce the theme and
some bars later restate it with the clarinet (2), which is sort of canonical is this kind
of composition. he question here is why Mozart chose the clarinet and not the
oboe? Or the horn, or the viol? In brief, what we are looking for is a “semiosis of the
timbre”, which is somehow independent from the piece in which it is actualized.
By doing so, we take timbre as a musical primitive. And what goes for timbre goes
for dynamics, tempo, and other melodic variables.
In order to examine this issue we have divided this article into three parts. he
subject of the irst part is melodic distinctiveness. Basically, we discuss how melodic
primitives are used to build up musical objects like melodies, chords, rhythmic
motives, and so on. In order to do so, we need to take a look at the kinship between
music and language from the viewpoint of the theory of distinctive features. Indeed,
timbre, intensity, pitch and duration are minimal features both in music and in
language.We will show that there are only two classes of features, one that comprises
pitch, duration and intensity, which is responsible for melodic distinctiveness, and
a class that comprises only timbre, which is responsible for musical enunciation.
he second part explores some properties of the vocal apparatus and of musical
instruments in order to explain why melodic primitives are organized in the way
they are. he subject of the third part, which is quite speculative, is the relationship
between timbre, connotation and enunciation.
26
Primitives
2. Melodic distinctiveness
Let us start our discussion by introducing the concept of distinctiveness in the
way it is understood in linguistics. he point here is to know whether or not
distinctiveness — a concept taken from phonology — can be applied to describe
how basic musical units are arranged in a system. According to phonologists, the
phoneme is the smallest unit of the speech chain which distinguishes meaning.
he English words “tea” and “key”, for instance, have both two phonemes each, /t/
and /i:/, and /k/ and /i:/ respectively. A pair of words like “tea” and “key” is called
a minimal pair because they difer in one single phoneme in the same position in
each word. he phonemes can be broken up into even smaller parts called phonetic
features, which are the smallest units of the system behind the speech chain.
Phonetic features are acoustic or articulatory attributes of the sounds of language
like Sonorant, Vocalic, Consonantal, etc. Figure 2 shows a matrix with all the
features necessary to build up the words “tea” and “key” and to distinguish them
from each other.
“tea”
/ti:/
Sonorant
Vocalic
Consonantal
Coronal
Anterior
High
Continuant
Voiced
/t/
+
+
+
-
“key”
/ki:/
/i:/
+
+
+
+
+
+
/k/
+
+
-
/i:/
+
+
+
+
+
+
Fig. 2. Matrix of phonetic features of the segments /t/, /k/ and /i:/.
We can see that /t/ and /k/ share the same features, but do not share their
values: /t/ is [+Coronal], [+Anterior] and [-High] while /k/ is [-Coronal],
[-Anterior] and [+High]. hus, although the phoneme is the smallest unit of the
speech chain used to diferentiate words, its distinctiveness is a property borrowed
from the phonetic features. Moreover, the inventory of phonemes is always a
restricted set of sounds that are actually employed in a particular language. he
inventory of phonetic features is universal, they are present in all known languages
and occupy the lowest degree in the phonological hierarchy, which means that
they are at the very endpoint of phonological analysis. In brief, phonetic features
are the primitives of the phonology of any and all everyday languages. 2
2.
Many aspects of the sound structure of languages cannot be described only through the theory
of distinctive features. Instances of this are the rhythmic patterns of speech, intonation, the
interaction between phonological units (words, feet, clitics, etc.) and syntactical structures, etc.
hus, since the late 80’s, new research pathways have been opening to explore these ields, like
Melodic Primitives and Semiosis
27
We are now in a position to introduce our irst hypothesis: the mechanism
behind the distinction between two melodic strings (a sequence of notes) and
between two words (a sequence of phonemes) is exactly the same, that is, the
concept of distinctiveness can be applied to music in the same way as it is applied
to language. For example, consider the fragments (3) and (4) in Figure 3 below.
Fig. 3. First bars of Happy Birthday (3) and Star-Spangled Banner (4).
he igure shows the irst bars of two unique melodies, Happy Birthday (3)
and he Star-Spangled Banner (4). No one will confuse these melodies despite the
fact that they difer from each other in nothing but one single attribute, namely,
pitch. Both have the same number of notes (six), with the same duration each,
and the more prominent notes of each string are placed in the same position.
Consequently, these fragments constitute a melodic minimal pair through which
we isolate pitch from other tone attributes, showing that it is a melodic feature —
from now on [Pitch] — whose distinctive function in music is similar to the one
played by phonetic features in language. he obvious diference between phonetic
features like Consonantal or Voiced, on one hand, and Pitch, on the other,
is that the former are binary, since they admit of just two values, + or -,while a
melodic feature like Pitch is n-ary, that is, it admits of multiple values.We will
come back to this point later on.
To say that pitch is a distinctive feature in music is to over-explain something
that everyone knows, and the extent to which such an approach could carry us
may be open to question.However, things are not that simple. Indeed, there are
melodies which can be easily distinguished from each other, but are made on the
same sequence of pitches. Consider the irst bars of the Beethoven’s “Ode to Joy”
(5) and Bach’s Cantata “Jesu, Joy of Man’s Desiring” (6). No one will confuse these
two melodies. However, they are built up on the very same chain of pitches, as
Figure 4 demonstrates.
metric phonology (Liberman & Prince 1977), intonational phonology (Pierrehumbert 1979),
auto-segmental phonology (Goldsmith 1976; Clements 1985), prosodic phonology (Nespor &
Vogel 1986), to mention a few. Although essential to our understanding of the way language
works as a whole, those analytical models have not changed a single detail in the theory of
distinctive features. Since the crucial point in our argument is distinctiveness, we have no need
to consider other phonological models in our analysis.
28
Primitives
Fig. 4. The first bars of Beethoven’s “Ode to joy” (5) and Bach’s “Jesu, joy of man’s desiring” (6).
hese two fragments constitute another melodic minimal pair. But now what
distinguishes them from each other is not the pitch of their tones but their duration.
In other words, we have managed to isolate duration from other features — from
now on [Duration] — showing that it plays a similar role to pitch in creating
melodic distinctiveness. Like pitch, duration is not binary, rather it is n-ary, which
seems to be an important characteristic of all melodic features and distinguishes
them from phonetic features, which are mainly binary.
Another tone attribute that plays a distinctive role in music is intensity —
for now on [Intensity]. Figure 5 below presents the theme of the Allegro from
the Brandenburg concerto BWV 1048 (7) and the same melodic material in a
compound rhythm (8). he single diference between the two fragments lies in the
notes which are stressed and the notes which are not. Since this detail is enough to
distinguish the two fragments, it is to be considered a melodic feature.
Fig. 5. Allegro from Bach’s Brandenburg Concerto BWV 1048 in the original version (7),
and in a compound rhythm (8).
It seems that intensity and duration are usually redundant features, that is, the
longer the tone is the stronger it tends to be and vice-versa. Finally, it is possible
to show that timbre, the remaining attribute of tone, is NOT a distinctive melodic
feature. We have already seen this indirectly in the fragments of Mozart’s clarinet
concerto (igure 1). he fact that timbre is not a distinctive feature explains why
there is not an accurate notational system for timbre. he same goes for dynamics,
Melodic Primitives and Semiosis
29
tempo, key, among others. As we will see later, the main characteristic of a nondistinctive feature is that it is able to be “translated” (in the sense Hjelmslev
understands this notion). 3 One can always play a melody at a slow or fast tempo,
and this will not change its identity. One can always play a melody on the violin or
on the clarinet, as we have seen from Mozart’s clarinet concerto. In these cases, we
are just “translating” one tempo into another, one timbre into another, and so on.
But when we change pitch, duration or intensity of the melodic tones, we change its
identity. It is for this reason that they are distinctive melodic features. Interestingly,
as with phonological features, one can build up a matrix with distinctive melodic
features. Indeed, one inds such a matrix in sotware for music composition and
edition. Figure 6 shows an example from the irst three bars of “Happy Birthday”.
We will come back to this point later.
Bar
0001.01.001
0001.01.007
0002.01.001
0002.02.001
0002.03.001
0003.01.001
Pitch
g3
g3
a3
g3
c4
b3
Duration
288
96
384
384
384
768
Fig. 6. A matrix from the first three bars of “Happy Birthday” taken from the software
MUSE.
3. Grading, context-sensitivity and tensiveness
Melodic distinctiveness is not a minor question. Indeed, distinctiveness per se is an
intriguing cognitive question whether from a theoretical or a practical perspective. 4
But our interest in distinctive melodic features lies on three particularities, namely
grading, context-sensitivity and tensiveness, which are absent from phonetic
features and which could help to explain some of the meaning efects of music. Let
3.
Hjelmslev (1961, p. 117).
4.
Every time we are faced with an object, no matter its nature, we are asked to identify and
recognize it, and the single way to do that is by inding out the distinctive feature(s) that makes
it diferent from other more or less similar objects. Whether we perceive an ordinary object, a
word, or a melody, we are always asked to ind out what gives them their speciicity. Although
distinctiveness is not among the main topics in semiotics, it is in the core of the so called artiicial
recognition systems, which are devices for the recognition of voices, characters (OCR), faces,
gestures, and melodies, among others. he economic interest in these devices and the technical
interest in how they may work are quite obvious. With 39% of the music market being made in
digital format, it is no wonder that companies like Sony, Google and Microsot have a strong
investment in research, mainly in Music Data Mining and Music Recommendation Systems,
ields in which a theory of the melodic distinctiveness is crucial.
30
Primitives
us briely examine each of these particularities, starting with grading. Phonological
systems are described as structures of numerous features (around twenty, according
Chomsky & Halle), and each one is deined as positive or negative. Any natural
language can be described as a special arrangement of those twenty binary features.
he structure behind music seems to be much more parsimonious, since no more
than three features are needed to characterize any melodic string. On the other
hand, these features are not binary, rather they are gradual, assuming several values
depending of each musical “idiom”. In other words, the phonological system and
the melodic system seem to be conversely oriented, the irst being a structure of
binary oppositions, 5 the second being a structure of gradual contrasts. his detail
afects the way semiosis is produced in language and in music.
Coming back to the examples we began with, the diference between /t/ in “tea”
and /k/ in “key” is due to the linguistic oppositions between phonological features
like [+/- Coronal], [+/- Anterior] and [+/-High]. he way language expresses
grammatical and semantic relations is also based on oppositions between binary
features, like [+/- Masculine], [+/- Singular]and so on. In a sense, the system
of phonological oppositions and the system of semantic and lexical oppositions
seem to suit each other. Both are organized on the same basis. hat is why binary
opposition is the fundamental relationship in linguistic structures. A quite
diferent picture emerges from the melodic system because, in this case, we have a
system exclusively based on degrees of contrast, rather than on binary oppositions.
Pitch is a relationship between two adjacent units in a string, and not simply an
opposition between two units. his is true not only for pitch, but for all melodic
features, be they primitive or derived. Duration, pitch, tempo, dynamics, etc. are
not bipolar categories, but contrastive and gradual in nature. his perhaps explains
why music lacks in representing referents and propositions. On the other hand
music is suitable to expressing correlations with meanings based on contrasts,
like afects and emotions. To sum up, the grading of distinctive melodic features
(pitch, duration, and intensity) and of their derived features (register, tempo and
dynamics), allows them to express any kind of meaning whose intrinsic character
is also gradual. As we have known since Sémiotique des passions (Greimas &
Fontanille 1991), grading is one of the basic characteristics of emotion, afects and
passions. 6
5.
“In view of the fact that phonological features are classiicatory devices, they are binary, as are
all other classiicatory features in the lexicon, for the natural way of indicating whether or not
an item belongs to a particular category by means of binary features. his does not mean that
the phonetic features into which the phonological features are mapped must also be binary. In
fact, the phonetic features are physical scales and may thus assume numerous coeicients, as
determined by the rules of the phonological component.” (Chomsky & Halle 1970, p. 297).
6.
We cannot develop this point further here as it would require an extended discussion about
tonality and harmony, and perhaps another essay. he reader interested in the semantic aspects
of this issue can refer to Claude Zilberberg’s recent work, in particular Zilberberg (2012).
Melodic Primitives and Semiosis
31
A second particularity which is present in melodic features — and absent from
phonological features — is context-sensitivity. 7 One cannot assign an absolute value
for the pitch, duration and intensity of a tone independently of the context where
it occurs. In order to make this point clear, let us consider the phonetic feature
[Voiced], for example. A segment can be [+ Voiced], like /b/ or [- Voiced], like
/p/. his characteristic is not dependent on the context where /b/ or /p/ occurs. 8 In
other words, /b/ is always /b/, in any circumstance, as well as /p/ is always /p/ in any
circumstance, and the same goes for any other phonetic feature, without exception.
his context-free characteristic makes the phonological system stable enough for
its purpose, namely, the communicative functions of the natural languages.
All melodic features, on the contrary, are context-sensitive. One cannot assert
that a musical sound is short or long, high or low, strong or weak by itself. As
the values are all relative, the sound that is long in one context can be short in
another context, and the same is true for intensity and pitch features. To say that a
sound is context-free is to say that it is not subject to any syntagmatic constraint.
Conversely, to say that a sound is context-sensitive is to say that it is subject to a
syntagmatic constraint.
his point is crucial for understanding why musical semiosis is fundamentally
iconic whereas verbal semiosis is fundamentally symbolic (see Santaella, in this
volume, page 92). Saussure showed that language is based on oppositions. 9 Two
phonemes, say /p/ and /b/, are opposed to each other from the point of view of the
paradigm, that is, the vertical axis that is behind the text. his means that /p/ and
/b/ presuppose each other. 10 However, there is no presupposition at all between
them from a syntagmatic point of view, i.e., there is no syntagmatic constraint
between /p/ and /b/.To sum up, thanks to the paradigmatic presupposition, /p/
and /b/ are placed in opposite sides in the system and can be used as a basis for
oppositions, like in “pack” versus “back”; however, thanks to the lack of syntagmatic
presupposition, the presence of /p/ does not imply the presence of /b/ and viceversa. So we ind words with both /p/ and /b/ (“bump”, “pub”), just /p/ (“papa”)
or just /b/ (“baby”). he basic mechanism of natural languages rests on this dual
property of the phonemes.
7.
We borrow this notion from semantics: “All linguistic morphemes are context-sensitive in the
way that their semantic value depends partly on their semantic environment (tender does not
have the same meaning in a tender steak and in a tender man)”. (Robert 2005, p. 123).
8.
Many languages present neutralization between phonological oppositions. For instance, under
certain circumstances, German plosive voiced consonants [b, d, g] become plosive unvoiced
consonants [p, t, k], respectively. hus, “ab” (from) is pronounced [ap], “Farhad” (bicycle) is
pronounced [faːɐ̯ˌʀaːt], and “Tag” (day) is pronounced [taːk]. Obviously, such a devoicing rule
does not imply that the speaker does not recognize the diference between voiced and unvoiced
consonants.
9.
“Tout le mécanisme du langage (…) repose sur des oppositions de ce genre et sur les diférences
phoniques et conceptuelles qu’elles impliquent.” (Saussure 1967, p. 165).
10.
he presupposition rests on the voiceless/voiced glosseme that underlies the opposition p/b.
32
Primitives
In music, on the contrary, things do not work in this way because musical
notes contract mutual presuppositions from both the paradigmatic and the
syntagmatic viewpoints. hus, a phoneme, say /p/, can be unequivocally established
through the paradigmatic functions contracted with other members of the system,
say /b/, /t/, /d/, etc. However, a musical tone cannot be unequivocally established
through only the paradigmatic functions it contracts with other members of
the system. In other words, the paradigmatic oppositions are necessary but not
suicient conditions to establish the value of a tone. To do so, it is necessary
to consider its syntagmatic environment, that is, its context. We are faced with
a relationship between two units in a string (one syntagma), and not simply an
opposition between two units (one paradigm). All and any musical segments are
subject to this kind of syntagmatic constraint. his is true not only for pitch, but
for all musical features, without exception. his property of musical features have
two remarkable consequences. On one hand, music is unable to express oppositive
contents like the ones we ind in natural languages. It is not possible to express the
diference between a “computer” and a “table” through musical sounds because
they are context-sensitive, that is, they are not autonomous from each other. 11
On the other hand, musical features can express iconically any content that itself
presents the same syntagmatic constraint. his is the basic reason why music can
express afects and emotions almost directly. hus, serenity and ecstasy, sadness
and joy, tension and relaxation are not only opposed afects (paradigmatic). We
can trace a continuous line between joy and sadness, as well as between fast and
low tempos, because joy and sadness, on one hand, and fast and low tempos, on
the other, establish cohesive relationships. Melody achieves its universality from
this structural parallelism between musical features on one hand and afects and
emotions on the other (see Zbikowski, in this volume, page 154).
Finally, the third particularity of distinctive melodic features, tensiveness, is
closely related to grading and context-sensitivity. All and any distinctive melodic
features are tensive categories. Let us see how Greimas & Courtés deine tensiveness
in the Dictionnaire I.
Tensivité : La tensivité est la relation que contracte le sème duratif d’un procès avec le sème terminatif : ce qui produit l’efet de “tension”, “progression”
(exemple : l’adverbe “presque”, ou l’expression aspectuelle “sur le point de”).
Cette relation aspectuelle surdétermine la coniguration aspectuelle et la dynamise en quelque sorte. Paradigmatiquement, la tensivité s’oppose à la détensivité.
(Greimas & Courtés 1979, p. 388)
Obviously, this deinition takes tensiveness as a category in the content plane.
Indeed, at the time the Dictionnaire I was being prepared (1979), the main task
of Greimas and Courtés was to establish a solid basis for a theory of discourse
11.
In fact, we are not faced with a real impossibility; we just do not know any everyday language
whose structure is based on this kind of syntagmatic relationship. By the way, such a
(hypothetical) language should be completely based on prosody.
Melodic Primitives and Semiosis
33
independently from its manifestation, that is, independently from the expression
plane. However, the Dictionnaire II (1986) restates the entry, now written by Claude
Zilberberg, who emphasizes that shortcoming in the theory. In any case, thanks
to the principle of structural analogy established by Hjelmslev, 12 it is possible to
regard tensiveness as a category of the expression plane as well.
he crucial point with tensiveness is that it establishes a strong relationship
between both the paradigmatic and the syntagmatic axes, which is far from being
usual among categories on both planes. On the contrary, adjacent units in the
syntagmatic axis usually do not inherit paradigmatic dependence. One example
will make this point clear. he category [+/- Voiced] constitutes a paradigmatic
dependence between the features [+ Voiced], or voiced and [- Voiced] or
voiceless. However, when projected onto the syntagmatic axis, these features
are NOT dependent on each other. he fact that a sound is [+ Voiced] does not
produce the expectation of a sound [- Voiced].
his kind of expectation is exactly the meaning efect produced by any
tensive category. It is obvious that tensiveness is one of the foundations of any
and all rhythmic phenomena. Indeed, we could deine rhythm as nothing but the
projection of a paradigmatic dependence onto the syntagma. his is the reason why
poetic rhythm is based on prosodic features, that is, the alternating between weak
and strong, high and low, short and long 13 syllables. All these prosodic categories
are tensive, since the presence of one of these terms creates the expectation of
its complement. Not surprisingly, the prosodic features cover exactly the same
substance as the distinctive melodic features do.
4. Musical instruments, vocal apparatus and melodic primitives
Before we go on, here is a summary of what we have seen so far. he expression
plan of both music and language is built up with sounds organized in a system
of opposite values, which can be analyzed in minimal elements, the so-called
primitives. hese elements are commutable, that is, the interchange between any
two of them changes the identity of an utterance. he properties of the primitives
are inherited by the sounds that compose the system of both language and music.
Since grading, context-sensitivity and tensiveness are properties of pitch, duration
and intensity, music is especially suitable to function as an analog (an icon) of any
object or process that presents the same properties. he expression plan of natural
language, on the other hand, is mainly structured on oppositions and discrete
12.
“…expression plane and content plane can be described exhaustively and consistently as being
structured in quite analogous fashions, so that quite identically deined categories are foreseen
in the two planes. his means a further essential conirmation of the correctness of conceiving
expression and content as coordinate and equal entities in every respect” (Hjelmslev 1961,
p. 60).
13.
See Jakobson, Fant & Halle (1952, p. 13).
34
Primitives
categories and, for this reason, it cannot refer directly to dynamic and tensive
objects and/or processes.
here are reasons to think that the roots of this diference between language
and music lie on their origin and development. Although language is as “natural” as
music and their origins are lost in the mists of time, it is possible to hypothesize that
speech and singing have a common background in the prehistory of mankind, but
at some point they diverted into opposing paths. his is the so called musilanguage
hypothesis sustained by Brown (2001). Although it is not possible to prove this
hypothesis, we can indirectly explore it, and its consequences, by considering some
properties of the vocal apparatus and of musical instruments. he idea behind
this approach is that the evolution and the development of language and music
could well be mapped against the mechanisms that produce linguistic and musical
sounds. Instead of considering language and music by themselves, we might now
consider the devices that produce language and music, with the hope that they can
shed light on the problems of melodic primitives.
According to Fant (1960), speech sounds result from two consecutive
processes: a source produces an initial sound and a ilter modiies it. At the larynx,
where the vocal folds are, sounds are produced whose spectrum contains diferent
frequencies. his spectrum is iltered by articulators like tongue, teeth, lips, velum,
etc. Basically, when a human speaks, the vibration of the vocal cords function as
a glottal source of energy and the displacement of the organs of the mouth and
throat function as the ilter. he combined action of source and ilter produces
the two phonological macro-categories, phonemes (consonants and vowels) and
prosodemes. With this mechanism, a human can produce thousands of syllables,
which are the basic building blocks of speech. he inventory of English syllables,
for example, runs to something around 16,000 syllables. he crucial point in Fant’s
model is that phonemes depend on the mobility of both source and ilter, whereas
prosodemes depend exclusively on the source’s mobility. Consequently, source
and ilter are functionally speciic.
Musical instruments are mechanisms fully comparable to the human vocal
tract. All and any melodic musical instruments have a source of energy and a
ilter. he source-function can be performed by the strings of a guitar, the wooden
bars of a xylophone, the reed of a clarinet, and so on, while the ilter-function
is performed by sound boxes and tubes in many diferent shapes and sizes, built
from a whole array of diferent materials. However, musical instruments present
a crucial diference in relation to the vocal tract. In musical instruments, the ilter
does not have movable parts that play a role similar to that played by the jaw,
tongue or velum. hat means that the ilter of musical instruments is always static.
One can pluck or bow the strings of a violin but one cannot change the shape of its
body in order to produce diferent timbres. he same is true for wind instruments.
On a saxophone, for example, the source is a single reed mouthpiece, and the ilter
is the metal pipe and the bell. he only movable part of the whole device is the reed.
Melodic Primitives and Semiosis
35
he keys only change the resonance frequency of the sound generated. We press
the keys in order to change nothing but the pitch.
As tones are the outcome of the source vibration, we have nothing but tones as
the output of a musical instrument. In this order of ideas, the mobility of the ilter
seems to be the key factor distinguishing musical and linguistic sounds, which are
therefore organized in systems conversely oriented. he melodic system can be
described as a singular case of organization of basic sound unit categories where (i)
the segments category is completely syncretized in one single value, the timbre, and
(ii) the supra-segments category presents maximal resolution, with a wide range
of pitch, duration, and intensity values. his organization is opposed to the one
observed in natural languages, where segments present maximal resolution and
supra-segments are reduced to a category with few terms. 14
From this point of view, what makes the relationship between music and
language so special is that they share the same basic units, but at the same time
these units are organized in systems conversely oriented. In language, the prosodic
categories are atrophied and play a marginal role; in music, they are hypertrophied
and constitute the core of the system. In language, the inventory of phonemes is
much richer than the inventory of prosodemes, while in music we have the opposite
scenario. hus, we could metaphorically say that the melodic system seems to be
an inverted mirror of the phonological system.
his reasoning shows that the systems of musical and linguistic sounds are not
diferent from each other in the substance of their basic units. Consequently, the
fact that they are specialized systems has to be attributed to the divergent ways their
evolution and development took. It is well known that several anatomical changes
in the vocal tract were decisive for the emergence of human language (de Boer
2010). Crucially, the lowered larynx and the angle between mouth and pharynx
created the room necessary for the improvement of tongue mobility, that is, the
ilter mobility. On the other hand, from the source-ilter model perspective, what
distinguishes a modern lute from its probably oldest ancestor (the disputed Divje
Babe lute), or a modern string instrument from the ancient lyra, is the range of
tones they produce (Figure 7). As instruments evolve, their range increases. If we
consider the three basic features of musical tones (pitch, duration and intensity),
the current violin has a range of about 2,400 sounds and a piano has 4,488 sounds,
within the same violin-like and piano-like timbres.
14.
he system of standard English sounds, for example, runs to forty-four phonemes but only
a few prosodemes, while the currently known languages number up to at least ninety-three
phonemes. On the other hand, the system of prosodemes can run to only ive levels of pitch
(Mandarin Chinese) and three levels of duration (Swedish).
36
Primitives
Fig. 7. The dashed line with arrows shows the divergence in the evolution of music and language.
his approach helps us understand why music and language share some
properties but not others, and strengthens the musilanguage hypothesis proposed
by Brown (2001). However, in contrast to Brown’s psycho-biological standpoint,
ours is semiotic — that is, we try to show that some properties of the vocal tract
and of musical instruments are at the root of the diferent ways language and music
produce meaning.
5. he psychoacoustics basis of timbre
It is time now to briely discuss the remaining melodic primitive: timbre. We have
just seen that, from a semiotic viewpoint, timbre seems to play a diferent role
than pitch, duration and intensity. Timbre is not a distinctive feature. In terms
of hjelmslevian semiotics, one can say that whereas pitch, duration and intensity
are melodic invariants, timbre is a melodic variant, thus emphasizing the fact that
timbre can be freely varied without changing the identity of a melody — something
which cannot be done with pitch, duration and intensity. he psychoacoustics
behind sound production help us igure out why things work that way and what
the consequences of this fact are for musical semiosis.
One of the most important tools for acoustic analysis (Henrique) was developed
in the early 19th century by the French mathematician and physicist Jean-Baptiste
Joseph Fourier, and is now called Fourier analysis. If applied to sound, it states that
any complex sound can be described as a composition of simpler sounds. hese
simpler sounds can be split into two groups: (i) the Fundamental, also called f1,
Melodic Primitives and Semiosis
37
and (ii) the harmonics, whose frequencies are integer multiples of f1. heoretically,
the number of harmonics is unlimited. For instance, given a sound whose f = 1000
Hz, its fundamental and harmonics are: f1 = 1000 Hz; f2 = 2000 Hz; f3 = 3000 Hz;
and so on.
One may ask what exactly difers between the timbre of two musical
instruments, say, the violin and the clarinet? he answer is that the harmonics
produced by these instruments present diferent intensities, even when f is exactly
the same. hus, suppose a violin and a clarinet producing the same note C, with the
same intensity and the same duration. We will perceive diferent sounds (timbres)
due to the harmonic content of these sounds. Moreover, a trained musician can
perceive and identify two violins because no two musical instruments produce
exactly the same timbre, no matter how similar they are to each other. he same
goes for any musical instrument. Furthermore, this principle of timbre complexity
also explains why we never ind two diferent people with the same voice. Indeed,
the structure of the vocal apparatus is quite similar to that of musical instruments,
and the physics behind them is exactly the same, both being composed of a
vibration system which produces a complex vibration pattern (f1 and harmonics)
and a ilter system which enhances certain harmonics and dampens others.
he violin is a good example of such a vibrating/ilter system, as we can trace
the sound path from the string to the plates. he vibrating string usually has a
lot of harmonics, each with a diferent intensity. From the string, the vibration
is transmitted along and down the bridge to the top plate. he bridge has its own
preferred frequencies and hence plays the role of a ilter that reinforces some
harmonics while sotening others. From the bridge, sound is transmitted to the
top plate of the violin. Like the bridge, the top and back plates also have their own
(diferent) preferred frequencies and again the ilter efect occurs. Let us illustrate
this point with a concrete example (see Figure 8) following Gough. A G note with a
fundamental frequency (f1) of 196 Hz is played by bowing the violin’s string (input
wave form). he intensity of each harmonic of this frequency decreases as its order
increases, which is represented by the decreasing height of the bars, as shown in
the graph on the right side of the input wave form. However, the bridge is selective
in frequency and emphasizes the frequencies between 2.5 kHz and 3.5 kHz (bridge
response). he body has a much more complex pattern of emphasized/dampened
frequencies that afects virtually all of the harmonics and the fundamental (body
response). To sum up, the inal sound produced by a violin (lower graph) has a
completely diferent set of relative intensities as compared with the ones of a single
vibrating string. he waveform of this inal sound is shown in Figure 8 (output
waveform).
38
Primitives
Fig. 8. The filter effects caused by the violin parts and the modifications they produce on the input
waveform.
What we have seen so far is related to the way sound is produced. But the
way sound is perceived also afects the meaning efect of the timbre. Indeed,
our perception of sound is not “lat” because the ear response varies according
to frequency. A clear indication of this is the impossibility to transpose timbres.
Suppose one records a female voice and then, with the aid of sotware, we transpose
this sound to a lower tone to make it sound like a male voice; the resulting voice
will not be a male voice, rather a caricature of a male voice with a comedic meaning
efect. As explains Leipp (2010, p. 156),
L’anatomie et la physiologie individuelles du système auditif jouent donc un rôle
déterminant dans la sensation de timbre et il n’est pas surprenant de constater
des divergences dans le jugement entre auditeurs d’âges diférents. Il est donc
clair que le mot « timbre » ne peut avoir de signiication intrinsèque, absolue,
physique.
What is the relevance of this mechanism behind sound production and sound
perception for musical semiosis? Above all, we note that what we call timbre, pitch,
intensity and duration are just meaning efects, that is, abstract representations that
result from our perception of certain physical magnitudes. hus, the frequency of
f1 produces the meaning efect of pitch, the amplitude of f1 produces the meaning
efect of intensity, the length in time of f1 produces the efect of duration. he
complex harmonic composition of multiples of f1 produces the efect of timbre.
It is important to note that, besides their frequencies and relative intensities, each
harmonic has its own duration.
We now have enough elements to justify our proposition that timbre, on one
hand, and the other melodic features, on the other, should be treated as two diferent
classes of primitives. We have mentioned the possibility of representing tone as a
Melodic Primitives and Semiosis
39
matrix of melodic features. However, due to the diferences we have pointed out
between pitch, duration and intensity, on one hand, and timbre, on the other, we
cannot represent timbre in a single matrix. Because distinctive melodic features are
inite values, a matrix for tone has a limited number of columns; and because the
number of harmonics that characterize timbre is theoretically unlimited, a matrix
for timbre would have an unlimited number of columns, which is far from being a
simple and practical system of representation. 15 To sum up, while pitch, duration
and intensity constitute a closed class, timbre constitutes an open class.
6. Conclusion
Now we will proceed to discuss how these particularities of the musical primitives
could guide future research in semiotics. Among the developments that could
be explored and about which we intend to write at length later, we now mention
just two, namely the kinship between music & language, including parallels and
divergences in their evolution and development, and the question of musical
enunciation and connotation.
One of the main tasks of contemporary semiotics is still the comparative
study of music and language. he chief reason for this is that there are no other
semiotics so similar to each other than those for music and language. At the same
time, the singularities presented by music can shed light on language and viceversa. Moreover, from a structural point of view, language and music are by far
the most complex semiotics we know. We have already seen that melodic features
present three particularities — namely grading, context-sensitivity and tensiveness
— which are absent from phonological features, and which are responsible for
some diferences between music and language. Moreover, the fact that music is a
prosodic-based semiotic (since pitch, intensity and timbre are the so-called prosodic
categories) opens the possibility for polyphony, which is strictly impossible in
language. Two or more people cannot speak at the same time, which is not a
problem at all when they sing or play. One important consequence of polyphony
is that it opens the doors for harmony. When we play two or more instruments at
the same time, the meaning efects of harmony are created. Any monody has an
implicit harmony, but the whole development of harmonic possibilities depends
on polyphony. hus, the roots of harmony and polyphony — to mention just
two relevant ields of music — are grounded on musical primitives. Another
question involving primitives relates to a striking asymmetry in the inventory of
the signiicants of both music and language. We know that musical instruments
and the vocal apparatus present structural similarities. Despite this fact, linguistic
15.
his is the reason behind the fact that one can represent a sequence of tones with a limited
number of symbols but one cannot do the same with timbre. Consequently, there is no notation
system for timbre, and we are limited to use labels like “violin” “cello” “lute” and so on.
40
Primitives
distinctiveness is organized around phonemes, with tones as a marginal class, even
in tone languages. On the other hand, we have nothing but tones as the output of
a musical instrument. his seems to be a key factor distinguishing musical and
linguistic sounds, which are therefore organized in systems conversely oriented.
In language — including tone languages — the prosodic features (pitch, duration
and intensity) play a secondary, marginal role; in music, they play a primary role
and, in fact, constitute the core of the system. From this viewpoint, the plan of
expression of music and language seem to be “inverted mirrors”. his phenomenon,
whose roots can be found in the evolution of music and language, can help us
understand why music is fundamentally iconic whereas language is primarily
symbolic (cf. Santaella, in this volume). Despite being a crucial question to our
understanding of music semiosis, no continued attempt has been made to explore
this research pathway, as far as we know.
Any and all enunciates — including melodic enunciates — are the result of an
enunciation act performed by an enunciator. Since pitch, duration and intensity
are the sole distinctive melodic features, they are the necessary and suicient
conditions to establish any and all melodic utterances, that is, a melodic utterance
is nothing but a string of pitch, duration and intensity values. Timbre has no place
in such a melodic utterance. On the other hand, there is a solidarity (Hjelmslev)
between enunciate and enunciator. hus, any melodic utterance can be performed
in an unlimited number of timbres, each one lending its particular character to it.
In other words, in each enunciation act, the enunciator manifests itself through
the timbre. It must be stressed that this is restricted to the melodic primitives and
their role in semiosis. Evidently there are derivates from these primitives, which
can play diferent roles. For instance, dynamics, which is a derivate from intensity,
is not a distinctive feature and consequently belongs in one of the enunciation
categories. he same goes for register and tempo, which are derived from pitch
and duration respectively. All these categories are controlled by the enunciator.
For instance, a singer or player can change the meaning efect of a whole piece by
changing the values of timbre (at least partially), tempo, register and dynamics.
By doing so, he/she will not change the piece in itself, but some of its meaning
efect, which is usually expressed verbally with connotative terms like “strong”,
“sweet”, “precise”, and so on. It is clear that those meaning efects are related to the
enunciator since those terms characterize a strong, sweet or precise enunciation.
In other words, all of those indexes connote something from the enunciator rather
than from the enunciate.
Melodic Primitives and Semiosis
41
Bibliographical References
Brown, S. (2001), “he ‘Musilanguage’ Model of Music Evolution”, in he Origins of
Music, Cambridge: MIT Press, pp. 271-301.
Carmo Jr., J.R. do (2012), “On some parallels between the vocal apparatus and musical
instruments, and their consequences for the evolution of language and music”, in
Scott-Phillips, Thomas et al. (eds.), Proceedings of the 9th international conference
on the evolution of language (EVOLANG9), Kyoto: World Scientiic, pp. 430-432.
Chomsky, N. & Halle, M. (1968), he sound patterns of English, New York: Harper and
Row.
De Boer, B.G. (2010), “Modelling vocal anatomy’s signiicant efect on speech”, Journal of
Evolutionary Psychology, 8(4), pp. 351-366.
Fant, G. (1960), Acoustic heory of Speech Production, Mouton: he Hague.
Gough, C. (2000), Science and the Stradivarius, Physics World, online: http://physicsworld.
com.
Greimas, A.J. & Courtés, J. (1979), Sémiotique. Dictionnaire raisonné de la théorie du
langage I, Paris: Hachette.
Greimas, A.J. & Fontanille, J. (1991), Sémiotique des passions, Paris: Seuil.
Helmholtz, H. (1895), On the Sensations of Tone as a Physiological Basis for the heory of
Music, London: Longman.
Henrique, L.L. (2011), Acústica Musical, Lisboa: Fundação Calouste Gulbekian.
Hjelmslev, L. (1961), Prolegomena to a heory of Language, Madison: he University of
Wisconsin Press.
Jakobson, R., Fant, G., Halle, M. (1952), Preliminaries to Speech Analysis: he Distinctive
CFeatures and their correlates, Cambridge: MIT Press.
Leipp, É. (2010), Acoustique et musique, Paris: Presses des Mines.
Robert, St. (2005), “he Challenge of Polygrammaticalization for Linguistic heory”, in
Frajzygier, Hodges & Rood (eds.), Linguistic Diversity and Language heories,
Amsterdam, John Benjamins, pp. 119-142.
Saussure, F. de (1913), Cours de linguistique générale, Paris: Payot.
Zilberberg, Cl. (2012), La Structure tensive, Liège: Presses Universitaires de Liège.