On the Formal Distinction between Literal and Figurative Language Udo Hahn Katja Markert Text Understanding Lab Freiburg University D-79085 Freiburg, Germany Language Technology Group HCRC, Edinburgh University Edinburgh EH8 9LW, U.K. http://www.coling.uni-freiburg.de [email protected] The distinction between literal and gurative language (metonymies, metaphors, etc.) is often not made formally explicit, or, if formal criteria exist, insuÆcient. This poses problems for an adequate computational treatment of these phenomena. The basic criterion for delineating literal from gurative speech we propose is centered around the notion of categorization conicts that follow from the context of the utterance. In addition, we consider the problem of granularity, which is posed by the dependence of our approach on the underlying ontology. Abstract. 1 Introduction Figurative speech comes in dierent varieties (e.g., the metonymy in example (2) and the metaphor in example (3) below), and is typically contrasted with literal language use (e.g., example (1)) on the basis of some notion of deviance. (1) \The man left without paying." (2) \The ham sandwich left without paying." (3) \The Internet is a gold mine." Currently, two approaches prevail, which spell out this distinction. The rst one [19] simply regards deviation from literal reference as a suÆcient condition for gurativeness. This does not explain, however, what the formal criteria for deviation actually are. Thus, the discrimination of literal and gurative meaning rests on subjective ascription. The second approach [5, 27, 24] introduces such a formal criterion, one which depends on the notion of \violation of norms", selectional restrictions of verbs and nouns, in particular. Each time these are violated, e.g., through type conicts, an instance of gurative speech is encountered. As a consequence, special reasoning patterns, such as type coercion (for metonymies [24]) or analogy-based structure mapping (for metaphors [2, 7]), are activated in order to cope with the triggering instance such that a reasonable interpretation can be derived that does no longer violate the underlying constraints. The proponents of this approach present a lot of supporting evidence for their methodological claims (cf., e.g., example (4)) but obviously fail to cover a wide range of residual phenomena (as can be seen from the example (5) that lacks any violation though being gurative without doubt): (4) \I read Chaucer." (5) \I like Chaucer." In this paper, we subscribe to the derivational paradigm, i.e., the requirement to generate gurative speech from specications of literal language use. Thus, we also regard gurative meaning as a deviation from literal meaning, but will not be content to leave the description at that. Rather than formalizing the notion of deviation with recurrence to selectional restrictions we will base our distinction on conceptual criteria that also incorporate the inuence from the context of an utterance. In addition, we share the assumption that the distinction between literal and gurative speech derives from one's conceptualization and is therefore subjective, but we will provide a formalization of this dependency on ontology. We will also identify at least one case of ontological dependency (the problem of granularity) where subjectiveness can be overcome by taking additional formal criteria into account. 2 Lexical Meaning We will base our considerations on the notion of context-independent lexical meaning of lexemes, from which the notions of literal and gurative meaning in context will be derived. Lexical meaning will be a function from lexemes to categories (concepts) of an ontology. So, let L be the set of lexemes of a given natural language and let L L be the subset of lexemes containing nouns, full verbs and adjectives only (e.g., man or policeman are elements of L ). We also assume an ontology composed of a set of concept types F = fMan, Policeman, Sandwich, : : : g, a set of instances I = fman-1, policeman-2, : : : g, and a set of relations R = fhas-part, partof, agent, : : : g (we take a settheoretical semantics for granted as is commonly assumed in description logics [32]). The lexical meaning lex can then be dened as a relation lex L fF [ Ig. While we refrain from considering the linkage between lexemes and ontological entities in depth (cf., e.g., [4] or [15]), we require the relation lex to fulll the following properties: 0 0 0 1. If lexeme 2 L is a proper name, then a unique lexeme.i 2 F [ I with (lexeme, lexeme.i) 2 lex exists such that lexeme.i 2 I . Thus, every proper name is linked to a single instance in the domain knowledge base. 2. If lexeme 2 L is not a proper name, then we require a concept lexeme.con 2 F to exist such that (lexeme; lexeme.con) 2 lex . In addition, no instance lexeme.i 2 I exists such that (lexeme, lexeme.i) 2 lex . 3. For reasons of simplicity, we will now restrict lex appropriately. If lexeme 2 L is not a proper name, we require for all i 2 lex (lexeme) that i can be referred to by lexeme in a context-independent way. So, we assume that reference to any i via lexeme is always possible. (We cannot, e.g., relate the lexeme fool to Man as not every man can be referenced by fool independent of the context.) The condition of context-independence may, however, still 0 0 0 hold for several concepts that stand in a subsumption relation to each other. When we regard the lexeme man, this condition holds for both the concepts Man and Policeman, as all i 2 Policeman and all i 2 man can be referenced by man. We then regard the most general concept (here, Man) as the lexical meaning, and, in general, consider lexical meaning as a function1 lex : L 7! F [ I . By convention, we denote lex (lexeme) by lexeme.con. 0 Lexical meaning is thus considered as a context-independent function from lexemes to categories (concepts) of an ontology. As there is no agreement on canonical ontologies, this mapping introduces subjective conceptualizations. Finally, we extend our denition to words w of a discourse so that their corresponding lexeme be w:lex 2 L . We simply assume lex (w) := lex (w:lex). We distinguish the range of that mapping by w.i for proper names and w.con in all other cases. Hence, the lexical meaning of the word \man" in example (1) is given by Man.2 0 3 Literal vs. Figurative Meaning While in the previous section we have been dealing with the isolated lexical meaning of a word only, we will now incorporate the context of an utterance in which a word appears. Hence (cf. Fig. 1), we here introduce the word w with respect to which word w is syntactically related { w is either head or modier of w. Such a dependency relation (either a direct one or a well-dened series of dependency relations) at the linguistic level induces a corresponding conceptual relation r 2 R at the ontological level [25]. The conceptual relation r links the conceptual correlates, w.sf and w .sf, of w and w , respectively. Accordingly, we may now say that w StandsF or a corresponding domain entity w.sf; alternatively, w.sf is called the (intended) meaning of w. The comparison of w.sf with w.con or w.i lies at the heart of the decision criterion we propose for judging whether a reading is literal or gurative. So, in the well-known example (2), \ham sandwich" (= w) StandsF or \the man who ordered the ham sandwich" (= w.sf), which is distinct from its lexical meaning, Ham-Sandwich (= w.con). We may now consider some examples to distinguish several cases how w.sf or w.consf can be related to w.con or w.i. This will also lead us to clarify the notion of distinctiveness between the items involved. Let w.sf be an instance from I , and let w.consf be the least general concept such that w.sf 2 w.consf .3 This assumption will be shortcut as w.sf instance-of w.consf . 0 0 0 1 2 3 0 In order to make lex a function we assume in the case of polysemy one of several meaning alternatives to be the primary one from which the others can be derived. In the case of homonymy, we assume the existence of dierent lexemes which can be mapped directly to mutually exclusive concepts. The lexical meaning of a word w must be distinguished from the concrete referent of w in the given discourse. The least general concept w.consf with w.sf 2 w.consf is the intersection of all concepts C 2 F with w.sf 2 C. Framework for Contextual Interpretation In the most simple case, w.sf and w.con / w.i are related by an instance-of relation. Then w.consf = w.con holds. In the utterance Fig. 1. (6) \A man left without paying." we have w = \man" and w = \left". Furthermore, w.sf = man-1 instance-of = w.con = w.consf . In the example (6), we note that lexical meaning and actual meaning coincide. If we consider all relations other than equality as deviant, we characterize a class of phenomena that is certainly larger than the one containing gurative speech only. Example (7) 0 Man (7) \A policeman left without paying. The man lost his job." illustrates an anaphoric relation between \the man" and \a policeman". With this, a subsumption relation holds between w.consf (= Policeman) and w.con (= Man), which means that w.con is either more general or equal with w.consf . In particular, we have (policeman-2 =) w.sf 2 w.con, but not w.sf instance-of w.con, in general (as in example (6)). Loosening ties a bit more, we may abandon the subsumption relation between w.consf and w.con as in example (8): (8) \A policeman left without paying. The fool lost his job." We have (policeman-2 =) w.sf 2 w.con (= Fool), but the specialization relation between w.consf (= Policeman) and w.con (= Fool) no longer holds. Instead, we are set back to w.sf 2 w.consf \ w.con and, therefore, w.consf \ w.con 6= ;. We say that w.consf and w.con are compatible, as no categorization conict arises. This also holds for the previous examples in this section. So, the notion of categorization conict turns out to become crucial for our distinction between literalness and gurativeness | the latter being based on an underlying categorization conict, whereas the former is not. We summarize these observations in the following denition: Denition 1 (Literalness via Syntactic Constraints). A word w in an utterance U is used according to its literal meaning, if for every instance w.sf 2 I which w StandsF or, one of the following two conditions hold: w:sf = w:i w:sf 2 w:con Especially, w.consf if w is a proper name else (1) (2) \ w.con 6= ; holds for non-proper nouns. We here restrict the notion of gurative speech to those relationships between w.sf and the lexical meaning of w in terms of w:con, which are not inclusive ones. A literal use of the word w for an instance w.sf instance-of w.consf is only possible, if w.consf \ w:con 6= ;. If, however, a categorization conict occurs, i.e., w.consf \ w:con = ; holds, we call the use of w gurative. Such a gurative use is illustrated by the word \ham sandwich" in example (2) or \gold mine" in (3). Two consequences of this denition deserve special mention: 1. We can determine exactly the place where subjectivity comes in when a distinction between literalness and gurativeness is made | it is mirrored by subjectivity in categorization. \fool" in example (8) can only be considered as literal, if the concepts Fool and Policeman are considered as being compatible (in the settheoretic sense introduced above). If one does not share this conceptualization, this usage of \fool" must be considered as gurative (or even absurd). Thus, we capture the subjectivity of gurativeness formally in the ontological premises, not via intuitive considerations. 2. It is also important to note that Denition 1 does not depend on the violation of selectional restrictions. The example (5) (\I like Chaucer.\) allows for the same analysis as example (4) (\I read Chaucer.\), because the intended patient of like are, in both cases, Writings-by-Chaucer (= w.sf), although this is not indicated by selectional restrictions at all. In both cases w.consf \ w:con = ;, i.e., gurativeness holds. 4 Granularity The (non-)inclusion criterion we have set up for the distinction between literal and gurative usage of words in Denition 1 introduces a strong tie to the underlying ontology. The problem this might cause lies in granularity phenomena of domain knowledge bases and in the general question whether every dierence in conceptualization induces dierent literal { gurative distinctions. Given different levels of granularity, it may well happen that a word w StandsF or an instance w.sf instance-of w.consf with w.consf \ w:con = ;, though, intuitively, one would rate the usage of w as a literal one. Let us illustrate this case with two examples. Assume we have a knowledge base KB1 in which Cpu happens to be part-of the Motherboard, while Motherboard itself turns out to be part-of Computer. If we analyze the example (9) \The CPU of the computer . . . " accordingly, we end up with the determination of a gurative usage for \computer", since Motherboard \ Computer = ; (cf. Fig. 2). Fig. 2. Example (9) Assuming KB1 If we assume, however, an ontological representation in a domain knowledge base KB2 such that Cpu is an immediate part-of the Computer, then we derive a literal usage for w (cf. Fig. 3). Fig. 3. Example (9) Assuming KB2 To get rid of the dependence on knowledge base granularity, to a certain degree at least, we may derive a weaker condition of literalness from Denition 1. To achieve this, we state that w.sf and w .sf are related by a conceptual relation r (technically, w .sf r w.sf). Thus, for literal usage of w we require: 0 0 w0 :sf r w:i if w is a proper name 9i 2 w:con: w0 :sf r i else (3) (4) (3) immediately follows from (1) in Denition 1, since w .sf r w.sf (= w.i) holds. (4) can be deduced from (2) by dening i := w.sf. Since these conditions provide weaker conditions of literal language use than the ones we have agreed upon in Denition 1, all literal usages determined by the strong condition still remain literal (in particular, example (9) is considered a literal usage of \computer" given KB2 ). Considering the granularity eects for example (9) with respect to KB1 , however, we may determine the literal usage of \computer" by the following consideration. As we know that Cpu is partof Motherboard, and Motherboard is part-of Computer, we conclude 0 with the transitivity of the part-of relation4 that Cpu is part-of Computer. Hence, criterion (4) is fullled. In contradistinction to the examples discussed previously, we do not have w.sf 2 w.con (cf. criterion (2) from Denition 1), as w.sf = motherboard-2 holds. So by moving from the strict criteria in Denition 1 to the weaker ones stated by criteria (3) and (4) we are able to incorporate granularity phenomena of knowledge bases. Note however that we cannot abandon completely our dependence on (the encodings of) the knowledge layer, since we exploit knowledge about the transitivity of relations to solve the problem. 5 Figurativeness and Reference One might argue that the problem just discussed, the dependence of the distinction between literal and gurative usage on knowledge base structures, follows from the denition of StandsF or. Accordingly, some researchers [19] have proposed to build the denition of gurative speech upon the notion of reference. The assumption being made is that w uniquely refers to a knowledge base item w.ref instance-of w.conref and that gurativeness results from the deviation of this reference from literal meaning. Although their notion of deviance is not formalized, referentially-based literalness can now be dened straightforwardly in our approach by proceeding exactly along the lines of Denition 1: Denition 2 (Literalness in the Referential Approach). A word w is called literal in the referential approach, w:ref = w:i w:ref 2 w:con if w is a proper name else if: (5) (6) Without doubt, we here circumvent the granularity problem, since no change in reference occurs for example (9), no matter whether KB1 or KB2 is assumed.5 But the reference approach runs into severe problems when one considers, e.g., classical examples of metonymies such as (10) \I like to read Chaucer. He was a great writer." We have w = \Chaucer" as a typical example for a writer-for-writings metonymy [19]. The assumption to link literal/gurative usage to reference relations is awed by the fact that w = \Chaucer" does not refer to the \works of Chaucer", because in this case the referentially determined anaphor \He" could not be resolved. In particular, we have Chaucer.ref = Chaucer, therefore w.ref = w.i. Hence, \Chaucer" must be considered, counterintuitively, as a literal use according to criterion (5) (similar problems have been discussed at length by Stallard 4 5 We are aware of empirical observations about the transitivity of part-whole relations [3, 31], in particular those pertaining to the claim that any of the subrelations of partwhole are transitive, while the general part-whole relation usually is not. When we talk about the transitivity of part-whole, we mean this constrained type of transitivity. Note that this denition is, nevertheless, still dependent on the knowledge base and on the lexical meaning of w. [27]). Given our context-dependent denitions (1 or 3), we get w.sf = Writingsby-Chaucer so that w.sf 6= Chaucer. Thus, w = \Chaucer" is analyzed gura- tively in our approach.6 Summarizing, we combine criteria (1) to (6) by the following conditions: 1. A word w is used in its literal meaning in all denitions, if w.sf 2 w.con (analogously, w.sf = w.i) for all w.sf, and w.ref 2 w.con hold (combining the referential Denition 2 and Denition 1 with respect to literal usage). 2. If w.sf 62 w.con (analogously, w.sf 6= w.i) for some w.sf (with respect to a relation r and another word w ), but w.ref 2 w:con (w.ref = w.i), two cases must be distinguished: { In cases of granularity eects criterion (4) holds. By this, an i 2 w.con exists with w .sf r i (analogously, w .sf r w.i). We can include this in our denition of literal usage as its analysis is only due to implications a particular ontology design brings to bear. { In cases of gurative speech like the one in example (10) the criteria (4) / (3) do not hold. We may include these cases into our denition of gurative usage: 3. A word w is used in its gurative meaning according to the syntactic and the referential denition, if w.ref 62 w.con holds and there exists a w.sf 62 w.con. This is the case, e.g., in example (2). So far, we have only considered the gurative usage of a word w. We may end up by dening U as a gurative utterance, if it contains at least one word w which is used in a gurative way. 0 0 6 0 Putting Theory to Practice { Metonymy Resolution So far, we considered the gurativeness from an entirely theoretical perspective. Our interest in this topic can be traced back, however, to practical requirements which emerge from SynDiKATe, a text understanding system for processing product reviews from the IT domain as well as medical nding reports [10]. The need to deal with a particular type of gurative language, metonymic expressions, becomes evident when one considers the quantitative distribution of referential text phenomena we determined for a random sample of 26 texts. In 606 utterances we encountered 103 metonymies together with 291 nominal anaphora and 351 bridging anaphora [20]. With a metonymic expression encountered in almost every sixth utterance, an uncontroversial need for dealing with this problem is demonstrated. We also collected empirical evidence that a 16% increase in the accuracy of the anaphora resolution procedure can be achieved by the incorporation of a metonymy resolution process (due to the overlapping of metonymies and nominal anaphora). SynDiKATe builds on a fully lexicalized dependency grammar [11]. Parse trees link lexical nodes via a well-dened set of dependency relations (cf. genitiveattribute as illustrated in Figures 2 and 3). For semantic interpretation [25], the 6 A methodology for intersentential anaphora resolution, given these interpretations, is described in [28]. conceptual correlates of the lexical items in the dependency graph need to be linked via conceptual relations. This linkage is based on conceptual constraints. One major source of these constraints are dependency relations, since they are often characterized by a hard-wired mapping onto general conceptual relations. For instance, the dependency relations subject or direct-object, by convention, map to the conceptual relations agent and patient or patient, respectively. There are types of dependency relations, e.g., genitive-attribute, which have no associated conceptual constraints. In such cases, a fairly unconstrained search of the knowledge base for proper linkage is performed resulting, e.g., in the determination of the part-of relation (chain) between Cpu-1 and Computer-2 in Fig. 3 for the lexical correspondents \Cpu" and \computer" linked via genitive-attribute. We compute possible StandsFor relationships relevant for the distinction between literal and gurative language by looking at the paths found as a result of the search of the knowledge base for conceptual relations. Especially, we look at the intermediate concept nodes a path contains in order to determine the compatibility between those concept and the lexical meaning. For illustration purposes, a few examples of paths will be discussed subsequently. When the dierent interpretations are computed and distinguished, these distinctions can then be used for disambiguation and preferential ranking.7 Consider the following two examples already introduced in this paper. (11) \The ham sandwich left (without paying)." (12) \(I) like Chaucer." In example (11) we concentrate on the subject dependency relation between \ham sandwich" and \left". As already remarked, subject maps conceptually to agent (for active voice) or patient (for passive voice) roles. Concentrating on the active voice case, the concept Leave has associated a conceptual role, leave-agent, a subrole of agent, which has a range constraint on Persons. A Person may order a Product, a possible product being eatables such as Ham-Sandwiches. A possible linkage between the conceptual correlates of \ham sandwich" and \left" is, therefore, given by the following path: p11 = (Leave { leave-agent { Person { orders { Ham-Sandwich). In example (12) we concentrate on the patient relation between \like" and \Chaucer" as induced by the direct-object relation. In this case, the concept Like is associated with the conceptual role, like-patient, a subrole of patient, which has no range constraint other than the most general concept AnyThing. AnyThing subsumes Persons, which Chaucer is an instance of, so that the path p112 = (Like { like-patient { Chaucer) just expresses that Chaucer as a person is liked. In addition, Anything also subsumes Books, which are written-by a particular Author, one of those being Chaucer. In this case, we have constructed another path: p212 = (Like { like-patient { Book { written-by { Chaucer). So when possible paths between w.con and w0.con have been determined the proposed gurativeness criterion can be used to determine whether the readings 7 Of course, we restrict the search through the knowledge base, e.g., by conceptual connectivity criteria (cf. [20] for details). that can be derived from the path are gurative or literal. Let us look at our example path p11 = (Leave { leave-agent { Person { orders { Ham-Sandwich) again. Possibilities for the StandsF or items left.sf and ham-sandwich.sf and the conceptual relation r corresponding to the syntactic subject relation are { left.sf 2 Leave (= Left.con), r = leave-agent and sandwich.sf 2 Person (with Person \ Ham-Sandwich = ;). { left.sf 2 Person (with Person \ Leave = ;), r = orders and sandwich.sf 2 Sandwich (= Sandwich.con). The rst possibility gives rise to a literal reading of left and a gurative reading of ham sandwich when compared to our criterion whereas the second gives rise to a gurative reading of left and a literal reading of ham sandwich. In the Chaucer example possibilities for the StandsF or items like.sf and chaucer.sf and the conceptual relation r corresponding to the syntactic directobject relation include { like.sf 2 Like (= like.con), r = like-patient and chaucer.sf = Chaucer (= Chaucer.i). This is the only reading that can arise by taking the path p112 into account as there are only two concepts and one relation in the path. The next two readings arise when looking at p212 : { like.sf 2 Like (= like.con), r = like-patient and chaucer.sf 2 Book (with Book and Chaucer.i being incompatible). { like.sf 2 Book (with Book \ Like = ;), r = written-by and chaucer.sf = Chaucer (= Chaucer.i). The rst possibility accounts for a literal reading of Chaucer and like, whereas the second and third account for a gurative reading for Chaucer or like. It is important to note that the disambiguation of dierent readings may proceed along several lines such that this process is not aected by the criterion of gurativeness itself. The determination of possible readings can, e.g., be combined with all sorts of disambiguation criteria or search heuristics such as 1. Preference of literal readings over gurative ones, thus preferring the rst reading in the Chaucer example over the other two readings. One can show that an approach based on this criterion only amounts to an approach considering selectional restriction violations as necessary for establishing gurative readings. 2. Additional world knowledge that determines that a literal reading in the Chaucer example is invalid when uttered in 1999. 3. Knowledge about anaphora and discourse relations might be helpful for the Ham-Sandwich example, e.g., such that a record of discourse entities is kept from which one may derive that a person has ordered a sandwich who has already been introduced into the discourse. So that person is likely to be referred to again, thus making the gurative reading of ham sandwich more likely than the gurative reading of left. 4. Statistical knowledge about linguistic structures that make it more likely that a noun phrase be used metonymically than a verb phrase, thus preferring the second reading over the third reading in the Chaucer example. All these criteria can be combined with our gurativeness criterion so that the criterion can also provide a testbed for preference heuristics and search restrictions when analyzing gurative language. In the framework of SynDiKATe, we have, rst, developed path patterns inferred from the gurativeness criterion that make the distinction between gurativeness and literalness based on paths between knowledge base concepts easier to determine. A simple example would be that paths of length 1 always mirror literal readings of both lexical items involved.8 We then developed various search and disambiguation heuristics which take into account the following features (for details, cf. [20]): { Common (or, typical) gurative relations are preferred over not so common ones, thus preferring paths including patterned relations. { Discourse criteria preferring readings allowing for anaphoric readings over ones that do not. { Only when those criteria do not apply, we prefer literal readings over gurative ones. It is important to note that these kinds of comparisons can only be achieved when a well-founded distinction of gurative and literal language is possible, which should not rely on the heuristics themselves, e.g., restricting gurative speech to cases where a sortal conict occurs. Instead, an independent ontological criterion is necessary that can be combined freely with reasonable heuristics. 7 Related Work We consider as the main contribution of this paper the introduction of a formal notion of deviance that is general and simple. To the best of our knowledge, no comparable work has been done so far on this issue. Although there exist formal characterizations of metaphors [13, 7] in the framework of analogical reasoning, these are entirely self-contained, i.e., they account for structural properties of metaphors (e.g., constraints on domain mappings, aptness conditions), rather than dealing with the distinction between literal and gurative speech. Even more serious from a natural language understanding point of view, concrete metaphorical utterances cannot be translated into the formal systems provided for metaphor explanation due to the lack of concrete mappings between the lexical and conceptual level [13]. The state of the art aiming at a clarication of what constitutes, e.g., a metonymy is characterized by a quote from Lako and Johnson [19]. They nd that it is characterized by the use of \one entity to refer to another that is related to it". It is not clear at all what kind of entities they are talking about. One must 8 There are other literal paths as well which account for granularity problems. assume that they are referring to some kind of signs because it is not obvious how other entities can refer to something. But even if we restrict the entities to linguistic signs there is no restriction on the kind of relatedness between the objects. For example, relatedness might include class inclusion, similarity or partwhole relations, but only the latter are included in metonymy in general and the examples Lako and Johnson put forward suggest that it is this conventional kind of metonymy they are talking about. The same shadowy denitions of gurative language are then often adopted by theoretical linguists [29, 17, 30, 23], as well as computational linguists [5, 21]. This leads to the fact that it is mostly not clear at all, which phenomena are treated by these approaches and how they discriminate dierent varieties of gurative speech from others. In addition, a tendency can be observed in more formal approaches { pressed by the need to nd computationally feasible denitions of metaphor or metonymy { to consider gurative language a violation of selectional restrictions [2, 5, 12, 24] or communicative norms [9, 26]. Such an approach equates an often used triggering condition, viz. constraint violation, with the phenomenon of gurative language (or, subsets, like metonymies). So, it confuses the possible, but not necessary eects of a phenomenon with the phenomenon to be explained. Despite the lack of formal rigor in previous work, it is worth to investigate how our formal criterion is compatible with other views on gurative speech from cognitive linguistics in general. The tendency to see gurative speech rooted in conceptual categories, as we do, is becoming consensus. The main trend is, e.g., to treat metaphors as a means of categorization by way of similarity [8] and, in principle, to retrace gurative speech to cognitive procedures involving categorization and (subjective) experience [19, 6, 18]. So, Lako and Johnson see metaphors rooted in our way of conceptualization via mappings. Kittay [17] and Turner [30] regard some kind of conceptual incompatibilities as the basis of metaphorization. Nevertheless, they do not explicate their theory of categorization and incompatibility nor do they recognize that these incompatibilities are relevant for other kinds of gurative speech, as well as for metaphors in the strict sense of the word. The dependence of lexical, literal and gurative meaning on ontologies is, therefore, realized, but no explicit formal treatment is given of particular problems resulting from the ontological presuppositions. We here regard the problem of granularity arising form ontological dependence and subjective experience and propose a solution by relaxing criteria for literal meaning. The second major contribution of the paper derives from the formal status of our distinction between literal and gurative meaning. Once such a formal notion is given, it allows us to characterize subjectivity, so far an entirely informal notion, by reference to the particular ontology underlying the language understanding process. The current perspective on the ontological basis is rather static, viz. dierent ontologies account for cases in which one person considers an utterance to be literal, while another considers it to be gurative. Our proposal introduces exibility into such an assessment. We aim at adapting dierent ontologies such that by way of abstracting away dierent granularities of representation structures (e.g., by generalizing more ne-grained representations to a coarser grain size, as in criterion (4)) disagreement might turn into consensus (e.g., considering example (9)). Contrary to that, the majority of researchers in our eld of study attribute the dierence in opinion to the existence of dierent, incompatible ontologies, and leave it with that explanation without further attempt at smoothing [19, 30]. The third major proposal we make relates to the contextual embedding of gurative speech. The criterion we formulate is based on syntactic relations only that guide conceptual interpretation. In particular, and unlike most algorithmic accounts [22, 5, 12, 24], it does not rely at all upon the violation of selectional restrictions (for a notable exception, cf. [21]), since this criterion accounts only for a subset of the phenomena naturally recognized as gurative language. In addition, the syntax-based proposal we make avoids to consider reference changes as an indicator of gurativeness as is commonly assumed [19]. It is inspired by Fauconnier's [6] \connector" function. Though this proposal aims at an embedding of gurative language into syntax, there exists no formalization of this notion in relation to an established grammar framework as we do nor is the notion of conceptual incompatibility formalized. A more formal criticism of the reference changes proposal was made by Stallard [27] who, nevertheless, then only dealt with gurative language, violating sortal constraints. Our approach is also compatible with viewing gurative language as regular and not violating linguistic norms. Whereas literal language is grounded in inclusion relations to lexical or context-independent meaning, gurative language is grounded in other relations to lexical meaning. These can, nonetheless, be systematic and conventionalized relations like part-whole relations or obvious similarities. Although we have not spelled out these relations, there is no need to claim that inclusion relations are prior or preferred to other relations. Literal as well as gurative speech are both grounded in structured relations in categorization. This is in accordance with the conventional metaphor view rst stipulated by Lako and Johnson. It is also in accordance with psycholinguistic research showing that gurative speech is in most cases as easily understood as literal speech. This is especially the case if the instance of gurative speech is conventional, i.e. grounded in conventional, systematic and pervasive ontological relationships [1]. The essence of this is that pervasive and structured relations or relations made salient by the context [14] may be as easily available to comprehension as inclusion relations. 8 Conclusion In this paper, we have drawn a distinction between literal and gurative speech which is based on formal criteria. These are grounded in the solid framework of description logics, in particular, by relying on its settheoretical semantics. We aimed at a comprehensive distinction, one that covers the whole range of word-centered gurative language use (i.e., metonymies, metaphors, irony) using general and simple criteria. A major determinant for gurative interpretation is given by the context of an utterance. As a consequence, contextual criteria are at the core of any adequate account of gurativeness. Our contextual criteria are independent of the notion of selectional restrictions, but dependent on the conceptual interpretation of syntactic relations, in general. Another crucial condition of whether language use is considered literal or gurative is introduced by the particular ontology referred to. While earlier formalist approaches appeal to semantic types, sortal constraints, etc., this is not fully convincing, since the entire structure and granularity of the theory of the domain being talked about contributes to the understanding process, whether literally or guratively based. In particular, we captured the notion of subjectivity in ontological premises and explained how granularity problems may be overcome. The model we presented does currently not account for neologisms, as those have no a priori lexical meaning, and many tricky cases of quantication and the use of proper names. From a more technical perspective, we have also not scrutinized the dierent kinds of relations that are still required to hold between w.consf and w.con, if w.consf \ w.con = ;. So, a necessary condition for gurative speech has been established that needs to be supplemented by suÆcient ones. We also have no criteria available right now to distinguish between various types of gurative speech (e.g., metaphors vs. irony). By this, we mean specications concerning similarity relations for metaphors, contrast relations for irony, and contiguity relations for metonymies (cf., e.g., [16]). Finally, we stop short of distinguishing between innovative gurative speech (like in the ham sandwich example) and conventionalized gurative speech (systematic polysemy [24, 23]). Acknowledgements. K. Markert was a member of the Graduate Program Human and Machine Intelligence at Freiburg University, funded by DFG. References 1. Dawn G. Blasko and Cynthia M. Connine. Eects of familiarity and aptness on metaphor processing. Journal of Experimental Psychology: Learning, Memory and Cognition, 19(2):295{308, 1993. 2. Jaime G. Carbonell. Metaphor: an inescapable phenomenon in natural-language comprehension. In W. G. Lehnert and M. H. Ringle, editors, Strategies for Natural Language Processing, pages 415{434. Hillsdale, N.J.: Lawrence Erlbaum, 1982. 3. Roger ChaÆn. The concept of a semantic relation. In A. Lehrer and E. F. Kittay, editors, Frames, Fields and Contrasts, pages 253{288. Hillsdale: Erlbaum, 1992. 4. D. A. Cruse. Lexical Semantics. Cambridge: Cambridge University Press, 1986. 5. Dan C. Fass. met : a method for discriminating metonymy and metaphor by computer. Computational Linguistics, 17(1):49{90, 1991. 6. Gilles Fauconnier. Espaces Mentaux. Paris: Les Editions de Minuit, 1984. 7. Dedre Gentner, Brian Falkenhainer, and Janice Skorstad. Viewing metaphors as analogy: the good, the bad, and the ugly. In Y. Wilks, editor, Theoretical Issues in Natural Language Processing, pages 171{177. Hillsdale: Lawrence Erlbaum, 1989. 8. Raymond Gibbs. Categorisation and metaphor understanding. Psychological Review, 99(3):572{577, 1992. 9. Herbert P. Grice. Logic and conversation. In P. Cole and J. Morgan, editors, Syntax and Semantics, volume 3, pages 41{58. New York: Academic Press, 1975. 10. Udo Hahn and Martin Romacker. Text understanding for knowledge base generation in the SynDiKATe system. In Database and Expert Systems Applications. Proceedings of the 10th International Conference { DEXA'99. Florence, Italy, August 30 - September 3, 1999. Berlin: Springer, 1999. 11. Udo Hahn, Susanne Schacht, and Norbert Broker. Concurrent, object-oriented natural language parsing: the ParseTalk model. International Journal of HumanComputer Studies, 41(1/2):179{222, 1994. 12. Jerry R. Hobbs, Mark E. Stickel, Douglas E. Appelt, and Paul Martin. Interpretation as abduction. Articial Intelligence, 63(1/2):69{142, 1993. 13. Bipin Indurkhya. Constrained semantic transference: a formal theory of metaphors. In A. Prieditis, editor, Analogica, pages 129{157. Morgan Kaufmann, 1988. 14. A. Inho, S.D. Lima, and P.J. Carroll. Contextual eects on metaphor comprehension in reading. Memory and Cognition, 12:558{567, 1984. 15. Ray Jackendo. Semantic Structures. Cambridge, MA: MIT Press, 1990. 16. Roman Jakobson and Morris Halle. Fundamentals of Language. The Hague, NL: Mouton Publishers, 1956. 17. Eva F. Kittay. Metaphor: Its Cognitive Force and Linguistic Structure. Oxford, U.K.: Clarendon Press, 1987. 18. George Lako. Women, Fire, and Dangerous Things. What Categories Reveal about the Mind. Chicago, IL: University of Chicago Press, 1987. 19. George Lako and M. Johnson. Metaphors We Live By. Chicago, IL: University of Chicago Press, 1980. 20. Katja Markert and Udo Hahn. On the interaction of metonymies and anaphora. In IJCAI'97 { Proceedings of the 15th International Joint Conference on Articial Intelligence, pages 1010{1015. Nagoya, Japan, August 23-29, 1997. 21. James Martin. Computer understanding of conventional metaphoric language. Cognitive Science, 16:233{270, 1992. 22. Peter Norvig. Marker passing as a weak method for inferencing. Cognitive Science, 13(4):569{620, 1989. 23. Georey Nunberg. Transfers of meaning. Journal of Semantics, 12:109{132, 1995. 24. James Pustejovsky. The Generative Lexicon. Cambridge, MA: MIT Press, 1995. 25. Martin Romacker, Katja Markert, and Udo Hahn. Lean semantic interpretation. In IJCAI'99 { Proceedings of the 16th International Joint Conference on Articial Intelligence. Stockholm, Sweden, July 31 - August 6, 1999. 26. John R. Searle. Metaphor. In Andrew Ortony, editor, Metaphor and thought, pages 93{123. Cambridge, U.K.: Cambridge University Press, 1979. 27. David Stallard. Two kinds of metonymy. In Proceedings of the 31st Annual Meeting of the Association for Computational Linguistics, pages 87{94. Columbus, Ohio, USA, 22-26 June, 1993. 28. Michael Strube and Udo Hahn. Functional centering. In ACL'96 { Proceedings of the 34th Annual Meeting of the Association for Computational Linguistics, pages 270{277. University of California at Santa Cruz, California, 24-27 June, 1996. 29. Roger Tourangeau and Robert Sternberg. Understanding and appreciating metaphors. Cognition, 11:203{244, 1982. 30. M. Turner. Categories and analogies. In D. H. Helman, editor, Analogical Reasoning, pages 3{24. Dordrecht, NL: Kluwer Academic Publishers, 1988. 31. M. Winston, R. ChaÆn, and D. Herrmann. A taxonomy of part-whole relationships. Cognitive Science, 11:417{444, 1987. 32. William A. Woods and James G. Schmolze. The Kl-One family. Computers & Mathematics with Applications, 23(2/5):133{177, 1992.
© Copyright 2025 Paperzz