2009 Eighth International Symposium on Natural Language Processing A Preliminary Study of Comparative and Evaluative Questions for Business Intelligence Nathalie Rose T. Lim, Patrick Saint-Dizier, Brigitte Gay, and Rachel Edita Roxas answer is not directly lifted from the source text. Instead, natural language text is constructed from the results of the processing. New types of questions like comparative and evaluative questions are targeted for research as indicated in [1]. It is of interest to study comparative and evaluative expressions (in questions) because of the challenges and issues associated with processing them. These include the following aspects: 1) Multiple styles: Comparative expressions may be expressed in different ways. They can be bipredicational expressions, cross-class comparisons, degree comparisons, and these can appear in nouns, verbs, adjectives, adverbs, and even implicitly denoted. For this research, the focus will be on degree comparisons and explicit and implicit denotations from nouns (like maturity from reached maturity), verbs (like win), adjectives (like good strategy), and adverbs (like fast in evolve fast), since these types appear in actual questions raised in the domain being considered. Degree comparisons refer to the extent of applicability of a certain comparative expression or predicate. Samples of which are the predicates better and active. 2) Inferencing synonymous terms: Determining the synonyms and what these entail is an issue not specific to comparative and evaluative QA, but also to general QAs. However, in certain domains, terminologies have different or more specific semantic meaning due to the context. For example, a hub entails different things depending on the context (eg., transportation hub where hub is a location versus transaction hub where hub is a company). 3) Accessing semantic dimension: The semantic dimension being referred to here is the list of quantifiable measures, properties, and criteria that are associated with the comparative expression or predicate. For example, expensive is associated with the property of cost. 4) Determining ranges and limits for comparison: Values of identified properties or criteria to be used for comparison between objects can be taken from various source texts. However, evaluation (not comparison) of certain criteria are more complex if there is no set standard of measurement and is dependent on the object being evaluated. Using expensive as the example, the ranges of values for determining if a book is expensive is different with Abstract—Comparative and evaluative question answering (QA) systems provide objective answers to questions that involve comparisons and evaluations based on a quantifiable set of criteria. As evaluations involve inferences and computations, answers are not lifted from source text. This entails the need for correct semantic interpretation of comparative expressions, converting them to quantifiable criteria before data can be obtained from source text, processing these information, and formulating natural language answers from the result of the processing. As business intelligence (BI) requires comparisons and interpretations of seemingly unrelated facts, a QA system for this domain would be beneficial. This paper presents a study of some comparative and evaluative questions that are raised in the domain of business intelligence. How these questions are processed is also discussed. I. INTRODUCTION C ONSIDER the following questions: “Which European companies had the most alliances in year 2008?” and “Did Company X take more risk than Company Y in the past year?”. The first question is an evaluative question. An evaluative question involves the computation or evaluation of at least one property or criteria. In this case, the criteria are explicitly stated in the question (i.e., most alliances which is equivalent to the number of alliances). However, in many cases, the properties involved are not explicitly stated, as in the case of the predicate take-risk in the second question. The predicate take-risk would have to be broken down to the properties such as number of transactions and the types of partners. Basis and constraints are defined by an expert in business intelligence. Evaluations and computations can be done for different objects for comparison purposes. And this comparative question is depicted in the second example question above. Thus, comparative and evaluative question-answering (QA) involves inferences in terminology, determining the properties involved for evaluation, and computation and comparison before an answer can be given. As such, the Manuscript received August 12, 2009. N. Rose T. Lim is affiliated with both De La Salle University in Manila, Philippines and Universite Paul Sabatier in Toulouse, France. She can be reached through phone: (632)524-0402; fax: (632)536-0278; e-mail: [email protected]. P. Saint-Dizier is with IRIT, France. (e-mail: [email protected]). B. Gay is with Groupe ESC Toulouse in France. (e-mail: [email protected]). R. Edita Roxas is with De La Salle University – Manila in the Philippines. (e-mail: [email protected]). 978-1-4244-4139-6/09/$25.00 ©2009 IEEE 35 extracted or derived from other web sources. From a basic question to compare company transactions like: “Which companies have the most number of transactions?”, there could be variations and additional constraints added to it. The following subsections list the different classifications of comparisons that may be combined to form a single comparative or evaluative question. determining if a house is expensive. The next section discusses the domain of business intelligence (BI) and types of questions that can be raised. II. BUSINESS INTELLIGENCE Business intelligence (BI) is an area in business and economy which aims at identifying trends in business and in any kind of strategic development (e.g., research themes, political orientations) from thousands of seemingly isolated facts. The globalization of markets for technology as well as fast innovation diffusion through complex networks of business relationships have created a major competitive challenge to corporate leaders. Companies and governments are experimenting with new approaches to the management of business relationships. Corporate strategies involving mergers, acquisitions, spin-offs, and a plethora of alliances are creating smaller, decentralized operational units within and across the boundaries of companies or countries. One role of BI is to help them (companies and governments) understand and master their position in global industries [2]. Software tools can be used to facilitate the analysis that they need to make. This entails that information is processed and structured to display extracted entities and semantic relationships among them. Therefore, there are at least two types of software tools that are necessary. One is a graphical tool that displays coarse-grained information, e.g., all commercial links between companies or countries. This is a kind of radiography of a situation over a certain period (generally, a year) with thousands of links between nodes representing companies. The graphs allow spatial analysis to identify business units or alliances which are larger than just companies. Evolution over a few years is often of much interest. A number of software are now able to handle this, among which is Tetralogie [3]. The other type of tool involves a more fine-grained analysis based on knowledge base constructed from news and other data. It includes determining requirements from the question, extracting the information from the knowledge base as per requirement, and processing these to derive the answer. These may be implemented through database queries (e.g., via SQL statements). However, this implies that the set of questions that can be raised are predetermined. Also, queries are far less natural and user-friendly than human language and these do not allow generation of cooperative responses. Thus, a QA system is better suited to the needs of the users in this domain. Though there are several QA systems, questions mainly focus on factoid, definitions, or lists. Comparative and evaluative questions are seldom tackled [4],[5]. For this study, the corpus is the set of news relevant to biotechnologies from years 2004 to 2007. The source information revolves around transactions between companies. Thus, answers to questions are mainly based on information extracted and processed from these news articles. Other needed information not present in the news is A. Spatial Scope Questions may include spatial qualification of the company or the transaction. Sample stem questions could be in (but not limited to) the form of: 1) Which companies in Asia … 2) Which … transactions in Europe … 3) Which cities in … 4) Which country … 5) Which continent … B. Categorial Scope Companies are categorized into public or private and products involved in the transactions fall under certain sectors. Sample stem questions could be in the form of: 1) Which <category> companies … 2) Which … in the <sector name> sector … C. Temporal Scope In BI, the temporal aspect is crucial. It specifies the scope of the analysis to be done. Sample stem questions could be in the form of: 1) … in <year>? 2) … from <start> to <end>? <start> and <end> may be exact dates, but usually is indicated as the inclusive years. The <end> may also be specified as “present”. Finally, it is also possible that the temporal scope is implicit depending on the criteria. D. Directly Quantifiable Criteria Questions that involve computations before the answer can be discerned may involve a combination of different directly quantifiable criteria, like: 1) Number of transactions (may be all transactions in general or a specific transaction) 2) Number of partners 3) Amount involved in the transaction 4) Number of products E. Non-Directly Translatable Criteria Some adjectives may be used to encompass a series of criteria. These have different semantic meanings and interpretations depending on the domain (or even the expert). Some terms include: 1) active (as in active companies) 2) stable (as in stable partners) 3) risky (as in risky transactions) 4) innovative (as in innovative products) 5) fast (as in fast evolution) 36 On the domain of BI, MUSING [12] aims to use the semantic web and combine with rule-based and statistical methods for knowledge acquisition and reasoning for providing financial analysis complying with Basel II requirements. The presentation in terms of input expected (whether these are natural language questions and whether these involve comparative expressions) and the output to be generated are undisclosed from the available documentation. In the next section, we show some related works and studies on the semantic meaning of comparatives, applications involving comparative expressions, and QA related to BI. In section 4, a discussion on how questions are processed is presented. The discussion includes details of how comparative and evaluative expressions are categorized and interpreted. Lastly, we conclude with issues that we have considered and the research directions we plan to take. IV. PROCESSING COMPARATIVE AND EVALUATIVE QUESTIONS III. RELATED WORKS General QA systems involve the processes of question analysis, information retrieval, answer determination, and response generation. For comparative and evaluative QA systems, the processes are redefined. The question analyzer must identify the comparative expressions in the question and decompose it into meaningful constituents, among which are those properties that will be evaluated. When predicates are decomposed into properties, then pertinent information can be extracted from sources (either already stored in database or additional information is mined from the web) and evaluation can be done in the answer determination phase. The properties and the evaluation criteria or rules are specified based on definitions given by an expert. Since the answer is not lifted from the source text, the response generator is in-charge of producing natural language text from the resulting computation and evaluation results. The succeeding subsections outline the processing of the source texts, the question, and the interpretation of a selected set of comparative and evaluative expressions. Comparisons may be in relation to properties within the same object, degree of comparisons of the same property between different objects, or different properties of different objects [6]. The properties at stake in the comparison are embedded in the semantics of the words in the question, and possibly in the context that comes with the question. To date, there is obviously no widely available lexical resource containing an exhaustive list of comparative predicates, applied to precise terms, together with the properties involved. These can possibly be derived, to a limited extent, from existing resources like FrameNet [7] or from an ontology where relationships between concepts and terms can be mapped. However, this is tractable for very simple situations, and in most cases, identifying those properties is a major challenge. Friedman [8] presents a general approach to process comparative expressions by syntactically treating them to conform to a standard form containing the comparative operator and the clauses that are involved in the comparison. Another approach would be to automatically extract comparative relations in sentences via machine learning. In [9], the approach used is to determine whether the expression is non-equal gradable, equative, or superlative. By identifying the type of expression, the type of comparison may be determined from the semantics of the predicate and the properties of the objects through the pairability constraints. What is missing is the exploration on semantic and conceptual issues and their dependence to context, users, and domains. Olawsky [10] attempts to study the semantic context by generating a set of candidate interpretations of comparative expressions. Then, the user is prompted to choose among these to specify his intent. Some QA systems, like [11], can handle comparative expressions including cross-class comparisons on a range of different domains. However, these involve having a different backend knowledge representation system and the frontend QA system has to be customized before it can answer queries in the new domain. In addition, both of these systems only consider comparisons based on quantifiable predicates (i.e., those measurable by count, mass, or value). Also, predicates with non-directly translatable properties that are dependent on domain or context, to our knowledge, have not been explored. A. Processing Source Text We are considering the set of economic news in biotechnologies as our main source of information. Each news article is between 80 and 200 words long and is written in English. An excerpt of a news article (from www.biospace.com) is as follows: IDDI (INTERNATIONAL DRUG DEVELOPMENT INSTITUTE) AND CYTEL INC. TODAY REPORTED ENTERING INTO A STRATEGIC TECHNOLOGY COLLABORATION. THE COMPANIES ARE COOPERATING TO DEVELOP INTEGRATED RANDOMIZATION OF SYSTEMS TREATMENT AND SERVICES ASSIGNMENTS FOR FOR THE PATIENTS PARTICIPATING IN CLINICAL TRIALS… As can be seen, sentences are long and verb forms may be quite complex and indirect. Each sentence is composed of a main predicate pred, which serves as a head, and arguments arg to the predicate are defined by their thematic roles t. The argument may be a string of words representing a noun phrase, a prepositional phrase, or a clause. ROL(s) = { t(arg , pred) | t ∈ {agent, theme, patient, goal, temporal, location, abstract-pos, amount} } Moreover, there exist rhetorical relations rel between sentences si and sj that comprise the news article. This is also used to identify which among the sentences contain relevant information. 37 REL = { rel(si, sj ) | rel ∈ { nucleus, elaboration(focus), justification, underspecified} Thus, for the given sample news article, the text is split into sentences. Let us call the first sentence S1 and the second sentence S2. Then these sentences are represented as: REL ={nucleus(S1, S2), elaboration[companies](S2, S1)} example, the news date is indicated, but not the transaction date. In this case, the date of the news is inherited as the transaction date. On other cases, information from other web source is used. An example would be for the case of location. The unprocessed location string (LocString) actually refers to the location of the companies involved. Identifying which of the companies is located in the first location and which is located in the second can be taken from other news sources or other sources like company profile (possibly from http://biz.yahoo.com). ROL(S1) = {agent( [IDDI (International Drug Development Institute) and Cytel Inc.], collaborate), temporal(today, collaborate), theme([strategic technology collaboration], collaborate) } B. Processing Questions Information from the question should be extracted for proper processing. We need to identify the type of question (question type), what we are looking for (question focus), and what the conditions are in our search (constraints). In our approach, we represent these into the following semantic representation: Q(<QUESTION TYPE>, <QUESTION FOCUS>,<BODY>) <Question Type> indicates the type of question (whether it is superlative or comparative) and its arguments. An example set of arguments for the superlative type of question would be the number of results (many or single) and search criteria. The <Question Focus> refers to what is expected as a result. The <Body> is the semantic dependency of the question defined by the main predicate and the thematic roles of its arguments. For the sample question “Which companies take the most risks?”, the semantic representation of the question will be the following: Q(SUPERLATIVE(MANY, HIGHEST), COMPANY, TAKE-RISK(AGENT: COMPANY)) This semantic representation is not enough to come up with the appropriate answer. We need other information to represent the basis for the evaluation. Thus, an operational representation of the question is constructed. An example format (in this case, for the superlative type of question) is: <SUPERLATIVE>(<VARIABLE>, <EVENT>, <RESPONSE>) Where <Superlative> could be highest or lowest depending on the search criteria in the semantic representation, the <Variable> is the basis of the search criteria, <Event> is the key concept determined from the semantic dependency in the question, and <Response> is the expected answer. In the above sample question, the operational representation will be: HIGHEST(RISK, TAKE-RISK(AGENT: COMPANY), COMPANY) Here, the <Event> is similar to the <Body> because takerisk is included in the identified key concepts that we can interpret. Other terms like “Which companies like to make risky investments?” are also mapped to the take-risk concept. To facilitate mapping of questions to the answers, we have a typed-feature representation for the question containing the following features: ROL(S2) = {agent([IDDI (International Drug Development Institute) and Cytel Inc.], develop), goal([integrated systems and services for patients], develop), abstract-pos([clinical trials], develop)} Notice that in S1 and S2 instead of the predicates “reported” (or “entering”) and “cooperating”, the main predicate used are collaborate and develop, respectively. This is because we are only concerned with predicates that are relevant to the transactions being reported in the news. Thus, the semantic dependency is simplified to model only those needed for the conceptual representation of the news. From the semantic representation of each sentence in the news, information is extracted to fill in the typed-feature structure (which is the conceptual representation of the news). It contains the following information: News Source Location Date Link Transaction TransCategory TransType Date Location Company (1..10) ContractedItem Such that Location is a complex type containing the LocString, City, State, Country, and Continent. The date consists of the month, day, and year. TransCategory and TransType are transaction categories and its transaction subtype. There can be at most ten companies. Each of the company information and the contracted item are complex types defined as follows: Company Name Location Role NewEntity SubsidiaryOf Category ContractedItem Item Sector Indication Stage Worth Not all the information that is stored into the typed-feature structure is available from one news article. Some processing has to be done. A set of inferencing rules is developed to retrieve and store the required information. For 38 Question Question Type Number of Results Search Criteria Question Focus Search Constraints Transact(c) = { (Transaction.Company1.Name : c, Transaction.Company2.Name: c2, Transaction.TransCategory:t, Transaction.Date.Year: y, Transaction.ContractedItem.Item:p) } NumTrans = { n | (c, n) ∈ CompanyTrans} To process a comparative question like “Does Company X take more risk than Company Y…?”, each of these entities will be tested based on the constraints. For a superlative question “Which companies take the most risk…?”, all companies will be tested and computations will be done to generate the top entities. Such that the <Question Type>, <Number of Results>, <Search Criteria>, and <Question Focus> are taken from the semantic representation, the <Search Constraints> is a complex type defined below. The <Duration> is the temporal scope of the search, while the <Location> and the <Transaction> are complex types, defined similar to that of the news. Search Contraints Duration D. Interpretation of Comparative and Evaluative Expressions Aside from active and take risk, other comparative and evaluative expressions have been studied from questions that can be raised in BI. The expression is studied from the predicate, identifying its basic properties, then looking at the nouns that it can modify, re-evaluating the properties if there are additional constraints or different constraints. From the study, the predicates are categorized into uni-dimensional, multi-dimensional, polysemous, and underspecified. 1) Uni-dimensional predicates: Some predicates have only one sense or definition. For example, expensive. It is essentially involving a high cost. In this case, the cost is the quantifiable property that we can use to evaluate or compare entities. However, for some uni-dimensional predicates, like innovative, it is difficult to quantify. Innovative is defined as “characterized by or introducing something new” [13]. In this case, we can look at the effect instead, i.e., something innovative is in demand. Thus, determining if a product is innovative would depend on the number of entities having an interest in it. And an innovative company is one with an innovative product. In this particular domain, the expert formally defines this as: Innovative(c) := ∃i, i ∈[1, m-1] ∀y:YEAR (c1, c, t, x, y, pi) ∈ SellTransact(c,y) ∧ pi = pi+1 ∧ |SellTransact(c,y)| ≥ 0.7 x n ∈ CompanyTrans-Per-Year DateStart DateEnd Location Transaction For criteria or properties that are already in the conceptual representation, these are used in the evaluation and/or comparison. For the sample question “Which Asian companies have the most number of transactions in year 2008?”. The company involved in the transaction should be located in Asia and the date (or year) should be 2008. Since these are search constraints indicated in the question, mapping the representation to the entries in the typed-feature representation of the news would provide a short-list of matching entries. Then with the most number of transactions, it is a matter of counting the occurrences of a certain company and comparing the values to determine the top companies. C. Complex Terms Other criteria that are non-directly quantifiable are referred to as complex terms. For these, the lexical knowledge is consulted to identify the term’s interpretation into quantifiable properties. For the take-risk example, the lexical knowledge represents a company that takes risks as one which is active, has transactions every year, have alliances every year with new and unstable partners. Take-risk(c) := Active(c) ∧ TransEveryYear(c) ∧ CompAllyEveryYearAndAlwaysNewPart(c) ∧ ¬HaveStablePartners(c) Again, the definition could consist of more key concepts or terms, which have to be evaluated first. Eventually, the key concept is broken down into values or quantifiable measure that can be extracted from the typed-feature structure of the news article. For example, the condition Active is also a key concept, defined to be a company that has above mean transactions in the duration of the search constraint. It is formally defined as: Active(c) := (c, n) ∈ CompanyTrans ∧ n ≥ ∑NumTrans / |NumTrans| SellTransact(c, y) = { (Transaction.Company1.Name : c1, Transaction.Company2.Name: c, Transaction.TransCategory:t, Transaction.TransType:x, Transaction.Date.Year : y, Transaction.ContractedItem.Item:p) | t = buy ∨ t = alliance ∧ (x = [exclusive licensing] ∨ x = [nonexclusive licensing]) } CompanyTrans-Per-Year = { (Company1.Name: c, Transaction.Date.Year: y, n) | n = | Transact-PerYear(c, y )| } CompanyTrans = { (Company1.Name: c, n) | n = | Transact(c) | } 2) Multi-dimensional predicate: Taking the example of 39 OnlyOncePartners(c) := ∀y1, y2 CompanyAlliance(c, c2, y1) ∧ CompanyAlliance(c, c2, y2 ) ⇒ y1 = y2 take-risk, it entails different dimensions from being conservative. Different aspects would then have to be considered. In the case of BI, this could be in terms of the amount of investments, types of products invested in, the partners being taken, or the overall strategy that is being employed. On the other hand, a stable partner is one which has alliances every year. And, a company has stable partners when it has alliances every year and always with new partners and the partners have alliances every year. HaveStablePartners(c) := CompAllyEveryYearAndAlwaysNewPart(c) ∧ AllianceEveryYear(c2) 3) Polysemous predicate: Many predicates have different senses and meanings. Taking the example stable, it is defined in [13] to have three meanings, namely: “firmly established”, the second is “steady in purpose”, and the third is “capability to resist motion”. Being able to identify which among these senses may depend on the noun that it is associated with or with the domain in question. CompAllyEveryYearAndAlwaysNewPart (c) := OnlyOncePartners(c) ∧ AllianceEveryYear(c,y) V. CONCLUSION AND DIRECTIONS Comparative and evaluative expressions in the domain of BI are complex because there are intricacies to the terminology used in BI where the criteria are predefined. As can be seen in the example predicates, expressions can be based on one criterion, can be based on multiple criteria, and/or can be underspecified. Currently, there are at least ten basic comparative and evaluative expressions in questions studied. Each of which have several variations considering aspects of polysemy, metonymy, underspecification, and multiple criteria. More comparative and evaluative expressions are yet to be explored in the context of BI. Expected form and style of answers from these questions will be taken into consideration. The research will also explore techniques to automatically determine the properties which are at stake in the evaluation and to automatically determine limits, ranges, and relative values of these properties from on-line sources, so that the technique can be portable to other domains. Evaluation will be carried out eventually. However, it is crucial first to identify evaluation metrics and processes and to which components the metrics and processes will be applied to, as the evaluation is not so straightforward. 4) Underspecified predicate and metonymy: Underspecification refers to a general criteria associated with the predicate, but will gain (more) context only when associated with the noun it modifies. Assuming that we consider only sense of stable as being “steady in purpose”, it is still underspecified because the properties associated to this meaning still depends on the context. Even within the domain of BI, the criteria for evaluating a stable company are different from a stable partner, even if the partner is also a company. This also leads to the issue of metonymy. The nouns associated to the predicate represent a class of objects that hold various properties. For example, a company can be quantified by the number of employees, the number of transactions, the types of transactions, the investments that it makes, and so on. By associating the predicate “stable” with “company”, determining which of these properties is to be used in the evaluation of “steady in purpose” is a challenge. In this case, the constraints are provided by an expert. A stable company is defined as one which is active, may not have alliances every year or have alliances every year but always with old partners. Stable(c) := Active(c) ∧ ¬AllianceEveryYear(c) ∨ AllianceEveryYear(c) ∧ ¬OnlyOncePartners(c) REFERENCES [1] [2] AllianceEveryYear(c) := ∀y: YEAR (c, y, n) ∈ CompanyAlliance [3] [4] CompanyAlliance = { (Company1.Name: c, Transaction.Date.Year: y , n) | n = | Alliance-PerYear(c, y ) | } [5] Alliance-Per-Year(c, y) = { (Transaction.Company1.Name : c, Transaction.Company1.Name : c2, Transaction.TransCategory:t, Transaction.Date.Year:y , Transaction.ContractedItem.Item: p ) | t = alliance } [6] [7] [8] 40 J. Burger, et al., “Issues, Tasks and Program Structures to Roadmap Research in Question & Answering,” 2009. [Online]. Available: www-nlpir.nist.gov/projects/duc/papers/qa.Roadmap-paper_v2.doc B. Gay and B. Dousset, “Innovation and Network Structural Dynamics: Study of the Alliance Network of a Major Sector of the Biotechnology Industry,” Research Policy,34(10), 1457-1475, Management Journal, Special Issue, 213, 203-215, 2005. Tetralogie. [Online]. Available: http://atlas.irit.fr/Tetralogie M. Maybury, New Directions in Question Answering, The MIT Press, Menlo Park. 2004. D. Moldovan, et al., “The Structure and Performance of an Open Domain Question Answering System,” in Proceedings of the 38th Meeting of the Association for Computational Linguistics (ACL), HongKong, 2000. C. Kennedy, “Comparatives, Semantics Of”, in K. Allen (section editor) Lexical and Logical Semantics; Encyclopedia of Language and Linguistics, 2nd Edition, Elsevier, Oxford, 2006. J. Ruppenhofer, et al. (2006). FrameNet II: Extended Theory and Practice. Available: http://framenet.icsi.berkeley.edu/ C. Friedman, “A General Computational Treatment of the Comparative,” in Proceedings of the 27th Annual Meeting of the ACL, [9] [10] [11] [12] [13] 1989. [Online]. Available: http://acl.ldc.upenn.edu/P/P89/P891020.pdf N. Jindal and B. Liu, “Mining Comparative Sentences and Relations,” in Proceedings of the 21st AAAI Conference on Artificial Intelligence, AAAI Press, USA, 2006. D. Olawsky, “The Lexical Semantics of Comparative Expressions in a Multi-level Semantic Processor,” in Proceedings of the 27th AnnualMeeting on ACL, USA, 1989. B. Ballard, “A General Computational Treatment of Comparatives for Natural Language Question Answering,” in Proceedings of the 26th Annual Meeting of the ACL, 1988. [Online]. Available: http://www.aclweb.org/anthology-new/P/P88/P88-1006.pdf MUSING Newsletter nr. 2 – Spring 2009. [Online]. Available : http://www.musing.eu/newsroom/newsletters/musing-newsletter-nr-2spring-2009 Merriam Webster Dictionary. [Online]. Available: http://www.merriam-webster.com/ 41
© Copyright 2025 Paperzz