Computers & Operations Research 49 (2014) 28–40 Contents lists available at ScienceDirect Computers & Operations Research journal homepage: www.elsevier.com/locate/caor Learning criteria weights of an optimistic ELECTRE TRI sorting rule Jun Zheng a, Stéphane Aimé Metchebon Takougang b, Vincent Mousseau a,n, Marc Pirlot c a b c Laboratoire Génie Industriel, Ecole Centrale Paris, Grande Voie des Vignes 92 295 Châtenay-Malabry, France Institut Supérieur d'Informatique et de Gestion, ISIG-International, 06 BP 9283 Ouagadougou 06, Burkina Faso UMONS, Faculté Polytechnique de Mons, 9 Rue de Houdain, 7000 Mons, Belgium art ic l e i nf o a b s t r a c t Available online 28 March 2014 Multiple criteria sorting methods assign alternatives to predefined ordered categories taking multiple criteria into consideration. The ELECTRE TRI method compares alternatives to several profiles separating the categories. Based on such comparisons, each alternative is assigned to the lowest (resp. highest) category for which it is at least as good as the lower profile (resp. is strictly preferred by the higher profile) of the category, and the corresponding assignment rule is called pessimistic (resp. optimistic). We propose algorithms for eliciting the criteria weights and majority threshold in a version of the optimistic ELECTRE TRI rule, which raises additional difficulties w.r.t. the pessimistic rule. We also describe an algorithm that computes robust alternatives' assignments from assignment examples. These algorithms proceed by solving mixed integer programs. Several numerical experiments are conducted to test the proposed algorithms on the following issues: learning ability of the algorithm to reproduce the DM's preference, robustness analysis and ability to identify conflicting preference information in case of inconsistencies in the learning set. Experiments show that eliciting the criteria weights in an accurate way requires quite a number of assignment examples. Furthermore, considering more criteria increases the information requirement. The present empirical study allows us to draw some lessons in view of practical applications of ELECTRE TRI using the optimistic rule. & 2014 Elsevier Ltd. All rights reserved. Keywords: Multiple criteria sorting Preference learning ELECTRE TRI Optimistic rule 1. Introduction Multiple criteria sorting methods have gained wide interest in the last few decades. Such methods aim at assigning alternatives to pre-defined ordered categories taking into account several criteria. Multiple criteria sorting methods differ from standard classification in two main features: (1) categories are predefined and ordered, (2) the sorting model integrates preferences of a Decision Maker (DM). Various methods have been proposed in the literature (see [35] for a review). Some methods are based on the use of value functions (e.g. [13,34]); some others involve the use of outranking relations (e.g. [21,17,33,29,27,1]); others make use of decision rules (e.g. [12]). The multiple criteria sorting models require to set many parameter values, which reflect the DM's value system. A preference elicitation process is therefore needed. In a direct aggregation paradigm, the parameter values are supposed to be provided by the DM through interactive communication with the analyst. n Corresponding author. E-mail addresses: [email protected] (J. Zheng), [email protected] (S.A. Metchebon Takougang), [email protected] (V. Mousseau), [email protected] (M. Pirlot). http://dx.doi.org/10.1016/j.cor.2014.03.012 0305-0548/& 2014 Elsevier Ltd. All rights reserved. The multiple criteria assignment model encompassing the DM's comprehensive preference is first constructed and then applied to the set of alternatives. Such a paradigm imposes a strong cognitive effort to the DM who needs to understand the meaning of and set appropriate values to the preference related parameters. In a disaggregation paradigm (see [15] for a review, and [4, Chapter 7]), the sorting model is inferred from a priori preferences expressed by the DM. The initial proposal of a disaggregation method called UTA was made by Jacquet-Lagrèze and Siskos [14] in the context of ranking problems. For additive value based sorting models, Doumpos and Zopounidis [11] present a disaggregation method to elicit a UTADIS model as a variant of the UTA model for sorting purposes. In the context of rule-based models, the disaggregation method proposed in [12] infers a set of “if…then…” decision rules derived from rough sets based approximations of decision examples. In this paper, we are interested in the disaggregation methodology for the ELECTRE TRI method [33,29]. ELECTRE TRI is a well known sorting method based on an outranking relation. Many applications have been reported (e.g. [24,30,18]). The ELECTRE TRI model involves several parameters including profiles that define the limits between the categories, criteria weights, discrimination thresholds, etc. ELECTRE TRI requires to select one of the two assignment rules proposed (optimistic or pessimistic). J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 Several authors have presented disaggregation methodologies to infer a pessimistic ELECTRE TRI rule from assignment examples provided by the DM [25,7,8,26]. These methodologies propose to infer the preference parameters that best match the DM's preference information and to compute a robust assignment, i.e. the range of categories to which an alternative can be assigned, considering all combinations of the parameters values compatible with the DM's preference statements. These disaggregation procedures lean on linear programming. Recently, an evolutionary approach has been presented considering both the optimistic and the pessimistic rules to infer the parameters of an ELECTRE TRI model [10]. However, all elicitation procedures (except for [10]) deal with the pessimistic rule only. The literature provides no mathematical programming formulation for the elicitation of an optimistic ELECTRE TRI rule, even though in practice, some applications require using the optimistic rule [20,19]. Recent experimental analysis of elicitation procedures, even for the pessimistic ELECTRE TRI model, has shown (e.g. [16,23]) that the number of assignment examples needed for specifying the parameters of the model with reasonable accuracy can be rather large and grows rapidly with the numbers of criteria and categories. Actually, learning methods for multicriteria sorting models, such as ELECTRE TRI, can be used in two very different contexts: decision aiding and machine learning. In the practice of the former, the number of assignment examples usually is very limited, a few dozens at best. In such a context, using a rule involving many parameters for learning purposes is inadequate since it leaves us with a highly undetermined situation, a large variety of models being compatible with the assignment examples. In machine learning, in contrast, one deals with large or huge sets of assignment examples (at least several hundreds). In such a case, the algorithms designed for eliciting all or part of the parameters of a classical ELECTRE TRI model are far too slow and complex. In both cases, while for very different reasons, we argue that studying the elicitation of a version of the ELECTRE TRI rule using fewer and less inter-related parameters is a fruitful option. The fact that such a rule is better understood thanks to the axiomatic work in [2,3] is also an asset. An exploration of the learning and expressive capabilities of such a procedure started with [16] using mixed integer linear programming formulations. In [31], the learning of all parameters of an ELECTRE TRI – like learning rule (called MRSort, for Majority Rule Sorting) is done by means of a specific metaheuristic. The method allows to deal with learning sets containing several hundreds of assignments. Both [16,31] deal with the pessimistic version of the assignment rule. In this paper, we explore the elicitation of a simple ELECTRE TRI rule in its optimistic version, without veto. Working with the optimistic rule raises specific difficulties. Even the mathematical programming formulation for the determination of the criteria weights, the limit profiles being given, requires the use of binary variables. Therefore this paper only deals with this issue. Note that a method for learning the criteria weights, assuming known profiles, can already help a lot in practice since the DM can often be directly questioned about profiles. The latter have a relatively clear interpretation in such models. When proposing a learning procedure it is crucial to perform numerical experiments that guarantee its practical usefulness. A basic requirement (seldom verified) is that it is possible to learn the “true model” whenever the data in the learning set have been assigned using this model. A related issue is the size of the learning set needed to retrieve approximately the true model. Alternatively, for a given size of the learning set, one can assess the degree to which the model remains undetermined by recording the average number of categories an alternative can be assigned to by the learned models. Another important question is whether it is possible to learn the model when noise has been added to the assignments in the learning set. Indeed as DMs cannot be expected to always provide reliable and consistent information, the learning 29 procedure should be as robust as possible with respect to the presence of erroneous assignments in the learning set. Finally comes the question of the expressiveness of the model, i.e. its capability to reproduce assignments made according to an unknown rule (artificial or real data sets). In this study we will not try to assess the expressiveness of the model but we will address the other issues in depth. The reason for this is that we concentrate here on the elicitation of the criteria weights in an optimistic ELECTRE TRI rule. The profiles are considered as given. This makes it difficult to study the expressiveness of the model since we did not develop tools for identifying appropriate profiles on the basis of a learning set generated by an arbitrary rule. Even in the case the assignments have been produced by means of a general ELECTRE TRI model (involving for instance indifference and preference thresholds) it is not clear whether using the same profiles in our simpler optimistic rule would be an optimal strategy. The study of the expressiveness of the simple ELECTRE TRI sorting rule (either in its optimistic or pessimistic version) is complex enough to deserve an in depth study in a separate paper. The introduction of vetoes and their impact both on the determination of the other parameters as well as the gain in expressiveness they will provide is also worth studying. This paper has two main objectives: Firstly, it establishes mathematical programming formulations for the elicitation of the criteria weights in an optimistic ELECTRE TRI rule. Our formulations allow us to deal with “assignment errors” in the learning set. Secondly, these formulations are tested in order to assess their practical usefulness in an interactive elicitation process, in particular their sensitivity to assignment errors. The paper is organized as follows. Section 2 briefly introduces the ELECTRE TRI method. Section 3 presents the mathematical programming formulations for eliciting the weights of the criteria in the optimistic ELECTRE TRI rule, computing robust assignments and dealing with infeasible learning sets. In Section 4, extensive numerical experiments are designed to test the computational behavior and study the learning capabilities of the proposed algorithms. The last section groups conclusions and issues for further research. 2. The ELECTRE TRI method ELECTRE TRI assigns the alternatives in a set A to one out of p predefined ordered categories C 1 ; C 2 ; …; C p . We assume w.l.o.g. that categories are ordered from the worst to the best, i.e. that C1 represents the lowest quality level and Cp the highest, while C 2 ; …; C p 1 are intermediate levels ranked in increasing order of their quality. We denote by C Z h , the set of categories Ch to Cp. One defines similarly the sets of categories C 4 h ; C r h and C o h . Each alternative is assessed w.r.t. m criteria. Let ðg 1 ðaÞ; …; g m ðaÞÞ denote the vector of evaluations of alternative a on the m criteria. P denotes the set f1; 2; …; pg and M, the set f1; 2; …; mg. In the ELECTRE TRI method, the assignment of an alternative a to a category results from the comparison of a with the profiles defining the “category limits”. To each category Ch are assigned h1 (the latter an upper limit profile bh and a lower limit profile b also being the upper limit of category C h 1 and the former, the lower limit of category C h þ 1 , provided such categories exist). A profile is just a vector of reference assessments on all criteria, h h h h i.e. b ¼ ðb1 ; …; bj ; …; bm Þ. The lower limit profile b0 of category C1 and the upper limit bp of category Cp are trivial profiles. Profile 0 0 0 0 b ¼ ðb1 ; …; bj ; …; bm Þ has the worst possible performance on all criteria, so that any alternative is at least as good as b0. Symmep p p p trically, b ¼ ðb1 ; …; bj ; …; bm Þ has the best possible performance on all criteria so that no alternative is better than bp. 30 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 2.1. Non-compensatory sorting models The assignment of an alternative to a category is based on rules in which the vector of evaluations associated with the alternative is compared to the upper and/or lower limit profile of the category. In the spirit of the non-compensatory sorting methods characterized by Bouyssou and Marchant [2,3], an alternative a is considered at least as good as profile bh if the coalition of criteria on which a receives evaluations at least as good as the profile is “strong enough”. A non-veto condition may be added but we do not consider such conditions in this study. Formally, one defines the set of strong enough coalitions of criteria as a subset F of the power set 2M. The only property that we need to impose on F is that it is stable, from above, for inclusion (a coalition containing a strong enough coalition is strong enough). Typically, the set of strong enough coalitions is determined by a weighted majority rule as we shall see below. For comparing a with bh we define the coalitions h h Sða; b Þ ¼ fj A M : g j ðaÞ Z bj g h ð2Þ We define the outranking relation ≿ between an alternative a and a profile bh, and conversely, as follows: a≿b h if Sða; b Þ A F h ð3Þ b ≿a h if Sðb ; aÞ A F : ð4Þ h The asymmetric part g of the relation ≿ is interpreted as a strict preference (better than relation). One has a ≻ b if a≿b h and h a ≻ b if Sða; b Þ A F h Not½b ≿a ð5Þ h ð6Þ Sðb ; aÞ2 =F; and h b ≻ a is defined similarly. The indifference relation is the symmetric part of ≿. Since the relation ≿ usually is not complete, h the incomparability relation a ? b is defined as the symmetric complement of ≿, i.e., h h a ? b if Not½a≿b h h a ? b if Sða; b Þ2 =F h Not½b ≿a and ð7Þ h Sðb ; aÞ2 =F: and and a≿b h1 ð13Þ It has been shown [29], and this is easy to check using the above definitions, that the category assigned to any alternative by the optimistic rule is guaranteed to be at least as high as that assigned by the pessimistic rule. The simple version of the pessimistic rule of ELECTRE TRI described above is in line with the work of Bouyssou and Marchant [2,3]. Our definition of the optimistic rule is in the same spirit, yet it has not been characterized in [2,3]. Note also that the definition of the outranking relation in (3) and (4) does not involve any non-discordance condition (see [28,29]). In this work, we focus on the analysis of the elicitation of the optimistic ELECTRE TRI rule without veto. 2.2. Introducing weights h h h a assigned to C h if Not½a≿b ð1Þ Sðb ; aÞ ¼ fj A M : bj Zg j ðaÞg h Pessimistic assignment rule: With this rule, alternative a is assigned to a category strictly below C h þ 1 if a is not at least as good as the lower h limit profile bh of C h þ 1 . Formally, we have a A C o ðh þ 1Þ if Not½a≿b . It is thus assigned to Ch if C h þ 1 is the lowest category such that h Not½a≿b , i.e. ð8Þ As already mentioned, it is often assumed in practice that the set of strong enough coalitions F can be represented by a set of nonnegative weights wj associated with the criteria gj with jA M and a majority rule. More precisely, we shall assume that there exist weights wj Z 0 with ∑m j ¼ 1 wj ¼ 1 and a majority level λ, with 0:5 r λ r 1, such that F AF iff ∑ wj Z λ: ð14Þ jAF In words, a coalition F is strong enough if the sum of the weights of the criteria belonging to F reaches at least some majority threshold λ. In order to avoid degenerate cases in which an evaluation inferior to that of a profile on a single criterion can prevent an alternative to be declared at least as good as a profile (or conversely), we impose that no criterion weight may exceed 12 . Such a condition has already been used previously in the literature so as to avoid dictatorship among a group of DMs (see [9]), and to forbid a single criterion to be decisive alone (see [25,8]). h h The weight cða; b Þ of a coalition Sða; b Þ is defined as the sum of h the weights of the criteria that belong to it: cða; b Þ ¼ ∑j A Sða;bh Þ wj . h A similar definition holds for the weight cðb ; aÞ of the coalition h Optimistic assignment rule: Using the optimistic rule, alternative a is assigned to category Ch or higher if the lower limit profile of category Ch is not strictly preferred to a. Formally, a assigned to C Z h if Not½b a assigned to C Z h iff Not½b h1 h1 h1 ≻ a ð9Þ h1 ≿a or a≿b : ð10Þ Hence the precise category which a is assigned to is Ch if the latter is the highest one satisfying the previous condition, i.e. h1 a assigned to C h if Not½b ≻ a h1 a assigned to C h iff ½Not½b h and b ≿a and Sðb ; aÞ. Using these weights, the optimistic and pessimistic assignment rules can be rephrased as follows. Optimistic assignment rule: h b ≻a and h1 ≿a or a≿b ð11Þ h Not½a≿b : ð12Þ Paraphrasing this somewhat awkward, yet understandable, condition, we say that a is assigned to the category of which the lower limit profile is not strictly better than a and the upper limit profile is strictly better than a. a assigned to C h if ½cðb h and cðb ; aÞ Z λ and h1 ; aÞ o λ or cða; b Þ Z λ h cða; b Þ oλ: ð15Þ Pessimistic assignment rule: h a assigned to C h if cða; b Þ o λ and h1 cða; b Þ Z λ: ð16Þ 2.3. Learning an ELECTRE TRI sorting rule In the above definitions, instead of considering the general ELECTRE TRI optimistic and pessimistic rules, which use the ELECTRE III definition of the concordance–discordance index (see [29]), we choose to work with the ELECTRE I version of the concordance index, without using veto, which amounts to compare alternatives and profiles using a majority rule. Moreover, this paper only deals with the elicitation of the criteria weights and majority threshold in the optimistic ELECTRE TRI sorting rule. Why do we focus on such an apparently limited objective? J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 First of all, let us observe that a method for eliciting the criteria weights, while the profiles are assumed to be known, is often useful in practice, be it for the pessimistic or the optimistic rule. For instance, in the application dealt with in Metchebon's doctoral dissertation [19], the profiles are spontaneously provided by the expert, so that the inference of the weights arises as the natural next step. More generally, it is advisable to approach the elicitation of the (numerous) parameters of an ELECTRE TRI rule in a stepwise manner. This allows us to control the interaction of the different categories of parameters (profiles, weights, veto thresholds). Learning all parameters at a time has two major drawbacks. It is often an intractable problem (due to the rapidly growing number of binary variables involved) and/or its solution is largely undetermined. Both problems were specifically studied for the pessimistic version of the simplified version of the ELECTRE TRI method (without veto), called MR-Sort (see [16]). A conclusion that forcefully emerges from this experimentation is that the simplified pessimistic rule based on MR-Sort is already quite expressive in the sense it allows not only to represent the learning set assignments when they are noisy but also when the assignments have been generated by another model (i.e. a rule based on an additive value function and some thresholds). Due to the additional complexity of the mathematical program required by the optimistic assignment rule, it was unrealistic, both from a practical and a methodological viewpoint, to try to learn the profiles and the weights at the same time. Hence the option we have taken in the present study. Having chosen to only learn the weights, the profiles being given, our experimentation shows e.g. that the number of assignment examples to be used to learn a model with 9 criteria and 4 categories with an accuracy of 90% requires more than 30 examples when the learning set is noiseless (no assignment “errors” w.r.t. to an unknown ELECTRE TRI model). Not to mention the computational effort, eliciting the 3 profiles at the same time would require the determination of about four times as many variables. This means that the size of the learning set needed to obtain a reasonable accuracy would clearly be at least 100 examples. Yet the variety of models providing approximately the same accuracy would be huge. This type of difficulty is amplified for noisy learning sets (with assignment errors), as shown in our experimentations. 3.1. Mixed integer linear programming formulation We use a regression-like technique to learn the parameters of such a model. The preference information (in the form of assignment examples) can be represented by linear constraints involving continuous and binary variables. A set of values for the parameters of the model can be obtained by solving a mixed integer program (MIP) whose feasible solutions satisfy these constraints. Such an approach has been previously followed for eliciting the parameters of an ELECTRE TRI model with the pessimistic rule [23,8]. However, when considering the optimistic rule, the conditions (15) for assigning alternative a to category Ch appear as a conjunction of conditions, one of them being a disjunction: h1 cðb ; aÞ o λ h1 or cða; b ð17Þ In the above, inequality δ1 þ δ2 r1 ensures that at least one of the two variables is set to 0, which entails that at least one of the two clauses in (17) is true. We aim at eliciting the parameter values of an ELECTRE TRI model which can restore the DM's assignment examples by using the optimistic rule. Provided such a model exists, the linear program below yields the values of these parameters. In the sequel, for each alternative ae ; eA L in the learning set An, we denote by C he the category to which ae is known to be assigned. α max ð19Þ h cðb e ; ae Þ βe ¼ λ; s:t: h he 1 We consider the optimistic assignment rule (15), assuming that the limits of the categories are given. Our aim is to elicit the h h criteria weights wj involved in cða; b Þ and cðb ; aÞ and the majority level λ on the basis of comprehensive preference statements made by the DM. The preference information consists of a number of assignment examples, which involve alternatives that the DM is able to assign to a category in a holistic way, i.e. without having to analyze their performances. We consider here a static set of assignment examples, i.e. we assume that all the information is available at the beginning of the learning process; in other words, the DM cannot be questioned to provide adequate or critical information in the course of the learning phase. The set of learning examples is denoted by An ¼ fa1 ; …; ae ; …; al g, where l is the number of alternatives in the learning set. We note L ¼ f1; …; e; …; lg. The assignment to a category of each alternative in An is supposed to be known. These examples will be used as a training set to establish an ELECTRE TRI model which can reproduce these assignment examples. The obtained model can then be applied to assign other alternatives to categories. Þ Z λ: Modeling this condition by a system of linear constraints can be done in a standard way by introducing two binary variables δ1, δ2 (see, e.g. [32]). Their values indicate, for each of the clauses in the disjunctive condition, whether it is satisfied or not. More precisely, h1 h1 δ1 (resp. δ2) equals 0 if cðb ; aÞ o λ (resp. cða; b Þ Z λ) and equals 1 otherwise. Condition (17) holds iff the following constraints are satisfied for some value of the positive variable ϵ: 8 h1 > cðb ; aÞ λ δ1 þ ϵ r0 > > > < h1 Þ λ þδ2 Z 0 cða; b ð18Þ > ; δ A f0; 1g δ > 1 2 > > : δ1 þ δ2 r1: cðae ; b e Þ þ γ e þϵ ¼ λ; 3. Eliciting the weights and majority threshold of the optimistic assignment rule 31 cðb ð20Þ 8eAL ð21Þ ; ae Þ λ δe1 þ ηe þ ϵ ¼ 0; he 1 cðae ; b 8eAL Þ λ þ δe2 μe ¼ 0; δe1 þ δe2 r 1; 8eAL δe1 ; δe2 A f0; 1g; 8 eA L 8 eA L 8eAL ð22Þ ð23Þ ð24Þ ð25Þ α r βe ; 8 eA L ð26Þ α r γe; 8eAL ð27Þ α r ηe ; 8eAL ð28Þ α r μe ; 8 eA L ð29Þ ∑ wj ¼ 1; jAM 0:5 r λ r 1 wj Z 0; 8 j A M ð30Þ ð31Þ In this program, the constraints (20)–(25) express condition (15) for each ae in An. In these constraints βe, γe, ηe and μe are the slack variables, and ϵ is a small positive value used to ensure that the appropriate inequalities are strict. The objective function consists 32 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 in maximizing the minimal value of the slack variables: constraints (26)–(29) express that α is not larger than the least slack variable. The usual constraints on the weights and the cutting level λ are expressed in (30) and (31). Since we have assumed that the h category limits are known, the constraints in which cðae ; b e Þ or he cðb ; ae Þ intervene are linear. If the previous program is feasible and its optimal value αn is non-negative, then there exists a combination of parameter values that satisfy all the constraints in the program and, consequently, there is an ELECTRE TRI model which assigns the alternatives in An as prescribed. Otherwise, either the DM should reconsider his/her statements or another type of assignment model should be used. In case αn is negative, an inconsistency resolution algorithm (see [22]) can be run to help the analyst identify the inconsistencies in the DM's statements w.r.t. the model considered. 3.2. Robustness analysis As the preference information in the form of assignment examples is represented by linear constraints, the set of feasible values for the weights satisfying the constraints is a (possibly empty) polytope. Whenever the latter is not empty, there are generally an infinite number of sets of weights compatible with the preference information. The objective function (19) in the MIP (19)–(31) helps us to select a particular set of weights in the polytope of feasible weight vectors, but this choice is rather arbitrary. By varying the weights in the feasible weight vectors set, an alternative a can often be assigned to more than one category. Finding the set of all categories to which an alternative can be assigned, given the information available, is the purpose of robustness analysis [8]. In this context, we are interested in the following question: “Does there exist a combination of parameters which would lead alternative a to be assigned to category Ch?”. Answering this question for all h is tantamount to solving the problem of computing the robust assignment of a. This is done by solving the following MIP (32)–(46). Alternative a is an alternative of interest selected from A, the set of all considered alternatives. We impose as a constraint that a is assigned to category Ch and we search for feasible weight vectors (i.e. for which this constraint is fulfilled as well as all constraints relative to the assignment of the alternatives in the learning set An). The mathematical program to be solved to compute whether a can possibly be assigned to category Ch runs as follows: ε max ð32Þ h s:t: cðb ; aÞ Z λ ð33Þ h ð34Þ cða; b Þ r λ ε h1 cðb cða; b ; aÞ r λ þ δ1 ε h1 ð37Þ δ1 ; δ2 A f0; 1g ð38Þ h 8eAL cðae ; b e Þ r λ ε; he 1 cðb ∑ wj ¼ 1; jAM 8eAL ð44Þ wj Z 0; 8 j A M ð45Þ 0:5 r λ r 1 ð46Þ The objective function (32), to be maximized, is the variable ε that is used for transforming strict inequalities into the non-strict ones (it was a constant in the MIP (19)–(31), here it is a variable). Constraints (33)–(38) express that a is assigned to Ch; constraints (39)–(44) enforce that the model correctly assigns all the alternatives in the learning set An; (45)–(46) impose the usual restrictions on the weights and the threshold λ. If the assignment of a to Ch is compatible with the preference information provided by the DM, the above program is feasible and its optimal value εn is strictly positive. Computing the robust assignment of an alternative a consists in repeatedly using the above MIP to check whether a can be assigned to category Ch, for each h ¼ 1; 2; …; p. Let Ra denote the set of categories Ch which a can be assigned to, i.e. the set of categories Ch for which the above program has a positive optimal solution εn . jRa j denotes the number of different categories in Ra . Computing Ra can be performed by using Algorithm 1. Remark on possible improvements in the search of a robust assignment: As observed in [8] for the pessimistic ELECTRE TRI rule, there may exist “gaps” in the set Ra of possible assignment classes for alternative a. In other words, it may happen that Ra contains classes Ch and Cl with h ol but not the class Cj with ho j o l. The same phenomenon appears with the optimistic rule. Consider the following example of an alternative a assessed on four criteria; the relative positions of these assessments (gi(a)) w.r.t. two profiles 1 2 b ; b determining three categories are shown in Fig. 1. We have 1 for i ¼ 1; 2 2 for i ¼ 3; 4: g i ðaÞ obi g i ðaÞ 4bi It is easy to see that such an alternative is never assigned to category C2 (neither by the optimistic nor by the pessimistic ELECTRE TRI rule) whatever the weights and threshold can be. Let us show it for the optimistic rule. Using (15), we see that a is 2 2 assigned to class C2 if and only if cða; b Þ oλ and cðb ; aÞ Zλ and 1 1 ½cða; b Þ Zλ or cðb ; aÞ oλ. With an alternative as illustrated in 2 1 2 1 Fig. 1, we have cða; b Þ ¼ cða; b Þ and cðb ; aÞ ¼ cðb ; aÞ, which makes ð39Þ 8eAL ; ae Þ rλ þδe1 ε; he 1 cðae ; b δe1 ; δe2 A f0; 1g; ð43Þ ð36Þ δ1 þδ2 r 1 h 8eAL ð35Þ Þ Z λ δ2 cðb e ; ae Þ Z λ; δe1 þ δe2 r 1; Þ Zλ δe2 ; ð40Þ 8 eA L 8eAL ð41Þ ð42Þ Fig. 1. Example of an alternative that can possibly be assigned to categories C1 or C3 but not C2. J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 it impossible to satisfy the conditions for assigning the alternative to C2. In contrast, it is clear that for some values of the weights and the threshold, such an alternative can be assigned to C3 (in case the last two criteria form a sufficient coalition) or C1 (in case the first two criteria form a sufficient coalition). The previous argument holds, in a general case, for any alternative having at least h1 one assessment above bh and one below b , but none between h1 h b and b . The algorithm we propose to compute Ra checks whether alternative a can be assigned to class Ch for all values of h. Since the presence of gaps in Ra cannot be excluded, it is not advisable to blindly apply a dichotomic search procedure to speed up the computation of Ra . Observe however that, in practice, the number p of categories is usually quite small (p is seldom greater than 5) so that solving the 0–1 program in Algorithm 1 p times instead of log 2 ðpÞ times is not an issue. Note also that the intuition that Ra should not involve gap is clearly linked with the prevalence of compensatory models (such as weighted sums and additive value functions) in Decision Analysis. If compensation is allowed, it is natural to consider that an alternative such as illustrated in Fig. 1 could be assigned to C2. On the contrary, in a non-compensatory approach (such as ELECTRE TRI), either criteria 3 and 4 form a “majority”, and a is assigned to C3, or they do not, and then the assessments of a on criteria 1 and 2 will prevail; therefore a will be assigned to C1. This is a striking illustration of the difference between compensatory and noncompensatory preferences; there are clearly examples of both in real-life situations. Algorithm 1. Procedure to compute the robust assignment of an alternative. Input: The set of assignment examples: ae -C he , for all e A L; The performance vector of the alternatives ae in the learning set An: ðg j ðae Þ; j A MÞ, for all eA L; The performance vector of alternative a: ðg j ðaÞ; jA MÞ; h The performance vector of each limit profile bh: ðbj Þ; j A M, for h ¼ 1; 2; …; p 1 Output: Robust assignment for alternative a: Ra ; 1: Ra ’∅ 2: for h¼ 1 to p do 3: write the constraints corresponding to a-C h in MIP solve the program 4: if εn 40 then 5: Ra ’Ra [ fhg 6: end if 7: end for 3.3. Infeasible learning sets In the course of a decision aiding process, inconsistency situations can occur when there is no ELECTRE TRI model which matches all the DM's preference information. The algorithms described in Mousseau et al. [22] have been designed to help the analyst to identify the conflicting pieces of information in the statements made by the DM. The algorithm identifies and suggests a subset of assignment examples which can be removed to make the remaining assignments consistent. However, the algorithm in Mousseau et al. [22] is dedicated to the resolution of inconsistencies arising when using the pessimistic ELECTRE TRI rule. We adapt this algorithm to the case of the optimistic rule. Additional binary variables are introduced to model the constraints stemming from assignment examples. 33 More precisely, for each ae A L, we introduce a binary variable γe which is equal to one if alternative ae is correctly assigned by the sorting model, and equal to zero otherwise. Given a set of inconsistent assignment examples, the following MIP computes the maximum number of them, which can be restored by an optimistic ELECTRE TRI model. The first four constraints impose the required conditions on the γe variables ∑ γe max ð47Þ eAL h cðb e ; ae Þ Z λ Mð1 γ e Þ; s:t: h cðae ; b e Þ rλ ε þ Mð1 γ e Þ; he 1 cðb 8eAL 8eAL ; ae Þ r λ þ δe1 ε þ Mð1 γ e Þ; h cðae ; b e Þ Zλ δe2 þ Mð1 γ e Þ; δe1 þ δe2 r 1; ∑ wj ¼ 1; jAM ð49Þ 8eAL 8eAL 8eAL δe1 ; δe2 ; γ e A f0; 1g; wj Z 0; ð48Þ ð50Þ ð51Þ ð52Þ 8eAL 8jAM 0:5 r λ r 1 ð53Þ ð54Þ ð55Þ The objective function of this MIP guarantees that we obtain a subset of alternatives of maximal cardinality in the learning set that can be represented by the sorting model. This model will be used in the third experiment (Section 4.2.3). 4. Experimental design and results In this section, we analyze the behavior of the proposed elicitation algorithm by means of numerical examples. Our aim is to check the empirical validity of the approach, to validate the practical usefulness and to provide guidance to the analyst who would like to use this procedure. In order to infer in a reliable way a weight vector, the optimization procedure requires information as input, i.e. on the set of assignment examples. What is the amount of information necessary to “calibrate” the model in a satisfactory way? How large should be An in order to derive the weights in a reliable manner? This question is essential for practical use to the inference model in real world decision problems. The analyst should have some simple guidelines to manage the interaction with the DM avoiding unnecessary questions, but collecting a sufficient information. In practical decision situations, real DMs do not always provide reliable information. Due to time constraints and cognitive limitations, DMs express contradictory information, their preferences change over time. The optimization procedure should be able to highlight the assignment examples that are contradictory or not representable through the ELECTRE TRI preference model. This experiment aims at investigating the ability of the tool to “identify” the inconsistencies in the DM's statements in order to help him/her in revising the preference information. How reliable is the optimization procedure to identify inconsistencies in the DM's judgments? 4.1. Experimental scheme The experiments are designed to address three issues: (1) the learning ability of the elicitation algorithm; (2) the behavior and performance of the robustness analysis algorithm; (3) the ability to deal with inconsistent information. Moreover, some factors 34 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 possibly influencing the performance of the algorithms are investigated, including the number of assignment examples, the number of criteria, and the number of categories. Similar experiments concerning the learning ability and inconsistency identification for ELECTRE TRI with the pessimistic rule can be found in [16]. We consider the following experimental scheme to answer the above questions. We implement a fictitious DM whose preferences are represented by an optimistic ELECTRE TRI model (called the original model). The model is characterized by the profiles bh ðh ¼ 1; 2; …; p 1Þ, a set of weights wj ðj ¼ 1; 2; …; mÞ and the majority level λ. The randomly generated original model assigns randomly generated alternatives to categories, and then these alternatives together with their assignments are used to infer an ELECTRE TRI model (called the inferred model). We are interested in how “close” the original model and the inferred model are. More precisely, 1. Firstly, a set An of l alternatives is generated; the evaluations of these alternatives are randomly drawn from the interval [0, 99] according to a uniform distribution. 2. Then the original model is generated randomly. The weight vectors are drawn randomly from the uniform distribution on the simplex in Rm . This is done in the following standard way. The sequence of m 1 random numbers are drawn uniformly from the [0, 1] interval. They are labeled in increasing order of their value as follows: x1 r x2 r ⋯ rxm 1 . The uniformly distributed derived weight vector is ðx1 ; x2 x1 ; x3 x2 ; …; xm 1 xm 2 ; 1 xm 1 Þ. Such a procedure is a standard one to obtain randomly distributed weights, see [5]. It is grounded on the distribution of the order statistics of the uniform random distribution in the unit interval, see [6]. 3. The majority threshold λ is randomly drawn from the [0.5, 1] interval. 4. We generate p profiles by partitioning the [0, 99] interval into p þ 1 equal intervals. Note that proceeding in this way leads to classes of similar “size”; an alternative option would be to draw the profiles at random; it is difficult to assess the impact of this choice on the result. During the elicitation process, the profiles are considered as known and are those of the original model. 5. Original assignments are compared to inferred assignments (see Fig. 2). To measure the closeness of the original and the inferred model, we also generate a large set A of test alternatives. To assess how close the original and the inferred models are, we count the number of alternatives that are assigned to the same category by both models. Such alternatives are called congruent. The proportion of congruent alternatives in the test set is what we will call model accuracy. A set of experiments are designed varying the complexity of the original model, i.e. the number of criteria and the number of categories. We also test the algorithms with different amounts of preference information, i.e. numbers of assignment examples. The parameter settings of the experiments are shown in Table 1. Our experimental strategy aims mainly at observing the effect of varying the learning set size. A secondary objective is to study the effect of other variables describing the dataset, namely, the number of criteria, the number of categories and, for noisy datasets (cf. Section 4.2.3), the proportion of wrong assignments in the learning set. To do so, we vary individually these three variables (setting the other two to “average” values). 4.2. Experiments and results 4.2.1. Learning ability Experiments: We study the ability of the elicitation algorithm to retrieve the original ELECTRE TRI model. It cannot be expected that the model can be accurately inferred when little input information is provided. In other words, if the number of assignment examples is limited, there should exist many ELECTRE TRI models compatible with the assignment examples, which means that the inferred model is almost arbitrary. It is expected that increasing the Table 1 Parameter settings for the experimental design. Parameters Values considered Number of test alternatives (set A) Number of assignment examples (set An) 10 000 5, 10, 15, 20, 30, …, 100 (Sections 4.2.1 and 4.2.2) 20, 30, …, 100 (Section 4.2.3) 3, 6, 9, 12 2, 4, 6, 8 Number of criteria Number of categories Fig. 3. Proportion of congruent assignments in the test set versus the number of assignment examples used for learning; variable number of criteria (3, 6, 9, 12), 4 categories. Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the number of criteria increases. Fig. 2. Workflow of the experimentation. J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 number of assignment examples (requiring an increased cognitive effort from the DM) will lead to inferred models closer to the original model. This investigation studies the tradeoff between cognitive effort and accuracy of the inferred models. We are interested in how many assignment examples are necessary to obtain an ELECTRE TRI model which is “close” enough to the original one. To compare the two models, we generate randomly 10 000 alternatives (called test alternatives) and use both the original and the elicited models to assign them to categories. The assignment results are then compared and the proportion of congruent assignments is calculated. The computation times are also recorded to assess the tractability of the elicitation algorithm. Experiments are run 500 times for each data size. Results: Fig. 3 presents the average proportion of congruent assignments in the test set as a function of the number of assignment examples when the original models involve 4 categories and a variable number of criteria. Fig. 5 displays the corresponding average computation time. In these figures, like in most others in the rest of the paper, the values taken by the variable of interest in the 500 repetitions of the experiment are summarized by a box and a whisker plot. The lower and upper sides of each box respectively represent the first and third quartiles of the cumulative distribution of these values. The small square inside each box indicates the median. The downward and upward whiskers respectively extend to the percentiles 10 and 90 of this distribution. In Fig. 3 we observe that the proportion of congruent assignments increases when more assignment examples are provided, which means that the inferred model is getting closer to the original model when more preference information is available. For instance, in the case of ELECTRE TRI models with 4 categories and 6 criteria, providing 20 assignment examples ensures, in about 75% of the cases, that the algorithm infers a model which is able to congruently re-assign 90% of the alternatives. The accuracy reaches 95% in about 50% of the cases. It almost never (less than 10% chance) falls below 80%. For obtaining a similar level of performance when there are 12 criteria, we need more, around 60, examples. This observation is not surprising because the more the assignment examples, the smaller the feasible polytope delimited by the constraints and thus the more determined the inferred model. Fig. 3 thus indicates that, with an increased Fig. 4. Average percentage of congruent assignments in the test set versus the number of assignment examples used for learning; variable number of criteria (3, 6, 9, 12), 4 categories. Whiskers indicate the limits of the 95% confidence interval for the mean percentage. 35 number of criteria, more assignment examples are required to obtain models with the same performance. If we hypothesize a constant “cost” for obtaining each assignment example and a tradeoff between such a cost and the improvement in classification accuracy, the curves and box plots in Fig. 3 can be used to determine an estimate of the “optimal” sample size, provided an acceptable level of risk has been fixed (e.g. 10% or 25% probability of not reaching the expected classification accuracy). Examples of exploitation of the information provided by these curves were given in the previous paragraph. Alternatively, given the size of the available learning set, we can predict the accuracy that can be expected. For instance, in the case of three criteria, if we use a learning set of size 15, we may expect that there is a 50% chance that the inferred model accuracy will lie between 85% and 95%. Another way of interpreting the information obtained through our experiment is in terms of average accuracy. The 500 values of the accuracy obtained by randomly sampling models and learning sets allow us to estimate the expected value of the accuracy. Fig. 4 shows the mean of the 500 accuracy values computed for each value of the experimental parameters. A 95% confidence interval around each mean value is also represented. This diagram can be used when the analyst formulates his/her requirements in terms of minimal average accuracy. For instance, if (s)he wants at least 90% classification accuracy on average (accepting a risk of 2.5% that the mean is actually worse than that), the required number of assignment examples is the smallest sample size for which the lower bound of the confidence intervals exceeds 0.9. Both Figs. 3 and 4 provide rich guidelines for the analyst, be it for helping to choose the size of the learning set or to make up his/her mind regarding the level of accuracy that can reasonably be expected. Fig. 4 also shows a decreasing improvement rate of the mean classification accuracy as the learning set grows. The confidence intervals represented in the figure show that this improvement in the mean value of the proportion of correct assignments is in general statistically significant whenever ten additional items are added to the learning set. It should be emphasized that the intervals of variation depicted in Figs. 3 and 4 definitely differ in nature: the boxplots in Fig. 3 represent the distributions of all proportions of correct assignments obtained in the 500 repetitions of our experiment for each datasize. In contrast, Fig. 4 depicts the 95% confidence interval obtained for the mean proportion of correct assignments. As expected the precision of the confidence interval for the mean proportion is quite good, since it decreases proportionally to the square root of the number of repetitions. The maximum average computation time of all the experiments with different settings is only 206 ms (Fig. 5), which is acceptable even for real time implementation in a decision support system. Furthermore, the computation time increases (but remains acceptable) with the number of assignment examples and the number of criteria. This increase remains reasonable given that the number of criteria impacts the number of variables in the MIP, and each assignment example is transformed to constraints and introduces two binary variables as well. Fig. 6 summarizes the results of experiments in which an inferred ELECTRE TRI model is used to assign 10 000 randomly generated test alternatives. In these experiments, the number of categories is varied while the number of criteria remains fixed to 7, which is approximately the middle of the criteria range (3–12) we considered. We observe that, when the number of assignment examples is limited (less than 15), the proportions of congruent assignments corresponding to experiments with different numbers of categories are clearly different. Due to the scarcity of input information, the inference of an ELECTRE TRI model involves a certain degree of arbitrariness. Thus a larger number of categories means more possibility to make mistakes. When the number of 36 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 Fig. 5. Computation time versus number of assignment examples; variable number of criteria (3,6,9,12), 4 categories. Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the number of criteria increases. Fig. 6. Proportion of congruent assignments versus the number of assignment examples; 7 criteria, variable number of categories (2,4,6,8). Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the number of categories increases. assignment examples is greater than 15, the proportions of right assignments for experiments with different number of categories become almost identical. The curves representing the proportions of congruent assignments for different numbers of categories converge to 100% in a similar way. We explain this phenomenon as follows. More categories make the constraints stemming from assignment examples tighter, resulting in a smaller feasible polytope. At the same time, assigning an alternative to the right category is more information demanding when there are more categories. The two effects cancel out so that the number of categories does not influence the number of assignment examples required to infer a model “close” to the DM's preference. 4.2.2. Robustness analysis Experiments: An important experimental issue refers to the robustness of the inferred sorting model. Our experiment studies Fig. 7. Number of possible assignments versus the number of assignment examples; 4 categories, variable number of criteria (3,6,9,12). Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the number of criteria increases. the relation between the amount of input preference information (number of assignment examples) and the robustness of the induced model. More precisely, we consider the number jRa j of categories to which an alternative a A A can be assigned to, considering the preference information (see Section 3.2). We compute the average value of jRa j, over the set A of test alternatives, as an indicator of the robustness of the sorting model. We vary the number of assignment examples and the complexity of the original model (the number of categories and criteria considered). For each parameter setting, the experiment is repeated 50 times. Results: The experimental results are shown in Fig. 7, where we consider 4 categories and a variable number of criteria. A significant decrease in the average number jRa j of assignable categories can be observed while the number of assignment examples increases. This observation is consistent with the results of the previous experiment in which the learning ability of the inference algorithm was tested. When the number of assignment examples is relatively limited, the average number of assignable categories jRa j is relatively large. This is because many ELECTRE TRI models are compatible with the preference information, so that the alternatives are possibly assigned to several categories when using different instances of the models compatible with the preference information. When the number of assignment examples grows to become relatively large, jRa j tends to decrease and becomes close to 1, which means that almost every alternative can only be assigned to a single category. In this case, only a few models conform to the preference information and the inferred model is close to the original one. The number of criteria also has a strong effect on jRa j. For a given number of assignment examples, jRa j increases with the number of criteria. This is due to the fact that more criteria means more variables in the MIP, giving more flexibility for the inferred models, and hence leaving more possibility for an alternative to be assigned to a category. Fig. 8 shows the impact on robustness of varying the number of categories (the number of criteria being kept fixed). We find that, with a limited number of assignment examples (less than 15), jRa j varies with the number of categories. When the number of assignment examples is not too small (larger than 15), the number of categories has no significant impact on jRa j. The curves corresponding to different category numbers converge to 1 in a J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 Fig. 8. Number of possible assignments versus the number of assignment examples; 7 criteria, variable number of categories (2,4,6,8). Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the number of categories increases. similar way, which means that the number of categories does not influence the robustness. The trend shown in Fig. 8 can be explained similarly to that in Fig. 6, and they both illustrate that the number of categories does not have any impact on the robustness level when sufficient input preference information is provided. 4.2.3. Inconsistency identification Experiments: The experiments are organized as follows. Firstly, the DM's inconsistent preference is simulated by introducing a certain proportion of assignment errors in the set of assignment examples. Just like in the two previous experiments, an original ELECTRE TRI optimistic model is randomly generated. Instead of assigning all alternatives in the learning set to the category suggested by the original optimistic ELECTRE TRI rule, we randomly assign a fixed number of them to one of the neighboring categories (for example, assigning an alternative to C1 or to C3 instead of C2). Secondly, the inconsistency resolution algorithm (Section 3.3, program (47)–(55)) identifies a maximum subset of assignment examples that can be represented in an ELECTRE TRI model. Lastly, we compare the assignments made by the original ELECTRE TRI models and the one inferred by the program (47)–(55) on a random set of 10 000 test alternatives. Note that the assignment examples which cannot be represented in the inferred model do not necessarily correspond to the errors introduced in the assignments made using the original model. We thus study the following two issues: 1. What is the proportion of assignment examples which can be represented by the inferred model and how is this proportion influenced by the number of assignment examples, the error rate, the numbers of criteria and categories? 2. What is the proportion of assignment examples which are assigned to the same category by both the original and the inferred model and how is this proportion influenced by the number of assignment examples, the error rate, the numbers of criteria and categories? The first issue is related to the capacity of the family of all ELECTRE TRI optimistic rules to represent assignments that depart from a 37 given original rule in the family. We refer to it as the representation issue. The second issue questions the capability of the learning algorithm to filter out errors and restate as much as possible the original model. We refer to it as the error correcting issue. Or to put it another way, the latter issue amounts to identify the conditions (size of the learning set, error rate, numbers of criteria and categories) in which the original model is almost determined as the model yielding the best approximation of the assignments in the learning set. To study these questions, for each parameter setting, we run the experiments 500 times. Three rates of assignment errors in the assignment examples are considered: 10%, 20%, and 30%. The number of assignment examples varies from 20 to 100 by steps of 10 examples. The impact of the number of criteria and categories is also studied. Computation times are recorded. Results relative to the representation issue: Fig. 9 displays the mean value (over 500 runs) of the maximum proportion of assignment examples that can be represented by an optimistic ELECTRE TRI model under different levels of assignment errors' rates (10%, 20%, 30%) when the original ELECTRE TRI models involve 7 criteria and 4 categories. Fig. 9 suggests that the maximum proportion of represented assignment examples asymptotically converges to (100 x)% where x stands for the rate of errors introduced in the assignment examples. When the size of the learning set is relatively small, the flexibility of the ELECTRE TRI model makes it possible to restore more than (100 x)% of assignment examples. When the size of the learning set is large, the proportion of assignment examples which cannot be represented by an optimistic ELECTRE TRI model almost equals the proportion of assignment errors introduced. This observation is consistent with the finding that more assignment examples produce more determined models, implying that it becomes much harder for an ELECTRE TRI model to “accommodate” assignment errors. Table 2 shows that the computation time increases with the proportion of errors and the number of assignment examples. Computation times remain acceptable even for the extreme cases of experiments which elicit ELECTRE TRI models with 7 criteria, 4 categories, using 100 assignment examples with 30% errors introduced. Fig. 9. Proportion of represented assignment examples for ELECTRE TRI models involving 7 criteria and 4 categories; 10%, 20%, 30% error rate. Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the error rate increases. 38 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 Table 2 Computation times (s) to identify inconsistencies in learning sets, 10%, 20%, 30% error rate; 7 criteria, 4 categories. % error 10 20 30 ♯ ass. ex. 20 30 40 50 60 70 80 90 100 0.04 0.04 0.08 0.07 0.10 0.22 0.15 0.18 0.34 0.22 0.43 0.90 0.37 0.65 2.77 0.44 1.04 2.93 0.64 1.31 7.87 0.69 1.64 7.61 0.86 2.04 25.51 Fig. 11. Proportion of represented erroneous assignment examples by ELECTRE TRI; 7 criteria, 4 categories, 10%, 20%, 30% error rate. Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the error rate increases. Fig. 10. Proportion of represented correct/erroneous assignment examples by ELECTRE TRI; 7 criteria, 4 categories, 10% error rate. Results relative to the “error correcting” issue: We have just observed that the inferred model tends to assign ð100 xÞ% of the examples to their prescribed category, where x is the error rate in the learning set. This does not imply that the approximate x% of examples incorrectly assigned by the inferred model indeed correspond to the errors introduced in the learning set. One may wonder whether the inferred model is close to the original one, assigning correctly the same examples and incorrectly the others. Fig. 10 depicts the proportion of represented assignment examples (for 7 criteria, 4 categories, 10% errors and a varying size of the learning set), and distinguishes among the represented assignment examples, those corresponding to the correct ones (gray part of the histogram) from those corresponding to errors (white part of the histogram). For instance, with a learning set of 30 assignment examples, on average approximately 92% of the assignment examples are represented, among which 89% correspond to correct assignment examples, and 3% correspond to errors. We observe that, when the learning set size increases, the proportion of correct represented assignment examples increases, while the proportion of incorrect assignment examples decreases. Another interesting question is the following: among the erroneous assignment examples, how does the proportion of represented erroneous assignment examples evolve with the size of the learning set? Fig. 11 shows such proportions for a varying size of the error rate, and varying size of the learning set. We observe that this proportion quickly decreases with the size of the learning set, thus showing a good ability of the algorithm to detect errors. For instance, with 20% errors and a learning set of size 80 Fig. 12. Proportion of congruent assignments versus the number of assignment examples in the learning set; original models involving 4 categories, 7 criteria; variable error rates in learning set (10%, 20%, 30%). Represented: median; box: 25–75%; whiskers: 10–90%. Boxes are slightly shifted from left to right when the error rate increases. (consequently with 16 erroneous assignment examples), this proportion is on average approximately 5%, which means that, on average, less than one erroneous assignment example is represented in the inferred model. Recall that an example assigned to the same category by the original and the inferred model is called a congruent assignment. To study whether the inferred model is close to the original one, we use a test set of 10 000 randomly generated alternatives and we count the number of congruent assignments in this test set. Fig. 12 displays the average proportion of congruent assignments for a varying size of the learning set with 10%, 20% and 30% error rates in the learning set. The original models involve J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 4 categories and 7 criteria. We observe that the proportion of congruent assignments increases with the size of the learning set. The inferred model is getting closer to the original one while more preference information is added. This holds true for all error rates although the convergence is slower for larger error rates. There is no indication that such a tendency will still hold for very noisy data sets (e.g. with an error rate larger than 50%); however, the sort of decision aiding applications targeted by our algorithm typically involve decision makers who provide “reasonably consistent” preference information (in any case, with error rates less than 30%). 5. Conclusion and further research This paper presents algorithms to infer the criteria weights and majority threshold of an optimistic ELECTRE TRI rule and compute robust assignments using this rule, in case the criteria importance vector is not precisely known, but contained in a polytope of acceptable values. This polytope is expressed by the DM through assignment examples. In contrast with the pessimistic rule, inferring an optimistic ELECTRE TRI rule from assignment examples induces disjunctive constraints. We linearize these conditions introducing binary variables, which permits to infer weights and compute robust assignments through MIP. Hence, our first contribution is the development of preference elicitation tools for the ELECTRE TRI optimistic rule, which learn parameter values and compute the corresponding robust assignment. Numerical experiments are conducted to investigate the performance of the algorithms with respect to learning ability, robustness and ability to identify conflicting preferences. The numerical tests prove the algorithms to be effective for realistic data sets' sizes. Furthermore, the experiments provide insights into the amount of input preference information needed to infer an optimistic ELECTRE TRI model in a reliable way. This study gives useful guidelines for a decision analyst involved with a DM in a real-world application when the analyst formulates the decision problem using the optimistic ELECTRE TRI rule. The paper assumes that the profiles of the categories are known a priori, and no veto threshold is taken into consideration. Further research should be pursued in view of relaxing these assumptions. Preliminary results have been obtained by Sobrie et al. [31] in view of eliciting the weights and the limit profiles in the case of the pessimistic rule. The algorithm used for the pessimistic case could be adapted to the optimistic one although this is not straightforward due to the added complexity of the optimistic assignment rule. However, as soon as an algorithm for estimating limit profiles will be available, it will be possible to assess the expressivity of the model and its ability to reproduce the assignments produced by various types of sorting rules as well as the sorting of real data into categories. This is an especially interesting question since, to the best of our knowledge, there has been no attempt at learning ELECTRE TRI – like rules or majority rules in machine learning to sort alternatives in ordered categories. This is due to the fact that learning such rules on the basis of large data sets is complicated. Yet the idea of comparing alternatives to limit profiles using a majority rule is an appealing one for sorting in ordered categories. It has the advantage of potentially involving all criteria in a single decision rule while more traditional decision rules used in classification tend to involve a small number of criteria. This is a good opportunity to see whether global rules, i.e. rules that simultaneously take into account all alternatives descriptors, have a competitive descriptive power at least for particular types of data. 39 Acknowledgments The authors are indebted to two anonymous referees for their remarks that greatly helped to improve previous versions of this manuscript. References [1] Belacel N. Multicriteria assignment method PROAFTN: methodology and medical application. Eur J Oper Res 2000;125(1):175–83. [2] Bouyssou D, Marchant T. An axiomatic approach to noncompensatory sorting methods in MCDM. I. The case of two categories. Eur J Oper Res 2007;178 (1):217–45. [3] Bouyssou D, Marchant T. An axiomatic approach to noncompensatory sorting methods in MCDM. II. More than two categories. Eur J Oper Res 2007;178 (1):246–76. [4] Bouyssou D, Marchant T, Pirlot M, Tsoukiàs A, Vincke P. Evaluation and decision models with multiple criteria: stepping stones for the analyst. New York: Springer; 2006. [5] Butler J, Jianmin J, Dyer J. Simulation techniques for the sensitivity analysis of multi-criteria decision models. Eur J Oper Res 1997;103(3):531–46. [6] Devroye L. Non-uniform random variate generation. New York: Springer; 1986. [7] Dias L, Climaco J. On computing ELECTRE's credibility indices under partial information. J Multi-Criteria Decis Anal 1999;8(2):74–92. [8] Dias L, Mousseau V, Figueira J, Clímaco J. An aggregation/disaggregation approach to obtain robust conclusions with ELECTRE TRI. Eur J Oper Res 2002;138(2):332–48. [9] Dias LC, Sarabando P. A note on a group preference axiomatization with cardinal utility. Decis Anal 2012;9(3):231–7. [10] Doumpos M, Marinakis Y, Marinaki M, Zopounidis C. An evolutionary approach to construction of outranking models for multicriteria classification: the case of the ELECTRE TRI method. Eur J Oper Res 2009;199:496–505. [11] Doumpos M, Zopounidis C. Developing sorting models using preference disaggregation analysis: an experimental investigation. Eur J Oper Res 2004;154(3):585–98. [12] Greco S, Matarazzo B, Słowiński R. Rough sets methodology for sorting problems in presence of multiple attributes and criteria. Eur J Oper Res 2002;138(2):247–59. [13] Jacquet-Lagrèze E. An application of the UTA discriminant model for the evaluation of R&D projects. In: Pardalos P, Siskos Y, Zopounidis C, editors. Advances in multicriteria analysis nonconvex optimization and its applications. Dordrecht: Kluwer Academic; 1995. p. 203–11. [14] Jacquet-Lagrèze E, Siskos Y. Assessing a set of additive utility functions for multicriteria decision making: the UTA method. Eur J Oper Res 1982;10:151–64. [15] Jacquet-Lagrèze E, Siskos Y. Preference disaggregation: 20 years of MCDA experience. Eur J Oper Res 2001;130(2):233–45. [16] Leroy A, Mousseau V, Pirlot M. Learning the parameters of a multiple criteria sorting method. In: Brafman R, Roberts F, Tsoukiàs A, editors. Algorithmic decision theory. Lecture notes in computer science, vol. 6992. Berlin, Heidelberg: Springer; 2011. p. 219–33. [17] Massaglia R, Ostanello A. N-tomic: a support system for multicriteria segmentation problems. In: Korhonen P, Lewandowski A, Wallenius J, editors. Multiple criteria decision support. Lecture notes in economics and mathematical systems, vol. 356. Berlin: Springer Verlag; 1991. p. 167–74. [18] Mavrotas G, Diakoulaki D, Capros P. Combined MCDA IP approach for project selection in the electricity market. Ann Oper Res 2003;120:159–70, http://dx. doi.org/10.1023/A:1023382514182. [19] Metchebon Takougang SA. Contribution à l’aide à la décision en matière de gestion spatialisée. Etude de cas en management environnemental et développement de nouveaux outils (PhD thesis). Belgium: Faculté polytechnique, Université de Mons; 2010. [20] Metchebon Takougang SA, Pirlot M, Some B, Yonkeu S. Assessing the response to land degradation risk: the case of the Loulouka catchment basin in Burkina Faso. In: Bisdorff R, Dias L, Meyer P, Mousseau V, Pirlot M, editors. Evaluation and decision models with multiple criteria: case studies. Berlin: Springer; 2014. [21] Moscarola J, Roy B. Procédure automatique d’examen de dossiers fondée sur une segmentation trichotomique en présence de critères multiples. RAIRO Rech Opér 1977;11(2):145–73. [22] Mousseau V, Dias L, Figueira J, Gomes C, Clímaco J. Resolving inconsistencies among constraints on the parameters of an MCDA model. Eur J Oper Res 2003;147(1):72–93. [23] Mousseau V, Figueira J, Naux JP. Using assignment examples to infer weights for ELECTRE TRI method: some experimental results. Eur J Oper Res 2001;130 (2):263–75. [24] Mousseau V, Roy B, Sommerlatt I. Development of a decision aiding tool for the evolution of public transport ticket pricing in the Paris region. In: Colorni A, MP, Roy B, editors. A-MCD-A aide multicritère à la décision—multiple criteria decision aiding. Luxembourg: Joint Research Center, European Commission; 2001. p. 213–30. [25] Mousseau V, Słowiński R. Inferring an ELECTRE TRI model from assignment examples. J Glob Optim 1998;12(2):157–74. 40 J. Zheng et al. / Computers & Operations Research 49 (2014) 28–40 [26] Mousseau V, Słowiński R, Zielniewicz P. A user-oriented implementation of the ELECTRE TRI method integrating preference elicitation support. Comput Oper Res 2000;27(7–8):757–77. [27] Perny P. Multicriteria filtering methods based on concordance/non-discordance principles. Ann Oper Res 1998;80:137–67. [28] Roy B. Méthodologie multicritère d’aide à la décision. Paris: Economica; 1985. [29] Roy B, Bouyssou D. Aide multicritère à la décision: méthodes et cas. Paris: Economica; 1993. [30] Siskos Y, Grigoroudis E, Krassadaki E, Matsatsinis N. A multicriteria accreditation system for information technology skills and qualifications. Eur J Oper Res 2007;182(2):867–85. [31] Sobrie O, Mousseau V, Pirlot M. Learning a majority rule model from large sets of assignment examples. In: Perny P, Pirlot M, Tsoukiàs A, editors. ADT. Lecture [32] [33] [34] [35] notes in computer science, vol. 8176. Berlin-Heidelberg: Springer; 2013. p. 336–50. Williams J. Model building in mathematical programming. Chichester: John Wiley and Sons; 1999. Yu W. Aide multicritère à la décision dans le cadre de la problématique du tri: méthodes et applications (PhD thesis). Paris: LAMSADE, Université Paris Dauphine; 1992. Zopounidis C, Doumpos M. A multicriteria decision aid methodology for sorting decision problems: the case of financial distress. Comput Econ 1999;14:197–218, http://dx.doi.org/10.1023/A:1008713823812. Zopounidis C, Doumpos M. Multicriteria classification and sorting methods: a literature review. Eur J Oper Res 2002;138(2):229–46.
© Copyright 2026 Paperzz