THE LINEAR MULTIPLE CHOICE KNAPSACK PROBLEM WITH

MCDM 2006, Chania, Greece, June 19-23, 2006
THE LINEAR MULTIPLE CHOICE KNAPSACK PROBLEM WITH TWO
CRITERIA: PROFIT AND EQUITY
George Kozanidis
Systems Optimization Laboratory
Dept. of Mechanical & Industrial Engineering
University of Thessaly
Pedion Areos, 38334 Volos, Greece
E-mail: [email protected]
Emanuel Melachrinoudis
Dept. of Mechanical & Industrial Engineering
Northeastern University
Boston, MA 02115, USA
E-mail: [email protected]
Keywords: linear multiple choice knapsack, balanced resource allocation, equity, multiobjective linear
programming, nondominated frontier
Summary: We study an extension of the Linear Multiple Choice Knapsack (LMCK) Problem
that considers two criteria. The problem can be used to find the optimal allocation of an
available resource to a group of disjoint sets of activities, while also ensuring that a certain
balance on the resource amounts allocated to the activity sets is attained. The first criterion
maximizes the profit incurred by the implementation of the considered activities. The second
criterion minimizes the maximum difference between the resource amounts allocated to any two
sets of activities. We present the mathematical formulation and explore the fundamental
properties of the problem. Based on these properties, we develop an efficient algorithm that
obtains the entire frontier of nondominated solutions. The algorithm is very efficient compared
to generic multiple objective linear programming (MOLP) algorithms. We present theoretical
findings which provide insight into the behavior of the algorithm, and report computational
results which demonstrate its efficiency.
1. Introduction
The multiple choice knapsack problem is a very commonly used variation of the classical knapsack
problem with numerous applications. Many solution techniques that have been developed for the integer
version of the problem rely on the existence of a fast solution algorithm to its linear relaxation. Besides
being used as a relaxation to its integer counterpart, the linear version of the problem models itself
practical applications too. Typical examples include location, production scheduling, and transportation
management problems (Lin, 1998).
In this work we present a solution algorithm for the linear multiple choice knapsack problem with two
criteria, profit and equity. The main advantage of the algorithm is that it is able to obtain the entire
frontier of nondominated solutions. The results of the computational experiments that were conducted
demonstrate that the algorithm exhibits high efficiency, and additionally, that the computational time
savings increase fast as the problem size increases. For the needs of this work, we term BLMCK
1
(Biobjective Linear Multiple Choice Knapsack) the problem under consideration, and we formulate it as
follows:
Max P = ∑ ∑ pki xki
(1)
Min f
(2)
k∈S i∈Rk
s.t.
∑∑c
k∈S i∈Rk
∑x
ki
i∈Rk
x ≤b
(3)
ki ki
≤ lk , k ∈ S
(4)
L ≤ ∑ cki xki ≤ U , k ∈ S
(5)
U −L≤ f
(6)
U , L, f ≥ 0; xki ≥ 0, i ∈ Rk , k ∈ S
(7)
i∈Rk
The decision variables xki represent continuous activities that belong to |S| disjoint sets. The specific set a
variable belongs to is denoted by its first index, k. Set S contains the indices of all these disjoint sets. For
a specific value of k in S, set Rk contains the indices of the decision variables that belong to set k. The
positive parameters pki and cki represent the profit and cost, respectively, incurred per unit of application
of the activity represented by xki. Parameter b is also a positive quantity, denoting the available amount of
resource that can be used towards the implementation of the considered activities.
The objective function (1) maximizes total profit, while constraint (3) limits the total resource amount
used to a maximum value b. The multiple choice constraints (4) restrict the sum of all variables within
each set k to at most lk. Constraints (5) define the auxiliary decision variables U and L in such a way that
cki xki , belongs to
the total resource amount allocated to any single set k (also called cost of the set),
∑
i∈Rk
the interval [L,U]. The width of this interval is restricted to at most f by constraint (6). The second
objective function (2) minimizes f. Constraints (7) ensure that all variables are nonnegative.
The remainder of this work is organized as follows: In Section 2 we explore the fundamental properties
of the problem, and based on them we introduce an efficient algorithm that obtains its entire set of
nondominated solutions. We also present theoretical results that give insight into the algorithm. In
Section 3 we examine the computational complexity of the algorithm and report computational results
which demonstrate its performance in the average case. Finally, in Section 4 we summarize the
contribution of the present work and point to directions for future research.
2. Solution Methodology
2.1 BLMCK Properties
The algorithm we propose for the solution of BLMCK is divided into two phases. In the first phase, the
second objective is ignored and an optimal solution to Problem LMCK is obtained. Starting from this
solution in the second phase, the algorithm incorporates the equity objective and evaluates the
nondominated tradeoff rate (see Steuer, 1986) between the two objectives, as the second objective is
allowed to decrease continuously to zero. Taking advantage of the problem’s special structure, the
algorithm returns its entire nondominated frontier upon termination.
2
For the solution of Problem LMCK in Phase I, we utilize Algorithm LMCK (see Kozanidis and
Melachrinoudis, 2004). Two important properties exist (see Kozanidis and Melachrinoudis, 2004) that
enable us to eliminate in advance several decision variables and reduce this way the size of the problem.
We can visualize graphically the decision variables of each set as points on a two-dimensional graph,
where the x-axis represents the cost of a variable and the y-axis its profit (see Fig. 1 in Kozanidis and
Melachrinoudis, 2004). The variables which are not eliminated in each set define the left upper hull of all
variables that belong to this set.
An important difference between Problems BLMCK and LMCK is that, for Problem BLMCK, variables
which qualify for elimination according to these two properties, but belong to the right upper hull of the
associated multiple choice coefficient space should not be apriori eliminated, because they may take a
positive value in a Pareto optimal solution. Therefore, Algorithm BLMCK, which is introduced in the
next section, does not eliminate in advance such variables. For the remainder of this work, we use the
terminology introduced by Kozanidis et al. (2005).
2.2 Solution Procedure
Let’s assume that the second objective is suppressed from the BLMCK formulation. Note that the
presence of constraints (5) and (6) alone does not affect the feasible region of the problem, since they
artificially enlarge the dimensionality of the original space. In the solution obtained when the resulting
LMCK problem is solved using Algorithm LMCK, f is greater or equal to the maximum difference
realized between the costs of any two sets. We use the terms maxcost and mincost to denote the
maximum and minimum cost, respectively over all sets. Internal is a set whose cost is different from
maxcost and mincost. Upper is a set whose cost is equal to maxcost and lower a set whose cost is equal to
mincost. Based on this terminology, at the end of Phase I, f is greater or equal to (maxcost-mincost).
After the equity objective is incorporated in Phase II, f should be minimized. In order to minimize f, the
algorithm needs to modify the current resource allocation to the multiple choice sets. There are five
candidate options for decreasing f, while also retaining optimality with respect to the first objective. Any
option different than these will lead to a suboptimal solution with respect to the first objective. These
options, displayed in Figure 1, are introduced next. A set is called increasing if the application of an
option increases its cost, and decreasing if the opposite is true. The increasing and decreasing slope of a
multiple choice set are defined in Kozanidis et al. (2005).
Option A
Option B
Option C
Option D
Cost
Cost
Cost
Cost
f
f
f
f
Option E
Cost
f
Figure 1: The five candidate options for decreasing f
Option A: Decrease the resource amount allocated to all upper sets and reallocate the recovered amount
to the internal or lower set with the maximum increasing slope.
Option B: Decrease the resource amount allocated to the internal or upper set with the minimum
decreasing slope and reallocate the recovered amount to the lower sets.
3
Option C: Decrease the resource amount allocated to all upper sets and reallocate the recovered amount
to the lower sets.
Option D: Decrease the resource amount allocated to all upper sets.
Option E: Increase the resource amount allocated to all lower sets (considered only if a positive resource
residual is available).
Let ∆P and ∆f be the marginal differences realized in total profit and in f, respectively, when any one of
these options is applied. Note that, by construction, ∆P is nonpositive and ∆f strictly negative, since the
initial solution is optimal for LMCK (∆P < 0), and we only consider ∆f < 0. Therefore, the ratio ∆P/∆f is
always nonnegative. Out of the five options, the algorithm should apply the one resulting in the minimum
total profit decrease per unit decrease in f. This will ensure that the first objective will have the maximum
possible value for the new improved value of the second objective.
After deciding on the option that should be applied by choosing the one with the minimum ∆P/∆f ratio,
the algorithm must determine when the application of this option should stop. The application of an
option stops when an appropriate stopping condition applies. This happens when at least one of the ∆P/∆f
ratios of the five options changes, or when f becomes equal to 0. In the latter case, this is because at that
point all possible values for f have been considered. In the former case, this is because at that point the
new ∆P/∆f ratios have to be computed and compared again.
At each Phase II iteration, the algorithm developed for Problem BLMCK repeats the following
procedure. First, it selects the option that will be applied. This is done by selecting out of the five
candidate options the one whose ratio ∆P/∆f is minimum (in what follows, we refer with ∆P/∆f to the
minimum such ratio). Next, the algorithm decides the extent of the current iteration by finding the
stopping condition that applies first. This determines the end of the current iteration and the beginning of
the next one. The algorithm terminates with the entire nondominated frontier when f drops to 0. The
following result is crucial for the development of the algorithm for Problem BLMCK.
Proposition 1: The initial solution, obtained at the end of Phase I, is Pareto optimal for Problem BLMCK
if and only if ∆P/∆f > 0.
Proof: Since this solution is optimal for the single objective problem, the result follows from the
definition of Pareto optimality and the fact that no other option exists, resulting in a smaller ∆P/∆f ratio.
The situation is illustrated in Figure 2, which displays the optimal solution obtained at the end of Phase I
in criterion space. In the first case, ∆P/∆f = 0 for this initial solution. This means that the single objective
problem has multiple optimal solutions and at least one of them is satisfied with f ’< f. Since we can move
to a solution with a better value for the second objective without worsening the value of the first
objective, the initial solution is not Pareto optimal. Therefore, the initial, together with all intermediate
solutions, are dominated by the solution obtained at the end of the first iteration (P = P1 for all of them).
In the second case, ∆P/∆f > 0. This means that in order to move to a solution with a better value for the
second objective, we must worsen the value of the first objective. Therefore, the initial solution is Pareto
optimal. Additionally, all intermediate solutions are Pareto optimal (nondominated line segment), since P
decreases strictly as f decreases. Of course, the ratio ∆P/∆f can never be negative, as explained above.
Next, the algorithm developed for Problem BLMCK is formally introduced.
Algorithm BLMCK
Phase I (Optimal solution of LMCK)
Without eliminating variables which belong to the right upper hull of the associated multiple choice
coefficient space, use Algorithm LMCK to find the optimal solution to the LMCK problem that results
when the second criterion is suppressed.
Phase ΙΙ (Solution of ΒLMCK)
while (f > 0) do {
select option that minimizes ΔP/Δf
find stopping condition and compute |Δf|
4
iterate, update solution and find new value of f
if ΔP/Δf > 0, add the initial and all intermediate solutions of the iteration that was
just completed to the set of nondominated solutions of the problem
}end while
P
P
Case 2: ΔP/Δf > 0
Initial solution Pareto optimal
Case 1: ΔP/Δf = 0
Initial solution not Pareto optimal
P1
P2
P1
f2
f1
f
f2
f1
f
Figure 2: Determining whether the solution obtained at the end of Phase I is Pareto optimal or not
2.3 Insight into Algorithm BLMCK
To get further insight into Algorithm BLMCK, assume that the second objective becomes Min (U − L).
Then, we can replace this objective by constraint (6), and parametrically change the value of f, in order to
get the entire nondominated frontier of the problem. Essentially, this is the parametric e-constraint
reduced feasible region method for obtaining the entire nondominated frontier of a MOLP problem (see
Steuer, 1986). Based on this observation, the validity of the algorithm is summarized in the following
result:
Proposition 2: Algorithm BLMCK correctly obtains the entire nondominated frontier of Problem
BLMCK.
Proof: Phase I of Algorithm BLMCK terminates with an optimal solution to Problem LMCK. If LMCK
has an alternative optimal solution with a better f-value, a nondominated criterion vector (P,f) will be
determined in Phase II. Otherwise, the starting criterion vector is nondominated. At each Phase II
iteration, the algorithm applies the option for which the marginal decrease in total profit is the minimum
possible per unit decrease in f. Therefore, at the end of each iteration, the first objective has the maximum
possible value for the current value of the second objective. This means that optimality is maintained
throughout for the first objective with respect to the current value of the second objective. Since f is
changed continuously, as soon as its value drops to 0, the algorithm has identified the optimal values of
the first objective for all possible values of the second objective. Therefore, the entire nondominated
frontier of Problem BLMCK has been obtained.
Algorithm BLMCK is related to Algorithm LMCKE (Kozanidis et al. 2005). The differences between the
two algorithms are the following. Algorithm LMCKE solves a single objective linear program and, as a
result, returns a unique optimal solution. On the other hand, Algorithm BLMCK solves a MOLP
problem, by computing the entire set of nondominated solutions, which in the general case is an infinite
set. In Algorithm LMCKE, f is a parameter with specific fixed value, while in Algorithm BLMCK it is a
decision variable. In Algorithm BLMCK, an iteration is carried out as long as f > 0. On the other hand,
Algorithm LMCKE terminates when the desired value of f is obtained. Additionally, the stopping
conditions that apply in each of the five options are not the same for the two algorithms. This is mainly
due to the fact that some of the stopping conditions of Algorithm LMCKE do not act as stopping
conditions in the case of Algorithm BLMCK. Finally, Algorithm LMCKE incorporates upper bounds on
the optimal cost of each set, while this is not true for Algorithm BLMCK.
5
At each Phase II iteration, Algorithm BLMCK moves to a new solution with a strictly lower f−value.
Implicitly, it visits an infinite number of intermediate solutions corresponding to the infinite number of
intermediate values that f takes between its initial and its final value. Proposition 1 clearly extends to each
of these iterations. Hence, there are two distinct cases. If ΔP/Δf = 0 at the beginning of an iteration, then
the initial and all intermediate solutions will be dominated by the solution obtained at the end of this
iteration. Whether this last solution will be Pareto optimal or not depends on the value of ΔP/Δf that will
be computed at the next iteration. If ΔP/Δf > 0, then the initial and all solutions encountered from that
point on, until termination of the algorithm, will be Pareto optimal. This is due to the following
interesting result:
Proposition 3: Let q and t be two Phase II iterations of Algorithm BLMCK, such that q is carried out
before t. Let also rq be the value of the ratio ΔP/Δf for iteration q and rt the value of the ratio ΔP/Δf for
iteration t. Then, rq ≤ rt.
Proof: Consider the feasible region of Problem BLMCK in criterion space. This region is an open
polyhedron in the nonnegative orthant of the two dimensional plane, since the two objectives always take
nonnegative values, and there always exists an infinite number of feasible values for variable f. The profit
objective takes its maximum value at the end of Phase I of Algorithm BLMCK. As the value of the equity
objective decreases in Phase II, the value of the profit objective is always nonincreasing. The validity of
the proposition results from the fact that the feasible region of Problem BLMCK in criterion space is
convex and the set of its nondominated solutions is connected, since this is a MOLP problem.
Proposition 3 implies that the ratio ΔP/Δf can only be equal to 0 for a sequence of iterations at the
beginning of Phase II. Once this ratio takes a positive value, it will remain positive until termination of
the algorithm. Additionally, this result states that, for any two Phase II iterations, the ratio of the one
which was performed earlier cannot be larger than the ratio of the one which was performed later. This is
an interesting result which agrees with practical experience. The intuition is that the more we need to
"squeeze" f in order to attain higher level of equity, the higher price (in terms of profit) we have to pay.
To put it in a different way, a unit of resource balance costs less when the solution at hand is unbalanced
than when it is somewhat balanced. As the balance level of the solution at hand increases, this unit cost
increases too.
When f drops to 0, the entire nondominated frontier of the problem has been obtained and the algorithm
terminates. Clearly, for this final solution, the following result holds, since we cannot improve any of the
two objectives without worsening the value of the other.
Corollary 1: The solution obtained from Algorithm BLMCK when f drops to 0 is Pareto optimal.
Two interesting observations with respect to Algorithm BLMCK should be made at this point. The first is
that it is possible for a solution in which money is simply “wasted” without reaching a higher profit in
one of the multiple choice sets to be Pareto optimal. It can be assumed that each set will prefer a solution
with a higher profit for this particular set as compared to a solution with a smaller individual profit, but a
higher individual cost. The decision maker should always keep this in mind when reviewing the set of
nondominated solutions in order to decide on the one of maximum utility. Even so, under certain
conditions this situation will never come up. For example, this situation will never come up when the
upper and the left upper hull in each multiple choice coefficient space coincide. This is because, in that
case, only nonnegative values will be possible for the increasing slope of each set, therefore the profit of
a set will never decrease when its cost increases.
The second interesting observation that should be pointed out is the following. Among the set of Pareto
optimal solutions of the problem, there may exist some with a positive resource residual, although
additional activities exist which have not been selected for implementation. When this situation is
encountered, a positive resource amount has been preserved, because there is no way this can be used to
increase total profit, while also maintaining the current level of equity. As the number of considered
activities in each set increases, the probability that this situation will come up decreases. This is because,
6
at the presence of more candidate activities in each set, it is more likely that it will be possible to divide
the unused resource amount among the sets in a way that total profit increases, while also maintaining the
current level of equity.
The above two observations imply that it is wiser to give to the profit objective higher priority than to the
equity one. Usually, very high level of equity should be sought when this is absolutely desirable, and the
resulting loss in profit can be tolerated.
3. Computational Implementation
3.1 Complexity Analysis
Let r be the number of multiple choice sets, Nk the number of variables in multiple choice set k, N
=
N k and Nmax = max N k . As shown in the analysis of Kozanidis et al. (2005), the worst case
∑
k∈S
k ∈S
complexity of Phase I is O(N logNmax) + O(N logr) = O(N logm), where m = max (Nmax, r). The work
needed at each Phase II iteration to find the ratios ΔP/Δf of the five options, find the optimal value of ∆f,
and update to get the new solution that results is O(r). The number of Phase II iterations depends on the
number of times each of the different stopping conditions applies. In what follows, the average-case
performance of the algorithm is examined by analyzing the computational experiments that were
conducted.
3.2 Computational Experiments
Algorithm BLMCK was coded in C/C++ and tested on a Pentium IV/2.5 GHz processor. The results
obtained are presented in Tables 1 and 2. In these tables, Nk was assumed to be the same for all k. Both r
and Nk varied between 150 and 600 in steps of 150. This means that the biggest problems solved
contained 360,000 decision variables. Two types of problems were tested. In type A problems, variables
that should be eliminated were also considered among the initial variables of the problem. Problems of
this type resemble the problems that arise in real world applications. In type B problems, the variable
coefficients were randomly generated in such a way that no variable qualified for elimination. Problems
of this type provide good insight into the performance of the algorithm, yet would rarely arise in practice.
The algorithm exhibited high variability for type A problems. The computational time for different
instances of the same size varied significantly for both Phases I and II. As a result, 50 random instances
were solved for each problem size. The results reported in columns 3-6 of Table 1 are the average and
maximum time in seconds needed for completion of Phase I and termination of the algorithm,
respectively. The last column of this table presents the average percentage of total variables that were
eliminated. It is clear from this column that the vast majority of the initial variables were eliminated.
Table 2 presents similar results as those of Table 1, for type B problems. The variance exhibited among
problems of the same size was much smaller in this case and as a result, 30 instances were solved for
each problem size. The random variate generation ensured that all variables belonged to the upper hull of
the associated coefficient space.
7
Table 1: Computational times (in seconds) for type A problems
Phase I
r
150
150
150
150
300
300
300
300
450
450
450
450
600
600
600
600
Nk
150
300
450
600
150
300
450
600
150
300
450
600
150
300
450
600
Avg
0.02002
0.03502
0.05390
0.06974
0.04064
0.07486
0.10914
0.14280
0.06108
0.11324
0.16804
0.21972
0.08288
0.15386
0.22058
0.29382
Total Time
Μax
0.040
0.041
0.100
0.100
0.060
0.110
0.150
0.200
0.101
0.171
0.221
0.291
0.120
0.211
0.291
0.380
Avg
0.04750
0.06528
0.08270
0.10114
0.28426
0.28992
0.35974
0.39756
0.98578
1.38052
1.22900
1.49870
2.43124
2.88496
2.99664
3.33722
Percentage
eliminated
Μax
0.090
0.101
0.130
0.141
0.561
0.631
0.681
0.942
2.573
3.204
3.415
6.059
6.409
10.786
8.833
9.654
95.2%
97.3%
98.1%
98.4%
95.3%
97.3%
98.1%
98.5%
95.1%
97.3%
98.0%
98.5%
95.1%
97.3%
98.1%
98.5%
Table 2: Computational times (in seconds) for type B problems
Phase I
r
150
150
150
150
300
300
300
300
450
450
450
450
600
600
600
600
Νk
150
300
450
600
150
300
450
600
150
300
450
600
150
300
450
600
Avg
0.11847
0.32083
0.64423
1.03803
0.26263
0.73273
1.39070
2.26900
0.40063
1.16713
2.27997
3.80067
0.58820
1.72463
3.47400
6.96337
Total Time
Max
0.160
0.331
0.751
1.162
0.350
0.872
1.503
2.384
0.471
1.282
2.364
3.925
0.701
1.813
4.287
9.754
8
Avg
0.13480
0.33817
0.65893
1.05607
0.39693
0.87617
1.52767
2.42580
0.88970
1.68743
2.79707
4.39387
1.82340
3.00337
4.79493
8.69147
Max
0.180
0.341
0.761
1.182
0.511
1.002
1.662
2.544
1.011
1.852
2.874
4.537
1.963
3.144
6.520
11.106
3.3 Discussion of Results
The variability exhibited for type A problems becomes clear from the results of Table 1. We observe that,
for some problem sizes, the average time needed for termination of the algorithm is almost 25% of the
corresponding maximum value. On the other hand, as it is clear from Table 2, type B problems exhibit a
much lower variability. This is because, in type B problems, a much larger number of activities are
considered in each set, since no variable is eliminated. Consequently, the sets look similar with each other
and the distribution of the available resource is much more balanced in the solution obtained after
termination of Phase I, which in turn simplifies the task of Phase II. Therefore, the total computational
effort does not differ very much for different instances of the same size.
For type A problems and fixed Nk, the percentage of total execution time devoted to Phase II seems to
increase as the total number of sets increases. On the other hand, for fixed r, the percentage of total
execution time devoted to Phase I seems to increase as the number of variables in each set increases. The
same behavior is observed for type B problems. Therefore, the number of variables in each set has a
greater impact on the computational effort of Phase I, while the number of sets on the computational
effort of Phase II.
Additionally, the percentage of total time devoted to Phase I is much smaller in type A problems than in
type B problems of the same size. This is not surprising, since the ordering of variables is the dominating
operation in Phase I, while the arrangement of sets is the dominating operation in Phase II. Hence, the
algorithm spends a larger percentage of the total time on ordering the variables in type B problems than
in type A problems. On the other hand, the algorithm spends a larger percentage of the total time on
arranging the sets in type A problems than in type B problems.
For the same total number of variables, the total computational effort increases as the number of sets
increases. This is more evident in type A problems, where for fixed total number of variables, an increase
in r results in an increase of the expected number of variables that are not eliminated (see Sinha and
Zoltners, 1979). On the other hand, for fixed r, the number of variables that remain after elimination does
not increase very fast for type A problems, even when the value of Nk increases from 150 to 600.
Therefore, the number of sets seems to be more crucial in the total computational effort of the algorithm
than the number of variables in each set.
Comparing the computational effort needed for the solution of a type A instance with that of a type B
instance of the same size, we observe that with very few exceptions, this is always smaller in type A
problems. This is mainly due to the fact that in type B problems no variable is eliminated and therefore a
larger computational effort is needed in Phase I for the ordering of the variables and the construction of
the multiple choice lists. On the other hand, when we compare the net time needed for execution of Phase
II this is smaller in type B than in type A problems of the same size.
The entire nondominated frontier of a 150X150 type A instance is shown in Figure 3. The feasible region
of the problem in objective space consists of the polyhedron formed by the nonnegative portions of the
two axes and this frontier, after it is extended to the right with a straight horizontal line. This is always an
open polyhedron, since there exists an infinite number of feasible values for f. The nondominated frontier
is always concave, ensuring that the feasible region of the problem in objective space is convex.
The result of Proposition 2 can be clearly visualized in these two diagrams. As we move towards the left
on the nondominated frontier, the marginal profit sacrificed in order to attain the same decrease on f
increases. Thus, the higher level of equity we want to attain, the higher price we have to pay. Let P(f ) be
the y-coordinate of the nondominated solution whose x-coordinate is equal to f. The above discussion
implies that the following result holds:
Corollary 2: P is an increasing piecewise linear concave function of f.
9
Figure 3: Nondominated frontier of a 150X150 type A problem
The graph of the nondominated frontier can be very useful in the decision making procedure. The
decision maker is presented with the whole nondominated set and the nondominated tradeoff rate (∆P/∆f
as a function of f) and can select the nondominated vector with his/her best tradeoff between the two
objectives. The nondominated vector could even be on a nondominated edge, not necessarily an extreme
point. Hence, the decision maker can decide directly on this diagram which solution gives the best
compromise between the two objectives. Of course, once this decision has been made, a specific Pareto
optimal solution can be found by modifying appropriately the terminating condition of Algorithm
BLMCK, if the history of iterations has not been stored.
In order to test the performance of Algorithm BLMCK further, we also developed an AMPL model (see
Fourer et al., 2002) for Problem BLMCK. To overcome the difficulty introduced by the second objective,
we used the e-constraint reduced feasible region method for multicriteria optimization (see Steuer, 1986)
and transformed it to a constraint by giving a specific value to f. This way, the second objective was
restricted to a maximum value, instead of being minimized. To ensure that the solution obtained this way
is Pareto optimal, we utilized Corollary 1 and set f equal to 0.
Of course, the main drawback of this approach is that, in contrast to Algorithm BLMCK, it is only able to
find a single nondominated solution and not the entire nondominated frontier of the problem. Even so,
the time Algorithm BLMCK needs to solve the problem is much smaller than the time required by
AMPL. In type A problems with r = 300, Nk = 300, AMPL (version 9.1, using CPLEX solver with
default values) needs approximately 1.5 seconds to find the optimal solution on the same machine. When
r = 450, Nk = 450, this time increases to approximately 5 seconds. For type B problems, the differences
between the algorithm and AMPL are drastically larger. When r = 300, Nk = 300, AMPL needs
approximately 10 seconds to find the optimal solution. When r = 450, Nk = 450, this time increases to 105
seconds. These time savings increase fast as the problem size increases. Therefore, the present algorithm
can be very useful to practitioners for solving real life applications with a large number of decision
variables. The reason that there is such a big difference in computational effort between the two problem
types is probably because in type A problems, the solver manages to detect the fact that the problem is
reducible in size which results in significant time savings.
Vector maximization algorithms (Steuer, 1986), that can alternatively be used to find the efficient set of
the problem, are reasonably expected to consume considerably more time than what AMPL does to find a
10
single nondominated solution. This observation makes the efficiency of Algorithm BMCK even clearer,
since it manages to obtain the entire nondominated frontier in a timely and straightforward fashion.
4. Conclusions
In this work, we introduced an extension of the Linear Multiple Choice Knapsack Problem, which
incorporates a second objective. This objective accounts for an equitable allocation of the available
resource. We presented the mathematical formulation of the problem and showed that this problem
structure exhibits several fundamental properties. These were used to develop an optimal two-phased
algorithm for the problem. In Phase I, the algorithm ignores the second objective and enhances an
existing method for the Linear Multiple Choice Knapsack Problem to obtain an initial solution. Phase II
incorporates the second objective and successively obtains the entire nondominated frontier of the
problem.
Our computational results illustrate the efficiency of the algorithm. It outperforms a commercial linear
programming package and its superiority increases with problem size. The algorithm performs well
because it exploits the special structure of the problem. It utilizes the special relationship between the two
objectives. In addition, typical software packages can only solve single objective problems, which means
that they can only find one out of the many Pareto optimal solutions of the problem. This makes the
multiple advantages of the algorithm even more evident.
The present work points to a number of directions for future research. We have illustrated a situation in
which direct tradeoffs between the two objectives can be developed in a biobjective linear program. The
concept of the algorithm can be extended to other multiobjective problems, in which direct tradeoffs
between the objectives can be derived. We believe that the insights gained into this problem structure and
the advantages of the algorithm will prove useful in real world applications that involve a large number
of decision variables.
References
Fourer, R., Gay, D.M. and Kernighan, B.W. (2002) AMPL: A Modeling Language for Mathematical
Programming, Duxbury Press.
Kozanidis, G. and Melachrinoudis, E. (2004) “A branch & bound algorithm for the 0-1 mixed integer
knapsack problem with linear multiple choice constraints” Computers & Operations Research, 31, 695711.
Kozanidis, G., Melachrinoudis, E. and Solomon, M. (2005) “The linear multiple choice knapsack
problem with equity constraints” International Journal of Operational Research, 1, 52-73.
Lin, E. Y-H. (1998) “A bibliographical survey on some well known non-standard knapsack problems”
INFOR, 36, 274-317.
Sinha, P. and Zoltners, A.A. (1979) “The multiple choice knapsack problem” Operations Research, 27,
503-515.
Steuer, R. E. (1986) Multiple Criteria Optimization: theory, computation and application, New York:
Wiley.
11