39
Learning the parameters of a multiple criteria sorting method from large sets of assignment examples Olivier Sobrie 1,2 - Vincent Mousseau 1 - Marc Pirlot 2 1 École Centrale de Paris - Laboratoire de Génie Industriel 2 University of Mons - Faculty of engineering July 2, 2013 University of Mons - Ecole Centrale Paris Olivier Sobrie 1,2 - Vincent Mousseau 1 - Marc Pirlot 2 - July 2, 2013 1 / 27

Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Learning the parameters of a multiple criteriasorting method from large sets of assignment

examples

Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2

1École Centrale de Paris - Laboratoire de Génie Industriel2University of Mons - Faculty of engineering

July 2, 2013

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 1 / 27

Page 2: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

1 Introduction

2 Algorithm

3 Experimentations

4 Conclusion

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 2 / 27

Page 3: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Introduction

Introductory example

Application : Lung cancer

Categories :C3 : No cancerC2 : Curable cancerC1 : Incurable cancer

C3 � C2 � C1

I 9394 patients analyzedI Monotone attributes (number of cigarettes per day, age, ...)I Output variable : no cancer, cancer, incurable cancerI Predict the risk to get a lung cancer for other patients on basis of

their attributes

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 3 / 27

Page 4: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Introduction

MR-Sort procedure

Main characteristics

I Sorting procedureI Simplified version of the ELECTRE TRI procedure [Yu, 1992]I Axioms based [Slowínski et al., 2002, Bouyssou and Marchant, 2007a,

Bouyssou and Marchant, 2007b]

Parameters

C1

C2

C3

Cp

Cp−1

Cp−2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2b3

bp−3

bp−2

bp−1

bp

I Profiles’ performances (bh,j forh = 1, ..., p − 1; j = 1, ..., n)

I Criteria weights (wj forn = 1, ..., n)

I Majority threshold (λ)

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 4 / 27

Page 5: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Introduction

MR-Sort procedure

Main characteristics

I Sorting procedureI Simplified version of the ELECTRE TRI procedure [Yu, 1992]I Axioms based [Slowínski et al., 2002, Bouyssou and Marchant, 2007a,

Bouyssou and Marchant, 2007b]

Parameters

C1

C2

C3

Cp

Cp−1

Cp−2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2b3

bp−3

bp−2

bp−1

bp

Assignment rule

a ∈ Ch

⇔∑j :aj≥bh−1,j

wj ≥ λ and∑

j :aj≥bh,j

wj < λ

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 4 / 27

Page 6: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Introduction

Inferring the parameters

What already exists to infer MR-Sort parameters ?

I Mixed Integer Program learning the parameters of an MR-Sort model[Leroy et al., 2011]

I Metaheuristic to learn the parameters of an ELECTRE TRI model[Doumpos et al., 2009]

I Not suitable for large problems : computing time becomes huge whenthe number of parameters or examples increases

Our objective

I Learn a MR-Sort model from a large set of assignment examplesI Efficient algorithm (i.e. can handle 1000 alternatives, 10 criteria, 5

categories)

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 5 / 27

Page 7: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Principe of the metaheuristic

Input parameters

I Assignment examplesI Performances of the examples on the n criteria

Objective

I Learn an MR-Sort model which is compatible with the highest numberof assignment examples, i.e. maximize the classification accuracy,

CA =Number of examples correctly restored

Total number of examples

Difficulty

I Learn all the parameters of an MR-Sort model at the same time

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 6 / 27

Page 8: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Metaheuristic to learn all the parameters

Algorithm

Generate a population of Nmodel models with profiles initialized with aheuristicrepeatfor all model M of the set doLearn the weights and majority threshold with a linear program,using the current profilesAdjust the profiles with a heuristic Nit times, using the currentweights and threshold.

end forReinitialize the

⌊Nmodel

2

⌋models giving the worst CA

until Stopping criterion is met

Stopping criterionStopping criterion is met when one model has a CA equal to 1 or when thealgorithm has run No times.University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 7 / 27

Page 9: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Profiles initialization

Principe

I By a heuristicI On each criterion j , give to the profile a performance such that CA

would be max for the alternatives belonging to h and h + 1 if wj = 1.I Take the probability to belong to a category into account

Example 1 : Where should the profile be set on criterion j ?

critj

a1,j

a2,j

a3,j

a4,j

a5,j

a6,j ai,j Category

a1,j C1

a2,j C1

a3,j C1

a4,j C2

a5,j C2

a6,j C2

Category P (ai ∈ Ch)

C112

C212

C2 � C1

a3,j < bh ≤ a4,j

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 8 / 27

Page 10: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Profiles initialization

Principe

I By a heuristicI On each criterion j , give to the profile a performance such that CA

would be max for the alternatives belonging to h and h + 1 if wj = 1.I Take the probability to belong to a category into account

Example 2 : Where should the profile be set on criterion j ?

critj

a1,j

a2,j

a3,j

a4,j

a5,j

a6,j ai,j Category

a1,j C1

a2,j C1

a3,j C1

a4,j C2

a5,j C1

a6,j C2

Category P (ai ∈ Ch)

C123

C213

C2 � C1

a3,j < bh ≤ a4,j

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 8 / 27

Page 11: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the weights and the majority threshold

Principe

I Maximizing the classification accuracy of the modelI Using a linear program with no binary variables

Linear program

Objective :min∑ai∈A

(x ′i + y ′i ) (1)

∑∀j|aiSjbh−1

wj − xi + x ′i = λ ∀ai ∈ Ah, h = {2, ..., p − 1} (2)

∑∀j|aiSjbh

wj + yi − y ′i = λ− δ ∀ai ∈ Ah, h = {1, ..., p − 2} (3)

n∑j=1

wj = 1 (4)

(5)University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 9 / 27

Page 12: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

Case 1 : Alternative a1 classified in C2 instead of C1 (C2 � C1)

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a1

δb1,1δb1,2

δb1,3 δb1,4

wj = 0.2 for j = 1, ..., 5 ; λ = 0.8

I a1 is classified by the DMinto category C1

I a1 is classified by the modelinto category C2

I a1 outranks b1

I Profile too low on one orseveral criteria (in red)

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 10 / 27

Page 13: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

Case 1 : Alternative a1 classified in C2 instead of C1 (C2 � C1)

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b2

b1

a1

wj = 0.2 for j = 1, ..., 5 ; λ = 0.8

I a1 is classified by the DMinto category C1

I a1 is classified by the modelinto category C2

I a1 outranks b1

I Profile too low on one orseveral criteria (in red)

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 10 / 27

Page 14: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

Case 2 : Alternative a2 classified in C1 instead of C2 (C2 � C1)

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a2

δb1,4δb1,5

wj = 0.2 for j = 1, ..., 5 ; λ = 0.8

I a2 is classified by the DMinto category C2

I a2 is classified by the modelinto category C1

I a2 doesn’t outrank b1

I Profile too high on one orseveral criteria (in blue)

I If profile moved by δb1,2,4 ong4 and/or by δb1,2,5 on g5,the alternative will be rightlyclassified

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 11 / 27

Page 15: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

Case 2 : Alternative a2 classified in C1 instead of C2 (C2 � C1)

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b2

b1

a2

wj = 0.2 for j = 1, ..., 5 ; λ = 0.8

I a2 is classified by the DMinto category C2

I a2 is classified by the modelinto category C1

I a2 doesn’t outrank b1

I Profile too high on one orseveral criteria (in blue)

I If profile moved by δb1,2,4 ong4 and/or by δb1,2,5 on g5,the alternative will be rightlyclassified

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 11 / 27

Page 16: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I V+δh,j (resp. V−δh,j ) : the sets of alternatives misclassified in Ch+1

instead of Ch (resp. Ch instead of Ch+1), for which moving the profilebh by +δ (resp. −δ) on j results in a correct assignment.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a3

δb1,4δb1,5

I C2 � C1

I wj = 0.2 for j = 1, ..., 5I λ = 0.8I a3 ∈ A1←Model

2←DM

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 17: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I V+δh,j (resp. V−δh,j ) : the sets of alternatives misclassified in Ch+1

instead of Ch (resp. Ch instead of Ch+1), for which moving the profilebh by +δ (resp. −δ) on j results in a correct assignment.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

I C2 � C1

I wj = 0.2 for j = 1, ..., 5I λ = 0.8I a3 ∈ A1←Model

2←DM

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 18: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I V+δh,j (resp. V−δh,j ) : the sets of alternatives misclassified in Ch+1

instead of Ch (resp. Ch instead of Ch+1), for which moving the profilebh by +δ (resp. −δ) on j results in a correct assignment.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2V +δ1,5

V −δ1,5

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 19: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profilesI W+δ

h,j (resp. W−δh,j ) : the sets of alternatives misclassified in Ch+1

instead of Ch (resp. Ch instead of Ch+1), for which moving the profilebh of +δ (resp. −δ) on j strengthens the criteria coalition in favor ofthe correct classification but will not by itself result in a correctassignment.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a4

δb1,3 δb1,4 δb1,5

I C2 � C1

I wj = 0.2 for j = 1, ..., 5I λ = 0.8I a4 ∈ A1←Model

2←DM

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 20: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profilesI W+δ

h,j (resp. W−δh,j ) : the sets of alternatives misclassified in Ch+1

instead of Ch (resp. Ch instead of Ch+1), for which moving the profilebh of +δ (resp. −δ) on j strengthens the criteria coalition in favor ofthe correct classification but will not by itself result in a correctassignment.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2W+δ

1,5

W−δ1,5

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 21: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I Q+δh,j (resp. Q−δh,j ) : the sets of alternatives correctly classified in Ch+1

(resp. Ch+1) for which moving the profile bh of +δ (resp. −δ) on jresults in a misclassification.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a5δb1,2δb1,3

δb1,4δb1,5 I C2 � C1

I wj = 0.2 for j = 1, ..., 5I λ = 0.8I a5 ∈ A2←Model

2←DM

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 22: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I Q+δh,j (resp. Q−δh,j ) : the sets of alternatives correctly classified in Ch+1

(resp. Ch+1) for which moving the profile bh of +δ (resp. −δ) on jresults in a misclassification.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2Q+δ

1,5

Q−δ1,5

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 23: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I R+δh,j (resp. R−δh,j ) : the sets of alternatives misclassified in Ch+1 instead

of Ch (resp. Ch instead of Ch+1), for which moving the profile bh of+δ (resp. −δ) on j weakens the criteria coalition in favor of thecorrect classification but does not induce misclassification by itself.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2

a6

δb1,5

I C2 � C1

I wj = 0.2 for j = 1, ..., 5I λ = 0.8I a6 ∈ A1←Model

2←DM

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 24: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

I R+δh,j (resp. R−δh,j ) : the sets of alternatives misclassified in Ch+1 instead

of Ch (resp. Ch instead of Ch+1), for which moving the profile bh of+δ (resp. −δ) on j weakens the criteria coalition in favor of thecorrect classification but does not induce misclassification by itself.

C1

C2

crit1 crit2 crit3 crit4 crit5

b0

b1

b2R+δ

1,5

R−δ1,5

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 12 / 27

Page 25: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

P(b+δ1,j ) =

kV |V+δ1,j |+ kW |W+δ

1,j |+ kT |T+δ1,j |

dV |V+δ1,j |+ dW |W+δ

1,j |+ dT |T+δ1,j |+ dQ |Q+δ

1,j |+ dR |R+δ1,j |

with : kV = 2, kW = 1, kT = 0.1, dV = dW = dT = 1, dQ = 5, dR = 1

P−δ1,5

P+δ1,5

V −δ1,5

V +δ1,5

W−δ1,5

W+δ1,5

Q−δ1,5

Q+δ1,5

R−δ1,5

R+δ1,5

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 13 / 27

Page 26: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Algorithm

Learning the profiles

Overview of the complete algorithmfor all profile bh do

for all criterion j chosen randomly doChoose, in a randomized manner, a set of positions in theinterval [bh−1,j , bh+1,j ]Select the one such that P(b∆

h,j) is maximalDraw uniformly a random number r from the interval [0, 1].if r ≤ P(b∆

h,j) thenMove bh,j to the position corresponding to bh,j + ∆Update the alternatives assignment

end ifend for

end for

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 14 / 27

Page 27: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Experimentations

1. What’s the efficiency of the algorithm ?2. How much alternatives are required to learn a good model ?3. What’s the capability of the algorithm to restore assignments when

there are errors in the examples ?4. How the algorithm behaves on real datasets ?

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 15 / 27

Page 28: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Algorithm efficiency

0 5 10 15 20 25 30

85

90

95

100

Number of iterations

CA

(in%)

2 categories; 10 criteria3 categories; 10 criteria4 categories; 10 criteria5 categories; 10 criteria

I Random model M generatedI Learning set : random alternatives assigned through the model MI Model M ′ learned with the metaheuristic from the learning set

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 16 / 27

Page 29: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Model retrieval

100 200 300 400 500 600 700 800 900 1,000

80

85

90

95

100

Number of assignment examples

CA

ofthegen

eralizationset(in%)

3 categories - 10 criteria

I Random model M generatedI Learning set : random alternatives assigned through model MI Model M ′ learned with the metaheuristic from the learning setI Generalization set : random alternatives assigned through M and M ′

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 17 / 27

Page 30: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Tolerance for errors

0 1 2 3 4 5 6 7 8 9 10

60

70

80

Number of iterations

CA

ofthelearningset(in%)

10 % of errors

20 % of errors

30 % of errors

40 % of errors

I Random model M generatedI Learning set : random alternatives assigned through model M + errorsI Model M ′ learned with the metaheuristic from the learning set

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 18 / 27

Page 31: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Tolerance for errors

5 10 15 20 25 30 35 40

0

10

20

30

Incompatible examples in the learning set (in %)

Errorsin

thegen

eralization

set(in%)

1000 assignment examples; 3 categories; 10 criteria

I Random model M generatedI Learning set : random alternatives assigned through model M + errorsI Model M ′ learned with the metaheuristic from the learning setI Generalization set : random alternatives assigned through M and M ′

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 19 / 27

Page 32: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Application on real datasets

Dataset #instances #attributes #categories

DBS 120 8 2CPU 209 6 4BCC 286 7 2MPG 392 7 36ESL 488 4 9MMG 961 5 2ERA 1000 4 4LEV 1000 4 5CEV 1728 6 4

I Instances split in two parts : learning and generalization setsI Binarization of the categories

Source : [Tehrani et al., 2012]University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 20 / 27

Page 33: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Application on real datasets - Binarized categories

Learning set Dataset MIP MR-SORT META MR-SORT LP UTADIS

20 %

DBS 0.8023 ± 0.0481 0.8012 ± 0.0469 0.7992 ± 0.0533CPU 0.9100 ± 0.0345 0.8960 ± 0.0433 0.9348 ± 0.0362BCC 0.7322 ± 0.0276 0.7196 ± 0.0302 0.7085 ± 0.0307MPG 0.7920 ± 0.0326 0.7855 ± 0.0383 0.7775 ± 0.0318ESL 0.8925 ± 0.0158 0.8932 ± 0.0159 0.9111 ± 0.0160MMG 0.8284 ± 0.0140 0.8235 ± 0.0135 0.8160 ± 0.0184ERA 0.7907 ± 0.0174 0.7915 ± 0.0146 0.7632 ± 0.0187LEV 0.8386 ± 0.0151 0.8327 ± 0.0221 0.8346 ± 0.0160CEV - 0.9214 ± 0.0045 0.9206 ± 0.0059

50 %

DBS 0.8373 ± 0.0426 0.8398 ± 0.0487 0.8520 ± 0.0421CPU 0.9360 ± 0.0239 0.9269 ± 0.0311 0.9770 ± 0.0238BCC - 0.7246 ± 0.0446 0.7146 ± 0.0246MPG - 0.8170 ± 0.0295 0.7910 ± 0.0236ESL 0.8982 ± 0.0155 0.8982 ± 0.0203 0.9217 ± 0.0163MMG - 0.8290 ± 0.0153 0.8242 ± 0.0152ERA 0.8042 ± 0.0137 0.7951 ± 0.0191 0.7658 ± 0.0171LEV 0.8554 ± 0.0151 0.8460 ± 0.0221 0.8444 ± 0.0132CEV - 0.9216 ± 0.0067 0.9201 ± 0.0091

80 %

DBS 0.8520 ± 0.0811 0.8712 ± 0.0692 0.8720 ± 0.0501CPU 0.9402 ± 0.0315 0.9476 ± 0.0363 0.9848 ± 0.0214BCC - 0.7486 ± 0.0640 0.7087 ± 0.0510MPG - 0.8152 ± 0.0540 0.7920 ± 0.0388ESL 0.8992 ± 0.0247 0.9017 ± 0.0276 0.9256 ± 0.0235MMG - 0.8313 ± 0.0271 0.8266 ± 0.0265ERA 0.8144 ± 0.0260 0.7970 ± 0.0272 0.7644 ± 0.0292LEV 0.8628 ± 0.0232 0.8401 ± 0.0321 0.8428 ± 0.0222CEV - 0.9204 ± 0.0130 0.9201 ± 0.0132

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 21 / 27

Page 34: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Experimentations

Application on real datasets

Dataset MIP MR-SORT META MR-SORT LP UTADIS

CPU 0.7542 ± 0.0506 0.7443 ± 0.0559 0.8679 ± 0.0488ERA - 0.5104 ± 0.0198 0.4856 ± 0.0169LEV - 0.5528 ± 0.0274 0.5775 ± 0.017520 %

CEV - 0.7761 ± 0.0183 0.7719 ± 0.0153

CPU - 0.8052 ± 0.0361 0.9340 ± 0.0266ERA - 0.5216 ± 0.0180 0.4833 ± 0.0171LEV - 0.5751 ± 0.0230 0.5889 ± 0.015850 %

CEV - 0.7833 ± 0.0180 0.7714 ± 0.0158

CPU - 0.8055 ± 0.0560 0.9512 ± 0.0351ERA - 0.5230 ± 0.0335 0.4824 ± 0.0332LEV - 0.5750 ± 0.0344 0.5933 ± 0.030580 %

CEV - 0.7895 ± 0.0203 0.7717 ± 0.0259

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 22 / 27

Page 35: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

Conclusion

Conclusions and further research

I Algorithm able to handle large datasetsI Adapted to the structure of the problem

I Comparison of AVF-Sort and MR-SortI Use MR-Sort models with vetoesI Test the algorithm on other datasets

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 23 / 27

Page 36: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of
Page 37: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

References

References I

Bouyssou, D. and Marchant, T. (2007a).An axiomatic approach to noncompensatory sorting methods inMCDM, I : The case of two categories.European Journal of Operational Research, 178(1) :217–245.

Bouyssou, D. and Marchant, T. (2007b).An axiomatic approach to noncompensatory sorting methods inMCDM, II : More than two categories.European Journal of Operational Research, 178(1) :246–276.

Doumpos, M., Marinakis, Y., Marinaki, M., and Zopounidis, C. (2009).

An evolutionary approach to construction of outranking models formulticriteria classification : The case of the ELECTRE TRI method.European Journal of Operational Research, 199(2) :496–505.

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 25 / 27

Page 38: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

References

References II

Leroy, A., Mousseau, V., and Pirlot, M. (2011).Learning the parameters of a multiple criteria sorting method.In Brafman, R., Roberts, F., and Tsoukiàs, A., editors, AlgorithmicDecision Theory, volume 6992 of Lecture Notes in Computer Science,pages 219–233. Springer Berlin / Heidelberg.

Slowínski, R., Greco, S., and Matarazzo, B. (2002).Axiomatization of utility, outranking and decision-rule preferencemodels for multiple-criteria classification problems under partialinconsistency with the dominance principle.Control and Cybernetics, 31(4) :1005–1035.

Tehrani, A. F., Cheng, W., Dembczynski, K., and Hüllermeier, E.(2012).Learning monotone nonlinear models using the choquet integral.Machine Learning, 89(1-2) :183–211.

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 26 / 27

Page 39: Learningtheparametersofamultiplecriteria ...olivier.sobrie.be/slides/slides-euro_2013_sobrie_et_al.pdf · 1 Introduction 2 Algorithm 3 Experimentations 4 Conclusion University of

References

References III

Yu, W. (1992).Aide multicritère à la décision dans le cadre de la problématique dutri : méthodes et applications.PhD thesis, LAMSADE, Université Paris Dauphine, Paris.

University of Mons - Ecole Centrale Paris Olivier Sobrie1,2 - Vincent Mousseau1 - Marc Pirlot2 - July 2, 2013 27 / 27