37
Introduction and context State of the art Conclusion & Perspectives Learning possibilistic networks from data: a survey Maroua Haddad 1,2 , Nahla Ben Amor 1 and Philippe Leray 2 1 Laboratoire de Recherche Opérationnelle de Décision et de Contrôle de Processus (LARODEC), ISG Tunis, Tunisie 2 Laboratoire d’Informatique de Nantes Atlantique (LINA), UMR CNRS 6241, Université de Nantes, France Maroua Haddad Learning possibilistic networks from data 1/21

Learning possibilistic networks from data: a survey

Embed Size (px)

DESCRIPTION

Talk during JFRB 2014, june 2014, IHP, Paris

Citation preview

Page 1: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Learning possibilistic networks from data: a survey

Maroua Haddad1,2, Nahla Ben Amor1 and Philippe Leray2

1 Laboratoire de Recherche Opérationnelle de Décision et de Contrôle deProcessus (LARODEC), ISG Tunis, Tunisie

2 Laboratoire d’Informatique de Nantes Atlantique (LINA), UMR CNRS 6241,Université de Nantes, France

Maroua Haddad Learning possibilistic networks from data 1/21

Page 2: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Outline ...

1. Introduction and context2. State of the art

2.1. Possibilistic network, inference, sampling2.2. Parameter learning2.3. Structure learning

3. Conclusion & Perspectives

Maroua Haddad Learning possibilistic networks from data 2/21

Page 3: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

Possibility theory [Zadeh, 1978], [Dubois and Prade, 1980]

another uncertainty theorycompleting probability theory in order to handle uncertain,imprecise and missing information

Possibility distributionπ ∶ ω → [0,1]

max is equal to 1, notthe integral

Possibility measure Π

is A coherent with π ?Π(A) = maxω∈A π(ω)

Necessity measure Nis A certainly implied by π ?N(A) = 1 −Π(¬A)

Maroua Haddad Learning possibilistic networks from data 3/21

Page 4: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

Possibility theory [Zadeh, 1978], [Dubois and Prade, 1980]

Possibility distributionπ ∶ ω → [0,1]

max is equal to 1, notthe integral

R

π

1

0

0.6

x

Possibility measure Π

is A coherent with π ?Π(A) = maxω∈A π(ω)

Necessity measure Nis A certainly implied by π ?N(A) = 1 −Π(¬A)

Maroua Haddad Learning possibilistic networks from data 3/21

Page 5: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

Possibility theory [Zadeh, 1978], [Dubois and Prade, 1980]

Possibility distributionπ ∶ ω → [0,1]

max is equal to 1, notthe integral

R

π

1

0

0.6

x

Possibility measure Π

is A coherent with π ?Π(A) = maxω∈A π(ω)

Necessity measure Nis A certainly implied by π ?N(A) = 1 −Π(¬A)

Maroua Haddad Learning possibilistic networks from data 3/21

Page 6: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

Possibilistic scaletwo distinct understandings of a possibility distribution

numerical interpretation (PROD)a possibility distribution may encode a piece of impreciseknowledge about a situation, as in approximate reasoning(Quantitative possibility theory)

ordinal interpretation (MIN)where the only important information is the ordering overpossibility values rather than their exact values, usually inorder to describe some user preferences(Qualitative possibility theory)

Maroua Haddad Learning possibilistic networks from data 4/21

Page 7: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

Possibilistic scaletwo distinct understandings of a possibility distribution

numerical interpretation (PROD)a possibility distribution may encode a piece of impreciseknowledge about a situation, as in approximate reasoning(Quantitative possibility theory)

ordinal interpretation (MIN)where the only important information is the ordering overpossibility values rather than their exact values, usually inorder to describe some user preferences(Qualitative possibility theory)

Maroua Haddad Learning possibilistic networks from data 4/21

Page 8: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

R

π

1

0

π1

π2

π3

Total ignorance

Complete knowledge

Specificityπ1 more specific than π2, π2 more specific than π3

from complete knowledgeto total ignorance

Maroua Haddad Learning possibilistic networks from data 5/21

Page 9: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

R

π

1

0

π1

π2

π3

Total ignorance

Complete knowledge

Specificityπ1 more specific than π2, π2 more specific than π3

from complete knowledge

to total ignorance

Maroua Haddad Learning possibilistic networks from data 5/21

Page 10: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Background

R

π

1

0

π1

π2

π3

Total ignorance

Complete knowledge

Specificityπ1 more specific than π2, π2 more specific than π3

from complete knowledgeto total ignorance

Maroua Haddad Learning possibilistic networks from data 5/21

Page 11: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Learning from data

C S R Wc1 s2 r3 w1c3 s1 r2 w2c1 s2 r3 w3c2 s1 r2 w3c2 s2 r2 w1c1 s2 r1 w1c2 s1 r2 w3c1 s2 r3 w3c3 s1 r3 w2c2 s1 r1 w3c2 s1 r2 w1

Which kind of data ?certain dataimprecise datapossibilistic data

Maroua Haddad Learning possibilistic networks from data 6/21

Page 12: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Learning from data

C S R Wc1 s2 r3 w1c3 s1 r2 w2c1 s2 r3 w3c2 s1,s2 r2 w2,w3c2 s2 r1,r2 w1c1 s2 r1c2 s1 r2 w3c1 s2 r3 w3c3 s1 r3 w2c2 s1 r2,r3 w3c2 s1,s2 r2 w1,w2,w3

Which kind of data ?certain dataimprecise datapossibilistic data

Maroua Haddad Learning possibilistic networks from data 7/21

Page 13: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Learning from data

C S R Wc1 s2 r3 w1c3 s1 r2 w2c1 s2 r3 w3c2 [1,0.8] r2 [1,1,1]c2 s2 [1,1,1] w1c1 s2 r1 [0.2,0.5,1]c2 s1 r2 w3c1 s2 r3 w3c3 s1 r3 w2c2 s1 [0.1,0.9,1] w3c2 [0.2,1] r2 [1,1,1]

Which kind of data ?certain dataimprecise datapossibilistic data

Maroua Haddad Learning possibilistic networks from data 8/21

Page 14: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Outline ...

1. Introduction and context2. State of the art

2.1. Possibilistic network, inference, sampling2.2. Parameter learning2.3. Structure learning

3. Conclusion & Perspectives

Maroua Haddad Learning possibilistic networks from data 9/21

Page 15: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Possibilistic networks [Fonck, 1994]

Sprinkler Rain

Wet Grass

Π (S|C) Π (R|C)

Π (W|SR)

Π (C)

Cloudy

Product-based possibilistic networks

π(ω∣pA) =⎧⎪⎪⎨⎪⎪⎩

π(ω)Π(A) if ω ∈ A0 otherwise.

Min-based possibilistic networks

π(ω∣mA) =⎧⎪⎪⎪⎨⎪⎪⎪⎩

1 if π(ω) = Π(A) and ω ∈ Aπ(ω) if π(ω) < Π(A) and ω ∈ A0 otherwise.

Possibilistic chain rule

π(V1, ..,VN) = ⊗i=1..NΠ(Vi ∣Pa(Vi))

Maroua Haddad Learning possibilistic networks from data 10/21

Page 16: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Possibilistic inference

Possibilistic inferencejunction tree [Fonck, 1992] [Borgelt and Kruse, 1998] PRODMINanytime propagation [Ben Amor et al., 2003] MINcompilation techniques [Ayachi et al., 2010] PROD MINloopy belief propagation [Ajroud and Benferhat, 2014] MIN

Maroua Haddad Learning possibilistic networks from data 11/21

Page 17: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Simulation

Sampling πin the Quantitative possibilistic frameworkby combining Monte Carlo random sampling and α-cutscertain data [Chanas and Nowakowski, 1988]imprecise data [Guyonnet, 2003]

and for a PROD possibilistic network ?

forward sampling (used for BNs) seems ok for certain databut how to deal with imprecise data, i.e. sampling data fromXi when its parents don’t have a certain value ?

Maroua Haddad Learning possibilistic networks from data 12/21

Page 18: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Simulation

Sampling πin the Quantitative possibilistic frameworkby combining Monte Carlo random sampling and α-cutscertain data [Chanas and Nowakowski, 1988]imprecise data [Guyonnet, 2003]

and for a PROD possibilistic network ?

forward sampling (used for BNs) seems ok for certain databut how to deal with imprecise data, i.e. sampling data fromXi when its parents don’t have a certain value ?

Maroua Haddad Learning possibilistic networks from data 12/21

Page 19: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Parameter learning

Objective

for a given structure, how can we estimate Π(Xi ∣Pa(Xi)) ?satisfying Maximum Uncertainty Principle (MUP) [Klir, 1990]When a problem solution is undetermined, the possiblesolution with the highest uncertainty should be chosenin Possibility theory : Minimize Non-Specificity

Maroua Haddad Learning possibilistic networks from data 13/21

Page 20: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... by using Probability-Possibility transformation

Direct transformationsseveral existing transformations with different properties[Klir and Parviz, 1992], [Dubois et al., 1993, 2004],[Mouchaweh et al., 2006], [Bouguelid, 2007]applicable to the joint possibility distribution

Parameter learning ?certain data → joint probability distributiontransformation into a joint possibility distributionthen marginalization in order to find the conditional possibilitydistributions

Inconvenientsworking with the joint distributions is not efficientsupposing that the probability estimation is perfect is notrealistic

Maroua Haddad Learning possibilistic networks from data 14/21

Page 21: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... by using Probability-Possibility transformation

Direct transformations

Parameter learning ?certain data → joint probability distributiontransformation into a joint possibility distributionthen marginalization in order to find the conditional possibilitydistributions

Inconvenientsworking with the joint distributions is not efficientsupposing that the probability estimation is perfect is notrealistic

Maroua Haddad Learning possibilistic networks from data 14/21

Page 22: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... by using Probability-Possibility transformation

Direct transformations

Parameter learning ?certain data → joint probability distributiontransformation into a joint possibility distributionthen marginalization in order to find the conditional possibilitydistributions

Inconvenientsworking with the joint distributions is not efficientsupposing that the probability estimation is perfect is notrealistic

Maroua Haddad Learning possibilistic networks from data 14/21

Page 23: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... by using Probability-Possibility transformation

Confidence interval[De Campos and Huete, 2001]mixture of [Klir and Parviz, 1992] and [Dubois et al., 1993,2004] transformationsdeals with 2 probability distributions min and max.

, applicable to local conditional possibility distributions

Inconvenients/ but do not conserve joint possibility distribution

Maroua Haddad Learning possibilistic networks from data 15/21

Page 24: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... by using Probability-Possibility transformation

Confidence interval[De Campos and Huete, 2001]mixture of [Klir and Parviz, 1992] and [Dubois et al., 1993,2004] transformationsdeals with 2 probability distributions min and max.

, applicable to local conditional possibility distributions

Inconvenients/ but do not conserve joint possibility distribution

Maroua Haddad Learning possibilistic networks from data 15/21

Page 25: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... directly from data

Using probability transformation with confidence intervals ?

[Masson and Denœux, 2006] Certain dataconsider a confidence interval for the estimation of the jointprobability distributionfind all the possibility distributions compatible with theseconstraints... but sometimes, return no solution

Using possibilistic histograms ?

Conclusionno existing solution for parameter learningsome possible solutions (possibilistic histograms) for PRODnetworks

Maroua Haddad Learning possibilistic networks from data 16/21

Page 26: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... directly from data

Using probability transformation with confidence intervals ?

Using possibilistic histograms ?

[Joslyn, 1994]certain data, imprecise data, interval dataconditional distributions

, satisfy Minimum Non-Specificity principle

Conclusionno existing solution for parameter learningsome possible solutions (possibilistic histograms) for PRODnetworks

Maroua Haddad Learning possibilistic networks from data 16/21

Page 27: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

... directly from data

Using probability transformation with confidence intervals ?

Using possibilistic histograms ?

Conclusionno existing solution for parameter learningsome possible solutions (possibilistic histograms) for PRODnetworks

Maroua Haddad Learning possibilistic networks from data 16/21

Page 28: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Structure learning

Constraint-based methodsno existing algorithmhow to measure possibilistic dependency ?

possibilistic mutual information (imprecise data) [Borgelt andKruse, 2003]possibilistic χ2 (imprecise data) [Borgelt and Kruse, 2003]

Score-based methods

Hybrid methods

Conclusionexisting but not satisfying solutions for PROD networkslot of open questions : for instance, what is Markovequivalence here ?

Maroua Haddad Learning possibilistic networks from data 17/21

Page 29: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Structure learning

Constraint-based methods

Score-based methodsadaptation of BN algorithms for certain data [Gebhardt andKruse, 1996] or imprecise data [Borgelt and Gebhardt, 1997][Borgelt and Kruse, 2003]

/ discordance between global and local scores/ do not take into account the Minimum Non-Specificity

principle

Hybrid methods

Conclusionexisting but not satisfying solutions for PROD networkslot of open questions : for instance, what is Markovequivalence here ?

Maroua Haddad Learning possibilistic networks from data 17/21

Page 30: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Structure learning

Constraint-based methods

Score-based methods

Hybrid methods

POSSCAUSE algorithm (certain data) [Sangüesa et al., 1998]/ incoherence in the definition of the independence test (one

different formula per paper)

Conclusionexisting but not satisfying solutions for PROD networkslot of open questions : for instance, what is Markovequivalence here ?

Maroua Haddad Learning possibilistic networks from data 17/21

Page 31: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Structure learning

Constraint-based methods

Score-based methods

Hybrid methods

Conclusionexisting but not satisfying solutions for PROD networkslot of open questions : for instance, what is Markovequivalence here ?

Maroua Haddad Learning possibilistic networks from data 17/21

Page 32: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

In practice

Applications

automotive Industry [Gebhardt and Kruse, 1995]aerospace [Kruze and Borgelt, 1998]information retrieval [Boughanem et al. 2008]

Implementationspossibilistic inference

POSSINFER [Gebhardt and Kruse, 1995]Pulcinella [Saffiotti and Umkehrer, 1991]Possibilistic networks Toolbox for Matlab [Ben Amor, 2012]

Maroua Haddad Learning possibilistic networks from data 18/21

Page 33: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

In practice

Applications

automotive Industry [Gebhardt and Kruse, 1995]aerospace [Kruze and Borgelt, 1998]information retrieval [Boughanem et al. 2008]

Implementationspossibilistic inference

POSSINFER [Gebhardt and Kruse, 1995]Pulcinella [Saffiotti and Umkehrer, 1991]Possibilistic networks Toolbox for Matlab [Ben Amor, 2012]

Maroua Haddad Learning possibilistic networks from data 18/21

Page 34: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Outline ...

1. Introduction and context2. State of the art

2.1. Possibilistic network, inference, sampling2.2. Parameter learning2.3. Structure learning

3. Conclusion & Perspectives

Maroua Haddad Learning possibilistic networks from data 19/21

Page 35: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Conclusion

Possibilistic networksanother uncertainty theory devoted to imprecision modellingpossibilistic networks essentially studied in term of possibilisticinferencelot of open problem for the learning taskMIN networks are more dedicated for expert elicitation ? ? ?

Perspectives

new learning algorithm (parameters + structure) based onconsistent theoretical propertiesvalidation ? we need benchmarks or data generation from apossibilistic distributionMatlab implementationusing more sophisticated imperfect data (imprecise data,possibilistic data, interval data...)

Maroua Haddad Learning possibilistic networks from data 20/21

Page 36: Learning possibilistic networks from data: a survey

Introduction and context State of the art Conclusion & Perspectives

Conclusion

Possibilistic networks

Perspectives

new learning algorithm (parameters + structure) based onconsistent theoretical propertiesvalidation ? we need benchmarks or data generation from apossibilistic distributionMatlab implementationusing more sophisticated imperfect data (imprecise data,possibilistic data, interval data...)

Maroua Haddad Learning possibilistic networks from data 20/21

Page 37: Learning possibilistic networks from data: a survey

Any question ?