Kullback-Leibler for Bayesian networks
In [1]:
import os
%matplotlib inline
from pylab import *
import matplotlib.pyplot as plt
Initialisation
importing pyAgrum
importing pyAgrum.lib tools
loading a BN
In [2]:
import pyAgrum as gum
import pyAgrum.lib.notebook as gnb
Create a first BN : bn
In [3]:
bn=gum.loadBN("res/asia.bif")
# randomly re-generate parameters for every Conditional Probability Table
bn.generateCPTs()
bn
Out[3]:
Create a second BN : bn2
In [4]:
bn2=gum.loadBN("res/asia.bif")
bn2.generateCPTs()
bn2
Out[4]:
bn vs bn2 : different parameters
In [5]:
gnb.flow.row(bn.cpt(3),bn2.cpt(3),
captions=["a CPT in bn","same CPT in bn2 (with different parameters)"])
|
| |
---|---|---|
0.1046 | 0.8954 | |
0.2958 | 0.7042 |
|
| |
---|---|---|
0.6247 | 0.3753 | |
0.3076 | 0.6924 |
Exact and (Gibbs) approximated KL-divergence
In order to compute KL-divergence, we just need to be sure that the 2 distributions are defined on the same domain (same variables, etc.)
Exact KL
In [6]:
g1=gum.ExactBNdistance(bn,bn2)
print(g1.compute())
{'klPQ': 2.693885010280355, 'errorPQ': 0, 'klQP': 5.322118996448689, 'errorQP': 0, 'hellinger': 0.9795669288987562, 'bhattacharya': 0.6534951836845349, 'jensen-shannon': 0.5757571920278003}
If the models are not on the same domain :
In [7]:
bn_different_domain=gum.loadBN("res/alarm.dsl")
# g=gum.BruteForceKL(bn,bn_different_domain) # a KL-divergence between asia and alarm ... :(
#
# would cause
#---------------------------------------------------------------------------
#OperationNotAllowed Traceback (most recent call last)
#
#OperationNotAllowed: this operation is not allowed : KL : the 2 BNs are not compatible (not the same vars : visit_to_Asia?)
Gibbs-approximated KL
In [8]:
g=gum.GibbsBNdistance(bn,bn2)
g.setVerbosity(True)
g.setMaxTime(120)
g.setBurnIn(5000)
g.setEpsilon(1e-7)
g.setPeriodSize(500)
In [9]:
print(g.compute())
print("Computed in {0} s".format(g.currentTime()))
{'klPQ': 2.693569456053914, 'errorPQ': 0, 'klQP': 4.170450481408639, 'errorQP': 0, 'hellinger': 0.9354607479362501, 'bhattacharya': 0.6557049176527434, 'jensen-shannon': 0.532485604630078}
Computed in 6.798801791 s
In [10]:
print("--")
print(g.messageApproximationScheme())
print("--")
print("Temps de calcul : {0}".format(g.currentTime()))
print("Nombre d'itérations : {0}".format(g.nbrIterations()))
--
stopped with epsilon=1e-07
--
Temps de calcul : 6.798801791
Nombre d'itérations : 1284500
In [11]:
p=plot(g.history(), 'g')
Animation of Gibbs KL
Since it may be difficult to know what happens during approximation algorithm, pyAgrum allows to follow the iteration using animated matplotlib figure
In [15]:
g=gum.GibbsBNdistance(bn,bn2)
g.setMaxTime(60)
g.setBurnIn(500)
g.setEpsilon(1e-7)
g.setPeriodSize(5000)
In [16]:
gnb.animApproximationScheme(g) # logarithmique scale for Y
g.compute()
Out[16]:
{'klPQ': 2.6980953717263163,
'errorPQ': 0,
'klQP': 4.294954692336129,
'errorQP': 0,
'hellinger': 0.9421255936192727,
'bhattacharya': 0.6566737627926815,
'jensen-shannon': 0.5390719372945214}
In [ ]: