| Class | Description | 
|---|---|
| Entropy | 
 Calculates entropy, joint entropy or conditional entropy, which can be used to determine the uncertainty in the states of a discrete distribution. 
 | 
| JensenShannon | 
 Methods for computing the Jensen Shannon divergence, which measures the similarity between probability distributions. 
 | 
| KullbackLeibler | 
 Calculate the Kullback–Leibler divergence between 2 distributions with the same variables, D(P||Q). 
 | 
| MutualInformation | 
 Calculates mutual information or conditional mutual information, which measures the dependence between two variables. 
 | 
| Enum | Description | 
|---|---|
| LogarithmBase | 
 Determines the base of the logarithm to use during calculations such as mutual information. 
 | 
Copyright © 2021. All rights reserved.