MyJournals Home  

RSS FeedsEntropy, Vol. 19, Pages 286: Assessing Probabilistic Inference by Comparing the Generalized Mean of the Model and Source Probabilities (Entropy)

 
 

19 june 2017 15:40:22

 
Entropy, Vol. 19, Pages 286: Assessing Probabilistic Inference by Comparing the Generalized Mean of the Model and Source Probabilities (Entropy)
 


An approach to the assessment of probabilistic inference is described which quantifies the performance on the probability scale. From both information and Bayesian theory, the central tendency of an inference is proven to be the geometric mean of the probabilities reported for the actual outcome and is referred to as the `Accuracy`. Upper and lower error bars on the accuracy are provided by the arithmetic mean and the -2/3 mean. The arithmetic is called the `Decisiveness` due to its similarity with the cost of a decision and the -2/3 mean is called the `Robustness`, due to its sensitivity to outlier errors. Visualization of inference performance is facilitated by plotting the reported model probabilities versus the histogram calculated source probabilities. The visualization of the calibration between model and source is summarized on both axes by the arithmetic, geometric, and -2/3 means. From information theory, the performance of the inference is related to the cross-entropy between the model and source distribution. Just as cross-entropy is the sum of the entropy and the divergence; the accuracy of a model can be decomposed into a component due to the source uncertainty and the divergence between the source and model. Translated to the probability domain these quantities are plotted as the average model probability versus the average source probability. The divergence probability is the average model probability divided by the average source probability. When an inference is over/under-confident, the arithmetic mean of the model increases/decreases, while the -2/3 mean decreases/increases, respectively.


 
125 viewsCategory: Informatics, Physics
 
Entropy, Vol. 19, Pages 265: Modeling Multi-Event Non-Point Source Pollution in a Data-Scarce Catchment Using ANN and Entropy Analysis (Entropy)
Entropy, Vol. 19, Pages 288: Inconsistency of Template Estimation by Minimizing of the Variance/Pre-Variance in the Quotient Space (Entropy)
 
 
blog comments powered by Disqus


MyJournals.org
The latest issues of all your favorite science journals on one page

Username:
Password:

Register | Retrieve

Search:

Physics


Copyright © 2008 - 2024 Indigonet Services B.V.. Contact: Tim Hulsen. Read here our privacy notice.
Other websites of Indigonet Services B.V.: Nieuws Vacatures News Tweets Nachrichten