Normalized Entropy

Normalized Entropy

Normalized Entropy is a metric that takes into account the maximum possible value of Entropy and returns a normalized measure of the uncertainty associated with the variable:

HN(X)=H(X)logā”2(SX){H_N}(X) = \frac{{H(X)}}{{{{\log }_2}({S_X})}}

Example

In this new example, we now compare the variables X1 and X2, which each represent ball colors:

  • X1 āˆˆ {blue, red}

  • X2 āˆˆ {blue, red, green, yellow, purple, orange, brown, black}

Normalized Entropy allows us to compare the degree of uncertainty even though these two variables have different numbers of states, i.e., two versus eight states:

X1X2

Usage

In BayesiaLab, the values of Entropy and Normalized Entropy can be accessed in a number of ways:

  • In Validation Mode , with the Information Mode activated, hovering over a Monitor with your cursor will bring up a Tooltip that includes Entropy and Normalized Entropy.

  • You can also sort the Monitors in the Monitor Panel according to their Normalized Entropy via Monitor Context Menu > Sort > Normalized Entropy.

  • The Normalized Entropy is also available as a Node Analysis metric for Size and Color in the 2D and 3D Mapping Tools.

  • In Function Nodes, Entropy and Normalized Entropy are available as Inference Functions in the Equation tab.

    • Entropy: Entropy(?X1?, False)

    • Normalized Entropy: Entropy(?X1?, True)

Demo Network

NormalizedEntropy.xbl

Last updated

Logo

Bayesia USA

info@bayesia.us

Bayesia S.A.S.

info@bayesia.com

Bayesia Singapore

info@bayesia.com.sg

Copyright Ā© 2024 Bayesia S.A.S., Bayesia USA, LLC, and Bayesia Singapore Pte. Ltd. All Rights Reserved.