Joint entropy is a measure of "the uncertainty" associated with a set of variables.
In order to calculate the joint entropy, you should enter the joint distribution matrix where the cell value for any i row and j column represents the probability of the outcome, . Joint entropy formula can be found below the calculator.
Joint Entropy Formula
The joint Shannon entropy (in bits) of two discrete random variables and with images and is defined as:
where and are particular values of and , respectively, is the joint probability of these values occurring together, and is defined to be 0 if 1