use cases:
- Use this formula to determine how “sure” a model is of the output. If the entropy is high, then the model is probably predicting an unknown class
- You can also use this for data cleaning. if there is abnormal amounts of entropy, maybe it’s corrupted.
- where each pi is the probability of outcome i (aka class prediction) happening
- the negative sign is because the logarithm of a number between 0 and 1 is negative
what makes a higher entropy
- Uniform distribution
- cause if you’re equally sure about all outcomes, you’re not sure of one
- Large Number of possible Outcomes
- even if they don’t have near probability
- cause the more potential outcomes there are, the more unpredictable the event is.