Im no math-viz. When reading about entropy I understand that it has something to do with the information contained.
High entropy means more randomness.
Low entropy means more compact structures/patterns etc?
Then I heard about the Kolmogorov complexity, the Shannon Entropy etc...
I viewed also a video explaining entropy. But I wasnt quite sure if I understood if one has a mathematical way of finding the low entropy of a chunk of information.
Say I have a file of 256 bytes. The data is computer opcodes, so its some patterns, but it looks random, but its not very random.
Is it possible to find the low-entropy value of these bytes (or 2048 bits)? meaning it contains all the information but the entropy-value tells how much it could have been compressed?
What is this mathematical formula called? Shannon or? (since I am no math-viz I was wondering if anyone here could give me a hint or direction, or if it is no such formula).