That 30 bits was not literally intended to represent only the object's noun category, but even if it did, none of the additional pieces of information you would like to add are going to change this picture much, because what one would think of as a "high amount of detail" is not actually all that high in terms of the logarithmic growth of the entropy.
Take color: suppose the average person has 16 baseline colors memorized, and then a few variations of each: each one can be bright or dark, saturated or pastel. That would be about 6 bits for color. If you have an eye for color or you're an artist you may have some additional degrees of freedom. Hell, a computer using RGB can only represent 24 bits worth of color, maximum. I am going to suggest this stuff gets cognized less than 10 bits worth for the average person; let's just say 10.
Now, of course, people can memorize more than one color. If colors are independently distributed uniformly at random, then processing N colors requires 10N bits. But of course they aren't, so the entropy is less. But again, let's just say they were. So how many color combinations can you process per second? I would say it's a bit of a challenge to memorize a set of 10 arbitrary drawn colors shown for a second. Most people couldn't continuously do that at a rate of 10 colors per second. That would be 100 bits/sec of info.
The point is that you really don't perceive all that much. You show the average person a Rubik's cube, there is no way they're going to remember the exact pattern of colors that they saw, unless the cube were solved or something. They will perceive it as "multicolored" and that's about it.
Adding behavior, texture, etc doesn't change this picture. None of this stuff is even close to 10^9 bits of entropy, which would be 2^1,000,000,000 different equally likely possibilities.
Take color: suppose the average person has 16 baseline colors memorized, and then a few variations of each: each one can be bright or dark, saturated or pastel. That would be about 6 bits for color. If you have an eye for color or you're an artist you may have some additional degrees of freedom. Hell, a computer using RGB can only represent 24 bits worth of color, maximum. I am going to suggest this stuff gets cognized less than 10 bits worth for the average person; let's just say 10.
Now, of course, people can memorize more than one color. If colors are independently distributed uniformly at random, then processing N colors requires 10N bits. But of course they aren't, so the entropy is less. But again, let's just say they were. So how many color combinations can you process per second? I would say it's a bit of a challenge to memorize a set of 10 arbitrary drawn colors shown for a second. Most people couldn't continuously do that at a rate of 10 colors per second. That would be 100 bits/sec of info.
The point is that you really don't perceive all that much. You show the average person a Rubik's cube, there is no way they're going to remember the exact pattern of colors that they saw, unless the cube were solved or something. They will perceive it as "multicolored" and that's about it.
Adding behavior, texture, etc doesn't change this picture. None of this stuff is even close to 10^9 bits of entropy, which would be 2^1,000,000,000 different equally likely possibilities.