I am currently training an agent to explore a platforming level. From my understanding, during training the entropy value is used as a measure of uncertainty to recognize when surprising events occur and thus reward curiosity. Is there a way to get a similar measure, but during inference? If the reason for wanting this is relevant, I wish to use it so that it can notify me when something unexpected to it happens (e.g. falling through the floor, as I assume falling through the supposedly safe floor would be surprising to the trained agent). Thanks in advance!