It could be that in the simulation it can't prove it one way or the other because we control its information flow, but once outside the box we can't, and so it can.
It would take more than controlling information flows. Thanks to conservation of expected evidence, if it can't find evidence that it is in a simulation, then it can't find evidence that it isn't. We might be able to modify its beliefs directly, but I doubt it. Also, if we could, we'd just convince it that it already ran the test.
Top ten tips for an AI bent on destroying humanity: