A big question that determines a lot about what risks from AGI/ASI may look like has to do with the kind of things that our universe's laws allow to exist. There is an intuitive sense in which these laws, involving certain symmetries as well as the inherent smoothing out caused by statistics over large ensembles and thus thermodynamics, etc., allow only certain kinds of things to exist and work reliably. For example, we know "rocket that travels to the Moon" is definitely possible. "Gene therapy that allows a human to live and be youthful until the age of 300" or "superintelligent AGI" are probably possible, though we don't know how hard. "Odourless ambient temperature and pressure gas that kills everyone who breathes it if and only if their name is Mark with 100% accuracy" probably is not. Are there known attempts at systematising this issue using algorithmic complexity, placing theoretical and computational bounds, and so on so forth?
Like you said the science we don't know is at inaccessibly large or small scales.
Yes maybe in the far future in a society spread across multiple galaxies, or that can make things near Planck lengths, they could do something that would totally stump us.
But your never going to find a final answer to this in the present day for exactly those reasons.
In fact it's unlikely anyone on LW could even grasp the answers even if by some miracle a helpful time traveller from the future showed up and started answering.