Is there a benefit in low capability AI Alignment research?
Basic Problem If I read about current AI Alignment research ideas, I see a lot of discussion about high capability scenarios, in which AI would be too smart to align simply. It would deceive operators, hack through communication channels and goodheart it's way over to infinity. However if I block...