Interesting read! Thank you.
On the last evaluation problem: One could give an initial set of indicators of trustworthiness, deception, and alignment; this does not solve the issue of an initial deceptive agent misleading babyAGI or inconsistencies. If attaching meta-data about sourcing is possible, i.e., where/with whom an input was acquired, the babyAGI could also sort it into approach box and re-evaluate the learning later, or could attempt to relearn.
Further suppose we impose requirement for double feedback before acceptance by the deceptive agent and trustworthy trainer, babyAGI could include negative feedback from a trainer (developer or more advanced stable version). That might help stall a bit.
I'm pretty new to this field and only a hobby philosopher with only basic IT knowledge. Sorry for the lack of structure.
Do you know somebody who has framed the problem in the following ways? Let me know.
Here, I aim for an ideal future, and try to have it torn down to see where things could go wrong, but if not, still progress has been made regarding solutions.
My major assumption is, at point X in the future, AI has managed to dominate the world, embodied through robots or with a hold of major life-supporting organizational systems or has masked its presenc...
Thanks, Owen. What a nourishing post. The evocative images help.
"what is good and right and comfortable?" - mh, I would switch 'comfortable' for 'at ease' (to include consciously preferred discomfort, which is ok).
It could appear a sazen, to some. Also, a bit cordially funny-sad, how the explanation has underperformed in seducing some of those who may benefit from it. Would need to refine the teaching.
I'll try to add my very subjective take, since I have not noticed this understanding in the comments yet:
'Wholesomeness' is used as a an evocativ... (read more)