I have a graph-based solution for decision-making in situations with multiple agents and stakeholders. I wonder if there's an elegant way to express it as a Mathematical formula. Background: I have oriented myself on the "true justified beliefs" framework from epistemology. Here, propositions are categorized into beliefs, and those beliefs...
I've noticed that there are 2 flavors of the alignment problem. One is about the technical how to and practical engineering, the other is about humanities, social sciences, human behavior and psychology. What they both have in common is doomsday scenarios about paperclip maximizers. I don't know about you, but...
The Effective Altruism movement seems to be mainly concerned with 2 broad objectives: One is a focus on reducing suffering and alleviating risks, the other is a focus on doing the most good and improving the world. A big concern is AI alignment, as it potentially poses an existential risk....
One of the main obstacles to building safe and aligned AI is that we don't know how to store human values on a computer. Why is that? Human values are abstract feelings and intuitions that can be described with words. For example: * Freedom or Liberty * Happiness or Welfare...