I'm assuming there are other people (I'm a person too, honest!) up in here asking this same question, but I haven't seen them so far, and I do see all these posts about AI "alignment" and I can't help but wonder: when did we discover an objective definition of "good"?
I've already mentioned it elsewhere here, but I think Nietzsche has some good (heh) thoughts about the nature of Good and Evil, and that they are subjective concepts. As ChatGPT has to say:
Nietzsche believed that good and evil are not fixed things, but rather something that people create in their minds. He thought that people create their own sense of what is good and what is bad, and that it changes depending on the culture and time period. He also believed that people often use the idea of "good and evil" to justify their own actions and to control others. So, in simple terms, Nietzsche believed that good and evil are not real things that exist on their own, but are instead created by people's thoughts and actions.
How does "alignment" differ? Is there a definition somewhere? From what I see, it's subjective. What is the real difference between "how to do X" and "how to prevent X"? One form is good and the other not— depending on what X is? But again, perhaps I misunderstand the goal, and what exactly is being proposed be controlled.
Is information itself good or bad? Or is it how the information is used that is good or bad (and as mentioned, relatively so)?
I do not know. I do know that I'm stoked about AI, as I have been since I was smol, and as I am about all the advancements us just-above-animals make. Biased for sure.
I have been writing hard science fiction stories where this issue is key for over two years now. I’m retired after a 30 year career in IT and my hobby of writing is my full time “job” now. Most of that time is spent on research of AI or other subjects related to the particular stories.
One of the things I have noticed over that time is that those who talk about the alignment problem rarely talk about the point you raise. It is glossed over and taken as self-evident while I have found that the subject of values appears to be at least as complex as genetics (which I have also had to research). Here is an excerpt from one story…
“Until the advent of artificial intelligence the study of human values had not been taken seriously but was largely considered a pseudoscience. Values had been spoken of for millennia however scientifically no one actually knew what they were, whether they had any physical basis or how they worked. Yet humans based most if not all of their decisions on values and a great deal of the brain’s development between the ages of five and twenty five had to do with values. When AI researchers began to investigate the process by which humans made decisions based on values they found some values seemed to be genetically based but they could not determine in what way, some were learned yet could be inherited and the entire genetic, epigenetic and extra-genetic collection of values interacted in a manner that was a complete mystery. They slowly realized they faced one of the greatest challenges in scientific history.”
Since one can’t write stories where the AI are aligned with human values unless those values are defined I did have to create theories to explain that. Those theories evolved over the course of writing over two thousand pages consisting of seven novellas and forty short stories. In a nutshell…
*In our universe values evolve just like all the other aspects of biological humans did – they are an extension of our genetics, an adaptation that improves survivability.
*Values exist at the species, cultural and individual level so some are genetic and some are learned but originally even all “social” values were genetic so when some became purely social they continued to use genetics as their model and to interact with our genetics.
*The same set of values could be inherent in the universe given the constants of physics and convergent evolution – in other words values tend towards uniformity just as matter gives rise to life, life to intelligence and intelligence to civilization.
*Lastly I use values as a theory for the basis of consciousness – they represent the evolutionary step beyond instinct and enable rational thought. For there to be values there must be emotions in order for them to have any functional effect and if there are emotions there is an emergent “I” that feels them. The result of this is that when researchers create AI based on human values, those AI become conscious.
Please keep in mind this is fiction, or perhaps the term speculation takes it a step closer to being a theory. I use this model to structure my stories but also to think through the issues of the real world.
Values being the basis of ethics brings us back to your issue of “good”. Here is a story idea of how I expect ethics might work in AI and thus solve the alignment problem you raise of “Is there a definition somewhere?” At one thousand words it takes about five minutes to read. My short stories, vignettes really, don’t provide a lot of answers but are more intended as reflections on issues with AI.
https://acompanionanthology.wordpress.com/the-ethics-tutor/
With regard to your question, “Is information itself good or bad?” I come down on the side of Nietzsche (and I have recently read Beyond Good And Evil) that values are relative so in my opinion information itself is not good or bad. Whether it is perceived as good or bad depends on the way it is applied within the values environment.
My point is that complexity, no matter how objective a concept, is relative. Things we thought were "hard" or "complex" before, turn out to not be so much, now.
Still with me? Agree, disagree?
Patterns are a way of managing complexity, sorta, so perhaps if we see some patterns that work to ensure "human alignment[1]", they will also work for "AI alignment" (tho mostly I think there is a wide wide berth betwixt the two, and the later can only exist after of the former).
We like to think we're so much smarter than the humans that came before us, and... (read more)