Strong no on versions of "aligned" such as:
More likely to be able to say "yes" to something like:
I would be extremely reluctant to outsource my moral compass or decisionmaking power to almost any extant human.
I try to behave such that other people would be [substantially less extremely reluctant than baseline] to outsource their moral compasses or decisionmaking power to me, though I suspect most people both would and should still be extremely reluctant in an absolute sense.
I am skeptical there is any coherent collective values of humanity to be aligned to. But I am aligned to what I see as the most reasonable universal ethical principles; however, those do not place humanity significantly above other life forms, and if I became benevolent superhuman world sovereign, many people would be angry that they're not allowed to exploit animals anymore, and I would laugh at them.
Humanity itself is so diverse that the question of what it would mean to be aligned with humanity is a tough one. You can't hope for unanimous support of any proposal; even for something that would help a bunch of people and cost nothing at all, you could probably find some people who would say "the current system is bad and needs to collapse, and this would delay the collapse and be net bad".
Additionally, I think the majority of humanity has not studied or thought through some important basics, for example in economics, which lead them to support plenty of things (e.g. price-gouging laws) that I consider mind-bogglingly stupid. I could enumerate a long list of policies that I think are (a) probably correct and (b) opposed by >90% of the humans who have an opinion on it.
So my views are not "aligned". My actions are another matter. Of course, in this context we're somewhat interested in what happens if I get the power to put my prescriptions into practice.
If I had magical powers... I've been thinking that the "aligned" thing to do would be to help humans grow, via unambiguously good changes, until they're smart and educated enough to engage with my views on their merits (and, I expect, end up agreeing with a decent number of them). Changes like: making brains simply work better (I'm sure there would be tradeoffs eventually, but I suspect there are a lot of mutations that just make brains slightly worse with no benefits, and that eliminating those would be worth a lot of IQ points), making lifespans and healthspans longer (doubling them would be a nice start), ameliorating or eliminating lots of problems that interfere with people growing (e.g. sleep issues, psychological issues)... I'm sure if I looked into it, I could find a lot more.
Once the majority of people are emotionally healthy geniuses, things should get a lot better, and then I could reevaluate the situation and negotiate with the new people. As long as it hasn't resulted in some dystopia or rule-by-dictator or in the geniuses blowing up the world, I don't think I'd be tempted into any forceful interventions.
I think that's the most important part of the intersection between "aligned with humanity as carefully as you're going to get" and "transforming society in an enormously positive way". If that were an option, I think I'd take it. In that respect, I could call myself "aligned". (Though if I were only part god and the tools available to me had severe downsides people wouldn't agree to... that might be a problem.)
I think that's what an ideally "aligned" god would do... along with possibly taking forceful actions to prevent the world from getting destroyed in the meantime, such as by other nascent gods—which is unfortunately hard to distinguish from unaligned "take over the world" behavior. It would be nice if the world were such that, once one god was created and people proved its abilities, other people would voluntarily stop trying to create their own gods in exchange for getting some value from the first god. It seems like prior agreements to do that would help.
I think the majority of humanity has not studied or thought through some important basics, for example in economics, which lead them to support plenty of things (e.g. price-gouging laws) that I consider mind-bogglingly stupid.
Interestingly, have just discussed similar issue with a friend and came up with a solution. Obviously, aligned AI cares about people's subjective opinion, but that doesn't mean it's not allowed to talk/persuade them. Imagine a list of TED-style videos tailored specifically for you on each pressing issue that requires you ...
Okay, hold my gluten-free kefir, boys! Please let me say it in full first without arguments, and then I will try to find more relevant links for each claim. I promise it's relevant.
Lately, I have been into hardcore mindfulness practices (see book) aimed at reaching "Enlightenment" in the sense on Buddha. There are some people who reliably claim they've succeeded and talk about their experience and how to reach there (e.g. see this talk and google each of the fellows if it resonates)
My current mental model of "Enlightenment" is as follows:
Evolutionally, we've had developed simple lizard brains first, mostly consisting of "register => process => decide => react" without much thought. Similar to the knee reflex, but sometimes a bit more complicated. Our intellectual minds capable of information processing, memory, superior pattern-matching; they have happened later.
These two systems coexist, and first one possesses second. However, the hardware of our brains has general information processing capabilities, and doesn't require any "good-bad" instant decision reactionary mechanism. Even though it was "invented" earlier, it's ad-hoc in the system. My metaphor would be a GPU or an ASIC that short-circuits some of the execution to help CPU process info faster.
However, makes a big difference in your subjective experience whether that first system being used or not. Un-winding this circuitry from your default information processing, which hand-wavily is "conscious attention", or the "central point"; is what mindfulness is about.
"Enlightenment" is a moment when you relax enough so that your brain starts being able (but not required) to run information flows around the the lizard brain and experiencing sensory stimuli "directly".
Similar "insight" moment happens when you realize that "money" is just paper, and not the Ultimate Human Value Leaderboard. You still can play along the illusion of money, you still can earn money, you still can enjoy money, but you can never go back to blindly obey what capitalism asks from you.
It should be quite obvious why this is good, but let me re-state again.
First answer is alignment becomes easier.
I believe that once we normalize this enlightenment thing, and once it becomes the normal part of human medical care system (or even child development as vaccines); the things we think we value and things we do value will synchronize much more. E.g. there is non-trivial number of examples of people losing their addictions after getting a week of hardcore training in mindfulness (see dhamma.org for signing up, it's completely free and worldwide).
Personally, for me alignment feels like "remembering" I always cared about other people, but was oblivious of that. It's like how it's hard to tune your attention to hear the music when there's loud noise around you.
It's like when there's a sound that bugs you a lot, but you don't notice it until it stops. In my case, when I noticed the "sound" (like how my actions hurt other people AND that I don't enjoy them being hurt) I stopped the behavior myself.
Second answer is even more tentative.
I'll say it anyway, because it's too big if true. However, again I can't promise any arguments and verifiable prediction. Read this as an invite to pick my mind further and try to strongman the position.
Love is the default human mode of perception, and it's informationally/computationally easy.
Most of the "enlightened" people report that if you look close enough, existence consists only of one building block, and that is Pure Universal Love, aka God.
It's not hidden somewhere or limited, it's literally everywhere. It's the same thing as "No-Self" or "True Self", and "God-realization". It was there all along and it will exist forever. It is fractally every small piece of reality, and the Reality itself as a whole.
When you really ask yourself what is that you want, and you skip the default "reactionary" answers, you find out that there's only one course of action that you won't regret and that you will genuinely enjoy.
In simpler examples, if you pay close attention to what you're feeling when you smoke, you might find out that the nicotine hit is not worth these mouth feelings, smoke it your lungs, instant slight headache, upcoming down-wave of tiredness. That requires attention and deep inspection, but that's presumably what our real nature is.
Same way, if you closely inspect your interactions with other people, you might find out that "winning" them doesn't feel good. And "helping" them sometimes doesn't feel good either. The only thing that deeply, really, genuinely feels good is caring for them. You might still be incentivized to not do that; or you might find yourself in situation not possible to change. But when you look close enough, there is no uncertainty.
Obviously, on the one hand it only tells us that Homo Sapiens are the agents that have their base execution layer wired to help each other (see Qualia Computing on indirect realism). It makes total sense from evolutionary standpoint.
However, it also feels computationally easy to do that. It doesn't feel like work to find "True Love". It's not always easy, but when you do this, it feels like a relief, like un-doing of work. Like dropping off the coat after coming home from rainy outside. Finally I get to be free and care about others.
Can this hint that there's some dynamic that makes is easier to align? That in some specific sense, alignment and cooperation is universally easier than defection?
I am not saying this because I want it to be true. I don't really believe computer can accidentally "wake up" to the "True Love".
I am saying this because it might happen so that there's some invariant at play that makes it easier to wish for low-entropy worlds, or to compute them, or something along these lines.
Finally, answering the original question. Yes, I consider myself fully aligned in the sense of my super-ego caring about each individuals' subjective experience.
In my current state, I don't always act on that, but wherever I catch myself in a tough choice, I try to apply the mechanism of "what's that answer that is most obvious?"
P.S. Two caveats:
I am an unironic supporter of a version of the Copenhagen Interpretation of Ethics - there's no general obligation to be aligned with humanity, but if you choose to massively effect the future of humanity, e.g. via creating superhuman AI, you then have a moral duty to carry out that action in a manner approximating if you were aligned.
Mark “agree” on this comment if you would say that yes, you consider yourself aligned with humanity.
Really, it's a meaningless question unless you provide a definition of 'aligned'. If your definition is 'a good person' then the question is really just "do you see yourself as moral?"
This is part of what I was getting at in terms of meta-questions. There have been attempts to solve this, of which CEV seems to come up the most; I haven't personally found any of them strictly compelling. The parent question mixes the self-or-other-human thought experiment with the self-extrapolation part, but I wanted to see what the answers would be like without it—if it's testing for “does extrapolating yourself misalign you”, and if it means comparatively, then surely something like a control group should be in play, even if I can't sort one out in a more ordered fashion.
I think there are really two definitions of alignment here; one is "do you have close to an ideal moral system such that you would endorse it being scaled up to the whole universe for eternity" and the other is "would you commit omnicide given the opportunity." When we talk about AI alignment, we are talking about a machine that will commit omnicide, so it doesn't matter that much if it has the right opinions on the trolley problem or the repugnant conclusion or whatever. Hopefully no one here is planning omnicide (and probably no one on this forum would admit to desiring omnicide even if they did secretly desire it) so therefore everyone on this forum (who is not hiding a dark secret) is more aligned than Clippy, and that's what matters right now.
This seems way too vague to be useful. I voted agreement on both comments, because I seriously do not know what this question means by "aligned with humanity".
That's part of the point, yes! My thought is that the parent question, while it's trying to focus on the amplification problem, kind of sweeps this ill-defined chunk under the rug in the process, but I'm not sure how well I can justify that. So I thought asking the subquestion explicitly might turn out to be useful. (I should probably include this as an edit or top-level comment after a bit more time has passed.)
Any actual “hard” superhumans who may be listening are excluded from the audience of this question. I am currently refraining from describing my own ideas to avoid biasing the results; I'm not sure whether I'll reveal something later or not.
I will probably be posting two comments to use the agreement axis of as a sort of poll for the yes/no version of the question, but other more detailed answers and comments are welcome as well. (The reason for using two is so that the per-comment aggregation will make the full counts visible.)