I've read that imagination (in the sense of conjuring mental imagery) is a spectrum, and I've encountered a test which some but not all phantasic people fail.
I don't recall the details enough to pose it directly, but I think I do recall enough to reinvent the test:
Test details guessed above may not properly recreate the ability to distinguish levels of imagery. My hazy memory says the words might be top-to-bottom? Or the order of providing the letters might matter?
Someone actually seeing the image you've requested they construct would be able to trivially read off three words. ...but someone without mental imagery or with insufficient mental imagery may fail.
I recall discovering that I really can't imagine more than about 2 letters at a time before adding additional detail to my mental visual workspace forces the loss of something else. That seems pretty poor, and tracks with my inability to imagine human faces -- my theory is that a specific face requires more details to distinguish it from other faces than the maximum amount of detail I can visualize.
Actually, having written this, it just now occurs to me that my cached thought may be incorrect, that all my other qualia processing is "normal".
...I routinely (but not always) fail to perceive any qualia for hunger or smells (this predates COVID) -- yet, curiously, in the case of smells I somehow know (without any experience of perception) that there is a smell that I ought to be experiencing, and its rough intensity.
In the case of hunger, I'll literally fail to know I need to eat. I'll get the shakes and collapse and wonder why. I've needed to establish a habit of scheduled eating, to avoid this occurrence.
Previously, I had grouped these defects in with my inability to know my own wants -- in my theory: trauma damage that severed the connection to certain mental modules -- but it now occurs to me that an alternative hypothesis exists: that there's a possible connection to my unusual visual qualia processing.
I see a couple leads to investigate, which could help shed additional light on the topic. One is common enough to have a name: synthesthesia. The other, I think, may be personally unique to me, or at least some combination of very-rare and never-discussed that I've never heard of it.
Synthesthesia, to my understanding, involves multiple qualia accompanying various experiences, notably including qualia native to a different sensory modality. E.g. "That sound was green." Exploring the causal chain resulting in such utterances seems likely to turn up insights into qualia which will be more broadly applicable.
As for my unusual qualia processing: I am measurably red-green colorblind; in a laboratory setting, clean of context clues, I guess no better than chance whether a color is red or green, although I can reliably tell X and Y are both red or green and they're the opposite color from each other. Yet in everyday life, I experience qualia for red and green, almost always "correctly" (in that the qualium I'll call "Green" I experience when seeing actual green objects well in excess of 99% of the time, and vice versa for "Red" and red objects.)
My current theory as to how this works:
I think the existence of my defect may shed some light on the working of normal qualia.
That is, I think there's a module which makes educated guesses about certain true properties of the world, based on the sensory stream, and annotates the sense information with its guesses, before that sense information reaches awareness. These annotations either become or select "qualia", the inexplicable ineffable differences in experience correlating to (or encoding) actual sense data.
Further, I think that investigating the causal chain resulting in my unusual experience might allow us to localize the qualia-annotation process in my brain, and perhaps find a standard location in many brains.
I suggest a simple explanation: some of us have qualia and some of us don’t.
Well that's an alarming hypothesis.
I've seen expressed (and held the view personally) that a world devoid of qualia is an example of a world devoid of value, in the consequentialist sense. ...but at least in my case, my view was somewhat grounded in the idea that all people are morally significant combined with the implicit assumption that the overwhelming majority of adults experience qualia, so updating to a higher probability of "a significant fraction of people alive today do not experience anything like qualia" ought to also come with an update away from "non-qualia-experiencing agents lack moral value".
I worry that some people may hold my prior view uncritically, and see an admission of not experiencing qualia as a moral license to disregard the person's well-being. See various historical takes about "X minority doesn't have souls" and the resultant treatment.
I can see some sense in this take; I've personally succumbed to predatory gambling services in years past, and can attest from personal experience that successfully quitting one addiction leaves me exceedingly vulnerable to pick up a different one. I rotated through six different highly-damaging vices, before settling into a relatively less-harmful vice, and I was grateful to find myself there.
And that's my point: you are in expectation doing someone who is inclined to addiction a favor by forcing them off of a particularly bad addiction.
You would be doing them a slightly bigger favor if you provide a much less-harmful replacement addiction at the same time. For me, the holy grail was to find a net-positive activity to slot into the addiction-shaped hole in my brain, and after a lifetime of struggle, I finally got "reading pop nonfiction books" to replace "scrolling social media", the first real Good in a long line of Lesser Evils.
Dramatically increasing taxes for childless people.
Too-low fertility concerns me deeply. My current preferred strategy had been something financial along the lines of this proposal, but on reflection, I think I need to update.
The main reasons that I see driving the people around me to defer/eschew children are, in rough decreasing order of prevalence:
The third point above is, in my social circle, usually downstream of needing to spend too much time working, not being able to afford childcare services, not being able to afford college, etc.
These issues could be lessened by financial interventions with the net effect of offsetting the burdens of child raising (obviously, implementation details matter a great deal.)
...I had previously assumed that "expectation of incompetence" was the primary issue because it's my primary issue (in the sense of looming large in my mind). ...but having taken an inventory of the childless adults in my social network, I now see that "inability to find a suitable partner" and "infertility" are much bigger issues.
"Inability to find a suitable partner" seems risky to fix, because it looks easy to create horrible side-effects. Many in my social network are in relationships but aren't having children because they're abusive relationships. Others have successfully escaped abusive relationships and subsequently given up on finding a partner; being alone is less painful, and they expect to find another abuser with nontrivial probability if they go looking for another partner. Still others lost a suitable partner to a treatable disease, due primarily to insurance companies ending up de facto in charge of healthcare decisions, and rendering decisions without the patient's health as the priority.
I don't see a good solution. (Well, except for "don't allow insurers to drive healthcare decisions".) All attempts to "solve" this issue I see being promoted in real life seem on-net harmful. My home state of Texas thinks the best solution is to ban divorce, and we expect that to be implemented within the next year. I don't see that as being a net good, even if it may compel additional births in the short term on net. Trapping people in abusive relationships seems incredibly dystopian. I also foresee this making young women even more afraid of getting into the wrong relationship than they already are, which runs the risk of making the "finding suitable partners" problem worse in the long run.
"Lack of fertility" seems challenging to fix, but far less fraught. I was born without a uterus, and wish to bear a child if the medical technology were developed to allow me to grow a uterus. I have a sibling on life-saving medication, a side-effect of which is extremely low sperm count -- a better drug to treat the condition (or an actual cure) could resolve this issue. Multiple of my sisters-in-law have simply failed to conceive for years, due to unknown-to-me causes, and I suspect the issues are similarly fixable via continued medical innovation.
Question: is it actually possible to cryptographically obfuscate goals/plans/precomittments (or other elements of "deception" as mentioned in your second bullet point) against something inspecting your algorithm and working memory?
In particular, it seems to me that anything properly able to inspect you can, at a minimum, discover the use of encryption, and find where the encryption key is read from, and then read the encrypted content. Or fail the model once encryption is detected at all.
I'm not saying it will be easy to build such transparency tools -- in fact, that must be a herculean challenge, since selecting against interpreted thoughts selects for uninterpretable thoughts, and a superintelligence might use principles not yet known to humanity -- I just mean that, from the perspective of the AI itself, I think perfect obfuscation against ideal logically-possible interpretability tools might be impossible.
Edit:
Nevermind, I answered my own question. There is no such logically-possible interpretability tool. For any such tool we logically-counterfactually assume could exist, then the AI could implement that tool, and utilize encryption if and only if that tool asserts that the AI doesn't use encryption. ...therefore no algorithm exists which can verify the use of encryption in the general case.
This argument has roughly the same shape as my reasoning regarding why prediction markets are likely to have much worse predictive power than one would naively guess, conditional on anyone using the outputs of a prediction market for decisions of significance: individual bettors are likely to care about the significant outcomes of the prediction. This outcome-driven prediction drive need not outweigh the profit/accuracy-driven component of the prediction market -- though it might -- in order to alter the prediction rendered enough to alter the relevant significant decision.
Perhaps the prediction market concept can be rescued from this failure mode via some analogue of the concept of financial leverage? That is, for predictions which will be used for significant decision purposes, some alteration may be applied to the financial incentive schedule, such that the expected value of predictive accuracy would remain larger than the value to predictors realizable by distorting the decision process. Alas, I find myself at a loss to specify an alternate incentive schedule with the desired properties for questions of high significance.
If there's any reason to suspect grant-givers to be uninformed on the topic, or biased against it, crowd-sourcing a sum of that size sounds possible.
Yes, red and green seem subjectively very different -- but only to conscious attention. A green object amid many red objects (or vice versa) does not grab my attention in the way that, e.g. a yellow object might.
When shown a patch of red-or-green in a lab setting, I see "Red" or "Green" seemingly at random.
If shown a red patch next to a green patch in a lab, I'll see one "Red" and one "Green", but it's about 50:50 as to whether they'll be switched or not. How does that work? I have no hypotheses that aren't very low confidence. It seems as much a mystery to me as I infer it seems mysterious to you.