Comment author: cameroncowan 26 September 2015 05:14:25AM 0 points [-]

Where is the cultural context in all of this? How does that play in? Pain and pleasure here in the West is different than in the East just as value systems are different. When it comes to creating AGI I think a central set of agreed upon tenets are important. What is valuable? How can we quantify that in a way that makes sense to create AGI? If we want to reward it for doing good things, we have to consider cultural validation. We don't steal, murder or assault people because we have significant cultural incentive not to do so, especially if you live in a stable country. I think that could help. If we can somehow show group approval of the AGI, like favorable opinions, verbal validation and other things that I intrinsically values as we do. We could use our own culture to reinforce norms within it's archetecture.

Comment author: johnsonmx 28 September 2015 10:14:56PM 0 points [-]

A rigorous theory of valence wouldn't involve cultural context, much as a rigorous theory of electromagnetism doesn't involve cultural context.

Cultural context may matter a great deal in terms of how to build a friendly AGI that preserves what's valuable about human civilization-- or this may mostly boil down to the axioms that 'pleasure is good' and 'suffering is bad'. I'm officially agnostic on whether value is simple or complex in this way.

One framework for dealing with the stuff you mention is Coherent Extrapolated Volition (CEV)- it's not the last word on anything but it seems like a good intuition pump.

Comment author: TheAncientGeek 19 September 2015 09:08:32AM 0 points [-]

nd will make the other parts of the problem easier down the line

That's what I am disputing. You are taking a problem we don;t know how to make a start on, and turning it into a smaller problem we also don't know how to make a start on. That is't an advance. Reducing or simplifying a problem isn't an unconditional, universal solvent, it only works where the simpler problem is one you can actually make progress on.

IIT isn't going toi be of any real use unless it is confirmed, and how are you goign to confirm it, as a theory of qualia, without qualiometers?

If we are going to continue not having qualiometers, we may have to give up on testing consciousness objectiively in favour oof subjective measures...phenomenology and heterophenomenology. But you can only do heterophenomenology on a system that can report its subjective sates. Starting with simpler systems, like a single simulated pain receptor, is not going to work.

Comment author: johnsonmx 19 September 2015 09:32:35PM 0 points [-]

We're not on the same page. Let's try this again.

  • The assertion I originally put forth is AI safety; it is not about reverse-engineering qualia. I'm willing to briefly discuss some intuitions on how one may make meaningful progress on reverse-engineering qualia as a courtesy to you, my anonymous conversation partner here, but since this isn't what I originally posted about I don't have a lot of time to address radical skepticism, especially when it seems like you want to argue against some strawman version of IIT.

  • You ask for references (in a somewhat rude monosyllabic manner) on "some of the empirical work on coma patients IIT has made possible" and I give you exactly that. You then ignore it as "not really qualia research"- which is fine. But I'm really not sure how you can think that this is completely irrelevant to supporting or refuting IIT: IIT made a prediction, Casali et al. tested the prediction, the prediction seemed to hold up. No qualiometer needed. (Granted, this would be a lot easier if we did have them.)

This apparently leads to you say,

You are taking a problem we don;t know how to make a start on, and turning it into a smaller problem we also don't know how to make a start on.

More precisely, I'm taking a problem you don't know how to make a start on, and am turning it into a smaller problem that you also don't seem to know how to make a start on. Which is fine, and I don't wish to be a jerk about it, and not merely because Tononi/Tegmark/Griffith could be wrong in how they're approaching consciousness, and I could be wrong in how I'm adapting their stuff to try to explain some specific things about qualia. But you seem to just want to give up, to put this topic beyond the reach of science, and criticize anyone trying to find clever indirect approaches. Needless to say I vehemently disagree with the productiveness of that attitude.

I think we are in agreement that valence could be a fairly simple property. I also agree that the brain is Vastly Complex, and that qualia research has some excruciatingly difficult methodological hurdles to overcome, and I agree that IIT is still a very speculative hypothesis which shouldn't be taken on faith. I think we differ radically on our understandings of IIT and related research. I guess it'll be an empirical question whether IIT morphs into something that can substantially address questions of qualia- based on my understandings and intuitions, I'm pretty optimistic about this.

Comment author: TheAncientGeek 18 September 2015 10:06:24AM 0 points [-]

You mean this?

But that isn't really saying anything about qualia. The authors can relate their PCI measure to consciousness as judged medically... in humans. But would that scale be applicable to very simple systems or artificial systems? There is a real possibility that qualia could go missing in computational simulations,even assuming strict physicalism. In fact , we standardly assume that AIs embedded in games don't suffer.

Comment author: johnsonmx 18 September 2015 08:38:40PM 0 points [-]

If you're looking for a Full, Complete Data-Driven And Validated Solution to the Qualia Problem, I fear we'll have to wait a long, long time. This seems squarely in the 'AI complete' realm of difficulty.

But if you're looking for clever ways of chipping away at the problem, then yes, Casali's Perturbational Complexity Index should be interesting. It doesn't directly say anything about qualia, but it does indirectly support Tononi's approach, which says much about qualia. (Of course, we don't yet know how to interpret most of what it says, nor can we validate IIT directly yet, but I'd just note that this is such a hard, multi-part problem that any interesting/predictive results are valuable, and will make the other parts of the problem easier down the line.)

Comment author: TheAncientGeek 17 September 2015 09:52:21PM 0 points [-]

But given some of the empirical work on coma patients IIT has made possible. I

refs?

Comment author: johnsonmx 18 September 2015 12:40:10AM 0 points [-]

The stuff by Casali is pretty topical, e.g. his 2013 paper with Tononi.

Comment author: TheAncientGeek 17 September 2015 09:20:48AM 0 points [-]

Im still not seeing how IIT would help with confirming that an attempt at reverse engineering had succeeded, absent circular reasoning along the lines of "IIT says the system will have qualia. therefore the system wil have qualia".

Comment author: johnsonmx 17 September 2015 05:41:45PM 0 points [-]

Testing hypotheses derived from or inspired by IIT will probably be on a case-by-case basis. But given some of the empirical work on coma patients IIT has made possible. I think it may be stretching things to critique IIT as wholly reliant on circular reasoning.

That said, yes there are deep methodological challenges with qualia that any approach will need to overcome. I do see your objection quite clearly- I'm confident that I address this in my research (as any meaningful research on this must do) but I don't expect you to take my word for it. The position that I'm defending here is simply that progress in valence research will have relevance to FAI research.

Out of curiosity, do you think valence has a large or small kolgoromov complexity?

Comment author: TheAncientGeek 16 September 2015 10:22:07AM 0 points [-]

There's a lot of philosophical research, and very little scientific research. That confirms the impression of philosophers qualia are a Hard Problem.

How do you reverse engineer a quale? how do you tell you have succeeded? I think that you have underestimated the hardness of the problem.

Comment author: johnsonmx 17 September 2015 03:18:03AM 0 points [-]

I do have some detailed thoughts on your two questions-- in short, given certain substantial tweaks, I think IIT (or variants by Tegmark/Griffiths) can probably be salvaged from its (many) problems in order to provide a crisp dataset on which to base testable hypotheses about qualia.

(If you're around the Bay Area I'd be happy to chat about this over a cup of coffee or something.)

I would emphasize, though, that this post only talks about the value results in this space would have for FAI, and tries to be as agnostic as possible on how any reverse-engineering may happen.

Comment author: TheAncientGeek 15 September 2015 11:54:05AM *  0 points [-]

In what way is current research into qualia not valency research?

In what way would the valency research we dont yet have prove more successful?

Comment author: johnsonmx 15 September 2015 06:44:43PM 0 points [-]

Are you referring to any specific "current research into qualia", or just the idea of qualia research in general? I definitely agree that valence research is a subset of qualia research- but there's not a whole lot of either going on at this point, or at least not much that has produced anything quantitative/predictive.

I suspect valence is actually a really great path to approach more 'general' qualia research, since valence could be a fairly simple property of conscious systems. If we can reverse-engineer one type of qualia (valence), it'll help us reverse other types.

Comment author: johnswentworth 08 September 2015 11:28:03PM 1 point [-]

How are humans supposed to generate cryptographic reward tokens in such a way that an AI could not duplicate the process?

Comment author: johnsonmx 09 September 2015 12:39:28AM 0 points [-]

It would probably be highly dependent on the AI's architecture. The basic idea comes from Shulman and Bostrom - Superintelligence, chapter 9, in the "Incentive methods" section (loc 3131 of 8770 on kindle).

My understanding is that such a strategy could help as part of a comprehensive strategy of limitations and inventivization but wouldn't be viable on its own.

Comment author: cosmicvoyager 23 June 2015 12:13:36PM *  1 point [-]

I am very surprised and pleased to find thinking that so closely parallels my own.

I typed this is Google: "what makes a particular arrangement of matter in the brain pain or pleasure"

This is what I have been thinking. The universe contains consciousness. Matter and energy (at least arranged in the form of a brain) are conscious. There are states of consciousness that feel good and states of consciousness that feel bad. What is the difference between the arrangements of matter (or the processes) in the brain that make some feel good and some feel bad? I think it would be extremely helpful and lead to answers if we actually knew exactly what these arrangements are and studied them, studied the physics of what is going on. Do we know yet? How much do we know?

Reasons I want to know

1 - curiosity 2 - increasing happiness and decreasing suffering in biological beings 3 - creating synthetic intelligence that is happy 4 - minimizing suffering in the universe and maximizing pleasure in the universe. it just occurred to me a few days ago that all the matter in the universe could be converted to whatever state is the most pleasure, convert the universe into bliss. hmm...many minds or one giant mind?

Comment author: johnsonmx 28 June 2015 01:20:33AM 1 point [-]

Right, absolutely. These are all things that we don't know, but should.

Are you familiar with David Pearce's Hedonistic Imperative movement? He makes a lot of the same points and arguments, basically outlining that it doesn't seem impossible that we could (and should) radically reduce, and eventually eliminate, suffering via technology.

But the problem is, we don't know what suffering is. So we have to figure that out before we can make much radical progress on this sort of work. I.e., I think a rigorous definition of suffering will be an information-theoretic one-- that it's a certain sort of pattern within conscious systems-- but we know basically nothing about what sort of pattern it is.

(I like the word "valence" instead of pain/pleasure, joy/suffering, eudaimonia, hedonic tone, etc. It's a term from psychology that just means 'the pleasantness or unpleasantness attached to any experience' and seems to involve less baggage than these other terms.)

I hope to have a formal paper on this out by this winter. In the meantime, if you're in the Bay Area, feel free to ping me and I can share some thoughts. You may also enjoy a recent blog post: Effective Altruism, and building a better QALY.

Comment author: minusdash 03 March 2015 01:23:04AM *  0 points [-]

Life, sin, disease, redness, maleness and indeed dogness "may" also be like electromagnetism. The English language may also be a fundamental part of the universe and maybe you could tell if "irregardless" or "wanna" are real English words by looking into a microscope or turning your telescope to certain parts of the sky, or maybe by looking at chicken intestines, who knows. I know some people think like this. Stuart Hameroff says that morality may be encoded into the universe at the Planck scale. So maybe that's where you should look for "good", maybe "pleasure" is there as well.

But anyway, research into electromagnetism was done using the scientific method, which means that the hypothesis had to produce predictions that were tested and replicated numerous times. What sort of experiment would you envision for testing something about "inherently pleasurable" arrangements of atoms? Would the atoms make you feel warm and fuzzy inside when you look at them? Or would you try to put that pattern into different living creatures and see if they react with their normal joyful reactions?

Comment author: johnsonmx 05 March 2015 06:44:35PM 0 points [-]

Although life, sin, disease, redness, maleness, and dogness are (I believe) inherently 'leaky' / 'fuzzy' abstractions that don't belong with electromagnetism, this is a good comment. If a hypothesis is scientific, it will make falsifiable predictions. I hope to have something more to share on this soon.

View more: Next