Comment author: John_Maxwell_IV 12 December 2015 03:18:06AM *  6 points [-]

I left this comment on Hacker News exploring whether "AI for everyone" will be a good thing or not. Interested to hear everyone's thoughts.

Comment author: danieldewey 12 December 2015 06:12:05PM 2 points [-]

Very thoughtful post! I was so impressed that I clicked the username to see who it was, only to see the link to your LessWrong profile :)

Comment author: danieldewey 01 September 2015 06:49:14PM 3 points [-]

Just wanted to mention that watching this panel was one of the things that convinced me to give AI safety research a try :) Thanks for re-posting, it's a good memory.

To at least try to address your question: one effect could be that there are coordination problems, where many people would be trying to "change the world" in roughly the same direction if they knew that other people would cooperate and work with them. This would result in less of the attention drain you suggest. This seems more like what I've experienced.

I'm more worried about people being stupid than mean, but that could be an effect of the bubble of non-mean people I'm part of.

Comment author: danieldewey 13 March 2015 12:33:03AM 2 points [-]

Is this sort of a way to get an agent with a DT that admits acausal trade (as we think the correct decision theory would) to act more like a CDT agent? I wonder how different the behaviors of the agent you specify are from those of a CDT agent -- in what kinds of situations would they come apart? When does "I only value what happens given that I exist" (roughly) differ from "I only value what I directly cause" (roughly)?

Comment author: spxtr 19 February 2015 08:54:01PM 3 points [-]

If you have a different version of QM (perhaps what Ted Bunn has called a “disappearing-world” interpretation), it must somehow differ from MWI, presumably by either changing the above postulates or adding to them. And in that case, if your theory is well-posed, we can very readily test those proposed changes. In a dynamical-collapse theory, for example, the wave function does not simply evolve according to the Schrödinger equation; it occasionally collapses (duh) in a nonlinear and possibly stochastic fashion. And we can absolutely look for experimental signatures of that deviation, thereby testing the relative adequacy of MWI vs. your collapse theory.

He asserts that such an experiment exists. I would love it if he were to expand on this assertion.

Comment author: danieldewey 20 February 2015 09:22:38AM 2 points [-]

I don't have the expertise to evaluate it, but Brian Greene suggests this experiment.

Comment author: Kaj_Sotala 08 February 2015 08:03:45AM 4 points [-]

Thanks, Mark. I'm definitely thinking about applying, but my current problem is that I have too many potential proposals that I could write:

  • There's the concept learning research program that you mentioned.
  • I also have an interest in figuring just what exactly human values and preferences are: this seems like a topic where there is likely to be plenty of low-hanging fruit, given that a number of fields touch upon the topic: it relates to concept learning, also in AI there's preference learning, philosophers have tried to ask the question of what value is, there's a bunch of stuff in neuroscience about motivation and preference, a bunch of stuff that I've read about emotion research lately is relevant, there's the economic definition of preferences and their research into the topic, there's intercultural comparisons about values in sociology and anthropology... It seems like if someone went through all of this stuff while keeping in mind the question of "okay, so exactly what part would we want an AI to extrapolate and how and why", one should be able to make considerable progress.
  • I was just recently talking to some friends about ontology identification and ontological crises, and based on some preliminary discussion, it seemed to us like one could make progress on it using an approach inspired by conceptual metaphors. Briefly, a conceptual metaphor is a mapping from one domain to another, "that allows us to use the inferential structure of one conceptual domain (say, geometry) to reason about another (say, arithmetic)" (Lakoff & Núñez, p. 6). Intuitively, the physicists who discovered quantum mechanics knew that their discovery didn't make everything they knew of classical mechanics obsolete - all the observations supporting CM were still there, so there had to exist some mapping from CM concepts to QM concepts in a way that allowed us to preserve most of what we knew of CM. That would suggest that an AI that discovered that its current understanding of the world was lacking could attempt to take its old world-model, and identify the things that it valued in the new model by looking for the inferential rules of the old model that could still be mapped into the new one, and using that to identify the corresponding mappings to entities. One of the friends I was talking with indicated that there's research on the deep learning side that might be able to do something like this.
  • At this rate, tomorrow I'll probably come up with a fourth thing that seems like a promising research direction.

I guess I could just write proposals for each of these and let FLI decide which one they find the most promising - their FAQ says that one can submit several proposals but they'll only invite one Full Proposal from a single PI.

Comment author: danieldewey 09 February 2015 06:57:11PM 3 points [-]

I would encourage you to apply, these ideas seem reasonable!

As far as choosing, I would advise you to choose the idea for which you can make the case most strongly that it is Topical and Impactful, as defined here.

Request for proposals for Musk/FLI grants

22 danieldewey 05 February 2015 05:04PM

As a follow-on to the recent thread on purchasing research effectively, I thought it'd make sense to post the request for proposals for projects to be funded by Musk's $10M donation. LessWrong's been a place for discussing long-term AI safety and research for quite some time, so I'd be happy to see some applications come out of LW members.

Here's the full Request for Proposals.

If you have questions, feel free to ask them in the comments or to contact me!

Here's the email FLI has been sending around:

Initial proposals (300–1000 words) due March 1, 2015

The Future of Life Institute, based in Cambridge, MA and headed by Max Tegmark (MIT), is seeking proposals for research projects aimed to maximize the future societal benefit of artificial intelligence while avoiding potential hazards. Projects may fall in the fields of computer science, AI, machine learning, public policy, law, ethics, economics, or education and outreach. This 2015 grants competition will award funds totaling $6M USD.

This funding call is limited to research that explicitly focuses not on the standard goal of making AI more capable, but on making AI more robust and/or beneficial; for example, research could focus on making machine learning systems more interpretable, on making high-confidence assertions about AI systems' behavior, or on ensuring that autonomous systems fail gracefully. Funding priority will be given to research aimed at keeping AI robust and beneficial even if it comes to greatly supersede current capabilities, either by explicitly focusing on issues related to advanced future AI or by focusing on near-term problems, the solutions of which are likely to be important first steps toward long-term solutions.

Please do forward this email to any colleagues and mailing lists that you think would be appropriate.

Proposals

Before applying, please read the complete RFP and list of example topics, which can be found online along with the application form:

    http://futureoflife.org/grants/large/initial

As explained there, most of the funding is for $100K–$500K project grants, which will each support a small group of collaborators on a focused research project with up to three years duration. For a list of suggested topics, see the complete RFP [1] and the Research Priorities document [2]. Initial proposals, which are intended to require merely a modest amount of preparation time, must be received on our website [1] on or before March 1, 2015.

Initial proposals should include a brief project summary, a draft budget, the principal investigator’s CV, and co-investigators’ brief biographies. After initial proposals are reviewed, some projects will advance to the next round, completing a Full Proposal by May 17, 2015. Public award recommendations will be made on or about July 1, 2015, and successful proposals will begin receiving funding in September 2015.

References and further resources

[1] Complete request for proposals and application form: http://futureoflife.org/grants/large/initial

[2] Research Priorities document: http://futureoflife.org/static/data/documents/research_priorities.pdf

[3] An open letter from AI scientists on research priorities for robust and beneficial AI: http://futureoflife.org/misc/open_letter

[4] Initial funding announcement: http://futureoflife.org/misc/AI

Questions about Project Grants: dewey@futureoflife.org

Media inquiries: tegmark@mit.edu

Comment author: ciphergoth 28 November 2014 09:57:53AM 6 points [-]

This looks to me like a misunderstanding of Müller & Bostrom 2014. The actual figure is that 50% of AI researchers give a 10% probability of HLMI by 2022.

Müller, V. C., & Bostrom, N. Future progress in artificial intelligence: A survey of expert opinion. In V. C. Müller (Ed.), Fundamental Issues of Artificial Intelligence. Berlin: Springer. 2014

Comment author: danieldewey 29 November 2014 12:36:20PM 4 points [-]

That's what I thought at first, too, but then I looked at the paper, and their figure looks right to me. Could you check my reasoning here?

On p.11 of Vincent's and Nick's survey, there's a graph "Proportion of experts with 10%/50%/90% confidence of HLMI by that date". At around the the 1 in 10 mark of proportion of experts -- the horizontal line from 0.1 -- the graph shows that 1 in 10 experts thought there was a 50% chance of HLAI by 2020 or so (the square-boxes-line), and 1 in 10 thought there was a 90% chance of HLAI by 2030 or so (the triangles-line). So, maybe 1 in 10 researchers think there's a 70% chance of HLAI by 2025 or so, which is roughly in line with the journalist's remark.

Did I do that right? Do you think the graph is maybe incorrect? I haven't checked the number against other parts of the paper.

There's a good chance that the reviewer got the right number by accident, I think, but it doesn't seem far enough away to call out.

Comment author: Vulture 14 October 2014 05:29:07PM 5 points [-]

So, the "polymath" thread seems to have ground to a halt. I can't tell whether the discussion just stopped going anywhere (possibly due to elimination of low-hanging fruit), or it dropped off the recent-posts list and people forgot about it, or what. Does anyone have any insight into what's going on?

Comment author: danieldewey 15 October 2014 08:52:41AM 3 points [-]

Lots of good stuff happened there, but it looks like it'll have to be curated fairly actively to continue to make progress, and unfortunately that doesn't fit with my current duties.

If someone else would like to act as a leader for it, I'd be happy for that! In any case, I'm glad we tried it, and thankful that so many people jumped in.

Comment author: Illano 30 September 2014 02:06:03PM 3 points [-]

I was thinking last night of how vote trading would work in a completely rational parliamentary system. To simplify things a bit, lets assume that each issue is binary, each delegate holds a position on every issue, and that position can be normalized to a 0.0 - 1.0 ranking. (e.g. If I have a 60% belief that I will gain 10 utility from this issue being approved, it may have a normalized score of .6, if it is a 100% belief that I will gain 10 utility it may be a .7, while a 40% chance of -1000 utility may be a .1) The mapping function doesn't really matter too much, as long as it can map to the 0-1 scale for simplification.

The first point that seems relatively obvious to me is that all rational agents will intentionally mis-state their utility functions as extremes for bargaining purposes. In a trade, you should be able to get a much better exchange by offering to update from 0 to 1 than you would for updating from 0.45 to 1, and as such, I would expect all utility function outputs to be reported to others as either 1 or 0, which simplifies things even further, though internally, each delegate would keep their true utlity function values. (As a sanity check, compare this to the current parliamentary models in the real world, where most politicians represent their ideals publicly as either strongly for or strongly against)

The second interesting point I noticed is that with the voting system as proposed, where every additional vote grants additional probability of the measure being enacted, every vote counts. This means it is always a good trade for me to exchange votes when my expected value of the issue you are changing position on is higher than my expected value of the position I am changing position on. This leads to a situation, where I am better off changing positions on every issue except the one that brings me the most utility in exchange for votes on the issue that brings me the most utility. Essentially, this means that the only issue that matters to an individual delegate is the issue that potentially brings them the most utility, and the rest of the issues are just fodder for trading.

Given the first point I mentioned, that all values should be externally represented as either 1 or 0, it seems that any vote trade will be a straight 1 for 1 trade. I haven't exactly worked out the math here, but I'm pretty sure that for an arbitrarily large parliament with an arbitrarily large number of issues (to be used for trading), the result of any given vote will be determined by the proportion of delegates holding that issue as either their highest or lowest utility issue, with the rest of the delegates trading their votes on that issue for votes on another issue they find to be higher utility. (As a second sanity check, this also seems to conform closely to reality with the way lobbyist groups push single issues and politicians trade votes to push their pet issues through the vote.)

This is probably an oversimplified case, but I thought I'd throw it for discussion to see if it sparked any new ideas.

Comment author: danieldewey 01 October 2014 01:31:50PM 1 point [-]

If what you say is true about all trades being 1-for-1, that seems more like a bug than a feature; if an agent doesn't have any votes valuable enough to sway others, it seems like I'd want them to be able (i.e. properly incentivized) to offer more votes, so that the system overall can reflect the aggregate's values more sensitively. I don't have a formal criterion that says why this would be better, but maybe that points towards one.

Comment author: Manfred 27 September 2014 04:12:08AM *  7 points [-]

My suspicion is that this just corresponds to some particular rule for normalizing preferences over strategies. The "amount of power" given to each faction is capped, so that even if some faction has an extreme opinion about one issue it can only express itself by being more and more willing to trade other things to get it.

If goodness numbers are normalized, and some moral theory wants to express a large relative preference for one thing over another, it can't just crank up the number on the thing it likes - it must flatten the contrast of things it cares less about in order to express a more extreme preference for one thing.

Comment author: danieldewey 28 September 2014 03:39:24PM 1 point [-]

This actually sounds plausible to me, but I'm not sure how to work it out formally. It might make for a suprising and interesting result.

View more: Next