All of Henry Prowbell's Comments + Replies

I'm afraid so. Sorry. We hope to run more in the future!

Answer by Henry Prowbell10

Some of the later levels on this?

https://en.wikipedia.org/wiki/Notpron

“Notpron is an online puzzle game and internet riddle created in 2004 by German game developer David Münnich. It has been named as ‘the hardest riddle available on the internet.’”

“Notpron follows a standard puzzle game layout, where the player is presented with a webpage containing a riddle and must find the answer to the riddle in order to proceed to the next webpage”

“Each level answer or solution is unique, often requiring specific skills such as decoding ciphers,&n... (read more)

My model of a non-technical layperson finds it really surprising that an AGI would turn rogue and kill everyone. For them it’s a big and crazy claim.

They imagine that an AGI will obviously be very human-like and the default is that it will be cooperative and follow ethical norms. They will say you need some special reason why it would decide to do something so extreme and unexpected as killing everyone.

When I’ve talked to family members and non-EA friends that’s almost always the first reaction I get.

If you don’t address that early in the introduction I th... (read more)

5adamShimi
Thanks for the comment! We'll consider this point for future releases, but personally, I would say that this kind of hedging also has a lot of downsides: it makes you sound far more uncertain and defensive than you really want to. This document tries to be both grounded and to the point, and so we by default don't want to put ourselves in a defensive position when arguing things that we think make sense and are supported by the evidence.

I’ll give it a go.

I’m not very comfortable with the term enlightened but I’ve been on retreats teaching non-dual meditation, received ‘pointing out instructions’ in the Mahamudra tradition and have experienced some bizarre states of mind where it seemed to make complete sense to think of a sense of awake awareness as being the ground thing that was being experienced spontaneously, with sensations, thoughts and emotions appearing to it — rather than there being a separate me distinct from awareness that was experiencing things ‘using my awareness’, which is... (read more)

4Kaj_Sotala
Great description. This sounds very similar to some of my experiences with non-dual states.

Great summary, and really happy that this helped you!

I'd recommend people read Rick Hanson's paper on HEAL, if they're interested too: https://rickhanson.net/wp-content/uploads/2021/12/LLPE-paper-final2.pdf

Does it make sense to put any money into a pension given your outlook on AGI?

1Vlad Sitalo
Also curious how this changes people's outlooks on putting money into 401k/IRA's/etc

I really like the way it handles headlines and bullet point lists!

In an ideal world I'd like the voice to sound less robotic. Something like https://elevenlabs.io/ or https://www.descript.com/overdub.  How much I enjoy listening to text-to-speech content depends a lot on how grating I find the voice after long periods of listening.

5PeterH
Thanks! We're currently using Azure TTS. Our plan is to review every couple months and update to use better voices when they become available on Azure or elsewhere. Elevenlabs is a good candidate but unfortunately they're ~10x more expensive per hour of narration than Azure ($10 vs $1).

Honestly, no plans at the moment. Writing these was a covid lockdown hobby. It's vaguely possible I'll finish it one day but I wouldn't hold your breath. Sorry.

2momom2
Heh, it's better to have this than nothing at all! I'll keep hoping for it. ^-^

But I rarely see anyone touch on the idea of "what if we only make something as smart as us?"

 

But why would intelligence reach human level and then halt there? There's no reason to think there's some kind of barrier or upper limit at that exact point.

Even in the weird case where that were true, aren't computers going to carry on getting faster? Just running a human level AI on a very powerful computer would be a way of creating a human scientist that can think at 1000x speed, create duplicates of itself, modify it's own brain. That's already a superintelligence isn't it?  

-2thefirechair
The assumption there is that the faste the hardware underneath, the faster the sentience running on it will be. But this isn't supported by evidence. We haven't produced a sentient AI to know whether this is true or not. For all we know, there may be a upper limit to "thinking" based on neural propagation of information. To understand and integrate a concept requires change and that change may move slowly across the mind and underlying hardware. Humans have sleep for example to help us learn and retain information. As for self modification - we don't have atomic level control over the meat we run on. A program or model doesn't have atomic level control over its hardware. It can't move an atom at will in its underlying circuitry to speed up processing for example. This level of control does not exist in nature in any way. We don't know so many things. For example, what if consciousness requires meat? That it is physically impossible on anything other than meat? We just assume it's possible using metal and silica.

A helpful way of thinking about 2 is imagining something less intelligent than humans trying to predict how humans will overpower it.

You could imagine a gorilla thinking "there's no way a human could overpower us. I would just punch it if it came into my territory." 

The actual way a human would overpower it is literally impossible for the gorilla to understand (invent writing, build a global economy, invent chemistry, build a tranquilizer dart gun...)

The AI in the AI takeover scenario is that jump of intelligence and creativity above us. There's literally no way a puny human brain could predict what tactics it would use. I'd imagine it almost definitely involves inventing new branches of science.

-2thefirechair
I'd suggest there may be an upper bound to intelligence because intelligence is bound by time and any AI lives in time like us. They can't gather information from the environment any faster. They cannot automatically gather all the right information. They cannot know what they do not know. The system of information, brain propagation, cellular change runs at a certain speed for us. We cannot know if it is even possible to run faster. One of the magical thinking criticisms I have of AI is that it suddenly is virtually omniscient. Is that AI observing mold cultures and about to discover penicillin? Is it doing some extremely narrow gut bateria experiment to reveal the source of some disease? No it's not. Because there are infinite experiments to run. It cannot know what it does not know. Some things are Petri dishes and long periods of time in the physical world and require a level of observation the AI may not possess.

I think that's true of people like: Steven Pinker and Neil deGrasse Tyson. They're intelligent but clearly haven't engaged with the core arguments because they're saying stuff like "just unplug it" and "why would it be evil?"

But there's also people like...

Robin Hanson. I don't really agree with him but he is engaging with the AI risk arguments, has thought about it a lot and is a clever guy.

Will MacAskill. One of the most thoughtful thinkers I know of, who I'm pretty confident will have engaged seriously with the AI Risk arguments. His p(doom) is far lower... (read more)

4Donald Hobson
Robin Hanson is weird. He paints a picture of a grim future where all nice human values are eroded away, replaced with endless frontier replicators optimized and optimizing only for more replication. And then he just accepts it as if that was fine.  Will Macaskill seems to think AI risk is real. He just thinks alignment is easy. He has a specific proposal involving making anthropomorphic AI and raising it like a human child that he seems keen on.

I find Eliezer and Nates' arguments compelling but I do downgrade my p(doom) somewhat (-30% maybe?) because there are intelligent people (inside and outside of LW/EA) who disagree with them.

I had some issues with the quote

Will continue to exist regardless of how well you criticize any one part of it.

I'd say LW folk are unusually open to criticism. I think if there were strong arguments they really would change people's minds here. And especially arguments that focus on one small part at a time.

But have there been strong arguments? I'd love to read them.

&nb... (read more)

1DavidW
I just posted a detailed explanation of why I am very skeptical of the traditional deceptive alignment story. I'd love to hear what you think of it!  Deceptive Alignment Skepticism - LessWrong
6Donald Hobson
There are intelligent people who disagree, but I was under the impression there was a shortage of intelligent disagreement. Most of the smart disagreement sounds like smart people who haven't thought in great depth about AI risk in particular, and are often shooting down garbled misunderstandings of the case for AI risk.

For me the core of it feels less like trying to "satisfying the values you think you should have, while neglecting the values you actually have" and more like having a hostile orientation to certain values I have.

I might be sitting at my desk working on my EA project and the parts of me that are asking to play video games, watch arthouse movies, take the day off and go hiking, find a girlfriend are like yapping dogs that won't shut up. I'll respond to their complaints once I've finished saving the world.

Through CFAR workshops, lots of goal factoring, journ... (read more)

6Duncan Sabien (Deactivated)
Just noting that I'm reasonably confident that neither Logan nor most CFAR staff would claim that values are immutable; just that they are not easily changeable. I think values do, indeed, shift; we can see this when e.g. people go through puberty or pregnancy or lose a limb or pass through a traumatic experience like a war zone. This puts a floor on how immutable values/needs can really be, and presumably they can be shifted via less drastic interventions.

You are however only counting one side here

 

In that comment I was only offering plausible counter-arguments to "the amount of people that were hurt by FTX blowing up is a rounding error."

How to model all the related factors is complicated. Saying that you easily know the right answer to whether the effects are negative or positive in expectation without running any numbers seems to me unjustified. 

I think we basically agree here.

I'm in favour of more complicated models that include more indirect effects, not less.

Maybe the difference is: I think ... (read more)

3ChristianKl
We already have an EA movement where the leading organization has no problem editing out elements of a picture it publishes on its website because of possible PR risks. While you can argue that it's not literally lying it comes very close and suggests the kind of environment that does not have the strong norms that would be desirable   I don't think FTX/Almeda doing this in secret strongly damaged general norms against lying, corruption, and fraud. Them blowing up like this actually is a chance for moving toward those norms. It's a chance to actually look into ethics in a different way to make it more clear that being honest and transparent is good.  Saying "poor messaging on our part" which resulted in "actions were negative in expectation in a purely utilitarian perspective" is a way to avoid having the actual conversation about the ethical norms that might produce change toward stronger norms for truth. 

If you believe that each future person is as valuable as each present person and there will be 10^100 people in the future lightcone, the amount of people that were hurt by FTX blowing up is a rounding error.

 

But you have to count the effect of the indirect harms on the future lightcone too. There's a longtermist argument that SBF's (alleged and currently very likely) crimes plausibly did more harm than all the wars and pandemics in history if...

  • Governments are now 10% less likely to cooperate with EAs on AI safety
  • The next 2 EA mega-donors decide to pass on EA
  • (Had he not been caught:) The EA movement drifted towards fraud and corruption
  • etc.
0ChristianKl
You are however only counting one side here. SBF appearing successful was a motivating example for others to start projects that would have made them Mega donors. I don't think that's likely to be the case.  There's an unclearness here about what "pass on EA means". Zvi wrote about Survival and Flourishing Fund not being an EA fund. How to model all the related factors is complicated. Saying that you easily know the right answer to whether the effects are negative or positive in expectation without running any numbers seems to me unjustified. 

The frequency with which datacenters, long range optical networks, and power plants, require human intervention to maintain their operations, should serve as a proxy to the risk an AGI would face in doing anything other than sustaining the global economy as is.

 

Probably those things are trivially easy for the AGI to solve itself e.g. with nanobots that can build and repair things.

I'm assuming this thing is to us what humans are to chimps, so it doesn't need our help in solving trivial 21 century engineering and logistics problems.

 

The strategic c... (read more)

5Ruby
It will be made!

I haven’t looked into his studies’ methodologies, but from my experience with them, I would put high odds that the 65% number is exaggerated.

 

From his sales page

"In our scientific study involving 245 people...

65% of participants who completed The 45 Days to Awakening Challenge and Experiment persistently awakened.

...

Another couple hundred people entered the program already in a place of Fundamental Wellbeing..."

Sounds like he's defining enlightenment as something that ~50% of people already experience.

Elsewhere he describes 'Location 1' enlightenment ... (read more)

Does anybody know if the Highlights From The Sequences are compiled in ebook format anywhere?

Something that takes 7 hours to read, I want to send to my Kindle and read in a comfy chair.

And maybe even have audio versions on a single podcast feed to listen to on my commute.

(Yes, I can print out the list of highlighted posts and skip to those chapters of the full ebook manually but I'm thinking about user experience, the impact of trivial inconveniences, what would make Lesswrong even more awesome.)

I love his books too. It's a real shame.

"...such as imagining that an intelligent tool will develop an alpha-male lust for domination."

It seems like he really hasn't understood the argument the other side is making here.

It's possible he simply hasn't read about instrumental convergence and the orthogonality thesis. What high quality widely-shared introductory resources do we have on those after all? There's Robert Miles, but you could easily miss him.

I'm imagining the CEO having a thought process more like...

- I have no idea how my team will actually react when we crack AGI 
- Let's quickly Google 'what would you do if you discovered AGI tomorrow?'*
- Oh Lesswrong.com, some of my engineering team love this website
- Wait what?!
- They would seriously try to [redacted]
- I better close that loophole asap

I'm not saying it's massively likely that things play out in exactly that way but a 1% increased chance that we mess up AI Alignment is quite bad in expectation.

*This post is already the top result on Google for that particular search

3Donald Hobson
Ok. Redacted part of my reply in response.

I immediately found myself brainstorming creative ways to pressure the CEO into delaying the launch (seems like strategically the first thing to focus on) and then thought 'is this the kind of thing I want to be available online for said CEOs to read if any of this happens?'

I'd suggest for those reasons people avoid posting answers along those lines.

5Donald Hobson
A CEO that has somehow read and understood that post, despite not reading any part of lesswrong warning that AI might be dangerous? 

Somebody else might be able to answer better than me. I don't know exactly what each researcher is working on right now.

“AI safety are now more focused on incidental catastrophic harms caused by a superintelligence on its way to achieve goals”

Basically, yes. The fear isn’t that AI will wipe out humanity because someone gave it the goal ‘kill all humans’.

For a huge number of innocent sounding goals ‘incapacitate all humans and other AIs’ is a really sensible precaution to take if all you care about is getting your chances of failure down to zero. As is hidi... (read more)

I read the article and I have to be honest I struggled to follow her argument or to understand why it impacts your decision to work on AI alignment. Maybe you can explain further?

The headline "Debating Whether AI is Conscious Is A Distraction from Real Problems" is a reasonable claim but the article also makes claims like...

"So from the moment we were made to believe, through semantic choices that gave us the phrase “artificial intelligence”, that our human intelligence will eventually contend with an artificial one, the competition began... The reality is... (read more)

1sidhe_they
I probably should have included the original Twitter thread that sparked the article link in which the author says bluntly that she will no longer discuss AI consciousness/superintelligence. Those two had become conflated, so thanks for pointing that out! With regards to instrumental convergence (just browsed the Arbitral page), are you saying the big names working on AI safety are now more focused on incidental catastrophic harms caused by a superintelligence on its way to achieve goals, rather than making sure artificial intelligence will understand and care about human values?

I suspect you should update the website with some of this? At the very least copying the above comment into a 2022 updates blog post.

The message 'CFAR did some awesome things that we're really proud of, now we're considering pivoting to something else, more details to follow' would be a lot better than the implicit message you may be sending currently 'nobody is updating this website, the CFAR team lost interest and it's not clear what the plan is or who's in charge anymore'

If somebody has time to pour into this I'd suggest recording an audio version of Mad Investor Chaos.

HPMOR reached a lot more people thanks to Eneasz Brodski's podcast recordings. That effect could be much more pronounced here if the weird glowfic format is putting people off.

I'd certainly be more likely to get through it if I could play it in the background whilst doing chores, commuting or falling asleep at night.

That's how I first listened to HPMOR, and then once I'd realised how good it was I went back and reread it slowly, taking notes, making an effort to internalize the lessons.

1EniScien
Hmm, funny, I usually listen to audiobooks, but this was not the case with HPMOR, I realized "how good it is" literally from the first chapter, which is extremely rare with books.

I have a sense of niggling confusion.

This immediately came to mind...

"The only way to get a good model of the world inside your head is to bump into the world, to let the light and sound impinge upon your eyes and ears, and let the world carve the details into your world-model. Similarly, the only method I know of for finding actual good plans is to take a bad plan and slam it into the world, to let evidence and the feedback impinge upon your strategy, and let the world tell you where the better ideas are." - Nate Soares, https://mindingourway.com/dive-in-... (read more)

2LoganStrohl
(this is the same question i was trying to ask in another comment but you did it better.)
7Duncan Sabien (Deactivated)
Yes, good point (and thanks). Perhaps we should change "come out" to "must report in," at least for some subset of 1,000-day monks who do indeed need to continually bump into the territory. EDIT: this led to an edit in response!  Double thank-you.

If you haven't already, I'd suggest you put a weekend aside and read through the guides on https://80000hours.org/

They have some really good analyses on when you should do a PhD, found a startup, etc.

what are some signs that someone isn’t doing frame control? [...]

  1. They give you power over them, like indications that they want your approval or unconditional support in areas you are superior to them. They signal to you that they are vulnerable to you.

 

There was a discussion on the Sam Harris podcast where he talks about the alarming frequency at which leaders of meditation communities end up abusing, controlling or sleeping with their students. I can't seem to find the episode name now.

But I remember being impressed with the podcast guest, a meditat... (read more)

8Spiracular
I think I have seen the "sanity-check"/"sanity-guillotine" thing done well. I have also seen it done poorly, in a way that mostly resembles the "finger-trap" targeting any close friends who notice problems. For actual accountability/protection? "Asking to have it reported publicly/to an outside third party" seems to usually work better than "Report it to me privately." (A very competent mass-crowd-controller might have a different dynamic, though; I haven't met one yet.) ---------------------------------------- For strong frame-controllers? "Encouraging their students to point out a vague category of issue in private," has a nasty tendency to speed up evaporative cooling, and burns out the fire of some of the people who might otherwise have reported misbehavior to a more-objective third-person. It can set up the frame-controller as the counter/arbiter of "how many real complains have been leveled their way about X" (...which they will probably learn to lie about...), frames them as "being careful about X," and gives the frame-controller one last pre-reporting opportunity to re-frame-control things in the sender. I think the "private reporting" variant is useful to protect a leader from unpleasant surprises, gives them a quick chance to update out of a bad pattern early on, and is slightly good for that reason. But I think as an "accountability method," this is simply not a viable protection against an even halfway-competent re-framer. ---------------------------------------- I think the gold-standard for actual accountability, is closer to the "outside HR firm" model. Having someone outside your circle, who people report serious issues to, and who is not primarily accountable to you. Not everyone has access to the gold-standard, though. When I single a person out for my future accountability? I pick people who I view as (high-integrity low-jealousy) peers-or-higher, AND/OR people on a totally different status-ladder. I want things set up such that even a m
6Alexei
Part of my model is that spiritual students tend to be a lot more prone to wanting to connect in a physical way. Sometimes to the point of almost literally throwing themselves at the teacher.
  • Erasable pens. Pens are clearly better than pencils in that you can write on more surfaces and have better colour selection. The only problem is you can’t erase them. Unless they’re erasable pens that is, then they strictly dominate. These are the best I’ve found that can erase well and write on the most surfaces.

 

I also loved these Frixion erasable pens when I discovered them. 

But another even better step up in my writing-by-hand experience was the reMarkable tablet. Genuinely feels like writing on paper — but with infinite pages, everything syn... (read more)

I've got this printed out on my desk at home but unfortunately I'm away on holiday for the next few weeks. I'll find it for you when I get back.

For what it's worth most of the ideas for this chapter comes from Stanislas Dehaene's book Consciousness and the Brain. Kaj Sotala has a great summary here and I'd recommend reading the whole book too if you've got the time and interest.

Well spotted! The Psychomagic for Beginners excerpt certainly takes some inspiration from that. I read that book a few years ago and really enjoyed it too.

Thanks Ustice!

I've already written first drafts of a couple more chapters which I'll be polishing and posting over the next few months.

So I can guarantee at least a few more installments. After that it will depend on what kind of response I get and whether I'm still enjoying the writing process.

Early in HPMOR there's a bit where Harry mentions the idea of using magic to improve his mind but it's never really taken much further.

I wanted to write about that: if you lived in a universe with magic how could you use it to improve your intelligence and rationali... (read more)