All of weverka's Comments + Replies

weverka32

Ukraine gave up its nuclear weapons in exchange for our assurance of its territorial integrity.

The Budapest Memorandum on Security Assurances commits us to help defend Ukraine.

What "non-appeasing de-escalation strategies" are you proposing.  We could remind Russia that it is a signatory to the Budapest Memorandum.  Do you think that will move them to withdraw?

This letter offers nothing.  

5Nathan Helm-Burger
The failure of the Budapest Memorandum is a very clear signal to other nations that they cannot risk nuclear dearmament, and indeed should push to become a nuclear power. This seems very bad for the world. I think the correct move would have been for the majority of nations of the world to object to Russia's invasion of Ukraine on these grounds, and then for meaningful pushback to happen. Instead, it was a case of too little too late. And you are arguing that the world should be even more submissive to Russia's aggressive treaty violation?
weverka-10

Why down votes and a statement that I am wrong because I misunderstood.  

This is a mean spirited reaction when I lead with admission that I could not follow the argument.  I offered a concrete example and stated that I could not follow the original thesis as applied to the concrete example.  No one took me up on this.  

Are you too advanced to stoop to my level of understanding and help me figure out how this abstract reasoning applies to a particular example?   Is the shut down mechanism suggested by Yudkowsky too simple?

3RHollerith
Yudkowsky's suggestion is for preventing the creation of a dangerous AI by people. Once a superhumanly-capable AI has been created and has had a little time to improve its situation, it is probably too late even for a national government with nuclear weapons to stop it (because the AI will have hidden copies of itself all around the world or taken other measures to protect itself, measures that might astonish all of us). The OP in contrast is exploring the hope that (before any dangerous AIs are created) a very particular kind of AI can be created that won't try to prevent people from shutting it down.
2EJT
Hi weverka, sorry for the downvotes (not mine, for the record). The answer is that Yudkowsky's proposal is aiming to solve a different 'shutdown problem' than the shutdown problem I'm discussing in this post. Yudkowsky's proposal is aimed at stopping humans developing potentially-dangerous AI. The problem I'm discussing in this post is the problem of designing artificial agents that both (1) pursue goals competently, and (2) never try to prevent us shutting them down.
-1weverka
Why down votes and a statement that I am wrong because I misunderstood.   This is a mean spirited reaction when I lead with admission that I could not follow the argument.  I offered a concrete example and stated that I could not follow the original thesis as applied to the concrete example.  No one took me up on this.   Are you too advanced to stoop to my level of understanding and help me figure out how this abstract reasoning applies to a particular example?   Is the shut down mechanism suggested by Yudkowsky too simple?

Compute is not the limiting factor for mammalian intelligence.  Mammalian brains are organized to maximize communication.  The gray matter, where most compute is done, is mostly on the surface  and the white matter which dominate long range communication, fills the interior, communicating in the third dimension.

If you plot volume of white matter vs. gray matter across the various mammal brains, you find that the volume of white matter grows super linearly with volume of gray matter.   https://www.pnas.org/doi/10.1073/pnas.1716956116

As b... (read more)

1Ted Sanders
Interesting! How do you think this dimension of intelligence should be calculated? Are there any good articles on the subject?

dy/dt = f(y) = m*y whose solution is the compound interest exponential, y = e^(m*t).

Why not estimate m?  

An exactly right law of diminishing returns that lets the system fly through the soft takeoff keyhole is unlikely.

This blog post contains a false dichotomy.  In the equation, m can take any value and there is no special keyhole value, and there is no line between fast and slow.

The description in the subsequent discussion is a distraction.  The posted equation is meaningful only if we have an estimate of the growth rate.  

1ZankerH
That’s Security Through Obscurity. Also, even if we decided we’re suddenly ok with that, it obviously doesn’t scale well to superhuman agents.

You said nothing about positive contributions.  When you throw away the positives, everything is negative.  

Why didn't you also compute the expectation this project contributes towards human flourishing?

If you only count the negative contributions, you will find that the expectation value of everything is negative. 

1WilliamKiely
The main benefits of the project are presumably known to the engineer engaging in it. It was the harm of the project (specifically the harm arising from how the project accelerates AI timelines) that the engineer was skeptical was significant that I wanted to look at more closely to determine whether it was large enough to make it questionable whether engaging in the project was good for the world. Given my finding that a 400-hour ML project (I stipulated the project takes 0.2 years of FTE work) would, via its effects on shortening AI timelines, shorten the lives of existing people by around 17 years, it seems like this harm is not only trivial, but likely dominates the expected value of engaging in the project. This works out to shortening peoples' lives by around 370 hours for every hour worked on the project. If someone thinks the known benefits of working on the project are being drastically underestimated as well, I'd be interested in seeing an analysis of the expected value of those benefits, and in particular and am curious which benefits that person thinks are surprisingly huge. Given the lack of safety angle to the project, I don't see what other benefit (or harm) would come close in magnitude to the harm caused via accelerating AI timelines and increasing extinction risk, but of course would love to hear if you have any idea.
weverka118

The ML engineer is developing an automation technology for coding and is aware of AI risks .  The engineers polite acknowledgment of the concerns is met with your long derivation of how many current and future people she will kill with this.

  Automating an aspect of coding is part of a long history of using computers to help design better computers, starting with Carver Mead's realization that you don't need humans to cut rubylith film to form each transistor.

    You haven't shown an argument that this project will accelerate the scenar... (read more)

3WilliamKiely
Thanks for the response and for the concern. To be clear, the purpose of this post was to explore how much a typical, small AI project would affect AI timelines and AI risk in expectation. It was not intended as a response to the ML engineer, and as such I did not send it or any of its contents to him, nor comment on the quoted thread. I understand how inappropriate it would be to reply to the engineer's polite acknowledgment of the concerns with my long analysis of how many additional people will die in expectation due to the project accelerating AI timelines. I also refrained from linking to the quoted thread specifically because again this post is not a contribution to that discussion. The thread merely inspired me to take a quantitative look at what the expected impacts of a typical ML project actually are. I included the details of the project for context in case others wanted to take them into account when forecasting the impact. I also included Jim and Raymond's comments because this post takes their claims as givens. While I understand the ML engineer may have been skeptical of their claims, and so elaborating on why the project is expected to accelerate AI timelines (and therefore increase AI risk) would be necessary to persuade them that their project is bad for the world, again that aim is outside of the scope of this post. I've edited the heading after "The trigger for this post" from "My response" to "My thoughts on whether small ML projects significantly affect AI timelines" to make clear that the contents are not intended as a response to the ML engineer, but rather are just my thoughts about the claim made by the ML engineer. I assume that heading is what led you to interpret this post as a response to the ML engineer, but if there's anything else that led you to interpret it that way, I'd appreciate you letting me know so I can improve it for others who might read it. Thanks again for reading and offering your thoughts.

The missile gap was a lie by the US Air Force to justify building more nukes, by falsely claiming that the Soviet Union had more nukes than the US

This statement is not supported by the link used as a reference.  Was it a lie?  The reference speaks to failed intelligence and political manipulation using the perceived gap. The phrasing above suggests conspiracy.

2jessicata
This implies an intentional, coordinated falsehood ("kept the American public in the dark", plus denying relevant true information).
4Lone Pine
If the AI is a commercial service like Google search or Wikipedia, that is so embedded into society that we have come to depend on it, or if the AI is seen as national security priority, do you really think we will turn it off?
8JBlack
If there is an AI much smarter than us, then it is almost certainly better at finding ways to render the off switch useless than we are at making sure that it works. For example, by secretly having alternative computing facilities elsewhere without any obvious off switch (distributed computing, datacentres that appear to be doing other things, alternative computing methods based on technology we don't know about). Maybe by acting in a way that we won't want to turn it off, such as by obviously doing everything we want while being very good at keeping the other things it's doing secret until it's too late. In the obvious literal sense of an "off switch", maybe by inducing an employee to replace the switch with a dummy. We don't know and in many ways can't know, because (at some point) it will be better than us at coming up with ideas.
4ZankerH
We have no idea how to make a useful, agent-like general AI that wouldn't want to disable its off switch or otherwise prevent people from using it.

You have more than an order of magnitude scatter in your plot, but you write 3 significant figures to your calculated doubling period. Is this precision of value?  

Also, your black data appears to have something different going on prior to 2008.  It would be worthwhile doing a separate fit to post 2008 data.  Eyeballing it, it is longer than 4 year doubling time.

AI is dependent on humans.  It gets power and data from humans and it cannot go on without humans.   We don't trade with it, we dictate terms. 

Do we fear a world where we have turned over mining, production and powering everything to the AI. Getting there would take a lot more than self amplifying feedback loop of a machine rewriting its own code. 

When I was doing runs in the dozens of miles, I found it better to cache water ahead of time at the ten mile points.  On a hot day, you need more water than you can comfortably carry.

4jefftk
The kind of running I'm thinking about here is much shorter stretches as you go about your daily life: short enough that you don't even break a sweat (more).
weverka*30

Ok, I could be that someone. here goes.  You and the paper author suggest a heat engine.  That needs a cold side and a hot side.  We build a heat engine where the hot side is kept hot by the incoming energy as described in this paper.  The cold side is a surface we have in radiative communication with the 3 degrees Kelvin temperature of deep space.  In order to keep the cold side from melting, we need to keep it below a few thousand degrees, so we have to make it really large so that it can still radiate the energy. 

From here,... (read more)

2Daniel Kokotajlo
I'm still interested in this question. I don't think you really did what I asked -- it seems like you were thinking 'how can I convince him that this is impossible' not 'how can I find a way to build a dyson swarm.' I'm interested in both but was hoping to have someone with more engineering and physics background than me take a stab at the latter. My current understanding of the situation is: There's no reason why we can't concentrate enough energy on the surface of Mercury, given enough orbiting solar panels and lasers; the problem instead seems to be that we need to avoid melting all the equipment on the surface. Or, in other words, the maximum amount of material we can launch off Mercury per second is limited by the maximum amount of heat that can be radiated outwards from Mercury (for a given operating temperature of the local equipment?) And you are claiming that this amount of heat radiation ability, for radiators only the size of Mercury's surface, is OOMs too small to enable dyson swarm construction. Is this right?

A billion times the energy flux from the surface of the sun, over any extended area is a lot to deal with.  It is hard to take this proposal seriously.

1Daniel Kokotajlo
Yeah, though not for the reason you originally said. I think I'd like to see someone make a revised proposal that addresses the thermal management problem, which does indeed seem to be a tricky though perhaps not insoluble problem. 
weverka121

For the record, I find that scientists make such errors routinely.  In public conferences when optical scientists propose systems that violate the constant radiance theorem, I have no trouble standing up and saying so.  It happens often enough that when I see a scientist propose such a system, It does not diminish my opinion of that scientist.  I have fallen into this trap myself at times.  Making this error should not be a source of embarrassment.  

either way, you are claiming Sandberg, a physicist who works with thermodynamic stu

... (read more)

I stand corrected.  Please forgive me.

Last summer I went on a week long backpacking trip where we had to carry out all our used toilet paper.  

This year, I got this bidet for Christmas:  https://www.garagegrowngear.com/products/portable-bidet-by-culoclean

You could carry one with you so you are no longer reliant on having them provided for you.

2philh
I have something similar, with a long neck which seems like it would be more convenient. I got it to see if I'd feel like getting an actual bidet, and decided I probably like it better than one of those, at least a reasonably cheap one. I can aim it easily, control the power, and if I want heated water I can fill it with that. If I didn't have a sink that I could reach while sitting on the toilet I might look into actual bidets more. I don't use it every time I poop, but it's a minor quality of life improvement.
5Richard_Kennaway
Those are also about 1/10 the cost of the things that you install in your bathroom, so a low-cost way of experimenting, before deciding if an installation would be worth it. There are Youtube videos explaining how to use them -- look for "backpacking bidet".

How much extra energy external energy is required to get an energy flux on Mercury of a billion times that leaving the sun?  I have an idea, but my statmech is rusty. (the fourth root of a billion?)

And do we have to receive the energy and convert it to useful work with 99.999999999% efficiency to avoid melting the apparatus on Mercury?

2Daniel Kokotajlo
I have no idea, I never took the relevant physics classes. For concreteness, suppose we do something like this: We have lots of solar panels orbiting the sun. They collect electricity (producing plenty of waste heat etc. in the process, they aren't 100% efficient) and then send it to lasers, which beam it at Mercury (producing plenty more waste heat etc. in the process, they aren't 100% efficient either). Let's suppose the efficiency is 10% in each case, for a total efficiency of 1%. So that means that if you completely surrounded the sun with a swarm of these things, you could get approximately 1% of the total power output of the sun concentrated down on Mercury in particular, in the form of laser beams. What's wrong with this plan? As far as I can tell it couldn't be used to make infinite power, because of the aforementioned efficiency losses. To answer your second question: Also an interesting objection! I agree melting the machinery is a problem & the authors should take that into account. I wonder what they'd say about it & hope they respond.

DK> "I don't see how they are violating the second law of thermodynamics"

Take a large body C, and a small body H.  Collect the thermal radiation from C in some manner and deposit that energy on H.  The power density emitted from C grows with temperature.  The temperature of H grows with the power density deposited.  If, without adding external energy, we concentrate the power density from the large body C to a higher power density on the small body H, H gets hotter than C.  We may then use a heat engine between H an C to make fr... (read more)

4Daniel Kokotajlo
But in laboratory experiments, haven't we produced temperatures greater than that of the surface of the sun? A quick google seems to confirm this. So, it is possible to take the power of the sun and concentrate it to a point H so as to make that point much hotter than the sun. (Since I assume that whatever experiment we ran, could have been run powered by solar panels if we wanted to) I think the key idea here is that we can add external energy -- specifically, we can lose energy. We collect X amount of energy from the sun, and use X/100 of it to heat our desired H, at the expense of the remaining 99X/100. If our scheme does something like this then no perpetual motion or infinite power generation is entailed.  

thanks for showing that Gwern's statement that I am "bad at reading" is misplaced. 

The conservation of etendué is merely a particular version of the second law of thermodynamics.  Now, You are trying to invoke a multistep photovoltaic/microwave/rectenna method of concentrating energy, but you are still violating the second law of thermodynamics.  

If one could concentrate the energy as you propose, one could build a perpetual motion machine.

2Daniel Kokotajlo
I don't see how they are violating the second law of thermodynamics -- "all that conversion back and forth induces losses." They are concentrating some of the power of the Sun in one small point, at the expense of further dissipating the rest of the power. No?

>Kokotajlo writes:Wouldn't that be enough to melt, and then evaporate, the entirety of Mercury within a few hours? After all isn't that what would happen if you dropped Mercury into the Sun?

How do you get hours?  

3Daniel Kokotajlo
I didn't do any calculation at all, I just visualized Mercury falling into the sun lol. Not the most scientific method. 

The sun emits light because it is hot.  You can't concentrate thermal emission to be brighter than the source.  (if you could, you could build a perpetual motion machine). 

Eternity in Six Hours describes very large lightweight mirrors concentrating solar radiation onto planet Mercury.

The most power you could deliver from the sun to Mercury is the power of the sun times the square of the ratio of the radius of Mercury to the radius of the sun.

The total solar output is 4*10^26 Watts.  The ratio of the sun's radius to that of mercury is half a million.  So you can focus about 10^15 Watts onto Mercury at most.

Figure 2 of Eternity in Six Hours projects getting 10^24 Watts to do the job.

-7gwern
8Arenamontanus
We do not assume mirrors. As you say, there are big limits due to conservation of etendué. We are assuming (if I remember right) photovoltaic conversion into electricity and/or microwave beams received by rectennas. Now, all that conversion back and forth induces losses, but they are not orders of magnitude large. In the years since we wrote that paper I have become much more fond of solar thermal conversion (use the whole spectrum rather than just part of it), and lightweight statite-style foil Dyson swarms rather than heavier collectors. The solar thermal conversion doesn't change things much (but allows for a more clean-cut analysis of entropy and efficiency; see Badescu's work). The statite style however reduces the material requirements many orders of magnitude: Mercury is safe, I only need the biggest asteroids.  Still, detailed modelling of the actual raw material conversion process would be nice. My main headache is not so much the energy input/waste heat removal (although they are by no means trivial and may slow things down for too concentrated mining operations - another reason to do it in the asteroid belt in many places), but how to solve the operations management problem of how many units of machine X to build at time t. Would love to do this in more detail!
4Daniel Kokotajlo
Awesome critique, thanks! I'm going to email the authors and ask what they think of this. I'll credit you of course.
weverka*20

I have read Eternity in Six Hours and I can say that it violates the Second Law of Thermodynamics through the violation of the Constant Radiance Theorem.  The Power density they deliver to Mercury exceeds the power density of radiation exiting the sun by 6 orders of magnitude!

6gwern
I don't follow. What does power density have to do with anything and how can any merely geometrical theorem matter? You are concentrating the power of the sun by the megaengineering (solar panels in this case), so the density can be whatever you want to pay for. (My CPU chip has much higher power density than the equivalent square inches of Earth's atmosphere receiving sunlight, but no one says it 'violates the laws of thermodynamics'.) Surely only the total power matters.

You should have a look at the conference on retrocausation.  And it would also be valuable to look at Garret Moddel's experiments on the subject.

You drew a right turn, the post is asking about a left turn.

Yes it is.  When I took Feynman's class on computation, he presented an argument on Landauer's limit.  It involved a multi-well quantum potential where the barrier between the wells was slowly lowered and the well depths adjusted.  During the argument, one of the students asked if he had not just introduced a Maxwell's demon.  Feynman got very defensive.

The normal distribution is baked into the scoring of intelligence tests. I do not know what the distribution of raw scores looks like, but the calculation of the IQ score is done by transforming the raw scores to make them normally distributed with a mean of 100. There is surely not enough data to do this transformation out to ±6 SD.

Answer by weverka10

Is it likely to do more good than harm?

Most AI safety criticisms carry a multitude of implicite assumptions.  This argument grants the assumption and attacks the wrong strategy.
  We are better off improving a single high-level AI than making a second one.  There is not battle between multiple high-level AIs if there is only one.

>What would you have said? 

Your comment is stronger without this sentence.

weverkaΩ0-40

Gwern asks"Why would you do that and ignore (mini literature review follows):"  

Thompson did not ignore the papers Gwern cites.  A number of them are in Thompson's tables comparing prior work on scaling.  Did Gwern tweet this criticism without even reading Thompson's paper?

5gwern
I did read it, and he did ignore them. Do you really think I criticized a paper publicly in harsh terms for not citing 12 different papers without even checking the bibliography or C-fing the titles/authors? Please look at the first 2020 paper version I was criticizing in 16 July 2020, when I wrote that comment, and don't lazily misread the version posted 2 years later on 27 July 2022 which, not being a time traveler, I obviously could not have read or have been referring to (and which may well have included those refs because of my comments there & elsewhere). (Not that I am impressed by their round 2 stuff which they tacked on - but at least now they acknowledge that prior scaling research exists and try to defend their very different approach at all.)
weverka1-2

Reliable?  Your hard disk will be unreadable before long, while the human brain has developed ways to pass information down over generations.

>I dont think you are calibrated properly about the ideas that are most commonly shared in the LW community. 

This is chastising him for failure to abide by groupthink.
The rest of your comment makes a point that is undermined by this statement.

-2Cervera
I dont think I wrote that statement with that particular intention in mind.  I'm not trying to imply he is wrong because he doenst know our "groupthink" I was just generally annoyed at how he started the post, so i wanted to be reasonably civil, but a bit mean.  Thanks for noticing, I'm not convinced I should have refrained from that particular comment tho. What would you have said? 
Answer by weverka-1-1

FSM

 

 

https://www.spaghettimonster.org/2015/08/mars-sighting/

weverka576

I must disagree.  I roasted a large plane for Thanksgiving yesterday and it was incomparable to a bird.  For tips on brining your plane, see here: https://en.wikipedia.org/wiki/US_Airways_Flight_1549

weverka3-1

No, Humans do not satisfy this assumptions adopted here, unless you make this more specific.

The definition of Generalize is given above as: "Generalizes, i.e., performs well in new domains, which were not optimized for during training, with no domain-specific tuning".

Whether you think humans do this depends on what you take for "new domains" and "perform well".  

Humans taught to crawl on hardwood floors can crawl on carpeted floors.  Humans taught to hunt fly larva will need further training to hunt big game.

When we substitute credentials for reason, we get nowhere.

  1.  Open the box when and iff you need the utilon.  This simple solution gets you the maximum utilons if you need a utilon and none if you don't need a utilon.
Answer by weverka10

This is difficult for people with no ML background.  The trouble with this is that one first has to explain timelines. Then explain what averages and ranges most researchers in the field maintain, and then explain why some discount that in favor of short AI timelines.  That is a long arc for a skeptical person.

Aren't we all skeptical people?  Carl Sagan said that extraordinary claims require extraordinary evidence.  Explaining a short timeline is a heavy lift by its very nature.