All of evand's Comments + Replies

evand40

Have you considered looking at the old Foresight Exchange / Ideosphere data? It should be available, and it's old enough that there might be a useful number of long-term forecasts there.

http://www.ideosphere.com/

2niplav
Thank you! If I pay more attention to my forecast dataset again I'll first add it there.
evand53

Yep, those are extremely different drugs with very different effects. They do in fact take a while and the effects can be much more subtle.

2Adam Zerner
Gotcha. In retrospect that actually makes a lot of sense, since kids in school are known to take Adderall for a night of studying.
evand80

having read a few books about ADHD, my best guess is that I have (and had) a moderate case of it.

when I'm behind the wheel of a car and X = "the stoplight in front of me" and Y = "that new Indian restaurant", it is bad

Which is one of the reasons why I don't drive.

If your ADHD is interfering with your driving, that does not sound like a moderate case!

But option #3 is much better: take medication for six weeks and see what happens.

My expectation is that you will likely get a lot of information from trying the meds for 6 days; 6 weeks sounds like a very long ... (read more)

8Adam Zerner
It shouldn't really matter whether it's one day or six weeks, but knowing now that it could be as little as one day makes it very salient to me how big of an error I made. It's plausible that medication would significantly improve my life, and it only would have taken (approximately) a day or so, a trip or two to the doctor, and a few dollars to find out.
2cata
I tried Adderall and Ritalin each just for one day and it was totally clear to me based on that that I wasn't interested in taking them on a regular basis.
2Adam Zerner
Oh, maybe. That's how long my psychiatrist had me try SSRIs before checking in so I chose that duration as a guesstimate.
Answer by evand20

Bitcoin Hivemind (nee Truthcoin) is the authority on doing this in a truly decentralized fashion. The original whitepaper is well worth a read. The fundamental insight: it's easier to coordinate on the truth than on something else; incentivizing defection to the truth works well.

Judges have reputation. If you judge against the consensus, you lose reputation, and the other (consensus) judges gain it. The amount you lose depends on the consensus: being the lone dissenter has a very small cost (mistakes have costs, but small ones), but being part of a near-ma... (read more)

evand20

Thank you!

I went ahead and created the 2024 version of one of the questions. If you're looking for high-liquidity questions to include, which seems like a good way to avoid false alarms / pranks, this one seems like a good inclusion.

There are a bunch of lower-liquidity questions; including a mix of those with some majority-rule type logic might or might not be worth it.

evand20

Thank you! Much to think about, but later...

If there are a large number of true-but-not-publicly-proven statements, does that impose a large computational cost on the market making mechanism?

I expect that the computers running this system might have to be fairly beefy, but they're only checking proofs.

They're not, though. They're making markets on all the interrelated statements. How do they know when they're done exhausting the standing limit orders and AMM liquidity pools? My working assumption is that this is equivalent to a full Bayesian network and ex... (read more)

2DaemonicSigil
There's no reason I can't just say: "I'm going to implement the rules listed above, and if anyone else wants to be a market-maker, they can go ahead and do that". The rules do prevent me from losing money (other than cases where I decide to subsidize a market). I think in some sense, this kind of market does run more on letting traders make their own deals, and less on each and every asset having a well-defined price at a given point in time. (Though market makers can maintain networks tracking which statements are equivalent, which helps with combining the limit orders on the different versions of the statement.) Good question, I'm still not sure how to handle this.
evand120

This is very neat work, thank you. One of those delightful things that seems obvious in retrospect, but that I've never seen expressed like this before. A few questions, or maybe implementation details that aren't obvious:

For complicated proofs, the fully formally verified statement all the way back to axioms might be very long. In practice, do we end up with markets for all of those? Do they each need liquidity from an automated market maker? Presumably not if you're starting from axioms and building a full proof, and that applies to implications and conj... (read more)

6DaemonicSigil
Thanks! I'll try to answer some of your questions. In practice, I think not. I expect interesting conjectures to end up thickly traded such that the concept of a market price makes sense. For most other propositions, I expect nobody to hold any shares in them at all most of the time. To take the example of Alice and ¬A, if we suppose that Alice is the only one with a proof of A, then everyone else has to tie up capital to invest in A, but Alice can just create shares of ¬A for free, and then sell them for profit. If A is an interesting conjecture, then Alice can sell on the open market and Bob can buy from the open market. If A is a niche thing, only useful as a step in the proof of B, then Bob might trade directly with Alice and the price will be determined by how hard Bob would find it to come up with a proof of A himself, or to find someone other than Alice to trade with. So Alice does have to keep her proof secret for as long as she wants to profit from it. Different axiom systems can encode each other. So long as we start from an axiom system that can fully describe Turing machines, statements about other axiom systems should be possible to encode as statements about certain specific Turing machines. The root system has to be consistent. If it's inconsistent, then people can just money pump the market mechanism. But systems built on top need not be consistent. If people try to prove things about some pet system that they've encoded, and that system happens to be inconsistent, that means that they're surprised by the outcomes of some markets they're trading on, but it doesn't blow up the entire rest of the market. I expect that the computers running this system might have to be fairly beefy, but they're only checking proofs. Coming up with proofs is a much harder task, and it's the traders who are responsible for that. Sorry if this is a bad answer, I'm not 100% sure what you're getting at here. For thickly traded propositions, I can make money by investing
1ProgramCrafter
Actually, I don't think that creates any problem? Just create shares of "ZFC axioms" and "not ZFC axioms" via logical share splitting. If you are unable to sell "not ZFC axioms", that only means that price of one main share is $1 (though, it's likely possible to prove something fun if we take these axioms as false).
evand51

My background: educated amateur. I can design simple to not-quite-simple analog circuits and have taken ordinary but fiddly material property measurements with electronics test equipment and gotten industrially-useful results.

One person alleges an online rumor that poorly connected electrical leads can produce the same graph.  Is that a conventional view?

I'm not seeing it. With a bad enough setup, poor technique can do almost anything. I'm not seeing the authors as that awful, though. I don't think they're immune from mistakes, but I give low odds on ... (read more)

evand30

The comparison between the calculations saying igniting the atmosphere was impossible and the catastrophic mistake on Castle Bravo is apposite as the initial calculations for both were done by the same people at the same gathering!

One out of two isn't bad, right?

https://twitter.com/tobyordoxford/status/1659659089658388545

evand83

Of course a superintelligence could read your keys off your computer's power light, if it found it worthwhile. Most of the time it would not need to, it would find easier ways to do whatever humans do by pressing keys. Or make the human press the keys.

FYI, the referenced thing is not about what keys are being pressed on a keyboard, it's about extracting the secret keys used for encryption or authentication. You're using the wrong meaning of "keys".

2Shmi
Yeah, good point. Would be easy in both meanings.
evand60

If you think the true likelihood is 10%, and are being offered odds of 50:1 on the bet, then the Kelly Criterion suggests you should be about 8% of your bankroll. For various reasons (mostly human fallibility and an asymmetry in the curve of the Kelly utility), lots of people recommend betting at fractions of the Kelly amount. So someone in the position you suggest might reasonably wish to be something like $2-5k per $100k of bankroll. That strategy, your proposed credences, and the behavior observed so far would imply a bankroll of a few hundred thousand ... (read more)

Answer by evand33

"Now" is the time at which you can make interventions. Subjective experience lines up with that because it can't be casually compatible with being in the future, and it maximizes the info available to make the decision with. Or rather, approximately maximizes subject to processing constraints: things get weird if you start really trying to ask whether "now" is "now" or "100 ms ago".

That's sort of an answer that seems like it depends on a concept of free will, though. To which my personal favorite response is... how good is your understanding of counterfact... (read more)

1MvB
I could not really make sense of your comment, though I had actually done what you proposed a couple of years ago, until I had read Lucius Bushnaq‘s comment. Did that imply what you were trying to tell me or is there another aspect to what you call an intuitive understanding?
evand53

My concern with conflating those two definitions of alignment is largely with the degree of reliability that's relevant.

The definition "does what the developer wanted" seems like it could cash out as something like "x% of the responses are good". So, if 99.7% of responses are "good", it's "99.7% aligned". You could even strengthen that as something like "99.7% aligned against adversarial prompting".

On the other hand, from a safety perspective, the relevant metric is something more like "probabilistic confidence that it's aligned against any input". So "99.... (read more)

evand31

If there are reasons to refuse bets in general, that apply to the LessWrong community in aggregate, something has gone horribly horribly wrong.

No one is requiring you personally to participate, and I doubt anyone here is going to judge you for reluctance to engage in bets with people from the Internet who you don't know. Certainly I wouldn't. But if no one took up this bet, it would have a meaningful impact on my view of the community as a whole.

8Jiro
It is my opinion that for the LessWrong community in aggregate, something has gone horribly horribly wrong. At a minimum, LWers should have 1) observed that normies don't bet like this and 2) applied Chesterton's Fence. It's often hard to give an exhaustive, bulletproof, explanation of why normies act in some way that does, in fact, make sense as a way to act. Rationalists have a habit of saying "well, I don't see a rational reason for X, so I can just discard X". That's what Chesterton's Fence is about avoiding.
evand20

I don't know how it prevents us from dying either! I don't have a plan that accomplishes that; I don't think anyone else does either. If I did, I promise I'd be trying to explain it.

That said, I think there are pieces of plans that might help buy time, or might combine with other pieces to do something more useful. For example, we could implement regulations that take effect above a certain model size or training effort. Or that prevent putting too many flops worth of compute in one tightly-coupled cluster.

One problem with implementing those regulations is... (read more)

evand20

I think neither. Or rather, I support it, but that's not quite what I had in mind with the above comment, unless there's specific stuff they're doing that I'm not aware of. (Which is entirely possible; I'm following this work only loosely, and not in detail. If I'm missing something, I would be very grateful for more specific links to stuff I should be reading. Git links to usable software packages would be great.)

What I'm looking for mostly, at the moment, is software tools that could be put to use. A library, a tutorial, a guide for how to incorporate th... (read more)

2Garrett Baker
I’m pretty confident the primary labs keep track of the number of flops used to train their models. I also don’t know how such a tool would prevent us all from dying.
evand105

One thing I'd like to see more of: attempts at voluntary compliance with proposed plans, and libraries and tools to support that.

I've seen suggestions to limit the compute power used on large training runs. Sounds great; might or might not be the answer, but if folks want to give it a try, let's help them. Where are the libraries that make it super easy to report the compute power used on a training run? To show a Merkle tree of what other models or input data that training run depends on? (Or, if extinction risk isn't your highest priority, to report whic... (read more)

2Garrett Baker
Do you disagree with Apollo or ARC evals's approaches to the voluntary compliance solutions?
evand20

Aren't the other used cars available nearby, and the potential other buyers should you walk away, relevant to that negotiation?

1Isaac King
Yeah, that could be relevant, but the system might be able to factor that in. For example, maybe it could be modeled as decreasing the maximum price the buyer is willing to pay (since they know they can have more attempts), and the system factors that in. I want it to be able to handle a broad range of real-world negotiations, so the exact details ideally shouldn't matter that much.
evand40

This was fantastic; thank you! I still haven't quite figured it out, I'll definitely have to watch it a second time (or at least some parts of it).

I think some sort of improved interface for your math annotations and diagrams would be a big benefit, whether that's a drawing tablet or typing out some LaTeX or something else.

I think the section on induction heads and how they work could have used a bit more depth. Maybe a couple more examples, maybe some additional demos of how to play around with PySvelte, maybe something else. That's the section I had the ... (read more)

4Neel Nanda
I appreciate the feedback! I have since bought a graphics tablet :) If you want to explore induction heads more, you may enjoy this tutorial Any papers you're struggling to find?
evand20

Yes, if Omega accurately simulates me and wants me to be wrong, Omega wins. But why do I need to get the answer exactly "right"? What does it matter if I'm slightly off?

This would be a (very slightly) more interesting problem if Omega was offering a bet or a reward and my goal was to maximize reward or utility or whatever. It sure looks like for this setup, combined with a non-adversarial reward schedule, I can get arbitrarily close to maximizing the reward.

2Chris_Leong
This problem is merely meant as an example of a phenomenon.
evand30

This feel reminiscent of:

If the human brain were so simple that we could understand it, we would be so simple that we couldn’t.

And while it's a well-constructed pithy quote, I don't think it's true. Can a system understand itself? Can a quining computer program exist? Where is the line between being able to recite itself and understand itself?

You need a model above some threshold of capability at which it can provide useful interpretations, yes, but I don't see any obvious reason why that threshold would move up with the size of the model under interpretat

... (read more)
evand20

I think that sounds about right. Collecting the arguments in one place is definitely helpful, and I think they carry some weight as initial heuristics, which this post helps clarify.

But I also think the technical arguments should (mostly) screen off the heuristics; the heuristics are better for evaluating whether it's worth paying attention to the details. By the time you're having a long debate, it's better to spend (at least some) time looking instead of continuing to rely on the heuristics. Rhymes with Argument Screens Off Authority. (And in both cases, only mostly screens off.)

9FinalFormal2
I think you're overestimating the strength of the arguments and underestimating the strength of the heuristic. All the Marxist arguments for why capitalism would collapse were probably very strong and intuitive, but they lost to the law of straight lines. I think you have to imagine yourself in that position and think about how you would feel and think about the problem.
evand20

That's the point. SpaceX can afford to fail at this; the decision makers know it. Eliezer can afford to fail at tweet writing and knows it. So they naturally ratchet up the difficulty of the problem until they're working on problems that maximize their expected return (in utility, not necessarily dollars). At least approximately. And then fail sometimes.

Or, for the trapeze artist... how long do they keep practicing? Do they do the no-net route when they estimate their odds of failure are 1/100? 1/10,000? 1e-6? They don't push them to zero, at some point th... (read more)

1dr_s
Well, to be fair, the post is making the point that perhaps they can afford less than they thought. They completely ignored the effects their failure would have on the surrounding communities (which reeks highly of conceit on their part) and now they're paying the price with the risk of a disproportionate crackdown. It'll cost them more than they expected for sure. You're right, but the analogy is also saying I think that if we were capable enough to one-shot AGI (which according to EY we need to), then we surely would be capable enough to also very cheaply one-shot a Starship launch, because it's a simpler problem. Failure may be a good teacher, but it's not a free one. If you're competent enough to one-shot things with only a tiny bit of additional effort, you do it. Having this failure rate instead shows that you're already straining yourself at the very limit of what's possible, and the very limit is apparently... launching big rockets. Which while awesome in a general sense is really, really child's play compared to getting superhuman AGI right, and on that estimate I do agree with Yud. I would add that a huge part of solving alignment requires being keenly aware of and caring about human values in general, and in that sense, the sort of mindset that leads to not foreseeing or giving a damn about how pissed off people would be by clouds of launchpad dust in their towns really isn't the culture you want to bring into AGI creation.
evand61

Perhaps I'm missing something obvious, and just continuing the misunderstanding, but...

It seems to me that if you're the sort of thing capable of one-shotting Starship launches, you don't just hang around doing so. You tackle harder problems. The basic Umeshism: if you're not failing sometimes, you're not trying hard enough problems.

Even the "existential" risk of SpaceX getting permanently and entirely shut down, or just Starship getting shut down, is much closer in magnitude to the payoff than is the case in AI risk scenarios.

Some problems are well calibr... (read more)

1dr_s
Well, or you're trying problems that you can't afford to fail at. If a trapeze artist doesn't fall off 50% of his no-net performances, should they try a harder performance?
evand174

Taboo "rationally".

I think the question you want is more like: "how can one have well-calibrated strong probabilities?". Or maybe "correct". I don't think you need the word "rationally" here, and it's almost never helpful at the object level -- it's a tool for meta-level discussions, training habits, discussing patterns, and so on.

To answer the object-level question... well, do you have well-calibrated beliefs in other domains? Did you test that? What do you think you know about your belief calibration, and how do you think you know it?

Personally, I think ... (read more)

7Shmi
I used that description very much intentionally. As in, use your best Bayesian estimate, formal or intuitive. As to the object level, "pre-paradigmatic" is essential. The field is full of unknown unknowns. Or, as you say "conditional on no near-term giant surprises..." -- and there have been "giant surprises" in both directions recently, and likely will be more soon. It seems folly to be very confident in any specific outcome at this point.
evand20

I'm still uncertain how I feel about a lot of the details on this (and am enough of a lurker rather than poster that I suspect it's not worth my time to figure that out / write it publicly), but I just wanted to say that I think this is an extremely good thing to include:

I will probably build something that let's people Opt Into More Said. I think it's fairly likely the mod team will probably generally do some more heavier handed moderation in the nearish future, and I think a reasonable countermeasure to build, to alleviate some downsides of this, is to a

... (read more)
evand20

i could believe this number’s within 3 orders of magnitude of truth, which is probably good enough for the point of this article

It's not. As best I can tell it's off by more like 4+ OOM. A very quick search suggests actual usage was maybe more like 1 GWh. Back of the envelope guess: thousands of GPUs, thousands of hours, < 1kW/GPU, a few GWh.

https://www.theregister.com/2020/11/04/gpt3_carbon_footprint_estimate/

https://www.numenta.com/blog/2022/05/24/ai-is-harming-our-planet/

i am a little surprised if you just took it 100% at face value.

Same. That doesn'... (read more)

1Logan Zoellner
Yeah, I should have double-checked. Editing post to reflect the correct values.  Does not affect the "two decades" bottom line conclusion.
evand40

I feel like the argumentation here is kinda misleading.

Here's a pattern that doesn't work very well: a tragedy catches our attention, we point to statistics to show it's an example of a distressingly common problem, and we propose laws to address the issue.

The post promises to discuss a pattern. It's obviously a culture-war-relevant pattern, and I can see a pretty good case for it being one where all the examples are going to be at least culture-war-adjacent. It's an important pattern, if true, so far that seems justified and worth worrying about how to im... (read more)

7Ben
An example I think fits the pattern: A serving London police officer used his badge and handcuffs to detain a young woman (Sarah Everard), he then raped her, murdered her and set fire to her body. Since this absolutely awful case caught headlines various changes have been made to police policy. Several hundred officers are being investigated for crimes that (for some reason) were not previously investigated [1]. And the government brought in new laws against street harassment [2].  I support the changes (as far as I understand them) and I really don't understand why the police were not already investigating officers accused of serious sex crimes. But, it is still clear that the case of Sarah Everard is not a typical example of "street harassment" (its not an example at all), so mentioning her case in that discussion feels vaguely dishonest. [1] https://www.bbc.co.uk/news/uk-63162881 [Police investigations] [2] https://www.bbc.co.uk/news/uk-politics-65065154 [Street harassment bill linked to the murder] [3] https://www.bbc.co.uk/news/topics/c8657zxk82wt?page=2 [A list of everything the BBC think links  to the murder]
8jefftk
Fair! I was thinking of it as three instances, where we have three different statistics being used to support policy changes that don't fit their underlying causes, but you're right that they're all on gun policy. Thinking a bit, here are two other cases where you see similar things: * Transportation, with people using "road deaths" to push for making buses safer after high-fatality crashes, even when buses are already far safer than cars and it's cars that are the reason for high road deaths (even adjusting for mode share) * Domestic hunger, where the examples people give are people starving, but the statistics are for food insecurity which is defined very broadly. I think it's likely that this is pushing us towards the wrong policies here but would need to look into this more.
evand62

Something like an h-index might be better than a total.

4Vladimir_Nesov
In some ways this sounds better than either my proposal or Raemon's. But there is still the spurious upvoting issue, so a metric should be able to not get too excited about a few highly upvoted things.
evand21

For an extremely brief summary of the problem, I like this from Zvi:

The core reason we so often have seen creation of things we value win over destruction is, once again, that most of the optimization pressure by strong intelligences was pointing in that directly, that it was coming from humans, and the tools weren’t applying intelligence or optimization pressure. That’s about to change.

https://thezvi.wordpress.com/2023/03/28/response-to-tyler-cowens-existential-risk-ai-and-the-inevitable-turn-in-human-history/

evand53

Intelligence has no upper limit, instead of diminishing sharply in relative utility

It seems to me that there is a large space of intermediate claims that I interpret the letter as falling into. Namely, that if there exists an upper limit to intelligence, or a point at which the utility diminishes enough to not be worth throwing more compute cycles at it, humans are not yet approaching that limit. Returns can diminish for a long time while still being worth pursuing.

you have NO EVIDENCE that AGI is hostile or is as capable as you claim or support for any of

... (read more)
8[anonymous]
For the first, the succinct argument for correctness is to consider the details of key barriers. Imagine the machine is trying to convince a human it doesn't know to do something in favor of the . machine. More and more intelligence you can model as allowing the machine to consider an ever wider search space of possible hidden state for the human or messages it can emit. But none of this does more than marginally improve the pSuccess. For this task I will claim the odds of success with human intelligence are 10 percent, and with infinite intelligence, 20 percent. It takes logarithmically more compute to approach 20 percent. Either way the machine is probably going to fail. I am claiming there are thousands of real world tasks on the way to conquering the planet with such high pFail. The way the machine wins is to have overwhelming force, same way you win any war. And that real world force has a bunch of barriers to obtaining. For the second, again, debates are one thing. Taking costly action (delays, nuclear war) is another. I am saying it is irrational to take costly actions without direct evidence.
evand20

It's always possible to fudge the numbers and decide that some values are unimportant and some are super important and lo and behold, the calculation turns in your favour! In the end it's no better than deontology or simply saying "I think this is good"; there is no point trying to vest it with a semblance of objectivity that just isn't there.

Is this not simply the fallacy of gray?

As saying goes, it's easy to lie with statistics, but even easier to lie without them. Certainly you can fudge the numbers to make the result say anything, but if you show your work then the fudging gets more obvious.

6dr_s
I agree that laying out your thinking at least forces you to specifically elucidate your values. That way people can criticise the precise assumptions they disagree with, and you can't easily back out of them. I don't think the "lying with statistics" saying applies in its original meaning because really this is entirely about subjective terminal values. "Because I like it this way" is essentially what it boils down to no matter how you slice it.
evand20

I think you missed a follow-on edit:

"Let’s unpack what that 0.36 bits means,"

2johnswentworth
Ah, yup, thankyou!
evand20

Metrics are only useful for comparison if they're accepted by a sufficient broad cross section of society. Since nearly everyone engages in discourse.

I note that "sufficiently broad" might mean something like "most of LessWrong users" or "most people attending this [set of] meetups". Just as communication is targeted at a particular audience, discourse norms are (presumably) intended for a specific context. That context probably includes things like intended users, audience, goals, and so on. I doubt "rationalist discourse" norms will align well with "tele... (read more)

evand7530

"Physicist motors" makes little sense because that position won out so completely that the alternative is not readily available when we think about "motor design". But this was not always so! For a long time, wind mills and water wheels were based on intuition.

But in fact one can apply math and physics and take a "physicist motors" approach to motor design, which we see appearing in the 18th and 19th centuries. We see huge improvements in the efficiency of things like water wheels, the invention of gas thermodynamics, steam engines, and so on, playing a ma... (read more)

-1M. Y. Zuo
If discourse has such high dimensionality, compared to motors, how can anyone be confident that any progress has been made at all? Now, or ever?
evand20

Obvious kinds of humans include:

Dead humans. (Who didn't manage to leave the coins to their heirs.)

Cryonically preserved humans hoping to use them later. (Including an obvious specific candidate.)

Humans optimistic enough about Bitcoin to think current prices are too low. (We know Nakamoto had resources, so it seems a safe bet that they could keep living on ordinary means for now.)

And the obvious: you don't know that all of Nakamoto's coins fit the standard assumed profile. It's entirely possible they intentionally mined some with the regular setup and are spending a few from that pool.

evand20

The advanced answer to this is to create conditional prediction markets. For example: a market for whether or not the Bank of Japan implements a policy, a market for the future GDP or inflation rate of Japan (or whatever your preferred metric is), and a conditional market for (GDP given policy) and (GDP given no policy).

Then people can make conditional bets as desired, and you can report your track record, and so on. Without a prediction market you can't, in general, solve the problem of "how good is this prediction track record really" except by looking at it in detail and making judgment calls.

evand10

I hope you have renter's insurance, knowledge of a couple evacuation routes, and backups for any important data and papers and such.

evand20

I'm not aware of any legal implications in the US. US gambling laws basically only apply when there is a "house" taking a cut or betting to their own advantage or similar. Bets between friends where someone wins the whole stake are permitted.

As for the shady implications... spend more time hanging out with aspiring rationalists and their ilk?

evand00

The richer structure you seek for those two coins is your distribution over their probabilities. They're both 50% likely to come up heads, given the information you have. You should be willing to make exactly the same bets about them, assuming the person offering you the bet has no more information than you do. However, if you flip each coin once and observe the results, your new probability estimate for next flips are now different.

For example, for the second coin you might have a uniform distribution (ignorance prior) over the set of all possible probabilities. In that case, if you observe a single flip that comes up heads, your probability that the next flip will be heads is now 2/3.

0Lumifer
Yes, I understand that. This subthread started when cousin_it said at which point I objected.
evand10

Well, in general, I'd say achieving that reliability through redundant means is totally reasonable, whether in engineering or people-based systems.

At a component level? Lots of structural components, for example. Airplane wings stay attached at fairly high reliability, and my impression is that while there is plenty of margin in the strength of the attachment, it's not like the underlying bolts are being replaced because they failed with any regularity.

I remember an aerospace discussion about a component (a pressure switch, I think?). NASA wanted documenta... (read more)

evand90

You might also want a mechanism to handle "staples" that individuals want. I have a few foods / ingredients I like to keep on hand at all times, and be able to rely on having. I'd have no objections to other people eating them, but if they did I'd want them to take responsibility for never leaving the house in a state of "no X on hand".

evand20

Those numbers sound like reasonable estimates and goals. Having taught classes at TechShop, that first handful of hours is important. 20 hours of welding instruction ought to be enough that you know whether you like it and can build some useful things, but probably not enough to get even an intro-level job. It should give you a clue as to whether signing up for a community college class is a good idea or not.

Also I'm really confused by your inclusion of EE in that list; I'd have put it on the other one.

evand80

However, I'm skeptical of systems that require 99.99% reliability to work. Heuristically, I expect complex systems to be stable only if they are highly fault-tolerant and degrade gracefully.

On the other hand... look at what happens when you simply demand that level of reliability, put in the effort, and get it. From my engineering perspective, that difference looks huge. And it doesn't stop at 99.99%; the next couple nines are useful too! The level of complexity and usefulness you can build from those components is breathtaking. It's what makes the 21st... (read more)

0Lumifer
Who exactly will be doing the demanding and what would be price for not delivering? Authoritarian systems are often capable of delivering short-term reliability by demanding the head of everyone who fails ("making the trains run on time"). Of course pretty soon they are left without any competent professionals.
0JacekLach
Do you have examples of systems that reach this kind of reliabilty internally? Most high-9 systems work by taking lots of low-9 components, and relying on not all of them failing at the same time. I.e. if you have 10 95% systems that fail completely independently, and you only need one of them to work, that gets you like eleven nines (99.9{11}%). Expecting a person to be 99% reliable is ridiculous. That's like two sick days per year, ignoring all other possible causes of failing to make a task. Instead you should build systems and organisations that have slack, so that one person failing at a particular point in time doesn't make a project/org fail.
evand00

What happens when the committed scorched-earth-defender meets the committed extortionist? Surely a strong precommitment to extortion by a powerful attacker can defeat a weak commitment to scorched earth by a defender?

It seems to me this bears a resemblence to Chicken or something, and that on a large scale we might reasonably expect to see both sets of outcomes.

evand30

What's that? If I don't give into your threat, you'll shoot me in the foot? Well, two can play at that game. If you shoot me in the foot, just watch, I'll shoot my other foot in revenge.

3Lumifer
And then I'll bleed on you!
evand20

On the other hand... what level do you want to examine this at?

We actually have pretty good control of our web browsers. We load random untrusted programs, and they mostly behave ok.

It's far from perfect, but it's a lot better than the desktop OS case. Asking why one case seems to be so much farther along than the other might be instructive.

2whpearson
In some ways Browser is better, it is also more limited. It still has things like CSRF and XSS which can be seen as failures of the user to control their systems. Those are getting better, for CSRF by making the server be more wary about what they accept as legitimate requests. I'll write an article this weekend on the two main system design patterns to avoid. *spoilers* Ambient authority because it causes the confused deputy problem and global namespaces. It is namespaces that web pages browsers have improved, web pages downloaded by the browser can't interact at all, so each one is a little island. It makes some things hard and the user very reliant on external servers.
evand00

Again, I'm going to import the "normal computer control" problem assumptions by analogy:

  • The normal control problem allows minor misbehaviour, but that it should not persist over time

Take a modern milling machine. Modern CNC mills can include a lot of QC. They can probe part locations, so that the setup can be imperfect. They can measure part features, in case a raw casting isn't perfectly consistent. They can measure the part after rough machining, so that the finish pass can account for imperfections from things like temperature variation.... (read more)

evand00

It bends the nails, leaves dents in the surface and given the slightest chance will even attack your fingers!

We've mostly solved that problem.

I'm not sure that being able to nearly perfectly replicate a fixed set of physical actions is the same thing as solving a control problem.

It's precisely what's required to solve the problem of a hammer that bends nails and leaves dents, isn't it?

Stuxnet-type attacks

I think that's outside the scope of the "hammer control problem" for the same reasons that "an unfriendly AI convinced my co-wor... (read more)

0Lumifer
Not quite. We mostly know how to go about it, but we didn't actually solve it -- otherwise there would be no need for QC and no industrial accidents. Still nope. The nails come in different shapes and sizes, the materials can be of different density and hardness, the space to swing a hammer can vary, etc. Replicating a fixed set of actions does not solve the general "control of the tool" problem. I don't think it is. If you are operating in the real world you have to deal with anything which affects the real-life outcomes, regardless of whether it fits your models and frameworks. The Iranians probably thought that malware was "outside the scope" of running the centrifuges -- it didn't work out well for them. Yes, they are. So if you treat the whole thing as an exercise in proper engineering, it's not that hard (by making-an-AI standards :-D) However the point of "agenty" tools is to be able to let the tool find a solution or achieve an outcome without you needing to specify precisely how to do it. In that sense the classic engineering control is all about specifying precise actions and "punishing" all deviations from them via feedback loops.
evand50

They usually don't have any way to leverage their models to increase the cost of not buying their product or service though; so such a situation is still missing at least one criterion.

Modern social networks and messaging networks would seem to be a strong counterexample. Any software with both network effects and intentional lock-in mechanisms, really.

And honestly, calling such products a blend of extortion and trade seems intuitively about right.

To try to get at the extortion / trade distinction a bit better:

Schelling gives us definitions of promises ... (read more)

0Lumifer
The cost of not buying is not the same thing as the cost of switching.
Load More