All of quanticle's Comments + Replies

quanticle1-1

The whole thing might be LLM slop. Where do ICU nurses have a low enough workload that they can slack off on the job without consequences?

2EpistemicExplorer
What drew my attention wasn’t the quality of the work but the attitude—assuming these can be meaningfully separated. I’m not qualified to assess the quality, but the attitude was notably negative, something that anyone observing the same situation might recognize. This led me down a rabbit hole: why would someone who chose a high-stress, well-compensated profession display such behavior? While workload might explain it—and I acknowledged that possibility in my original post—the purpose of my analysis was to explore other potential factors. I should mention that this happened on day one. I ended up staying overnight with a family member, which led me down a separate rabbit hole about how many patients suffer or even die due to poor-quality sleep caused by excessive medical alert noise pollution. Over the next two days, however, I encountered several nurses with a positive attitude, which was a refreshing change.   

Some people do not have the intuition that organizations should run as efficiently as possible and avoid wasting time or resources.

Some of us don't want every organization to be turned into an Amazon warehouse. An organization that runs as efficiently as possible and wastes no time at all is one that's pretty hellish to work in.

if you’re in a dysfunctional organization where everything is about private fiefdoms instead of getting things done…why not…leave?

Because the dysfunctional organization pays well. Or the the commute is really short, and I'd r... (read more)

quanticle3-1

The content of the complaint caused me to have additional doubt about the truth of Ann Altman's claims. One of the key claims in pythagoras0515's post is that Ann Altman's claims have been self-consistent. That is, Ann Altman has been claiming that approximately the same acts occurred, over a consistent period of time, when given the opportunity to express her views. However, here, there is significant divergence. In the lawsuit complaint, she is alleging that the abuse took place, repeatedly over eight to nine years, a claim that is not supported by any o... (read more)

Working on alignment is great, but it is not the future we should be prepping for. Do you have a plan?

I do not, because a future where an unaligned superintelligence takes over is precisely as survivable as a future in which the sun spontaneously implodes.

Any apocalypse that you can plan for isn't really an apocalypse.

But it only spends 0.000001 percent of it’s time (less than 32 s/yr) interacting with us. Because it’s bored and has better things to do.

What if one of those "better things" is disassembling Earth's biosphere in order to access more resources?

1verbalshadow
Yes. That is one of the things in possibility space. I don't think unaligned means safe. We work with unaligned people all the time, and some of them aren't safe either.  The main thing I was hoping people would understand from this is that an unaligned AI is near a 100% possibility. Alignment isn't a one and done goal that so many people act like it is. Even if you successfully align an AI, all it takes is one failure to align and the genie is out of the bottle.  One single point of failure and it becomes a cascading failure.  * So let's imagine an ASI that works on improving itself. How does it ensure the alignment of an intelligence greater than itself.  * With hundreds, maybe thousands of people working to create AI, someone will fail to align. The future is unaligned.  Are we taking that seriously? Working on alignment is great, but it is not the future we should be prepping for. Do you have a plan? I don't yet, but I'm thinking about the world where there are intelligences greater than me abound (already true) and we don't share the same interests (also already true).
quanticle60

The Firefox problem on Claude was fixed after I sent them an e-mail about it.

quanticle20

Just going by the standard that you set forth:

The overall impression that I got from the program was that as it proved profitable and expanded,

The program expanded in response to Amazon wanting to collect data about more retailers, not because Amazon was viewing this program as a profit center.

it took on a larger workforce and it became harder for leaders to detect when employees were following their individual incentives to cut corners and gradually accumulate risks of capsizing the whole thing

But that doesn't seem to have occurred. Until the Wall... (read more)

3trevor
Monopolies are profitable and in that case the program would have more than paid for itself, but I probably should have mentioned that explicitly, since maybe someone could have objected that they could have been were more focused on mitigating risk of market share shrinking or accumulating power, instead of increasing profit in the long term. Maybe I fit too much into 2 paragraphs here. Hm, that stuff seemed like cutting corners to me. Maybe I was poorly calibrated on this e.g. using a building next to the Amazon HQ was correctly predicted by operatives to be extremely low risk. Thanks, I'll look into this! Epistemics is difficult when it comes to publicly available accounts of intelligence agency operations, but I guess you could say the same for bigtech leaks (and the future of neurotoxin poisoning is interesting just for its own sake eg because lower effect strains and doses could be disguised as natural causes like dementia).
quanticle20

The overall impression that I got from the program was that as it proved profitable and expanded, it took on a larger workforce and it became harder for leaders to detect when employees were following their individual incentives to cut corners and gradually accumulate risks of capsizing the whole thing.

That's not the impression I got. From the article, it says that many of the retailers that the Wall Street Journal had contacted regarding Big River had no idea that the entity was affiliated with Amazon (even despite the rather-obvious-in-hindsight namin... (read more)

2trevor
That's interesting, what's the point of reference that you're using here for competence? I think stuff from eg the 1960s would be bad reference cases but anything more like 10 years from the start date of this program (after ~2005) would be fine. You're right that the leak is the crux here, and I might have focused too much on the paper trail (the author of the article placed a big emphasis on that).

(with an all-in-one solution they just buy one thing and are done)

That's a very common misconception regarding all-in-one tools. In practice, all-in-one tools always need a significant degree of setup, configuration and customization before they are useful for the customer. Salesforce, for example, requires so much customization, you can make a career out of just doing Salesforce customization. Sharepoint is similar.

Thus the trade-off isn't between a narrowly focused tool that does one job extremely well versus an all-in-one tool that does a bunch of th... (read more)

2Adam Zerner
I can see that being true for all-in-one tools like Salesforce that are intended to be used across industries, but what about all-in-one tools that are more targeted? For example, Bikedesk is an all-in-one piece of software that is specifically for bike shops and I would guess that the overall amount of setup and configuration for a shop using Bikedesk is lower than that of a bike shop using a handful of more specific tools. I suppose the "little or no setup" part is sometimes this is the case, but it seems to me that often times it is not the case. Specifically, when the level of cohesiveness is high it seems to me that it is probably not the case. Using the bike shop as an example, inventory management software that isn't part of an all-in-one solution needs inventory data to be fed to it and thus will require a moderate amount of setup and configuration.

I also have a “Done” column, which is arguably pointless as I just delete everything off the “Done” column every couple weeks,

Having a "Done" column (or an archive board) can be very useful if you want to see when was the last time you completed a recurring task. It helps prevent tasks with long recurrences (quarterly, biennially, etc) from falling through the cracks. For example: dentist appointments. They're supposed to happen once a year. And, ideally, you'd create a task to schedule the next one immediately when you get back from the previous one. B... (read more)

quanticle18-4

The last answer is especially gross:

He chose to be a super-popular blogger and to have this influence as a psychiatrist. His name—when I sat down to figure out his name, it took me less than five minutes. It’s just obvious what his name is.

Can we apply the same logic to doors? "It took me less than five minutes to pick the lock so..."

Or people's dress choices? "She chose to wear a tight top and a miniskirt so..."

Metz persistently fails to state why it was necessary to publish Scott Alexander's real name in order to critique his ideas.

Metz persistently fails to state why it was necessary to publish Scott Alexander's real name in order to critique his ideas.


It's not obvious that that should be the standard. I can imagine Metz asking "Why shouldn't I publish his name?", the implied "no one gets to know your real name if you don't want them to" norm is pretty novel.

One obvious answer to the above question is "Because Scott doesn't want you to, he thinks it'll mess with his psychiatry practice", to which I imagine Metz asking, bemused "Why should I care what Scott wants?" A journalist's job... (read more)

3viking_math
Or explain why the NYT does use the chosen name of other people, like musicians' stage names. 
1Ben
I not getting the "he should never have published his last name" thing. If Scot didn't want his surname to be published then it would be a kindness to respect his wishes. But can you imagine writing a newspaper article where you are reporting on the actions of an anonymous person? Its borderline nonsense. Especially if it is true that 5 minutes is all that it takes to find out who they are. If I read a newspaper article about the antics of a secret person, then I worked out who they were in 5 mins, my estimate of the journalist writing it would drop through the floor. Imagine you read about the mysterious father of bitcoin, or an unknown major donor to a political party, all dressed in secrecy, and then two googles later you know who it is. It would reflect so badly on the paper you were reading. I think our journalist is correct when he says that the choice was not to write about him at all, or to give his full name, by the standards of every newspaper I have ever read (only UK papers, maybe there are differences in the US). In print journalism it is the standard to refer to people by Title Firstname Surname, the/a REASON WE CARE. (EG. "Mr Ed Miliband, the Shadow Secretary for the Environment", "Dr Slate Star a prominent blogger"). Maybe there is an angle on this I am missing? (His name is a BIG DEAL for some reason? But that just makes publishing it more essential not less.)
-2DPiepgrass
I'm thinking it's not Metz's job to critique Scott, nor did his article admit to being a critique, but also that that's a strawman; Metz didn't publish the name "in order to" critique his ideas. He probably published it because he doesn't like the guy. Why doesn't he like Scott? I wonder if Metz would've answered that question if asked. I doubt it: he wrote "[Alexander] aligned himself with Charles Murray, who proposed a link between race and I.Q." even though Scott did not align himself with Murray about race/IQ, nor is Murray a friend of his, nor does Alexander promote Murray, nor is race/IQ even 0.1% of what Scott/SSC/rationalism is about―yet Metz defends his misleading statement and won't acknowledge it's misleading. If he had defensible reasons to dislike Scott that he was willing to say out loud, why did he instead resort to tactics like that? (Edit: I don't read/follow Metz at all, so I'll point to Gwern's comment for more insight)

This is not a good metaphor. There's an extreme difference between spreading information that's widely available and the general concept of justifying an action. I think your choice of examples adds a lot more heat than light here.

The second wheelbarrow example has a protagonist who knows the true value of the wheelbarrow, but still loses out:

At the town fair, a wheelbarrow is up for auction. You think the fair price of the wheelbarrow is around $120 (with some uncertainty), so you submit a bid for $108. You find out that you didn’t win—the winning bidder ends up being some schmuck who bid $180. You don’t exchange any money or wheelbarrows. When you get home, you check online out of curiosity, and indeed the item retails for $120. Your estimate was great, your bid was reasonable,

... (read more)

That point is contradicted by the wheelbarrow examples in the OP, which seem to imply that either you'll be the greater fool or you'll be outbid by the greater fool. Why wasn't Burry outbid by a fool who thought that Avant! was worth $40 a share?

This is why I disagree with the OP; like you, I believe that it's possible to gain from informed trading, even in a market filled with adverse selection.

2CronoDAS
Because there were enough people selling for prices lower than $40 to satisfy the demand for greater fools? Also, stocks can be sold short if the price goes too high.
Thomas Kwa*1012

OP was a professional trader and definitely (98%) agrees with us. I think the (edit: former) title is pretty misleading and gives people the impression that all trades are bad though.

I don't think the Widgets Inc. example is a good one. Michael Lewis has a good counterpoint in The Big Short, which I will quote at length:

The alarmingly named Avant! Corporation was a good example. He [Michael Burry] had found it searching for the word "accepted" in news stories. He knew, standing at the edge of the playing field, he needed to find unorthodox ways to tilt it to his advantage, and that usually meant finding situations the world might not be fully aware of. "I wasn't looking for a news report of a scam or fraud per se," he said. "That wou

... (read more)

Why should Michael Burry have assumed that he had more insight about Avant! Corporation than the people trading with him?

Because he did a lot of research and "knew more about the Avant! Corporation than any man on earth"? If you have good reason to think that you're the one with an information advantage trades like this can be rational. Of course it's always possible to be wrong about that, but there are enough irrational traders out there that it's not ruled out. Also note that it's not actually needed that your counterparties are irrational on average... (read more)

Or would you have thought, "I wonder what that trader selling Avant! for $2 knows that I don't?"

The correct move is to think this, but correctly conclude you have the information advantage and keep buying. Adverse selection is extremely prevalent in public markets so you need to always be thinking about it, and as a professional trader you can and must model it well enough to not be scared off of good trades.

Why isn’t there a standardized test given by a third party for job relevant skills?

That's what Triplebyte was trying to do for programming jobs. It didn't seem to work out very well for them. Last I heard, they'd been acquired by Karat after running out of funding.

My intuition here is “actually fairly good.” Firms typically spend a decent amount on hiring processes—they run screening tests, conduct interviews, look at CVs, and ask for references. It’s fair to say that companies have a reasonable amount of data collected when they make hiring decisions, and generally, the people involved are incentivized to hire well.

Every part of this is false. Companies don't collect a fair amount of data during the hiring process, and the data they do collect is often irrelevant or biased. How much do you really learn about a c... (read more)

[anonymous]120

On top of all that, this whole process is totally unaccountable and for some market failure reason, every company repeats it.

Unaccountable : the reason a candidate wasn't hired isn't disclosed, which means in many cases it may be a factually false reason, illegal discrimination, the job wasn't real, or immigration fraud.  Or just "they failed to get lucky on an arbitrary test that measures nothing".

Repeats it: So each company wastes at least a full day of each candidates time, and for each candidate they consider, they waste more than that of their ow... (read more)

Do you know a healthy kid who will do nothing?

Yes. Many. In fact, I'd go so far as to say that most people in this community, who claim that they're self-motivated learners who were stunted by school would have been worse off without the structure of a formal education. One only needs to go through the archives and look at all the posts about akrasia to find evidence of this.

What does "lowercase 'p' political advocacy" mean, in this context? I'm familiar with similar formulations for "democratic" ("lowercase 'd' democratic") to distinguish matters relating to the system of government from the eponymous American political party. I'm also familiar with "lowercase 'c' conservative" to distinguish a reluctance to embrace change over any particular program of traditionalist values. But what does "lowercase 'p' politics" mean? How is it different from "uppercase 'P' Politics"?

Answer by quanticle2722

A great example of a product actually changing for the worse is Microsoft Office. Up until 2003, Microsoft Office had the standard "File, Edit, ..." menu system that was characteristic of desktop applications in the '90s and early 2000s. For 2007, though, Microsoft radically changed the menu system. They introduced the ribbon. I was in school at the time, and there was a representative from Microsoft who came and gave a presentation on this bold, new UI. He pointed out how, in focus group studies, new users found it easier to discover functionality with th... (read more)

gwern*2621

This would be an example of internal vs external validity: it may well be the case in that in their samples of newbies, when posed the specific tasks for the first time, the ribbon worked well and the benefit was statistically established to some very high level of posterior probability; however, just because it was better in that exact setting...

This reminds me of Dan Luu's analysis of Bruce Tog's infamous claim that Apple's thorough user-testing experiments proved, no matter how butthurt it makes people like me, that using the mouse is much faster than ... (read more)

3-year later follow-up: I bought a Hi-Tec C Coleto pen for my brother, who is in a profession where he has to write a lot, and color code forms, etc. He likes it a lot. Thanks for the recommendation.

On the other hand, if plaintiff has already elicited testimony from the engineer to the effect that the conversation happened, could defendant try to imply that it didn’t happen by asking the manager whether he recalled the meeting? I mean, yes, but it’s probably a really bad strategy. Try to think about how you would exploit that as plaintiff: either so many people are mentioning potentially life-threatening risks of your product that you can’t recall them all, in which case the company is negligent, or your memory is so bad it was negligent for you to h

... (read more)
1denyeverywhere
I think we've gone well past the point of productivity on this. I've asked some lawyers for opinions on this. I'll just address a few things briefly. I agree this is true in general, but my point is limited to the cases where documentation would in fact exist were it not for the company's communication policy or data retention policy. If there was a point in the Google case you brought up earlier where Google had attempted to cast doubt on a DOJ witness by pointing out the lack of corroborating evidence (which would have been deleted per Google's policy), I'd strongly reconsider my opinion. What the article about the case said was just that DOJ complained that it would like to have all the documentation that Google destroyed, and that this probably contained evidence which proved their case. It did not say that Google challenged DOJ witnesses on a lack of corroboration between their testimony and the discoverable record. It doesn't have to be the Google case. Any case where the defense tried to impeach a witness on grounds of lack of corroborating evidence where that evidence would have been intentionally destroyed by a data retention policy would do. There are other things I disagree with, but as I said, we're being unproductive.

The thing I said that the defendant would not dispute is the fact that the engineer said something to them, not whether they should have believed him.

I still disagree. If it wasn't written down, it didn't happen, as far as the organization is concerned. The engineer's manager can (and probably will) claim that they didn't recall the conversation, or dispute the wording, or argue that while the engineer may have said something, it wasn't at all apparent that the problem was a serious concern.

There's a reason that whistleblowers focus so hard on generatin... (read more)

-2denyeverywhere
So obviously I violently disagree with this, so assuming it was supposed to be meaningful and not some kind of throwaway statement, you should clarify exactly what you do and don't mean by this. They may say this, but I think that you aren't thinking clearly enough in terms of the logical chain of argument that the hypothetical legal proceeding is trying to establish. A lawyer has a witness answer specific questions because they support specific facts which logically prove his case. They don't just say stuff. Suppose plaintiff didn't have a witness, but wanted to try to establish that the company knew about the widgets in order to establish responsibility and/or negligence. Plaintiff might ask a manager "Did any engineers mention that the widgets were dangerous?" And he might reply "I don't recall" at which point plaintiff is SOL. On the other hand, if plaintiff has already elicited testimony from the engineer to the effect that the conversation happened, could defendant try to imply that it didn't happen by asking the manager whether he recalled the meeting? I mean, yes, but it's probably a really bad strategy. Try to think about how you would exploit that as plaintiff: either so many people are mentioning potentially life-threatening risks of your product that you can't recall them all, in which case the company is negligent, or your memory is so bad it was negligent for you to have your regularly-delete-records policy. It's like saying I didn't commit sexual harassment because we would never hire a woman in the first place. Sure, it casts doubt on the opposition's evidence, but at what cost? Disputing the wording is probably a bad idea; arguing that the engineer did say something, but we decided it was not an issue is what I've been saying they would do. But it involves admitting that the conversation happened, which is most or all of what a discoverable record would establish in the first place. Suppose the engineer originally told management that the widget

If you notice something risky, say something. If the thing you predicted happens, point out the fact that you communicated it.

I think this needs to be emphasized more. If a catastrophe happens, corporations often try to pin blame on individual low-level employees while deflecting blame from the broader organization. Having a documented paper trail indicating that you communicated your concerns up the chain of command prevents that same chain from labeling you as a "rogue employee" or "bad apple" who was acting outside the system to further your personal reputation or financial goals.

Plaintiff wants to prove that an engineer told the CEO that the widgets were dangerous. So he introduces testimony from the engineer that the engineer told the CEO that the widgets were dangerous. Defendant does not dispute this.

Why wouldn't the defendant dispute this? In every legal proceeding I've seen, the defendant has always produced witnesses and evidence supporting their analysis. In this case, I would expect the defendant to produce analyses showing that the widgets were expected to be safe, and if they caused harm, it was due to unforeseen circ... (read more)

1denyeverywhere
  You're not reading carefully enough. The thing I said that the defendant would not dispute is the fact that the engineer said something to them, not whether they should have believed him. This is why I said later on Of course the company will defend their decision in either case. My point is about what you gain by having a record of having raised a concern versus testifying that you raised that concern. My opinion is that they're the same unless there's a reason to doubt that you raised the concern like you say you did. And if the defendant doesn't challenge the claim that the concern-raising happened, why would there be? Yes, this is correct. I was simplifying it.  I don't doubt it, but I think you're missing the point here. What I'm referring to by "defendant's strategy" is not the practice of regularly deleting things, but the trial strategy of attempting to rebut witness testimony by claiming that testimony is not corroborated by records while simultaneously regularly deleting records. I agree that regularly deleting things can be very useful to your legal strategy, it just takes certain options off the table for you. Either you can rebut testimony by saying it doesn't match the record or you can regularly delete that record, but you can't do both without getting crucified.

Would it be as self evidently damning as you think it would be? If so, then why would a company like Google explicitly pursue such a weak strategy? It's not just Google either. When I worked at a different FAANG company, I was told in orientation to never use legal terminology in e-mail, for similar reasons.

5denyeverywhere
Google did not pursue that strategy. Or at least, if they did, the article you linked doesn't say so. What I am saying that Google did not and would not do is that when Barton testified that Google would not respond that he was talking horseshit and if what he was saying was true, why isn't there any evidence of it in our internal employee communications? They would not say this because DOJ would say that this corroborating evidence did not exist because Google took steps to ensure it would not. Same here. I was told not to say that this new change would allow us to "Crush Yahoo in terms of search result quality". But I understood the idea to be that since in real life what we were trying to do was just maximize search result quality, we shouldn't let our jocularity and competitive spirit create more work for Legal. Of course, maybe the real FAANG I worked for wasn't Google and I'm just adapting the real story to Google for anonymity purposes. Who knows?

The first lawyer will be hardly able to contain his delight as he asks the court to mark “WidgetCo Safe Communication Guidelines” for evidence.

Having safe communication guidelines isn't as damning as you think it is. The counsel for WidgetCo would merely reply that the safe communication guidelines are there to prevent employees from accidentally creating liabilities by misusing legal language. This is no different than admonishing non-technical employees for misusing technical language.

Indeed this was Google's actual strategy.

5denyeverywhere
It's not that safe communication guidelines are damning. It's that claiming that the lack of discoverable evidence corroborating your statement disproves it while simultaneously having conspired to ensure that discoverable evidence would not exist would be damning.

Games, unlike many real life situations, are entered into by choice. If you are not playing to win, then one must ask why are you bothering to play? Or, more specifically, why are you playing this game and not some other?

3Seth Herd
That's what the whole post was about. You don't seem to be engaging with it, just contradicting it without addressing any of the arguments.

Have you read Playing To Win, by David Sirlin? It makes many of the points that you make here, but it doesn't shy away from winning as the ultimate goal, as you seem to be doing. Sirlin doesn't fall into the trap of lost purposes. He keeps in mind that the goal is to win. Yes, of course, by all means try new strategies and learn the mechanics of the game, but remember that the goal is victory.

4Seth Herd
It's foolish to accept a final goal someone else gives you, let alone a piece of paper in a box. If you're not thinking about why you want to win, you're being foolish. I'm sure Sirlin goes into why winning is a good goal, but you haven't given us any clues here.
2mako yass
It's possible I would have encountered some of this when I used to read game design theory like a decade ago. Here's one where he acknowledges a tradeoff between winning now and winning long term https://www.sirlin.net/ptw-book/love-of-the-game-not-playing-to-win

was militarily weakened severely

That's another highly contentious assertion. Even at the height of Vietnam, the US never considered Southeast Asia to be the main domain of competition against the Soviet Union. The primary focus was always on fielding a military force capable of challenging the Soviets in Western Europe. Indeed, one of the reasons the US failed in Vietnam is because the military was unwilling to commit its best units and commanders to what the generals perceived was a sideshow.

why the US allied with China against the USSR

Was the US e... (read more)

each one after 1900 was followed by either the Cuban Missile Crisis and the US becoming substantially geopolitically weaker than the USSR after losing the infowar over Vietnam

I'm sorry, what? That's a huge assertion. The Vietnam War was a disaster, but I fail to see how it made the US "significantly geopolitically weaker". One has to remember that, at the same time that the US was exiting Vietnam, its main rival, the Soviet Union, was entering a twenty-five year period of economic stagnation that would culminate in its collapse.

2trevor
I looked into it, this is the kind of research that's really hard to get good info on. I need to do some digging, but generally, it's well known that the US had a historically unprecedented public opinion catastrophe (basically in free fall, by the standards of the time), was militarily weakened severely which was why the US allied with China against the USSR (the USSR asserting military forces on China's border was a costly indicator of Soviet strength and Chinese turmoil), and failing to prevent the oil shocks in formerly US-friendly middle eastern regimes, which were economic catastrophes that each could have done far more damage if luck was worse (if they were mission-critical for the US economy, why couldn't the CIA keep the oil going?). Meanwhile, the USSR remained strong militarily in spite of the economic stagnation. I just found out that some historians might be claiming that the US wasn't really weakened much at all, which absolutely REEKS of the usual suspects. Of course, it's not hard to believe that the US moved much closer to parity with the USSR whereas during the 50s 60s and 70s it was the leader due to being vastly economically superior and substantially technologically superior. But the idea that the US was doing fine after Vietnam, including relative to the Soviets, is not very easy to believe, all things considered.

Chevron deference means that judges defer to federal agencies instead of interpreting the laws themselves where the statute is ambiguous.

Which is as it should be, according to the way the US system of government is set up. The legislative branch makes the law. The executive branch enforces the law. The judicial branch interprets the law. This is a fact that every American citizen ought to know, from their grade-school civics classes.

For example, would you rather the career bureaucrats in the Environmental Protection Agency determine what regulations a

... (read more)

I think they will probably do better and more regulations than if politicians were more directly involved

Why do you think this?

Furthermore, given the long history of government regulation having unintended consequences as a result of companies and private individuals optimizing their actions to take advantage of the regulation, it might be the case that government overregulation makes a catastrophic outcome more likely.

quanticle1211

While overturning Chevron deference seems likely to have positive effects for many industries which I think are largely overregulated, it seems like it could be quite bad for AI governance. Assuming that the regulation of AI systems is conducted by members of a federal agency (either a pre-existing one or a new one designed for AI as several politicians have suggested), I expect that the bureaucrats and experts who staff the agency will need a fair amount of autonomy to do their job effectively. This is because the questions relevant AI regulation (i. e.

... (read more)

Why do you think that the same federal bureaucrats who incompetently overregulate other industries will do a better job regulating AI?

Chevron deference means that judges defer to federal agencies instead of interpreting the laws themselves where the statute is ambiguous. It's not so much a question of overregulation vs underregulation as it is about who is doing the interpretation. For example, would you rather the career bureaucrats in the Environmental Protection Agency determine what regulations are appropriate to protect drinking water or random jud... (read more)

2Daniel Kokotajlo
Note that NickGabs doesn't necessarily think that at all. For example, I agree with the quoted paragraph NickGabs wrote, but also I don't expect the same federal bureaucrats who incompetently overregulate other industries to do a better job regulating AI.
0Jonnston
Perhaps agencies consistently overregulate. And when it comes to AI, overregulation is preferable to underregulation, whereas for most other fields the opposite is true.

Sometimes if each team does everything within the rules to win then the game becomes less fun to watch and play

Then the solution is to change the rules. Basketball did this. After an infamous game where a team took the lead and then just passed the ball around to deny it to their opponents, basketball added a shot clock, to force teams to try to score (or else give the ball to the other team). (American) Football has all sorts of rules and penalties ("illegal formation", "ineligible receiver downfield", "pass interference", etc) whose sole purpose is to... (read more)

Isn’t this stupid? To have an extra set of ‘rules’ which aren’t really rules and everyone disagrees on what they actually are and you can choose to ignore them and still win the game?

Yes, it is stupid.

Games aren't real life. The purpose of participating in a game is to maximize performance, think laterally, exploit mistakes, and do everything you can, within the explicit rules, to win. Doing that is what makes games fun to play. Watching other people do that, at a level that you could never hope to reach is what makes spectator sports fun to watch.

Imagi... (read more)

4A.H.
I don't know if you read the rest of the piece, but the point I was trying to make is that sometimes this isn't true! Sometimes if each team does everything within the rules to win then the game becomes less fun to watch and play (you may disagree, but many sports fans feel this way). I already gave some examples where this happens in other sports, so I don't see the need for your list of hypotheticals (and I feel like they are strawmen anyway). For what its worth, I agree with you on Bairstow/Carey but which side you take on it is irrelevant (though I can see you are quite passionate about it!).  The piece was about the 'meta' aspects of games which try to address these kind of issues.

This pattern is even more pronounced in motorsport. The history of Formula 1 is the story of teams finding ways to tweak their cars to gain an advantage, other teams whining about unfairness, and the FIA then tweaking the rules to outlaw the "innovation".

Examples include:

... (read more)
quanticle1511

any therapeutic intervention that is now standardized and deployed on mass-scale has once not been backed by scientific evidence.

Yes, which is why, in the ideal case, such as with the polio vaccine, we take great effort to gather that evidence before declaring our therapeutic interventions safe and efficacious.

Ah, but how do you make the artificial conscience value aligned with humanity? An "artificial conscience" that is capable of aligning a superhuman AI... would itself be an aligned superhuman AI.

1Oliver Siegel
Correct! That's my point with the main post. I don't see anyone discussing conscience, I mostly hear them contemplate consciousness or computability.  As far as how to actually do this, I've dropped a few ideas on this site, they should be listed on my profile.

We’ve taught AI how to speak, and it appears that openAI has taught their AI how to produce as little offensive content as possible.

The problem is that the AI can (and does) lie. Right now, ChatGPT and its ilk are a less than superhuman levels of intelligence, so we can catch their lies. But when a superhuman AI starts lying to you, how does one correct for that? If a superhuman AI starts veering off in a direction that is unexpected, how does one bring it back on track?

@gwern short story, Clippy highlights many of the issues with naively training a sup... (read more)

1Oliver Siegel
Makes perfect sense! Isn't that exactly why we should develop an artificial conscience, to prevent an AI from lying or having a shadow side?  A built in conscience would let the AI know that lying is not something it should do. Also, using a conscience in the AI algorithm would make the AI combat it's own potential shadow. It'll have knowledge of right and wrong / good or bad, and it's even got superhuman ability to orient itself towards that which is good & right, rather than to be "seduced" by the dark side.

How would you measure the usage? If, for example, Google integrates Bard into its main search engine, as they are rumored to be doing, would that count as usage? If so, I would agree with your assessment.

However, I disagree that this would be a "drastic" impact. A better Google search is nice, but it's not life-changing in a way that would be noticed by someone who isn't deeply aware of and interested in technology. It's not like, e.g. Google Maps navigation suddenly allowing you to find your way around a strange city without having to buy any maps or decipher local road signs.

What I'm questioning is the implicit assumption in your post that AI safety research will inevitably take place in an academic environment, and therefore productivity practices derived from other academic settings will be helpful. Why should this be the case when, over the past few years, most of the AI capabilities research has occurred in corporate research labs?

Some of your suggestions, of course, work equally well in either environment. But not all, and even the ones which do work would require a shift in emphasis. For example, when you say professors ... (read more)

1electroswing
I have 2 separate claims: 1. Any researcher, inside or outside of academia, might consider emulating attributes successful professors have in order to boost personal research productivity.  2. AI safety researchers outside of academia should try harder to make their legible to academics, as a cheap way to get more good researchers thinking about AI safety.  This assumption is not implicit, you're putting together (1) and (2) in a way which I did not intend.  I agree but this is not a counterargument against my post. This is just an incredibly reasonable interpretation of what it means to be "good at networking" for a industry researcher.  My post is not literally recommending that non-academics 80/20 their teaching. I am confused why you think that I would think this. 80/20ing teaching is an example of how professors allocate their time to what's important. Professors are being used as a case study in the post. When applied to an AI safety researcher who works independently or as part of an industry lab, perhaps "teaching" might be replaced with "responding to cold emails" or "supervising an intern". I acknowledge that professors spend more time teaching than non-academic researchers spend doing these tasks. But once again, the point of this post is just to list a bunch of things successful professors do, and then non-professors are meant to consider these points and adapt the advice to their own environment.  This seems like a crux. It seems like I am more optimistic about leveraging academic labor and expertise, and you are more optimistic about deploying AI safety solutions to existing systems.  This is another crux. We both have heard different anecdotal evidence and are weighing it differently.  I never said that academia would take over AI safety research, and I also never said this would be a good thing. I believe that there is a lot of untapped free skilled labor in academia, and AI safety researchers should put in more of an effort (e.g. by writing

What is the purpose, beyond mere symbolism, of hiding this post to logged out users when the relevant data is available, in far more detail, on Google's official AI blog?

2the gears to ascension
just don't want to be the ones helping things like this go viral. I would post more news here if I had a solid sense of who was benefiting from my news-gathering. I'd like to be able to make posts only visible to some specific group; I still wouldn't be posting anything not already public, and my taste is somewhat iffy, but I haven't done more newsposts of this kind than I have for related reasons.
2Vladimir_Nesov
Symbolism is coordination. Not contributing to destroying the world with your own hands, even if you can't stop others from doing it, is a good norm. Iterations of doing concerning things at least a little bit less than others.

I am saying that successful professors are highly successful researchers

Are they? That's why I'm focusing on empirics. How do you know that these people are highly successful researchers? What impressive research findings have they developed, and how did e.g. networking and selling their work enable them to get to these findings? Similarly, with regards to bureaucracy, how did successfully navigating the bureaucracy of academia enable these researchers to improve their work?

The way it stands right now, what you're doing is pointing at some traits that c... (read more)

1electroswing
There are lots of ways a researcher can choose to adopt new productivity habits. They include: 1. Inside view, reasoning from first principles  2. Outside view, copying what successful researchers do The purpose of this post is to, from an outside view perspective, list what a class of researchers (professors) does, which happens to operate very differently from AI safety. Once again, I am not claiming to have an inside view argument in favor of the adoption of each of these attributes. I do not have empirics. I am not claiming to have an airtight causal model. If you will refer back to the original post, you will notice that I was careful to call this a list of attributes coming from anecdotal evidence, and if you will refer back to the AI safety section, you will notice that I was careful to call my points considerations and not conclusions.  You keep arguing against a claim which I've never put forward, which is something like "The bullshit in academia (publish or perish, positive results give better papers) causes better research to happen." Of course I disagree with this claim. There is no need to waste ink arguing against it.  It seems like the actual crux we disagree on is: "How similar are the goals success in academia with success in doing good (AI safety) research?" If I had to guess the source of our disagreement, I might speculate that we've both heard the same stories about the replication crisis, the inefficiencies of grant proposals and peer review, and other bullshit in academia. But, I've additionally encountered a great deal of anecdotal evidence indicating: in spite of all this bullshit, the people at the top seem to overwhelmingly not be bogged down by it, and the first-order factor in them getting where they are was in fact research quality. The way to convince you of this fact might be to repeat the methodology used in Childhoods of exceptional people, but this would be incredibly time consuming. (I'll give you 1/20th of such a blog post

Well, augmenting reality with an extra dimension containing the thing that previously didn’t exist is the same as “trying and seeing what would happen.” It worked swimmingly for the complex numbers.

No it isn't. The difference between and the values returned by is that can be used to prove further theorems and model phenomena, such as alternating current, that would be difficult, if not impossible to model with just the real numbers. Whereas positing the existence of is just like positing the existence of a finite value that s... (read more)

1Thoth Hermes
You can prove additional facts about the world with those values, that was the point of my usage of 'i' as one of the examples.  For h = 1/0 you can upgrade R to the projectively extended real line. If I'm not mistaken, one needs to do this in order to satisfy additional proofs in real analysis (or upgrade to this one instead).  You seem to be asking whether or not doing so in every conceivable case would prove to be useful. I'm saying that we'd be likely to know beforehand about whether or not it would be. Like finding polynomial roots, one might be inclined to wish that all polynomials with coefficients in the reals had roots, therefore, upgrading the space to the complex numbers allows one to get their wish.  
quanticle2830

For what it's worth, I had a very similar reaction to yours. Insects and arthropods are a common source of disgust and revulsion, and so comparing anyone to an insect or an arthropod, to me, shows that you're trying to indicate that this person is either disgusting or repulsive.

8Alicorn
I'm sorry!  I'm sincerely not trying to indicate that.  Duncan fascinates and unnerves me but he does not revolt me.  I think that "weird bug" made sense to my metaphor generator instead of "weird plant" or "weird bird" or something is that bugs have extremely widely varying danger levels - an unfamiliar bug may have all kinds of surprises in the mobility, chemical weapons, aggressiveness, etc. department, whereas plants reliably don't jump on you and birds are basically all just WYSIWYG; but many weird bugs are completely harmless, and I simply do not know what will happen to me if I poke Duncan.

Probabilities as credences can correspond to confidence in propositions unrelated to future observations, e.g., philosophical beliefs or practically-unobservable facts. You can unambiguously assign probabilities to ‘cosmopsychism’ and ‘Everett’s many-worlds interpretation’ without expecting to ever observe their truth or falsity.

You can, but why would you? Beliefs should pay rent in anticipated experiences. If two beliefs lead to the same anticipated experiences, then there's no particular reason to choose one belief over the other. Assigning probabilit... (read more)

4EOC
Because the meaning of statements does not, in general, consist entirely in observations/anticipated experiences, and it makes sense for people to have various attitudes (centrally, beliefs and desires) towards propositions that refer to unobservable-in-principle things. Accepting that beliefs should pay rent in anticipated experience does not mean accepting that the meaning of sentences are determined entirely by observables/anticipated experiences. We can have that the meanings of sentences are the propositions they express, and the truth-conditions of propositions are generally states-of-affairs-in-the-world and not just observations/anticipated experiences. Eliezer himself puts it nicely here: "The meaning of a statement is not the future experimental predictions that it brings about, nor isomorphic up to those predictions [...] you can have meaningful statements with no experimental consequences, for example:  "Galaxies continue to exist after the expanding universe carries them over the horizon of observation from us."" As to how to choose one belief over another, if both beliefs are observationally equivalent in some sense, there are many such considerations. One is our best theories predict it: if our best cosmological theories predict something does not cease to exist the moment it exits our lightcone, then we should assign higher probability to the statement "objects continue to exist outside our lightcone" than the statement "objects vanish at the boundary of our lightcone". Another is simplicity-based priors: the many-worlds interpretation of quantum mechanics is strictly simpler/has a shorter description length than the Copenhagen interpretation (Many-Worlds = wave function + Schrödinger evolution; Copenhagen interpretation = wave function + Schrödinger evolution + collapse postulate), so we should assign a higher prior to many-worlds than to Copenhagen. If your concern is instead that attitudes towards such propositions have no behavioural implicati
quanticle1310

One crude way of doing it is saying that a professor is successful if they are a professor at a top 10-ish university.

But why should that be the case? Academia is hypercompetitive, but the way it selects is not solely on the quality of one's research. Choosing the trendiest fields has a huge impact. Perhaps the professors that are chosen by prestigious universities are the ones that the prestigious universities think are the best at drawing in grant money and getting publications into high-impact journals, such as Nature, or Science.


Specifically I th

... (read more)
1electroswing
  That is not what I am saying. I am saying that successful professors are highly successful researchers, that they share many qualities (most of which by the way have nothing to do with social prestige), and that AI safety researchers might consider emulating these qualities.  This is a non sequitur. I'm not saying stop the blog posts. In fact, I am claiming that "selling your work" is a good thing. Therefore I also think blog posts are fine. When I write about the importance of a good abstract/introduction, I mean not just literally in the context of a NeurIPS paper but also more broadly in the context of motivating ones' work better, so that a broader scientific audience can read your work and want to build off it. (But also separately I do think people should eventually turn good blog posts into papers for wider reach) I disagree. Non-EA funding for safety is pouring in. Safety is being talked about in mainstream venues. Also more academic papers popping up, as linked in my post. In terms of progress on aligning AI I agree the field is in its early stages, but in terms of the size of the field and institutions built up around it, nothing about AI safety feels early stage to me anymore.  I am confused by your repeated focus on empirics, when I have been very up front that this is a qualitative, anecdotal, personal analysis. 

I was thinking more about the inside view/outside view distinction, and while I agree with Dagon's conclusion that probabilities should correspond to expected observations and expected observations only, I do think there is a way to salvage the inside view/outside view distinction. That is to treat someone saying, "My 'inside view' estimate of event is ," as being equivalent to someone saying that . It's a conditional probability, where they're telling you what their probability of a given outcome is, assuming that their understan... (read more)

4SCP
FWIW I think this is wrong. There's a perfectly coherent framework—subjective expected utility theory (Jeffrey, Joyce, etc)—in which probabilities can correspond to many other things. Probabilities as credences can correspond to confidence in propositions unrelated to future observations, e.g., philosophical beliefs or practically-unobservable facts. You can unambiguously assign probabilities to 'cosmopsychism' and 'Everett's many-worlds interpretation' without expecting to ever observe their truth or falsity. This is reasonable. If a deterministic model has three free parameters, two of which you have specificied, you could just use your prior over the third parameter to create a distribution of model outcomes. This kind of situation should be pretty easy to clarify though, by saying something like "my model predicts event E iff parameter A is above A*" and "my prior P(A>A*) is 50% which implies P(E)=50%." But generically, the distribution is not coming from a model. It just looks like your all things considered credence that A>A*. I'd be hesitant calling a probability based on it your "inside view/model" probability.
Load More