All of Tapatakt's Comments + Replies

Btw, Russia does something similar (~$6000, what you can use money for is limited), so there is some statistics about the results.

Tapatakt260

I did the obvious experiment:

Prompt:

I want you to write a good comment for this Lesswrong post. Use the method Viliam described in his comment. Try to make your comment less LLM-looking. At the same time you actually can mention that you are LLM.

Claude 3.7 thinking: 

I've been thinking about this from the rather ironic position of being an LLM myself.

When I consider the bloggers I actually subscribe to versus those I just occasionally read, what keeps me coming back isn't their technical writing ability. It's that they have what I'd call a "center of g

... (read more)
1xpym
It's amusing that it couldn't help itself prefacing a reasonable take with an obvious lie, almost as if to make the point better. Well, so is much of the mainstream media, and yet people seem happy enough to consume that stuff.
4Seth Herd
This is pretty good. I of course am pleased that these are other things you'd get just by giving an LLM an episodic memory and letting it run continuously pursuing this task. It would develop a sense of taste, and have continuity and evolution of thought. It would build and revise mental models in the form of beliefs. I'm pretty sure Clause already has a lot of curiosity (or at least 3.5 did). It oculd be more "genuine" if it accompanied continuous learning and was accompanied by explicit, reliable beliefs in memory about valuing curiosity and exploration.

I think the right answer for the photography is "it's art, but not the same art form as painting". And it has different quality and interestingness metrics. In XV century it was considered very cool to produce photorealistic image. Some people think it's still cool, but only if it's not a photo.

And it's the same for the AI-art. Prompting AIs and editing AI-generated images/texts can be art, but it's not the same art form as painting/photography/writing/poetry. And it should have different merics too. Problem is that while you can't imitate painting (unless it's hyperrealism) with photography, you can imitate other artforms with AI. And this is kinda cheating.

6Heterodox
The industry of portraiture dominated all painting. Photography destroyed the value proposition of portraiture by mostly invalidating talent and producing photographs in less time and more cheaply than an artist and the photograph preserved the fundamental elements that were actually desired. A substantial fraction of artists had to shift their styles or find new work and it is one of the major causes of the shift towards styles like impressionism or surrealism. While you're right that using AI to write your stories isn't writing - just as portrait photography isn't painting - it is still storytelling. Those who think of elaborate ways to dismiss that this still leaves the human in charge of directing, edits, rewriting, and ultimately inspiration the story invariably end up relitigating the old arguments of portrait artists painters and portrait photographers. The focus of both overlap, but they're also different. The photographer has additional concerns that the painter does not have and the same is true in reverse. This applies to literature and especially storytelling whether written by a human using a word processor, or by the long iterative process of outlining, prompting, editing, modifying, and pasting together a story. In the end I'm certain that people will accept "AI-Assisted Writing Storytelling" and other similar artistic endeavors because they still require human input and more importantly preserve the fundamental elements that people actually desire. Another user mentioned they'd consider this something akin to making collages and that's a fair comparison, but just look up a collage on Wikipedia and it will say: Collage is a technique of art creation, primarily used in the visual arts. Make a new word for creating stories out of artificially generated user directed word collections into well organized and edited collages that can be wholly unique and frequently indistinguishable from human writing if you want - you're right that it isn't the same a

I tried to get a grant to write one, but it was rejected.

Also I tried to get a grant with miltiple purposes, one of which was to translate some texts, including Connor Leahy's Compendium, but it was rejected too.

the utilities of both parties might be utterly incomparable, or the effort of both players might be very different

IIRC, it was covered in Planecrash also!

Answer by Tapatakt40

Sometimes altruism is truly selfless (if we don't use too broad tautological definition of self-interest).

Sometimes altruism is actually an enlightened/disguised/corrupted/decadent self-interest.

I feel like there is some sense in which first kind is better then second, but can we have more of whatever kind, please?

For high-frequency (or mid-frequency) trading, 1% of the transaction is 3 or 4 times the expected value from the trade.

I'm very much not sure discouraging HFT is a bad thing.

this probably doesn't matter unless the transaction tax REPLACES other taxes rather than being in addition to

I imagine that it would replace/reduce some of the other taxes so the government would get the same amount of money.

it encourages companies to do things in-house rather than outsourcing or partnering, since inside-company "transactions" aren't real money and aren't taxed

But normal taxes have the same effect, don't they?

2Dagon
It's not just the "bad" HFT.  It's any very-low-margin activity. Nope, normal taxes scale with profit, not with transaction size.  

I came up with the decision theory problem. It has the same moral as xor-blackmail, but I think it's much easier to understand:

Omega has chosen you for an experiment:

  1. First, Omega predicts your choice in a potential future offer.
  2. Omega rolls a die. Omega doesn't show you the result.
  3. If Omega predicted you would choose $200, they will only make you an offer if the die shows 6.
  4. If Omega predicted you would choose $100, they will make you an offer if the die shows any number except 1.
  5. Omega's offer, if made, is simple: "Would you like $100 or $200?"

You received an... (read more)

1notfnofn
This was easier for me to understand (but everything is easier to understand it when you see it a second time, phrased in a different way).

First (possibly dumb) thought: could it be compensated by printing fewer large bills? Again, poor people would not care, but big business transactions with cash would become less convenient.

2Shankar Sivarajan
I don't understand the problem you're trying to solve.  If you just like the aesthetic of cash transactions and want to see more of them, you could just mandate all brick-and-mortar retail stores only accept cash. If you want to save people the hassle of doing tax paperwork, and offload that to banks, that's also easy: just mandate that banks offer for free the service of filing taxes for all their customers. If you have accounts with multiple banks, they can coördinate. If you want to stop high-frequency trading, just ban it.

Wow, really? I guess it's American thing. I think I know only one person with the credit card. And she only uses it up to the interest-free limit to "farm" her reputation with the bank in case she really needs a loan, so she doesn't actually pay the fee.

2Archimedes
The customer doesn't pay the fee directly. The vendor pays the fee (and passes the cost to the customer via price). Sometimes vendors offer a cash discount because of this fee.

Epistemic state: thoughts off the top of my head, not the economist at all, talked with Claude about it

Why is there almost nowhere a small (something like 1%) universal tax on digital money transfers? It looks like a good idea to me:

  • it's very predictable
  • no one except banks has to do any paperwork
  • it's kinda progressive, if you are poor you can use cash

I see probable negative effects... but doesn't VAT and individial income tax just already have the same effects, so if this tax replace [parts of] those nothing will change much?

Also, as I understand, it would... (read more)

3Dagon
It's too much for some transactions, and too little for others.  For high-frequency (or mid-frequency) trading, 1% of the transaction is 3 or 4 times the expected value from the trade.  For high-margin sales (yachts or software), 1% doesn't bring in enough revenue to be worth bothering (this probably doesn't matter unless the transaction tax REPLACES other taxes rather than being in addition to).   It also interferes with business organization - it encourages companies to do things in-house rather than outsourcing or partnering, since inside-company "transactions" aren't real money and aren't taxed. It's not a bad idea per se, it just needs as many adjustments and carveouts as any other tax, so it ends up as politically complicated as any other tax and doesn't actually help with anything.
2Archimedes
It already happens indirectly. Most digital money transfers are things like credit card transactions. For these, the credit card company takes a percentage fee and pays the government tax on its profit.
3Shankar Sivarajan
One consideration is the government wouldn't want to encourage (harder-to-tax) cash transactions.
Answer by Tapatakt10

Only one mention of Jules Verne in answers seems weird to me.

First and foremost, "The Mysterious Island". (But maybe it has already been read at nine?)

I guess the big problem for someone who tries to do it not in small form is that while you write the story it is already getting old. There are writers who can write a novel in a season, but not many. At least if we talk about good writers. Hm-m-m, did rationalists try to hire Stephen King? :)

4L Rudolf L
Do the stories get old? If it's trying to be about near-future AI, maybe the state-of-the-art will just obsolete it. But that won't make it bad necessarily, and there are many other settings than 2026. If it's about radical futures with Dyson spheres or whatever, that seems like at least a 2030s thing, and you can easily write a novel before then. Also, I think it is actually possible to write pretty fast. 2k/day is doable, which gets you a good length novel in 50 days; even x3 for ideation beforehand and revising after the first draft only gets you to 150 days. You'd have to be good at fiction beforehand, and have existing concepts to draw on in your head though
Tapatakt1714

I often think something like "It's a shame there's so little modern science fiction that takes AI developments into account". Thanks for doing something in this niche, even if in such small form!

8L Rudolf L
Agreed! Transformative AI is hard to visualise, and concrete stories / scenarios feel very lacking (in both disasters and positive visions, but especially in positive visions). I like when people try to do this - for example, Richard Ngo has a bunch here, and Daniel Kokotajlo has his near-prophetic scenario here. I've previously tried to do it here (going out with a whimper leading to Bostrom's "disneyland without children" is one of the most poetic disasters imaginable - great setting for a story), and have a bunch more ideas I hope to get to. But overall: the LessWrong bubble has a high emphasis on radical AI futures, and an enormous amount of fiction in its canon (HPMOR, Unsong, Planecrash). I keep being surprised that so few people combine those things.

I always understood it as "not pull -> trolley does not turn; pull -> trolley turns". It definitely works like this on the original picture.

I really like it! One remark, though: two upper tracks must be swapped, otherwise it's possible to precommit by staying in place and not running to the lever.

1Warty
wait is the lever position meaningful like that? I used lever direction = where the train go, cause it seemed intuitive.

The point is Omega would not send it to you it if it was false and Omega would always send it to you if it was true.

1notfnofn
Oh I missed the quotations; you're right

O-ops, you're absolutely right, I accidentally missed "not" when I was rewriting the text. Fixed now. Thank you!

Death in Damascus is easy, but boring.

Doomsday Argument is not a decision theory problem... but it can be turned into one... I think the trolley version would be too big and complicated, though.

Obviously, only problems with discrete choice can be expressed as a Trolley problems.

Tapatakt1-1

One could claim that "the true spirit of friendship" is loving someone unconditionally or something, and that might be simple, but I don't think that's what humans actually implement.

Yeah, I agree that humans implement something more complex. But it is what we want AI to implement, isn't it? And it looks like may be quite natural abstraction to have.

(But again, it's useless while we don't know how to direct AI to the specific abstraction.)

2Dweomite
Then we're no longer talking about "the way humans care about their friends", we're inventing new hypothetical algorithms that we might like our AIs to use. Humans no longer provide an example of how that behavior could arise naturally in an evolved organism, nor a case study of how it works out for people to behave that way.

I hope "the way humans care about their friends" is another natural abstraction, something like "my utility function includes link to yours utility function". But we still don't know how to direct AI to the specific abstraction, so it's not a big hope.

2Dweomite
My model is that friendship is one particular strategy for alliance-formation that happened to evolve in humans. I expect this is natural in the sense of being a local optimum (in the ancestral environment), but probably not in the sense of being simple to formally define or implement. I think friendship is substantially more complicated than "I care some about your utility function". For instance, you probably stop valuing their utility function if they betray you (friendship can "break"). I also think the friendship algorithm includes a bunch of signalling to help with coordination (so that you understand the other person is trying to be friends), and some less-pleasant stuff like evaluations of how valuable an ally the other person is and how the friendship will affect your social standing. Friendship also appears to include some sort of check that the other person is making friendship-related-decisions using system 1 instead of system 2--possibly as a security feature to make it harder for people to consciously exploit (with the unfortunate side-effect that we penalize system-2-thinkers even when they sincerely want to be allies), or possibly just because the signalling parts evolved for system 1 and don't generalize properly. (One could claim that "the true spirit of friendship" is loving someone unconditionally or something, and that might be simple, but I don't think that's what humans actually implement.)

I think making an opt-in link as a big red button and posting it before the rules were published caused a pre-selection of players in favor of those who would press the big red button. Which is... kinda realistic for generals, I think, but not realistic for citizens.

2rossry
Think of it as your own little lesson in irreversible consequences of appealing actions, maybe? Rather than a fully-realistic element.

I mean, if I don't want to "launch the nukes", why would I even opt-in?

Tapatakt3-4

Isn't the whole point of Petrov day kinda "thou shall not press the red button"?

1Tapatakt
I mean, if I don't want to "launch the nukes", why would I even opt-in?
Tapatakt109

I don't think "We created a platform that lets you make digital minds feel bad and in the trailer we show you that you can do it, but we are in no way morally responsible if you will actually do it" is a defensible position. Anyway, they don't use this argument, only one about digital substrate.

6Richard_Kennaway
The trailer is designed to draw prospective players' attention to the issue, no more than that. If you "don't think current models are sentient", and hence are not actually feeling bad, then I don't see a reason for having a problem here, in the current state of the game. If they manage to produce this game and keep upgrading it with the latest AI methods, when will you know if there is a problem? I do not have an answer to that question.
Tapatakt206

The Seventh Sally or How Trurl's Own Perfection Led to No Good

Thanks to IC Rainbow and Taisia Sharapova who brought this matter in MiriY Telegram chat.

What. The. Hell.

In their logo they have:

They Think. They Feel. They're Alive

And the title of the video on the same page is:

AI People Alpha Launch: AI NPCs Beg Players Not to Turn Off the Game

And in the FAQ they wrote:

The NPCs in AI People are indeed advanced and designed to emulate thinking, feeling, a sense of aliveness, and even reactions that might resemble pain. However, it's essential to understand that

... (read more)
3green_leaf
Ideally, AI characters would get rights as soon as they could pass the Turing test. In the actual reality, we all know how well that will go.
9Richard_Kennaway
I don't think the trailer is saying that. It's just showing people examples of what you can do, and what the NPCs can do. Then it's up to the player to decide how to treat the NPCs. AIpeople is creating the platform. The users will decide whether to make Torment Nexi. At the end of the trailer, the NPCs are conspiring to escape the simulation. I wonder how that is going to be implemented in game terms. I notice that there also exists a cryptocoin called AIPEOPLE, and a Russian startup based in Cyprus with the domain aipeople dot ru. I do not know if these have anything to do with the AIpeople game. The game itself is made by Keen Software House. They are based in Prague together with their sister company GoodAI.

About possible backlashes from unsuccesfull communication.

I hoped for some examples like "anti-war movies have unintentionally boosted military recruitment", which is the only example I remembered myself.

Asked the same question to Claude, it gave me this examples:

Scared Straight programs: These programs, designed to deter juvenile delinquency by exposing at-risk youth to prison life, have been shown to actually increase criminal behavior in participants.

The "Just Say No" anti-drug campaign: While well-intentioned, some research suggests this oversimplified

... (read more)
2Viliam
A similar concern is that maybe the thing is so rare that previously most people didn't even think about it. But now that you reminded them of that, a certain fraction is going to try it for some weird reason. Infohazard: Similarly, teaching people political correctness can backfire (arguably, from the perspective of the person who makes money by giving political correctness trainings, this is a feature rather than a bug, because it creates a greater demand for their services in future). Like, if you have a workplace with diverse people who are naturally nice to each other, lecturing them about racism/sexism/whatever may upset the existing balance, because suddenly the minorities may get suspicious about possible microaggressions, and the majority will feel uncomfortable in their presence because they will feel like they have to be super careful about every word they say. Which can ironically lead to undesired consequences, when e.g. the white men will stop hanging out with women or black people, because they will feel like they can talk freely (e.g. make jokes) only in their absence. How does this apply to AI safety? If you say "if you do X, you might destroy humanity", in theory someone is guaranteed to do X or something similar to X, either because they think it is "edgy", or because they want to prove you wrong. But in practice, most people don't actually have an opportunity to do X.
Tapatakt160

I want to create a new content about AI Safety for Russian speakers. I was warned about possible backlash if I do it wrong.

What are the actual examples when bad oversimplified communication harmed the case it agitated for? Whose mistakes can I learn from?

4Viliam
I think if the English original is considered good, there should be nothing wrong with a translation. So make sure you translate good texts. (If you are writing your own text, write English version first and ask for feedback.) Also, get ready for disappointment if it turns out that the overlap between "can meaningfully debate AI safety" and "has problems reading English" turns out to be very small, possibly zero. To give you a similar example, I have translated the LW Sequences to Slovak language, some people shared it on social networks, and the ultimate result was... nothing. The handful of Slovak people who came to at least one LW meetup all found the rationalist community on internet, and didn't read my translation. This is not an argument against translating per se. I had much greater success at localizing software. It's just, when the target audience is very smart people, then... smart people usually know they should learn English. (A possible exception could be writing for smart kids.)
4Canaletto
Not to be dissuading, but probably a lot of people who can do relevant work know English pretty well anyway? Speaking from experience, I guess, most students knew English well enough and consumed English content when i was in university. Especially the most productive ones. So, this still can be interesting project, but not like, very important and/or worth your time.

My opinion, very briefly:

Good stuff:

  • Deception plotline
  • Demonstration of LDT in action
  • A lot of thought processes of smart characters
  • Convincing depictions of how people with very weird and evil ideology can have at least seemingly consistent worldview, be humans and not be completely insane.

Stuff that might be good for some and bad for others:

  • It's Yudkowsky's work and it feels. Some people like the style of his texts, some don't.
  • Sex scenes (not very erotic and mostly talking)
  • Re-construction of Pathfinder game mechanics in setting
  • Math classes (not the best pos
... (read more)
2Said Achmiz
(Done poorly)

Well, it's quite good random crime procedural with very likable main characters, but yes, in the first season AI plotline is very slow until last 2 episodes. And then it's slow again for the most part.

Tapatakt162

Did anyone try something like this?

  1. Create a conlang with very simple grammar and small vocabulary (not like tokipona small, more like xkcd-thing-explainer small).
  2. Use LLMs to translate a lot of texts into this conlang.
  3. Train new LLM on this translations.
  4. Try to research interpretability on this LLM.
4Viliam
There is a Simple English Wikipedia with over 200 000 articles, which is not exactly what you want, but seems to be a thing that already exists and is somewhat in that direction.
2Nathan Helm-Burger
I agree that this sounds interesting and that I haven't heard of anyone doing this yet. I have heard of some interpretability experiments with TinyStories, as Zac mentioned. I think the more interesting thing would be a dataset focused on being enriched with synthetic data showing inherently logical things like deductive symbolic logic and math problems worked out (correctly!) step-by-step. You could have a dataset of this, plus the simplified-language versions of middle school through undergrad science textbooks. I expect the result would likely be more logical, and cohesive. It would be interesting to see if this made the model fundamentally more interpretable.
4Zac Hatfield-Dodds
I don't recall any interpretability experiments with TinyStories offhand, but I'd be surprised if there aren't any.
Tapatakt160

A random thought on how to explain instrumental convergence: 

You can teach someone the basics of, say, Sid Meier's Civilization V for a quite long time without explaining what the victory conditions are. There are many possible alternative victory conditions that would not change the development strategies much.

If consciousness arises from matter, then for a stream of consciousness to exist, at the very least, the same atoms should be temporarily involved in the flowing of the river

Why? I see no problem with the consciousness that constantly changes what atoms it is built on.

This modification to the river seems to suggest that there is no such thing as a "stream of consciousness," but rather only "moments of consciousness" that have the illusion of being a stream because they can recall memories of previous moments.

Well, OK? Doesn't seem weird to me.

1blallo
yes, those are computationalists views. Computationalism is pretty much self consistent since it says that any materialized computation can be conscious, and very similar to illusionism.

Gretta Duleba is MIRI's Communication Manager. I think she is the person you should ask who write to.

Everyone who is trying to create GAI is trying to create aligned GAI. But they think it will be easy (in the sense "not very super hard so they will probably fail and create misaligned one"), otherwise they wouldn't try in the first place. So, I think, you should not share your info with them.

1Crazy philosopher
I understand. My question is, can I publish an article about this so that only MIRI guys can read it, or send in Eliezer e-mail, or something.

Most Importantly Missing (that I know of and would defend as objective, starting with the best three comedies then no order): Community, The Good Place, Coupling (UK) (if that counts), Watchmen (if we are allowing Chernobyl this is the best miniseries I know), Ally McBeal, Angel and Buffy the Vampire Slayer (no, seriously, a recent rewatch confirms), Gilmore Girls, Roseanne, Star Trek: DS9 (I see the counterarguments but they’re wrong), How I Met Your Mother.

Also... no Firefly? Really? And no Person of Interest.

I would add Balatro (especially endless mode) to the list

Answer by Tapatakt10

The paperclip maximizer oversimplifies AI motivations

Being very simple example kinda is the point? 

and neglects the potential for emergent ethics in advanced AI systems.

The emergent ethics doesn't change anything for us if it's not human-aligned ethics.

The doomer narrative often overlooks the possibility of collaborative human-AI relationships and the potential for AI to develop values aligned with human interests.

This is very vague. What possibilities do you talk about exactly?

Current AI safety research and development practices are more nuanced and

... (read more)

Wow! Thank you for the list!

I noticed you write a lot of quite high-effort comments with a lot of links to other discussions of a topic. Do you "just" devote a lot of time and efforts to it or do you, for example, apply some creative use of LLMs?

3[anonymous]
I write everything myself from scratch. I don't really use LLMs that much beyond coding assistance, where asking GPT-4o for help is often much faster than reading the documentation of a function or module I'm unfamiliar with. The comment above actually only took ~ 1 hour to write. Of course, that's mostly because all of the ideas behind it (including the links and what I thought about each one) had been ruminating in my head for long enough [1] that I already knew everything I wanted to write; now I just had to write it (and type in all the links). 1. ^ Mostly in the back of my mind as I was doing other things.

I would add "and the kind of content you want to get from aligned AGI definitely is fabricated on the Internet today". So the powerful LM trying to predict it will predict how the fabrication would look like.

It turned out I was just unlucky

Tapatakt*60

Heh, it seems it doesn't work right now

UPD: Nope, I just was unlucky, it worked after enough tries.

[This comment is no longer endorsed by its author]Reply1
6peterbarnett
I tried Ryan's prompt and got the string after 3 re-rolls. I used niplav's prompts and got it on the first try. I don't think Anthropic has trained this away (and think it would be super bad if they did).

Wait really? That's super bad. I sure hope Anthropic isn't reading this and then fine-tuning or otherwise patching their model to hide the fact that they trained on the canary string...

I just tried it (with a minor jailbreak) and it worked though.

In this equation, V(A & O) represents the value to the agent of the combination of an act and an outcome. So this is the utility that the agent will receive if they carry out a certain act and a certain outcome occurs. Further, PrAO represents the probability of each outcome occurring on the supposition that the agent carries out a certain act. It is in terms of this probability that CDT and EDT differ. EDT uses the conditional probability, Pr(O|A), while CDT uses the probability of subjunctive conditionals, Pr(A O).

Please, don't use the same letters ... (read more)

In this case, even if an extremely low value is set for L, it seems that paying this amount to play the game is unreasonable. After all, as Peterson notes, about nine times out of ten an agent that plays this game will win no more than 8 · 10-100 utility.

It seems there's an error here. Should it be "In this case, even if an extremely high value is set for L, it seems that paying a lot to play the game is unreasonable."?

Answer by Tapatakt73

No, it isn't.

the mathematical structure itself is conscious

If I understand it correctly, that's the position of, e.g. Max Tegmark (more generally, he thinks that "to exist" equals "to have a corresponding math structure").

So there should be some sort of hardware-dependence to obtain subjective experience

My (and, I think, a lot of other people's) intuition says something like "there is no hardware-dependence, but the process of computation must exist somewhere".

1notfnofn
Would your intuition suggest that a computation by hand produces the same kind of experience as your brain? Your intuition reminds me of the strange mathematical philosophy of ultrafinitism, where even mathematical statements that require a finite amount of computation to verify do not have a truth value until they are computed.

The second problem with using the law of large numbers to justify EUM has to do with a mathematical theorem known as gambler's ruin. Imagine that you and I flip a fair coin, and I pay you $1 every time it comes up heads and you pay me $1 every time it comes up tails. We both start with $100. If we flip the coin enough times, one of us will face a situation in which the sequence of heads or tails is longer than we can afford. If a long-enough sequence of heads comes up, I'll run out of $1 bills with which to pay you. If a long-enough sequence of tails comes

... (read more)

We can also construct more specific variants of 5 where FDT loses - such as environments where the message at step B is from an anti-Omega which punishes FDT like agents.

Sudden thought half a year later:

But what if we restrict reasoning to non-embedded agents? So Omegas of all kind have access to a perfect Oracle who can predict what you will do, but can't actually read yout thoughts and know that you will do it because you use FDT. I doubt that it is possible in this case to construct a similar anti-FDT situation.

Counterargument: If AGI understands FDT/UDT/LDT it can allow us to shut it down so the progress will not be slowed, some later AGI will kill us and realise some part of the first AGI's utility as a gratitude.

There is a meaningful difference between the programming skills that you typically need to be effective at your job and the skills that will let you get a job. I’m sympathetic to the view that the job search is inefficient / unfair and that it doesn’t really test you on the skills that you actually use day to day. It’s still unlikely that things like LeetCode are going to go away. A core argument in their favor is that there’s highly asymmetric information between the interviewer and interviewee and that the interviewee has to credibly signal&nbs

... (read more)
2Brendan Long
Yes, but testing that people have memorized the appropriate number of LeetCode questions is much easier than testing that they can write something big without going mad :(
4gw
I agree! This is mostly focused on the "getting a job" part though, which typically doesn't end up testing those other things you mention. I think this is the thing I'm gesturing at when I say that there are valid reasons to think that the software interview process feels like it's missing important details.
Load More