"So I found this counterexample, and saw that my attempted disproof was false, along with my dreams of fame and glory."
I know how that feels. When I was 14 or so, I took a course on cryptography, and the textbook proclaimed that modular inverses were the basis of public-key algorithms like RSA. I felt that modular inverses were crackable, and I plodded along on the problem for a few weeks, until I finally discovered a polynomial-time algorithms for doing modular inverses. It turned out that I had reinvented Euclid's algorithm, and the textbook authors were idiots.
Phil: Build the set from the used exponents of the powers of two. For instance, 1101[2] = 20 + 22 + 2**3
So I found this counterexample, and saw that my attempted disproof was false, along with my dreams of fame and glory.
Feels familiar - when I was younger, I proved the Poincaré conjecture, and Fermat's last theorem (twice). I generally managed to slay my proofs by myself, though I felt not regret at being wrong, just frustration and anger at myself.
Even now, as a mathematical researcher, it's very hard to give up a nice result that can't be proved. But I manage. And I do feel that there is a silver lining: greater, more confident accuracy.
After the fact you could see you made a mistake. But the key question is: what were the clearest signals at the time, the sort of signals you had a chance to notice and recognize? What is the warning to others? Presumably it is not to give up after your first failure.
But the key question is: what were the clearest signals at the time, the sort of signals you had a chance to notice and recognize?
In my case, it was the fact that brilliant mathematicians had tried to prove these results for generations. No matter how brilliant I think myself, it would be unlikely for me to have found a simple proof where everyone else had failed.
Minor quibble: since binary 0.1111... is 1, you need a number like 0.1010101... to get an actual counterexample.
It seems like one of the key factors in your story, Eliezer, is that you had read that book on math cranks. You were able to make the leap from your project of disproving Cantor and see its implications for the rest of your life thanks in part to having the example of the math crank in your mind.
Seeking evidence outside the immediate domain of inquiry can be tricky because it might lead one to include evidence that has no bearing on the actual problem, but because human endeavors don't happen in a vacuum, it's a great way of checking yourself for more general errors (like tilting at windmills).
I was not displaying ... any ... virtue
Most math teachers would be delighted if a student was able to understand Cantor's proof, think critically enough to search for a counter-proof, think creatively enough to describe a counter-proof (and based on different mathematical constructs at that), even though the proof was wrong at some critical steps.
This would be quite an achievement even for those who do not go on to the crucial last step of thinking self-critically enough to find the mistake in that "proof."
Minor quibble: since binary 0.1111... is 1, you need a number like 0.1010101... to get an actual counterexample.
Afaict, the original post doesn't contain any mention of binary fractions. An infinite binary sequence consisting entirely of ones doesn't represent any finite integer.
It seems to be a common childhood experience on this list to have tried to disprove famous mathematical theorems.
Me, I tried to disprove the four-color map conjecture when I was 10 or 11. At that point it was a conjecture, not a theorem. I came up with a nice moderate size map that, after a apparently free initial labelling and a sequence of apparently forced moves, required a fifth color.
Fortunately the first thing that occured to me was to double-check my result, and of course I found a 4-color coloring.
I expect that many people who grew up to be scientists and mathematicians attempted to create famous proofs when they were young, but I also expect that for many engineers such as myself our youthful folly went more along the direction of perpetual motion machines. I'd actually like to see some research on what the correlations really are.
LOL. Color me for both, Andrew. Perpetual motion using magnetic levitation in a vacuum at 10. Attempting to come up with a simple proof of Fermat's Theorem at 20 (if there was an easy way to determine n-roots of non-primes, I'd have been SET! :-) )
Something seems out of kilter about this, Eliezer.
When I was 13, I thought I had a proof in principle that there must be a minimum possible distance-- because to move is to move a finite distance, but no sum of infitesimal distances can compose a finite distance. I shared my idea with a professional physicist, who dismissed my idea using an appeal to authority. I don't care how fabulous the authority was, nor how ignorant I may have been, it was a terrible thing to for him to do that. It killed my enthusiasm for questioning physics, or math, at the time.
Re...
"I challenge the "rules" set out by whomever thinks he's the know-all on what can be done with a compass and straight edge."
I would be interested to see what you can get out of a compass and straightedge if you change the allowable operations. You could wind up with something much more complex than the things the ancient Greeks studied (think of how much more complex a Riemannian manifold is than a Euclidean n-space, once you remove a few of Euclid's axioms).
I know this is an old comment, but the answer is actually quite nice.
What the compass and straight-edge basically give you is the capacity for solving quadratic equations. There's a field of numbers between the rational and real numbers called the Constructible numbers that completely characterizes what can be done there.
Alternative techniques (e.g., folding) can allow one to solve cubic equations, and so the field of numbers that can be constructed in this way is an extension of the Constructible numbers.
So the full answer to "what you can get if you change the allowable operations" is that construction techniques correspond to field extensions of the rational numbers, and this characterizes their expressive power.
The ancient Greeks themselves played around with the rules. Archimedes used a "marked straightedge" to trisect an angle.
The first hit on google for trisect an angle is about ways to do it, not discussions of impossibility.
It seems to me that unless Eliezer was unusual in some other important way not described, he was not at close risk of becoming a math crank.
I'm also getting that impression, Robin. I'd say, "But there may be a selection effect in the people who comment at Overcoming Bias", but perhaps that would be, well, clinging.
This of course begs the question of where math cranks do come from.
While many people have mentioned similar disappointments, no one has echoed "I'll get that theorem eventually...even though my first try failed!" That was what seemed like a really bad sign when I read the essay before the comments. But I think we're really bad at communicating feelings, so I don't know how the feelings relate, how strong they were, and especially, how the commenters see the parallels with their reactions.
While many people have mentioned similar disappointments, no one has echoed "I'll get that theorem eventually...even though my first try failed!" That was what seemed like a really bad sign when I read the essay before the comments.
I think it's worse than that. Many people mentioned that they have tried to solve open conjectures, which is something that would require exceptional intelligence, expecially without many years of experience. But if you are a smart teenager, thinking that you are exceptionally intelligent falls in the range of normal juvenile hubris.
Yudkowsky didn't try to solve an open conjecture. He tried to disprove a theorem. A theorem that was proved one hundred years ago, and has been known by pretty much everybody who had a math education since then. Thus, Yudkowsky didn't just think he was exceptionally intelligent, he thought that everyone else was basically an idiot.
That's actually a bad symptom of crackpot thought patterns, IMHO.
This argument that one should admit when they're wrong doesn't generalize beyond the exact reasoning of mathematical proofs and the like. In probablistic reasoning one can be, indeed usually is, wrong but close. The whole Bayesian worldview is predicated on the assumption that being a little bit wrong, or less wrong than the next guy, means you are probably on a more correct track towards the truth. But it doesn't and can't prove that, given just a few more important bits of information, the guy who's currently "more wrong" is right after all...
Would any regular commenters/readers object if I deleted comments like those from "a woo just like you"? I've always been nervous around censorship, especially where it carries the appearance of conflict of interest, but lack of censorship also carries its penalties. If I don't get any requests not to do so, I'll delete the comment tomorrow.
As I'm not much of a contributor, you can take my suggestion with a grain of salt but: Why not file away all deleted non-spam comments to a place where they can be read, but are out of the way? That way, moderators don't have to worry so much about censoring people and can instead focus on keeping discussions civil/troll-free.
Do you remember the title of the book? It sounds interesting, speaking as a lapsed mathematician.
Not sure if this is cranky or not, but when I was youthful I noticed that the Lorentz transformation of space-time due relativistic effects, square root of one minuc v squared over c squared, implies an imaginary solution for an v greater than c, that is for traveling faster than the speed of light. Now, most sci fi stories suggest that one would go backwards in time if one exceeded the speed of light, but I deduced that one would go into a second time dimension.
Of course the problem is that as long as Einstein is right, it is simply impossible to exceed the speed of light, thereby making the entire speculation irrelevant.
I don't like the formulation: "A thought you should never have thought."
I'd prefer, "An idea you should have quickly rejected."
I suspect that many genuine innovations might first appear to be mistakes or unwarranted challenges to the prevailing wisdom. They should be thought. And they should be considered and criticized. But, we should be ready to reject them if they don't survive the criticism.
Don't know what your blogging software allows, but richarddawkins.net now has a separate thread for off-topic posts; you click on a label at the end of the article to get to the off-topic thread.
I love this site. Found it when looking at a piece of crackpot science on the internet and, wondering, typed "crackpot" into google. I am trying to argue with someone who's my nemesis in most every way, and I'm trying to do it honestly. I feel his vested interest in the preferred answer vastly biases his judgment & wonder what biases do I have, and how did they get there. You seem to address a key one I liken to tree roots, growing in deep and steadfast wherever you first happen to fall, whether it's good ground or not.
Not unlike that analogy, I landed here first, on your post, and found it very good ground indeed.
If you want another couple threads to start exploring, one very good starting place is What Do We Mean By Rationality? and its links; then there is the massive collection of posts accumulated in the Sequences which you can pick over for interesting nuggets. A lot of posts (and comments!) will have links back to related material, both at the top and throughout the text.
Just to make it explicit: I really appreciate your "welcome" comments, they're good for the site. Thanks.
Let me echo ciphergoth. The effect is broader than you might think; it was because of one of these sorts of comments that I (years later) found the introduction thread when I did.
Admittedly, most readers probably don't start from the beginning and work their way forward. But some of us do!
This is one of my favorite crackpot writings. It does seem plausible that held breath underwater swimming is really good exercise. http://www.winwenger.com/ebooks/guaran.htm
Since everyone is sharing their stories, here's mine. When I was around 10, a family friend introduced me to the four-color map problem. I spent months trying to draw a map that required five colors, and one time I thought I had it. I dreamed of fame and glory for a few hours, then I showed the map to a relative who colored it with four colors. Shortly after, I accepted that I wasn't going to get it and stopped.
what a terribly unfair test to visit upon a child of thirteen. That I had to be that rational, already, at that age, or fail.
I always find it odd that you seem to write as though there is no hope of redemption when one makes a mistake of this magnitude. Certainly, lifetimes can be lost to such mistakes. But then, sometimes, it only takes a week to realise our folly, neh?
I fear that I might be currently trapped in this error: I've always resented Gödel's Incompleteness Theorems. When I was about 17 I thought I'd disproved 1IT (turned out I'd just reconstructed the proof of 2IT and missed the detail that Con(T)≠ProvT(Con(T))). It took me about a year after that to realise that, no, I wasn't going to disprove the ITs no matter how much I wanted to, and I accepted that trying to disprove them anyway would be a crackpot thing to do. Since then I've been trying to construct a philosophical framework of mathematics in which the ITs become irrelevant. Have I, in fact, taken the Crackpot Offer?
I don't remember ever coming up with a false disproof in math, though I did manage to "solve" perpetual motion machines. I did successfully prove a trivial result in solving quadratic equations in modular arithmetic.
Eliezer, did you realize at the time that what you had done was construct the basic outline of the proof that 2^aleph0 = aleph1? There was an interesting gem hiding in your disproof, had you looked. Reversed stupidity is not intelligence, and all that :)
With all due respect to Eliezer, there exists an institution that can protect you from the danger described in the post. It's called "math school". Sometime in tenth grade, I came up with a proof of the continuum hypothesis that my teacher couldn't immediately overturn. We had a fun time finding the catch, then moved on to other things.
I once thought I could prove that the set of all natural numbers is as large as its power set. However, I was smart enough to acknowledge my limitations (What‘s more likely: That I made a mistake in my thinking I haven‘t yet noticed, or that a theorem pretty much any professional mathematician accepts as true is actually false?), so I activly searched for errors in my thinking. Eventually, I noticed that my methods only works for finite sub sets (The set of all natural numbers is, indeed, as large as the set of all FINITE subsets), but not for infinite subsets.
Eliziers method also works for all finite subsets, but not for infinite subsets
What the hell, I'll tack mine on.
As a kid, I was introduced to an interesting contraption (that was supposed to be a party trick kind of thing) that I then dreamed for years of using to generate free energy (and therefore to be able to fight against entropy with). It was a ramp with the middle cut out so it looks like a "V" that you have a roller for made of 2 cones base-to-base. The cones look like they roll up the ramp. This does not work, but it took me half a decade to learn this.
My other one is that I heard about the angle trisection thing and "found ...
perhaps I should apply Cantor’s Diagonal Argument to my clever construction, and of course it found a counterexample—the binary number (. . . 1111), which does not correspond to any finite whole number.
I’m not following despite having recently reviewed Cantor’s Diagonal Argument. I can imagine constructing a matrix such that the diagonal is all ones… but I don’t see how this connects up to the counterexample claim above.
Also, why worry that an infinite binary representation (of any kind) doesn’t correspond to a finite whole number? I suspect I’m missing something here. A little help please to help close this inferential distance?
When I was very young—I think thirteen or maybe fourteen—I thought I had found a disproof of Cantor's Diagonal Argument, a famous theorem which demonstrates that the real numbers outnumber the rational numbers. Ah, the dreams of fame and glory that danced in my head!
My idea was that since each whole number can be decomposed into a bag of powers of 2, it was possible to map the whole numbers onto the set of subsets of whole numbers simply by writing out the binary expansion. 13, for example, 1101, would map onto {0, 2, 3}. It took a whole week before it occurred to me that perhaps I should apply Cantor's Diagonal Argument to my clever construction, and of course it found a counterexample—the binary number ...1111, which does not correspond to any finite whole number.
So I found this counterexample, and saw that my attempted disproof was false, along with my dreams of fame and glory.
I was initially a bit disappointed.
The thought went through my mind: "I'll get that theorem eventually! Someday I'll disprove Cantor's Diagonal Argument, even though my first try failed!" I resented the theorem for being obstinately true, for depriving me of my fame and fortune, and I began to look for other disproofs.
And then I realized something. I realized that I had made a mistake, and that, now that I'd spotted my mistake, there was absolutely no reason to suspect the strength of Cantor's Diagonal Argument any more than other major theorems of mathematics.
I saw then very clearly that I was being offered the opportunity to become a math crank, and to spend the rest of my life writing angry letters in green ink to math professors. (I'd read a book once about math cranks.)
I did not wish this to be my future, so I gave a small laugh, and let it go. I waved Cantor's Diagonal Argument on with all good wishes, and I did not question it again.
And I don't remember, now, if I thought this at the time, or if I thought it afterward... but what a terribly unfair test to visit upon a child of thirteen. That I had to be that rational, already, at that age, or fail.
The smarter you are, the younger you may be, the first time you have what looks to you like a really revolutionary idea. I was lucky in that I saw the mistake myself; that it did not take another mathematician to point it out to me, and perhaps give me an outside source to blame. I was lucky in that the disproof was simple enough for me to understand. Maybe I would have recovered eventually, otherwise. I've recovered from much worse, as an adult. But if I had gone wrong that early, would I ever have developed that skill?
I wonder how many people writing angry letters in green ink were thirteen when they made that first fatal misstep. I wonder how many were promising minds before then.
I made a mistake. That was all. I was not really right, deep down; I did not win a moral victory; I was not displaying ambition or skepticism or any other wondrous virtue; it was not a reasonable error; I was not half right or even the tiniest fraction right. I thought a thought I would never have thought if I had been wiser, and that was all there ever was to it.
If I had been unable to admit this to myself, if I had reinterpreted my mistake as virtuous, if I had insisted on being at least a little right for the sake of pride, then I would not have let go. I would have gone on looking for a flaw in the Diagonal Argument. And, sooner or later, I might have found one.
Until you admit you were wrong, you cannot get on with your life; your self-image will still be bound to the old mistake.
Whenever you are tempted to hold on to a thought you would never have thought if you had been wiser, you are being offered the opportunity to become a crackpot—even if you never write any angry letters in green ink. If no one bothers to argue with you, or if you never tell anyone your idea, you may still be a crackpot. It's the clinging that defines it.
It's not true. It's not true deep down. It's not half-true or even a little true. It's nothing but a thought you should never have thought. Not every cloud has a silver lining. Human beings make mistakes, and not all of them are disguised successes. Human beings make mistakes; it happens, that's all. Say "oops", and get on with your life.