Less Wrong is a large community of very smart people with a wide spectrum of expertise, and I think relatively little of that value has been tapped.
Like my post The Best Textbooks on Every Subject, this is meant to be a community-driven post. The first goal is to identify topics the Less Wrong community would like to read more about. The second goal is to encourage Less Wrongers to write on those topics. (Respecting, of course, the implicit and fuzzy guidelines for what should be posted to Less Wrong.)
One problem is that those with expertise on a subject don't necessarily feel competent to write a front-page post on it. If that's the case, please comment here explaining that you might be able to write one of the requested posts, but you'd like a writing collaborator. We'll try to find you one.
Rules
You may either:
- Post the title of the post you want someone to write for Less Wrong. If the title itself isn't enough to specify the content, include a few sentences of explanation. "How to Learn a Language Quickly" probably needs no elaboration, but "Normative Theory and Coherent Extrapolated Volition" certainly does. Do not post two proposed post titles in the same comment, because that will confuse voting. Please put the title in bold.
or... - Vote for a post title that has already been suggested, indicating that you would like to read that post, too. Vote with karma ('Vote Up' or 'Vote Down' on the comment that contains the proposed post title).
I will regularly update the list of suggested Less Wrong posts, ranking them in descending order of votes (like this).
The List So Far (updated 02/11/11)
- (35) Conversation Strategies for Spreading Rationality Without Annoying People
- (32) Smart Drugs: Which Ones to Use for What, and Why
- (30) A Survey of Upgrade Paths for the Human Brain
- (29) Trusting Your Doctor: When and how to be skeptical about medical advice and medical consensus
- (25) Rational Homeschool Education
- (25) Field Manual: What to Do If You're Stranded in a Level 1 (Base Human Equivalent) Brain in a pre-Singularity Civilization
- (20) Entrepreneurship
- (20) Detecting And Bridging Inferential Distance For Teachers
- (19) Detecting And Bridging Inferential Distance For Learners
- (18) Teaching Utilizable Rationality Skills by Exemplifying the Application of Rationality
- (13) Open Thread: Offers of Help, Requests for Help
- (13) Open Thread: Math
- (12) How to Learn a Language Quickly
- (12) True Answers for Every Philosophical Question
- (10) The "Reductionism" Sequence in One Lesson
- (10) The "Map and Territory" Sequence in One Lesson
- (10) The "Mysterious Answers to Mysterious Questions" Sequence in One Lesson
- (10) Lecture Notes on Personal Rationality
- (10) The "Joy in the Merely Real" Sequence in One Lesson
Believe it or not, I actually started an article on this around "17 October 2009" (per the date stamp) and never finished it. (I actually had the more ambitious idea of summarizing every chapter in one article, but figured Chapter 3 would be enough.) Might as well post what I have (formatting and links don't carry over; I've corrected the worst issues) ...
Here I attempt to summarize the points laid out in Gary Drescher's Good and Real: Demystifying Paradoxes from Physics to Ethics (discussed previously on Less Wrong), chapter 3, which explores the apparent flow of time and gives a reductionist account of it. To [...] What follows is a restating of the essential points and the arguments behind them in my own words, which I hope to make faithful to the text. It's long, but a lot shorter than reading the chapter, a lot cheaper than buying the book, and a lot less subjuntively self-defeating than pirating it.
The focus of the chapter is to solve three interrelated paradoxes: If the laws of physics are time-symmetric:
1) Why does entropy increase in only one direction?
2) Why do we perceive a directional flow of time?
3) Why do we remember the past but not the future?
Starting from the first: why does entropy -- the total disorder in the universe -- increase asymmetrically? To answer, start with a simple case: the billiard ball simulation, where balls have a velocity and position and inelastically bounce off each other as per the standard equations predicated on the (time-symmetric) conservation of linear momentum. For a good example of entropy's increase, let's initialize it with a non-uniformity: there will be a few large, fast balls, and many small, slow balls.
What happens? Well, as time goes by, they bounce off each other, and the larger balls transfer their momentum to balls with less. We see the standard increase in entropy as time increases. So if you were to watch a video of the simulation in action, there would be telltale signs of which is the positive and which is the negative direction: in the positive direction, large balls would plow through groups of smaller balls, leaving a "wake" during which it increases their speeds. But if we watch it in reverse, going back to the start, entropy, of course, decreases: highly-ordered wakes spontaneously form before the large balls go into them.
Hence, the asymmetry: entropy increases in only one direction.
The mystery dissolves when you consider what happens when you continue to view the simulation backwards, and proceed through the initial time, onward to t= -1, -2, -3, ... . You see the exact same thing happen going in the direction of negative time from t=0. So, we see our confusion: entropy does not increase in just the positive direction: it increases as you move away from zero, even if that direction isn't positive.
So, we need to reframe our understanding: instead of thinking in terms of positive and negative time directions, we should think in terms of "pastward" and "futureward" directions. Pastward means in the direction of the initial state, and futureward means away from it. Both the sequences t= 1, 2, 3, ... and t= -1, -2, -3, ... go into the future. (Note the parallel here to the reframing of "up" and "down" once your model of the earth goes from flat to round: "down" no longer means a specific vector, but the vector from where you are to the center of the earth. So you change your model of "down" and "up" to "centerward" and "anticenterward" [my terms, not Drescher's], respectively.)
Okay, that gets us a correct statement of the conditions under which entropy increases, but still doesn't say why entropy increases in only the futureward direction. For that, we need to identify what the positive-time futureward direction and the negative-time futureward direction have in common. For one thing, the balls become correlated. Previously (pastwardly), knowing a ball's state did not allow you to infer much about the other balls' states, as the velocities were set independently of one another. But the accumulation of collisions causes the balls to become correlated -- in effect, to share information with each other. [Rephrase to discuss elimination of gradients/exchange of information of all parts of system?...]
Note that the entropy does not need to increase uniformly: this model still permits local islands of lower entropy in the futureward direction, as long as the total entropy still increases. Consider the "wakes" left by the large balls that were mentioned above. In that case, the large balls will "plow" right through the small balls and leave a (low entropy) wake. (Even as they do this, the large balls transfer momentum to the smaller balls and increase total entropy.) The wakes allow you to identify time's direction: a wake is always located where the large ball was in an immediately pastward state. This relationship also implies that wake contains a "record" of sorts, giving physical form to the information in the current timewise state, regarding a pastward state.
This process is similar to what goes on in the brain. Just as wakes are islands of low entropy containing information about pastward states, so too is your brain an island of low entropy containing information about pastward states. (Life forms are already known to be dissipative systems that maintain an island of low entropy at the cost of a counterbalancing increase elsewhere.) [...]
So it's not that "gee, we notice time goes forward, and we notice that entropy happens to always increase". Rather, the increase of entropy determines what we will identify as the future, since any time slice will only contain versions of ourselves with memories of pastward states.
BTW, Sean Carroll just wrote an entire popular-level book on this subject.